Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
5,960
13,397,846,787
IssuesEvent
2020-09-03 12:17:06
openpower-cores/a2i
https://api.github.com/repos/openpower-cores/a2i
closed
The order of writing the GPRs
core RTL/Architecture
Hi All, In the pipeline, the A2I core writes the data into GPRs (general purpose registers) after the ex7 stage which is named as ex8 or rf1. My current understanding of A2I is that different instructions of one thread can actually write the data into the register file out-of-order. For example, there are two instructions, i.e., "Load RegA, addr1"; and "Add RegB, RegC, RegD". If the load instruction misses in the L1 cache, it will be stored in the LMQ and waiting to access the L2 cache. In this case, if the add instruction keeps executing, it will write the data into GPR before the load instruction. If this is the case, the order of writing the register file for these two instructions, i.e.., load and add, is out of sequence. Is it the right understanding? If I understand correctly, how can A2I guarantee the precise interrupt or exception? For example, one exception happens after the add instruction writing the GPRs. Thanks!
1.0
The order of writing the GPRs - Hi All, In the pipeline, the A2I core writes the data into GPRs (general purpose registers) after the ex7 stage which is named as ex8 or rf1. My current understanding of A2I is that different instructions of one thread can actually write the data into the register file out-of-order. For example, there are two instructions, i.e., "Load RegA, addr1"; and "Add RegB, RegC, RegD". If the load instruction misses in the L1 cache, it will be stored in the LMQ and waiting to access the L2 cache. In this case, if the add instruction keeps executing, it will write the data into GPR before the load instruction. If this is the case, the order of writing the register file for these two instructions, i.e.., load and add, is out of sequence. Is it the right understanding? If I understand correctly, how can A2I guarantee the precise interrupt or exception? For example, one exception happens after the add instruction writing the GPRs. Thanks!
non_priority
the order of writing the gprs hi all in the pipeline the core writes the data into gprs general purpose registers after the stage which is named as or my current understanding of is that different instructions of one thread can actually write the data into the register file out of order for example there are two instructions i e load rega and add regb regc regd if the load instruction misses in the cache it will be stored in the lmq and waiting to access the cache in this case if the add instruction keeps executing it will write the data into gpr before the load instruction if this is the case the order of writing the register file for these two instructions i e load and add is out of sequence is it the right understanding if i understand correctly how can guarantee the precise interrupt or exception for example one exception happens after the add instruction writing the gprs thanks
0
209,374
16,191,904,646
IssuesEvent
2021-05-04 09:40:09
microsoft/microsoft-ui-xaml
https://api.github.com/repos/microsoft/microsoft-ui-xaml
closed
Can you please keep the roadmap more up to date?
documentation question winui3preview
The latest preview is not on the [roadmap](https://github.com/microsoft/microsoft-ui-xaml/blob/master/docs/roadmap.md). The "[Feature Roadmap](https://github.com/microsoft/microsoft-ui-xaml/blob/master/docs/roadmap.md#winui-30-feature-roadmap)" only has two categories left: "Reunion 0.5" (in two weeks) and "Planned for a future update" (which could not be any vaguer.) I understand not wanting to promise too much. But at this point you are promising nothing that is more than 2 weeks away from shipping. Are there going to be future previews, I guess so, I remember @ryandemopoulos speaking about roughly monthly previews at some point? Maybe the "Milestones" in GitHub could be used to track some of the progress? (Currently they are clearly [very outdated](https://github.com/microsoft/microsoft-ui-xaml/milestones).) (Btw. the project [reunion roadmap](https://github.com/microsoft/ProjectReunion/blob/main/docs/roadmap.md) contains the "unpackaged desktop apps WinUI 3" as part of the Q4 2021 milestone, but the WinUI 3 roadmap does not give and date for that)
1.0
Can you please keep the roadmap more up to date? - The latest preview is not on the [roadmap](https://github.com/microsoft/microsoft-ui-xaml/blob/master/docs/roadmap.md). The "[Feature Roadmap](https://github.com/microsoft/microsoft-ui-xaml/blob/master/docs/roadmap.md#winui-30-feature-roadmap)" only has two categories left: "Reunion 0.5" (in two weeks) and "Planned for a future update" (which could not be any vaguer.) I understand not wanting to promise too much. But at this point you are promising nothing that is more than 2 weeks away from shipping. Are there going to be future previews, I guess so, I remember @ryandemopoulos speaking about roughly monthly previews at some point? Maybe the "Milestones" in GitHub could be used to track some of the progress? (Currently they are clearly [very outdated](https://github.com/microsoft/microsoft-ui-xaml/milestones).) (Btw. the project [reunion roadmap](https://github.com/microsoft/ProjectReunion/blob/main/docs/roadmap.md) contains the "unpackaged desktop apps WinUI 3" as part of the Q4 2021 milestone, but the WinUI 3 roadmap does not give and date for that)
non_priority
can you please keep the roadmap more up to date the latest preview is not on the the only has two categories left reunion in two weeks and planned for a future update which could not be any vaguer i understand not wanting to promise too much but at this point you are promising nothing that is more than weeks away from shipping are there going to be future previews i guess so i remember ryandemopoulos speaking about roughly monthly previews at some point maybe the milestones in github could be used to track some of the progress currently they are clearly btw the project contains the unpackaged desktop apps winui as part of the milestone but the winui roadmap does not give and date for that
0
694,063
23,800,597,512
IssuesEvent
2022-09-03 08:06:23
kubernetes/ingress-nginx
https://api.github.com/repos/kubernetes/ingress-nginx
closed
Allow SSL certificate expiration warning threshold to be adjusted
kind/feature lifecycle/rotten triage/accepted needs-priority
<!-- Welcome to ingress-nginx! For a smooth feature request process, try to answer the following questions. Don't worry if they're not all applicable; just try to include what you can :-) If you need to include code snippets or logs, please put them in fenced code blocks. If they're super-long, please use the details tag like <details><summary>super-long log</summary> lots of stuff </details> --> <!-- What do you want to happen? --> Some PKI implementations like https://github.com/smallstep/certificates issue aggressively short-lived certificates by default (24 hours, for example). In situations where this is the desired/intended configuration for Ingress TLS certs, this causes a disproportionate amount of warning messages. ``` W0122 17:44:59.804433 9 controller.go:1339] SSL certificate for server "prometheus.k8s.home.arpa" is about to expire (2022-01-23 16:50:14 +0000 UTC) W0122 17:44:59.804670 9 controller.go:1339] SSL certificate for server "grafana.k8s.home.arpa" is about to expire (2022-01-23 16:50:15 +0000 UTC) W0122 17:44:59.804801 9 controller.go:1339] SSL certificate for server "alertmanager.k8s.home.arpa" is about to expire (2022-01-23 16:50:15 +0000 UTC) ``` The warning threshold is currently hard-coded here: https://github.com/kubernetes/ingress-nginx/blob/abdece6e80b6d54d177cf3f51e43d1f8220c1b1c/internal/ingress/controller/controller.go#L1349 It would be useful to make this an adjustable value. <!-- Is there currently another issue associated with this? --> <!-- Does it require a particular kubernetes version? --> <!-- If this is actually about documentation, uncomment the following block --> <!-- /kind documentation /remove-kind feature -->
1.0
Allow SSL certificate expiration warning threshold to be adjusted - <!-- Welcome to ingress-nginx! For a smooth feature request process, try to answer the following questions. Don't worry if they're not all applicable; just try to include what you can :-) If you need to include code snippets or logs, please put them in fenced code blocks. If they're super-long, please use the details tag like <details><summary>super-long log</summary> lots of stuff </details> --> <!-- What do you want to happen? --> Some PKI implementations like https://github.com/smallstep/certificates issue aggressively short-lived certificates by default (24 hours, for example). In situations where this is the desired/intended configuration for Ingress TLS certs, this causes a disproportionate amount of warning messages. ``` W0122 17:44:59.804433 9 controller.go:1339] SSL certificate for server "prometheus.k8s.home.arpa" is about to expire (2022-01-23 16:50:14 +0000 UTC) W0122 17:44:59.804670 9 controller.go:1339] SSL certificate for server "grafana.k8s.home.arpa" is about to expire (2022-01-23 16:50:15 +0000 UTC) W0122 17:44:59.804801 9 controller.go:1339] SSL certificate for server "alertmanager.k8s.home.arpa" is about to expire (2022-01-23 16:50:15 +0000 UTC) ``` The warning threshold is currently hard-coded here: https://github.com/kubernetes/ingress-nginx/blob/abdece6e80b6d54d177cf3f51e43d1f8220c1b1c/internal/ingress/controller/controller.go#L1349 It would be useful to make this an adjustable value. <!-- Is there currently another issue associated with this? --> <!-- Does it require a particular kubernetes version? --> <!-- If this is actually about documentation, uncomment the following block --> <!-- /kind documentation /remove-kind feature -->
priority
allow ssl certificate expiration warning threshold to be adjusted welcome to ingress nginx for a smooth feature request process try to answer the following questions don t worry if they re not all applicable just try to include what you can if you need to include code snippets or logs please put them in fenced code blocks if they re super long please use the details tag like super long log lots of stuff some pki implementations like issue aggressively short lived certificates by default hours for example in situations where this is the desired intended configuration for ingress tls certs this causes a disproportionate amount of warning messages controller go ssl certificate for server prometheus home arpa is about to expire utc controller go ssl certificate for server grafana home arpa is about to expire utc controller go ssl certificate for server alertmanager home arpa is about to expire utc the warning threshold is currently hard coded here it would be useful to make this an adjustable value kind documentation remove kind feature
1
192,207
6,847,630,952
IssuesEvent
2017-11-13 15:58:24
CS2103AUG2017-W15-B2/main
https://api.github.com/repos/CS2103AUG2017-W15-B2/main
closed
Command Prediction for CommandBox
new feature Priority: HIGH
## TODO - [x] Implement basic functionality - [x] Rename components to `CommandPrediction` for clarity (instead of `SearchPrediction` or `TextPrediction` - [x] Add tests - [x] Add more cases to support new commands that will be introduced up till V1.5 - [x] Update UG and DG ### Motivation The command box is going to be the main way of navigating through CYNC. It'd be great if we can make the text bar more powerful with text prediction / autocomplete. ### Description I'ma do this in a few parts. First part will only be involved with searching through commands, Next part will be involved with making the find command more powerful I'll try my best to touch only the UI component. ### UI Mockup ![image](https://user-images.githubusercontent.com/18271085/31314595-cf8395aa-ac36-11e7-9041-c440f3615049.png) ### Class Diagram ![image](https://user-images.githubusercontent.com/18271085/31314543-be81aeaa-ac35-11e7-8439-9fcbd17896dd.png)
1.0
Command Prediction for CommandBox - ## TODO - [x] Implement basic functionality - [x] Rename components to `CommandPrediction` for clarity (instead of `SearchPrediction` or `TextPrediction` - [x] Add tests - [x] Add more cases to support new commands that will be introduced up till V1.5 - [x] Update UG and DG ### Motivation The command box is going to be the main way of navigating through CYNC. It'd be great if we can make the text bar more powerful with text prediction / autocomplete. ### Description I'ma do this in a few parts. First part will only be involved with searching through commands, Next part will be involved with making the find command more powerful I'll try my best to touch only the UI component. ### UI Mockup ![image](https://user-images.githubusercontent.com/18271085/31314595-cf8395aa-ac36-11e7-9041-c440f3615049.png) ### Class Diagram ![image](https://user-images.githubusercontent.com/18271085/31314543-be81aeaa-ac35-11e7-8439-9fcbd17896dd.png)
priority
command prediction for commandbox todo implement basic functionality rename components to commandprediction for clarity instead of searchprediction or textprediction add tests add more cases to support new commands that will be introduced up till update ug and dg motivation the command box is going to be the main way of navigating through cync it d be great if we can make the text bar more powerful with text prediction autocomplete description i ma do this in a few parts first part will only be involved with searching through commands next part will be involved with making the find command more powerful i ll try my best to touch only the ui component ui mockup class diagram
1
291,414
8,924,955,792
IssuesEvent
2019-01-21 20:40:14
alakajam-team/alakajam
https://api.github.com/repos/alakajam-team/alakajam
closed
PSQL errors are no longer caught
bug high priority
I can't seem to be able to catch PSQL errors anymore, to display 500s in a pretty page. I've had the same issue with multer as well, but I've been able to [wrap multer](https://github.com/alakajam-team/alakajam/blob/master/controllers/index.js#L170) to force catching any errors. I feel something's wrong with our integration of `express-promise-router` but I can't put my finger on it. Back to jamming for now...
1.0
PSQL errors are no longer caught - I can't seem to be able to catch PSQL errors anymore, to display 500s in a pretty page. I've had the same issue with multer as well, but I've been able to [wrap multer](https://github.com/alakajam-team/alakajam/blob/master/controllers/index.js#L170) to force catching any errors. I feel something's wrong with our integration of `express-promise-router` but I can't put my finger on it. Back to jamming for now...
priority
psql errors are no longer caught i can t seem to be able to catch psql errors anymore to display in a pretty page i ve had the same issue with multer as well but i ve been able to to force catching any errors i feel something s wrong with our integration of express promise router but i can t put my finger on it back to jamming for now
1
19,001
2,616,017,148
IssuesEvent
2015-03-02 00:59:11
jasonhall/bwapi
https://api.github.com/repos/jasonhall/bwapi
closed
Get Order Type from Tech Type
auto-migrated Component-Logic Component-Persistence Priority-High Type-Enhancement
``` This is mostly for internal use so we don't have huge duplicate switch statements just to map orders to techs. ``` Original issue reported on code.google.com by `AHeinerm` on 9 Jun 2011 at 7:03
1.0
Get Order Type from Tech Type - ``` This is mostly for internal use so we don't have huge duplicate switch statements just to map orders to techs. ``` Original issue reported on code.google.com by `AHeinerm` on 9 Jun 2011 at 7:03
priority
get order type from tech type this is mostly for internal use so we don t have huge duplicate switch statements just to map orders to techs original issue reported on code google com by aheinerm on jun at
1
115,135
11,866,018,992
IssuesEvent
2020-03-26 02:19:59
laurenriddle/Hive-Mind-Back-End-Capstone-Client
https://api.github.com/repos/laurenriddle/Hive-Mind-Back-End-Capstone-Client
closed
Update README.md for Client
documentation
Add ERD, Wireframes, Technologies, Pictures of App with instructions
1.0
Update README.md for Client - Add ERD, Wireframes, Technologies, Pictures of App with instructions
non_priority
update readme md for client add erd wireframes technologies pictures of app with instructions
0
629,763
20,052,333,974
IssuesEvent
2022-02-03 08:20:12
Climate-Refugee-Stories/crs-website
https://api.github.com/repos/Climate-Refugee-Stories/crs-website
opened
Netlify: swap account for CMS
low priority
change account to Main owner's account for netlify integration and GH Oauth app.
1.0
Netlify: swap account for CMS - change account to Main owner's account for netlify integration and GH Oauth app.
priority
netlify swap account for cms change account to main owner s account for netlify integration and gh oauth app
1
461,066
13,222,993,917
IssuesEvent
2020-08-17 16:26:18
magento/magento2
https://api.github.com/repos/magento/magento2
opened
[Issue] Convert MSRP currency of configurable product options
Component: ConfigurableProduct Priority: P2 Severity: S2
This issue is automatically created based on existing pull request: magento/magento2#27446: Convert MSRP currency of configurable product options --------- ### Description The MSRP of configurable products is updated in Javascript when an option is chosen. Javascript uses an object with prices of the options. The prices in this object are converted to the chosen currency. MSRP prices however were not converted, leading to the price being displayed in a wrong currency. ### Manual testing scenarios 1. Enable MAP (Configuration > Sales > Sales > Minimum Advertised Price > Enable MAP). 2. Enable multiple currencies (Configuration > General > Currency Setup > Allowed Currencies) 3. Create a configurable product with options that have an MSRP 4. Go to the frontend and choose a currency that is not the base currency 5. View the product on the frontend a choose an option. Expected result: The MSRP is converted to the chosen currency Actual result: The MSRP is not converted but it is labeled as the chosen currency ![Screenshot from 2020-03-26 11-13-38](https://user-images.githubusercontent.com/45412158/77610665-4eed3980-6f5e-11ea-8dbb-bb44991af8ba.png) Note that the MSRP is configured in euro but displayed in dollars with the euro amount. After applying the change the MSRP is converted: ![Screenshot from 2020-03-26 12-52-14](https://user-images.githubusercontent.com/45412158/77611623-c58b3680-6f60-11ea-9c65-b101913ff225.png) Rates in table directory_currency_rate: ![Screenshot from 2020-03-26 12-57-40](https://user-images.githubusercontent.com/45412158/77611862-6bd73c00-6f61-11ea-9bb0-7381ac332fc2.png)
1.0
[Issue] Convert MSRP currency of configurable product options - This issue is automatically created based on existing pull request: magento/magento2#27446: Convert MSRP currency of configurable product options --------- ### Description The MSRP of configurable products is updated in Javascript when an option is chosen. Javascript uses an object with prices of the options. The prices in this object are converted to the chosen currency. MSRP prices however were not converted, leading to the price being displayed in a wrong currency. ### Manual testing scenarios 1. Enable MAP (Configuration > Sales > Sales > Minimum Advertised Price > Enable MAP). 2. Enable multiple currencies (Configuration > General > Currency Setup > Allowed Currencies) 3. Create a configurable product with options that have an MSRP 4. Go to the frontend and choose a currency that is not the base currency 5. View the product on the frontend a choose an option. Expected result: The MSRP is converted to the chosen currency Actual result: The MSRP is not converted but it is labeled as the chosen currency ![Screenshot from 2020-03-26 11-13-38](https://user-images.githubusercontent.com/45412158/77610665-4eed3980-6f5e-11ea-8dbb-bb44991af8ba.png) Note that the MSRP is configured in euro but displayed in dollars with the euro amount. After applying the change the MSRP is converted: ![Screenshot from 2020-03-26 12-52-14](https://user-images.githubusercontent.com/45412158/77611623-c58b3680-6f60-11ea-9c65-b101913ff225.png) Rates in table directory_currency_rate: ![Screenshot from 2020-03-26 12-57-40](https://user-images.githubusercontent.com/45412158/77611862-6bd73c00-6f61-11ea-9bb0-7381ac332fc2.png)
priority
convert msrp currency of configurable product options this issue is automatically created based on existing pull request magento convert msrp currency of configurable product options description the msrp of configurable products is updated in javascript when an option is chosen javascript uses an object with prices of the options the prices in this object are converted to the chosen currency msrp prices however were not converted leading to the price being displayed in a wrong currency manual testing scenarios enable map configuration sales sales minimum advertised price enable map enable multiple currencies configuration general currency setup allowed currencies create a configurable product with options that have an msrp go to the frontend and choose a currency that is not the base currency view the product on the frontend a choose an option expected result the msrp is converted to the chosen currency actual result the msrp is not converted but it is labeled as the chosen currency note that the msrp is configured in euro but displayed in dollars with the euro amount after applying the change the msrp is converted rates in table directory currency rate
1
7,425
9,669,439,696
IssuesEvent
2019-05-21 17:22:36
jenkinsci/configuration-as-code-plugin
https://api.github.com/repos/jenkinsci/configuration-as-code-plugin
closed
provide build_timestamp configuration demo
plugin-compatibility
[jenkins-jira]: https://issues.jenkins-ci.org [dashboard]: https://issues.jenkins.io/secure/Dashboard.jspa?selectPageId=18341 [contributing]: ../blob/master/docs/CONTRIBUTING.md [compatibility]: ../blob/master/docs/COMPATIBILITY.md ### Your checklist for this issue 🚨 Please review the [guidelines for contributing][contributing] to this repository. 💡 To better understand plugin compatibility issues, you can [read more here][compatibility] <!-- Here is a link to get you started with creating the issue over at Jenkins JIRA https://issues.jenkins-ci.org/secure/CreateIssueDetails!init.jspa?pid=10172&issuetype=1&summary=Cannot+configure+X+plugin+with+JCasC&labels=jcasc-compatibility --> - [ ] Create an issue on [issues.jenkins-ci.org][jenkins-jira], set the component to the plugin you are reporting it against - [ ] Before creating an issue on [Jenkins JIRA][jenkins-jira], check for [an existing one via dashboard][dashboard] - [ ] Link to [Jenkins JIRA issue][jenkins-jira] - [ ] Ensure [Jenkins JIRA issue][jenkins-jira] has the label `jcasc-compatibility` - [ ] Link to plugin's GitHub repository - [x] Link to Plugin Compatibility Tracker #809 <!-- Put an `x` into the [ ] to show you have filled the information below Describe your issue below --> ### Description Please add the https://wiki.jenkins.io/display/JENKINS/Build+Timestamp+Plugin to the casc demo and to the plugin itself.
True
provide build_timestamp configuration demo - [jenkins-jira]: https://issues.jenkins-ci.org [dashboard]: https://issues.jenkins.io/secure/Dashboard.jspa?selectPageId=18341 [contributing]: ../blob/master/docs/CONTRIBUTING.md [compatibility]: ../blob/master/docs/COMPATIBILITY.md ### Your checklist for this issue 🚨 Please review the [guidelines for contributing][contributing] to this repository. 💡 To better understand plugin compatibility issues, you can [read more here][compatibility] <!-- Here is a link to get you started with creating the issue over at Jenkins JIRA https://issues.jenkins-ci.org/secure/CreateIssueDetails!init.jspa?pid=10172&issuetype=1&summary=Cannot+configure+X+plugin+with+JCasC&labels=jcasc-compatibility --> - [ ] Create an issue on [issues.jenkins-ci.org][jenkins-jira], set the component to the plugin you are reporting it against - [ ] Before creating an issue on [Jenkins JIRA][jenkins-jira], check for [an existing one via dashboard][dashboard] - [ ] Link to [Jenkins JIRA issue][jenkins-jira] - [ ] Ensure [Jenkins JIRA issue][jenkins-jira] has the label `jcasc-compatibility` - [ ] Link to plugin's GitHub repository - [x] Link to Plugin Compatibility Tracker #809 <!-- Put an `x` into the [ ] to show you have filled the information below Describe your issue below --> ### Description Please add the https://wiki.jenkins.io/display/JENKINS/Build+Timestamp+Plugin to the casc demo and to the plugin itself.
non_priority
provide build timestamp configuration demo blob master docs contributing md blob master docs compatibility md your checklist for this issue 🚨 please review the to this repository 💡 to better understand plugin compatibility issues you can here is a link to get you started with creating the issue over at jenkins jira create an issue on set the component to the plugin you are reporting it against before creating an issue on check for link to ensure has the label jcasc compatibility link to plugin s github repository link to plugin compatibility tracker put an x into the to show you have filled the information below describe your issue below description please add the to the casc demo and to the plugin itself
0
20,203
15,112,109,359
IssuesEvent
2021-02-08 21:21:54
stripe/vscode-stripe
https://api.github.com/repos/stripe/vscode-stripe
opened
Allow user to easily skip https verification on `listen` command
usability
From customer: "I use the extension to run the `listen` command, but then I cancel, press ctrl-p to get last command, and add --skip-verify, because I use https:// for my local server. Would be nice if I could tick a box to "skip https verification". See https://github.com/stripe/stripe-cli/issues/108#issuecomment-521806984"
True
Allow user to easily skip https verification on `listen` command - From customer: "I use the extension to run the `listen` command, but then I cancel, press ctrl-p to get last command, and add --skip-verify, because I use https:// for my local server. Would be nice if I could tick a box to "skip https verification". See https://github.com/stripe/stripe-cli/issues/108#issuecomment-521806984"
non_priority
allow user to easily skip https verification on listen command from customer i use the extension to run the listen command but then i cancel press ctrl p to get last command and add skip verify because i use https for my local server would be nice if i could tick a box to skip https verification see
0
129,070
18,070,769,057
IssuesEvent
2021-09-21 02:26:43
JoePep09/WebGoat
https://api.github.com/repos/JoePep09/WebGoat
opened
CVE-2021-3805 (Medium) detected in object-path-0.9.2.tgz
security vulnerability
## CVE-2021-3805 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>object-path-0.9.2.tgz</b></p></summary> <p>Access deep properties using a path</p> <p>Library home page: <a href="https://registry.npmjs.org/object-path/-/object-path-0.9.2.tgz">https://registry.npmjs.org/object-path/-/object-path-0.9.2.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/object-path/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - eazy-logger-3.0.2.tgz - tfunk-3.1.0.tgz - :x: **object-path-0.9.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> object-path is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3805>CVE-2021-3805</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/571e3baf-7c46-46e3-9003-ba7e4e623053/">https://huntr.dev/bounties/571e3baf-7c46-46e3-9003-ba7e4e623053/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: object-path - 0.11.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3805 (Medium) detected in object-path-0.9.2.tgz - ## CVE-2021-3805 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>object-path-0.9.2.tgz</b></p></summary> <p>Access deep properties using a path</p> <p>Library home page: <a href="https://registry.npmjs.org/object-path/-/object-path-0.9.2.tgz">https://registry.npmjs.org/object-path/-/object-path-0.9.2.tgz</a></p> <p>Path to dependency file: WebGoat/docs/package.json</p> <p>Path to vulnerable library: WebGoat/docs/node_modules/object-path/package.json</p> <p> Dependency Hierarchy: - browser-sync-2.26.3.tgz (Root Library) - eazy-logger-3.0.2.tgz - tfunk-3.1.0.tgz - :x: **object-path-0.9.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> object-path is vulnerable to Improperly Controlled Modification of Object Prototype Attributes ('Prototype Pollution') <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3805>CVE-2021-3805</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/571e3baf-7c46-46e3-9003-ba7e4e623053/">https://huntr.dev/bounties/571e3baf-7c46-46e3-9003-ba7e4e623053/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: object-path - 0.11.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in object path tgz cve medium severity vulnerability vulnerable library object path tgz access deep properties using a path library home page a href path to dependency file webgoat docs package json path to vulnerable library webgoat docs node modules object path package json dependency hierarchy browser sync tgz root library eazy logger tgz tfunk tgz x object path tgz vulnerable library found in base branch develop vulnerability details object path is vulnerable to improperly controlled modification of object prototype attributes prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution object path step up your open source security game with whitesource
0
59,807
24,879,034,168
IssuesEvent
2022-10-27 22:09:09
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
"Authenticate the app" section tells you to follow principle of least privilege, but doesn't do that itself
service-bus-messaging/svc triaged assigned-to-author docs-experience Pri1
In the "Authenticate the app to Azure" section, in the "passwordless" alternative, it says "The following example assigns the Azure Service Bus Data Owner role to your user account, which provides full access to Azure Service Bus resources. In a real scenario, follow the [Principle of Least Privilege](https://learn.microsoft.com/en-us/azure/active-directory/develop/secure-least-privileged-access)" ... so why doesn't it just say what the minimum things you need are? At the moment, this means it's making life harder for the people who are trying to do things properly: they will all have to do further searches to find the answer (good luck with that on a topic that you're just learning). And it's worth saying that I don't have permission to grant permissions on our Azure, so I can't even go and look what's in there. Which makes it very hard to ask the administrator to grant me permissions that I don't know. I think I've found it here: https://learn.microsoft.com/en-us/azure/service-bus-messaging/service-bus-sas#rights-required-for-service-bus-operations and it looks like I need "Send" to send messages and "Listen" to receive them; and "Manage" if I wanted to do admin-like things. --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 443c9051-c55e-1202-7b75-ce86f6977017 * Version Independent ID: 677b6375-196f-fed8-7300-b9823c1031d5 * Content: [Get started with Azure Service Bus topics (.NET) - Azure Service Bus](https://learn.microsoft.com/en-us/azure/service-bus-messaging/service-bus-dotnet-how-to-use-topics-subscriptions?tabs=passwordless%2Croles-azure-portal%2Csign-in-azure-cli%2Cidentity-visual-studio#authenticate-the-app-to-azure) * Content Source: [articles/service-bus-messaging/service-bus-dotnet-how-to-use-topics-subscriptions.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/service-bus-messaging/service-bus-dotnet-how-to-use-topics-subscriptions.md) * Service: **service-bus-messaging** * GitHub Login: @spelluru * Microsoft Alias: **spelluru**
1.0
"Authenticate the app" section tells you to follow principle of least privilege, but doesn't do that itself - In the "Authenticate the app to Azure" section, in the "passwordless" alternative, it says "The following example assigns the Azure Service Bus Data Owner role to your user account, which provides full access to Azure Service Bus resources. In a real scenario, follow the [Principle of Least Privilege](https://learn.microsoft.com/en-us/azure/active-directory/develop/secure-least-privileged-access)" ... so why doesn't it just say what the minimum things you need are? At the moment, this means it's making life harder for the people who are trying to do things properly: they will all have to do further searches to find the answer (good luck with that on a topic that you're just learning). And it's worth saying that I don't have permission to grant permissions on our Azure, so I can't even go and look what's in there. Which makes it very hard to ask the administrator to grant me permissions that I don't know. I think I've found it here: https://learn.microsoft.com/en-us/azure/service-bus-messaging/service-bus-sas#rights-required-for-service-bus-operations and it looks like I need "Send" to send messages and "Listen" to receive them; and "Manage" if I wanted to do admin-like things. --- #### Document Details ⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.* * ID: 443c9051-c55e-1202-7b75-ce86f6977017 * Version Independent ID: 677b6375-196f-fed8-7300-b9823c1031d5 * Content: [Get started with Azure Service Bus topics (.NET) - Azure Service Bus](https://learn.microsoft.com/en-us/azure/service-bus-messaging/service-bus-dotnet-how-to-use-topics-subscriptions?tabs=passwordless%2Croles-azure-portal%2Csign-in-azure-cli%2Cidentity-visual-studio#authenticate-the-app-to-azure) * Content Source: [articles/service-bus-messaging/service-bus-dotnet-how-to-use-topics-subscriptions.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/service-bus-messaging/service-bus-dotnet-how-to-use-topics-subscriptions.md) * Service: **service-bus-messaging** * GitHub Login: @spelluru * Microsoft Alias: **spelluru**
non_priority
authenticate the app section tells you to follow principle of least privilege but doesn t do that itself in the authenticate the app to azure section in the passwordless alternative it says the following example assigns the azure service bus data owner role to your user account which provides full access to azure service bus resources in a real scenario follow the so why doesn t it just say what the minimum things you need are at the moment this means it s making life harder for the people who are trying to do things properly they will all have to do further searches to find the answer good luck with that on a topic that you re just learning and it s worth saying that i don t have permission to grant permissions on our azure so i can t even go and look what s in there which makes it very hard to ask the administrator to grant me permissions that i don t know i think i ve found it here and it looks like i need send to send messages and listen to receive them and manage if i wanted to do admin like things document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service service bus messaging github login spelluru microsoft alias spelluru
0
349,797
31,831,810,835
IssuesEvent
2023-09-14 11:05:33
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachprod: azure.List should retry on "ExpiredAuthenticationToken"
C-bug X-stale A-roachprod no-issue-activity T-testeng O-cloudreport
Azure provider currently uses CLI Authorizer (see `auth.NewAuthorizerFromCLI`) which essentially executes `az account get-access-token`. The returned auth token can expire after 5 minutes according to official doc. [1], > The token will be valid for at least 5 minutes with the maximum at 60 minutes. If the subscription argument isn't specified, the current account is used. Below, we see that VMs took longer than 5 minutes to start (see `SyncedCluster.Wait`). Subsequently, `cleanupFailedCreate` is invoked which in turn invokes `azure.List`. The latter fails because the auth token has already expired. Consequently, `DestroyCluster` is skipped, leaving the resources dangling. ``` 07:23:24 cluster_cloud.go:155: stan-cldrprt23-Standard-E32s-v5-3228080358-125-1-0001 10.1.0.28 10.1.0.28 20.55.66.124 07:23:24 cluster_cloud.go:155: stan-cldrprt23-Standard-E32s-v5-3228080358-125-1-0002 10.1.0.29 10.1.0.29 20.55.66.159 07:23:24 cluster_cloud.go:155: stan-cldrprt23-Standard-E32s-v5-3228080358-125-1-0003 10.1.0.31 10.1.0.31 20.55.66.178 07:23:24 cluster_cloud.go:155: stan-cldrprt23-Standard-E32s-v5-3228080358-125-1-0004 10.1.0.30 10.1.0.30 20.55.66.160 stan-cldrprt23-Standard-E32s-v5-3228080358-125-1: waiting for nodes to start... 08:18:29 cluster_synced.go:775: 3: timed out after 5m Cleaning up partially-created cluster (prev err: not all nodes booted successfully) Error while cleaning up partially-created cluster: compute.VirtualMachinesClient#ListAll: Failure responding to request: StatusCode=401 -- Original Error: autorest/azure: Service returned an error. Status=401 Code="ExpiredAuthenticationToken" Message="The access token expiry UTC time '2/20/2022 8:14:57 AM' is earlier than current UTC time '2/20/2022 8:18:29 AM'." Error: UNCLASSIFIED_PROBLEM: not all nodes booted successfully (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Wait | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:780 | github.com/cockroachdb/cockroach/pkg/roachprod.SetupSSH | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:587 | github.com/cockroachdb/cockroach/pkg/roachprod.Create | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:1247 | main.glob..func1 | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:129 | main.wrap.func1 | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:68 | github.com/spf13/cobra.(*Command).execute | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:860 | github.com/spf13/cobra.(*Command).ExecuteC | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:974 | github.com/spf13/cobra.(*Command).Execute | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:902 | main.main | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:988 | runtime.main | /usr/local/go/src/runtime/proc.go:255 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1581 Wraps: (3) not all nodes booted successfully ``` [1] https://docs.microsoft.com/en-us/cli/azure/account?view=azure-cli-latest Epic: CRDB-10428 Jira issue: CRDB-13679
1.0
roachprod: azure.List should retry on "ExpiredAuthenticationToken" - Azure provider currently uses CLI Authorizer (see `auth.NewAuthorizerFromCLI`) which essentially executes `az account get-access-token`. The returned auth token can expire after 5 minutes according to official doc. [1], > The token will be valid for at least 5 minutes with the maximum at 60 minutes. If the subscription argument isn't specified, the current account is used. Below, we see that VMs took longer than 5 minutes to start (see `SyncedCluster.Wait`). Subsequently, `cleanupFailedCreate` is invoked which in turn invokes `azure.List`. The latter fails because the auth token has already expired. Consequently, `DestroyCluster` is skipped, leaving the resources dangling. ``` 07:23:24 cluster_cloud.go:155: stan-cldrprt23-Standard-E32s-v5-3228080358-125-1-0001 10.1.0.28 10.1.0.28 20.55.66.124 07:23:24 cluster_cloud.go:155: stan-cldrprt23-Standard-E32s-v5-3228080358-125-1-0002 10.1.0.29 10.1.0.29 20.55.66.159 07:23:24 cluster_cloud.go:155: stan-cldrprt23-Standard-E32s-v5-3228080358-125-1-0003 10.1.0.31 10.1.0.31 20.55.66.178 07:23:24 cluster_cloud.go:155: stan-cldrprt23-Standard-E32s-v5-3228080358-125-1-0004 10.1.0.30 10.1.0.30 20.55.66.160 stan-cldrprt23-Standard-E32s-v5-3228080358-125-1: waiting for nodes to start... 08:18:29 cluster_synced.go:775: 3: timed out after 5m Cleaning up partially-created cluster (prev err: not all nodes booted successfully) Error while cleaning up partially-created cluster: compute.VirtualMachinesClient#ListAll: Failure responding to request: StatusCode=401 -- Original Error: autorest/azure: Service returned an error. Status=401 Code="ExpiredAuthenticationToken" Message="The access token expiry UTC time '2/20/2022 8:14:57 AM' is earlier than current UTC time '2/20/2022 8:18:29 AM'." Error: UNCLASSIFIED_PROBLEM: not all nodes booted successfully (1) UNCLASSIFIED_PROBLEM Wraps: (2) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Wait | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:780 | github.com/cockroachdb/cockroach/pkg/roachprod.SetupSSH | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:587 | github.com/cockroachdb/cockroach/pkg/roachprod.Create | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:1247 | main.glob..func1 | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:129 | main.wrap.func1 | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:68 | github.com/spf13/cobra.(*Command).execute | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:860 | github.com/spf13/cobra.(*Command).ExecuteC | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:974 | github.com/spf13/cobra.(*Command).Execute | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:902 | main.main | /home/srosenberg/go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:988 | runtime.main | /usr/local/go/src/runtime/proc.go:255 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1581 Wraps: (3) not all nodes booted successfully ``` [1] https://docs.microsoft.com/en-us/cli/azure/account?view=azure-cli-latest Epic: CRDB-10428 Jira issue: CRDB-13679
non_priority
roachprod azure list should retry on expiredauthenticationtoken azure provider currently uses cli authorizer see auth newauthorizerfromcli which essentially executes az account get access token the returned auth token can expire after minutes according to official doc the token will be valid for at least minutes with the maximum at minutes if the subscription argument isn t specified the current account is used below we see that vms took longer than minutes to start see syncedcluster wait subsequently cleanupfailedcreate is invoked which in turn invokes azure list the latter fails because the auth token has already expired consequently destroycluster is skipped leaving the resources dangling cluster cloud go stan standard cluster cloud go stan standard cluster cloud go stan standard cluster cloud go stan standard stan standard waiting for nodes to start cluster synced go timed out after cleaning up partially created cluster prev err not all nodes booted successfully error while cleaning up partially created cluster compute virtualmachinesclient listall failure responding to request statuscode original error autorest azure service returned an error status code expiredauthenticationtoken message the access token expiry utc time am is earlier than current utc time am error unclassified problem not all nodes booted successfully unclassified problem wraps attached stack trace stack trace github com cockroachdb cockroach pkg roachprod install syncedcluster wait home srosenberg go src github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod setupssh home srosenberg go src github com cockroachdb cockroach pkg roachprod roachprod go github com cockroachdb cockroach pkg roachprod create home srosenberg go src github com cockroachdb cockroach pkg roachprod roachprod go main glob home srosenberg go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home srosenberg go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home srosenberg go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home srosenberg go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home srosenberg go src github com cockroachdb cockroach vendor github com cobra command go main main home srosenberg go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps not all nodes booted successfully epic crdb jira issue crdb
0
49,967
26,410,678,994
IssuesEvent
2023-01-13 11:58:11
conda-forge/status
https://api.github.com/repos/conda-forge/status
closed
Feedstock uploads need token reset
degraded performance
Our admin migrations bots lost a significant number of feedstock tokens, disabling uploads of feedstock builds. If your feedstock is affected, please send a request for a token reset to <a href="https://github.com/conda-forge/admin-requests">conda-forge/admin-requests</a>. We are working to reset the tokens ourselves but do not know when this will be complete.
True
Feedstock uploads need token reset - Our admin migrations bots lost a significant number of feedstock tokens, disabling uploads of feedstock builds. If your feedstock is affected, please send a request for a token reset to <a href="https://github.com/conda-forge/admin-requests">conda-forge/admin-requests</a>. We are working to reset the tokens ourselves but do not know when this will be complete.
non_priority
feedstock uploads need token reset our admin migrations bots lost a significant number of feedstock tokens disabling uploads of feedstock builds if your feedstock is affected please send a request for a token reset to a href we are working to reset the tokens ourselves but do not know when this will be complete
0
212,284
16,437,892,826
IssuesEvent
2021-05-20 11:21:44
ethersphere/bee
https://api.github.com/repos/ethersphere/bee
closed
TestDelivery is flaking
flaky-test
``` ok github.com/ethersphere/bee/pkg/resolver/multiresolver/multierror 0.030s === RUN TestDelivery ##[error] retrieval_test.go:133: unexpected balance on server. want 10 got 0 --- FAIL: TestDelivery (0.00s) FAIL ``` from the CI
1.0
TestDelivery is flaking - ``` ok github.com/ethersphere/bee/pkg/resolver/multiresolver/multierror 0.030s === RUN TestDelivery ##[error] retrieval_test.go:133: unexpected balance on server. want 10 got 0 --- FAIL: TestDelivery (0.00s) FAIL ``` from the CI
non_priority
testdelivery is flaking ok github com ethersphere bee pkg resolver multiresolver multierror run testdelivery retrieval test go unexpected balance on server want got fail testdelivery fail from the ci
0
180,748
6,652,990,981
IssuesEvent
2017-09-29 05:58:31
ocf/slackbridge
https://api.github.com/repos/ocf/slackbridge
closed
Make IRC bots show as away when the Slack user is inactive
enhancement medium-priority
Using [this API method](https://api.slack.com/events/presence_change), it could probably be done.
1.0
Make IRC bots show as away when the Slack user is inactive - Using [this API method](https://api.slack.com/events/presence_change), it could probably be done.
priority
make irc bots show as away when the slack user is inactive using it could probably be done
1
257,096
27,561,768,595
IssuesEvent
2023-03-07 22:45:10
samqws-marketing/walmartlabs-concord
https://api.github.com/repos/samqws-marketing/walmartlabs-concord
closed
CVE-2022-24785 (High) detected in moment-2.24.0.tgz - autoclosed
security vulnerability
## CVE-2022-24785 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.24.0.tgz</b></p></summary> <p>Parse, validate, manipulate, and display dates</p> <p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p> <p>Path to dependency file: /console2/package.json</p> <p>Path to vulnerable library: /console2/node_modules/moment/package.json</p> <p> Dependency Hierarchy: - semantic-ui-calendar-react-0.15.3.tgz (Root Library) - :x: **moment-2.24.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/walmartlabs-concord/commit/b9420f3b9e73a9d381266ece72f7afb756f35a76">b9420f3b9e73a9d381266ece72f7afb756f35a76</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js. <p>Publish Date: 2022-04-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24785>CVE-2022-24785</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4">https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4</a></p> <p>Release Date: 2022-04-04</p> <p>Fix Resolution: moment - 2.29.2</p> </p> </details> <p></p>
True
CVE-2022-24785 (High) detected in moment-2.24.0.tgz - autoclosed - ## CVE-2022-24785 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>moment-2.24.0.tgz</b></p></summary> <p>Parse, validate, manipulate, and display dates</p> <p>Library home page: <a href="https://registry.npmjs.org/moment/-/moment-2.24.0.tgz">https://registry.npmjs.org/moment/-/moment-2.24.0.tgz</a></p> <p>Path to dependency file: /console2/package.json</p> <p>Path to vulnerable library: /console2/node_modules/moment/package.json</p> <p> Dependency Hierarchy: - semantic-ui-calendar-react-0.15.3.tgz (Root Library) - :x: **moment-2.24.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/walmartlabs-concord/commit/b9420f3b9e73a9d381266ece72f7afb756f35a76">b9420f3b9e73a9d381266ece72f7afb756f35a76</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Moment.js is a JavaScript date library for parsing, validating, manipulating, and formatting dates. A path traversal vulnerability impacts npm (server) users of Moment.js between versions 1.0.1 and 2.29.1, especially if a user-provided locale string is directly used to switch moment locale. This problem is patched in 2.29.2, and the patch can be applied to all affected versions. As a workaround, sanitize the user-provided locale name before passing it to Moment.js. <p>Publish Date: 2022-04-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-24785>CVE-2022-24785</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4">https://github.com/moment/moment/security/advisories/GHSA-8hfj-j24r-96c4</a></p> <p>Release Date: 2022-04-04</p> <p>Fix Resolution: moment - 2.29.2</p> </p> </details> <p></p>
non_priority
cve high detected in moment tgz autoclosed cve high severity vulnerability vulnerable library moment tgz parse validate manipulate and display dates library home page a href path to dependency file package json path to vulnerable library node modules moment package json dependency hierarchy semantic ui calendar react tgz root library x moment tgz vulnerable library found in head commit a href found in base branch master vulnerability details moment js is a javascript date library for parsing validating manipulating and formatting dates a path traversal vulnerability impacts npm server users of moment js between versions and especially if a user provided locale string is directly used to switch moment locale this problem is patched in and the patch can be applied to all affected versions as a workaround sanitize the user provided locale name before passing it to moment js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution moment
0
830,663
32,020,446,121
IssuesEvent
2023-09-22 03:46:17
dimensionhq/infralink
https://api.github.com/repos/dimensionhq/infralink
reopened
`deploy` Command
todo high-priority
This deploys the stack required to host your application using the previously generated terraform configuration. This step is also responsible for uploading the required Docker images that were previously generated and hosting it on the cloud provider of choice.
1.0
`deploy` Command - This deploys the stack required to host your application using the previously generated terraform configuration. This step is also responsible for uploading the required Docker images that were previously generated and hosting it on the cloud provider of choice.
priority
deploy command this deploys the stack required to host your application using the previously generated terraform configuration this step is also responsible for uploading the required docker images that were previously generated and hosting it on the cloud provider of choice
1
86,397
3,711,498,949
IssuesEvent
2016-03-02 10:35:16
CS2103jan2016-w13-3j/main
https://api.github.com/repos/CS2103jan2016-w13-3j/main
opened
Storage unit to store the completed tasks into a separate file
priority.medium type.enhancement
...so that the file reading process speeds up.
1.0
Storage unit to store the completed tasks into a separate file - ...so that the file reading process speeds up.
priority
storage unit to store the completed tasks into a separate file so that the file reading process speeds up
1
295,145
25,457,934,428
IssuesEvent
2022-11-24 15:41:52
Jamesflynn1/CS344-Opponent-Modelling-Poker
https://api.github.com/repos/Jamesflynn1/CS344-Opponent-Modelling-Poker
closed
Define what game exactly is being used for testing
Project Management Research Testing
What Poker ruleset? How many players? How many rounds? What kind of opponents? ### ANSWER What I am doing: Limit Texas Hold 'em (stretch goal would be the full NL HE but this is almost certainly intractable on DCS machines and 2 days of compute time). 2 player, this is the minimum number of players that we can chose, increasing this number would again add alot of complexity to the project. I have resources limited to the DCS batch compute system. Computer opponents, this will allow for efficient learning and circuments potential ethical issues and admin issues in administering such tests. See #17 for more details.
1.0
Define what game exactly is being used for testing - What Poker ruleset? How many players? How many rounds? What kind of opponents? ### ANSWER What I am doing: Limit Texas Hold 'em (stretch goal would be the full NL HE but this is almost certainly intractable on DCS machines and 2 days of compute time). 2 player, this is the minimum number of players that we can chose, increasing this number would again add alot of complexity to the project. I have resources limited to the DCS batch compute system. Computer opponents, this will allow for efficient learning and circuments potential ethical issues and admin issues in administering such tests. See #17 for more details.
non_priority
define what game exactly is being used for testing what poker ruleset how many players how many rounds what kind of opponents answer what i am doing limit texas hold em stretch goal would be the full nl he but this is almost certainly intractable on dcs machines and days of compute time player this is the minimum number of players that we can chose increasing this number would again add alot of complexity to the project i have resources limited to the dcs batch compute system computer opponents this will allow for efficient learning and circuments potential ethical issues and admin issues in administering such tests see for more details
0
648,891
21,212,551,733
IssuesEvent
2022-04-11 01:51:32
EeeeG-Inc/OKR-web-app
https://api.github.com/repos/EeeeG-Inc/OKR-web-app
opened
アプリ名称を設定画面で決められるようにする
priority: low
<img width="1248" alt="スクリーンショット 2022-04-11 10 49 57" src="https://user-images.githubusercontent.com/27577954/162651894-c639a25a-4fd5-44f9-92fd-42379c161ea8.png"> 現状 `.env` の値を参照しているので、アプリ的には設定画面から参照できたほうがいい
1.0
アプリ名称を設定画面で決められるようにする - <img width="1248" alt="スクリーンショット 2022-04-11 10 49 57" src="https://user-images.githubusercontent.com/27577954/162651894-c639a25a-4fd5-44f9-92fd-42379c161ea8.png"> 現状 `.env` の値を参照しているので、アプリ的には設定画面から参照できたほうがいい
priority
アプリ名称を設定画面で決められるようにする img width alt スクリーンショット src 現状 env の値を参照しているので、アプリ的には設定画面から参照できたほうがいい
1
673,184
22,951,599,253
IssuesEvent
2022-07-19 07:56:35
celo-org/celo-monorepo
https://api.github.com/repos/celo-org/celo-monorepo
closed
[Protocol Economics] Cap on cUSD CP-DOTO tank size
economics Priority: P2 stablecoin-issuance
### Expected Behavior As a safeguard against oracle manipulation or other issues related to an incorrect cUSD/CELO exchange rate, there should be a cap on the cUSD tank size. The rationale is that if other safeguards fail, the impact is limited ### Current Behavior There is no cap. ### Definition of Done - We have a design doc detailing why the cap is necessary and how we come up with a value for it - Cap is implemented and live on mainnet.
1.0
[Protocol Economics] Cap on cUSD CP-DOTO tank size - ### Expected Behavior As a safeguard against oracle manipulation or other issues related to an incorrect cUSD/CELO exchange rate, there should be a cap on the cUSD tank size. The rationale is that if other safeguards fail, the impact is limited ### Current Behavior There is no cap. ### Definition of Done - We have a design doc detailing why the cap is necessary and how we come up with a value for it - Cap is implemented and live on mainnet.
priority
cap on cusd cp doto tank size expected behavior as a safeguard against oracle manipulation or other issues related to an incorrect cusd celo exchange rate there should be a cap on the cusd tank size the rationale is that if other safeguards fail the impact is limited current behavior there is no cap definition of done we have a design doc detailing why the cap is necessary and how we come up with a value for it cap is implemented and live on mainnet
1
148,117
11,838,976,340
IssuesEvent
2020-03-23 16:26:45
istio/istio
https://api.github.com/repos/istio/istio
closed
StartupProbe's are removed from Deployments
area/config area/test and release area/user experience lifecycle/needs-triage
**Bug description** When creating a Deployment in Kubernetes with a StartupProbe included the (as I believe) sidecar injector service which rewrites the readiness/liveness probes will remove the StartupProbe. I think this issue is related to https://github.com/istio/istio/issues/19324. **Expected behavior** StartupProbes are able to be set against a Deployment **Steps to reproduce the bug** Create a Deployment with a StartupProbe set, check the created Deployment and notice the StartupProbe is no longer then. **Version (include the output of `istioctl version --remote` and `kubectl version` and `helm version` if you used Helm)** ``` > istioctl version --remote client version: 1.4.3 citadel version: 1.4.3 citadel version: 1.4.3 citadel version: 1.4.3 galley version: 1.4.3 galley version: 1.4.3 galley version: 1.4.3 ingressgateway version: 1.4.3 ingressgateway version: 1.4.3 ingressgateway version: 1.4.3 nodeagent version: nodeagent version: nodeagent version: pilot version: 1.4.3 pilot version: 1.4.3 pilot version: 1.4.3 policy version: 1.4.3 policy version: 1.4.3 policy version: 1.4.3 sidecar-injector version: 1.4.3 sidecar-injector version: 1.4.3 sidecar-injector version: 1.4.3 telemetry version: 1.4.3 telemetry version: 1.4.3 telemetry version: 1.4.3 data plane version: 1.4.3 (32 proxies) ``` **How was Istio installed?** Legacy shipped helm charts inside the istio/istio repository **Environment where bug was observed (cloud vendor, OS, etc)** Kubernetes v1.17.0 running on CoreOS virtualised inside VMWare
1.0
StartupProbe's are removed from Deployments - **Bug description** When creating a Deployment in Kubernetes with a StartupProbe included the (as I believe) sidecar injector service which rewrites the readiness/liveness probes will remove the StartupProbe. I think this issue is related to https://github.com/istio/istio/issues/19324. **Expected behavior** StartupProbes are able to be set against a Deployment **Steps to reproduce the bug** Create a Deployment with a StartupProbe set, check the created Deployment and notice the StartupProbe is no longer then. **Version (include the output of `istioctl version --remote` and `kubectl version` and `helm version` if you used Helm)** ``` > istioctl version --remote client version: 1.4.3 citadel version: 1.4.3 citadel version: 1.4.3 citadel version: 1.4.3 galley version: 1.4.3 galley version: 1.4.3 galley version: 1.4.3 ingressgateway version: 1.4.3 ingressgateway version: 1.4.3 ingressgateway version: 1.4.3 nodeagent version: nodeagent version: nodeagent version: pilot version: 1.4.3 pilot version: 1.4.3 pilot version: 1.4.3 policy version: 1.4.3 policy version: 1.4.3 policy version: 1.4.3 sidecar-injector version: 1.4.3 sidecar-injector version: 1.4.3 sidecar-injector version: 1.4.3 telemetry version: 1.4.3 telemetry version: 1.4.3 telemetry version: 1.4.3 data plane version: 1.4.3 (32 proxies) ``` **How was Istio installed?** Legacy shipped helm charts inside the istio/istio repository **Environment where bug was observed (cloud vendor, OS, etc)** Kubernetes v1.17.0 running on CoreOS virtualised inside VMWare
non_priority
startupprobe s are removed from deployments bug description when creating a deployment in kubernetes with a startupprobe included the as i believe sidecar injector service which rewrites the readiness liveness probes will remove the startupprobe i think this issue is related to expected behavior startupprobes are able to be set against a deployment steps to reproduce the bug create a deployment with a startupprobe set check the created deployment and notice the startupprobe is no longer then version include the output of istioctl version remote and kubectl version and helm version if you used helm istioctl version remote client version citadel version citadel version citadel version galley version galley version galley version ingressgateway version ingressgateway version ingressgateway version nodeagent version nodeagent version nodeagent version pilot version pilot version pilot version policy version policy version policy version sidecar injector version sidecar injector version sidecar injector version telemetry version telemetry version telemetry version data plane version proxies how was istio installed legacy shipped helm charts inside the istio istio repository environment where bug was observed cloud vendor os etc kubernetes running on coreos virtualised inside vmware
0
377,132
11,164,581,225
IssuesEvent
2019-12-27 05:45:22
space-wizards/RobustToolbox
https://api.github.com/repos/space-wizards/RobustToolbox
closed
Transforms should be linked to their grid
Area: Entities Priority: 2-medium Type: Refactor
Transforms should be made relative to a grid when a transform is located on that grid, and when a transform moves off that grid it should switch to using world space instead. What that allows: Grids themselves become able to move without affecting the transform of anything on the grid itself.
1.0
Transforms should be linked to their grid - Transforms should be made relative to a grid when a transform is located on that grid, and when a transform moves off that grid it should switch to using world space instead. What that allows: Grids themselves become able to move without affecting the transform of anything on the grid itself.
priority
transforms should be linked to their grid transforms should be made relative to a grid when a transform is located on that grid and when a transform moves off that grid it should switch to using world space instead what that allows grids themselves become able to move without affecting the transform of anything on the grid itself
1
15,204
19,278,265,956
IssuesEvent
2021-12-10 14:24:03
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
opened
Upgrading from 20.3 to 21.3: Partition key cannot contain non-deterministic functions, but contains function now
backward compatibility
**Describe the issue** When non-deterministic function is used in 20.3, upgrade to 21.3 can't be performed **How to reproduce** * When upgrading from 20.3.21.2 to 21.3.18.4 * Interface: clickhouse-client * No unusual settings (used config.xml and users.xml from [20.3](https://github.com/ClickHouse/ClickHouse/tree/20.3/programs/server) without any other files) Do this on 20.3: * `CREATE TABLE test_coalescence (id Int, name String) Engine=MergeTree() ORDER BY id PARTITION BY toDate(coalesce(id, now())); ` * `INSERT INTO test_coalescence VALUES (5, 'hello')` Upgrade to 21.3, start server and select: * `INSERT INTO test_coalescence` **Error message and/or stacktrace** ``` 2021.12.10 17:09:04.517045 [ 4509 ] {} <Error> Application: Caught exception while loading metadata: Code: 36, e.displayText() = DB::Exception: Partition key cannot contain non-deterministic functions, but contains function now: Cannot attach table `default`.`test_coalescence` from metadata file /var/lib/clickhouse/metadata/default/test_coalescence.sql from query ATTACH TABLE default.test_coalescence (`id` Int32, `name` String) ENGINE = MergeTree PARTITION BY toDate(coalesce(id, now())) ORDER BY id SETTINGS index_granularity = 8192: while loading database `default` from path /var/lib/clickhouse/metadata/default, Stack trace (when copying this message, always include the lines below): 0. ? @ 0xf39fb87 in /usr/bin/clickhouse 1. DB::MergeTreeData::checkPartitionKeyAndInitMinMax(DB::KeyDescription const&) @ 0xf39b74d in /usr/bin/clickhouse 2. DB::MergeTreeData::MergeTreeData(DB::StorageID const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::StorageInMemoryMetadata const&, DB::Context&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::MergeTreeData::MergingParams const&, std::__1::unique_ptr<DB::MergeTreeSettings, std::__1::default_delete<DB::MergeTreeSettings> >, bool, bool, std::__1::function<void (std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&)>) @ 0xf396435 in /usr/bin/clickhouse 3. DB::StorageMergeTree::StorageMergeTree(DB::StorageID const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::StorageInMemoryMetadata const&, bool, DB::Context&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::MergeTreeData::MergingParams const&, std::__1::unique_ptr<DB::MergeTreeSettings, std::__1::default_delete<DB::MergeTreeSettings> >, bool) @ 0xf16170e in /usr/bin/clickhouse 4. ? @ 0xf59730d in /usr/bin/clickhouse 5. DB::StorageFactory::get(DB::ASTCreateQuery const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::Context&, DB::Context&, DB::ColumnsDescription const&, DB::ConstraintsDescription const&, bool) const @ 0xf0dbe19 in /usr/bin/clickhouse 6. DB::createTableFromAST(DB::ASTCreateQuery, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::Context&, bool) @ 0xe782789 in /usr/bin/clickhouse 7. ? @ 0xe781911 in /usr/bin/clickhouse 8. ThreadPoolImpl<ThreadFromGlobalPool>::worker(std::__1::__list_iterator<ThreadFromGlobalPool, void*>) @ 0x853bac8 in /usr/bin/clickhouse 9. ThreadFromGlobalPool::ThreadFromGlobalPool<void ThreadPoolImpl<ThreadFromGlobalPool>::scheduleImpl<void>(std::__1::function<void ()>, int, std::__1::optional<unsigned long>)::'lambda1'()>(void&&, void ThreadPoolImpl<ThreadFromGlobalPool>::scheduleImpl<void>(std::__1::function<void ()>, int, std::__1::optional<unsigned long>)::'lambda1'()&&...)::'lambda'()::operator()() @ 0x853d4df in /usr/bin/clickhouse 10. ThreadPoolImpl<std::__1::thread>::worker(std::__1::__list_iterator<std::__1::thread, void*>) @ 0x853911f in /usr/bin/clickhouse 11. ? @ 0x853c593 in /usr/bin/clickhouse 12. start_thread @ 0x76db in /lib/x86_64-linux-gnu/libpthread-2.27.so 13. __clone @ 0x12171f in /lib/x86_64-linux-gnu/libc-2.27.so (version 21.3.18.4 (official build)) ```
True
Upgrading from 20.3 to 21.3: Partition key cannot contain non-deterministic functions, but contains function now - **Describe the issue** When non-deterministic function is used in 20.3, upgrade to 21.3 can't be performed **How to reproduce** * When upgrading from 20.3.21.2 to 21.3.18.4 * Interface: clickhouse-client * No unusual settings (used config.xml and users.xml from [20.3](https://github.com/ClickHouse/ClickHouse/tree/20.3/programs/server) without any other files) Do this on 20.3: * `CREATE TABLE test_coalescence (id Int, name String) Engine=MergeTree() ORDER BY id PARTITION BY toDate(coalesce(id, now())); ` * `INSERT INTO test_coalescence VALUES (5, 'hello')` Upgrade to 21.3, start server and select: * `INSERT INTO test_coalescence` **Error message and/or stacktrace** ``` 2021.12.10 17:09:04.517045 [ 4509 ] {} <Error> Application: Caught exception while loading metadata: Code: 36, e.displayText() = DB::Exception: Partition key cannot contain non-deterministic functions, but contains function now: Cannot attach table `default`.`test_coalescence` from metadata file /var/lib/clickhouse/metadata/default/test_coalescence.sql from query ATTACH TABLE default.test_coalescence (`id` Int32, `name` String) ENGINE = MergeTree PARTITION BY toDate(coalesce(id, now())) ORDER BY id SETTINGS index_granularity = 8192: while loading database `default` from path /var/lib/clickhouse/metadata/default, Stack trace (when copying this message, always include the lines below): 0. ? @ 0xf39fb87 in /usr/bin/clickhouse 1. DB::MergeTreeData::checkPartitionKeyAndInitMinMax(DB::KeyDescription const&) @ 0xf39b74d in /usr/bin/clickhouse 2. DB::MergeTreeData::MergeTreeData(DB::StorageID const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::StorageInMemoryMetadata const&, DB::Context&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::MergeTreeData::MergingParams const&, std::__1::unique_ptr<DB::MergeTreeSettings, std::__1::default_delete<DB::MergeTreeSettings> >, bool, bool, std::__1::function<void (std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&)>) @ 0xf396435 in /usr/bin/clickhouse 3. DB::StorageMergeTree::StorageMergeTree(DB::StorageID const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::StorageInMemoryMetadata const&, bool, DB::Context&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::MergeTreeData::MergingParams const&, std::__1::unique_ptr<DB::MergeTreeSettings, std::__1::default_delete<DB::MergeTreeSettings> >, bool) @ 0xf16170e in /usr/bin/clickhouse 4. ? @ 0xf59730d in /usr/bin/clickhouse 5. DB::StorageFactory::get(DB::ASTCreateQuery const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::Context&, DB::Context&, DB::ColumnsDescription const&, DB::ConstraintsDescription const&, bool) const @ 0xf0dbe19 in /usr/bin/clickhouse 6. DB::createTableFromAST(DB::ASTCreateQuery, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, DB::Context&, bool) @ 0xe782789 in /usr/bin/clickhouse 7. ? @ 0xe781911 in /usr/bin/clickhouse 8. ThreadPoolImpl<ThreadFromGlobalPool>::worker(std::__1::__list_iterator<ThreadFromGlobalPool, void*>) @ 0x853bac8 in /usr/bin/clickhouse 9. ThreadFromGlobalPool::ThreadFromGlobalPool<void ThreadPoolImpl<ThreadFromGlobalPool>::scheduleImpl<void>(std::__1::function<void ()>, int, std::__1::optional<unsigned long>)::'lambda1'()>(void&&, void ThreadPoolImpl<ThreadFromGlobalPool>::scheduleImpl<void>(std::__1::function<void ()>, int, std::__1::optional<unsigned long>)::'lambda1'()&&...)::'lambda'()::operator()() @ 0x853d4df in /usr/bin/clickhouse 10. ThreadPoolImpl<std::__1::thread>::worker(std::__1::__list_iterator<std::__1::thread, void*>) @ 0x853911f in /usr/bin/clickhouse 11. ? @ 0x853c593 in /usr/bin/clickhouse 12. start_thread @ 0x76db in /lib/x86_64-linux-gnu/libpthread-2.27.so 13. __clone @ 0x12171f in /lib/x86_64-linux-gnu/libc-2.27.so (version 21.3.18.4 (official build)) ```
non_priority
upgrading from to partition key cannot contain non deterministic functions but contains function now describe the issue when non deterministic function is used in upgrade to can t be performed how to reproduce when upgrading from to interface clickhouse client no unusual settings used config xml and users xml from without any other files do this on create table test coalescence id int name string engine mergetree order by id partition by todate coalesce id now insert into test coalescence values hello upgrade to start server and select insert into test coalescence error message and or stacktrace application caught exception while loading metadata code e displaytext db exception partition key cannot contain non deterministic functions but contains function now cannot attach table default test coalescence from metadata file var lib clickhouse metadata default test coalescence sql from query attach table default test coalescence id name string engine mergetree partition by todate coalesce id now order by id settings index granularity while loading database default from path var lib clickhouse metadata default stack trace when copying this message always include the lines below in usr bin clickhouse db mergetreedata checkpartitionkeyandinitminmax db keydescription const in usr bin clickhouse db mergetreedata mergetreedata db storageid const std basic string std allocator const db storageinmemorymetadata const db context std basic string std allocator const db mergetreedata mergingparams const std unique ptr bool bool std function std allocator const in usr bin clickhouse db storagemergetree storagemergetree db storageid const std basic string std allocator const db storageinmemorymetadata const bool db context std basic string std allocator const db mergetreedata mergingparams const std unique ptr bool in usr bin clickhouse in usr bin clickhouse db storagefactory get db astcreatequery const std basic string std allocator const db context db context db columnsdescription const db constraintsdescription const bool const in usr bin clickhouse db createtablefromast db astcreatequery std basic string std allocator const std basic string std allocator const db context bool in usr bin clickhouse in usr bin clickhouse threadpoolimpl worker std list iterator in usr bin clickhouse threadfromglobalpool threadfromglobalpool scheduleimpl std function int std optional void void threadpoolimpl scheduleimpl std function int std optional lambda operator in usr bin clickhouse threadpoolimpl worker std list iterator in usr bin clickhouse in usr bin clickhouse start thread in lib linux gnu libpthread so clone in lib linux gnu libc so version official build
0
687,467
23,527,372,296
IssuesEvent
2022-08-19 12:15:10
inbo/vespa-watch
https://api.github.com/repos/inbo/vespa-watch
reopened
nests not longer added to management module
🔥 priority
Since the 8th of August, nests are not longer added to the management module on Vespa-Watch. Still, nests are synchronized with iNaturalist so it is unclear why the management module is not being updated. The only thing I notice is that total number of nests reached 99 tabs, and maybe this is a maximum so that no extra nests can be added? Is it possible to solve this? Reimbursement of management by VBI requires registration of the management action on the Vespa-Watch website so this should be solved as quickly as possible! Thank you for your help
1.0
nests not longer added to management module - Since the 8th of August, nests are not longer added to the management module on Vespa-Watch. Still, nests are synchronized with iNaturalist so it is unclear why the management module is not being updated. The only thing I notice is that total number of nests reached 99 tabs, and maybe this is a maximum so that no extra nests can be added? Is it possible to solve this? Reimbursement of management by VBI requires registration of the management action on the Vespa-Watch website so this should be solved as quickly as possible! Thank you for your help
priority
nests not longer added to management module since the of august nests are not longer added to the management module on vespa watch still nests are synchronized with inaturalist so it is unclear why the management module is not being updated the only thing i notice is that total number of nests reached tabs and maybe this is a maximum so that no extra nests can be added is it possible to solve this reimbursement of management by vbi requires registration of the management action on the vespa watch website so this should be solved as quickly as possible thank you for your help
1
84,946
16,581,661,202
IssuesEvent
2021-05-31 12:42:03
libjxl/libjxl
https://api.github.com/repos/libjxl/libjxl
closed
GIF / JPEG -> Lossless JXL: Different results with and without `-q 100` / `-d 0`
cjxl encoder
**Summary** For GIF and JPEG, libjxl will use lossless mode (GIF) / lossless JPEG transcode (JPEG) by default. However, if you explicitly ask libjxl to use lossless mode by `-q 100` or `-d 0`, the results would be different. **Steps to reproduce** * `cjxl -s 3 $infile $outfile` * `cjxl -s 3 -d 0 $infile $outfile` * `cjxl -s 3 -q 100 $infile $outfile` **Observed behavior** For JPEG transcode, no parameter or `-q 100` produces 280305, but `-d 0` produces 280246. For GIF, no parameter produces 47431412, while `-d 0` or `-q 100` produces output 26518874. **Expected behavior** The output files should be the same. **Test files** * https://www.ganganonline.com/contents/slime/img/slime_7cover.gif * https://jpegxl.info/fallbacklogo.jpg **Environment** - OS: Linux - Compiler version: clang - CPU type: x86_64 - cjxl/djxl version string: `cjxl [v0.3.7 | SIMD supported: SSE4,Scalar]` **Additional context** For the test JPEG file, the difference is small, but for the GIF file, the difference is very large.
1.0
GIF / JPEG -> Lossless JXL: Different results with and without `-q 100` / `-d 0` - **Summary** For GIF and JPEG, libjxl will use lossless mode (GIF) / lossless JPEG transcode (JPEG) by default. However, if you explicitly ask libjxl to use lossless mode by `-q 100` or `-d 0`, the results would be different. **Steps to reproduce** * `cjxl -s 3 $infile $outfile` * `cjxl -s 3 -d 0 $infile $outfile` * `cjxl -s 3 -q 100 $infile $outfile` **Observed behavior** For JPEG transcode, no parameter or `-q 100` produces 280305, but `-d 0` produces 280246. For GIF, no parameter produces 47431412, while `-d 0` or `-q 100` produces output 26518874. **Expected behavior** The output files should be the same. **Test files** * https://www.ganganonline.com/contents/slime/img/slime_7cover.gif * https://jpegxl.info/fallbacklogo.jpg **Environment** - OS: Linux - Compiler version: clang - CPU type: x86_64 - cjxl/djxl version string: `cjxl [v0.3.7 | SIMD supported: SSE4,Scalar]` **Additional context** For the test JPEG file, the difference is small, but for the GIF file, the difference is very large.
non_priority
gif jpeg lossless jxl different results with and without q d summary for gif and jpeg libjxl will use lossless mode gif lossless jpeg transcode jpeg by default however if you explicitly ask libjxl to use lossless mode by q or d the results would be different steps to reproduce cjxl s infile outfile cjxl s d infile outfile cjxl s q infile outfile observed behavior for jpeg transcode no parameter or q produces but d produces for gif no parameter produces while d or q produces output expected behavior the output files should be the same test files environment os linux compiler version clang cpu type cjxl djxl version string cjxl additional context for the test jpeg file the difference is small but for the gif file the difference is very large
0
170,124
26,905,676,110
IssuesEvent
2023-02-06 18:52:37
webb-tools/webb-experiences
https://api.github.com/repos/webb-tools/webb-experiences
closed
Transfer Component
design 🎨
## Product Design Goals - Enable users to easily transfer shielded funds from one registered address to another ## Deliverables - Create wireframes that represent and organize the user actions for making a transfer to another user - Design a dedicated transfer component that displays required input boxes, success modal, and notifications - Create high fidelity prototype of a transfer component and flow ## Designs and Product Flow ![Image](https://user-images.githubusercontent.com/29983536/191097019-676dda87-5059-4872-8cc8-c7eee48e6813.png) **Figure 1.1 -** *Transfer UI flow.* ### User Flow Transfer 1. Login with registered Note Account 2. Select token asset to transfer 1. These token types will only be webb wrapped assets (e.g. webbUSDC, webbETH) 3. Enter an amount 4. Input other registered addresses 1. Validate address is registered within input box 5. Select relayer 1. Same component used on bridge 6. Send transfer **Note:** The shielded balance available will be informed by the current connected chain. For example, a user makes two deposits into the bridge for USDC, ETH where the destination chain selected was Arbitrum. When the user navigates to the Transfer tab, and is connected to the Arbitrum chain, the available shielded balance to transfer will consist of the webbUSDC, and webbETH previously deposited. However, if they are connected to Optimism and have not deposited anything into the bridge where Optimism is the destination chain the available shielded balance will be 0. ## User Selection Inputs 1. Token type 2. Amount 3. Registered address 4. Relayer selection ## Notifications - Failed transfer - Successful transfer - Invalid recipient address ## Components List - Transfer UI interface for above mentioned inputs - Successful / unsuccessful indicator ## Alternative Transfer UI’s <img src="https://user-images.githubusercontent.com/29983536/191096998-a4fa28bd-51cb-4667-81b2-e35cca12fd6a.png" height=450 /> ## Future Feature Considerations 1. Contact / address book for pre-saved registered addresses 2. Notification informing recipient of transferred funds 1. Currently we do not have anything in place that informs the user that they received funds via transfer # Open Questions
1.0
Transfer Component - ## Product Design Goals - Enable users to easily transfer shielded funds from one registered address to another ## Deliverables - Create wireframes that represent and organize the user actions for making a transfer to another user - Design a dedicated transfer component that displays required input boxes, success modal, and notifications - Create high fidelity prototype of a transfer component and flow ## Designs and Product Flow ![Image](https://user-images.githubusercontent.com/29983536/191097019-676dda87-5059-4872-8cc8-c7eee48e6813.png) **Figure 1.1 -** *Transfer UI flow.* ### User Flow Transfer 1. Login with registered Note Account 2. Select token asset to transfer 1. These token types will only be webb wrapped assets (e.g. webbUSDC, webbETH) 3. Enter an amount 4. Input other registered addresses 1. Validate address is registered within input box 5. Select relayer 1. Same component used on bridge 6. Send transfer **Note:** The shielded balance available will be informed by the current connected chain. For example, a user makes two deposits into the bridge for USDC, ETH where the destination chain selected was Arbitrum. When the user navigates to the Transfer tab, and is connected to the Arbitrum chain, the available shielded balance to transfer will consist of the webbUSDC, and webbETH previously deposited. However, if they are connected to Optimism and have not deposited anything into the bridge where Optimism is the destination chain the available shielded balance will be 0. ## User Selection Inputs 1. Token type 2. Amount 3. Registered address 4. Relayer selection ## Notifications - Failed transfer - Successful transfer - Invalid recipient address ## Components List - Transfer UI interface for above mentioned inputs - Successful / unsuccessful indicator ## Alternative Transfer UI’s <img src="https://user-images.githubusercontent.com/29983536/191096998-a4fa28bd-51cb-4667-81b2-e35cca12fd6a.png" height=450 /> ## Future Feature Considerations 1. Contact / address book for pre-saved registered addresses 2. Notification informing recipient of transferred funds 1. Currently we do not have anything in place that informs the user that they received funds via transfer # Open Questions
non_priority
transfer component product design goals enable users to easily transfer shielded funds from one registered address to another deliverables create wireframes that represent and organize the user actions for making a transfer to another user design a dedicated transfer component that displays required input boxes success modal and notifications create high fidelity prototype of a transfer component and flow designs and product flow figure transfer ui flow user flow transfer login with registered note account select token asset to transfer these token types will only be webb wrapped assets e g webbusdc webbeth enter an amount input other registered addresses validate address is registered within input box select relayer same component used on bridge send transfer note the shielded balance available will be informed by the current connected chain for example a user makes two deposits into the bridge for usdc eth where the destination chain selected was arbitrum when the user navigates to the transfer tab and is connected to the arbitrum chain the available shielded balance to transfer will consist of the webbusdc and webbeth previously deposited however if they are connected to optimism and have not deposited anything into the bridge where optimism is the destination chain the available shielded balance will be user selection inputs token type amount registered address relayer selection notifications failed transfer successful transfer invalid recipient address components list transfer ui interface for above mentioned inputs successful unsuccessful indicator alternative transfer ui’s future feature considerations contact address book for pre saved registered addresses notification informing recipient of transferred funds currently we do not have anything in place that informs the user that they received funds via transfer open questions
0
351,427
25,026,287,907
IssuesEvent
2022-11-04 08:20:10
Tiny-Metro/Tiny-Metro
https://api.github.com/repos/Tiny-Metro/Tiny-Metro
opened
Usage for GridBuilder
documentation
- Content/GridBuilder에 해당 콘텐츠 있음 - Blueprints의 BP_GridBuilder 및 BP_GridManager를 레벨에 배치 - 이 때, BP_GridManager를 먼저 배치해야 함 - 그리드의 사용에 관하여는 C++ Sample의 SampleGridAccess를 참조 - GridManager의 GerGridCellData함수를 통해 접근 및 참조하는 구조 **중요** - Grid의 위치는 BP_GridBuilder에 종속적 - 하지만, GridManager를 통해 접근 가능한 Grid의 좌표는 그렇지 않음 - BP_GridManager의 좌표와 무관하게 <0, 0, 0>에 대한 좌표값이 설정되어 있음 - BP_GridBuilder를 <0, 0, 0>에 맞춰놓아야 함
1.0
Usage for GridBuilder - - Content/GridBuilder에 해당 콘텐츠 있음 - Blueprints의 BP_GridBuilder 및 BP_GridManager를 레벨에 배치 - 이 때, BP_GridManager를 먼저 배치해야 함 - 그리드의 사용에 관하여는 C++ Sample의 SampleGridAccess를 참조 - GridManager의 GerGridCellData함수를 통해 접근 및 참조하는 구조 **중요** - Grid의 위치는 BP_GridBuilder에 종속적 - 하지만, GridManager를 통해 접근 가능한 Grid의 좌표는 그렇지 않음 - BP_GridManager의 좌표와 무관하게 <0, 0, 0>에 대한 좌표값이 설정되어 있음 - BP_GridBuilder를 <0, 0, 0>에 맞춰놓아야 함
non_priority
usage for gridbuilder content gridbuilder에 해당 콘텐츠 있음 blueprints의 bp gridbuilder 및 bp gridmanager를 레벨에 배치 이 때 bp gridmanager를 먼저 배치해야 함 그리드의 사용에 관하여는 c sample의 samplegridaccess를 참조 gridmanager의 gergridcelldata함수를 통해 접근 및 참조하는 구조 중요 grid의 위치는 bp gridbuilder에 종속적 하지만 gridmanager를 통해 접근 가능한 grid의 좌표는 그렇지 않음 bp gridmanager의 좌표와 무관하게 에 대한 좌표값이 설정되어 있음 bp gridbuilder를 에 맞춰놓아야 함
0
187,549
6,758,891,598
IssuesEvent
2017-10-24 15:26:26
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
games.espn.com - site is not usable
browser-firefox priority-important status-needstriage
<!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://games.espn.com/ffl/freeagency?leagueId=1862820&teamId=10&seasonId=2017#&seasonId=2017 **Browser / Version**: Firefox 58.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: won't load page **Steps to Reproduce**: layout.css.servo.enabled: true [![Screenshot Description](https://webcompat.com/uploads/2017/10/536d0e8e-7788-464b-9c84-89838a752f10-thumb.jpg)](https://webcompat.com/uploads/2017/10/536d0e8e-7788-464b-9c84-89838a752f10.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
games.espn.com - site is not usable - <!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: desktop-reporter --> **URL**: https://games.espn.com/ffl/freeagency?leagueId=1862820&teamId=10&seasonId=2017#&seasonId=2017 **Browser / Version**: Firefox 58.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Site is not usable **Description**: won't load page **Steps to Reproduce**: layout.css.servo.enabled: true [![Screenshot Description](https://webcompat.com/uploads/2017/10/536d0e8e-7788-464b-9c84-89838a752f10-thumb.jpg)](https://webcompat.com/uploads/2017/10/536d0e8e-7788-464b-9c84-89838a752f10.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
games espn com site is not usable url browser version firefox operating system windows tested another browser yes problem type site is not usable description won t load page steps to reproduce layout css servo enabled true from with ❤️
1
398,678
11,742,173,396
IssuesEvent
2020-03-11 23:51:35
thaliawww/concrexit
https://api.github.com/repos/thaliawww/concrexit
closed
API information leakage
priority: low security technical change
In GitLab by @joren485 on Jun 25, 2017, 23:12 ### One-sentence description Some private information is leaked from the database through the API. ### Why? When the ThaliApp requests events information from the website, private information is revealed. For example, the registration dates of every participant of an event and their member ids. ### Current implementation When querying the API, data is returned that is private and seemingly unnecessary. ### Desired implementation The API should only return necessary data.
1.0
API information leakage - In GitLab by @joren485 on Jun 25, 2017, 23:12 ### One-sentence description Some private information is leaked from the database through the API. ### Why? When the ThaliApp requests events information from the website, private information is revealed. For example, the registration dates of every participant of an event and their member ids. ### Current implementation When querying the API, data is returned that is private and seemingly unnecessary. ### Desired implementation The API should only return necessary data.
priority
api information leakage in gitlab by on jun one sentence description some private information is leaked from the database through the api why when the thaliapp requests events information from the website private information is revealed for example the registration dates of every participant of an event and their member ids current implementation when querying the api data is returned that is private and seemingly unnecessary desired implementation the api should only return necessary data
1
499,799
14,479,482,833
IssuesEvent
2020-12-10 09:52:18
fidelity/kconnect
https://api.github.com/repos/fidelity/kconnect
closed
Environment based authentication for EKS
area/authentication good first issue help wanted kind/feature priority/important
**Describe the solution you'd like:** I want to be able to use environment/config file credentials for my identity when discovering EKS clusters **Why do you want this feature:** **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.]
1.0
Environment based authentication for EKS - **Describe the solution you'd like:** I want to be able to use environment/config file credentials for my identity when discovering EKS clusters **Why do you want this feature:** **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.]
priority
environment based authentication for eks describe the solution you d like i want to be able to use environment config file credentials for my identity when discovering eks clusters why do you want this feature anything else you would like to add
1
43,723
11,806,529,746
IssuesEvent
2020-03-19 09:42:31
PowerDNS/pdns
https://api.github.com/repos/PowerDNS/pdns
closed
Service is not reloaded on systemd during %post on updates
auth defect dnsdist rec
### Short description <!-- Explain in a few sentences what the issue/request is --> the builder-support/specs/dnsdist.spec is using `%systemd_post` in the `%post` section, but unfortunately on several distro that macro does not enforce a reload of systemd. We should consider adding `systemctl daemon-reload` on update otherwise any change to the service file will be ignored by systemd. ``` %post ... %if 0%{?rhel} >= 7 %systemd_post %{name}.service %endif ``` Example on CentOS7 for the macro `%systemd_post` ``` > rpm --eval "%systemd_post foo" if [ $1 -eq 1 ] ; then # Initial installation systemctl preset foo >/dev/null 2>&1 || : fi ``` ### Expected behaviour call `systemctl daemon-reload` on reload
1.0
Service is not reloaded on systemd during %post on updates - ### Short description <!-- Explain in a few sentences what the issue/request is --> the builder-support/specs/dnsdist.spec is using `%systemd_post` in the `%post` section, but unfortunately on several distro that macro does not enforce a reload of systemd. We should consider adding `systemctl daemon-reload` on update otherwise any change to the service file will be ignored by systemd. ``` %post ... %if 0%{?rhel} >= 7 %systemd_post %{name}.service %endif ``` Example on CentOS7 for the macro `%systemd_post` ``` > rpm --eval "%systemd_post foo" if [ $1 -eq 1 ] ; then # Initial installation systemctl preset foo >/dev/null 2>&1 || : fi ``` ### Expected behaviour call `systemctl daemon-reload` on reload
non_priority
service is not reloaded on systemd during post on updates short description the builder support specs dnsdist spec is using systemd post in the post section but unfortunately on several distro that macro does not enforce a reload of systemd we should consider adding systemctl daemon reload on update otherwise any change to the service file will be ignored by systemd post if rhel systemd post name service endif example on for the macro systemd post rpm eval systemd post foo if then initial installation systemctl preset foo dev null fi expected behaviour call systemctl daemon reload on reload
0
66,765
14,798,952,362
IssuesEvent
2021-01-13 01:03:59
jgeraigery/argo
https://api.github.com/repos/jgeraigery/argo
opened
CVE-2020-36067 (High) detected in github.com/tidwall/gjson-v1.6.1
security vulnerability
## CVE-2020-36067 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/tidwall/gjson-v1.6.1</b></p></summary> <p>Get JSON values quickly - JSON parser for Go</p> <p> Dependency Hierarchy: - :x: **github.com/tidwall/gjson-v1.6.1** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> GJSON <=v1.6.5 allows attackers to cause a denial of service (panic: runtime error: slice bounds out of range) via a crafted GET call. <p>Publish Date: 2021-01-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36067>CVE-2020-36067</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36067">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36067</a></p> <p>Release Date: 2021-01-05</p> <p>Fix Resolution: v1.6.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/tidwall/gjson","packageVersion":"v1.6.1","isTransitiveDependency":false,"dependencyTree":"github.com/tidwall/gjson:v1.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.6.6"}],"vulnerabilityIdentifier":"CVE-2020-36067","vulnerabilityDetails":"GJSON \u003c\u003dv1.6.5 allows attackers to cause a denial of service (panic: runtime error: slice bounds out of range) via a crafted GET call.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36067","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-36067 (High) detected in github.com/tidwall/gjson-v1.6.1 - ## CVE-2020-36067 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>github.com/tidwall/gjson-v1.6.1</b></p></summary> <p>Get JSON values quickly - JSON parser for Go</p> <p> Dependency Hierarchy: - :x: **github.com/tidwall/gjson-v1.6.1** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> GJSON <=v1.6.5 allows attackers to cause a denial of service (panic: runtime error: slice bounds out of range) via a crafted GET call. <p>Publish Date: 2021-01-05 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36067>CVE-2020-36067</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36067">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36067</a></p> <p>Release Date: 2021-01-05</p> <p>Fix Resolution: v1.6.6</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"GO","packageName":"github.com/tidwall/gjson","packageVersion":"v1.6.1","isTransitiveDependency":false,"dependencyTree":"github.com/tidwall/gjson:v1.6.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.6.6"}],"vulnerabilityIdentifier":"CVE-2020-36067","vulnerabilityDetails":"GJSON \u003c\u003dv1.6.5 allows attackers to cause a denial of service (panic: runtime error: slice bounds out of range) via a crafted GET call.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36067","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in github com tidwall gjson cve high severity vulnerability vulnerable library github com tidwall gjson get json values quickly json parser for go dependency hierarchy x github com tidwall gjson vulnerable library found in base branch master vulnerability details gjson allows attackers to cause a denial of service panic runtime error slice bounds out of range via a crafted get call publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails gjson allows attackers to cause a denial of service panic runtime error slice bounds out of range via a crafted get call vulnerabilityurl
0
733,613
25,314,210,422
IssuesEvent
2022-11-17 20:05:23
red-hat-storage/ocs-ci
https://api.github.com/repos/red-hat-storage/ocs-ci
closed
Memory does not free between tests
bug High Priority performance lifecycle/stale
When we are running test suits of several tests (multiple tests classes and multiple tests in a class)the memory which consume by a specific test doesn't getting free in the end of the test, so, with a long suite (e.g. performance - 70 tests) the total memory consumption become huge (more then 5 GiB)
1.0
Memory does not free between tests - When we are running test suits of several tests (multiple tests classes and multiple tests in a class)the memory which consume by a specific test doesn't getting free in the end of the test, so, with a long suite (e.g. performance - 70 tests) the total memory consumption become huge (more then 5 GiB)
priority
memory does not free between tests when we are running test suits of several tests multiple tests classes and multiple tests in a class the memory which consume by a specific test doesn t getting free in the end of the test so with a long suite e g performance tests the total memory consumption become huge more then gib
1
367,571
10,855,288,160
IssuesEvent
2019-11-13 18:03:39
ntop/ntopng
https://api.github.com/repos/ntop/ntopng
closed
Slow HT Update on Raspberry
low-priority bug priority ticket
03/Nov/2019 17:11:33 [NetworkInterface.cpp:2933] ERROR: Deadline exceeded [periodicHTStateUpdate][lo][expected: 17:11:16][off by: 17 secs] 03/Nov/2019 17:11:48 [NetworkInterface.cpp:2933] ERROR: Deadline exceeded [periodicHTStateUpdate][eth0][expected: 17:11:31][off by: 17 secs]
2.0
Slow HT Update on Raspberry - 03/Nov/2019 17:11:33 [NetworkInterface.cpp:2933] ERROR: Deadline exceeded [periodicHTStateUpdate][lo][expected: 17:11:16][off by: 17 secs] 03/Nov/2019 17:11:48 [NetworkInterface.cpp:2933] ERROR: Deadline exceeded [periodicHTStateUpdate][eth0][expected: 17:11:31][off by: 17 secs]
priority
slow ht update on raspberry nov error deadline exceeded nov error deadline exceeded
1
33,791
2,772,530,494
IssuesEvent
2015-05-02 19:21:56
GoogleCloudPlatform/kubernetes
https://api.github.com/repos/GoogleCloudPlatform/kubernetes
closed
`kubectl describe` should be using resource.Builder so you can describe multiple things
area/usability component/CLI priority/P2 team/UX
For debugging it would be useful to be able to $ kubectl describe pods --all or $ kubectl describe pods,services -s mylabel=value Just needs to support resource.Builder
1.0
`kubectl describe` should be using resource.Builder so you can describe multiple things - For debugging it would be useful to be able to $ kubectl describe pods --all or $ kubectl describe pods,services -s mylabel=value Just needs to support resource.Builder
priority
kubectl describe should be using resource builder so you can describe multiple things for debugging it would be useful to be able to kubectl describe pods all or kubectl describe pods services s mylabel value just needs to support resource builder
1
320,641
9,784,177,375
IssuesEvent
2019-06-08 16:53:49
RobotLocomotion/drake
https://api.github.com/repos/RobotLocomotion/drake
closed
Need to Add RoadGeometry::FindLane(const std::string& name).
priority: medium team: automotive type: feature request
And remove it from `AutomotiveSimulator`. See: https://reviewable.io/reviews/robotlocomotion/drake/5699#-Kgok7MShFFHKbN4xISv
1.0
Need to Add RoadGeometry::FindLane(const std::string& name). - And remove it from `AutomotiveSimulator`. See: https://reviewable.io/reviews/robotlocomotion/drake/5699#-Kgok7MShFFHKbN4xISv
priority
need to add roadgeometry findlane const std string name and remove it from automotivesimulator see
1
323,697
23,961,682,866
IssuesEvent
2022-09-12 19:43:48
pluralsight/tva
https://api.github.com/repos/pluralsight/tva
closed
[Docs?]: Host PR Builds on Netlify
documentation
### Latest version - [X] I have checked the latest version ### Summary 💡 Codesandbox doesn't support building anything outside of basic settings so we'll use Netflify (free) to host PR builds when there is a change to the docs site. ### Motivation 🔦 _No response_
1.0
[Docs?]: Host PR Builds on Netlify - ### Latest version - [X] I have checked the latest version ### Summary 💡 Codesandbox doesn't support building anything outside of basic settings so we'll use Netflify (free) to host PR builds when there is a change to the docs site. ### Motivation 🔦 _No response_
non_priority
host pr builds on netlify latest version i have checked the latest version summary 💡 codesandbox doesn t support building anything outside of basic settings so we ll use netflify free to host pr builds when there is a change to the docs site motivation 🔦 no response
0
246,096
18,821,153,486
IssuesEvent
2021-11-10 08:28:57
daviddellarossa/CM3025-virtual-reality
https://api.github.com/repos/daviddellarossa/CM3025-virtual-reality
opened
Game Design Document
documentation long-lasting issue
Fill the following sections - [ ] Game overview - [ ] Game title - [ ] Game concept - [ ] Target audience - [ ] Background - [ ] Game mechanics - [ ] Procedures - [ ] Rules - [ ] Game world - [ ] General look and feel - [ ] Levels in game mode - [ ] Levels in learning mode - [ ] Gameplay - [ ] Game modes - [ ] Game - [ ] Learning tool - [ ] Therapy tool - [ ] User interface - [ ] Visual system - [ ] HUD - [ ] Menu - [ ] Camera model - [ ] Audio, music, sound effects - [ ] Control system - [ ] Game art - [ ] Help system - [ ] References
1.0
Game Design Document - Fill the following sections - [ ] Game overview - [ ] Game title - [ ] Game concept - [ ] Target audience - [ ] Background - [ ] Game mechanics - [ ] Procedures - [ ] Rules - [ ] Game world - [ ] General look and feel - [ ] Levels in game mode - [ ] Levels in learning mode - [ ] Gameplay - [ ] Game modes - [ ] Game - [ ] Learning tool - [ ] Therapy tool - [ ] User interface - [ ] Visual system - [ ] HUD - [ ] Menu - [ ] Camera model - [ ] Audio, music, sound effects - [ ] Control system - [ ] Game art - [ ] Help system - [ ] References
non_priority
game design document fill the following sections game overview game title game concept target audience background game mechanics procedures rules game world general look and feel levels in game mode levels in learning mode gameplay game modes game learning tool therapy tool user interface visual system hud menu camera model audio music sound effects control system game art help system references
0
508,587
14,703,045,742
IssuesEvent
2021-01-04 14:30:27
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.klix.ba - design is broken
browser-fenix engine-gecko priority-important
<!-- @browser: Firefox Mobile 86.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:86.0) Gecko/86.0 Firefox/86.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64816 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.klix.ba/ **Browser / Version**: Firefox Mobile 86.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Items are overlapped **Steps to Reproduce**: Site nit working <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201225095506</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/1/20a438cc-748e-49c7-b4e6-995694bd48b1) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.klix.ba - design is broken - <!-- @browser: Firefox Mobile 86.0 --> <!-- @ua_header: Mozilla/5.0 (Android 10; Mobile; rv:86.0) Gecko/86.0 Firefox/86.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/64816 --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.klix.ba/ **Browser / Version**: Firefox Mobile 86.0 **Operating System**: Android **Tested Another Browser**: Yes Chrome **Problem type**: Design is broken **Description**: Items are overlapped **Steps to Reproduce**: Site nit working <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201225095506</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/1/20a438cc-748e-49c7-b4e6-995694bd48b1) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
design is broken url browser version firefox mobile operating system android tested another browser yes chrome problem type design is broken description items are overlapped steps to reproduce site nit working browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
1
44,319
2,903,535,076
IssuesEvent
2015-06-18 13:55:48
greenlion/PHP-SQL-Parser
https://api.github.com/repos/greenlion/PHP-SQL-Parser
closed
Problem with parser and datatables
bug imported Priority-Medium
_From [thalesor...@gmail.com](https://code.google.com/u/103561826670313614456/) on June 17, 2015 22:51:29_ as you can see here: http://stackoverflow.com/questions/30732590/need-advice-to-make-php-server-side-datatables-work-properly00 , i have a huge problem going on _Original issue: http://code.google.com/p/php-sql-parser/issues/detail?id=159_
1.0
Problem with parser and datatables - _From [thalesor...@gmail.com](https://code.google.com/u/103561826670313614456/) on June 17, 2015 22:51:29_ as you can see here: http://stackoverflow.com/questions/30732590/need-advice-to-make-php-server-side-datatables-work-properly00 , i have a huge problem going on _Original issue: http://code.google.com/p/php-sql-parser/issues/detail?id=159_
priority
problem with parser and datatables from on june as you can see here i have a huge problem going on original issue
1
207,682
7,132,482,520
IssuesEvent
2018-01-22 14:45:25
pmem/issues
https://api.github.com/repos/pmem/issues
opened
test: vmem_check/TEST0: SETUP (all/none/debug/helgrind) fails
Exposure: Low OS: Linux Priority: 4 low Type: Bug
Found on 1a8b9ce8125852cb50cbdcc65faa9aa1172daa68 > vmem_check/TEST0: SETUP (all/none/debug/helgrind) > vmem_check/TEST0 failed with Valgrind. See helgrind0.log. First 20 lines below. > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E5116D: malloc_init_base_pool (jemalloc.c:434) > vmem_check/TEST0 helgrind0.log ==11493== Address 0x50a7880 is 0 bytes inside data symbol "je_vmem_je_pools_lock" > vmem_check/TEST0 helgrind0.log ==11493== > vmem_check/TEST0 helgrind0.log ==11493== Lock at 0x64022F8 was first observed > vmem_check/TEST0 helgrind0.log ==11493== at 0x4C37CBE: pthread_mutex_init (hg_intercepts.c:779) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E7ECAB: je_vmem_je_malloc_mutex_init (mutex.c:116) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E6B291: pool_boot (pool.c:16) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E6B406: je_vmem_je_pool_new (pool.c:72) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E55037: pool_create_empty (jemalloc.c:1664) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E55646: je_vmem_pool_create (jemalloc.c:1795) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E47B63: vmem_create_in_region (vmem.c:289) > vmem_check/TEST0 helgrind0.log ==11493== by 0x109B3F: main (vmem_check.c:60) > vmem_check/TEST0 helgrind0.log ==11493== Address 0x64022f8 is in a rw- anonymous segment > vmem_check/TEST0 helgrind0.log ==11493== > vmem_check/TEST0 helgrind0.log ==11493== > vmem_check/TEST0 helgrind0.log ==11493== > vmem_check/TEST0 helgrind0.log ==11493== For counts of detected and suppressed errors, rerun with: -v > vmem_check/TEST0 helgrind0.log ==11493== Use --history-level=approx or =none to gain increased speed, at > vmem_check/TEST0 helgrind0.log ==11493== the cost of reduced accuracy of conflicting-access information > vmem_check/TEST0 helgrind0.log ==11493== ERROR SUMMARY: 3 errors from 1 contexts (suppressed: 0 from 0) > RUNTESTS: stopping: vmem_check/TEST0 failed, TEST=all FS=none BUILD=debug > helgrind0.log > ==11493== Helgrind, a thread error detector > ==11493== Copyright (C) 2007-2015, and GNU GPL'd, by OpenWorks LLP et al. > ==11493== Using Valgrind-3.12.0 and LibVEX; rerun with -h for copyright info > ==11493== Command: ./vmem_check > ==11493== Parent PID: 11468 > ==11493== > ==11493== ---Thread-Announcement------------------------------------------ > ==11493== > ==11493== Thread #1 is the program's root thread > ==11493== > ==11493== ---------------------------------------------------------------- > ==11493== > ==11493== Thread #1: lock order "0x50A7880 before 0x64022F8" violated > ==11493== > ==11493== Observed (incorrect) order is: acquisition of lock at 0x64022F8 > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E55C68: je_vmem_pool_check (jemalloc.c:1986) > ==11493== by 0x4E47E08: vmem_check (vmem.c:359) > ==11493== by 0x109BBA: main (vmem_check.c:69) > ==11493== > ==11493== followed by a later acquisition of lock at 0x50A7880 > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E55D4F: je_vmem_pool_check (jemalloc.c:2010) > ==11493== by 0x4E47E08: vmem_check (vmem.c:359) > ==11493== by 0x109BBA: main (vmem_check.c:69) > ==11493== > ==11493== Required order was established by acquisition of lock at 0x50A7880 > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E55491: je_vmem_pool_create (jemalloc.c:1757) > ==11493== by 0x4E47B63: vmem_create_in_region (vmem.c:289) > ==11493== by 0x109B3F: main (vmem_check.c:60) > ==11493== > ==11493== followed by a later acquisition of lock at 0x64022F8 > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E55259: pool_create_empty (jemalloc.c:1699) > ==11493== by 0x4E55646: je_vmem_pool_create (jemalloc.c:1795) > ==11493== by 0x4E47B63: vmem_create_in_region (vmem.c:289) > ==11493== by 0x109B3F: main (vmem_check.c:60) > ==11493== > ==11493== Lock at 0x50A7880 was first observed > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E893F8: je_vmem_je_tcache_enabled_set (tcache.h:210) > ==11493== by 0x4E8ABCE: je_vmem_je_tcache_get_hard (tcache.c:289) > ==11493== by 0x4E89627: je_vmem_je_tcache_get (tcache.h:265) > ==11493== by 0x4E5BE0B: je_vmem_je_arena_malloc (arena.h:1102) > ==11493== by 0x4E4F4DF: je_vmem_je_imalloct (jemalloc_internal.h:713) > ==11493== by 0x4E4F54F: je_vmem_je_imalloc (jemalloc_internal.h:723) > ==11493== by 0x4E82AF6: je_vmem_je_quarantine_init (quarantine.c:31) > ==11493== by 0x4E82ABB: je_vmem_je_quarantine_alloc_hook (quarantine.h:61) > ==11493== by 0x4E5116D: malloc_init_base_pool (jemalloc.c:434) > ==11493== Address 0x50a7880 is 0 bytes inside data symbol "je_vmem_je_pools_lock" > ==11493== > ==11493== Lock at 0x64022F8 was first observed > ==11493== at 0x4C37CBE: pthread_mutex_init (hg_intercepts.c:779) > ==11493== by 0x4E7ECAB: je_vmem_je_malloc_mutex_init (mutex.c:116) > ==11493== by 0x4E6B291: pool_boot (pool.c:16) > ==11493== by 0x4E6B406: je_vmem_je_pool_new (pool.c:72) > ==11493== by 0x4E55037: pool_create_empty (jemalloc.c:1664) > ==11493== by 0x4E55646: je_vmem_pool_create (jemalloc.c:1795) > ==11493== by 0x4E47B63: vmem_create_in_region (vmem.c:289) > ==11493== by 0x109B3F: main (vmem_check.c:60) > ==11493== Address 0x64022f8 is in a rw- anonymous segment > ==11493== > ==11493== > ==11493== > ==11493== For counts of detected and suppressed errors, rerun with: -v > ==11493== Use --history-level=approx or =none to gain increased speed, at > ==11493== the cost of reduced accuracy of conflicting-access information > ==11493== ERROR SUMMARY: 3 errors from 1 contexts (suppressed: 0 from 0) >
1.0
test: vmem_check/TEST0: SETUP (all/none/debug/helgrind) fails - Found on 1a8b9ce8125852cb50cbdcc65faa9aa1172daa68 > vmem_check/TEST0: SETUP (all/none/debug/helgrind) > vmem_check/TEST0 failed with Valgrind. See helgrind0.log. First 20 lines below. > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E5116D: malloc_init_base_pool (jemalloc.c:434) > vmem_check/TEST0 helgrind0.log ==11493== Address 0x50a7880 is 0 bytes inside data symbol "je_vmem_je_pools_lock" > vmem_check/TEST0 helgrind0.log ==11493== > vmem_check/TEST0 helgrind0.log ==11493== Lock at 0x64022F8 was first observed > vmem_check/TEST0 helgrind0.log ==11493== at 0x4C37CBE: pthread_mutex_init (hg_intercepts.c:779) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E7ECAB: je_vmem_je_malloc_mutex_init (mutex.c:116) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E6B291: pool_boot (pool.c:16) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E6B406: je_vmem_je_pool_new (pool.c:72) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E55037: pool_create_empty (jemalloc.c:1664) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E55646: je_vmem_pool_create (jemalloc.c:1795) > vmem_check/TEST0 helgrind0.log ==11493== by 0x4E47B63: vmem_create_in_region (vmem.c:289) > vmem_check/TEST0 helgrind0.log ==11493== by 0x109B3F: main (vmem_check.c:60) > vmem_check/TEST0 helgrind0.log ==11493== Address 0x64022f8 is in a rw- anonymous segment > vmem_check/TEST0 helgrind0.log ==11493== > vmem_check/TEST0 helgrind0.log ==11493== > vmem_check/TEST0 helgrind0.log ==11493== > vmem_check/TEST0 helgrind0.log ==11493== For counts of detected and suppressed errors, rerun with: -v > vmem_check/TEST0 helgrind0.log ==11493== Use --history-level=approx or =none to gain increased speed, at > vmem_check/TEST0 helgrind0.log ==11493== the cost of reduced accuracy of conflicting-access information > vmem_check/TEST0 helgrind0.log ==11493== ERROR SUMMARY: 3 errors from 1 contexts (suppressed: 0 from 0) > RUNTESTS: stopping: vmem_check/TEST0 failed, TEST=all FS=none BUILD=debug > helgrind0.log > ==11493== Helgrind, a thread error detector > ==11493== Copyright (C) 2007-2015, and GNU GPL'd, by OpenWorks LLP et al. > ==11493== Using Valgrind-3.12.0 and LibVEX; rerun with -h for copyright info > ==11493== Command: ./vmem_check > ==11493== Parent PID: 11468 > ==11493== > ==11493== ---Thread-Announcement------------------------------------------ > ==11493== > ==11493== Thread #1 is the program's root thread > ==11493== > ==11493== ---------------------------------------------------------------- > ==11493== > ==11493== Thread #1: lock order "0x50A7880 before 0x64022F8" violated > ==11493== > ==11493== Observed (incorrect) order is: acquisition of lock at 0x64022F8 > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E55C68: je_vmem_pool_check (jemalloc.c:1986) > ==11493== by 0x4E47E08: vmem_check (vmem.c:359) > ==11493== by 0x109BBA: main (vmem_check.c:69) > ==11493== > ==11493== followed by a later acquisition of lock at 0x50A7880 > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E55D4F: je_vmem_pool_check (jemalloc.c:2010) > ==11493== by 0x4E47E08: vmem_check (vmem.c:359) > ==11493== by 0x109BBA: main (vmem_check.c:69) > ==11493== > ==11493== Required order was established by acquisition of lock at 0x50A7880 > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E55491: je_vmem_pool_create (jemalloc.c:1757) > ==11493== by 0x4E47B63: vmem_create_in_region (vmem.c:289) > ==11493== by 0x109B3F: main (vmem_check.c:60) > ==11493== > ==11493== followed by a later acquisition of lock at 0x64022F8 > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E55259: pool_create_empty (jemalloc.c:1699) > ==11493== by 0x4E55646: je_vmem_pool_create (jemalloc.c:1795) > ==11493== by 0x4E47B63: vmem_create_in_region (vmem.c:289) > ==11493== by 0x109B3F: main (vmem_check.c:60) > ==11493== > ==11493== Lock at 0x50A7880 was first observed > ==11493== at 0x4C33E7D: mutex_lock_WRK (hg_intercepts.c:901) > ==11493== by 0x4C37CE1: pthread_mutex_lock (hg_intercepts.c:917) > ==11493== by 0x4E7EBCB: je_vmem_je_malloc_mutex_lock (mutex.h:107) > ==11493== by 0x4E893F8: je_vmem_je_tcache_enabled_set (tcache.h:210) > ==11493== by 0x4E8ABCE: je_vmem_je_tcache_get_hard (tcache.c:289) > ==11493== by 0x4E89627: je_vmem_je_tcache_get (tcache.h:265) > ==11493== by 0x4E5BE0B: je_vmem_je_arena_malloc (arena.h:1102) > ==11493== by 0x4E4F4DF: je_vmem_je_imalloct (jemalloc_internal.h:713) > ==11493== by 0x4E4F54F: je_vmem_je_imalloc (jemalloc_internal.h:723) > ==11493== by 0x4E82AF6: je_vmem_je_quarantine_init (quarantine.c:31) > ==11493== by 0x4E82ABB: je_vmem_je_quarantine_alloc_hook (quarantine.h:61) > ==11493== by 0x4E5116D: malloc_init_base_pool (jemalloc.c:434) > ==11493== Address 0x50a7880 is 0 bytes inside data symbol "je_vmem_je_pools_lock" > ==11493== > ==11493== Lock at 0x64022F8 was first observed > ==11493== at 0x4C37CBE: pthread_mutex_init (hg_intercepts.c:779) > ==11493== by 0x4E7ECAB: je_vmem_je_malloc_mutex_init (mutex.c:116) > ==11493== by 0x4E6B291: pool_boot (pool.c:16) > ==11493== by 0x4E6B406: je_vmem_je_pool_new (pool.c:72) > ==11493== by 0x4E55037: pool_create_empty (jemalloc.c:1664) > ==11493== by 0x4E55646: je_vmem_pool_create (jemalloc.c:1795) > ==11493== by 0x4E47B63: vmem_create_in_region (vmem.c:289) > ==11493== by 0x109B3F: main (vmem_check.c:60) > ==11493== Address 0x64022f8 is in a rw- anonymous segment > ==11493== > ==11493== > ==11493== > ==11493== For counts of detected and suppressed errors, rerun with: -v > ==11493== Use --history-level=approx or =none to gain increased speed, at > ==11493== the cost of reduced accuracy of conflicting-access information > ==11493== ERROR SUMMARY: 3 errors from 1 contexts (suppressed: 0 from 0) >
priority
test vmem check setup all none debug helgrind fails found on vmem check setup all none debug helgrind vmem check failed with valgrind see log first lines below vmem check log by malloc init base pool jemalloc c vmem check log address is bytes inside data symbol je vmem je pools lock vmem check log vmem check log lock at was first observed vmem check log at pthread mutex init hg intercepts c vmem check log by je vmem je malloc mutex init mutex c vmem check log by pool boot pool c vmem check log by je vmem je pool new pool c vmem check log by pool create empty jemalloc c vmem check log by je vmem pool create jemalloc c vmem check log by vmem create in region vmem c vmem check log by main vmem check c vmem check log address is in a rw anonymous segment vmem check log vmem check log vmem check log vmem check log for counts of detected and suppressed errors rerun with v vmem check log use history level approx or none to gain increased speed at vmem check log the cost of reduced accuracy of conflicting access information vmem check log error summary errors from contexts suppressed from runtests stopping vmem check failed test all fs none build debug log helgrind a thread error detector copyright c and gnu gpl d by openworks llp et al using valgrind and libvex rerun with h for copyright info command vmem check parent pid thread announcement thread is the program s root thread thread lock order before violated observed incorrect order is acquisition of lock at at mutex lock wrk hg intercepts c by pthread mutex lock hg intercepts c by je vmem je malloc mutex lock mutex h by je vmem pool check jemalloc c by vmem check vmem c by main vmem check c followed by a later acquisition of lock at at mutex lock wrk hg intercepts c by pthread mutex lock hg intercepts c by je vmem je malloc mutex lock mutex h by je vmem pool check jemalloc c by vmem check vmem c by main vmem check c required order was established by acquisition of lock at at mutex lock wrk hg intercepts c by pthread mutex lock hg intercepts c by je vmem je malloc mutex lock mutex h by je vmem pool create jemalloc c by vmem create in region vmem c by main vmem check c followed by a later acquisition of lock at at mutex lock wrk hg intercepts c by pthread mutex lock hg intercepts c by je vmem je malloc mutex lock mutex h by pool create empty jemalloc c by je vmem pool create jemalloc c by vmem create in region vmem c by main vmem check c lock at was first observed at mutex lock wrk hg intercepts c by pthread mutex lock hg intercepts c by je vmem je malloc mutex lock mutex h by je vmem je tcache enabled set tcache h by je vmem je tcache get hard tcache c by je vmem je tcache get tcache h by je vmem je arena malloc arena h by je vmem je imalloct jemalloc internal h by je vmem je imalloc jemalloc internal h by je vmem je quarantine init quarantine c by je vmem je quarantine alloc hook quarantine h by malloc init base pool jemalloc c address is bytes inside data symbol je vmem je pools lock lock at was first observed at pthread mutex init hg intercepts c by je vmem je malloc mutex init mutex c by pool boot pool c by je vmem je pool new pool c by pool create empty jemalloc c by je vmem pool create jemalloc c by vmem create in region vmem c by main vmem check c address is in a rw anonymous segment for counts of detected and suppressed errors rerun with v use history level approx or none to gain increased speed at the cost of reduced accuracy of conflicting access information error summary errors from contexts suppressed from
1
403,758
11,846,162,433
IssuesEvent
2020-03-24 09:42:39
IlchCMS/Ilch-2.0
https://api.github.com/repos/IlchCMS/Ilch-2.0
closed
Events-Modul: Art der Veranstaltung
Priority: Medium Status: Completed Type: Enhancement
> und zwar wäre es möglich im Backend (Admincenter) im Veranstaltungsmodul ebenfalls die Auswahl Möglichkeit hinzu zu fügen, sodass man direkt im Admincenter ( im Modul) die art der Veranstaltung erstellen kann. > Damit meine ich zb , Messe, Tanzabend, Karaoke Abend, Jubiläum, usw. > Dann braucht der Ersteller der Veranstaltung nur noch das passende auszuwählen. > Im Frontend sollte dieses dann auch sichtbar sein zb = Art der Veranstaltung: Messe https://www.ilch.de/forum-showposts-55688-p1.html#398546 ToDo: - Möglichkeit im Admincenter bieten Arten für Veranstaltungen zu erstellen, welche dann beim Erstellen einer Veranstaltung ausgewählt werden können. - Die Art der Veranstaltung sollte dann natürlich im Frontend angezeigt werden. - Filtermöglichkeit im Frontend: Der Besucher kann sich z.B. alle zukünftigen Messen anzeigen lassen. - ...
1.0
Events-Modul: Art der Veranstaltung - > und zwar wäre es möglich im Backend (Admincenter) im Veranstaltungsmodul ebenfalls die Auswahl Möglichkeit hinzu zu fügen, sodass man direkt im Admincenter ( im Modul) die art der Veranstaltung erstellen kann. > Damit meine ich zb , Messe, Tanzabend, Karaoke Abend, Jubiläum, usw. > Dann braucht der Ersteller der Veranstaltung nur noch das passende auszuwählen. > Im Frontend sollte dieses dann auch sichtbar sein zb = Art der Veranstaltung: Messe https://www.ilch.de/forum-showposts-55688-p1.html#398546 ToDo: - Möglichkeit im Admincenter bieten Arten für Veranstaltungen zu erstellen, welche dann beim Erstellen einer Veranstaltung ausgewählt werden können. - Die Art der Veranstaltung sollte dann natürlich im Frontend angezeigt werden. - Filtermöglichkeit im Frontend: Der Besucher kann sich z.B. alle zukünftigen Messen anzeigen lassen. - ...
priority
events modul art der veranstaltung und zwar wäre es möglich im backend admincenter im veranstaltungsmodul ebenfalls die auswahl möglichkeit hinzu zu fügen sodass man direkt im admincenter im modul die art der veranstaltung erstellen kann damit meine ich zb messe tanzabend karaoke abend jubiläum usw dann braucht der ersteller der veranstaltung nur noch das passende auszuwählen im frontend sollte dieses dann auch sichtbar sein zb art der veranstaltung messe todo möglichkeit im admincenter bieten arten für veranstaltungen zu erstellen welche dann beim erstellen einer veranstaltung ausgewählt werden können die art der veranstaltung sollte dann natürlich im frontend angezeigt werden filtermöglichkeit im frontend der besucher kann sich z b alle zukünftigen messen anzeigen lassen
1
741,900
25,827,438,212
IssuesEvent
2022-12-12 13:56:34
unfoldingWord/tc-create-app
https://api.github.com/repos/unfoldingWord/tc-create-app
closed
Pasting inserts line breaks that can't be removed immediately
bug Field Reported Priority/High
~~**Note**: This needs to be tested after Preview editing has been disabled. #871~~ This was tested and it is a problem in raw markdown mode. tC Create has the following behavior that should be removed: when something is pasted into the edit field (from anywhere), it adds line breaks. In order to remove those line breaks, the user has to click out of the field and then back in again, then can proceed to delete the line breaks. That adds a lot of time to the edit process. Doing a 'unformatted paste' does not add the line break. DoD: --- Pasting in text does not add a line break unless the line break is in the clipboard. Details: --- From https://unfoldingword.zulipchat.com/#narrow/stream/207526-tools/topic/tC.20Create.20App.20.28GLtS.29/near/236361042
1.0
Pasting inserts line breaks that can't be removed immediately - ~~**Note**: This needs to be tested after Preview editing has been disabled. #871~~ This was tested and it is a problem in raw markdown mode. tC Create has the following behavior that should be removed: when something is pasted into the edit field (from anywhere), it adds line breaks. In order to remove those line breaks, the user has to click out of the field and then back in again, then can proceed to delete the line breaks. That adds a lot of time to the edit process. Doing a 'unformatted paste' does not add the line break. DoD: --- Pasting in text does not add a line break unless the line break is in the clipboard. Details: --- From https://unfoldingword.zulipchat.com/#narrow/stream/207526-tools/topic/tC.20Create.20App.20.28GLtS.29/near/236361042
priority
pasting inserts line breaks that can t be removed immediately note this needs to be tested after preview editing has been disabled this was tested and it is a problem in raw markdown mode tc create has the following behavior that should be removed when something is pasted into the edit field from anywhere it adds line breaks in order to remove those line breaks the user has to click out of the field and then back in again then can proceed to delete the line breaks that adds a lot of time to the edit process doing a unformatted paste does not add the line break dod pasting in text does not add a line break unless the line break is in the clipboard details from
1
418,521
12,198,941,293
IssuesEvent
2020-04-30 00:10:48
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
GlassFish submits wrong Client certificate and throws bad_certificate SSL error from Webservice
Axis Component: security ERR: Assignee Priority: Major Stale Type: Bug bad_certificate certificate chooseClientAlias client handshake https s1as ssl
### <a name="Background"></a>Background This problem happens especially in GFv3.x as the following JVM options is added -Dcom.sun.enterprise.security.httpsOutboundKeyAlias=s1as The impact of this is that it sets the default client cert to be sent is s1as. Although this option is there is GFv2, it was not enabled in the JVM options. So no issues unless one set it. Now the internals of GFv3 opensource suggest that on startup 1) GFv301 will replace the HttpsURLConnection setDefault to have a KeyManager that will sent "httpsOutboundKeyAlias"/s1as is client cert is requested. 2) Gfv312 also does this 3) Now in GFv301, it does not set the SSLContext but in GFv312x due to change #15369, it seems that SSLContext.setDefault() is called to make the SSLContext have a default key manager that will submit s1as as the client cert if requested Well the behaviour is that things that work say in GFv2 may not work in GFv301 (if URLConnection to a [https://](https://)... that does optional client cert request). For example ==> [https://www.java.net//forum/topic/glassfish/glassfish/axis2-generated-stub-soap-webservice-call-not-working-glassfish-312?force=516](https://www.java.net//forum/topic/glassfish/glassfish/axis2-generated-stub-soap-webservice-call-not-working-glassfish-312?force=516) where javax.net.ssl.SSLHandshakeException: Received fatal alert: bad_certificate may happen in GFV312 (but may work in GFv301) due to (3). ### <a name="Nowtheproblemisthis"></a>Now the problem is this The issue is that a) GFv3 set the default alias to sent the client cert globally b) In GFv312, SSL artifact created from HttpsURLConnection and SSLContext will have this KeyManager that does chooseClientAlias() from X509ExtendedKeymanager and always returns s1as irregardless if the issuer matches with s1as <ins>Look at J2EEKeyManager.java</ins> [https://java.net/projects/glassfish/sources/svn/content/tags/3.1.2/security/core/src/main/java/com/sun/enterprise/security/ssl/J2EEKeyManager.java](https://java.net/projects/glassfish/sources/svn/content/tags/3.1.2/security/core/src/main/java/com/sun/enterprise/security/ssl/J2EEKeyManager.java) As you may see, from the code, the fix/issues are: a) Not setting com.sun.enterprise.security.httpsOutboundKeyAlias may help (if you do not ever use client cert) but the code may always return null for some conditions (ie: !server&!acc) b) Even if the property is explicitly set, the chooseClientAlias should still check if this alias is compatible with what the SSL server requested otherwise it should return NULL <ins>Symptom</ins> For this issue even if you added all the SSL cert to the client trust store, if the remote SSL server ask for say an optional client cert (or required), the handshake will fail since the wrong client cert (self-signed s1as) is submitted which is totally different from the valid cert the server accepts. ``` javax.net.ssl.SSLHandshakeException: Received fatal alert: bad_certificate at sun.security.ssl.Alerts.getSSLException(Alerts.java:192) at sun.security.ssl.Alerts.getSSLException(Alerts.java:154) at sun.security.ssl.SSLSocketImpl.recvAlert(SSLSocketImpl.java:1961) at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:1077) at sun.security.ssl.SSLSocketImpl.waitForClose(SSLSocketImpl.java:1707) at sun.security.ssl.HandshakeOutStream.flush(HandshakeOutStream.java:122) at sun.security.ssl.Handshaker.sendChangeCipherSpec(Handshaker.java:972) at sun.security.ssl.ClientHandshaker.sendChangeCipherAndFinish(ClientHandshaker.java:1087) at sun.security.ssl.ClientHandshaker.serverHelloDone(ClientHandshaker.java:1006) at sun.security.ssl.ClientHandshaker.processMessage(ClientHandshaker.java:285) at sun.security.ssl.Handshaker.processLoop(Handshaker.java:868) at sun.security.ssl.Handshaker.process_record(Handshaker.java:804) at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:1016) at sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1312) at sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1339) at sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1323) at sun.net.www.protocol.https.HttpsClient.afterConnect(HttpsClient.java:515) at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:185) ``` #### Environment * GlassFish 2.1.x (if httpsOutBoundKeyAlias is set) * GFv301x * GFv312x * GFv4 #### Affected Versions [3.1.2.2]
1.0
GlassFish submits wrong Client certificate and throws bad_certificate SSL error from Webservice - ### <a name="Background"></a>Background This problem happens especially in GFv3.x as the following JVM options is added -Dcom.sun.enterprise.security.httpsOutboundKeyAlias=s1as The impact of this is that it sets the default client cert to be sent is s1as. Although this option is there is GFv2, it was not enabled in the JVM options. So no issues unless one set it. Now the internals of GFv3 opensource suggest that on startup 1) GFv301 will replace the HttpsURLConnection setDefault to have a KeyManager that will sent "httpsOutboundKeyAlias"/s1as is client cert is requested. 2) Gfv312 also does this 3) Now in GFv301, it does not set the SSLContext but in GFv312x due to change #15369, it seems that SSLContext.setDefault() is called to make the SSLContext have a default key manager that will submit s1as as the client cert if requested Well the behaviour is that things that work say in GFv2 may not work in GFv301 (if URLConnection to a [https://](https://)... that does optional client cert request). For example ==> [https://www.java.net//forum/topic/glassfish/glassfish/axis2-generated-stub-soap-webservice-call-not-working-glassfish-312?force=516](https://www.java.net//forum/topic/glassfish/glassfish/axis2-generated-stub-soap-webservice-call-not-working-glassfish-312?force=516) where javax.net.ssl.SSLHandshakeException: Received fatal alert: bad_certificate may happen in GFV312 (but may work in GFv301) due to (3). ### <a name="Nowtheproblemisthis"></a>Now the problem is this The issue is that a) GFv3 set the default alias to sent the client cert globally b) In GFv312, SSL artifact created from HttpsURLConnection and SSLContext will have this KeyManager that does chooseClientAlias() from X509ExtendedKeymanager and always returns s1as irregardless if the issuer matches with s1as <ins>Look at J2EEKeyManager.java</ins> [https://java.net/projects/glassfish/sources/svn/content/tags/3.1.2/security/core/src/main/java/com/sun/enterprise/security/ssl/J2EEKeyManager.java](https://java.net/projects/glassfish/sources/svn/content/tags/3.1.2/security/core/src/main/java/com/sun/enterprise/security/ssl/J2EEKeyManager.java) As you may see, from the code, the fix/issues are: a) Not setting com.sun.enterprise.security.httpsOutboundKeyAlias may help (if you do not ever use client cert) but the code may always return null for some conditions (ie: !server&!acc) b) Even if the property is explicitly set, the chooseClientAlias should still check if this alias is compatible with what the SSL server requested otherwise it should return NULL <ins>Symptom</ins> For this issue even if you added all the SSL cert to the client trust store, if the remote SSL server ask for say an optional client cert (or required), the handshake will fail since the wrong client cert (self-signed s1as) is submitted which is totally different from the valid cert the server accepts. ``` javax.net.ssl.SSLHandshakeException: Received fatal alert: bad_certificate at sun.security.ssl.Alerts.getSSLException(Alerts.java:192) at sun.security.ssl.Alerts.getSSLException(Alerts.java:154) at sun.security.ssl.SSLSocketImpl.recvAlert(SSLSocketImpl.java:1961) at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:1077) at sun.security.ssl.SSLSocketImpl.waitForClose(SSLSocketImpl.java:1707) at sun.security.ssl.HandshakeOutStream.flush(HandshakeOutStream.java:122) at sun.security.ssl.Handshaker.sendChangeCipherSpec(Handshaker.java:972) at sun.security.ssl.ClientHandshaker.sendChangeCipherAndFinish(ClientHandshaker.java:1087) at sun.security.ssl.ClientHandshaker.serverHelloDone(ClientHandshaker.java:1006) at sun.security.ssl.ClientHandshaker.processMessage(ClientHandshaker.java:285) at sun.security.ssl.Handshaker.processLoop(Handshaker.java:868) at sun.security.ssl.Handshaker.process_record(Handshaker.java:804) at sun.security.ssl.SSLSocketImpl.readRecord(SSLSocketImpl.java:1016) at sun.security.ssl.SSLSocketImpl.performInitialHandshake(SSLSocketImpl.java:1312) at sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1339) at sun.security.ssl.SSLSocketImpl.startHandshake(SSLSocketImpl.java:1323) at sun.net.www.protocol.https.HttpsClient.afterConnect(HttpsClient.java:515) at sun.net.www.protocol.https.AbstractDelegateHttpsURLConnection.connect(AbstractDelegateHttpsURLConnection.java:185) ``` #### Environment * GlassFish 2.1.x (if httpsOutBoundKeyAlias is set) * GFv301x * GFv312x * GFv4 #### Affected Versions [3.1.2.2]
priority
glassfish submits wrong client certificate and throws bad certificate ssl error from webservice background this problem happens especially in x as the following jvm options is added dcom sun enterprise security httpsoutboundkeyalias the impact of this is that it sets the default client cert to be sent is although this option is there is it was not enabled in the jvm options so no issues unless one set it now the internals of opensource suggest that on startup will replace the httpsurlconnection setdefault to have a keymanager that will sent httpsoutboundkeyalias is client cert is requested also does this now in it does not set the sslcontext but in due to change it seems that sslcontext setdefault is called to make the sslcontext have a default key manager that will submit as the client cert if requested well the behaviour is that things that work say in may not work in if urlconnection to a that does optional client cert request for example where javax net ssl sslhandshakeexception received fatal alert bad certificate may happen in but may work in due to now the problem is this the issue is that a set the default alias to sent the client cert globally b in ssl artifact created from httpsurlconnection and sslcontext will have this keymanager that does chooseclientalias from and always returns irregardless if the issuer matches with look at java as you may see from the code the fix issues are a not setting com sun enterprise security httpsoutboundkeyalias may help if you do not ever use client cert but the code may always return null for some conditions ie server acc b even if the property is explicitly set the chooseclientalias should still check if this alias is compatible with what the ssl server requested otherwise it should return null symptom for this issue even if you added all the ssl cert to the client trust store if the remote ssl server ask for say an optional client cert or required the handshake will fail since the wrong client cert self signed is submitted which is totally different from the valid cert the server accepts javax net ssl sslhandshakeexception received fatal alert bad certificate at sun security ssl alerts getsslexception alerts java at sun security ssl alerts getsslexception alerts java at sun security ssl sslsocketimpl recvalert sslsocketimpl java at sun security ssl sslsocketimpl readrecord sslsocketimpl java at sun security ssl sslsocketimpl waitforclose sslsocketimpl java at sun security ssl handshakeoutstream flush handshakeoutstream java at sun security ssl handshaker sendchangecipherspec handshaker java at sun security ssl clienthandshaker sendchangecipherandfinish clienthandshaker java at sun security ssl clienthandshaker serverhellodone clienthandshaker java at sun security ssl clienthandshaker processmessage clienthandshaker java at sun security ssl handshaker processloop handshaker java at sun security ssl handshaker process record handshaker java at sun security ssl sslsocketimpl readrecord sslsocketimpl java at sun security ssl sslsocketimpl performinitialhandshake sslsocketimpl java at sun security ssl sslsocketimpl starthandshake sslsocketimpl java at sun security ssl sslsocketimpl starthandshake sslsocketimpl java at sun net at sun net environment glassfish x if httpsoutboundkeyalias is set affected versions
1
12,615
14,917,403,106
IssuesEvent
2021-01-22 19:48:34
ldtteam/minecolonies
https://api.github.com/repos/ldtteam/minecolonies
closed
[Server] java.lang.IndexOutOfBoundsException: Index: 0, Size: 0 involving Listeners
Bug Compatibility: Mod
<!-- PLEASE DO NOT DELETE TOPICS AS YOUR ISSUE WILL GET CLOSED --> ### Prerequisites <!-- Replace spaces with "x" to tick boxes. --> - [x] I am running the latest **alpha** version of MineColonies and Structurize for my Minecraft version. - [X] I checked the MineColonies/Structurize [wiki](https://wiki.minecolonies.ldtteam.com/) and made sure my issue is not covered there. - [X] I made sure that **this issue is not a duplicate of any existing issue.** ### Context <!-- Exact MineColonies and Structurize versions (e.g. 0.9.126-ALPHA or 0.9.2-RELEASE) *and* the Minecraft version you're playing, please.--> - Minecraft Version: 1.6.4 - MineColonies Version: 0.13.570 - Structurize Version: 0.13.104 - Related Mods and their Versions: ## Expected behavior World gens and server is stable. ## Actual behavior World generates (as noted by spawn gen and world files presence) but then crashes. ## Steps to reproduce the issue On server, haven't tried client yet. minecolonies-0.13.570-ALPHA-universal structurize-0.13.104-ALPHA-universal May be sufficient but if not I had them present within the Monumental Experience mod pack. It may simple be the mod doesn't play well due to volume of other mods (no expert talking here). ### Logs <!-- Add your latest.log and crashlog (if you have one) to https://gist.github.com/ and put the link below. These are often important in figuring out where issues are. --> - latest.log: https://gist.githubusercontent.com/thorsanvil/2c5071247dc3d8abb728e5279bc3d0e2/raw/6c846856f2494ca37d2402f6b93ab9b3d94252a0/gistfile1.txt ### Notes <!-- Add any other context about the problem here. --> --- ## Viewers * Add a thumbs-up to the bug report if you are also affected. This helps the bug report become more visible to the team and doesn't clutter the comments. * Add a comment if you have any insights or background information that isn't already part of the conversation.
True
[Server] java.lang.IndexOutOfBoundsException: Index: 0, Size: 0 involving Listeners - <!-- PLEASE DO NOT DELETE TOPICS AS YOUR ISSUE WILL GET CLOSED --> ### Prerequisites <!-- Replace spaces with "x" to tick boxes. --> - [x] I am running the latest **alpha** version of MineColonies and Structurize for my Minecraft version. - [X] I checked the MineColonies/Structurize [wiki](https://wiki.minecolonies.ldtteam.com/) and made sure my issue is not covered there. - [X] I made sure that **this issue is not a duplicate of any existing issue.** ### Context <!-- Exact MineColonies and Structurize versions (e.g. 0.9.126-ALPHA or 0.9.2-RELEASE) *and* the Minecraft version you're playing, please.--> - Minecraft Version: 1.6.4 - MineColonies Version: 0.13.570 - Structurize Version: 0.13.104 - Related Mods and their Versions: ## Expected behavior World gens and server is stable. ## Actual behavior World generates (as noted by spawn gen and world files presence) but then crashes. ## Steps to reproduce the issue On server, haven't tried client yet. minecolonies-0.13.570-ALPHA-universal structurize-0.13.104-ALPHA-universal May be sufficient but if not I had them present within the Monumental Experience mod pack. It may simple be the mod doesn't play well due to volume of other mods (no expert talking here). ### Logs <!-- Add your latest.log and crashlog (if you have one) to https://gist.github.com/ and put the link below. These are often important in figuring out where issues are. --> - latest.log: https://gist.githubusercontent.com/thorsanvil/2c5071247dc3d8abb728e5279bc3d0e2/raw/6c846856f2494ca37d2402f6b93ab9b3d94252a0/gistfile1.txt ### Notes <!-- Add any other context about the problem here. --> --- ## Viewers * Add a thumbs-up to the bug report if you are also affected. This helps the bug report become more visible to the team and doesn't clutter the comments. * Add a comment if you have any insights or background information that isn't already part of the conversation.
non_priority
java lang indexoutofboundsexception index size involving listeners prerequisites i am running the latest alpha version of minecolonies and structurize for my minecraft version i checked the minecolonies structurize and made sure my issue is not covered there i made sure that this issue is not a duplicate of any existing issue context minecraft version minecolonies version structurize version related mods and their versions expected behavior world gens and server is stable actual behavior world generates as noted by spawn gen and world files presence but then crashes steps to reproduce the issue on server haven t tried client yet minecolonies alpha universal structurize alpha universal may be sufficient but if not i had them present within the monumental experience mod pack it may simple be the mod doesn t play well due to volume of other mods no expert talking here logs latest log notes viewers add a thumbs up to the bug report if you are also affected this helps the bug report become more visible to the team and doesn t clutter the comments add a comment if you have any insights or background information that isn t already part of the conversation
0
567,590
16,887,113,360
IssuesEvent
2021-06-23 02:43:19
leighmacdonald/uncletopia
https://api.github.com/repos/leighmacdonald/uncletopia
closed
Please reset the Seattle player limit to 24
normal priority
**A 32 player limit creates chaos and lowers the framerate of the game. It feels too crowded and hurts my latency. It makes it less likely for a player to get their preferred class and makes certain classes harder/easier to play.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] **I would prefer a 24 player limit.** A clear and concise description of what you want to happen.
1.0
Please reset the Seattle player limit to 24 - **A 32 player limit creates chaos and lowers the framerate of the game. It feels too crowded and hurts my latency. It makes it less likely for a player to get their preferred class and makes certain classes harder/easier to play.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] **I would prefer a 24 player limit.** A clear and concise description of what you want to happen.
priority
please reset the seattle player limit to a player limit creates chaos and lowers the framerate of the game it feels too crowded and hurts my latency it makes it less likely for a player to get their preferred class and makes certain classes harder easier to play a clear and concise description of what the problem is ex i m always frustrated when i would prefer a player limit a clear and concise description of what you want to happen
1
31,791
2,736,875,958
IssuesEvent
2015-04-19 21:03:43
x360ce/x360ce
https://api.github.com/repos/x360ce/x360ce
closed
[Saitek Cyborg Rumble Pad] No vibrations at all in Prince of Persia 2008
imported Priority-Low Type-Other wontfix x360ce-Core
_From [Veila...@gmail.com](https://code.google.com/u/117785195174804742025/) on May 06, 2010 12:11:08_ What steps will reproduce the problem? 1. Run the game 2. No vibrations at all [should appear for example when jumping from bar/stick/timber [whatever :] to another bar]. What is the expected output? What do you see instead? Expected: Vibrations Instead: No vibrations What version of the product are you using? On what operating system? 2.0.2.51; Windows 7 64bit What gamepad are you using? On what version of drivers? Saitek Cyborg Rumble Pad. 6.0.5.11 Of course vibrations work okay in Saitek properties "check tool". Please set third label to Core if issue affect emulator (dll file) or App if issue affect x360ce.App application (exe file). Also if you want you can provide any additional information below. Vibrations have worked with older version of xinput test, but then havent' worked right analog stick. Problem seems to lie within dll. Despite that - everything is working flawlessly at last... just... no vibrations. _Original issue: http://code.google.com/p/x360ce/issues/detail?id=15_
1.0
[Saitek Cyborg Rumble Pad] No vibrations at all in Prince of Persia 2008 - _From [Veila...@gmail.com](https://code.google.com/u/117785195174804742025/) on May 06, 2010 12:11:08_ What steps will reproduce the problem? 1. Run the game 2. No vibrations at all [should appear for example when jumping from bar/stick/timber [whatever :] to another bar]. What is the expected output? What do you see instead? Expected: Vibrations Instead: No vibrations What version of the product are you using? On what operating system? 2.0.2.51; Windows 7 64bit What gamepad are you using? On what version of drivers? Saitek Cyborg Rumble Pad. 6.0.5.11 Of course vibrations work okay in Saitek properties "check tool". Please set third label to Core if issue affect emulator (dll file) or App if issue affect x360ce.App application (exe file). Also if you want you can provide any additional information below. Vibrations have worked with older version of xinput test, but then havent' worked right analog stick. Problem seems to lie within dll. Despite that - everything is working flawlessly at last... just... no vibrations. _Original issue: http://code.google.com/p/x360ce/issues/detail?id=15_
priority
no vibrations at all in prince of persia from on may what steps will reproduce the problem run the game no vibrations at all should appear for example when jumping from bar stick timber to another bar what is the expected output what do you see instead expected vibrations instead no vibrations what version of the product are you using on what operating system windows what gamepad are you using on what version of drivers saitek cyborg rumble pad of course vibrations work okay in saitek properties check tool please set third label to core if issue affect emulator dll file or app if issue affect app application exe file also if you want you can provide any additional information below vibrations have worked with older version of xinput test but then havent worked right analog stick problem seems to lie within dll despite that everything is working flawlessly at last just no vibrations original issue
1
67,713
13,018,206,729
IssuesEvent
2020-07-26 16:12:15
EdenServer/community
https://api.github.com/repos/EdenServer/community
closed
Clear Mind V Recovery Rate
in-code-review
### Checklist <!-- Don't edit or delete this section, but tick the boxes after you have submitted your issue. If you are impatient and want to tick boxes before you submit your issue, use an x WITHOUT spaces like this: - [x] If there are unticked boxes a developer may not address the issue and/or may close the issue without comment. Make sure you comply with the checklist (as in you're not just ticking boxes you actually do the checklist) and then start writing in the details section below. --> - [x] I have searched for existing issues for issues like this one. The issue has not been posted. (Duplicate reports slow down development.) - [x] I have provided reproducable steps. (No "as the title says" posts please. Provide reproducable steps even if it seems like a waste of time.) - [ ] I have provided my client version in the details. (type /ver into your game window) ### Details https://github.com/EdenServer/community/issues/479 @tigo9 > Testing as a 75 blm right now, so Clear Mind V. > > The first tick is 25, followed by a 28, etc. This should be 27, followed by 30, etc. Source: > all revisions of the following page https://ffxiclopedia.fandom.com/wiki/Clear_Mind?oldid=378082
1.0
Clear Mind V Recovery Rate - ### Checklist <!-- Don't edit or delete this section, but tick the boxes after you have submitted your issue. If you are impatient and want to tick boxes before you submit your issue, use an x WITHOUT spaces like this: - [x] If there are unticked boxes a developer may not address the issue and/or may close the issue without comment. Make sure you comply with the checklist (as in you're not just ticking boxes you actually do the checklist) and then start writing in the details section below. --> - [x] I have searched for existing issues for issues like this one. The issue has not been posted. (Duplicate reports slow down development.) - [x] I have provided reproducable steps. (No "as the title says" posts please. Provide reproducable steps even if it seems like a waste of time.) - [ ] I have provided my client version in the details. (type /ver into your game window) ### Details https://github.com/EdenServer/community/issues/479 @tigo9 > Testing as a 75 blm right now, so Clear Mind V. > > The first tick is 25, followed by a 28, etc. This should be 27, followed by 30, etc. Source: > all revisions of the following page https://ffxiclopedia.fandom.com/wiki/Clear_Mind?oldid=378082
non_priority
clear mind v recovery rate checklist don t edit or delete this section but tick the boxes after you have submitted your issue if you are impatient and want to tick boxes before you submit your issue use an x without spaces like this if there are unticked boxes a developer may not address the issue and or may close the issue without comment make sure you comply with the checklist as in you re not just ticking boxes you actually do the checklist and then start writing in the details section below i have searched for existing issues for issues like this one the issue has not been posted duplicate reports slow down development i have provided reproducable steps no as the title says posts please provide reproducable steps even if it seems like a waste of time i have provided my client version in the details type ver into your game window details testing as a blm right now so clear mind v the first tick is followed by a etc this should be followed by etc source all revisions of the following page
0
272,812
29,795,092,318
IssuesEvent
2023-06-16 01:10:20
billmcchesney1/pacbot
https://api.github.com/repos/billmcchesney1/pacbot
closed
CVE-2022-0122 (Medium) detected in node-forge-0.10.0.tgz - autoclosed
Mend: dependency security vulnerability
## CVE-2022-0122 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p> <p>Path to dependency file: /webapp/package.json</p> <p>Path to vulnerable library: /webapp/node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - cli-1.6.8.tgz (Root Library) - webpack-dev-server-2.11.5.tgz - selfsigned-1.10.8.tgz - :x: **node-forge-0.10.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> forge is vulnerable to URL Redirection to Untrusted Site Mend Note: Converted from WS-2022-0007, on 2022-11-07. <p>Publish Date: 2022-01-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0122>CVE-2022-0122</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-gf8q-jrpm-jvxq">https://github.com/advisories/GHSA-gf8q-jrpm-jvxq</a></p> <p>Release Date: 2022-01-06</p> <p>Fix Resolution: node-forge - 1.0.0</p> </p> </details> <p></p>
True
CVE-2022-0122 (Medium) detected in node-forge-0.10.0.tgz - autoclosed - ## CVE-2022-0122 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary> <p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p> <p>Path to dependency file: /webapp/package.json</p> <p>Path to vulnerable library: /webapp/node_modules/node-forge/package.json</p> <p> Dependency Hierarchy: - cli-1.6.8.tgz (Root Library) - webpack-dev-server-2.11.5.tgz - selfsigned-1.10.8.tgz - :x: **node-forge-0.10.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/pacbot/commit/acf9a0620c1a37cee4f2896d71e1c3731c5c7b06">acf9a0620c1a37cee4f2896d71e1c3731c5c7b06</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> forge is vulnerable to URL Redirection to Untrusted Site Mend Note: Converted from WS-2022-0007, on 2022-11-07. <p>Publish Date: 2022-01-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-0122>CVE-2022-0122</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-gf8q-jrpm-jvxq">https://github.com/advisories/GHSA-gf8q-jrpm-jvxq</a></p> <p>Release Date: 2022-01-06</p> <p>Fix Resolution: node-forge - 1.0.0</p> </p> </details> <p></p>
non_priority
cve medium detected in node forge tgz autoclosed cve medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file webapp package json path to vulnerable library webapp node modules node forge package json dependency hierarchy cli tgz root library webpack dev server tgz selfsigned tgz x node forge tgz vulnerable library found in head commit a href found in base branch master vulnerability details forge is vulnerable to url redirection to untrusted site mend note converted from ws on publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge
0
776,223
27,252,126,494
IssuesEvent
2023-02-22 08:56:36
marrlab/DomainLab
https://api.github.com/repos/marrlab/DomainLab
closed
jigen failure
bug priority
``` RuleException: IndexError in file /home/ubuntu/georg/DomainLab/domainlab/exp_protocol/benchmark.smk, line 59: index 9 is out of bounds for axis 0 with size 9 File "/home/ubuntu//DomainLab/domainlab/exp_protocol/benchmark.smk", line 59, in __rule_run_experiment File "/home/ubuntu/DomainLab/domainlab/exp_protocol/run_experiment.py", line 87, in run_experiment File "/home/ubuntu/DomainLab/domainlab/compos/exp/exp_main.py", line 48, in execute File "/home/ubuntu/DomainLab/domainlab/algos/trainers/train_visitor.py", line 43, in tr_epoch File "/home/ubuntu/DomainLab/domainlab/algos/trainers/train_basic.py", line 26, in tr_epoch File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 681, in __next__ File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 721, in _next_da ta File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 49, in fetch File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 49, in <listco mp> File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/dataset.py", line 235, in __getitem__ File "/home/ubuntu/DomainLab/domainlab/dsets/utils_wrapdset_patches.py", line 108, in __getitem__ File "/home/ubuntu/DomainLab/domainlab/dsets/utils_wrapdset_patches.py", line 108, in <listcomp> File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/concurrent/futures/thread.py", line 57, in run Shutting down, this might take some time. Exiting because a job execution failed. Look above for error message ```
1.0
jigen failure - ``` RuleException: IndexError in file /home/ubuntu/georg/DomainLab/domainlab/exp_protocol/benchmark.smk, line 59: index 9 is out of bounds for axis 0 with size 9 File "/home/ubuntu//DomainLab/domainlab/exp_protocol/benchmark.smk", line 59, in __rule_run_experiment File "/home/ubuntu/DomainLab/domainlab/exp_protocol/run_experiment.py", line 87, in run_experiment File "/home/ubuntu/DomainLab/domainlab/compos/exp/exp_main.py", line 48, in execute File "/home/ubuntu/DomainLab/domainlab/algos/trainers/train_visitor.py", line 43, in tr_epoch File "/home/ubuntu/DomainLab/domainlab/algos/trainers/train_basic.py", line 26, in tr_epoch File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 681, in __next__ File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/dataloader.py", line 721, in _next_da ta File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 49, in fetch File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/_utils/fetch.py", line 49, in <listco mp> File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/site-packages/torch/utils/data/dataset.py", line 235, in __getitem__ File "/home/ubuntu/DomainLab/domainlab/dsets/utils_wrapdset_patches.py", line 108, in __getitem__ File "/home/ubuntu/DomainLab/domainlab/dsets/utils_wrapdset_patches.py", line 108, in <listcomp> File "/home/ubuntu/mambaforge/envs/gdl/lib/python3.8/concurrent/futures/thread.py", line 57, in run Shutting down, this might take some time. Exiting because a job execution failed. Look above for error message ```
priority
jigen failure ruleexception indexerror in file home ubuntu georg domainlab domainlab exp protocol benchmark smk line index is out of bounds for axis with size file home ubuntu domainlab domainlab exp protocol benchmark smk line in rule run experiment file home ubuntu domainlab domainlab exp protocol run experiment py line in run experiment file home ubuntu domainlab domainlab compos exp exp main py line in execute file home ubuntu domainlab domainlab algos trainers train visitor py line in tr epoch file home ubuntu domainlab domainlab algos trainers train basic py line in tr epoch file home ubuntu mambaforge envs gdl lib site packages torch utils data dataloader py line in next file home ubuntu mambaforge envs gdl lib site packages torch utils data dataloader py line in next da ta file home ubuntu mambaforge envs gdl lib site packages torch utils data utils fetch py line in fetch file home ubuntu mambaforge envs gdl lib site packages torch utils data utils fetch py line in listco mp file home ubuntu mambaforge envs gdl lib site packages torch utils data dataset py line in getitem file home ubuntu domainlab domainlab dsets utils wrapdset patches py line in getitem file home ubuntu domainlab domainlab dsets utils wrapdset patches py line in file home ubuntu mambaforge envs gdl lib concurrent futures thread py line in run shutting down this might take some time exiting because a job execution failed look above for error message
1
330,909
10,057,358,632
IssuesEvent
2019-07-22 11:25:28
code-ready/crc
https://api.github.com/repos/code-ready/crc
closed
Add hyperkit support
kind/bug kind/task os/macos priority/major size/L
- [x] Extend the existing darwin preflight checks for any additional check which is needed for hyperkit (#46) - [x] Import/adapt https://github.com/machine-drivers/docker-machine-driver-hyperkit - [x] Add hyperkit support to the crc binary We'll also more metadata associated to the disk image for proper support, see https://github.com/code-ready/snc/issues/39
1.0
Add hyperkit support - - [x] Extend the existing darwin preflight checks for any additional check which is needed for hyperkit (#46) - [x] Import/adapt https://github.com/machine-drivers/docker-machine-driver-hyperkit - [x] Add hyperkit support to the crc binary We'll also more metadata associated to the disk image for proper support, see https://github.com/code-ready/snc/issues/39
priority
add hyperkit support extend the existing darwin preflight checks for any additional check which is needed for hyperkit import adapt add hyperkit support to the crc binary we ll also more metadata associated to the disk image for proper support see
1
557,423
16,509,689,949
IssuesEvent
2021-05-26 01:22:12
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
[fishcat] codehaus.jackson.map.exc.UnrecognizedPropertyException: Unrecognized field "name" (Class com.demo.User), not marked as ignorable
4_0-exclude Component: jax-rs Priority: Major Stale Type: Bug fishcat jax-rs rest
I used Netbeans to generate a jax-rs resources and config to test a jax-rs application. I also used Netbeans to generate the jax-rs client from the HelloResources. The problem is when I try to consume a "application/json" into a User object. I obtain this stacktrace in GF 4.0 latest promoted build. I'll provide the source code at the end of the stacktrace. INFO: class com.demo.CDIManagedBean post construct INFO: class com.demo.HelloResource post construct WARNING: StandardWrapperValve[org.netbeans.rest.application.config.ApplicationConfig]: Servlet.service() for servlet org.netbeans.rest.application.config.ApplicationConfig threw exception org.codehaus.jackson.map.exc.UnrecognizedPropertyException: Unrecognized field "name" (Class com.demo.User), not marked as ignorable at [Source: org.glassfish.jersey.message.internal.EntityInputStream@d03ff; line: 1, column: 10] (through reference chain: com.demo.User["name"]) at org.codehaus.jackson.map.exc.UnrecognizedPropertyException.from(UnrecognizedPropertyException.java:53) at org.codehaus.jackson.map.deser.StdDeserializationContext.unknownFieldException(StdDeserializationContext.java:267) at org.codehaus.jackson.map.deser.std.StdDeserializer.reportUnknownProperty(StdDeserializer.java:673) at org.codehaus.jackson.map.deser.std.StdDeserializer.handleUnknownProperty(StdDeserializer.java:659) at org.codehaus.jackson.map.deser.BeanDeserializer.handleUnknownProperty(BeanDeserializer.java:1365) at org.codehaus.jackson.map.deser.BeanDeserializer._handleUnknown(BeanDeserializer.java:725) at org.codehaus.jackson.map.deser.BeanDeserializer.deserializeFromObject(BeanDeserializer.java:703) at org.codehaus.jackson.map.deser.BeanDeserializer.deserialize(BeanDeserializer.java:580) at org.codehaus.jackson.map.ObjectMapper._readValue(ObjectMapper.java:2704) at org.codehaus.jackson.map.ObjectMapper.readValue(ObjectMapper.java:1315) at org.codehaus.jackson.jaxrs.JacksonJsonProvider.readFrom(JacksonJsonProvider.java:419) at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor$TerminalReaderInterceptor.aroundReadFrom(ReaderInterceptorExecutor.java:181) at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor.proceed(ReaderInterceptorExecutor.java:134) at org.glassfish.jersey.server.internal.MappableExceptionWrapperInterceptor.aroundReadFrom(MappableExceptionWrapperInterceptor.java:72) at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor.proceed(ReaderInterceptorExecutor.java:134) at org.glassfish.jersey.message.internal.MessageBodyFactory.readFrom(MessageBodyFactory.java:828) at org.glassfish.jersey.message.internal.InboundMessageContext.readEntity(InboundMessageContext.java:833) at org.glassfish.jersey.server.ContainerRequest.readEntity(ContainerRequest.java:245) at org.glassfish.jersey.server.internal.inject.EntityParamValueFactoryProvider$EntityValueFactory.get(EntityParamValueFactoryProvider.java:96) at org.glassfish.jersey.server.internal.inject.AbstractHttpContextValueFactory.provide(AbstractHttpContextValueFactory.java:66) at org.glassfish.jersey.server.spi.internal.ParameterValueHelper.getParameterValues(ParameterValueHelper.java:81) at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$AbstractMethodParamInvoker.getParamValues(JavaResourceMethodDispatcherProvider.java:121) at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$VoidOutInvoker.doDispatch(JavaResourceMethodDispatcherProvider.java:136) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.dispatch(AbstractJavaResourceMethodDispatcher.java:91) at org.glassfish.jersey.server.model.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:346) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:341) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:101) at org.glassfish.jersey.server.ServerRuntime$1.run(ServerRuntime.java:224) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:271) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:267) at org.glassfish.jersey.internal.Errors.process(Errors.java:315) at org.glassfish.jersey.internal.Errors.process(Errors.java:297) at org.glassfish.jersey.internal.Errors.process(Errors.java:267) at org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:317) at org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:198) at org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:946) at org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:323) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:372) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:335) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:218) at org.apache.catalina.core.StandardWrapper.service(StandardWrapper.java:1682) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:318) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:160) at org.apache.catalina.core.StandardPipeline.doInvoke(StandardPipeline.java:734) at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:673) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:174) at org.apache.catalina.connector.CoyoteAdapter.doService(CoyoteAdapter.java:357) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:260) at com.sun.enterprise.v3.services.impl.ContainerMapper.service(ContainerMapper.java:188) at org.glassfish.grizzly.http.server.HttpHandler.runService(HttpHandler.java:191) at org.glassfish.grizzly.http.server.HttpHandler.doHandle(HttpHandler.java:168) at org.glassfish.grizzly.http.server.HttpServerFilter.handleRead(HttpServerFilter.java:189) at org.glassfish.grizzly.filterchain.ExecutorResolver$9.execute(ExecutorResolver.java:119) at org.glassfish.grizzly.filterchain.DefaultFilterChain.executeFilter(DefaultFilterChain.java:288) at org.glassfish.grizzly.filterchain.DefaultFilterChain.executeChainPart(DefaultFilterChain.java:206) at org.glassfish.grizzly.filterchain.DefaultFilterChain.execute(DefaultFilterChain.java:136) at org.glassfish.grizzly.filterchain.DefaultFilterChain.process(DefaultFilterChain.java:114) at org.glassfish.grizzly.ProcessorExecutor.execute(ProcessorExecutor.java:77) at org.glassfish.grizzly.nio.transport.TCPNIOTransport.fireIOEvent(TCPNIOTransport.java:838) at org.glassfish.grizzly.strategies.AbstractIOStrategy.fireIOEvent(AbstractIOStrategy.java:113) at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.run0(WorkerThreadIOStrategy.java:115) at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.access$100(WorkerThreadIOStrategy.java:55) at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy$WorkerThreadRunnable.run(WorkerThreadIOStrategy.java:135) at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.doWork(AbstractThreadPool.java:564) at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.run(AbstractThreadPool.java:544) at java.lang.Thread.run(Thread.java:722) ``` package com.demo; import javax.xml.bind.annotation.XmlAccessType; import javax.xml.bind.annotation.XmlAccessorType; import javax.xml.bind.annotation.XmlRootElement; /** * * @author sdionne1 */ @XmlRootElement @XmlAccessorType(XmlAccessType.FIELD) public class User { public String name; public String lastname; } ``` ``` package com.demo; import javax.annotation.PostConstruct; import javax.inject.Inject; import javax.ws.rs.Consumes; import javax.ws.rs.GET; import javax.ws.rs.POST; import javax.ws.rs.Path; import javax.ws.rs.Produces; /** * * @author sdionne1 */ @Path("hello") public class HelloResource { @Inject CDIManagedBean bean; @Inject EJBManagedBean ejb; @PostConstruct public void init(){ System.out.println(HelloResource.class + " post construct"); } @GET @Produces("text/plain") public String getHello(){ return "hello : " + bean.getOutput() + " : " + ejb.getOutput(); } @POST @Consumes("text/plain") public void sayHello(String message){ System.out.println("Say Hello : " + message); } @POST @Consumes("application/json") public void helloJson(User user){ System.out.println("helloJson user : " + user); } } ``` ``` /* * To change this template, choose Tools | Templates * and open the template in the editor. */ package com.demo; import com.sun.jersey.api.client.Client; import com.sun.jersey.api.client.UniformInterfaceException; import com.sun.jersey.api.client.WebResource; /** * Jersey REST client generated for REST resource:HelloResource [hello]<br> * USAGE: * <pre> * NewJerseyClient client = new NewJerseyClient(); * Object response = client.XXX(...); * // do whatever with response * client.close(); * </pre> * * @author sdionne1 */ public class NewJerseyClient { private WebResource webResource; private Client client; private static final String BASE_URI = "http://localhost:8080/WebApplication1/resources"; public NewJerseyClient() { com.sun.jersey.api.client.config.ClientConfig config = new com.sun.jersey.api.client.config.DefaultClientConfig(); client = Client.create(config); webResource = client.resource(BASE_URI).path("hello"); } public void helloJson(Object requestEntity) throws UniformInterfaceException { webResource.type(javax.ws.rs.core.MediaType.APPLICATION_JSON).post(requestEntity); } public String getHello() throws UniformInterfaceException { WebResource resource = webResource; return resource.accept(javax.ws.rs.core.MediaType.TEXT_PLAIN).get(String.class); } public void sayHello(Object requestEntity) throws UniformInterfaceException { webResource.type(javax.ws.rs.core.MediaType.TEXT_PLAIN).post(requestEntity); } public void close() { client.destroy(); } public static void main(String[] args){ NewJerseyClient client = new NewJerseyClient(); User user = new User(); user.name="name1"; user.lastname="lastname2"; client.helloJson(user); client.close(); } } ``` #### Environment Product Version: NetBeans IDE Dev (Build 201305052300) Java: 1.7.0_13; Java HotSpot(TM) Client VM 23.7-b01 Runtime: Java(TM) SE Runtime Environment 1.7.0_13-b20 System: Windows XP version 5.1 running on x86; Cp1252; fr_CA (nb) User directory: C:\Documents and Settings\sdionne1\Application Data\NetBeans\dev Cache directory: C:\Documents and Settings\sdionne1\Local Settings\Application Data\NetBeans\Cache\dev #### Affected Versions [4.0_dev]
1.0
[fishcat] codehaus.jackson.map.exc.UnrecognizedPropertyException: Unrecognized field "name" (Class com.demo.User), not marked as ignorable - I used Netbeans to generate a jax-rs resources and config to test a jax-rs application. I also used Netbeans to generate the jax-rs client from the HelloResources. The problem is when I try to consume a "application/json" into a User object. I obtain this stacktrace in GF 4.0 latest promoted build. I'll provide the source code at the end of the stacktrace. INFO: class com.demo.CDIManagedBean post construct INFO: class com.demo.HelloResource post construct WARNING: StandardWrapperValve[org.netbeans.rest.application.config.ApplicationConfig]: Servlet.service() for servlet org.netbeans.rest.application.config.ApplicationConfig threw exception org.codehaus.jackson.map.exc.UnrecognizedPropertyException: Unrecognized field "name" (Class com.demo.User), not marked as ignorable at [Source: org.glassfish.jersey.message.internal.EntityInputStream@d03ff; line: 1, column: 10] (through reference chain: com.demo.User["name"]) at org.codehaus.jackson.map.exc.UnrecognizedPropertyException.from(UnrecognizedPropertyException.java:53) at org.codehaus.jackson.map.deser.StdDeserializationContext.unknownFieldException(StdDeserializationContext.java:267) at org.codehaus.jackson.map.deser.std.StdDeserializer.reportUnknownProperty(StdDeserializer.java:673) at org.codehaus.jackson.map.deser.std.StdDeserializer.handleUnknownProperty(StdDeserializer.java:659) at org.codehaus.jackson.map.deser.BeanDeserializer.handleUnknownProperty(BeanDeserializer.java:1365) at org.codehaus.jackson.map.deser.BeanDeserializer._handleUnknown(BeanDeserializer.java:725) at org.codehaus.jackson.map.deser.BeanDeserializer.deserializeFromObject(BeanDeserializer.java:703) at org.codehaus.jackson.map.deser.BeanDeserializer.deserialize(BeanDeserializer.java:580) at org.codehaus.jackson.map.ObjectMapper._readValue(ObjectMapper.java:2704) at org.codehaus.jackson.map.ObjectMapper.readValue(ObjectMapper.java:1315) at org.codehaus.jackson.jaxrs.JacksonJsonProvider.readFrom(JacksonJsonProvider.java:419) at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor$TerminalReaderInterceptor.aroundReadFrom(ReaderInterceptorExecutor.java:181) at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor.proceed(ReaderInterceptorExecutor.java:134) at org.glassfish.jersey.server.internal.MappableExceptionWrapperInterceptor.aroundReadFrom(MappableExceptionWrapperInterceptor.java:72) at org.glassfish.jersey.message.internal.ReaderInterceptorExecutor.proceed(ReaderInterceptorExecutor.java:134) at org.glassfish.jersey.message.internal.MessageBodyFactory.readFrom(MessageBodyFactory.java:828) at org.glassfish.jersey.message.internal.InboundMessageContext.readEntity(InboundMessageContext.java:833) at org.glassfish.jersey.server.ContainerRequest.readEntity(ContainerRequest.java:245) at org.glassfish.jersey.server.internal.inject.EntityParamValueFactoryProvider$EntityValueFactory.get(EntityParamValueFactoryProvider.java:96) at org.glassfish.jersey.server.internal.inject.AbstractHttpContextValueFactory.provide(AbstractHttpContextValueFactory.java:66) at org.glassfish.jersey.server.spi.internal.ParameterValueHelper.getParameterValues(ParameterValueHelper.java:81) at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$AbstractMethodParamInvoker.getParamValues(JavaResourceMethodDispatcherProvider.java:121) at org.glassfish.jersey.server.model.internal.JavaResourceMethodDispatcherProvider$VoidOutInvoker.doDispatch(JavaResourceMethodDispatcherProvider.java:136) at org.glassfish.jersey.server.model.internal.AbstractJavaResourceMethodDispatcher.dispatch(AbstractJavaResourceMethodDispatcher.java:91) at org.glassfish.jersey.server.model.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:346) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:341) at org.glassfish.jersey.server.model.ResourceMethodInvoker.apply(ResourceMethodInvoker.java:101) at org.glassfish.jersey.server.ServerRuntime$1.run(ServerRuntime.java:224) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:271) at org.glassfish.jersey.internal.Errors$1.call(Errors.java:267) at org.glassfish.jersey.internal.Errors.process(Errors.java:315) at org.glassfish.jersey.internal.Errors.process(Errors.java:297) at org.glassfish.jersey.internal.Errors.process(Errors.java:267) at org.glassfish.jersey.process.internal.RequestScope.runInScope(RequestScope.java:317) at org.glassfish.jersey.server.ServerRuntime.process(ServerRuntime.java:198) at org.glassfish.jersey.server.ApplicationHandler.handle(ApplicationHandler.java:946) at org.glassfish.jersey.servlet.WebComponent.service(WebComponent.java:323) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:372) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:335) at org.glassfish.jersey.servlet.ServletContainer.service(ServletContainer.java:218) at org.apache.catalina.core.StandardWrapper.service(StandardWrapper.java:1682) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:318) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:160) at org.apache.catalina.core.StandardPipeline.doInvoke(StandardPipeline.java:734) at org.apache.catalina.core.StandardPipeline.invoke(StandardPipeline.java:673) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:174) at org.apache.catalina.connector.CoyoteAdapter.doService(CoyoteAdapter.java:357) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:260) at com.sun.enterprise.v3.services.impl.ContainerMapper.service(ContainerMapper.java:188) at org.glassfish.grizzly.http.server.HttpHandler.runService(HttpHandler.java:191) at org.glassfish.grizzly.http.server.HttpHandler.doHandle(HttpHandler.java:168) at org.glassfish.grizzly.http.server.HttpServerFilter.handleRead(HttpServerFilter.java:189) at org.glassfish.grizzly.filterchain.ExecutorResolver$9.execute(ExecutorResolver.java:119) at org.glassfish.grizzly.filterchain.DefaultFilterChain.executeFilter(DefaultFilterChain.java:288) at org.glassfish.grizzly.filterchain.DefaultFilterChain.executeChainPart(DefaultFilterChain.java:206) at org.glassfish.grizzly.filterchain.DefaultFilterChain.execute(DefaultFilterChain.java:136) at org.glassfish.grizzly.filterchain.DefaultFilterChain.process(DefaultFilterChain.java:114) at org.glassfish.grizzly.ProcessorExecutor.execute(ProcessorExecutor.java:77) at org.glassfish.grizzly.nio.transport.TCPNIOTransport.fireIOEvent(TCPNIOTransport.java:838) at org.glassfish.grizzly.strategies.AbstractIOStrategy.fireIOEvent(AbstractIOStrategy.java:113) at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.run0(WorkerThreadIOStrategy.java:115) at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy.access$100(WorkerThreadIOStrategy.java:55) at org.glassfish.grizzly.strategies.WorkerThreadIOStrategy$WorkerThreadRunnable.run(WorkerThreadIOStrategy.java:135) at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.doWork(AbstractThreadPool.java:564) at org.glassfish.grizzly.threadpool.AbstractThreadPool$Worker.run(AbstractThreadPool.java:544) at java.lang.Thread.run(Thread.java:722) ``` package com.demo; import javax.xml.bind.annotation.XmlAccessType; import javax.xml.bind.annotation.XmlAccessorType; import javax.xml.bind.annotation.XmlRootElement; /** * * @author sdionne1 */ @XmlRootElement @XmlAccessorType(XmlAccessType.FIELD) public class User { public String name; public String lastname; } ``` ``` package com.demo; import javax.annotation.PostConstruct; import javax.inject.Inject; import javax.ws.rs.Consumes; import javax.ws.rs.GET; import javax.ws.rs.POST; import javax.ws.rs.Path; import javax.ws.rs.Produces; /** * * @author sdionne1 */ @Path("hello") public class HelloResource { @Inject CDIManagedBean bean; @Inject EJBManagedBean ejb; @PostConstruct public void init(){ System.out.println(HelloResource.class + " post construct"); } @GET @Produces("text/plain") public String getHello(){ return "hello : " + bean.getOutput() + " : " + ejb.getOutput(); } @POST @Consumes("text/plain") public void sayHello(String message){ System.out.println("Say Hello : " + message); } @POST @Consumes("application/json") public void helloJson(User user){ System.out.println("helloJson user : " + user); } } ``` ``` /* * To change this template, choose Tools | Templates * and open the template in the editor. */ package com.demo; import com.sun.jersey.api.client.Client; import com.sun.jersey.api.client.UniformInterfaceException; import com.sun.jersey.api.client.WebResource; /** * Jersey REST client generated for REST resource:HelloResource [hello]<br> * USAGE: * <pre> * NewJerseyClient client = new NewJerseyClient(); * Object response = client.XXX(...); * // do whatever with response * client.close(); * </pre> * * @author sdionne1 */ public class NewJerseyClient { private WebResource webResource; private Client client; private static final String BASE_URI = "http://localhost:8080/WebApplication1/resources"; public NewJerseyClient() { com.sun.jersey.api.client.config.ClientConfig config = new com.sun.jersey.api.client.config.DefaultClientConfig(); client = Client.create(config); webResource = client.resource(BASE_URI).path("hello"); } public void helloJson(Object requestEntity) throws UniformInterfaceException { webResource.type(javax.ws.rs.core.MediaType.APPLICATION_JSON).post(requestEntity); } public String getHello() throws UniformInterfaceException { WebResource resource = webResource; return resource.accept(javax.ws.rs.core.MediaType.TEXT_PLAIN).get(String.class); } public void sayHello(Object requestEntity) throws UniformInterfaceException { webResource.type(javax.ws.rs.core.MediaType.TEXT_PLAIN).post(requestEntity); } public void close() { client.destroy(); } public static void main(String[] args){ NewJerseyClient client = new NewJerseyClient(); User user = new User(); user.name="name1"; user.lastname="lastname2"; client.helloJson(user); client.close(); } } ``` #### Environment Product Version: NetBeans IDE Dev (Build 201305052300) Java: 1.7.0_13; Java HotSpot(TM) Client VM 23.7-b01 Runtime: Java(TM) SE Runtime Environment 1.7.0_13-b20 System: Windows XP version 5.1 running on x86; Cp1252; fr_CA (nb) User directory: C:\Documents and Settings\sdionne1\Application Data\NetBeans\dev Cache directory: C:\Documents and Settings\sdionne1\Local Settings\Application Data\NetBeans\Cache\dev #### Affected Versions [4.0_dev]
priority
codehaus jackson map exc unrecognizedpropertyexception unrecognized field name class com demo user not marked as ignorable i used netbeans to generate a jax rs resources and config to test a jax rs application i also used netbeans to generate the jax rs client from the helloresources the problem is when i try to consume a application json into a user object i obtain this stacktrace in gf latest promoted build i ll provide the source code at the end of the stacktrace info class com demo cdimanagedbean post construct info class com demo helloresource post construct warning standardwrappervalve servlet service for servlet org netbeans rest application config applicationconfig threw exception org codehaus jackson map exc unrecognizedpropertyexception unrecognized field name class com demo user not marked as ignorable at through reference chain com demo user at org codehaus jackson map exc unrecognizedpropertyexception from unrecognizedpropertyexception java at org codehaus jackson map deser stddeserializationcontext unknownfieldexception stddeserializationcontext java at org codehaus jackson map deser std stddeserializer reportunknownproperty stddeserializer java at org codehaus jackson map deser std stddeserializer handleunknownproperty stddeserializer java at org codehaus jackson map deser beandeserializer handleunknownproperty beandeserializer java at org codehaus jackson map deser beandeserializer handleunknown beandeserializer java at org codehaus jackson map deser beandeserializer deserializefromobject beandeserializer java at org codehaus jackson map deser beandeserializer deserialize beandeserializer java at org codehaus jackson map objectmapper readvalue objectmapper java at org codehaus jackson map objectmapper readvalue objectmapper java at org codehaus jackson jaxrs jacksonjsonprovider readfrom jacksonjsonprovider java at org glassfish jersey message internal readerinterceptorexecutor terminalreaderinterceptor aroundreadfrom readerinterceptorexecutor java at org glassfish jersey message internal readerinterceptorexecutor proceed readerinterceptorexecutor java at org glassfish jersey server internal mappableexceptionwrapperinterceptor aroundreadfrom mappableexceptionwrapperinterceptor java at org glassfish jersey message internal readerinterceptorexecutor proceed readerinterceptorexecutor java at org glassfish jersey message internal messagebodyfactory readfrom messagebodyfactory java at org glassfish jersey message internal inboundmessagecontext readentity inboundmessagecontext java at org glassfish jersey server containerrequest readentity containerrequest java at org glassfish jersey server internal inject entityparamvaluefactoryprovider entityvaluefactory get entityparamvaluefactoryprovider java at org glassfish jersey server internal inject abstracthttpcontextvaluefactory provide abstracthttpcontextvaluefactory java at org glassfish jersey server spi internal parametervaluehelper getparametervalues parametervaluehelper java at org glassfish jersey server model internal javaresourcemethoddispatcherprovider abstractmethodparaminvoker getparamvalues javaresourcemethoddispatcherprovider java at org glassfish jersey server model internal javaresourcemethoddispatcherprovider voidoutinvoker dodispatch javaresourcemethoddispatcherprovider java at org glassfish jersey server model internal abstractjavaresourcemethoddispatcher dispatch abstractjavaresourcemethoddispatcher java at org glassfish jersey server model resourcemethodinvoker invoke resourcemethodinvoker java at org glassfish jersey server model resourcemethodinvoker apply resourcemethodinvoker java at org glassfish jersey server model resourcemethodinvoker apply resourcemethodinvoker java at org glassfish jersey server serverruntime run serverruntime java at org glassfish jersey internal errors call errors java at org glassfish jersey internal errors call errors java at org glassfish jersey internal errors process errors java at org glassfish jersey internal errors process errors java at org glassfish jersey internal errors process errors java at org glassfish jersey process internal requestscope runinscope requestscope java at org glassfish jersey server serverruntime process serverruntime java at org glassfish jersey server applicationhandler handle applicationhandler java at org glassfish jersey servlet webcomponent service webcomponent java at org glassfish jersey servlet servletcontainer service servletcontainer java at org glassfish jersey servlet servletcontainer service servletcontainer java at org glassfish jersey servlet servletcontainer service servletcontainer java at org apache catalina core standardwrapper service standardwrapper java at org apache catalina core standardwrappervalve invoke standardwrappervalve java at org apache catalina core standardcontextvalve invoke standardcontextvalve java at org apache catalina core standardpipeline doinvoke standardpipeline java at org apache catalina core standardpipeline invoke standardpipeline java at org apache catalina core standardhostvalve invoke standardhostvalve java at org apache catalina connector coyoteadapter doservice coyoteadapter java at org apache catalina connector coyoteadapter service coyoteadapter java at com sun enterprise services impl containermapper service containermapper java at org glassfish grizzly http server httphandler runservice httphandler java at org glassfish grizzly http server httphandler dohandle httphandler java at org glassfish grizzly http server httpserverfilter handleread httpserverfilter java at org glassfish grizzly filterchain executorresolver execute executorresolver java at org glassfish grizzly filterchain defaultfilterchain executefilter defaultfilterchain java at org glassfish grizzly filterchain defaultfilterchain executechainpart defaultfilterchain java at org glassfish grizzly filterchain defaultfilterchain execute defaultfilterchain java at org glassfish grizzly filterchain defaultfilterchain process defaultfilterchain java at org glassfish grizzly processorexecutor execute processorexecutor java at org glassfish grizzly nio transport tcpniotransport fireioevent tcpniotransport java at org glassfish grizzly strategies abstractiostrategy fireioevent abstractiostrategy java at org glassfish grizzly strategies workerthreadiostrategy workerthreadiostrategy java at org glassfish grizzly strategies workerthreadiostrategy access workerthreadiostrategy java at org glassfish grizzly strategies workerthreadiostrategy workerthreadrunnable run workerthreadiostrategy java at org glassfish grizzly threadpool abstractthreadpool worker dowork abstractthreadpool java at org glassfish grizzly threadpool abstractthreadpool worker run abstractthreadpool java at java lang thread run thread java package com demo import javax xml bind annotation xmlaccesstype import javax xml bind annotation xmlaccessortype import javax xml bind annotation xmlrootelement author xmlrootelement xmlaccessortype xmlaccesstype field public class user public string name public string lastname package com demo import javax annotation postconstruct import javax inject inject import javax ws rs consumes import javax ws rs get import javax ws rs post import javax ws rs path import javax ws rs produces author path hello public class helloresource inject cdimanagedbean bean inject ejbmanagedbean ejb postconstruct public void init system out println helloresource class post construct get produces text plain public string gethello return hello bean getoutput ejb getoutput post consumes text plain public void sayhello string message system out println say hello message post consumes application json public void hellojson user user system out println hellojson user user to change this template choose tools templates and open the template in the editor package com demo import com sun jersey api client client import com sun jersey api client uniforminterfaceexception import com sun jersey api client webresource jersey rest client generated for rest resource helloresource usage newjerseyclient client new newjerseyclient object response client xxx do whatever with response client close author public class newjerseyclient private webresource webresource private client client private static final string base uri public newjerseyclient com sun jersey api client config clientconfig config new com sun jersey api client config defaultclientconfig client client create config webresource client resource base uri path hello public void hellojson object requestentity throws uniforminterfaceexception webresource type javax ws rs core mediatype application json post requestentity public string gethello throws uniforminterfaceexception webresource resource webresource return resource accept javax ws rs core mediatype text plain get string class public void sayhello object requestentity throws uniforminterfaceexception webresource type javax ws rs core mediatype text plain post requestentity public void close client destroy public static void main string args newjerseyclient client new newjerseyclient user user new user user name user lastname client hellojson user client close environment product version netbeans ide dev build java java hotspot tm client vm runtime java tm se runtime environment system windows xp version running on fr ca nb user directory c documents and settings application data netbeans dev cache directory c documents and settings local settings application data netbeans cache dev affected versions
1
788,509
27,755,443,242
IssuesEvent
2023-03-16 01:50:58
brandondombrowsky/BastCastle
https://api.github.com/repos/brandondombrowsky/BastCastle
closed
Assist with local vent setup process
priority-high
As a developer, I want to assist in the vent setup process so that I know they are working correctly. - [ ] Deliver vents to Kyle - [ ] Wait for Kyle to install them - For testing or time purposes, they might just be left in the open to demo - [ ] Assist in making sure they are set up and connected properly - [ ] Assist with connecting smart plugs to HA - [ ] Verify that plugs<->HA works as expected - This issue does not include physical installation, that is left up to the client.
1.0
Assist with local vent setup process - As a developer, I want to assist in the vent setup process so that I know they are working correctly. - [ ] Deliver vents to Kyle - [ ] Wait for Kyle to install them - For testing or time purposes, they might just be left in the open to demo - [ ] Assist in making sure they are set up and connected properly - [ ] Assist with connecting smart plugs to HA - [ ] Verify that plugs<->HA works as expected - This issue does not include physical installation, that is left up to the client.
priority
assist with local vent setup process as a developer i want to assist in the vent setup process so that i know they are working correctly deliver vents to kyle wait for kyle to install them for testing or time purposes they might just be left in the open to demo assist in making sure they are set up and connected properly assist with connecting smart plugs to ha verify that plugs ha works as expected this issue does not include physical installation that is left up to the client
1
52,548
10,880,950,570
IssuesEvent
2019-11-17 14:40:51
square/moshi
https://api.github.com/repos/square/moshi
closed
Suppress deprecation warnings if any properties or target classes are deprecated
Kotlin - code gen enhancement
This prevents using `warningsAsError` or `-Werror` in compilations otherwise Thanks to @erichkleung for the report
1.0
Suppress deprecation warnings if any properties or target classes are deprecated - This prevents using `warningsAsError` or `-Werror` in compilations otherwise Thanks to @erichkleung for the report
non_priority
suppress deprecation warnings if any properties or target classes are deprecated this prevents using warningsaserror or werror in compilations otherwise thanks to erichkleung for the report
0
334,780
29,984,708,184
IssuesEvent
2023-06-25 04:45:47
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix jax_numpy_math.test_jax_numpy_sin
JAX Frontend Sub Task Failing Test
| | | |---|---| |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix jax_numpy_math.test_jax_numpy_sin - | | | |---|---| |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5367467501/jobs/9737641566"><img src=https://img.shields.io/badge/-success-success></a>
non_priority
fix jax numpy math test jax numpy sin torch a href src numpy a href src jax a href src tensorflow a href src paddle a href src
0
210,988
7,197,309,315
IssuesEvent
2018-02-05 08:36:54
kubernetes-incubator/kube-arbitrator
https://api.github.com/repos/kubernetes-incubator/kube-arbitrator
closed
Race condition between cache and policy
kind/bug priority/important-soon sig/batchd
There is a race condition between cache and policy, which will cause policy schedule more pods on the same host. Here are the details: #### The brief workflow of each thread/component: ##### T1. Cache thread: 1. Update Pod/Node/Consumer information from api-server ##### T2.Policy - resource alloc thread: 1. Get Pod/Node/Consumer snapshot from Cache 2. Allocate resources for Pod/PodSet in the cache snapshot 2.1. Assign MinAvailable to each PodSet one by one 2.2. Assign left resources to each PodSet by DRF 3. Add allocate resources result to a queue, name it as `AllocationQueue` ##### T3.Policy - process alloc decision thread: 1. Get the scheduler result from `AllocationQueue` and update it into api-server ##### T4.Kubelet 1. Get Pod information from api-server, start/stop it and then update status to api-server ### Case 01: 1. In step T2.2.1: Policy allocate MinAvailable to PodSet 2. In step T2.3: Policy add allocate result to `AllocationQueue` (Pods still pending) 3. In step T3.1: Policy update allocate result to api-server (Pods still pending, with Nodename) 4. In step T1.1: Cache update information from api-server in step T1.1 (Pods still pending, with Nodename) 5. In step T2.1: Policy get cache snapshot (Pods still pending, with Nodename) 6. In step T2.2.1: Policy allocate MinAvailable to PodSet, however, the policy doesn't calculate the number of pending pods with Nodename, still get pods from pending queue one by one. 7. In step T3.1: Policy update allocate result to api-server (more Pods still pending, with Nodename) 8. In step T4.1: Kubelet start pods ### Case 02: 1. In step T2.2.1: Policy allocate MinAvailable to PodSet 2. In step T2.3: Policy add allocate result to `AllocationQueue` (Pods still pending) 3. In step T3.1: Policy update allocate result to api-server (Pods still pending, with Nodename) 4. In step T2.1: Policy get cache snapshot (Pods still pending) 5. In step T1.1: Cache update information from api-server (Pods still pending, with Nodename) 6. In step T2.2.1: Policy allocate MinAvailable to PodSet, however, all the pods in the pending queue without nodename, it may cause policy allocate more MinAvailable in this step. 7. In step T3.1: Policy update allocate result to api-server (more Pods still pending, with Nodename) 8. In step T4.1: Kubelet start pods
1.0
Race condition between cache and policy - There is a race condition between cache and policy, which will cause policy schedule more pods on the same host. Here are the details: #### The brief workflow of each thread/component: ##### T1. Cache thread: 1. Update Pod/Node/Consumer information from api-server ##### T2.Policy - resource alloc thread: 1. Get Pod/Node/Consumer snapshot from Cache 2. Allocate resources for Pod/PodSet in the cache snapshot 2.1. Assign MinAvailable to each PodSet one by one 2.2. Assign left resources to each PodSet by DRF 3. Add allocate resources result to a queue, name it as `AllocationQueue` ##### T3.Policy - process alloc decision thread: 1. Get the scheduler result from `AllocationQueue` and update it into api-server ##### T4.Kubelet 1. Get Pod information from api-server, start/stop it and then update status to api-server ### Case 01: 1. In step T2.2.1: Policy allocate MinAvailable to PodSet 2. In step T2.3: Policy add allocate result to `AllocationQueue` (Pods still pending) 3. In step T3.1: Policy update allocate result to api-server (Pods still pending, with Nodename) 4. In step T1.1: Cache update information from api-server in step T1.1 (Pods still pending, with Nodename) 5. In step T2.1: Policy get cache snapshot (Pods still pending, with Nodename) 6. In step T2.2.1: Policy allocate MinAvailable to PodSet, however, the policy doesn't calculate the number of pending pods with Nodename, still get pods from pending queue one by one. 7. In step T3.1: Policy update allocate result to api-server (more Pods still pending, with Nodename) 8. In step T4.1: Kubelet start pods ### Case 02: 1. In step T2.2.1: Policy allocate MinAvailable to PodSet 2. In step T2.3: Policy add allocate result to `AllocationQueue` (Pods still pending) 3. In step T3.1: Policy update allocate result to api-server (Pods still pending, with Nodename) 4. In step T2.1: Policy get cache snapshot (Pods still pending) 5. In step T1.1: Cache update information from api-server (Pods still pending, with Nodename) 6. In step T2.2.1: Policy allocate MinAvailable to PodSet, however, all the pods in the pending queue without nodename, it may cause policy allocate more MinAvailable in this step. 7. In step T3.1: Policy update allocate result to api-server (more Pods still pending, with Nodename) 8. In step T4.1: Kubelet start pods
priority
race condition between cache and policy there is a race condition between cache and policy which will cause policy schedule more pods on the same host here are the details the brief workflow of each thread component cache thread update pod node consumer information from api server policy resource alloc thread get pod node consumer snapshot from cache allocate resources for pod podset in the cache snapshot assign minavailable to each podset one by one assign left resources to each podset by drf add allocate resources result to a queue name it as allocationqueue policy process alloc decision thread get the scheduler result from allocationqueue and update it into api server kubelet get pod information from api server start stop it and then update status to api server case in step policy allocate minavailable to podset in step policy add allocate result to allocationqueue pods still pending in step policy update allocate result to api server pods still pending with nodename in step cache update information from api server in step pods still pending with nodename in step policy get cache snapshot pods still pending with nodename in step policy allocate minavailable to podset however the policy doesn t calculate the number of pending pods with nodename still get pods from pending queue one by one in step policy update allocate result to api server more pods still pending with nodename in step kubelet start pods case in step policy allocate minavailable to podset in step policy add allocate result to allocationqueue pods still pending in step policy update allocate result to api server pods still pending with nodename in step policy get cache snapshot pods still pending in step cache update information from api server pods still pending with nodename in step policy allocate minavailable to podset however all the pods in the pending queue without nodename it may cause policy allocate more minavailable in this step in step policy update allocate result to api server more pods still pending with nodename in step kubelet start pods
1
113,899
14,501,779,641
IssuesEvent
2020-12-11 20:00:59
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
VSP Customer Engagement Strategy
Epic service-design super-epic
### High Level User Stories - As a VFS team member, I want VSP to engage with me to ensure I am aware of platform changes that could impact my work. - As a VFS team member, I want to be able to easily provide feedback to VSP on platform products. - As a VSP team member, I want VSP customer engagements to contribute to VFS teams adopting platform standards. ### Problem Statement VSP currently communicates platform changes to all VFS team members through Slack. We have heard from VFS teams that announcements on Slack are often missed or get lost resulting in delay and frustration as they run into avoidable issues or waste time searching for an old Slack message. We have heard from VSP teams that these communication gaps lead to redundant support requests and a lack of awareness of platform improvements. _How might we improve VSP customer engagement processes to keep pace with the scale of our operation?_ ### Hypothesis or Bet Defining and executing a customer engagement strategy will improve VFS teams knowledge of and satisfaction with VSP products. ### OKR VSP Q4 '20 - O4: Get VFS teams to know, use, and love the DEPO Platform products and services. ### Success metrics - Increase in customer satisfaction rating - Decrease in manual customer support requests ### We will know we're done when... VSP has a documented and approved comms strategy that defines: - This is how customers learn of a new or updated product - This is how we remind them of a new or updated product - This is how we provide information for customers to learn more about a new or updated product - This is how customers provide feedback on a new or updated product - This is how we respond to customer feedback on a new or updated product ### Additional Details/Notes - [Customer feedback loop mural](https://app.mural.co/t/adhocvetsgov9623/m/adhocvetsgov9623/1580853363126/ef5036613c38f1afeffbc082f28d2d8016c66a4a) - [Customer feedback loop working doc](https://docs.google.com/document/d/1pvKMCciI_WkVxLvhnVLJVAKVeELUKNdwYl1gkOs-Jgg/edit?usp=sharing) ### Artifacts Product outline to come ### Communication - Team that owns this super epic: Service Design - Teams contributing to this super epic:
1.0
VSP Customer Engagement Strategy - ### High Level User Stories - As a VFS team member, I want VSP to engage with me to ensure I am aware of platform changes that could impact my work. - As a VFS team member, I want to be able to easily provide feedback to VSP on platform products. - As a VSP team member, I want VSP customer engagements to contribute to VFS teams adopting platform standards. ### Problem Statement VSP currently communicates platform changes to all VFS team members through Slack. We have heard from VFS teams that announcements on Slack are often missed or get lost resulting in delay and frustration as they run into avoidable issues or waste time searching for an old Slack message. We have heard from VSP teams that these communication gaps lead to redundant support requests and a lack of awareness of platform improvements. _How might we improve VSP customer engagement processes to keep pace with the scale of our operation?_ ### Hypothesis or Bet Defining and executing a customer engagement strategy will improve VFS teams knowledge of and satisfaction with VSP products. ### OKR VSP Q4 '20 - O4: Get VFS teams to know, use, and love the DEPO Platform products and services. ### Success metrics - Increase in customer satisfaction rating - Decrease in manual customer support requests ### We will know we're done when... VSP has a documented and approved comms strategy that defines: - This is how customers learn of a new or updated product - This is how we remind them of a new or updated product - This is how we provide information for customers to learn more about a new or updated product - This is how customers provide feedback on a new or updated product - This is how we respond to customer feedback on a new or updated product ### Additional Details/Notes - [Customer feedback loop mural](https://app.mural.co/t/adhocvetsgov9623/m/adhocvetsgov9623/1580853363126/ef5036613c38f1afeffbc082f28d2d8016c66a4a) - [Customer feedback loop working doc](https://docs.google.com/document/d/1pvKMCciI_WkVxLvhnVLJVAKVeELUKNdwYl1gkOs-Jgg/edit?usp=sharing) ### Artifacts Product outline to come ### Communication - Team that owns this super epic: Service Design - Teams contributing to this super epic:
non_priority
vsp customer engagement strategy high level user stories as a vfs team member i want vsp to engage with me to ensure i am aware of platform changes that could impact my work as a vfs team member i want to be able to easily provide feedback to vsp on platform products as a vsp team member i want vsp customer engagements to contribute to vfs teams adopting platform standards problem statement vsp currently communicates platform changes to all vfs team members through slack we have heard from vfs teams that announcements on slack are often missed or get lost resulting in delay and frustration as they run into avoidable issues or waste time searching for an old slack message we have heard from vsp teams that these communication gaps lead to redundant support requests and a lack of awareness of platform improvements how might we improve vsp customer engagement processes to keep pace with the scale of our operation hypothesis or bet defining and executing a customer engagement strategy will improve vfs teams knowledge of and satisfaction with vsp products okr vsp get vfs teams to know use and love the depo platform products and services success metrics increase in customer satisfaction rating decrease in manual customer support requests we will know we re done when vsp has a documented and approved comms strategy that defines this is how customers learn of a new or updated product this is how we remind them of a new or updated product this is how we provide information for customers to learn more about a new or updated product this is how customers provide feedback on a new or updated product this is how we respond to customer feedback on a new or updated product additional details notes artifacts product outline to come communication team that owns this super epic service design teams contributing to this super epic
0
374,303
11,083,711,028
IssuesEvent
2019-12-13 15:01:56
cms-gem-daq-project/gem-plotting-tools
https://api.github.com/repos/cms-gem-daq-project/gem-plotting-tools
opened
Mapping change for GE21 problem
Priority: High Type: Bug
<!--- Provide a general summary of the issue in the Title above --> ## Brief summary of issue <!--- Provide a description of the issue, including any other issues or pull requests it references --> After changing chamber_iEta2VFATPos to be a list of dicts with keys `["gemType"]` not all the analysis tools are working. E.g. anaSBitReadout.py is broken now ### Types of issue <!--- Propsed labels (see CONTRIBUTING.md) to help maintainers label your issue: --> - [x] Bug report (report an issue with the code) - [ ] Feature request (request for change which adds functionality) ## Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> ## Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ### Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> ## Possible Solution (for bugs) <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> ## Context (for feature requests) <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used: * Shell used: <!--- Template thanks to https://www.talater.com/open-source-templates/#/page/98 -->
1.0
Mapping change for GE21 problem - <!--- Provide a general summary of the issue in the Title above --> ## Brief summary of issue <!--- Provide a description of the issue, including any other issues or pull requests it references --> After changing chamber_iEta2VFATPos to be a list of dicts with keys `["gemType"]` not all the analysis tools are working. E.g. anaSBitReadout.py is broken now ### Types of issue <!--- Propsed labels (see CONTRIBUTING.md) to help maintainers label your issue: --> - [x] Bug report (report an issue with the code) - [ ] Feature request (request for change which adds functionality) ## Expected Behavior <!--- If you're describing a bug, tell us what should happen --> <!--- If you're suggesting a change/improvement, tell us how it should work --> ## Current Behavior <!--- If describing a bug, tell us what happens instead of the expected behavior --> <!--- If suggesting a change/improvement, explain the difference from current behavior --> ### Steps to Reproduce (for bugs) <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug. Include code to reproduce, if relevant --> ## Possible Solution (for bugs) <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- or ideas how to implement the addition or change --> ## Context (for feature requests) <!--- How has this issue affected you? What are you trying to accomplish? --> <!--- Providing context helps us come up with a solution that is most useful in the real world --> ## Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * Version used: * Shell used: <!--- Template thanks to https://www.talater.com/open-source-templates/#/page/98 -->
priority
mapping change for problem brief summary of issue after changing chamber to be a list of dicts with keys not all the analysis tools are working e g anasbitreadout py is broken now types of issue bug report report an issue with the code feature request request for change which adds functionality expected behavior current behavior steps to reproduce for bugs possible solution for bugs context for feature requests your environment version used shell used
1
13,342
10,213,534,767
IssuesEvent
2019-08-14 22:33:05
dotnet/core-setup
https://api.github.com/repos/dotnet/core-setup
opened
VS insertion federation: manufacture and push MSI insertion nupkgs for 2.1/2.2
area-Infrastructure
Port https://github.com/dotnet/core-setup/pull/6913 to 2.1/2.2. @dleeapho, @MichaelSimons, this tracks the work to port MSI nupkg creation back to `release/2.1` and `release/2.2`. We'll have to figure out when to make this happen. /cc @wtgodbe (Core-Setup tracking issue for https://github.com/dotnet/core-sdk/issues/2148 and https://github.com/dotnet/core-sdk/issues/2150.)
1.0
VS insertion federation: manufacture and push MSI insertion nupkgs for 2.1/2.2 - Port https://github.com/dotnet/core-setup/pull/6913 to 2.1/2.2. @dleeapho, @MichaelSimons, this tracks the work to port MSI nupkg creation back to `release/2.1` and `release/2.2`. We'll have to figure out when to make this happen. /cc @wtgodbe (Core-Setup tracking issue for https://github.com/dotnet/core-sdk/issues/2148 and https://github.com/dotnet/core-sdk/issues/2150.)
non_priority
vs insertion federation manufacture and push msi insertion nupkgs for port to dleeapho michaelsimons this tracks the work to port msi nupkg creation back to release and release we ll have to figure out when to make this happen cc wtgodbe core setup tracking issue for and
0
250,638
27,108,649,251
IssuesEvent
2023-02-15 13:56:09
Gal-Doron/Remediate-missing-PR
https://api.github.com/repos/Gal-Doron/Remediate-missing-PR
opened
CVE-2023-25577 (High) detected in Werkzeug-1.0.1-py2.py3-none-any.whl
security vulnerability
## CVE-2023-25577 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary> <p>The comprehensive WSGI web application library.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - ganga-8.5.9.tar.gz (Root Library) - Flask_Login-0.6.2-py3-none-any.whl - :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Remediate-missing-PR/commit/a7570e25bee62662dbc5bd392f664bf6305e5488">a7570e25bee62662dbc5bd392f664bf6305e5488</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Werkzeug is a comprehensive WSGI web application library. Prior to version 2.2.3, Werkzeug's multipart form data parser will parse an unlimited number of parts, including file parts. Parts can be a small amount of bytes, but each requires CPU time to parse and may use more memory as Python data. If a request can be made to an endpoint that accesses `request.data`, `request.form`, `request.files`, or `request.get_data(parse_form_data=False)`, it can cause unexpectedly high resource usage. This allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it. The amount of CPU time required can block worker processes from handling legitimate requests. The amount of RAM required can trigger an out of memory kill of the process. Unlimited file parts can use up memory and file handles. If many concurrent requests are sent continuously, this can exhaust or kill all available workers. Version 2.2.3 contains a patch for this issue. <p>Publish Date: 2023-02-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25577>CVE-2023-25577</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25577">https://www.cve.org/CVERecord?id=CVE-2023-25577</a></p> <p>Release Date: 2023-02-14</p> <p>Fix Resolution: Werkzeug - 2.2.3</p> </p> </details> <p></p>
True
CVE-2023-25577 (High) detected in Werkzeug-1.0.1-py2.py3-none-any.whl - ## CVE-2023-25577 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Werkzeug-1.0.1-py2.py3-none-any.whl</b></p></summary> <p>The comprehensive WSGI web application library.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/cc/94/5f7079a0e00bd6863ef8f1da638721e9da21e5bacee597595b318f71d62e/Werkzeug-1.0.1-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt</p> <p> Dependency Hierarchy: - ganga-8.5.9.tar.gz (Root Library) - Flask_Login-0.6.2-py3-none-any.whl - :x: **Werkzeug-1.0.1-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/Gal-Doron/Remediate-missing-PR/commit/a7570e25bee62662dbc5bd392f664bf6305e5488">a7570e25bee62662dbc5bd392f664bf6305e5488</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Werkzeug is a comprehensive WSGI web application library. Prior to version 2.2.3, Werkzeug's multipart form data parser will parse an unlimited number of parts, including file parts. Parts can be a small amount of bytes, but each requires CPU time to parse and may use more memory as Python data. If a request can be made to an endpoint that accesses `request.data`, `request.form`, `request.files`, or `request.get_data(parse_form_data=False)`, it can cause unexpectedly high resource usage. This allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it. The amount of CPU time required can block worker processes from handling legitimate requests. The amount of RAM required can trigger an out of memory kill of the process. Unlimited file parts can use up memory and file handles. If many concurrent requests are sent continuously, this can exhaust or kill all available workers. Version 2.2.3 contains a patch for this issue. <p>Publish Date: 2023-02-14 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25577>CVE-2023-25577</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25577">https://www.cve.org/CVERecord?id=CVE-2023-25577</a></p> <p>Release Date: 2023-02-14</p> <p>Fix Resolution: Werkzeug - 2.2.3</p> </p> </details> <p></p>
non_priority
cve high detected in werkzeug none any whl cve high severity vulnerability vulnerable library werkzeug none any whl the comprehensive wsgi web application library library home page a href path to dependency file requirements txt path to vulnerable library requirements txt dependency hierarchy ganga tar gz root library flask login none any whl x werkzeug none any whl vulnerable library found in head commit a href found in base branch main vulnerability details werkzeug is a comprehensive wsgi web application library prior to version werkzeug s multipart form data parser will parse an unlimited number of parts including file parts parts can be a small amount of bytes but each requires cpu time to parse and may use more memory as python data if a request can be made to an endpoint that accesses request data request form request files or request get data parse form data false it can cause unexpectedly high resource usage this allows an attacker to cause a denial of service by sending crafted multipart data to an endpoint that will parse it the amount of cpu time required can block worker processes from handling legitimate requests the amount of ram required can trigger an out of memory kill of the process unlimited file parts can use up memory and file handles if many concurrent requests are sent continuously this can exhaust or kill all available workers version contains a patch for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution werkzeug
0
600,647
18,348,979,893
IssuesEvent
2021-10-08 10:03:28
betagouv/service-national-universel
https://api.github.com/repos/betagouv/service-national-universel
closed
feat: add phase2ValidationDate for young
priority-HIGH
Il faut ajouter une propriété au modèle Young : `phase2ValidationDate` (ou un autre nom plus cohérent avec le reste, à checker) qui contient une date qui est mise à jour à chaque fois que le statut de phase 2 d'un jeune passe à "Effectuée". Cela permettra ensuite d'utiliser la VRAIE date (et non pas la date du jour) quand elle est présente dans l'attestation de fin de phase 2 : <img width="1078" alt="Capture d’écran 2021-10-07 à 16 48 04" src="https://user-images.githubusercontent.com/1575946/136408819-c169822f-f62b-4d7b-a380-0d22b6a45c23.png"> Autrement dit dans l'ordre : - Ajouter le champ au modèle - Vérifier dans les controllers, pour que partout ou le statut de phase 2 est mis à jour mettre la date au moment où on enregistre (on met toujours la dernière date, même si le statut à changé plusieurs fois) - Mettre à jour le certificat en prenant cette date, et à défaut la date du jour pour les comptes où ce n'est pas renseigné Source : https://trello.com/c/pcawO2Bt/823-admin-app-figer-la-date-fait-le-de-lattestation-phase-2-%C3%A0-la-date-de-validation-de-la-phase-2-sur-la-plateforme
1.0
feat: add phase2ValidationDate for young - Il faut ajouter une propriété au modèle Young : `phase2ValidationDate` (ou un autre nom plus cohérent avec le reste, à checker) qui contient une date qui est mise à jour à chaque fois que le statut de phase 2 d'un jeune passe à "Effectuée". Cela permettra ensuite d'utiliser la VRAIE date (et non pas la date du jour) quand elle est présente dans l'attestation de fin de phase 2 : <img width="1078" alt="Capture d’écran 2021-10-07 à 16 48 04" src="https://user-images.githubusercontent.com/1575946/136408819-c169822f-f62b-4d7b-a380-0d22b6a45c23.png"> Autrement dit dans l'ordre : - Ajouter le champ au modèle - Vérifier dans les controllers, pour que partout ou le statut de phase 2 est mis à jour mettre la date au moment où on enregistre (on met toujours la dernière date, même si le statut à changé plusieurs fois) - Mettre à jour le certificat en prenant cette date, et à défaut la date du jour pour les comptes où ce n'est pas renseigné Source : https://trello.com/c/pcawO2Bt/823-admin-app-figer-la-date-fait-le-de-lattestation-phase-2-%C3%A0-la-date-de-validation-de-la-phase-2-sur-la-plateforme
priority
feat add for young il faut ajouter une propriété au modèle young ou un autre nom plus cohérent avec le reste à checker qui contient une date qui est mise à jour à chaque fois que le statut de phase d un jeune passe à effectuée cela permettra ensuite d utiliser la vraie date et non pas la date du jour quand elle est présente dans l attestation de fin de phase img width alt capture d’écran à src autrement dit dans l ordre ajouter le champ au modèle vérifier dans les controllers pour que partout ou le statut de phase est mis à jour mettre la date au moment où on enregistre on met toujours la dernière date même si le statut à changé plusieurs fois mettre à jour le certificat en prenant cette date et à défaut la date du jour pour les comptes où ce n est pas renseigné source
1
135,166
12,677,176,930
IssuesEvent
2020-06-19 07:12:28
actions/virtual-environments
https://api.github.com/repos/actions/virtual-environments
closed
Details of included Go version is missing in the included software documentation for windows-2019 and vs2017-win2016 images
OS: Windows awaiting-deployment documentation
**Describe the bug** The included software documents for [`windows-2019`](https://github.com/actions/virtual-environments/blob/master/images/win/Windows2019-Readme.md) and [`vs2017-win2016`](https://github.com/actions/virtual-environments/blob/master/images/win/Windows2016-Readme.md) images are missing details about included Go version as well as Go related environment variables. Running `go version` in the pipeline using both images produces `go version go1.14.3 windows/amd64`, confirming Go is still included. Commits that broke documentation are: `windows-2019`: [44f2342c9e796a4d0c3bd36aff870b38abd6a71b](https://github.com/actions/virtual-environments/commit/44f2342c9e796a4d0c3bd36aff870b38abd6a71b) `vs2017-win2016`: [e91f2138c0b7304831e9aaa43ea47279d4160ef4](https://github.com/actions/virtual-environments/commit/e91f2138c0b7304831e9aaa43ea47279d4160ef4) **Area for Triage**: Go **Question, Bug, or Feature?**: Bug **Virtual environments affected** - [ ] macOS 10.15 - [ ] Ubuntu 16.04 LTS - [ ] Ubuntu 18.04 LTS - [ ] Ubuntu 20.04 LTS - [x] Windows Server 2016 R2 - [x] Windows Server 2019
1.0
Details of included Go version is missing in the included software documentation for windows-2019 and vs2017-win2016 images - **Describe the bug** The included software documents for [`windows-2019`](https://github.com/actions/virtual-environments/blob/master/images/win/Windows2019-Readme.md) and [`vs2017-win2016`](https://github.com/actions/virtual-environments/blob/master/images/win/Windows2016-Readme.md) images are missing details about included Go version as well as Go related environment variables. Running `go version` in the pipeline using both images produces `go version go1.14.3 windows/amd64`, confirming Go is still included. Commits that broke documentation are: `windows-2019`: [44f2342c9e796a4d0c3bd36aff870b38abd6a71b](https://github.com/actions/virtual-environments/commit/44f2342c9e796a4d0c3bd36aff870b38abd6a71b) `vs2017-win2016`: [e91f2138c0b7304831e9aaa43ea47279d4160ef4](https://github.com/actions/virtual-environments/commit/e91f2138c0b7304831e9aaa43ea47279d4160ef4) **Area for Triage**: Go **Question, Bug, or Feature?**: Bug **Virtual environments affected** - [ ] macOS 10.15 - [ ] Ubuntu 16.04 LTS - [ ] Ubuntu 18.04 LTS - [ ] Ubuntu 20.04 LTS - [x] Windows Server 2016 R2 - [x] Windows Server 2019
non_priority
details of included go version is missing in the included software documentation for windows and images describe the bug the included software documents for and images are missing details about included go version as well as go related environment variables running go version in the pipeline using both images produces go version windows confirming go is still included commits that broke documentation are windows area for triage go question bug or feature bug virtual environments affected macos ubuntu lts ubuntu lts ubuntu lts windows server windows server
0
366,214
10,818,362,500
IssuesEvent
2019-11-08 11:53:10
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.google.com - desktop site instead of mobile site
browser-chrome priority-critical
<!-- @browser: Chrome 78.0.3904 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/78.0.3904.87 Safari/537.36 --> <!-- @reported_with: --> **URL**: https://www.google.com **Browser / Version**: Chrome 78.0.3904 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Desktop site instead of mobile site **Description**: Test bug **Steps to Reproduce**: Test bug <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.google.com - desktop site instead of mobile site - <!-- @browser: Chrome 78.0.3904 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/78.0.3904.87 Safari/537.36 --> <!-- @reported_with: --> **URL**: https://www.google.com **Browser / Version**: Chrome 78.0.3904 **Operating System**: Windows 10 **Tested Another Browser**: No **Problem type**: Desktop site instead of mobile site **Description**: Test bug **Steps to Reproduce**: Test bug <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
desktop site instead of mobile site url browser version chrome operating system windows tested another browser no problem type desktop site instead of mobile site description test bug steps to reproduce test bug browser configuration none from with ❤️
1
14,469
17,577,460,540
IssuesEvent
2021-08-15 22:04:10
2i2c-org/pilot-hubs
https://api.github.com/repos/2i2c-org/pilot-hubs
opened
Status Page for our clusters
type: enhancement :label: team-process :label: hub administrator prio: low
# Summary There are many cases where our clusters might be down for one reason or another (e.g. upgrades, outages, etc). In those cases, it's helpful if there is a source of truth for "is 2i2c's infrastructure down, or is it just me?". We should have a place to point users to so that they can quickly answer this question. # Important information The most common service I've seen for this is [`statuspage.io`](https://www.atlassian.com/software/statuspage), which [even has a non-profit discount](https://support.atlassian.com/statuspage/docs/apply-for-a-community-open-source-or-academic-license/). # Tasks to complete - [ ] Decide what kind of service we'd like to use for a status page - [ ] ...figure out steps to implement this - [ ] Document the new status page in our user and team documentation
1.0
Status Page for our clusters - # Summary There are many cases where our clusters might be down for one reason or another (e.g. upgrades, outages, etc). In those cases, it's helpful if there is a source of truth for "is 2i2c's infrastructure down, or is it just me?". We should have a place to point users to so that they can quickly answer this question. # Important information The most common service I've seen for this is [`statuspage.io`](https://www.atlassian.com/software/statuspage), which [even has a non-profit discount](https://support.atlassian.com/statuspage/docs/apply-for-a-community-open-source-or-academic-license/). # Tasks to complete - [ ] Decide what kind of service we'd like to use for a status page - [ ] ...figure out steps to implement this - [ ] Document the new status page in our user and team documentation
non_priority
status page for our clusters summary there are many cases where our clusters might be down for one reason or another e g upgrades outages etc in those cases it s helpful if there is a source of truth for is s infrastructure down or is it just me we should have a place to point users to so that they can quickly answer this question important information the most common service i ve seen for this is which tasks to complete decide what kind of service we d like to use for a status page figure out steps to implement this document the new status page in our user and team documentation
0
30,849
13,340,653,863
IssuesEvent
2020-08-28 14:45:12
terraform-providers/terraform-provider-azurerm
https://api.github.com/repos/terraform-providers/terraform-provider-azurerm
closed
Fixup pull request #7749 (`azurerm_*_virtual_machine` - `allow_extensions_operations` can now be updated)
bug service/virtual-machine
### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform (and AzureRM Provider) Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * `azurerm_linux_virtual_machine` * `azurerm_windows_virtual_machine` The code change for https://github.com/terraform-providers/terraform-provider-azurerm/issues/7070 via pull request https://github.com/terraform-providers/terraform-provider-azurerm/pull/7749 introduced a regression. I already tried to explain it here: https://github.com/terraform-providers/terraform-provider-azurerm/pull/7749#discussion_r454835017 This code in azurerm/internal/services/compute/{linux|windows}_virtual_machine_resource.go is wrong: linux: https://github.com/terraform-providers/terraform-provider-azurerm/blob/55e02f697557906dd4dae43657d5e6381138b491/azurerm/internal/services/compute/linux_virtual_machine_resource.go#L390 windows: https://github.com/terraform-providers/terraform-provider-azurerm/blob/55e02f697557906dd4dae43657d5e6381138b491/azurerm/internal/services/compute/windows_virtual_machine_resource.go#L414 ``` if !provisionVMAgent && allowExtensionOperations { return fmt.Errorf("`allow_extension_operations` cannot be set to `true` when `provision_vm_agent` is set to `false`") } ``` You can change the allow_extension_operations value, even if provision_vm_agent is false. If you did not deploy with provision_vm_agent you can still install it manually. This is explained here: https://docs.microsoft.com/en-us/azure/virtual-machines/troubleshooting/install-vm-agent-offline After that, you can enable extensions, even with provision_vm_agent set to false. ### Expected Behavior Allow to set allow_extension_operations independent of provision_vm_agent. The AZURE API should fail if conditions are not fullfilled. ### Actual Behavior If provision_vm_agent is false, you cannot set allow_extension_operations to true. The provider does not allow it. ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. Create an instance with provision_vm_agent = false and allow_extension_operations = false 2. Install vm agent manually as documented by AZURE (see link above) 3. Set allow_extension_operations = true ### References * #7070 * #7749
1.0
Fixup pull request #7749 (`azurerm_*_virtual_machine` - `allow_extensions_operations` can now be updated) - ### Community Note * Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform (and AzureRM Provider) Version <!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). ---> ### Affected Resource(s) <!--- Please list the affected resources and data sources. ---> * `azurerm_linux_virtual_machine` * `azurerm_windows_virtual_machine` The code change for https://github.com/terraform-providers/terraform-provider-azurerm/issues/7070 via pull request https://github.com/terraform-providers/terraform-provider-azurerm/pull/7749 introduced a regression. I already tried to explain it here: https://github.com/terraform-providers/terraform-provider-azurerm/pull/7749#discussion_r454835017 This code in azurerm/internal/services/compute/{linux|windows}_virtual_machine_resource.go is wrong: linux: https://github.com/terraform-providers/terraform-provider-azurerm/blob/55e02f697557906dd4dae43657d5e6381138b491/azurerm/internal/services/compute/linux_virtual_machine_resource.go#L390 windows: https://github.com/terraform-providers/terraform-provider-azurerm/blob/55e02f697557906dd4dae43657d5e6381138b491/azurerm/internal/services/compute/windows_virtual_machine_resource.go#L414 ``` if !provisionVMAgent && allowExtensionOperations { return fmt.Errorf("`allow_extension_operations` cannot be set to `true` when `provision_vm_agent` is set to `false`") } ``` You can change the allow_extension_operations value, even if provision_vm_agent is false. If you did not deploy with provision_vm_agent you can still install it manually. This is explained here: https://docs.microsoft.com/en-us/azure/virtual-machines/troubleshooting/install-vm-agent-offline After that, you can enable extensions, even with provision_vm_agent set to false. ### Expected Behavior Allow to set allow_extension_operations independent of provision_vm_agent. The AZURE API should fail if conditions are not fullfilled. ### Actual Behavior If provision_vm_agent is false, you cannot set allow_extension_operations to true. The provider does not allow it. ### Steps to Reproduce <!--- Please list the steps required to reproduce the issue. ---> 1. Create an instance with provision_vm_agent = false and allow_extension_operations = false 2. Install vm agent manually as documented by AZURE (see link above) 3. Set allow_extension_operations = true ### References * #7070 * #7749
non_priority
fixup pull request azurerm virtual machine allow extensions operations can now be updated community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version affected resource s azurerm linux virtual machine azurerm windows virtual machine the code change for via pull request introduced a regression i already tried to explain it here this code in azurerm internal services compute linux windows virtual machine resource go is wrong linux windows if provisionvmagent allowextensionoperations return fmt errorf allow extension operations cannot be set to true when provision vm agent is set to false you can change the allow extension operations value even if provision vm agent is false if you did not deploy with provision vm agent you can still install it manually this is explained here after that you can enable extensions even with provision vm agent set to false expected behavior allow to set allow extension operations independent of provision vm agent the azure api should fail if conditions are not fullfilled actual behavior if provision vm agent is false you cannot set allow extension operations to true the provider does not allow it steps to reproduce create an instance with provision vm agent false and allow extension operations false install vm agent manually as documented by azure see link above set allow extension operations true references
0
19,910
10,429,703,500
IssuesEvent
2019-09-17 03:53:42
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
How to Set Email for Security Recommendations?
Pri2 cxp product-question security-center/svc triaged
Security threats have options to configure email, but i could not find an option to email send security Recommendations when detected. Please let me know how to do so. Thanks. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 06951b9a-8933-8307-371a-e9b008eb8f65 * Version Independent ID: 7a5b1269-5397-17a5-634d-4e39773dbf93 * Content: [Use Azure Security Center recommendations to enhance security](https://docs.microsoft.com/en-us/azure/security-center/security-center-using-recommendations#feedback) * Content Source: [articles/security-center/security-center-using-recommendations.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/security-center-using-recommendations.md) * Service: **security-center** * GitHub Login: @monhaber * Microsoft Alias: **v-mohabe**
True
How to Set Email for Security Recommendations? - Security threats have options to configure email, but i could not find an option to email send security Recommendations when detected. Please let me know how to do so. Thanks. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 06951b9a-8933-8307-371a-e9b008eb8f65 * Version Independent ID: 7a5b1269-5397-17a5-634d-4e39773dbf93 * Content: [Use Azure Security Center recommendations to enhance security](https://docs.microsoft.com/en-us/azure/security-center/security-center-using-recommendations#feedback) * Content Source: [articles/security-center/security-center-using-recommendations.md](https://github.com/Microsoft/azure-docs/blob/master/articles/security-center/security-center-using-recommendations.md) * Service: **security-center** * GitHub Login: @monhaber * Microsoft Alias: **v-mohabe**
non_priority
how to set email for security recommendations security threats have options to configure email but i could not find an option to email send security recommendations when detected please let me know how to do so thanks document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service security center github login monhaber microsoft alias v mohabe
0
272,986
20,766,561,933
IssuesEvent
2022-03-15 21:15:27
adtzlr/felupe
https://api.github.com/repos/adtzlr/felupe
closed
Re-write documentation
documentation
### Technical improvements - use a documentation system with API generation (Sphinx) - generate version-based documentations (Read the docs) ### some links - https://www.writethedocs.org/guide/writing/beginners-guide-to-docs/ - https://diataxis.fr/ - https://www.sphinx-doc.org/en/master/tutorial/index.html ### structure suggested by [diataxis](https://diataxis.fr/) - tutorials (learning-oriented): no specific code options, straight-forward *lessons to learn* - how-to guides (task-oriented): e.g. how to setup linear- and bilinearforms, external pressure contributions, etc. - explanation (understanding-oriented): former *theory guide* - reference (information-oriented): (*will be added*) **hopefully** auto-generated by sphinx
1.0
Re-write documentation - ### Technical improvements - use a documentation system with API generation (Sphinx) - generate version-based documentations (Read the docs) ### some links - https://www.writethedocs.org/guide/writing/beginners-guide-to-docs/ - https://diataxis.fr/ - https://www.sphinx-doc.org/en/master/tutorial/index.html ### structure suggested by [diataxis](https://diataxis.fr/) - tutorials (learning-oriented): no specific code options, straight-forward *lessons to learn* - how-to guides (task-oriented): e.g. how to setup linear- and bilinearforms, external pressure contributions, etc. - explanation (understanding-oriented): former *theory guide* - reference (information-oriented): (*will be added*) **hopefully** auto-generated by sphinx
non_priority
re write documentation technical improvements use a documentation system with api generation sphinx generate version based documentations read the docs some links structure suggested by tutorials learning oriented no specific code options straight forward lessons to learn how to guides task oriented e g how to setup linear and bilinearforms external pressure contributions etc explanation understanding oriented former theory guide reference information oriented will be added hopefully auto generated by sphinx
0
359,973
10,683,051,428
IssuesEvent
2019-10-22 07:26:04
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.unionesarda.it - Unable to access article link
ML Correct ML OFF browser-fenix engine-gecko priority-normal severity-important status-needsinfo-cipriansv
<!-- @browser: Firefox Mobile 70.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.unionesarda.it/ **Browser / Version**: Firefox Mobile 70.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: Can't click **Steps to Reproduce**: Click on link <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.unionesarda.it - Unable to access article link - <!-- @browser: Firefox Mobile 70.0 --> <!-- @ua_header: Mozilla/5.0 (Android 9; Mobile; rv:70.0) Gecko/70.0 Firefox/70.0 --> <!-- @reported_with: --> <!-- @extra_labels: browser-fenix --> **URL**: https://www.unionesarda.it/ **Browser / Version**: Firefox Mobile 70.0 **Operating System**: Android **Tested Another Browser**: No **Problem type**: Site is not usable **Description**: Can't click **Steps to Reproduce**: Click on link <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
unable to access article link url browser version firefox mobile operating system android tested another browser no problem type site is not usable description can t click steps to reproduce click on link browser configuration none from with ❤️
1
497,724
14,383,566,057
IssuesEvent
2020-12-02 09:18:05
towaquimbayo/kiku
https://api.github.com/repos/towaquimbayo/kiku
opened
Gallery lightbox photo shows a cropped image
Priority: Low Severity: Low enhancement
### Issue Gallery images in the lightbox/carousel don't show the whole images. They're zoomed in and cropped. ### Steps Click on any of the gallery images. ### Suggestion Expand the width of the lightbox/carousel. ![Screenshot_2020-12-02 KIKU](https://user-images.githubusercontent.com/24981064/100853056-2944f800-343c-11eb-8826-03a6770fd79b.png)
1.0
Gallery lightbox photo shows a cropped image - ### Issue Gallery images in the lightbox/carousel don't show the whole images. They're zoomed in and cropped. ### Steps Click on any of the gallery images. ### Suggestion Expand the width of the lightbox/carousel. ![Screenshot_2020-12-02 KIKU](https://user-images.githubusercontent.com/24981064/100853056-2944f800-343c-11eb-8826-03a6770fd79b.png)
priority
gallery lightbox photo shows a cropped image issue gallery images in the lightbox carousel don t show the whole images they re zoomed in and cropped steps click on any of the gallery images suggestion expand the width of the lightbox carousel
1
615,551
19,256,885,945
IssuesEvent
2021-12-09 12:18:01
epam/ketcher
https://api.github.com/repos/epam/ketcher
closed
Settings - Preview images of structures (Templates Library, FG Library) do not respect global display settings
bug priority: low epic: functional groups
**Steps to Reproduce** 1. Launch Ketcher with the following display settings (settings: 'Atoms', 'Show hydrogen labels' - on). 2. Open FG Library. 3. Add FG on canvas and Expanded FG abb. 4. Change settings: 'Atoms', 'Show hydrogen labels' - on). 5. Open Templates Library. 6. Add any Template on canvas. **Expected behavior** When Launch Ketcher with the following display settings (settings: 'Atoms', 'Show hydrogen labels' - on) all structures (in FG Library and in Templates Library) should be displayed according to these settings. **Actual behavior** In FG Library and Templates Library, preview images of structures do not respect global display settings. https://user-images.githubusercontent.com/90637536/139446193-5013bf9e-d71a-40c3-92f2-91fb43cc6f85.mp4 **Desktop:** OS: Windows 10 Browser: Chrome Version 93.0.4577.63, Firefox Version 92.0, Microsoft Edge Version 93.0.961.47
1.0
Settings - Preview images of structures (Templates Library, FG Library) do not respect global display settings - **Steps to Reproduce** 1. Launch Ketcher with the following display settings (settings: 'Atoms', 'Show hydrogen labels' - on). 2. Open FG Library. 3. Add FG on canvas and Expanded FG abb. 4. Change settings: 'Atoms', 'Show hydrogen labels' - on). 5. Open Templates Library. 6. Add any Template on canvas. **Expected behavior** When Launch Ketcher with the following display settings (settings: 'Atoms', 'Show hydrogen labels' - on) all structures (in FG Library and in Templates Library) should be displayed according to these settings. **Actual behavior** In FG Library and Templates Library, preview images of structures do not respect global display settings. https://user-images.githubusercontent.com/90637536/139446193-5013bf9e-d71a-40c3-92f2-91fb43cc6f85.mp4 **Desktop:** OS: Windows 10 Browser: Chrome Version 93.0.4577.63, Firefox Version 92.0, Microsoft Edge Version 93.0.961.47
priority
settings preview images of structures templates library fg library do not respect global display settings steps to reproduce launch ketcher with the following display settings settings atoms show hydrogen labels on open fg library add fg on canvas and expanded fg abb change settings atoms show hydrogen labels on open templates library add any template on canvas expected behavior when launch ketcher with the following display settings settings atoms show hydrogen labels on all structures in fg library and in templates library should be displayed according to these settings actual behavior in fg library and templates library preview images of structures do not respect global display settings desktop os windows browser chrome version firefox version microsoft edge version
1
418,121
12,193,601,216
IssuesEvent
2020-04-29 14:38:11
domialex/Sidekick
https://api.github.com/repos/domialex/Sidekick
opened
Limit maximum requests in a given time
Priority: High Status: Available Type: Enhancement
This is now a priority since spamming the servers can be detected as malicious intent by GGG and lock accounts. This is happening a lot with POE-Overlay right now. We need to add a debounce on requests and/or a search button. Mentioned in #362.
1.0
Limit maximum requests in a given time - This is now a priority since spamming the servers can be detected as malicious intent by GGG and lock accounts. This is happening a lot with POE-Overlay right now. We need to add a debounce on requests and/or a search button. Mentioned in #362.
priority
limit maximum requests in a given time this is now a priority since spamming the servers can be detected as malicious intent by ggg and lock accounts this is happening a lot with poe overlay right now we need to add a debounce on requests and or a search button mentioned in
1
69,446
13,249,928,754
IssuesEvent
2020-08-19 21:45:52
aws/aws-cdk
https://api.github.com/repos/aws/aws-cdk
closed
Region config for S3DeployAction
@aws-cdk/aws-codepipeline @aws-cdk/core effort/medium feature-request in-progress
<!-- short description of the feature you are proposing: --> Please add `region` configuration in S3DeployAction for cross-region deployment. ### Use Case <!-- why do you need this feature? --> I have a pipeline that supports cross-region deployment in us-west-2. A deployment stage of my pipeline looks like: ``` *Pipeline is in us-west-2 ... Deploy stage: CF CF (us-east-1) (us-west-2) | S3 (us-east-1) ... ``` I can set `region` for CloudFormationCreateUpdateStackAction, so it can support cross region deployment. However, S3DeployAction does not have `region` config in S3DeployActionProps. The target S3 bucket is created by the us-east-1 CF action above and the bucket is located in us-east-1. Therefore, S3DeployAction fails with the below message in console: ``` Invalid action configuration The bucket named [BUCKET_NAME] is not located in the us-west-2 AWS region ``` I manually changed the region of the action in the console to resolve the problem. ### Proposed Solution <!-- Please include prototype/workaround/sketch/reference implementation: --> It would be great if you can add `region` to S3DeployActionProps and support cross-region deployment for S3DeployAction. ### Other <!-- e.g. detailed explanation, stacktraces, related issues, suggestions on how to fix, links for us to have context, eg. associated pull-request, stackoverflow, gitter, etc --> * [ ] :wave: I may be able to implement this feature request * [ ] :warning: This feature might incur a breaking change --- This is a :rocket: Feature Request
1.0
Region config for S3DeployAction - <!-- short description of the feature you are proposing: --> Please add `region` configuration in S3DeployAction for cross-region deployment. ### Use Case <!-- why do you need this feature? --> I have a pipeline that supports cross-region deployment in us-west-2. A deployment stage of my pipeline looks like: ``` *Pipeline is in us-west-2 ... Deploy stage: CF CF (us-east-1) (us-west-2) | S3 (us-east-1) ... ``` I can set `region` for CloudFormationCreateUpdateStackAction, so it can support cross region deployment. However, S3DeployAction does not have `region` config in S3DeployActionProps. The target S3 bucket is created by the us-east-1 CF action above and the bucket is located in us-east-1. Therefore, S3DeployAction fails with the below message in console: ``` Invalid action configuration The bucket named [BUCKET_NAME] is not located in the us-west-2 AWS region ``` I manually changed the region of the action in the console to resolve the problem. ### Proposed Solution <!-- Please include prototype/workaround/sketch/reference implementation: --> It would be great if you can add `region` to S3DeployActionProps and support cross-region deployment for S3DeployAction. ### Other <!-- e.g. detailed explanation, stacktraces, related issues, suggestions on how to fix, links for us to have context, eg. associated pull-request, stackoverflow, gitter, etc --> * [ ] :wave: I may be able to implement this feature request * [ ] :warning: This feature might incur a breaking change --- This is a :rocket: Feature Request
non_priority
region config for please add region configuration in for cross region deployment use case i have a pipeline that supports cross region deployment in us west a deployment stage of my pipeline looks like pipeline is in us west deploy stage cf cf us east us west us east i can set region for cloudformationcreateupdatestackaction so it can support cross region deployment however does not have region config in the target bucket is created by the us east cf action above and the bucket is located in us east therefore fails with the below message in console invalid action configuration the bucket named is not located in the us west aws region i manually changed the region of the action in the console to resolve the problem proposed solution it would be great if you can add region to and support cross region deployment for other e g detailed explanation stacktraces related issues suggestions on how to fix links for us to have context eg associated pull request stackoverflow gitter etc wave i may be able to implement this feature request warning this feature might incur a breaking change this is a rocket feature request
0
527,414
15,342,304,256
IssuesEvent
2021-02-27 15:41:46
webpack-contrib/copy-webpack-plugin
https://api.github.com/repos/webpack-contrib/copy-webpack-plugin
closed
ERROR in The "path" argument must be of type string. Received an instance of Object
priority: 4 (important) semver: Patch severity: 3 (broken) type: Bug upstream
When compiling after updating from copy-webpack-plugin@6.0.0 to 6.0.1, I got this error (that seems be thrown by mini-css-extract-plugin). """ TypeError [ERR_INVALID_ARG_TYPE]: The "path" argument must be of type string. Received an instance of Object """ I am not sure if it's an actual bug, but I post this here since the only thing that changed when I got this error is the version for mini-css-extract-plugin. I can try to make a minimal project to reproduce the errors, but seems overkill for me right now. - Operating System: MacOS 10.15.4 (happen also int he Google Cloud build environment) - Node Version: 14.2.0 (and node:13-alpine on Gcloud build) - NPM Version: 6.14.5 - webpack Version: 4.43.0 - copy-webpack-plugin Version: 6.0.1
1.0
ERROR in The "path" argument must be of type string. Received an instance of Object - When compiling after updating from copy-webpack-plugin@6.0.0 to 6.0.1, I got this error (that seems be thrown by mini-css-extract-plugin). """ TypeError [ERR_INVALID_ARG_TYPE]: The "path" argument must be of type string. Received an instance of Object """ I am not sure if it's an actual bug, but I post this here since the only thing that changed when I got this error is the version for mini-css-extract-plugin. I can try to make a minimal project to reproduce the errors, but seems overkill for me right now. - Operating System: MacOS 10.15.4 (happen also int he Google Cloud build environment) - Node Version: 14.2.0 (and node:13-alpine on Gcloud build) - NPM Version: 6.14.5 - webpack Version: 4.43.0 - copy-webpack-plugin Version: 6.0.1
priority
error in the path argument must be of type string received an instance of object when compiling after updating from copy webpack plugin to i got this error that seems be thrown by mini css extract plugin typeerror the path argument must be of type string received an instance of object i am not sure if it s an actual bug but i post this here since the only thing that changed when i got this error is the version for mini css extract plugin i can try to make a minimal project to reproduce the errors but seems overkill for me right now operating system macos happen also int he google cloud build environment node version and node alpine on gcloud build npm version webpack version copy webpack plugin version
1
123,552
4,864,776,904
IssuesEvent
2016-11-14 18:58:18
stillmuseum/worksonpaper
https://api.github.com/repos/stillmuseum/worksonpaper
closed
New photo in essay needs to be formatted correctly
A priority
We've added a new inline image to the essay page (http://pubs.clyffordstillmuseum.org/worksonpaperdev/how-to-render-the-sublime/). The image needs to be formatted correctly or have styles applied consistent with the treatment of the other inline images. ![screen shot 2016-11-14 at 11 45 49 am](https://cloud.githubusercontent.com/assets/7491703/20277905/daf877e0-aa5f-11e6-8aea-406ead317546.png)
1.0
New photo in essay needs to be formatted correctly - We've added a new inline image to the essay page (http://pubs.clyffordstillmuseum.org/worksonpaperdev/how-to-render-the-sublime/). The image needs to be formatted correctly or have styles applied consistent with the treatment of the other inline images. ![screen shot 2016-11-14 at 11 45 49 am](https://cloud.githubusercontent.com/assets/7491703/20277905/daf877e0-aa5f-11e6-8aea-406ead317546.png)
priority
new photo in essay needs to be formatted correctly we ve added a new inline image to the essay page the image needs to be formatted correctly or have styles applied consistent with the treatment of the other inline images
1
1,181
2,508,712,393
IssuesEvent
2015-01-13 06:51:36
RomanHargrave/cuelib
https://api.github.com/repos/RomanHargrave/cuelib
closed
TrackCutter should be configurable though a properties file.
enhancement imported Milestone-Release1.3.0 Priority-High Usability
_From [monsterm...@gmail.com](https://code.google.com/u/114086941947790110678/) on May 11, 2008 08:25:57_ TrackCutter accepts many parameters, and some of these parameters can become quite long. Since many parameter will often be reused, it would be useful if these could be configured though a properties file, avoiding have to retype them each time. _Original issue: http://code.google.com/p/cuelib/issues/detail?id=9_
1.0
TrackCutter should be configurable though a properties file. - _From [monsterm...@gmail.com](https://code.google.com/u/114086941947790110678/) on May 11, 2008 08:25:57_ TrackCutter accepts many parameters, and some of these parameters can become quite long. Since many parameter will often be reused, it would be useful if these could be configured though a properties file, avoiding have to retype them each time. _Original issue: http://code.google.com/p/cuelib/issues/detail?id=9_
priority
trackcutter should be configurable though a properties file from on may trackcutter accepts many parameters and some of these parameters can become quite long since many parameter will often be reused it would be useful if these could be configured though a properties file avoiding have to retype them each time original issue
1
260,106
8,204,323,879
IssuesEvent
2018-09-03 05:55:09
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
github.com - see bug description
browser-firefox priority-critical type-stylo
<!-- @browser: Firefox 59.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:59.0) Gecko/20100101 Firefox/59.0 --> <!-- @reported_with: desktop-reporter --> <!-- @extra_labels: type-stylo --> **URL**: https://github.com/sessions/two-factor/security-key **Browser / Version**: Firefox 59.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: User agent detection prevents U2F usage **Steps to Reproduce**: Despite U2F working with my google account, facebook, and Yubico's test page, Github relies on user agent detection instead of feature detection to decide whether or not to enable use of U2F login. layout.css.servo.enabled: true [![Screenshot Description](https://webcompat.com/uploads/2018/1/0fae2ab2-443d-4786-844e-117fe68df3a6-thumb.jpg)](https://webcompat.com/uploads/2018/1/0fae2ab2-443d-4786-844e-117fe68df3a6.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
github.com - see bug description - <!-- @browser: Firefox 59.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:59.0) Gecko/20100101 Firefox/59.0 --> <!-- @reported_with: desktop-reporter --> <!-- @extra_labels: type-stylo --> **URL**: https://github.com/sessions/two-factor/security-key **Browser / Version**: Firefox 59.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: User agent detection prevents U2F usage **Steps to Reproduce**: Despite U2F working with my google account, facebook, and Yubico's test page, Github relies on user agent detection instead of feature detection to decide whether or not to enable use of U2F login. layout.css.servo.enabled: true [![Screenshot Description](https://webcompat.com/uploads/2018/1/0fae2ab2-443d-4786-844e-117fe68df3a6-thumb.jpg)](https://webcompat.com/uploads/2018/1/0fae2ab2-443d-4786-844e-117fe68df3a6.jpg) _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
github com see bug description url browser version firefox operating system windows tested another browser yes problem type something else description user agent detection prevents usage steps to reproduce despite working with my google account facebook and yubico s test page github relies on user agent detection instead of feature detection to decide whether or not to enable use of login layout css servo enabled true from with ❤️
1
72,367
8,724,238,153
IssuesEvent
2018-12-10 03:54:10
hackforla/heart
https://api.github.com/repos/hackforla/heart
opened
Design the intake form
design
Design the intake form using bootstrap or a similar set of UI components. **Notes** 1) Form needs to include everything listed in the IA up to Profile: https://github.com/hackforla/heart/wiki/Information-Architecture Should match the paper form. 2) Put the clinic attended at the top as two fields Location and Date (where date autopopulates but is editable) Design intent is for this to save for every user until changed and saved 3) CTA to submit form is "Save and Complete" 4) Add the checkboxes for all the agreements (we'll hold on signature for now) 5) Include email, phone OR mailing address as a way to get notified about status (with the agreement) 6) When form submits successfully, should say "firstName lastName intake saved" as a toast on top of a new blank form. 7) If form is not saved successfully, show "Intake not saved. Please retry." AND keep them on the form with all the info filled in 8) Form should cache info so a page refresh doesn't clear it out. Only clear on submit. 9) If any of the required fields are not filled in, kick them up to the first unfilled field and show the error near the field
1.0
Design the intake form - Design the intake form using bootstrap or a similar set of UI components. **Notes** 1) Form needs to include everything listed in the IA up to Profile: https://github.com/hackforla/heart/wiki/Information-Architecture Should match the paper form. 2) Put the clinic attended at the top as two fields Location and Date (where date autopopulates but is editable) Design intent is for this to save for every user until changed and saved 3) CTA to submit form is "Save and Complete" 4) Add the checkboxes for all the agreements (we'll hold on signature for now) 5) Include email, phone OR mailing address as a way to get notified about status (with the agreement) 6) When form submits successfully, should say "firstName lastName intake saved" as a toast on top of a new blank form. 7) If form is not saved successfully, show "Intake not saved. Please retry." AND keep them on the form with all the info filled in 8) Form should cache info so a page refresh doesn't clear it out. Only clear on submit. 9) If any of the required fields are not filled in, kick them up to the first unfilled field and show the error near the field
non_priority
design the intake form design the intake form using bootstrap or a similar set of ui components notes form needs to include everything listed in the ia up to profile should match the paper form put the clinic attended at the top as two fields location and date where date autopopulates but is editable design intent is for this to save for every user until changed and saved cta to submit form is save and complete add the checkboxes for all the agreements we ll hold on signature for now include email phone or mailing address as a way to get notified about status with the agreement when form submits successfully should say firstname lastname intake saved as a toast on top of a new blank form if form is not saved successfully show intake not saved please retry and keep them on the form with all the info filled in form should cache info so a page refresh doesn t clear it out only clear on submit if any of the required fields are not filled in kick them up to the first unfilled field and show the error near the field
0
217,922
7,329,036,604
IssuesEvent
2018-03-05 02:11:56
InQuest/ThreatKB
https://api.github.com/repos/InQuest/ThreatKB
closed
(Staging) Release notes incorrectly display metadata values
bug high-priority
After creating a Release and exporting the Release Notes text file, the metadata values do not actually show the Yara signatures true metadata values - instead they display the values assigned by Metadata's mappings: - `Confidence`, `Severity`, and `Description` are incorrect - `EventID`, `Name`, and `Category` **are** correct **Examples from exported Release notes:** ``` EventID: 1000008 Name: FID_Java_Archive Category: FileID Confidence: confidence Severity: severity Description: description EventID: 1000007 Name: FID_ELF_Binary Category: FileID Confidence: confidence Severity: severity Description: description ```
1.0
(Staging) Release notes incorrectly display metadata values - After creating a Release and exporting the Release Notes text file, the metadata values do not actually show the Yara signatures true metadata values - instead they display the values assigned by Metadata's mappings: - `Confidence`, `Severity`, and `Description` are incorrect - `EventID`, `Name`, and `Category` **are** correct **Examples from exported Release notes:** ``` EventID: 1000008 Name: FID_Java_Archive Category: FileID Confidence: confidence Severity: severity Description: description EventID: 1000007 Name: FID_ELF_Binary Category: FileID Confidence: confidence Severity: severity Description: description ```
priority
staging release notes incorrectly display metadata values after creating a release and exporting the release notes text file the metadata values do not actually show the yara signatures true metadata values instead they display the values assigned by metadata s mappings confidence severity and description are incorrect eventid name and category are correct examples from exported release notes eventid name fid java archive category fileid confidence confidence severity severity description description eventid name fid elf binary category fileid confidence confidence severity severity description description
1
126,293
10,417,572,119
IssuesEvent
2019-09-14 23:02:17
neovim/neovim
https://api.github.com/repos/neovim/neovim
closed
tests: test_python{2,3}.vim fails
bug tests
`make oldtest TEST_FILE=test_python3.res` fails: > AttributeError: 'LegacyVim' object has no attribute 'bindeval' ``` From test_python3.vim: Found errors in Test_vim_function(): function RunTheTest[37]..Test_vim_function line 13: Expected False but got 'Vim(return):Error invoking ''python_execute'' on channel 3 (python3-script-host):\nTraceback (most recent call last):\n File "<string>", line 1, in <module>\nAttributeError: ''LegacyVim'' object has no attribute ''bindeval''' function RunTheTest[37]..Test_vim_function line 20: Expected False but got 'Vim(return):Error invoking ''python_execute'' on channel 3 (python3-script-host):\nTraceback (most recent call last):\n File "<string>", line 1, in <module>\nAttributeError: ''LegacyVim'' object has no attribute ''Function''' Caught exception in Test_vim_function(): Vim(return):Error invoking 'python_execute' on channel 3 (python3-script-host):Traceback (most recent call last): File "<string>", line 1, in <module>NameError: name 'f' is not defined @ function RunTheTest[37]..Test_vim_function[23]..provider#python3#Call, line 18 ``` This is not catched on CI due to: 1. `make oldtest` passing silently on Windows/AppVeyor (a real waste on resources btw) 2. `pynvim` not being installed properly on Travis, and the test being skipped there Started working on this / debugging it in https://github.com/neovim/neovim/pull/10269.
1.0
tests: test_python{2,3}.vim fails - `make oldtest TEST_FILE=test_python3.res` fails: > AttributeError: 'LegacyVim' object has no attribute 'bindeval' ``` From test_python3.vim: Found errors in Test_vim_function(): function RunTheTest[37]..Test_vim_function line 13: Expected False but got 'Vim(return):Error invoking ''python_execute'' on channel 3 (python3-script-host):\nTraceback (most recent call last):\n File "<string>", line 1, in <module>\nAttributeError: ''LegacyVim'' object has no attribute ''bindeval''' function RunTheTest[37]..Test_vim_function line 20: Expected False but got 'Vim(return):Error invoking ''python_execute'' on channel 3 (python3-script-host):\nTraceback (most recent call last):\n File "<string>", line 1, in <module>\nAttributeError: ''LegacyVim'' object has no attribute ''Function''' Caught exception in Test_vim_function(): Vim(return):Error invoking 'python_execute' on channel 3 (python3-script-host):Traceback (most recent call last): File "<string>", line 1, in <module>NameError: name 'f' is not defined @ function RunTheTest[37]..Test_vim_function[23]..provider#python3#Call, line 18 ``` This is not catched on CI due to: 1. `make oldtest` passing silently on Windows/AppVeyor (a real waste on resources btw) 2. `pynvim` not being installed properly on Travis, and the test being skipped there Started working on this / debugging it in https://github.com/neovim/neovim/pull/10269.
non_priority
tests test python vim fails make oldtest test file test res fails attributeerror legacyvim object has no attribute bindeval from test vim found errors in test vim function function runthetest test vim function line expected false but got vim return error invoking python execute on channel script host ntraceback most recent call last n file line in nattributeerror legacyvim object has no attribute bindeval function runthetest test vim function line expected false but got vim return error invoking python execute on channel script host ntraceback most recent call last n file line in nattributeerror legacyvim object has no attribute function caught exception in test vim function vim return error invoking python execute on channel script host traceback most recent call last file line in nameerror name f is not defined function runthetest test vim function provider call line this is not catched on ci due to make oldtest passing silently on windows appveyor a real waste on resources btw pynvim not being installed properly on travis and the test being skipped there started working on this debugging it in
0
435,871
12,542,469,353
IssuesEvent
2020-06-05 14:05:46
geosolutions-it/MapStore2
https://api.github.com/repos/geosolutions-it/MapStore2
opened
Improve feature info template editor to support video, iframe and audio tags.
Priority: High enhancement investigation
## Description <!-- A few sentences describing new feature --> <!-- screenshot, video, or link to mockup/prototype are welcome --> Currently video, iframe and audio tags are not supported in the template editor of feature info. Contextually we should evaluate the replacement of `react-quill` with `react-draft-wysiwyg` (new editor used in GeoStory) **What kind of improvement you want to add?** (check one with "x", remove the others) - [x] Other... Please describe: Feature improvement ## Other useful information
1.0
Improve feature info template editor to support video, iframe and audio tags. - ## Description <!-- A few sentences describing new feature --> <!-- screenshot, video, or link to mockup/prototype are welcome --> Currently video, iframe and audio tags are not supported in the template editor of feature info. Contextually we should evaluate the replacement of `react-quill` with `react-draft-wysiwyg` (new editor used in GeoStory) **What kind of improvement you want to add?** (check one with "x", remove the others) - [x] Other... Please describe: Feature improvement ## Other useful information
priority
improve feature info template editor to support video iframe and audio tags description currently video iframe and audio tags are not supported in the template editor of feature info contextually we should evaluate the replacement of react quill with react draft wysiwyg new editor used in geostory what kind of improvement you want to add check one with x remove the others other please describe feature improvement other useful information
1
436,828
12,554,034,026
IssuesEvent
2020-06-07 00:23:40
eclipse-ee4j/glassfish
https://api.github.com/repos/eclipse-ee4j/glassfish
closed
The attribute [...] of class [...] is mapped to a primary key column in the database. Updates are not allowed
Component: entity-persistence ERR: Assignee Priority: Minor Stale Type: Bug
<p>This is a reopen of issue 3868, as the defect (or a related defect) is still present as of Toplink Essentials 2.1 (Build 34 (05/21/2008)). I'd have posted this on that issue, but for doesn't appear I have rights to comment on issues I didn't create.</p> <p>Looking under the debugger, what I see is Toplink Essentials applying a changeSet entry, with identical values for the old and new primary key. The simple and effective fix would be to recognize when old and new values match, and not apply the update and not throw the ValidationException.</p> <p>In my scenerio, I've got a Menu object with: <pre> @Entity @Table(name = "MENU") public class Menu extends PersistableObject { @ManyToOne(fetch = FetchType.EAGER, optional = true) @JoinColumn(name = "parentId") private Menu parent = null; @OneToMany(cascade=CascadeType.ALL, fetch = FetchType.LAZY, mappedBy = "parent") private List<Menu> subMenus = new ArrayList<Menu>(); </pre></p> <p>The trace is: <pre> Exception [TOPLINK-7251] (Oracle TopLink Essentials - 2.1 (Build 34 (05/21/2008))): oracle.toplink.essentials.exceptions.ValidationException Exception Description: The attribute [id] of class [com.dst.hps.web.config.Menu] is mapped to a primary key column in the database. Updates are not allowed. at oracle.toplink.essentials.exceptions.ValidationException.primaryKeyUpdateDisallo wed(ValidationException.java:2222) at oracle.toplink.essentials.mappings.foundation.AbstractDirectMapping.writeFromObj ectIntoRowWithChangeRecord(AbstractDirectMapping.java:750) </pre></p> #### Environment Operating System: All Platform: All #### Affected Versions [v2.1]
1.0
The attribute [...] of class [...] is mapped to a primary key column in the database. Updates are not allowed - <p>This is a reopen of issue 3868, as the defect (or a related defect) is still present as of Toplink Essentials 2.1 (Build 34 (05/21/2008)). I'd have posted this on that issue, but for doesn't appear I have rights to comment on issues I didn't create.</p> <p>Looking under the debugger, what I see is Toplink Essentials applying a changeSet entry, with identical values for the old and new primary key. The simple and effective fix would be to recognize when old and new values match, and not apply the update and not throw the ValidationException.</p> <p>In my scenerio, I've got a Menu object with: <pre> @Entity @Table(name = "MENU") public class Menu extends PersistableObject { @ManyToOne(fetch = FetchType.EAGER, optional = true) @JoinColumn(name = "parentId") private Menu parent = null; @OneToMany(cascade=CascadeType.ALL, fetch = FetchType.LAZY, mappedBy = "parent") private List<Menu> subMenus = new ArrayList<Menu>(); </pre></p> <p>The trace is: <pre> Exception [TOPLINK-7251] (Oracle TopLink Essentials - 2.1 (Build 34 (05/21/2008))): oracle.toplink.essentials.exceptions.ValidationException Exception Description: The attribute [id] of class [com.dst.hps.web.config.Menu] is mapped to a primary key column in the database. Updates are not allowed. at oracle.toplink.essentials.exceptions.ValidationException.primaryKeyUpdateDisallo wed(ValidationException.java:2222) at oracle.toplink.essentials.mappings.foundation.AbstractDirectMapping.writeFromObj ectIntoRowWithChangeRecord(AbstractDirectMapping.java:750) </pre></p> #### Environment Operating System: All Platform: All #### Affected Versions [v2.1]
priority
the attribute of class is mapped to a primary key column in the database updates are not allowed this is a reopen of issue as the defect or a related defect is still present as of toplink essentials build i d have posted this on that issue but for doesn t appear i have rights to comment on issues i didn t create looking under the debugger what i see is toplink essentials applying a changeset entry with identical values for the old and new primary key the simple and effective fix would be to recognize when old and new values match and not apply the update and not throw the validationexception in my scenerio i ve got a menu object with entity table name menu public class menu extends persistableobject manytoone fetch fetchtype eager optional true joincolumn name parentid private menu parent null onetomany cascade cascadetype all fetch fetchtype lazy mappedby parent private list submenus new arraylist the trace is exception oracle toplink essentials build oracle toplink essentials exceptions validationexception exception description the attribute of class is mapped to a primary key column in the database updates are not allowed at oracle toplink essentials exceptions validationexception primarykeyupdatedisallo wed validationexception java at oracle toplink essentials mappings foundation abstractdirectmapping writefromobj ectintorowwithchangerecord abstractdirectmapping java environment operating system all platform all affected versions
1
145,822
11,709,138,610
IssuesEvent
2020-03-08 17:09:41
jordan528/issues-demo
https://api.github.com/repos/jordan528/issues-demo
opened
Unsecured on GET:/example/v1/hotels/{id}
GitHub vul test
Title : Unsecured Vulnerability on GET:/example/v1/hotels/{id} Project : GitHub vul test Description : The unsecured exploit gives an attacker full access to the vulnerable endpoint without credentials. Risk : Unsecured Severity : Major API Endpoint : http://localhost:8080/example/v1/hotels/395617409 Environment : Master Playbook : ExampleV1HotelsIdGetAnonymousInvalid Researcher : [APISec Bot] Quick Tips : Suggestion : Make sure the endpoint is secured as part of the authentication framework. Effort Estimate : 2.0 Wire logs : 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : URL [http://localhost:8080/example/v1/hotels/395617409] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Method [GET] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Auth [] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Request [] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Request-Headers [{Content-Type=[application/json], Accept=[application/json]}] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Response [{ "timestamp" : "2020-03-08T17:07:38.384+0000", "status" : 401, "error" : "Unauthorized", "message" : "Unauthorized", "path" : "/example/v1/hotels/395617409" }] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Response-Headers [{WWW-Authenticate=[Basic realm="Realm"], X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Sun, 08 Mar 2020 17:07:37 GMT]}] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : StatusCode [401] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Time [200] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Size [143] 2020-03-08 05:07:38 ERROR [ExampleV1HotelsIdGetAnonymousInvalid] : Assertion [@StatusCode == 401 AND @StatusCode == 403] resolved-to [401 == 401 AND 401 == 403] result [Failed] Important Links : Vulnerability Details : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/recommendations/8a80cb8170babc450170bb1df7c900ef/details Project : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/jobs Environment : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/environments/8a80cb8170babc450170bad6a765000b/edit Scan Dashboard : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/jobs/8a80cb8170babc450170bad81e58004c/runs/8a80cb8170babc450170bb1dce6600eb Playbook : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/template/ExampleV1HotelsIdGetAnonymousInvalid Coverage : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/configuration Code Sample : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/recommendations/8a80cb8170babc450170bb1df7c900ef/codesamples PS: : Please contact admin@fxlabs.local For APISec access and login issues. --- APISec Bot ---
1.0
Unsecured on GET:/example/v1/hotels/{id} - Title : Unsecured Vulnerability on GET:/example/v1/hotels/{id} Project : GitHub vul test Description : The unsecured exploit gives an attacker full access to the vulnerable endpoint without credentials. Risk : Unsecured Severity : Major API Endpoint : http://localhost:8080/example/v1/hotels/395617409 Environment : Master Playbook : ExampleV1HotelsIdGetAnonymousInvalid Researcher : [APISec Bot] Quick Tips : Suggestion : Make sure the endpoint is secured as part of the authentication framework. Effort Estimate : 2.0 Wire logs : 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : URL [http://localhost:8080/example/v1/hotels/395617409] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Method [GET] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Auth [] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Request [] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Request-Headers [{Content-Type=[application/json], Accept=[application/json]}] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Response [{ "timestamp" : "2020-03-08T17:07:38.384+0000", "status" : 401, "error" : "Unauthorized", "message" : "Unauthorized", "path" : "/example/v1/hotels/395617409" }] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Response-Headers [{WWW-Authenticate=[Basic realm="Realm"], X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Sun, 08 Mar 2020 17:07:37 GMT]}] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : StatusCode [401] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Time [200] 2020-03-08 05:07:38 DEBUG [ExampleV1HotelsIdGetAnonymousInvalid] : Size [143] 2020-03-08 05:07:38 ERROR [ExampleV1HotelsIdGetAnonymousInvalid] : Assertion [@StatusCode == 401 AND @StatusCode == 403] resolved-to [401 == 401 AND 401 == 403] result [Failed] Important Links : Vulnerability Details : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/recommendations/8a80cb8170babc450170bb1df7c900ef/details Project : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/jobs Environment : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/environments/8a80cb8170babc450170bad6a765000b/edit Scan Dashboard : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/jobs/8a80cb8170babc450170bad81e58004c/runs/8a80cb8170babc450170bb1dce6600eb Playbook : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/template/ExampleV1HotelsIdGetAnonymousInvalid Coverage : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/configuration Code Sample : https://localhost:8080/#/app/projects/8a80cb8170babc450170bad6a6620009/recommendations/8a80cb8170babc450170bb1df7c900ef/codesamples PS: : Please contact admin@fxlabs.local For APISec access and login issues. --- APISec Bot ---
non_priority
unsecured on get example hotels id title unsecured vulnerability on get example hotels id project github vul test description the unsecured exploit gives an attacker full access to the vulnerable endpoint without credentials risk unsecured severity major api endpoint environment master playbook researcher quick tips suggestion make sure the endpoint is secured as part of the authentication framework effort estimate wire logs debug url debug method debug auth debug request debug request headers accept debug response timestamp status error unauthorized message unauthorized path example hotels debug response headers x content type options x xss protection cache control pragma expires x frame options content type transfer encoding date debug statuscode debug time debug size error assertion resolved to result important links vulnerability details project environment scan dashboard playbook coverage code sample ps please contact admin fxlabs local for apisec access and login issues apisec bot
0
19,852
10,546,043,622
IssuesEvent
2019-10-02 20:31:40
tensorflow/tensorflow
https://api.github.com/repos/tensorflow/tensorflow
closed
Prediction speed worse than community keras due to training related code running during inference
TF 1.14 TF 2.0.0-beta0 comp:keras stat:awaiting response type:performance
**System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): YES - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 18.04.2 (but we believe this to be irrelevant) - TensorFlow installed from (source or binary): pip (inside a docker image) - TensorFlow version (use command below): 1.14 (likely affects 1.12 - current 2.0 beta) - Python version: 3.6.8 (system python, we believe irrelevant) - CUDA/cuDNN version: 10.0.130 (but we believe irrelevant) - GPU model and memory: 1080 Ti **Describe the current behavior** Calling model.predict() includes calls to various non-prediction related things, in order of decreasing severity: * reset_metrics * get_progbar * standardize_user_data * validate_or_infer_batch_size In our project, this results in a 2X prediction speed regression: 35 ms per call for community keras 70 ms per call for tensorflow keras A Snakeviz flamegraph for prediction of our network in Tensorflow Keras, showing unnecessary overhead: ![Tensorflow Keras](https://user-images.githubusercontent.com/190617/63684638-ab884a80-c7fd-11e9-9dbe-5404eda2d70c.png) **Describe the expected behavior** Only do things necessary for prediction while doing prediction. A Snakeviz flamegraph for prediction of our network in Community Keras, showing no unnecessary overhead: ![Community Keras](https://user-images.githubusercontent.com/190617/63684637-ab884a80-c7fd-11e9-810d-6341eba99070.png) **Code to reproduce the issue** This code is active for all calls to predict(). This is for a commercial project we are trying to migrate from community keras to tensorflow keras. It is a speed critical real-time robotics application where the 30ms is enough to miss our hard deadlines. If predict() is intended for production inference, we feel this should be resolved in tensorflow keras. If it is only intended for use during training, this should be noted in the mainline documentation, along with a suggestion for what ~is intended for production inference. As a temporary workaround, we disabled some of these function calls in a local tensorflow fork, and the performance regression went away as expected. We took care to make sure this is not a profiling glitch. We excluded warmup time from the profile, and average over a number of samples.
True
Prediction speed worse than community keras due to training related code running during inference - **System information** - Have I written custom code (as opposed to using a stock example script provided in TensorFlow): YES - OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Linux Ubuntu 18.04.2 (but we believe this to be irrelevant) - TensorFlow installed from (source or binary): pip (inside a docker image) - TensorFlow version (use command below): 1.14 (likely affects 1.12 - current 2.0 beta) - Python version: 3.6.8 (system python, we believe irrelevant) - CUDA/cuDNN version: 10.0.130 (but we believe irrelevant) - GPU model and memory: 1080 Ti **Describe the current behavior** Calling model.predict() includes calls to various non-prediction related things, in order of decreasing severity: * reset_metrics * get_progbar * standardize_user_data * validate_or_infer_batch_size In our project, this results in a 2X prediction speed regression: 35 ms per call for community keras 70 ms per call for tensorflow keras A Snakeviz flamegraph for prediction of our network in Tensorflow Keras, showing unnecessary overhead: ![Tensorflow Keras](https://user-images.githubusercontent.com/190617/63684638-ab884a80-c7fd-11e9-9dbe-5404eda2d70c.png) **Describe the expected behavior** Only do things necessary for prediction while doing prediction. A Snakeviz flamegraph for prediction of our network in Community Keras, showing no unnecessary overhead: ![Community Keras](https://user-images.githubusercontent.com/190617/63684637-ab884a80-c7fd-11e9-810d-6341eba99070.png) **Code to reproduce the issue** This code is active for all calls to predict(). This is for a commercial project we are trying to migrate from community keras to tensorflow keras. It is a speed critical real-time robotics application where the 30ms is enough to miss our hard deadlines. If predict() is intended for production inference, we feel this should be resolved in tensorflow keras. If it is only intended for use during training, this should be noted in the mainline documentation, along with a suggestion for what ~is intended for production inference. As a temporary workaround, we disabled some of these function calls in a local tensorflow fork, and the performance regression went away as expected. We took care to make sure this is not a profiling glitch. We excluded warmup time from the profile, and average over a number of samples.
non_priority
prediction speed worse than community keras due to training related code running during inference system information have i written custom code as opposed to using a stock example script provided in tensorflow yes os platform and distribution e g linux ubuntu linux ubuntu but we believe this to be irrelevant tensorflow installed from source or binary pip inside a docker image tensorflow version use command below likely affects current beta python version system python we believe irrelevant cuda cudnn version but we believe irrelevant gpu model and memory ti describe the current behavior calling model predict includes calls to various non prediction related things in order of decreasing severity reset metrics get progbar standardize user data validate or infer batch size in our project this results in a prediction speed regression ms per call for community keras ms per call for tensorflow keras a snakeviz flamegraph for prediction of our network in tensorflow keras showing unnecessary overhead describe the expected behavior only do things necessary for prediction while doing prediction a snakeviz flamegraph for prediction of our network in community keras showing no unnecessary overhead code to reproduce the issue this code is active for all calls to predict this is for a commercial project we are trying to migrate from community keras to tensorflow keras it is a speed critical real time robotics application where the is enough to miss our hard deadlines if predict is intended for production inference we feel this should be resolved in tensorflow keras if it is only intended for use during training this should be noted in the mainline documentation along with a suggestion for what is intended for production inference as a temporary workaround we disabled some of these function calls in a local tensorflow fork and the performance regression went away as expected we took care to make sure this is not a profiling glitch we excluded warmup time from the profile and average over a number of samples
0
30,766
14,670,368,532
IssuesEvent
2020-12-30 04:36:13
sergiorribeiro/webmetry
https://api.github.com/repos/sergiorribeiro/webmetry
opened
[SidekiqJob] OtherTransaction/SidekiqJob/SyncAccountsJob/perform
needs squad prodops sidekiq-job transaction-performance
The transaction **`OtherTransaction/SidekiqJob/SyncAccountsJob/perform` (SidekiqJob)** violated a performance threshold. ## Violations: - [2020-12-30] Maximum execution duration during the current week was exceeded. Duration: **`2.9 hr`**. Limit: **`1.0 hr`**. <!-- /// --> ## Weekly transaction performance: ### Evolution graph (percentile 95): ``` [2020-12-23] ~ [2020-12-30] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️ 2.8 hr [2020-12-16] ~ [2020-12-23] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️ 3.1 hr [2020-12-09] ~ [2020-12-16] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦 3.7 hr [2020-12-02] ~ [2020-12-09] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️ 3.2 hr ``` ### Weekly indicators: | Indicator | Week -3 | Week -2 | Week -1 | Week 0 | |-|-|-|-|-| | Above average hits | 22 | 21 | 23 | 25 | | Max | 3.3 hr | 3.7 hr | 3.6 hr | 2.9 hr | | Average | 2.4 hr | 2.1 hr | 2.2 hr | 2.0 hr | | Percentile 95 | 3.2 hr | 3.7 hr | 3.1 hr | 2.8 hr | ### Month totals: | Indicator | Value | |-|-| | Max | 3.7 hr | | Average | 2.1 hr | | Percentile 95 | 3.2 hr | <!-- [EPID:4c8e6bbe3cc3ecaae12d95052e3a0a000f6ff836] -->
True
[SidekiqJob] OtherTransaction/SidekiqJob/SyncAccountsJob/perform - The transaction **`OtherTransaction/SidekiqJob/SyncAccountsJob/perform` (SidekiqJob)** violated a performance threshold. ## Violations: - [2020-12-30] Maximum execution duration during the current week was exceeded. Duration: **`2.9 hr`**. Limit: **`1.0 hr`**. <!-- /// --> ## Weekly transaction performance: ### Evolution graph (percentile 95): ``` [2020-12-23] ~ [2020-12-30] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️ 2.8 hr [2020-12-16] ~ [2020-12-23] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️ 3.1 hr [2020-12-09] ~ [2020-12-16] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦 3.7 hr [2020-12-02] ~ [2020-12-09] 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️ 3.2 hr ``` ### Weekly indicators: | Indicator | Week -3 | Week -2 | Week -1 | Week 0 | |-|-|-|-|-| | Above average hits | 22 | 21 | 23 | 25 | | Max | 3.3 hr | 3.7 hr | 3.6 hr | 2.9 hr | | Average | 2.4 hr | 2.1 hr | 2.2 hr | 2.0 hr | | Percentile 95 | 3.2 hr | 3.7 hr | 3.1 hr | 2.8 hr | ### Month totals: | Indicator | Value | |-|-| | Max | 3.7 hr | | Average | 2.1 hr | | Percentile 95 | 3.2 hr | <!-- [EPID:4c8e6bbe3cc3ecaae12d95052e3a0a000f6ff836] -->
non_priority
othertransaction sidekiqjob syncaccountsjob perform the transaction othertransaction sidekiqjob syncaccountsjob perform sidekiqjob violated a performance threshold violations maximum execution duration during the current week was exceeded duration hr limit hr weekly transaction performance evolution graph percentile 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️⬜️ hr 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️⬜️ hr 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦 hr 🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦🟦⬜️⬜️⬜️ hr weekly indicators indicator week week week week above average hits max hr hr hr hr average hr hr hr hr percentile hr hr hr hr month totals indicator value max hr average hr percentile hr
0
192,037
6,846,068,646
IssuesEvent
2017-11-13 10:38:09
status-im/status-react
https://api.github.com/repos/status-im/status-react
closed
App crashes if continue to Reddit app while browsing reddit.com
android bounty bug medium-priority
### Description [comment]: # (Feature or Bug? i.e Type: Bug) *Type*: Bug [comment]: # (Describe the feature you would like, or briefly summarise the bug and what you did, what you expected to happen, and what actually happens. Sections below) *Summary*: App crashes if browse reddit.com and Continue to download reddit app ![reddit](https://cloud.githubusercontent.com/assets/6584810/26007314/414d4ad0-3740-11e7-86cd-ae891d05a4a4.png) #### Expected behavior [comment]: # (Describe what you expected to happen.) Play Store is opened with reddit app #### Actual behavior [comment]: # (Describe what actually happened.) App crashes ### Reproduction [comment]: # (Describe how we can replicate the bug step by step.) Video: https://drive.google.com/open?id=0Bz3t9zSg1wb7bmRWR2ZYMFJBTFE - Open Status - Open console - @browse reddit.com - after page is loaded popup is shown "View Reddit in the app..." - tap on button Continue on the popup ### Additional Information [comment]: # (Please do your best to fill this out.) * Status version: 0.9.7 * Operating System: Android only #### Logs [20171205181427-logcat.txt](https://github.com/status-im/status-react/files/997357/20171205181427-logcat.txt)
1.0
App crashes if continue to Reddit app while browsing reddit.com - ### Description [comment]: # (Feature or Bug? i.e Type: Bug) *Type*: Bug [comment]: # (Describe the feature you would like, or briefly summarise the bug and what you did, what you expected to happen, and what actually happens. Sections below) *Summary*: App crashes if browse reddit.com and Continue to download reddit app ![reddit](https://cloud.githubusercontent.com/assets/6584810/26007314/414d4ad0-3740-11e7-86cd-ae891d05a4a4.png) #### Expected behavior [comment]: # (Describe what you expected to happen.) Play Store is opened with reddit app #### Actual behavior [comment]: # (Describe what actually happened.) App crashes ### Reproduction [comment]: # (Describe how we can replicate the bug step by step.) Video: https://drive.google.com/open?id=0Bz3t9zSg1wb7bmRWR2ZYMFJBTFE - Open Status - Open console - @browse reddit.com - after page is loaded popup is shown "View Reddit in the app..." - tap on button Continue on the popup ### Additional Information [comment]: # (Please do your best to fill this out.) * Status version: 0.9.7 * Operating System: Android only #### Logs [20171205181427-logcat.txt](https://github.com/status-im/status-react/files/997357/20171205181427-logcat.txt)
priority
app crashes if continue to reddit app while browsing reddit com description feature or bug i e type bug type bug describe the feature you would like or briefly summarise the bug and what you did what you expected to happen and what actually happens sections below summary app crashes if browse reddit com and continue to download reddit app expected behavior describe what you expected to happen play store is opened with reddit app actual behavior describe what actually happened app crashes reproduction describe how we can replicate the bug step by step video open status open console browse reddit com after page is loaded popup is shown view reddit in the app tap on button continue on the popup additional information please do your best to fill this out status version operating system android only logs
1
303,985
9,312,805,576
IssuesEvent
2019-03-26 02:54:17
nhnent/tui.chart
https://api.github.com/repos/nhnent/tui.chart
closed
safari browser date parsing issue
Category: ETC NHN Priority: High Type: Bug
<!-- Thank you for your contribution. When it comes to write an issue, please, use the template below. To use the template is mandatory for submit new issue and we won't reply the issue that without the template. --> <!-- TEMPLATE --> ## Version <!-- Write the version of the grid you are currently using. --> v3.5.1 ## Development Environment <!-- Write the browser type, OS and so on --> OSX, safari browser ## Current Behavior <!-- Write a description of the current operation. You can add sample code, 'CodePen' or 'jsfiddle' links. --> ```js var container = document.getElementById('chart-area'); var data = { series: [ { name: 'SiteA', data: [ ['2016-08-22T10:00:00', 202], ['2016-08-22T10:05:00', 212], ['2016-08-22T10:10:00', 222], ['2016-08-22T10:15:00', 351], ['2016-08-22T10:20:00', 412], ['2016-08-22T10:25:00', 420], ['2016-08-22T10:30:00', 300], ['2016-08-22T10:35:00', 213], ['2016-08-22T10:40:00', 230], ['2016-08-22T10:45:00', 220], ['2016-08-22T10:50:00', 234], ['2016-08-22T10:55:00', 210], ['2016-08-22T11:00:00', 220] ] }, { name: 'SiteB', data: [ ['2016-08-22T10:00:00', 312], ['2016-08-22T10:10:00', 320], ['2016-08-22T10:20:00', 295], ['2016-08-22T10:30:00', 300], ['2016-08-22T10:40:00', 320], ['2016-08-22T10:50:00', 30], ['2016-08-22T11:00:00', 20] ] } ] }; var options = { chart: { width: 1160, height: 540, title: 'Concurrent users' }, xAxis: { title: 'minutes', type: 'datetime', dateFormat: 'hh:mm' }, yAxis: { title: 'users', pointOnColumn: true }, series: { showDot: false }, plot: { bands: [ { range: ['08/22/2016 10:40:00', '08/22/2016 11:00:00'], color: 'gray', opacity: 0.2 } ], lines: [ { value: '08/22/2016 10:10:00', color: '#fa2828' } ] } }; var theme = { series: { colors: [ '#83b14e', '#458a3f', '#295ba0', '#2a4175', '#289399', '#289399', '#617178', '#8a9a9a', '#516f7d', '#dddddd' ] } }; // For apply theme // tui.chart.registerTheme('myTheme', theme); // options.theme = 'myTheme'; var chart = tui.chart.lineChart(container, data, options); ``` ![image](https://user-images.githubusercontent.com/35371660/53936748-48d7b100-40ee-11e9-9525-848b7d77c071.png) ## Expected Behavior ![image](https://user-images.githubusercontent.com/35371660/53936745-45442a00-40ee-11e9-9273-c5c573332429.png)
1.0
safari browser date parsing issue - <!-- Thank you for your contribution. When it comes to write an issue, please, use the template below. To use the template is mandatory for submit new issue and we won't reply the issue that without the template. --> <!-- TEMPLATE --> ## Version <!-- Write the version of the grid you are currently using. --> v3.5.1 ## Development Environment <!-- Write the browser type, OS and so on --> OSX, safari browser ## Current Behavior <!-- Write a description of the current operation. You can add sample code, 'CodePen' or 'jsfiddle' links. --> ```js var container = document.getElementById('chart-area'); var data = { series: [ { name: 'SiteA', data: [ ['2016-08-22T10:00:00', 202], ['2016-08-22T10:05:00', 212], ['2016-08-22T10:10:00', 222], ['2016-08-22T10:15:00', 351], ['2016-08-22T10:20:00', 412], ['2016-08-22T10:25:00', 420], ['2016-08-22T10:30:00', 300], ['2016-08-22T10:35:00', 213], ['2016-08-22T10:40:00', 230], ['2016-08-22T10:45:00', 220], ['2016-08-22T10:50:00', 234], ['2016-08-22T10:55:00', 210], ['2016-08-22T11:00:00', 220] ] }, { name: 'SiteB', data: [ ['2016-08-22T10:00:00', 312], ['2016-08-22T10:10:00', 320], ['2016-08-22T10:20:00', 295], ['2016-08-22T10:30:00', 300], ['2016-08-22T10:40:00', 320], ['2016-08-22T10:50:00', 30], ['2016-08-22T11:00:00', 20] ] } ] }; var options = { chart: { width: 1160, height: 540, title: 'Concurrent users' }, xAxis: { title: 'minutes', type: 'datetime', dateFormat: 'hh:mm' }, yAxis: { title: 'users', pointOnColumn: true }, series: { showDot: false }, plot: { bands: [ { range: ['08/22/2016 10:40:00', '08/22/2016 11:00:00'], color: 'gray', opacity: 0.2 } ], lines: [ { value: '08/22/2016 10:10:00', color: '#fa2828' } ] } }; var theme = { series: { colors: [ '#83b14e', '#458a3f', '#295ba0', '#2a4175', '#289399', '#289399', '#617178', '#8a9a9a', '#516f7d', '#dddddd' ] } }; // For apply theme // tui.chart.registerTheme('myTheme', theme); // options.theme = 'myTheme'; var chart = tui.chart.lineChart(container, data, options); ``` ![image](https://user-images.githubusercontent.com/35371660/53936748-48d7b100-40ee-11e9-9525-848b7d77c071.png) ## Expected Behavior ![image](https://user-images.githubusercontent.com/35371660/53936745-45442a00-40ee-11e9-9273-c5c573332429.png)
priority
safari browser date parsing issue thank you for your contribution when it comes to write an issue please use the template below to use the template is mandatory for submit new issue and we won t reply the issue that without the template version development environment osx safari browser current behavior js var container document getelementbyid chart area var data series name sitea data name siteb data var options chart width height title concurrent users xaxis title minutes type datetime dateformat hh mm yaxis title users pointoncolumn true series showdot false plot bands range color gray opacity lines value color var theme series colors dddddd for apply theme tui chart registertheme mytheme theme options theme mytheme var chart tui chart linechart container data options expected behavior
1
204,415
7,087,512,030
IssuesEvent
2018-01-11 18:04:45
DjangoChained/Boogle
https://api.github.com/repos/DjangoChained/Boogle
closed
Modifier le message affiché quand un record est battu
bug low priority
Quand un record est battu, le joueur étant indiqué comme battu est celui avec le score le plus bas même si le nouveau record est le plus élevé, à revoir.
1.0
Modifier le message affiché quand un record est battu - Quand un record est battu, le joueur étant indiqué comme battu est celui avec le score le plus bas même si le nouveau record est le plus élevé, à revoir.
priority
modifier le message affiché quand un record est battu quand un record est battu le joueur étant indiqué comme battu est celui avec le score le plus bas même si le nouveau record est le plus élevé à revoir
1
406,210
11,888,652,577
IssuesEvent
2020-03-28 09:44:32
AY1920S2-CS2103T-W17-2/main
https://api.github.com/repos/AY1920S2-CS2103T-W17-2/main
opened
Implement Parser (Markdown to HTML compiler)
priority.High status.Ongoing type.Enhancement
Part of #162 Proposed Markdown to HTML compiler architecture: String -> Tokens -> Abstract Syntax Tree (AST) -> HTML Let's now implement the Tokens -> Abstract Syntax Tree (AST) portion. We'll call this a `Parser`
1.0
Implement Parser (Markdown to HTML compiler) - Part of #162 Proposed Markdown to HTML compiler architecture: String -> Tokens -> Abstract Syntax Tree (AST) -> HTML Let's now implement the Tokens -> Abstract Syntax Tree (AST) portion. We'll call this a `Parser`
priority
implement parser markdown to html compiler part of proposed markdown to html compiler architecture string tokens abstract syntax tree ast html let s now implement the tokens abstract syntax tree ast portion we ll call this a parser
1
81,871
3,599,768,205
IssuesEvent
2016-02-03 00:53:39
HeavensGate/Eternal
https://api.github.com/repos/HeavensGate/Eternal
closed
Passive Mapping
low priority mapping simple/easy
Add neat, harmless details that flavour the station, so long as they wouldn't conspicuously not belong for months at a time.
1.0
Passive Mapping - Add neat, harmless details that flavour the station, so long as they wouldn't conspicuously not belong for months at a time.
priority
passive mapping add neat harmless details that flavour the station so long as they wouldn t conspicuously not belong for months at a time
1
55,320
13,610,412,049
IssuesEvent
2020-09-23 07:19:35
Iridescent-CM/technovation-app
https://api.github.com/repos/Iridescent-CM/technovation-app
closed
Add option to copy/paste a custom invite link to supplement student email invitations
0 - Icebox Discuss [theme] team building it-32 it-33
Is it a one-time use code or can it be used multiple times? (Presumably once, like the links in the autogenerated emails) <!--- @huboard:{"order":4.776000305994726e-31,"milestone_order":7.5735656109917e-45} -->
1.0
Add option to copy/paste a custom invite link to supplement student email invitations - Is it a one-time use code or can it be used multiple times? (Presumably once, like the links in the autogenerated emails) <!--- @huboard:{"order":4.776000305994726e-31,"milestone_order":7.5735656109917e-45} -->
non_priority
add option to copy paste a custom invite link to supplement student email invitations is it a one time use code or can it be used multiple times presumably once like the links in the autogenerated emails huboard order milestone order
0
495,099
14,272,116,768
IssuesEvent
2020-11-21 15:37:45
UC-Davis-molecular-computing/scadnano
https://api.github.com/repos/UC-Davis-molecular-computing/scadnano
closed
add "rope select" mode to select items by drawing an arbitrary polygon
closed in dev enhancement high priority
Currently, the selection box is parallel to the x and y axes. It would be easier to select some strands (e.g., in single-stranded tile designs) if one could draw an arbitrary polygon, by clicking three or more points, then clicking near the start point. The selection box is drawn by pressing Ctrl or Shift, and then clicking and dragging. One possible way to implement this is as a separate Edit Mode called "rope select". It acts like select mode, but when pressing Shift or Click, there's no "drag" response. Instead, each click after the first draws a new edge of the polygon, and there is always an edge from the last point clicked to the first point as well. Edges that cross existing edges are disallowed (i.e, the click is ignored). Lifting up on the Shift/Ctrl key closes the polygon. Then all objects whose bounding rectangles lie inside the polygon are selected. Note that this will cause some problems if something is a diagonal line, since its bounding rectangle is much larger; currently I don't know how to get a smallest convex polygon, or a smallest rectangle, surrounding an arbitrary SVG object.
1.0
add "rope select" mode to select items by drawing an arbitrary polygon - Currently, the selection box is parallel to the x and y axes. It would be easier to select some strands (e.g., in single-stranded tile designs) if one could draw an arbitrary polygon, by clicking three or more points, then clicking near the start point. The selection box is drawn by pressing Ctrl or Shift, and then clicking and dragging. One possible way to implement this is as a separate Edit Mode called "rope select". It acts like select mode, but when pressing Shift or Click, there's no "drag" response. Instead, each click after the first draws a new edge of the polygon, and there is always an edge from the last point clicked to the first point as well. Edges that cross existing edges are disallowed (i.e, the click is ignored). Lifting up on the Shift/Ctrl key closes the polygon. Then all objects whose bounding rectangles lie inside the polygon are selected. Note that this will cause some problems if something is a diagonal line, since its bounding rectangle is much larger; currently I don't know how to get a smallest convex polygon, or a smallest rectangle, surrounding an arbitrary SVG object.
priority
add rope select mode to select items by drawing an arbitrary polygon currently the selection box is parallel to the x and y axes it would be easier to select some strands e g in single stranded tile designs if one could draw an arbitrary polygon by clicking three or more points then clicking near the start point the selection box is drawn by pressing ctrl or shift and then clicking and dragging one possible way to implement this is as a separate edit mode called rope select it acts like select mode but when pressing shift or click there s no drag response instead each click after the first draws a new edge of the polygon and there is always an edge from the last point clicked to the first point as well edges that cross existing edges are disallowed i e the click is ignored lifting up on the shift ctrl key closes the polygon then all objects whose bounding rectangles lie inside the polygon are selected note that this will cause some problems if something is a diagonal line since its bounding rectangle is much larger currently i don t know how to get a smallest convex polygon or a smallest rectangle surrounding an arbitrary svg object
1
818,009
30,667,147,961
IssuesEvent
2023-07-25 19:12:09
paleobot/pbot-dev
https://api.github.com/repos/paleobot/pbot-dev
closed
Reference order, prepopulate first entry to 1
enhancement low priority Hack-A-Thon
Can the reference order be pre-filled to "1" for the first entry? Like with the states.
1.0
Reference order, prepopulate first entry to 1 - Can the reference order be pre-filled to "1" for the first entry? Like with the states.
priority
reference order prepopulate first entry to can the reference order be pre filled to for the first entry like with the states
1
788,969
27,774,999,062
IssuesEvent
2023-03-16 16:38:41
DataScienceScotland/intro_to_r
https://api.github.com/repos/DataScienceScotland/intro_to_r
opened
Stick to ~80 characters per line
low priority
To help ensure that everything is visible when we zoom in. Quite a bit of work so low priority
1.0
Stick to ~80 characters per line - To help ensure that everything is visible when we zoom in. Quite a bit of work so low priority
priority
stick to characters per line to help ensure that everything is visible when we zoom in quite a bit of work so low priority
1
752,697
26,295,824,682
IssuesEvent
2023-01-08 23:49:57
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth-2
opened
Adapt MaA Triggers in 00_cultural_triggers.txt
priority high :exclamation: balance :balance_scale: cultural :mortar_board:
`# Cultural MAA triggers - Used in various places for triggers, AI weights, etc.` Relevant triggers are described from line 201 to 302
1.0
Adapt MaA Triggers in 00_cultural_triggers.txt - `# Cultural MAA triggers - Used in various places for triggers, AI weights, etc.` Relevant triggers are described from line 201 to 302
priority
adapt maa triggers in cultural triggers txt cultural maa triggers used in various places for triggers ai weights etc relevant triggers are described from line to
1
7,608
7,016,639,135
IssuesEvent
2017-12-21 05:18:12
srinikoganti/sams
https://api.github.com/repos/srinikoganti/sams
closed
College Infrastructure: Server breakdown when click on view button(P1)
Infrastructure P1
Error: Getting P1server side exception when we click on the View button of any college in the Infrastructure grid. Please refer the below screen shots. ![infra1](https://user-images.githubusercontent.com/33898419/33661337-834db508-daad-11e7-82da-843026baae04.png) ![infra2](https://user-images.githubusercontent.com/33898419/33661344-8878a5c4-daad-11e7-9f35-4084aac99d32.png)
1.0
College Infrastructure: Server breakdown when click on view button(P1) - Error: Getting P1server side exception when we click on the View button of any college in the Infrastructure grid. Please refer the below screen shots. ![infra1](https://user-images.githubusercontent.com/33898419/33661337-834db508-daad-11e7-82da-843026baae04.png) ![infra2](https://user-images.githubusercontent.com/33898419/33661344-8878a5c4-daad-11e7-9f35-4084aac99d32.png)
non_priority
college infrastructure server breakdown when click on view button error getting side exception when we click on the view button of any college in the infrastructure grid please refer the below screen shots
0
277,504
24,077,557,737
IssuesEvent
2022-09-19 00:42:45
Princeton-CDH/geniza
https://api.github.com/repos/Princeton-CDH/geniza
closed
As a content editor, I want to add and edit transcriptions for records without all IIIF images available so that transcriptions aren't limited to records with all images.
🗜️ awaiting testing
## testing notes (QA) (these notes are similar to #1039, but refer to the transcription editor and not the public viewer) Navigate to the QA public site document details page for one or more documents with: - no images, but transcription content (e.g. PGPID 6176) - no images and no transcription content (e.g. PGPID 33472) - both, with some but not all images (i.e. some transcription content extends past the images). (e.g. PGPID 1390) In the transcription editor ("edit X's transcription"/"add a new transcription", visible when logged into admin): - For a document with no images and no transcription: - [x] confirm that you see two gray blank rectangles as placeholder images where images would go - [x] shift+click on them to create new annotations on the image, should result in new transcription blocks in the right column - For a document with transcription content but no images: - [x] Confirm you see placeholder images for each block of transcription - [x] Edit and save some transcription text - For a document with transcription content and some but not all images: - [x] Confirm you see placeholder images for each block of transcription that extends past the number of images present ## dev notes - [x] load transcription editor even if document has no images (ex. PGPID 6176) - [x] display placeholder canvases in editor of document has no images (maybe recto/verso, canvas 1 and 2 ?) - [x] use placeholder canvases for documents with some but not all images, ex. PGPID 1390 may have some overlap with #1039
1.0
As a content editor, I want to add and edit transcriptions for records without all IIIF images available so that transcriptions aren't limited to records with all images. - ## testing notes (QA) (these notes are similar to #1039, but refer to the transcription editor and not the public viewer) Navigate to the QA public site document details page for one or more documents with: - no images, but transcription content (e.g. PGPID 6176) - no images and no transcription content (e.g. PGPID 33472) - both, with some but not all images (i.e. some transcription content extends past the images). (e.g. PGPID 1390) In the transcription editor ("edit X's transcription"/"add a new transcription", visible when logged into admin): - For a document with no images and no transcription: - [x] confirm that you see two gray blank rectangles as placeholder images where images would go - [x] shift+click on them to create new annotations on the image, should result in new transcription blocks in the right column - For a document with transcription content but no images: - [x] Confirm you see placeholder images for each block of transcription - [x] Edit and save some transcription text - For a document with transcription content and some but not all images: - [x] Confirm you see placeholder images for each block of transcription that extends past the number of images present ## dev notes - [x] load transcription editor even if document has no images (ex. PGPID 6176) - [x] display placeholder canvases in editor of document has no images (maybe recto/verso, canvas 1 and 2 ?) - [x] use placeholder canvases for documents with some but not all images, ex. PGPID 1390 may have some overlap with #1039
non_priority
as a content editor i want to add and edit transcriptions for records without all iiif images available so that transcriptions aren t limited to records with all images testing notes qa these notes are similar to but refer to the transcription editor and not the public viewer navigate to the qa public site document details page for one or more documents with no images but transcription content e g pgpid no images and no transcription content e g pgpid both with some but not all images i e some transcription content extends past the images e g pgpid in the transcription editor edit x s transcription add a new transcription visible when logged into admin for a document with no images and no transcription confirm that you see two gray blank rectangles as placeholder images where images would go shift click on them to create new annotations on the image should result in new transcription blocks in the right column for a document with transcription content but no images confirm you see placeholder images for each block of transcription edit and save some transcription text for a document with transcription content and some but not all images confirm you see placeholder images for each block of transcription that extends past the number of images present dev notes load transcription editor even if document has no images ex pgpid display placeholder canvases in editor of document has no images maybe recto verso canvas and use placeholder canvases for documents with some but not all images ex pgpid may have some overlap with
0
217,847
24,351,630,898
IssuesEvent
2022-10-03 01:03:48
saurockSaurav/weather-information-api
https://api.github.com/repos/saurockSaurav/weather-information-api
opened
CVE-2022-42003 (Medium) detected in jackson-databind-2.8.11.3.jar
security vulnerability
## CVE-2022-42003 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /weather-information-api/weather-rest-api-service/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.20.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In FasterXML jackson-databind before 2.14.0-rc1, resource exhaustion can occur because of a lack of a check in primitive value deserializers to avoid deep wrapper array nesting, when the UNWRAP_SINGLE_VALUE_ARRAYS feature is enabled. <p>Publish Date: 2022-10-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42003>CVE-2022-42003</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-42003 (Medium) detected in jackson-databind-2.8.11.3.jar - ## CVE-2022-42003 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.11.3.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /weather-information-api/weather-rest-api-service/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.11.3/jackson-databind-2.8.11.3.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-1.5.20.RELEASE.jar (Root Library) - :x: **jackson-databind-2.8.11.3.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In FasterXML jackson-databind before 2.14.0-rc1, resource exhaustion can occur because of a lack of a check in primitive value deserializers to avoid deep wrapper array nesting, when the UNWRAP_SINGLE_VALUE_ARRAYS feature is enabled. <p>Publish Date: 2022-10-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42003>CVE-2022-42003</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file weather information api weather rest api service pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library vulnerability details in fasterxml jackson databind before resource exhaustion can occur because of a lack of a check in primitive value deserializers to avoid deep wrapper array nesting when the unwrap single value arrays feature is enabled publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
0