Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
1k
labels
stringlengths
4
1.38k
body
stringlengths
1
262k
index
stringclasses
16 values
text_combine
stringlengths
96
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
52,812
13,771,902,152
IssuesEvent
2020-10-07 23:04:32
jgeraigery/please-del-this-repo
https://api.github.com/repos/jgeraigery/please-del-this-repo
opened
CVE-2017-16138 (High) detected in mime-1.3.4.tgz
security vulnerability
## CVE-2017-16138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.3.4.tgz</b></p></summary> <p>A comprehensive library for mime-type mapping</p> <p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.3.4.tgz">https://registry.npmjs.org/mime/-/mime-1.3.4.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/mime/package.json</p> <p> Dependency Hierarchy: - mailparser-0.6.2.tgz (Root Library) - :x: **mime-1.3.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/please-del-this-repo/commit/e701d80cb459747be96e9c0e79c8640eca26c232">e701d80cb459747be96e9c0e79c8640eca26c232</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 1.4.1,2.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.3.4","isTransitiveDependency":true,"dependencyTree":"mailparser:0.6.2;mime:1.3.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-16138 (High) detected in mime-1.3.4.tgz - ## CVE-2017-16138 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.3.4.tgz</b></p></summary> <p>A comprehensive library for mime-type mapping</p> <p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.3.4.tgz">https://registry.npmjs.org/mime/-/mime-1.3.4.tgz</a></p> <p>Path to dependency file: please-del-this-repo/package.json</p> <p>Path to vulnerable library: please-del-this-repo/node_modules/mime/package.json</p> <p> Dependency Hierarchy: - mailparser-0.6.2.tgz (Root Library) - :x: **mime-1.3.4.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/please-del-this-repo/commit/e701d80cb459747be96e9c0e79c8640eca26c232">e701d80cb459747be96e9c0e79c8640eca26c232</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input. <p>Publish Date: 2018-06-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p> <p>Release Date: 2018-06-07</p> <p>Fix Resolution: 1.4.1,2.0.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.3.4","isTransitiveDependency":true,"dependencyTree":"mailparser:0.6.2;mime:1.3.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in mime tgz cve high severity vulnerability vulnerable library mime tgz a comprehensive library for mime type mapping library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules mime package json dependency hierarchy mailparser tgz root library x mime tgz vulnerable library found in head commit a href found in base branch master vulnerability details the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input vulnerabilityurl
0
71,198
18,518,853,908
IssuesEvent
2021-10-20 13:11:50
astropy/astropy
https://api.github.com/repos/astropy/astropy
reopened
Add support to release aarch64 wheels
Feature Request Release build
Problem --------- On aarch64, pip install astropy builds the wheels from source code and then install it. It requires user to have development environment installed on his system. also, it take some time to build the wheels than downloading and extracting the wheels from pypi. Resolution ----------- On aarch64, pip install astropy should download the wheels from pypi @saimn, Please let me know your interest on releasing aarch64 wheels. I can help in this.
1.0
Add support to release aarch64 wheels - Problem --------- On aarch64, pip install astropy builds the wheels from source code and then install it. It requires user to have development environment installed on his system. also, it take some time to build the wheels than downloading and extracting the wheels from pypi. Resolution ----------- On aarch64, pip install astropy should download the wheels from pypi @saimn, Please let me know your interest on releasing aarch64 wheels. I can help in this.
non_priority
add support to release wheels problem on pip install astropy builds the wheels from source code and then install it it requires user to have development environment installed on his system also it take some time to build the wheels than downloading and extracting the wheels from pypi resolution on pip install astropy should download the wheels from pypi saimn please let me know your interest on releasing wheels i can help in this
0
386,062
11,431,152,318
IssuesEvent
2020-02-04 11:31:31
w3c/epubcheck
https://api.github.com/repos/w3c/epubcheck
closed
Testing Tools for EPUBCheck
priority: low status: in discussion type: feature
I have created a simple script for running a folder of files through EPUBCheck and outputting results in a single text file. https://github.com/wareid/EPUB3-tests/tree/master/EPUBCheck%20Testing%20Tools Would this be worthwhile including in the EPUBCheck repo for potential testers? Also, please let me know if there's other tools or changes I can make to improve this :). My personal version runs the same files against 2 versions of EPUBCheck for comparison.
1.0
Testing Tools for EPUBCheck - I have created a simple script for running a folder of files through EPUBCheck and outputting results in a single text file. https://github.com/wareid/EPUB3-tests/tree/master/EPUBCheck%20Testing%20Tools Would this be worthwhile including in the EPUBCheck repo for potential testers? Also, please let me know if there's other tools or changes I can make to improve this :). My personal version runs the same files against 2 versions of EPUBCheck for comparison.
priority
testing tools for epubcheck i have created a simple script for running a folder of files through epubcheck and outputting results in a single text file would this be worthwhile including in the epubcheck repo for potential testers also please let me know if there s other tools or changes i can make to improve this my personal version runs the same files against versions of epubcheck for comparison
1
143,520
22,061,036,591
IssuesEvent
2022-05-30 17:52:44
DXgovernance/dxvote
https://api.github.com/repos/DXgovernance/dxvote
closed
Add tenderly verification modal to add action flow
Enhancement Design
Guilds are able to do on-chain function calls which we call actions. We are going to use tenderly to simulate the defined actions to check if they are actually executable once a proposal has passed. Please go through the playground before you start working on this issue [HERE](https://www.figma.com/proto/cjLng0AbOKuDj9r5oZVDmR/Wireframe?page-id=4442%3A17030&node-id=4442%3A17031&viewport=325%2C48%2C0.1&scaling=min-zoom&starting-point-node-id=4442%3A17031). To make this step transparent we are gonna add a `Verify Action - Modal` to the `Add Action - Flow`: ### Verify Action ![Verify](https://user-images.githubusercontent.com/5337809/159005438-3d4d4f2a-5aaa-44a4-b950-c81b535e6e15.png) - Always appears once the user clicks on `Add Action - Button` in the add action process. - Anytime a user edits, adds, removes an action we need to verify the whole package of actions via tenderly. - The `Go to Tenderly - Button` is opening up a new tab with the actual tenderly simulation. - [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5281%3A17505) ### Verified ![Verified](https://user-images.githubusercontent.com/5337809/159005645-de068b61-d9af-4cc7-898c-1ead24bb7728.png) - If tenderly's simulation passed successfully, the user can dismiss or close the modal with the close icon. - User can still go to tenderly with the `Go to tenderly - Button`. - After clicking dismiss user should see the defined action in the `Actions - Widget` - [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5281%3A17836) ### Verification Failed ![VerificationFailed](https://user-images.githubusercontent.com/5337809/159007663-98b3abbe-de40-4882-991d-33e377928975.png) - If tenderly's simulation fails, the user can go back to the action modal, go to tenderly via the `Go to tenderly - Button` or close the modal with the close icon. - We will not allow the user add an action which has not passed the simulation of tenderly. - [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5353%3A17674)
1.0
Add tenderly verification modal to add action flow - Guilds are able to do on-chain function calls which we call actions. We are going to use tenderly to simulate the defined actions to check if they are actually executable once a proposal has passed. Please go through the playground before you start working on this issue [HERE](https://www.figma.com/proto/cjLng0AbOKuDj9r5oZVDmR/Wireframe?page-id=4442%3A17030&node-id=4442%3A17031&viewport=325%2C48%2C0.1&scaling=min-zoom&starting-point-node-id=4442%3A17031). To make this step transparent we are gonna add a `Verify Action - Modal` to the `Add Action - Flow`: ### Verify Action ![Verify](https://user-images.githubusercontent.com/5337809/159005438-3d4d4f2a-5aaa-44a4-b950-c81b535e6e15.png) - Always appears once the user clicks on `Add Action - Button` in the add action process. - Anytime a user edits, adds, removes an action we need to verify the whole package of actions via tenderly. - The `Go to Tenderly - Button` is opening up a new tab with the actual tenderly simulation. - [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5281%3A17505) ### Verified ![Verified](https://user-images.githubusercontent.com/5337809/159005645-de068b61-d9af-4cc7-898c-1ead24bb7728.png) - If tenderly's simulation passed successfully, the user can dismiss or close the modal with the close icon. - User can still go to tenderly with the `Go to tenderly - Button`. - After clicking dismiss user should see the defined action in the `Actions - Widget` - [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5281%3A17836) ### Verification Failed ![VerificationFailed](https://user-images.githubusercontent.com/5337809/159007663-98b3abbe-de40-4882-991d-33e377928975.png) - If tenderly's simulation fails, the user can go back to the action modal, go to tenderly via the `Go to tenderly - Button` or close the modal with the close icon. - We will not allow the user add an action which has not passed the simulation of tenderly. - [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5353%3A17674)
non_priority
add tenderly verification modal to add action flow guilds are able to do on chain function calls which we call actions we are going to use tenderly to simulate the defined actions to check if they are actually executable once a proposal has passed please go through the playground before you start working on this issue to make this step transparent we are gonna add a verify action modal to the add action flow verify action always appears once the user clicks on add action button in the add action process anytime a user edits adds removes an action we need to verify the whole package of actions via tenderly the go to tenderly button is opening up a new tab with the actual tenderly simulation verified if tenderly s simulation passed successfully the user can dismiss or close the modal with the close icon user can still go to tenderly with the go to tenderly button after clicking dismiss user should see the defined action in the actions widget verification failed if tenderly s simulation fails the user can go back to the action modal go to tenderly via the go to tenderly button or close the modal with the close icon we will not allow the user add an action which has not passed the simulation of tenderly
0
106,827
9,188,302,881
IssuesEvent
2019-03-06 06:55:06
CARTAvis/carta-backend
https://api.github.com/repos/CARTAvis/carta-backend
closed
loading a file without specifying an HDU fails
bug ready to test
The frontend has a secondary method of loading files, via URL parameters. In this case, the HDU field will be left blank. However, this results in the file not loading, because the `casacore::string::fromString` call has the `chk` flag set to `true`, resulting in an exception being thrown. The other instance of `fromString` being used (in `getFITShdu`) explicitly guards against empty strings. This should be done in `openFile` as well.
1.0
loading a file without specifying an HDU fails - The frontend has a secondary method of loading files, via URL parameters. In this case, the HDU field will be left blank. However, this results in the file not loading, because the `casacore::string::fromString` call has the `chk` flag set to `true`, resulting in an exception being thrown. The other instance of `fromString` being used (in `getFITShdu`) explicitly guards against empty strings. This should be done in `openFile` as well.
non_priority
loading a file without specifying an hdu fails the frontend has a secondary method of loading files via url parameters in this case the hdu field will be left blank however this results in the file not loading because the casacore string fromstring call has the chk flag set to true resulting in an exception being thrown the other instance of fromstring being used in getfitshdu explicitly guards against empty strings this should be done in openfile as well
0
485,837
14,000,209,192
IssuesEvent
2020-10-28 11:59:07
ihhub/fheroes2
https://api.github.com/repos/ihhub/fheroes2
closed
Game screens, all "EVIL" windows are not centered vertically in the screen
Hacktoberfest good first issue low priority ui
In fheroes2, with the "EVIL" interface, all windows of this type are not centered vertically in the screen. In fheroes2: ![Quit confirmation window evil fh2 01](https://user-images.githubusercontent.com/43583994/94999522-323f5780-05ba-11eb-90f8-c5c69b765b71.png) ![Adventure Evil fh2 01](https://user-images.githubusercontent.com/43583994/94999527-39666580-05ba-11eb-8322-8e2ffd04a35f.png) ![Hero Evil fh2 03](https://user-images.githubusercontent.com/43583994/94999528-3f5c4680-05ba-11eb-8dc8-95c0dda64920.png) ![Hero meeting Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999534-46835480-05ba-11eb-9b8b-aa907e70b050.png) ![Castle Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999544-4b480880-05ba-11eb-8379-8d881d90ec6f.png) ![Castle construction Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999550-51d68000-05ba-11eb-9e49-7c44fbca187c.png) ![Castle mage guild Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999554-5864f780-05ba-11eb-84d2-50bf062d1192.png) ![Castle pop well Evil fh2 03](https://user-images.githubusercontent.com/43583994/94999560-5ef36f00-05ba-11eb-8a2a-733cb36f50e7.png) ![Kingdom Hero Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999566-6581e680-05ba-11eb-9cec-82debed56859.png) ![Kingdom Town Evil fh2 04](https://user-images.githubusercontent.com/43583994/94999571-6c105e00-05ba-11eb-95dc-b7d5bcf510ca.png)
1.0
Game screens, all "EVIL" windows are not centered vertically in the screen - In fheroes2, with the "EVIL" interface, all windows of this type are not centered vertically in the screen. In fheroes2: ![Quit confirmation window evil fh2 01](https://user-images.githubusercontent.com/43583994/94999522-323f5780-05ba-11eb-90f8-c5c69b765b71.png) ![Adventure Evil fh2 01](https://user-images.githubusercontent.com/43583994/94999527-39666580-05ba-11eb-8322-8e2ffd04a35f.png) ![Hero Evil fh2 03](https://user-images.githubusercontent.com/43583994/94999528-3f5c4680-05ba-11eb-8dc8-95c0dda64920.png) ![Hero meeting Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999534-46835480-05ba-11eb-9b8b-aa907e70b050.png) ![Castle Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999544-4b480880-05ba-11eb-8379-8d881d90ec6f.png) ![Castle construction Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999550-51d68000-05ba-11eb-9e49-7c44fbca187c.png) ![Castle mage guild Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999554-5864f780-05ba-11eb-84d2-50bf062d1192.png) ![Castle pop well Evil fh2 03](https://user-images.githubusercontent.com/43583994/94999560-5ef36f00-05ba-11eb-8a2a-733cb36f50e7.png) ![Kingdom Hero Evil fh2 02](https://user-images.githubusercontent.com/43583994/94999566-6581e680-05ba-11eb-9cec-82debed56859.png) ![Kingdom Town Evil fh2 04](https://user-images.githubusercontent.com/43583994/94999571-6c105e00-05ba-11eb-95dc-b7d5bcf510ca.png)
priority
game screens all evil windows are not centered vertically in the screen in with the evil interface all windows of this type are not centered vertically in the screen in
1
565,604
16,765,476,708
IssuesEvent
2021-06-14 08:17:48
enso-org/ide
https://api.github.com/repos/enso-org/ide
opened
Undo Redo does not work with upload progress updates.
Category: Controllers Difficulty: Core Contributor Priority: High Type: Bug
### Summary During uploading file we update the node's metadata about progress to keep information that file has been uploading when IDE has been closed, and also the node's expression is updated when uploading finishes - because we use a special expression during upload to have correct error messages in dependent nodes (see also # These operations should be transparent for the Undo Redo, because those are not user activities. However, with current UR architecture it is impossible, because each UR stack element have a full snapshot of AST - so even if the operations of uploading will be ignored, the other transactions committed during upload (user activity during upload) will have a trail of old metadata and node's expression. ### Value <!-- - This section should describe the value of this task. - This value can be for users, to the team, etc. --> ### Specification To be specified. ### Acceptance Criteria & Test Cases <!-- - Any criteria that must be satisfied for the task to be accepted. - The test plan for the feature, related to the acceptance criteria. -->
1.0
Undo Redo does not work with upload progress updates. - ### Summary During uploading file we update the node's metadata about progress to keep information that file has been uploading when IDE has been closed, and also the node's expression is updated when uploading finishes - because we use a special expression during upload to have correct error messages in dependent nodes (see also # These operations should be transparent for the Undo Redo, because those are not user activities. However, with current UR architecture it is impossible, because each UR stack element have a full snapshot of AST - so even if the operations of uploading will be ignored, the other transactions committed during upload (user activity during upload) will have a trail of old metadata and node's expression. ### Value <!-- - This section should describe the value of this task. - This value can be for users, to the team, etc. --> ### Specification To be specified. ### Acceptance Criteria & Test Cases <!-- - Any criteria that must be satisfied for the task to be accepted. - The test plan for the feature, related to the acceptance criteria. -->
priority
undo redo does not work with upload progress updates summary during uploading file we update the node s metadata about progress to keep information that file has been uploading when ide has been closed and also the node s expression is updated when uploading finishes because we use a special expression during upload to have correct error messages in dependent nodes see also these operations should be transparent for the undo redo because those are not user activities however with current ur architecture it is impossible because each ur stack element have a full snapshot of ast so even if the operations of uploading will be ignored the other transactions committed during upload user activity during upload will have a trail of old metadata and node s expression value this section should describe the value of this task this value can be for users to the team etc specification to be specified acceptance criteria test cases any criteria that must be satisfied for the task to be accepted the test plan for the feature related to the acceptance criteria
1
396,598
11,711,190,704
IssuesEvent
2020-03-09 04:06:45
aitorres/firelink
https://api.github.com/repos/aitorres/firelink
closed
Fix union access verification
medium priority
Union property access verification is being triggered on assignments (while unions are LVALs). They should ONLY be triggered on RVAL accesses, or particular LVAL accesses. Actually, a probable fix is to ONLY trigger them while parsing an Access expression. (maybe).
1.0
Fix union access verification - Union property access verification is being triggered on assignments (while unions are LVALs). They should ONLY be triggered on RVAL accesses, or particular LVAL accesses. Actually, a probable fix is to ONLY trigger them while parsing an Access expression. (maybe).
priority
fix union access verification union property access verification is being triggered on assignments while unions are lvals they should only be triggered on rval accesses or particular lval accesses actually a probable fix is to only trigger them while parsing an access expression maybe
1
248,228
7,928,294,213
IssuesEvent
2018-07-06 11:05:49
aharwood2/JBlock2D
https://api.github.com/repos/aharwood2/JBlock2D
closed
Error when trying to use larger batch file
bug high priority
Program won't run when a larger batch file of the same format is used
1.0
Error when trying to use larger batch file - Program won't run when a larger batch file of the same format is used
priority
error when trying to use larger batch file program won t run when a larger batch file of the same format is used
1
65,128
14,708,442,645
IssuesEvent
2021-01-04 23:43:14
SmartBear/readyapi-swagger-assertion-plugin
https://api.github.com/repos/SmartBear/readyapi-swagger-assertion-plugin
opened
CVE-2017-1000208 (High) detected in swagger-parser-1.0.8.jar
security vulnerability
## CVE-2017-1000208 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>swagger-parser-1.0.8.jar</b></p></summary> <p>Sonatype helps open source projects to set up Maven repositories on https://oss.sonatype.org/</p> <p>Library home page: <a href="http://nexus.sonatype.org/oss-repository-hosting.html/swagger-parser-project/modules/swagger-parser">http://nexus.sonatype.org/oss-repository-hosting.html/swagger-parser-project/modules/swagger-parser</a></p> <p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/swagger/swagger-parser/1.0.8/swagger-parser-1.0.8.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.7.0.jar (Root Library) - :x: **swagger-parser-1.0.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability in Swagger-Parser's (version <= 1.0.30) yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml Open-API specification is parsed. This in particular, affects the 'generate' and 'validate' command in swagger-codegen (<= 2.2.2) and can lead to arbitrary code being executed when these commands are used on a well-crafted yaml specification. <p>Publish Date: 2017-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000208>CVE-2017-1000208</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000208">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000208</a></p> <p>Release Date: 2017-11-17</p> <p>Fix Resolution: io.swagger:swagger-parser:1.0.31</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.swagger","packageName":"swagger-parser","packageVersion":"1.0.8","isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;io.swagger:swagger-parser:1.0.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.swagger:swagger-parser:1.0.31"}],"vulnerabilityIdentifier":"CVE-2017-1000208","vulnerabilityDetails":"A vulnerability in Swagger-Parser\u0027s (version \u003c\u003d 1.0.30) yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml Open-API specification is parsed. This in particular, affects the \u0027generate\u0027 and \u0027validate\u0027 command in swagger-codegen (\u003c\u003d 2.2.2) and can lead to arbitrary code being executed when these commands are used on a well-crafted yaml specification.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000208","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2017-1000208 (High) detected in swagger-parser-1.0.8.jar - ## CVE-2017-1000208 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>swagger-parser-1.0.8.jar</b></p></summary> <p>Sonatype helps open source projects to set up Maven repositories on https://oss.sonatype.org/</p> <p>Library home page: <a href="http://nexus.sonatype.org/oss-repository-hosting.html/swagger-parser-project/modules/swagger-parser">http://nexus.sonatype.org/oss-repository-hosting.html/swagger-parser-project/modules/swagger-parser</a></p> <p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/swagger/swagger-parser/1.0.8/swagger-parser-1.0.8.jar</p> <p> Dependency Hierarchy: - ready-api-soapui-pro-1.7.0.jar (Root Library) - :x: **swagger-parser-1.0.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability in Swagger-Parser's (version <= 1.0.30) yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml Open-API specification is parsed. This in particular, affects the 'generate' and 'validate' command in swagger-codegen (<= 2.2.2) and can lead to arbitrary code being executed when these commands are used on a well-crafted yaml specification. <p>Publish Date: 2017-11-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000208>CVE-2017-1000208</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000208">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000208</a></p> <p>Release Date: 2017-11-17</p> <p>Fix Resolution: io.swagger:swagger-parser:1.0.31</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.swagger","packageName":"swagger-parser","packageVersion":"1.0.8","isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;io.swagger:swagger-parser:1.0.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.swagger:swagger-parser:1.0.31"}],"vulnerabilityIdentifier":"CVE-2017-1000208","vulnerabilityDetails":"A vulnerability in Swagger-Parser\u0027s (version \u003c\u003d 1.0.30) yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml Open-API specification is parsed. This in particular, affects the \u0027generate\u0027 and \u0027validate\u0027 command in swagger-codegen (\u003c\u003d 2.2.2) and can lead to arbitrary code being executed when these commands are used on a well-crafted yaml specification.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000208","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_priority
cve high detected in swagger parser jar cve high severity vulnerability vulnerable library swagger parser jar sonatype helps open source projects to set up maven repositories on library home page a href path to dependency file readyapi swagger assertion plugin pom xml path to vulnerable library home wss scanner repository io swagger swagger parser swagger parser jar dependency hierarchy ready api soapui pro jar root library x swagger parser jar vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability in swagger parser s version yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml open api specification is parsed this in particular affects the generate and validate command in swagger codegen and can lead to arbitrary code being executed when these commands are used on a well crafted yaml specification publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io swagger swagger parser isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a vulnerability in swagger parser version yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml open api specification is parsed this in particular affects the and command in swagger codegen and can lead to arbitrary code being executed when these commands are used on a well crafted yaml specification vulnerabilityurl
0
229,434
25,343,418,115
IssuesEvent
2022-11-19 01:02:28
MidnightBSD/src
https://api.github.com/repos/MidnightBSD/src
closed
CVE-2022-41916 (High) detected in heimdalheimdal-7.7.0 - autoclosed
security vulnerability
## CVE-2022-41916 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>heimdalheimdal-7.7.0</b></p></summary> <p> <p>Heimdal</p> <p>Library home page: <a href=https://github.com/heimdal/heimdal.git>https://github.com/heimdal/heimdal.git</a></p> <p>Found in base branches: <b>stable/2.1, stable/2.2, master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/heimdal/lib/wind/normalize.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Heimdal is an implementation of ASN.1/DER, PKIX, and Kerberos. Versions prior to 7.7.1 are vulnerable to a denial of service vulnerability in Heimdal's PKI certificate validation library, affecting the KDC (via PKINIT) and kinit (via PKINIT), as well as any third-party applications using Heimdal's libhx509. Users should upgrade to Heimdal 7.7.1 or 7.8. There are no known workarounds for this issue. <p>Publish Date: 2022-11-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41916>CVE-2022-41916</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/heimdal/heimdal/security/advisories/GHSA-mgqr-gvh6-23cx">https://github.com/heimdal/heimdal/security/advisories/GHSA-mgqr-gvh6-23cx</a></p> <p>Release Date: 2022-11-15</p> <p>Fix Resolution: heimdal-7.7.1 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-41916 (High) detected in heimdalheimdal-7.7.0 - autoclosed - ## CVE-2022-41916 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>heimdalheimdal-7.7.0</b></p></summary> <p> <p>Heimdal</p> <p>Library home page: <a href=https://github.com/heimdal/heimdal.git>https://github.com/heimdal/heimdal.git</a></p> <p>Found in base branches: <b>stable/2.1, stable/2.2, master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/heimdal/lib/wind/normalize.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Heimdal is an implementation of ASN.1/DER, PKIX, and Kerberos. Versions prior to 7.7.1 are vulnerable to a denial of service vulnerability in Heimdal's PKI certificate validation library, affecting the KDC (via PKINIT) and kinit (via PKINIT), as well as any third-party applications using Heimdal's libhx509. Users should upgrade to Heimdal 7.7.1 or 7.8. There are no known workarounds for this issue. <p>Publish Date: 2022-11-15 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41916>CVE-2022-41916</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/heimdal/heimdal/security/advisories/GHSA-mgqr-gvh6-23cx">https://github.com/heimdal/heimdal/security/advisories/GHSA-mgqr-gvh6-23cx</a></p> <p>Release Date: 2022-11-15</p> <p>Fix Resolution: heimdal-7.7.1 </p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in heimdalheimdal autoclosed cve high severity vulnerability vulnerable library heimdalheimdal heimdal library home page a href found in base branches stable stable master vulnerable source files crypto heimdal lib wind normalize c vulnerability details heimdal is an implementation of asn der pkix and kerberos versions prior to are vulnerable to a denial of service vulnerability in heimdal s pki certificate validation library affecting the kdc via pkinit and kinit via pkinit as well as any third party applications using heimdal s users should upgrade to heimdal or there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution heimdal step up your open source security game with mend
0
128,133
10,516,402,570
IssuesEvent
2019-09-28 17:17:26
Vachok/ftpplus
https://api.github.com/repos/Vachok/ftpplus
closed
testSyncData [D172]
Medium TestQuality bug resolution_Dupe
Execute DBUploadUniversalTest::testSyncData\*\*testSyncData\*\* \*DBUploadUniversalTest\* \*expected [0 rows uploaded to test.test] but found [2 rows uploaded to test.test]\* \*java.lang.AssertionError\*
1.0
testSyncData [D172] - Execute DBUploadUniversalTest::testSyncData\*\*testSyncData\*\* \*DBUploadUniversalTest\* \*expected [0 rows uploaded to test.test] but found [2 rows uploaded to test.test]\* \*java.lang.AssertionError\*
non_priority
testsyncdata execute dbuploaduniversaltest testsyncdata testsyncdata dbuploaduniversaltest expected but found java lang assertionerror
0
549
2,499,784,057
IssuesEvent
2015-01-08 05:48:51
chessmasterhong/WaterEmblem
https://api.github.com/repos/chessmasterhong/WaterEmblem
closed
Bug with ending battle early
bug high priority
If the player kills the last attacking enemy (The enemy initiates the attack on the enemy phase, and is the last enemy on the map) and the attacking enemy had a droppable item (and if the map condition is to kill all enemies), the map will fade out before the player can choose what item to discard.
1.0
Bug with ending battle early - If the player kills the last attacking enemy (The enemy initiates the attack on the enemy phase, and is the last enemy on the map) and the attacking enemy had a droppable item (and if the map condition is to kill all enemies), the map will fade out before the player can choose what item to discard.
priority
bug with ending battle early if the player kills the last attacking enemy the enemy initiates the attack on the enemy phase and is the last enemy on the map and the attacking enemy had a droppable item and if the map condition is to kill all enemies the map will fade out before the player can choose what item to discard
1
514,056
14,932,336,628
IssuesEvent
2021-01-25 07:34:22
sct/overseerr
https://api.github.com/repos/sct/overseerr
closed
Use the Plex API to pre-populate the server information
enhancement priority:medium
**Is your feature request related to a problem? Please describe.** Plex.tv has an API call that can be made to get a list of servers a user has access to. Pulling this list would make the initial setup for users much easier since it is the same method the players use to discover the server. **Describe the solution you'd like** Use the Plex API to pre-populate the Plex server information. Additional details can be found here. https://support.plex.tv/articles/206721658-using-plex-tv-resources-information-to-troubleshoot-app-connections/
1.0
Use the Plex API to pre-populate the server information - **Is your feature request related to a problem? Please describe.** Plex.tv has an API call that can be made to get a list of servers a user has access to. Pulling this list would make the initial setup for users much easier since it is the same method the players use to discover the server. **Describe the solution you'd like** Use the Plex API to pre-populate the Plex server information. Additional details can be found here. https://support.plex.tv/articles/206721658-using-plex-tv-resources-information-to-troubleshoot-app-connections/
priority
use the plex api to pre populate the server information is your feature request related to a problem please describe plex tv has an api call that can be made to get a list of servers a user has access to pulling this list would make the initial setup for users much easier since it is the same method the players use to discover the server describe the solution you d like use the plex api to pre populate the plex server information additional details can be found here
1
4,875
7,753,598,812
IssuesEvent
2018-05-31 01:38:07
kubernetes-sigs/kubebuilder
https://api.github.com/repos/kubernetes-sigs/kubebuilder
closed
Create a SECURITY_CONTACTS file.
process
As per the email sent to kubernetes-dev[1], please create a SECURITY_CONTACTS file. The template for the file can be found in the kubernetes-template repository[2]. A description for the file is in the steering-committee docs[3], you might need to search that page for "Security Contacts". Please feel free to ping me on the PR when you make it, otherwise I will see when you close this issue. :) Thanks so much, let me know if you have any questions. (This issue was generated from a tool, apologies for any weirdness.) [1] https://groups.google.com/forum/#!topic/kubernetes-dev/codeiIoQ6QE [2] https://github.com/kubernetes/kubernetes-template-project/blob/master/SECURITY_CONTACTS [3] https://github.com/kubernetes/community/blob/master/committee-steering/governance/sig-governance-template-short.md
1.0
Create a SECURITY_CONTACTS file. - As per the email sent to kubernetes-dev[1], please create a SECURITY_CONTACTS file. The template for the file can be found in the kubernetes-template repository[2]. A description for the file is in the steering-committee docs[3], you might need to search that page for "Security Contacts". Please feel free to ping me on the PR when you make it, otherwise I will see when you close this issue. :) Thanks so much, let me know if you have any questions. (This issue was generated from a tool, apologies for any weirdness.) [1] https://groups.google.com/forum/#!topic/kubernetes-dev/codeiIoQ6QE [2] https://github.com/kubernetes/kubernetes-template-project/blob/master/SECURITY_CONTACTS [3] https://github.com/kubernetes/community/blob/master/committee-steering/governance/sig-governance-template-short.md
non_priority
create a security contacts file as per the email sent to kubernetes dev please create a security contacts file the template for the file can be found in the kubernetes template repository a description for the file is in the steering committee docs you might need to search that page for security contacts please feel free to ping me on the pr when you make it otherwise i will see when you close this issue thanks so much let me know if you have any questions this issue was generated from a tool apologies for any weirdness
0
398,202
11,739,246,193
IssuesEvent
2020-03-11 17:22:41
georchestra/mapstore2-georchestra
https://api.github.com/repos/georchestra/mapstore2-georchestra
closed
Documentation extension
Accepted Priority: Low
An extension must be provided to reach the online documentation related to the viewer and its installed extensions. - The URL of the documentation must be parametrizable by the administrator in backoffice. ![image](https://user-images.githubusercontent.com/1280027/76071818-24c7dd80-5f97-11ea-9446-7f97a74fb13b.png)
1.0
Documentation extension - An extension must be provided to reach the online documentation related to the viewer and its installed extensions. - The URL of the documentation must be parametrizable by the administrator in backoffice. ![image](https://user-images.githubusercontent.com/1280027/76071818-24c7dd80-5f97-11ea-9446-7f97a74fb13b.png)
priority
documentation extension an extension must be provided to reach the online documentation related to the viewer and its installed extensions the url of the documentation must be parametrizable by the administrator in backoffice
1
235,963
19,474,869,774
IssuesEvent
2021-12-24 10:09:24
optuna/optuna
https://api.github.com/repos/optuna/optuna
opened
Unify the tests between Plotly and matplotlib
test feature v3
<!-- Please write a clear and concise description of the feature proposal. --> ## Motivation This is a sub-issue of https://github.com/optuna/optuna/issues/2959. Tests cases are inconsistent between two visualisation backends: Plotly or Matplotlib. Due to this inconsistency, this might cause an expected bug. In addition, we find it difficult to maintain asymmetric tests. ## Description Please unify the test cases between both functions. For example, if either test file does not perform a test that has been done by the other, please add a test case. Note that the missing test case might be function level or fine-grained level. <!-- Please write a detailed description of the new feature. --> ## Alternatives (optional) <!-- Please write a clear and concise description of any alternative solutions or features you've considered. --> ## Additional context (optional) <!-- Please add any other context or screenshots about the feature request here. -->
1.0
Unify the tests between Plotly and matplotlib - <!-- Please write a clear and concise description of the feature proposal. --> ## Motivation This is a sub-issue of https://github.com/optuna/optuna/issues/2959. Tests cases are inconsistent between two visualisation backends: Plotly or Matplotlib. Due to this inconsistency, this might cause an expected bug. In addition, we find it difficult to maintain asymmetric tests. ## Description Please unify the test cases between both functions. For example, if either test file does not perform a test that has been done by the other, please add a test case. Note that the missing test case might be function level or fine-grained level. <!-- Please write a detailed description of the new feature. --> ## Alternatives (optional) <!-- Please write a clear and concise description of any alternative solutions or features you've considered. --> ## Additional context (optional) <!-- Please add any other context or screenshots about the feature request here. -->
non_priority
unify the tests between plotly and matplotlib motivation this is a sub issue of tests cases are inconsistent between two visualisation backends plotly or matplotlib due to this inconsistency this might cause an expected bug in addition we find it difficult to maintain asymmetric tests description please unify the test cases between both functions for example if either test file does not perform a test that has been done by the other please add a test case note that the missing test case might be function level or fine grained level alternatives optional additional context optional
0
125,768
10,352,818,652
IssuesEvent
2019-09-05 10:03:58
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
Nothing happens after clicking 'Create Disk' button
:gear: disks 🧪 testing
**Storage Explorer Version**: 1.10.0 **Build:** 20190905.11 **Branch:** master **Platform/OS:** Windows 10/Linux Ubuntu 18.04/macOS High Sierra **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand 'Disks' -> Open one resource group. 2. Click 'Create Disk' button on the toolbar. **Expect Experience:** The 'Create Disk' dialog displays. **Actual Experience:** Nothing happens. **More Info:** This issue also reproduces for 'Paste' button.
1.0
Nothing happens after clicking 'Create Disk' button - **Storage Explorer Version**: 1.10.0 **Build:** 20190905.11 **Branch:** master **Platform/OS:** Windows 10/Linux Ubuntu 18.04/macOS High Sierra **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Expand 'Disks' -> Open one resource group. 2. Click 'Create Disk' button on the toolbar. **Expect Experience:** The 'Create Disk' dialog displays. **Actual Experience:** Nothing happens. **More Info:** This issue also reproduces for 'Paste' button.
non_priority
nothing happens after clicking create disk button storage explorer version build branch master platform os windows linux ubuntu macos high sierra architecture regression from not a regression steps to reproduce expand disks open one resource group click create disk button on the toolbar expect experience the create disk dialog displays actual experience nothing happens more info this issue also reproduces for paste button
0
752,138
26,274,752,884
IssuesEvent
2023-01-06 20:41:54
zowe/imperative
https://api.github.com/repos/zowe/imperative
closed
Plugin validation fails when using chained handlers
bug priority-low severity-high
Plugin validation fails for command with chainedHandlers defined with message "CmdError: Command name = '***' has no 'handler' property", resulting in plugin commands not available.
1.0
Plugin validation fails when using chained handlers - Plugin validation fails for command with chainedHandlers defined with message "CmdError: Command name = '***' has no 'handler' property", resulting in plugin commands not available.
priority
plugin validation fails when using chained handlers plugin validation fails for command with chainedhandlers defined with message cmderror command name has no handler property resulting in plugin commands not available
1
154,751
19,758,348,713
IssuesEvent
2022-01-16 01:18:40
lemurchop/pancake-frontend
https://api.github.com/repos/lemurchop/pancake-frontend
opened
CVE-2022-0155 (High) detected in follow-redirects-1.14.3.tgz, follow-redirects-1.14.1.tgz
security vulnerability
## CVE-2022-0155 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>follow-redirects-1.14.3.tgz</b>, <b>follow-redirects-1.14.1.tgz</b></p></summary> <p> <details><summary><b>follow-redirects-1.14.3.tgz</b></p></summary> <p>HTTP and HTTPS modules that follow redirects.</p> <p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.3.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/axios/node_modules/follow-redirects/package.json</p> <p> Dependency Hierarchy: - start-server-and-test-1.14.0.tgz (Root Library) - wait-on-6.0.0.tgz - axios-0.21.3.tgz - :x: **follow-redirects-1.14.3.tgz** (Vulnerable Library) </details> <details><summary><b>follow-redirects-1.14.1.tgz</b></p></summary> <p>HTTP and HTTPS modules that follow redirects.</p> <p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.1.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/follow-redirects/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - webpack-dev-server-3.11.1.tgz - http-proxy-middleware-0.19.1.tgz - http-proxy-1.18.1.tgz - :x: **follow-redirects-1.14.1.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> follow-redirects is vulnerable to Exposure of Private Personal Information to an Unauthorized Actor <p>Publish Date: 2022-01-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155>CVE-2022-0155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/">https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: follow-redirects - v1.14.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0155 (High) detected in follow-redirects-1.14.3.tgz, follow-redirects-1.14.1.tgz - ## CVE-2022-0155 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>follow-redirects-1.14.3.tgz</b>, <b>follow-redirects-1.14.1.tgz</b></p></summary> <p> <details><summary><b>follow-redirects-1.14.3.tgz</b></p></summary> <p>HTTP and HTTPS modules that follow redirects.</p> <p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.3.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/axios/node_modules/follow-redirects/package.json</p> <p> Dependency Hierarchy: - start-server-and-test-1.14.0.tgz (Root Library) - wait-on-6.0.0.tgz - axios-0.21.3.tgz - :x: **follow-redirects-1.14.3.tgz** (Vulnerable Library) </details> <details><summary><b>follow-redirects-1.14.1.tgz</b></p></summary> <p>HTTP and HTTPS modules that follow redirects.</p> <p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.1.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/follow-redirects/package.json</p> <p> Dependency Hierarchy: - react-scripts-4.0.3.tgz (Root Library) - webpack-dev-server-3.11.1.tgz - http-proxy-middleware-0.19.1.tgz - http-proxy-1.18.1.tgz - :x: **follow-redirects-1.14.1.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> follow-redirects is vulnerable to Exposure of Private Personal Information to an Unauthorized Actor <p>Publish Date: 2022-01-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155>CVE-2022-0155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/">https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: follow-redirects - v1.14.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in follow redirects tgz follow redirects tgz cve high severity vulnerability vulnerable libraries follow redirects tgz follow redirects tgz follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file package json path to vulnerable library node modules axios node modules follow redirects package json dependency hierarchy start server and test tgz root library wait on tgz axios tgz x follow redirects tgz vulnerable library follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file package json path to vulnerable library node modules follow redirects package json dependency hierarchy react scripts tgz root library webpack dev server tgz http proxy middleware tgz http proxy tgz x follow redirects tgz vulnerable library found in base branch develop vulnerability details follow redirects is vulnerable to exposure of private personal information to an unauthorized actor publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution follow redirects step up your open source security game with whitesource
0
736,634
25,481,185,446
IssuesEvent
2022-11-25 21:19:38
diffgram/diffgram
https://api.github.com/repos/diffgram/diffgram
closed
Helm timeout
lowpriority helm
Add timeout parameter documentation for cases where the index creation db migration is too big.
1.0
Helm timeout - Add timeout parameter documentation for cases where the index creation db migration is too big.
priority
helm timeout add timeout parameter documentation for cases where the index creation db migration is too big
1
263,939
8,303,411,949
IssuesEvent
2018-09-21 17:29:11
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
nRF52: MPU Fault issue
bug nRF priority: high
Hi, This **zephyr/samples/boards/nrf52/mesh/onoff_level_lighting_vnd_app** App in latest master branch works perfectly normal with Zephyr v1.12.99 (with last commit ba6763a187a347cfc825a2bece78e7d1ef28772d). But with latest master branch or after v1.13 onward, I am facing issue of MPU FAULT intermittently. If we set "LIGHT_CTL_TT" in publisher.c & configure buttons to publish Light CTL set (acknowledged) messages then it get easily encountered while playing with on-boards buttons on #nRF52840_PDK boards.
1.0
nRF52: MPU Fault issue - Hi, This **zephyr/samples/boards/nrf52/mesh/onoff_level_lighting_vnd_app** App in latest master branch works perfectly normal with Zephyr v1.12.99 (with last commit ba6763a187a347cfc825a2bece78e7d1ef28772d). But with latest master branch or after v1.13 onward, I am facing issue of MPU FAULT intermittently. If we set "LIGHT_CTL_TT" in publisher.c & configure buttons to publish Light CTL set (acknowledged) messages then it get easily encountered while playing with on-boards buttons on #nRF52840_PDK boards.
priority
mpu fault issue hi this zephyr samples boards mesh onoff level lighting vnd app app in latest master branch works perfectly normal with zephyr with last commit but with latest master branch or after onward i am facing issue of mpu fault intermittently if we set light ctl tt in publisher c configure buttons to publish light ctl set acknowledged messages then it get easily encountered while playing with on boards buttons on pdk boards
1
331,900
29,170,244,980
IssuesEvent
2023-05-19 00:37:40
ray-project/ray
https://api.github.com/repos/ray-project/ray
closed
[Ray release glue.py] `Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.`
P1 testing
Seeing this in the following release tests: * dask_on_ray_1tb_sort https://buildkite.com/ray-project/release-tests-branch/builds/1304#0185d156-acdf-4b5c-b63f-9fc4f6b2f7c2 * rllib_stress_tests https://buildkite.com/ray-project/release-tests-branch/builds/1305#0185d81d-437e-49ff-9103-4bf92a4d11ea * shuffle_1tb_5000_partitions https://buildkite.com/ray-project/release-tests-branch/builds/1304#0185d156-ace6-4f9a-8255-25901c91daa7 * map_batches_benchmark_single_node https://buildkite.com/ray-project/release-tests-branch/builds/1333#0186130c-2057-4ecf-8be2-25bc4159fade ``` [ERROR 2023-01-21 22:46:41,582] run_release_test.py: 157 Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507. --   | Traceback (most recent call last):   | File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 69, in prepare_remote_env   | self.file_manager.upload()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 122, in upload   | self._push_local_dir()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 107, in _push_local_dir   | f"Error uploading local dir to session "   | ray_release.exception.FileUploadError: Error uploading local dir to session rllib_stress_tests_1674368507.   |     | The above exception was the direct cause of the following exception:   |     | Traceback (most recent call last):   | File "ray_release/scripts/run_release_test.py", line 153, in main   | no_terminate=no_terminate,   | File "/tmp/release-4K7NOSGKrN/release/ray_release/glue.py", line 404, in run_release_test   | raise pipeline_exception   | File "/tmp/release-4K7NOSGKrN/release/ray_release/glue.py", line 275, in run_release_test   | command_runner.prepare_remote_env()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 73, in prepare_remote_env   | ) from e   | ray_release.exception.RemoteEnvSetupError: Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.   | ERROR:ray_release.logger:Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.   | Traceback (most recent call last):   | File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 69, in prepare_remote_env   | self.file_manager.upload()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 122, in upload   | self._push_local_dir()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 107, in _push_local_dir   | f"Error uploading local dir to session "   | ray_release.exception.FileUploadError: Error uploading local dir to session rllib_stress_tests_1674368507. ``` cc @krfricke
1.0
[Ray release glue.py] `Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.` - Seeing this in the following release tests: * dask_on_ray_1tb_sort https://buildkite.com/ray-project/release-tests-branch/builds/1304#0185d156-acdf-4b5c-b63f-9fc4f6b2f7c2 * rllib_stress_tests https://buildkite.com/ray-project/release-tests-branch/builds/1305#0185d81d-437e-49ff-9103-4bf92a4d11ea * shuffle_1tb_5000_partitions https://buildkite.com/ray-project/release-tests-branch/builds/1304#0185d156-ace6-4f9a-8255-25901c91daa7 * map_batches_benchmark_single_node https://buildkite.com/ray-project/release-tests-branch/builds/1333#0186130c-2057-4ecf-8be2-25bc4159fade ``` [ERROR 2023-01-21 22:46:41,582] run_release_test.py: 157 Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507. --   | Traceback (most recent call last):   | File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 69, in prepare_remote_env   | self.file_manager.upload()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 122, in upload   | self._push_local_dir()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 107, in _push_local_dir   | f"Error uploading local dir to session "   | ray_release.exception.FileUploadError: Error uploading local dir to session rllib_stress_tests_1674368507.   |     | The above exception was the direct cause of the following exception:   |     | Traceback (most recent call last):   | File "ray_release/scripts/run_release_test.py", line 153, in main   | no_terminate=no_terminate,   | File "/tmp/release-4K7NOSGKrN/release/ray_release/glue.py", line 404, in run_release_test   | raise pipeline_exception   | File "/tmp/release-4K7NOSGKrN/release/ray_release/glue.py", line 275, in run_release_test   | command_runner.prepare_remote_env()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 73, in prepare_remote_env   | ) from e   | ray_release.exception.RemoteEnvSetupError: Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.   | ERROR:ray_release.logger:Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.   | Traceback (most recent call last):   | File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 69, in prepare_remote_env   | self.file_manager.upload()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 122, in upload   | self._push_local_dir()   | File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 107, in _push_local_dir   | f"Error uploading local dir to session "   | ray_release.exception.FileUploadError: Error uploading local dir to session rllib_stress_tests_1674368507. ``` cc @krfricke
non_priority
error setting up remote environment error uploading local dir to session rllib stress tests seeing this in the following release tests dask on ray sort rllib stress tests shuffle partitions map batches benchmark single node run release test py error setting up remote environment error uploading local dir to session rllib stress tests   traceback most recent call last   file tmp release release ray release command runner sdk runner py line in prepare remote env   self file manager upload   file tmp release release ray release file manager job file manager py line in upload   self push local dir   file tmp release release ray release file manager job file manager py line in push local dir   f error uploading local dir to session   ray release exception fileuploaderror error uploading local dir to session rllib stress tests       the above exception was the direct cause of the following exception       traceback most recent call last   file ray release scripts run release test py line in main   no terminate no terminate   file tmp release release ray release glue py line in run release test   raise pipeline exception   file tmp release release ray release glue py line in run release test   command runner prepare remote env   file tmp release release ray release command runner sdk runner py line in prepare remote env   from e   ray release exception remoteenvsetuperror error setting up remote environment error uploading local dir to session rllib stress tests   error ray release logger error setting up remote environment error uploading local dir to session rllib stress tests   traceback most recent call last   file tmp release release ray release command runner sdk runner py line in prepare remote env   self file manager upload   file tmp release release ray release file manager job file manager py line in upload   self push local dir   file tmp release release ray release file manager job file manager py line in push local dir   f error uploading local dir to session   ray release exception fileuploaderror error uploading local dir to session rllib stress tests cc krfricke
0
11,878
5,104,087,964
IssuesEvent
2017-01-04 23:38:16
angular/material
https://api.github.com/repos/angular/material
closed
md-chips has aria-hidden='true' which makes it hidden from the screenreader
has: Pull Request P0 - Critical Fixes / Build or Deployment Issues ui: a11y
**Actual Behavior**: - aria-hidden='true' is applied to the md-chips directive, which causes the the content inside md-chips does not get announced by the screenreader. - Expect no aria-hidden='true' on the md-chips, so that the items inside it can be read by screenreader. **Angular Versions**: * - `Angular Version:` 1.5.8 - `Angular Material Version:` 1.5.8 **Additional Information**: - `Browser Type: *` Chrome
1.0
md-chips has aria-hidden='true' which makes it hidden from the screenreader - **Actual Behavior**: - aria-hidden='true' is applied to the md-chips directive, which causes the the content inside md-chips does not get announced by the screenreader. - Expect no aria-hidden='true' on the md-chips, so that the items inside it can be read by screenreader. **Angular Versions**: * - `Angular Version:` 1.5.8 - `Angular Material Version:` 1.5.8 **Additional Information**: - `Browser Type: *` Chrome
non_priority
md chips has aria hidden true which makes it hidden from the screenreader actual behavior aria hidden true is applied to the md chips directive which causes the the content inside md chips does not get announced by the screenreader expect no aria hidden true on the md chips so that the items inside it can be read by screenreader angular versions angular version angular material version additional information browser type chrome
0
122,090
4,827,290,583
IssuesEvent
2016-11-07 13:09:16
LuxoftAKutsan/RequirementsTest
https://api.github.com/repos/LuxoftAKutsan/RequirementsTest
opened
[GENIVI] SDL should respond "IGNORED" with correct result code for UnSubscribeVehicleData in case vi interface isn't available
bug high_priority
**Requirement:** APPLINK-20043 **Occurrence:** Always **Preconditions:** 1. SubscribeVehicleData and UnSubscribeVehicleData are allowed in PT **Steps to reproduce:** 1. RegisterApp in case HMI does not respond to VehicleInfo.IsReady. 2. Activate app 3. Send SubscribeVehicleData with gps = true -> SDL responds with SUCCESS. 4. Send UnSubscribeVehicleData with gps = true **Expected result:** SDL should respond " {success = false, resultCode = "IGNORED", info = "Some provided VehicleData was not subscribed."} " code to mobile app. **Actual result:** SDL respond " {success = true, resultCode = "IGNORED", info = "Already subscribed on some provided VehicleData."} ". **Problem can be reproduced with ATF script:** 1. Move IsReady_Template folder into sdl_atf\user_modules 2. Go to atf folder -> Run command: ./start.sh ../ATF_VehicleInfo_IsReady_available_false_SingleRPC.lua **Attachments:**
1.0
[GENIVI] SDL should respond "IGNORED" with correct result code for UnSubscribeVehicleData in case vi interface isn't available - **Requirement:** APPLINK-20043 **Occurrence:** Always **Preconditions:** 1. SubscribeVehicleData and UnSubscribeVehicleData are allowed in PT **Steps to reproduce:** 1. RegisterApp in case HMI does not respond to VehicleInfo.IsReady. 2. Activate app 3. Send SubscribeVehicleData with gps = true -> SDL responds with SUCCESS. 4. Send UnSubscribeVehicleData with gps = true **Expected result:** SDL should respond " {success = false, resultCode = "IGNORED", info = "Some provided VehicleData was not subscribed."} " code to mobile app. **Actual result:** SDL respond " {success = true, resultCode = "IGNORED", info = "Already subscribed on some provided VehicleData."} ". **Problem can be reproduced with ATF script:** 1. Move IsReady_Template folder into sdl_atf\user_modules 2. Go to atf folder -> Run command: ./start.sh ../ATF_VehicleInfo_IsReady_available_false_SingleRPC.lua **Attachments:**
priority
sdl should respond ignored with correct result code for unsubscribevehicledata in case vi interface isn t available requirement applink occurrence always preconditions subscribevehicledata and unsubscribevehicledata are allowed in pt steps to reproduce registerapp in case hmi does not respond to vehicleinfo isready activate app send subscribevehicledata with gps true sdl responds with success send unsubscribevehicledata with gps true expected result sdl should respond success false resultcode ignored info some provided vehicledata was not subscribed code to mobile app actual result sdl respond success true resultcode ignored info already subscribed on some provided vehicledata problem can be reproduced with atf script move isready template folder into sdl atf user modules go to atf folder run command start sh atf vehicleinfo isready available false singlerpc lua attachments
1
130,330
18,155,766,998
IssuesEvent
2021-09-27 01:12:25
benlazarine/cas-overlay
https://api.github.com/repos/benlazarine/cas-overlay
opened
CVE-2019-17531 (High) detected in jackson-databind-2.9.5.jar
security vulnerability
## CVE-2019-17531 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: cas-overlay/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - cas-server-support-oauth-webflow-5.3.7.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload. <p>Publish Date: 2019-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531>CVE-2019-17531</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531</a></p> <p>Release Date: 2019-10-12</p> <p>Fix Resolution: 2.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-17531 (High) detected in jackson-databind-2.9.5.jar - ## CVE-2019-17531 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: cas-overlay/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - cas-server-support-oauth-webflow-5.3.7.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload. <p>Publish Date: 2019-10-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531>CVE-2019-17531</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531</a></p> <p>Release Date: 2019-10-12</p> <p>Fix Resolution: 2.10</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file cas overlay pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy cas server support oauth webflow jar root library x jackson databind jar vulnerable library vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the apache extra version x jar in the classpath and an attacker can provide a jndi service to access it is possible to make the service execute a malicious payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
334,570
10,142,506,143
IssuesEvent
2019-08-04 01:34:14
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
jx install is skipped if (some undetermined) options specified to jx
area/install kind/bug lifecycle/rotten priority/important-soon
### Summary I attempted to install jx with `jx create terraform --environment-git-owner my-org --verbose -o testing --cluster goodclustername=gke --git-username=my-bot --git-api-token=somelonghash` after terraform created the cluster I expected jx to be installed. ### Steps to reproduce the behavior `jx create terraform --environment-git-owner my-org --verbose -o testing --cluster goodclustername=gke --git-username=my-bot --git-api-token=somelonghash` ### Jx version The output of `jx version` is: ``` ``` ### Kubernetes cluster What kind of Kubernetes cluster are you using & how did you create it? ### Operating system / Environment ### Expected behavior the cluster is created using the options supplied **and jx is installed**. ### Actual behavior The install of jx was "skipped" after the creation of the cluster ``` <snip> google_container_node_pool.jx-node-pool: Still creating... (2m50s elapsed) google_container_node_pool.jx-node-pool: Creation complete after 2m50s (ID: europe-west2-a/jamesntesting-jamesntesting/default-pool) Apply complete! Resources: 2 added, 0 changed, 0 destroyed. Outputs: cluster_endpoint = 35.242.180.127 cluster_master_version = 1.9.7-gke.11 Fetching cluster endpoint and auth data. kubeconfig entry generated for jamesntesting-jamesntesting.Skipping jx install ```
1.0
jx install is skipped if (some undetermined) options specified to jx - ### Summary I attempted to install jx with `jx create terraform --environment-git-owner my-org --verbose -o testing --cluster goodclustername=gke --git-username=my-bot --git-api-token=somelonghash` after terraform created the cluster I expected jx to be installed. ### Steps to reproduce the behavior `jx create terraform --environment-git-owner my-org --verbose -o testing --cluster goodclustername=gke --git-username=my-bot --git-api-token=somelonghash` ### Jx version The output of `jx version` is: ``` ``` ### Kubernetes cluster What kind of Kubernetes cluster are you using & how did you create it? ### Operating system / Environment ### Expected behavior the cluster is created using the options supplied **and jx is installed**. ### Actual behavior The install of jx was "skipped" after the creation of the cluster ``` <snip> google_container_node_pool.jx-node-pool: Still creating... (2m50s elapsed) google_container_node_pool.jx-node-pool: Creation complete after 2m50s (ID: europe-west2-a/jamesntesting-jamesntesting/default-pool) Apply complete! Resources: 2 added, 0 changed, 0 destroyed. Outputs: cluster_endpoint = 35.242.180.127 cluster_master_version = 1.9.7-gke.11 Fetching cluster endpoint and auth data. kubeconfig entry generated for jamesntesting-jamesntesting.Skipping jx install ```
priority
jx install is skipped if some undetermined options specified to jx summary i attempted to install jx with jx create terraform environment git owner my org verbose o testing cluster goodclustername gke git username my bot git api token somelonghash after terraform created the cluster i expected jx to be installed steps to reproduce the behavior jx create terraform environment git owner my org verbose o testing cluster goodclustername gke git username my bot git api token somelonghash jx version the output of jx version is kubernetes cluster what kind of kubernetes cluster are you using how did you create it operating system environment expected behavior the cluster is created using the options supplied and jx is installed actual behavior the install of jx was skipped after the creation of the cluster google container node pool jx node pool still creating elapsed google container node pool jx node pool creation complete after id europe a jamesntesting jamesntesting default pool apply complete resources added changed destroyed outputs cluster endpoint cluster master version gke fetching cluster endpoint and auth data kubeconfig entry generated for jamesntesting jamesntesting skipping jx install
1
827,562
31,778,715,985
IssuesEvent
2023-09-12 15:54:33
opentensor/validators
https://api.github.com/repos/opentensor/validators
closed
Diversity model throwing exception
bug low priority
There are some runs logged in wandb that points to an error in the diversity model: https://wandb.ai/opentensor-dev/openvalidators/runs/3em5sbqf/logs?workspace=user-opentensor-pedro By the look of it, it seems to be the very same error that some of models in the reward/mask stack were facing: Whenever the text > context window of the model, the validator throws with this exception. I believe it's worth investigating, even though the cases for this type of this exception is not that common
1.0
Diversity model throwing exception - There are some runs logged in wandb that points to an error in the diversity model: https://wandb.ai/opentensor-dev/openvalidators/runs/3em5sbqf/logs?workspace=user-opentensor-pedro By the look of it, it seems to be the very same error that some of models in the reward/mask stack were facing: Whenever the text > context window of the model, the validator throws with this exception. I believe it's worth investigating, even though the cases for this type of this exception is not that common
priority
diversity model throwing exception there are some runs logged in wandb that points to an error in the diversity model by the look of it it seems to be the very same error that some of models in the reward mask stack were facing whenever the text context window of the model the validator throws with this exception i believe it s worth investigating even though the cases for this type of this exception is not that common
1
825,635
31,465,145,649
IssuesEvent
2023-08-30 00:58:04
War-Brokers/.github
https://api.github.com/repos/War-Brokers/.github
closed
FPS settings goes to 1
type:bug priority:1 - high
When players close the game with vsync turned off and their fps set at 500, their FPS is set to 1 when they reopen the game
1.0
FPS settings goes to 1 - When players close the game with vsync turned off and their fps set at 500, their FPS is set to 1 when they reopen the game
priority
fps settings goes to when players close the game with vsync turned off and their fps set at their fps is set to when they reopen the game
1
670,115
22,675,260,059
IssuesEvent
2022-07-04 03:22:27
bitfoundation/bitplatform
https://api.github.com/repos/bitfoundation/bitplatform
closed
The `AutoInject` source generator ignores non-partial classes
high priority area / source generator
The `AutoInject` source generator should not ignore non-partial classes. It should always generate the source code for all occurrences of the attribute.
1.0
The `AutoInject` source generator ignores non-partial classes - The `AutoInject` source generator should not ignore non-partial classes. It should always generate the source code for all occurrences of the attribute.
priority
the autoinject source generator ignores non partial classes the autoinject source generator should not ignore non partial classes it should always generate the source code for all occurrences of the attribute
1
110,491
11,704,451,503
IssuesEvent
2020-03-07 09:33:36
bitfocus/companion-module-requests
https://api.github.com/repos/bitfocus/companion-module-requests
closed
SONOS
Hardware Missing documentation
I would like to user streamdeck for sonos volume up/down and previous/next song in the playlist. ...maybe choose a playlist but that isn´t much important.
1.0
SONOS - I would like to user streamdeck for sonos volume up/down and previous/next song in the playlist. ...maybe choose a playlist but that isn´t much important.
non_priority
sonos i would like to user streamdeck for sonos volume up down and previous next song in the playlist maybe choose a playlist but that isn´t much important
0
375,354
11,103,100,011
IssuesEvent
2019-12-17 02:31:59
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
[0.33.7.1] Cannot create new users or send emails (pulses/alerts/invitations)
Administration/ Priority:P0 Reporting/Pulses Type:Bug
**Describe the bug** Cannot create new users on 0.33.7.1, or sending pulse mails. **Logs** When trying to create a user: ``` 12-15 11:43:04 ERROR middleware.log :: POST /api/user 500 247.0 ms (9 DB calls) {:message "No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil", :type java.lang.IllegalArgumentException, :stacktrace ("clojure.core$_cache_protocol_fn.invokeStatic(core_deftype.clj:583)" "clojure.core$_cache_protocol_fn.invoke(core_deftype.clj:575)" "stencil.ast$fn__30303$G__30298__30312.invoke(ast.clj:19)" "stencil.core$render.invokeStatic(core.clj:80)" "stencil.core$render.invoke(core.clj:74)" "stencil.core$render_file.invokeStatic(core.clj:87)" "stencil.core$render_file.invoke(core.clj:83)" "--> email.messages$send_new_user_email_BANG_.invokeStatic(messages.clj:67)" "email.messages$send_new_user_email_BANG_.invoke(messages.clj:63)" "models.user$send_welcome_email_BANG_.invokeStatic(user.clj:176)" "models.user$send_welcome_email_BANG_.invoke(user.clj:172)" "models.user$fn__32292$create_and_invite_user_BANG___32297$fn__32298.invoke(user.clj:208)" "models.user$fn__32292$create_and_invite_user_BANG___32297.invoke(user.clj:203)" "api.user$fn__63972$fn__63975.invoke(user.clj:101)" "api.user$fn__63972.invokeStatic(user.clj:100)" "api.user$fn__63972.invoke(user.clj:89)" "middleware.auth$enforce_authentication$fn__64360.invoke(auth.clj:14)" "routes$fn__65737$fn__65738.doInvoke(routes.clj:56)" "middleware.exceptions$catch_uncaught_exceptions$fn__64459.invoke(exceptions.clj:104)" "middleware.exceptions$catch_api_exceptions$fn__64456.invoke(exceptions.clj:92)" "middleware.log$log_api_call$fn__66111$fn__66112.invoke(log.clj:170)" "middleware.log$log_api_call$fn__66111.invoke(log.clj:164)" "middleware.security$add_security_headers$fn__64422.invoke(security.clj:122)" "middleware.json$wrap_json_body$fn__65816.invoke(json.clj:61)" "middleware.json$wrap_streamed_json_response$fn__65834.invoke(json.clj:97)" "middleware.session$bind_current_user$fn__60984$fn__60985.invoke(session.clj:209)" "middleware.session$do_with_current_user.invokeStatic(session.clj:184)" "middleware.session$do_with_current_user.invoke(session.clj:177)" "middleware.session$bind_current_user$fn__60984.invoke(session.clj:208)" "middleware.session$wrap_current_user_id$fn__60969.invoke(session.clj:161)" "middleware.session$wrap_session_id$fn__60954.invoke(session.clj:123)" "middleware.auth$wrap_api_key$fn__64368.invoke(auth.clj:27)" "middleware.misc$maybe_set_site_url$fn__66138.invoke(misc.clj:56)" "middleware.misc$bind_user_locale$fn__66141.invoke(misc.clj:72)" "middleware.misc$add_content_type$fn__66126.invoke(misc.clj:28)" "middleware.misc$disable_streaming_buffering$fn__66149.invoke(misc.clj:87)")} ``` When trying to send pulse test email: ``` 12-15 12:49:41 ERROR middleware.log :: POST /api/pulse/test 500 187.6 ms (15 DB calls) {:message "No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil", :type java.lang.IllegalArgumentException, :stacktrace ... ``` When sending scheduled pulse: ``` 12-15 13:00:00 INFO task.send-pulses :: Sending scheduled pulses... 12-15 13:00:00 ERROR task.send-pulses :: Error sending Pulse 1 java.lang.IllegalArgumentException: No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil ... ``` **To Reproduce** - Go to Admin > People > Add someone > fill in details and click Create - Go to a pulse and click Send test email - Go to a pulse and schedule a send, and check the log when pulse should have been sent - ...perhaps there's other sections not working - will update as I find them... **Information about your Metabase Installation:** Metabase 0.33.7.1 Docker and JAR on various hosting **Severity** P0 **Additional context** https://discourse.metabase.com/t/cant-add-a-new-user-no-implementation-of-method-render-error/8350
1.0
[0.33.7.1] Cannot create new users or send emails (pulses/alerts/invitations) - **Describe the bug** Cannot create new users on 0.33.7.1, or sending pulse mails. **Logs** When trying to create a user: ``` 12-15 11:43:04 ERROR middleware.log :: POST /api/user 500 247.0 ms (9 DB calls) {:message "No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil", :type java.lang.IllegalArgumentException, :stacktrace ("clojure.core$_cache_protocol_fn.invokeStatic(core_deftype.clj:583)" "clojure.core$_cache_protocol_fn.invoke(core_deftype.clj:575)" "stencil.ast$fn__30303$G__30298__30312.invoke(ast.clj:19)" "stencil.core$render.invokeStatic(core.clj:80)" "stencil.core$render.invoke(core.clj:74)" "stencil.core$render_file.invokeStatic(core.clj:87)" "stencil.core$render_file.invoke(core.clj:83)" "--> email.messages$send_new_user_email_BANG_.invokeStatic(messages.clj:67)" "email.messages$send_new_user_email_BANG_.invoke(messages.clj:63)" "models.user$send_welcome_email_BANG_.invokeStatic(user.clj:176)" "models.user$send_welcome_email_BANG_.invoke(user.clj:172)" "models.user$fn__32292$create_and_invite_user_BANG___32297$fn__32298.invoke(user.clj:208)" "models.user$fn__32292$create_and_invite_user_BANG___32297.invoke(user.clj:203)" "api.user$fn__63972$fn__63975.invoke(user.clj:101)" "api.user$fn__63972.invokeStatic(user.clj:100)" "api.user$fn__63972.invoke(user.clj:89)" "middleware.auth$enforce_authentication$fn__64360.invoke(auth.clj:14)" "routes$fn__65737$fn__65738.doInvoke(routes.clj:56)" "middleware.exceptions$catch_uncaught_exceptions$fn__64459.invoke(exceptions.clj:104)" "middleware.exceptions$catch_api_exceptions$fn__64456.invoke(exceptions.clj:92)" "middleware.log$log_api_call$fn__66111$fn__66112.invoke(log.clj:170)" "middleware.log$log_api_call$fn__66111.invoke(log.clj:164)" "middleware.security$add_security_headers$fn__64422.invoke(security.clj:122)" "middleware.json$wrap_json_body$fn__65816.invoke(json.clj:61)" "middleware.json$wrap_streamed_json_response$fn__65834.invoke(json.clj:97)" "middleware.session$bind_current_user$fn__60984$fn__60985.invoke(session.clj:209)" "middleware.session$do_with_current_user.invokeStatic(session.clj:184)" "middleware.session$do_with_current_user.invoke(session.clj:177)" "middleware.session$bind_current_user$fn__60984.invoke(session.clj:208)" "middleware.session$wrap_current_user_id$fn__60969.invoke(session.clj:161)" "middleware.session$wrap_session_id$fn__60954.invoke(session.clj:123)" "middleware.auth$wrap_api_key$fn__64368.invoke(auth.clj:27)" "middleware.misc$maybe_set_site_url$fn__66138.invoke(misc.clj:56)" "middleware.misc$bind_user_locale$fn__66141.invoke(misc.clj:72)" "middleware.misc$add_content_type$fn__66126.invoke(misc.clj:28)" "middleware.misc$disable_streaming_buffering$fn__66149.invoke(misc.clj:87)")} ``` When trying to send pulse test email: ``` 12-15 12:49:41 ERROR middleware.log :: POST /api/pulse/test 500 187.6 ms (15 DB calls) {:message "No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil", :type java.lang.IllegalArgumentException, :stacktrace ... ``` When sending scheduled pulse: ``` 12-15 13:00:00 INFO task.send-pulses :: Sending scheduled pulses... 12-15 13:00:00 ERROR task.send-pulses :: Error sending Pulse 1 java.lang.IllegalArgumentException: No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil ... ``` **To Reproduce** - Go to Admin > People > Add someone > fill in details and click Create - Go to a pulse and click Send test email - Go to a pulse and schedule a send, and check the log when pulse should have been sent - ...perhaps there's other sections not working - will update as I find them... **Information about your Metabase Installation:** Metabase 0.33.7.1 Docker and JAR on various hosting **Severity** P0 **Additional context** https://discourse.metabase.com/t/cant-add-a-new-user-no-implementation-of-method-render-error/8350
priority
cannot create new users or send emails pulses alerts invitations describe the bug cannot create new users on or sending pulse mails logs when trying to create a user error middleware log post api user ms db calls message no implementation of method render of protocol stencil ast astnode found for class nil type java lang illegalargumentexception stacktrace clojure core cache protocol fn invokestatic core deftype clj clojure core cache protocol fn invoke core deftype clj stencil ast fn g invoke ast clj stencil core render invokestatic core clj stencil core render invoke core clj stencil core render file invokestatic core clj stencil core render file invoke core clj email messages send new user email bang invokestatic messages clj email messages send new user email bang invoke messages clj models user send welcome email bang invokestatic user clj models user send welcome email bang invoke user clj models user fn create and invite user bang fn invoke user clj models user fn create and invite user bang invoke user clj api user fn fn invoke user clj api user fn invokestatic user clj api user fn invoke user clj middleware auth enforce authentication fn invoke auth clj routes fn fn doinvoke routes clj middleware exceptions catch uncaught exceptions fn invoke exceptions clj middleware exceptions catch api exceptions fn invoke exceptions clj middleware log log api call fn fn invoke log clj middleware log log api call fn invoke log clj middleware security add security headers fn invoke security clj middleware json wrap json body fn invoke json clj middleware json wrap streamed json response fn invoke json clj middleware session bind current user fn fn invoke session clj middleware session do with current user invokestatic session clj middleware session do with current user invoke session clj middleware session bind current user fn invoke session clj middleware session wrap current user id fn invoke session clj middleware session wrap session id fn invoke session clj middleware auth wrap api key fn invoke auth clj middleware misc maybe set site url fn invoke misc clj middleware misc bind user locale fn invoke misc clj middleware misc add content type fn invoke misc clj middleware misc disable streaming buffering fn invoke misc clj when trying to send pulse test email error middleware log post api pulse test ms db calls message no implementation of method render of protocol stencil ast astnode found for class nil type java lang illegalargumentexception stacktrace when sending scheduled pulse info task send pulses sending scheduled pulses error task send pulses error sending pulse java lang illegalargumentexception no implementation of method render of protocol stencil ast astnode found for class nil to reproduce go to admin people add someone fill in details and click create go to a pulse and click send test email go to a pulse and schedule a send and check the log when pulse should have been sent perhaps there s other sections not working will update as i find them information about your metabase installation metabase docker and jar on various hosting severity additional context
1
276,695
30,521,836,937
IssuesEvent
2023-07-19 08:36:57
kseb49/my_blog
https://api.github.com/repos/kseb49/my_blog
closed
Handle the expired confirmed link
backend security
Generate a new link to confirmed an account `// Expired link if ($register->newToken() !== false) { $mail = new Mail(); $message = $this->twig->render("templates/mail/validation-mail.twig", ["link" => $register->link,]); $mail->mail($register->user['email'],$message,"Confirmez votre compte", $register->user['f_name']." ".$register->user['l_name'], 'Recopier ce lien pour valider votre compte : '.$register->link); Flash::flash('success', 'Ce lien est expiré, vous avez reçu un nouveau mail pour confirmer votre compte'); $this->redirect(); }` 18/07/2023
True
Handle the expired confirmed link - Generate a new link to confirmed an account `// Expired link if ($register->newToken() !== false) { $mail = new Mail(); $message = $this->twig->render("templates/mail/validation-mail.twig", ["link" => $register->link,]); $mail->mail($register->user['email'],$message,"Confirmez votre compte", $register->user['f_name']." ".$register->user['l_name'], 'Recopier ce lien pour valider votre compte : '.$register->link); Flash::flash('success', 'Ce lien est expiré, vous avez reçu un nouveau mail pour confirmer votre compte'); $this->redirect(); }` 18/07/2023
non_priority
handle the expired confirmed link generate a new link to confirmed an account expired link if register newtoken false mail new mail message this twig render templates mail validation mail twig mail mail register user message confirmez votre compte register user register user recopier ce lien pour valider votre compte register link flash flash success ce lien est expiré vous avez reçu un nouveau mail pour confirmer votre compte this redirect
0
800,674
28,374,784,354
IssuesEvent
2023-04-12 19:54:53
GoogleCloudPlatform/microservices-demo
https://api.github.com/repos/GoogleCloudPlatform/microservices-demo
closed
Unable to build adservice
type: bug priority: p2
### Describe the bug build [adservice] failed: exit status 1. Docker build ran into internal error. Please retry. ### To Reproduce <!-- Steps to reproduce the behavior: --> <!-- 1. Built image '...' --> <!-- 2. Ran command '....' --> <!-- 3. See error --> ### Logs Generating tags... - productcatalogservice -> productcatalogservice:v0.5.0-30-gb2ef53f - recommendationservice -> recommendationservice:v0.5.0-30-gb2ef53f - shippingservice -> shippingservice:v0.5.0-30-gb2ef53f - checkoutservice -> checkoutservice:v0.5.0-30-gb2ef53f - paymentservice -> paymentservice:v0.5.0-30-gb2ef53f - currencyservice -> currencyservice:v0.5.0-30-gb2ef53f - cartservice -> cartservice:v0.5.0-30-gb2ef53f - frontend -> frontend:v0.5.0-30-gb2ef53f - adservice -> adservice:v0.5.0-30-gb2ef53f - loadgenerator -> loadgenerator:v0.5.0-30-gb2ef53f Checking cache... - emailservice: Not found. Building - productcatalogservice: Not found. Building - recommendationservice: Not found. Building - shippingservice: Found Locally - checkoutservice: Found Locally - paymentservice: Found Locally - currencyservice: Found Locally - cartservice: Found Locally - frontend: Found Locally - adservice: Not found. Building - loadgenerator: Found Locally Starting build... Found [minikube] context, using local docker daemon. Building [adservice]... Target platforms: [linux/amd64] #1 [internal] load build definition from Dockerfile #1 sha256:4fd7c5b169637ade6dc46fa79f5ebe497faea8466f82e301c0fc5d1799fe1763 #1 transferring dockerfile: 38B done #1 DONE 0.0s #2 [internal] load .dockerignore #2 sha256:3a3376a110bfc911c611e7778255bed1a6fcc5f12331b0f88bd3f8a17922905e #2 transferring context: 2B done #2 DONE 0.0s #4 [internal] load metadata for docker.io/library/eclipse-temurin:19@sha256:17e3d3b61ca4a7606490f596feb77f69980939fecacf91a1f13ea7b17147058f #4 sha256:fa4367848242f6486a2e5650a6e73e7b59134bc9cb3dc1bc2d5c9165b8260a0c #4 DONE 0.0s #3 [internal] load metadata for docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a #3 sha256:095a08a5e87c0f08ca7fe08671a1bd8b85fa3840a51991c36774312eb5e5a0c7 #3 DONE 0.0s #9 [builder 1/9] FROM docker.io/library/eclipse-temurin:19@sha256:17e3d3b61ca4a7606490f596feb77f69980939fecacf91a1f13ea7b17147058f #9 sha256:e6598a80b4806e6f96c6639d605cf4c6eca4174e9d77b4bae443fffb1f2dcc2c #9 DONE 0.0s #11 [internal] load build context #11 sha256:14ea7f78c196e3fc2a5c4cafe5ae47a5c9c8babae792578ef62a0fb5a60e3aaf #11 transferring context: 840B done #11 DONE 0.0s #13 [builder 4/9] COPY gradle gradle #13 sha256:fdaeb43ef35a139a2ecd1cead76d33dfdf08ec928d056377bb6d0e4fa7ba5adc #13 CACHED #10 [builder 2/9] WORKDIR /app #10 sha256:c8376895fb2b0a3b0283686ae8db694693504e2b0084b35d3dad35064057b43b #10 CACHED #12 [builder 3/9] COPY [build.gradle, gradlew, ./] #12 sha256:83c710fc667545608bfc816075122245e3b257bcc1277345bb9b3282cfe1fefe #12 CACHED #14 [builder 5/9] RUN chmod +x gradlew #14 sha256:1a500f6f159b665ab926871cb66d0e20677a0a5f0b055537bcd781f6c6395e61 #14 CACHED #5 [without-grpc-health-probe-bin 1/5] FROM docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a 3a #5 sha256:3ec4fd675030cb4080679827d76012c9836cfa8e143f949ed6cf49b3d748d907 #5 resolve docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a done #5 sha256:bfba53c59e931f5d57c41ba291b90b1ab7e3034d4df5a30ba2ff56f63583beef 0B / 161B 0.2s #5 sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a 1.16kB / 1.16kB done #5 sha256:0fdafb034be2a2cb5ed98aa1b8116b05184efb869e1d7026bb395dc7f7be8f50 4.22kB / 4.22kB done #5 sha256:d8e5acd5897d762b9a83758d4ceae374df7b8b0367a48cc14b8a00e33998b3bf 0B / 12.02MB 0.2s #5 sha256:b231e5a3c8c25582f722e10270be76c230ecdb668b26782fe0b4cd29037021c6 0B / 49.14MB 0.2s #5 CANCELED #15 [builder 6/9] RUN ./gradlew downloadRepos #15 sha256:3997494df4830672a9af711404a7be3af2e05bafbda5e4154acd7d8958e0c28e #15 0.366 /bin/sh: 1: ./gradlew: not found #15 ERROR: executor failed running [/bin/sh -c ./gradlew downloadRepos]: exit code: 127 ------ > [builder 6/9] RUN ./gradlew downloadRepos: ------ executor failed running [/bin/sh -c ./gradlew downloadRepos]: exit code: 127 Building [productcatalogservice]... Target platforms: [linux/amd64] Build [productcatalogservice] was canceled Building [emailservice]... Target platforms: [linux/amd64] Build [emailservice] was canceled Building [recommendationservice]... Target platforms: [linux/amd64] Build [recommendationservice] was canceled build [adservice] failed: exit status 1. Docker build ran into internal error. Please retry. If this keeps happening, please open an issue.. ### Screenshots <!-- If applicable, add screenshots to help explain your problem --> ### Environment - OS: Windows 10 (Run on WSL 2) - Kubernetes distribution, version: minikube version: v1.28.0 - Any relevant tool version: Docker Desktop v4.15.0 ### Additional context <!-- Add any other context about the problem here --> ### Exposure <!-- Is the bug intermittent, persistent? Is it widespread, local? -->
1.0
Unable to build adservice - ### Describe the bug build [adservice] failed: exit status 1. Docker build ran into internal error. Please retry. ### To Reproduce <!-- Steps to reproduce the behavior: --> <!-- 1. Built image '...' --> <!-- 2. Ran command '....' --> <!-- 3. See error --> ### Logs Generating tags... - productcatalogservice -> productcatalogservice:v0.5.0-30-gb2ef53f - recommendationservice -> recommendationservice:v0.5.0-30-gb2ef53f - shippingservice -> shippingservice:v0.5.0-30-gb2ef53f - checkoutservice -> checkoutservice:v0.5.0-30-gb2ef53f - paymentservice -> paymentservice:v0.5.0-30-gb2ef53f - currencyservice -> currencyservice:v0.5.0-30-gb2ef53f - cartservice -> cartservice:v0.5.0-30-gb2ef53f - frontend -> frontend:v0.5.0-30-gb2ef53f - adservice -> adservice:v0.5.0-30-gb2ef53f - loadgenerator -> loadgenerator:v0.5.0-30-gb2ef53f Checking cache... - emailservice: Not found. Building - productcatalogservice: Not found. Building - recommendationservice: Not found. Building - shippingservice: Found Locally - checkoutservice: Found Locally - paymentservice: Found Locally - currencyservice: Found Locally - cartservice: Found Locally - frontend: Found Locally - adservice: Not found. Building - loadgenerator: Found Locally Starting build... Found [minikube] context, using local docker daemon. Building [adservice]... Target platforms: [linux/amd64] #1 [internal] load build definition from Dockerfile #1 sha256:4fd7c5b169637ade6dc46fa79f5ebe497faea8466f82e301c0fc5d1799fe1763 #1 transferring dockerfile: 38B done #1 DONE 0.0s #2 [internal] load .dockerignore #2 sha256:3a3376a110bfc911c611e7778255bed1a6fcc5f12331b0f88bd3f8a17922905e #2 transferring context: 2B done #2 DONE 0.0s #4 [internal] load metadata for docker.io/library/eclipse-temurin:19@sha256:17e3d3b61ca4a7606490f596feb77f69980939fecacf91a1f13ea7b17147058f #4 sha256:fa4367848242f6486a2e5650a6e73e7b59134bc9cb3dc1bc2d5c9165b8260a0c #4 DONE 0.0s #3 [internal] load metadata for docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a #3 sha256:095a08a5e87c0f08ca7fe08671a1bd8b85fa3840a51991c36774312eb5e5a0c7 #3 DONE 0.0s #9 [builder 1/9] FROM docker.io/library/eclipse-temurin:19@sha256:17e3d3b61ca4a7606490f596feb77f69980939fecacf91a1f13ea7b17147058f #9 sha256:e6598a80b4806e6f96c6639d605cf4c6eca4174e9d77b4bae443fffb1f2dcc2c #9 DONE 0.0s #11 [internal] load build context #11 sha256:14ea7f78c196e3fc2a5c4cafe5ae47a5c9c8babae792578ef62a0fb5a60e3aaf #11 transferring context: 840B done #11 DONE 0.0s #13 [builder 4/9] COPY gradle gradle #13 sha256:fdaeb43ef35a139a2ecd1cead76d33dfdf08ec928d056377bb6d0e4fa7ba5adc #13 CACHED #10 [builder 2/9] WORKDIR /app #10 sha256:c8376895fb2b0a3b0283686ae8db694693504e2b0084b35d3dad35064057b43b #10 CACHED #12 [builder 3/9] COPY [build.gradle, gradlew, ./] #12 sha256:83c710fc667545608bfc816075122245e3b257bcc1277345bb9b3282cfe1fefe #12 CACHED #14 [builder 5/9] RUN chmod +x gradlew #14 sha256:1a500f6f159b665ab926871cb66d0e20677a0a5f0b055537bcd781f6c6395e61 #14 CACHED #5 [without-grpc-health-probe-bin 1/5] FROM docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a 3a #5 sha256:3ec4fd675030cb4080679827d76012c9836cfa8e143f949ed6cf49b3d748d907 #5 resolve docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a done #5 sha256:bfba53c59e931f5d57c41ba291b90b1ab7e3034d4df5a30ba2ff56f63583beef 0B / 161B 0.2s #5 sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a 1.16kB / 1.16kB done #5 sha256:0fdafb034be2a2cb5ed98aa1b8116b05184efb869e1d7026bb395dc7f7be8f50 4.22kB / 4.22kB done #5 sha256:d8e5acd5897d762b9a83758d4ceae374df7b8b0367a48cc14b8a00e33998b3bf 0B / 12.02MB 0.2s #5 sha256:b231e5a3c8c25582f722e10270be76c230ecdb668b26782fe0b4cd29037021c6 0B / 49.14MB 0.2s #5 CANCELED #15 [builder 6/9] RUN ./gradlew downloadRepos #15 sha256:3997494df4830672a9af711404a7be3af2e05bafbda5e4154acd7d8958e0c28e #15 0.366 /bin/sh: 1: ./gradlew: not found #15 ERROR: executor failed running [/bin/sh -c ./gradlew downloadRepos]: exit code: 127 ------ > [builder 6/9] RUN ./gradlew downloadRepos: ------ executor failed running [/bin/sh -c ./gradlew downloadRepos]: exit code: 127 Building [productcatalogservice]... Target platforms: [linux/amd64] Build [productcatalogservice] was canceled Building [emailservice]... Target platforms: [linux/amd64] Build [emailservice] was canceled Building [recommendationservice]... Target platforms: [linux/amd64] Build [recommendationservice] was canceled build [adservice] failed: exit status 1. Docker build ran into internal error. Please retry. If this keeps happening, please open an issue.. ### Screenshots <!-- If applicable, add screenshots to help explain your problem --> ### Environment - OS: Windows 10 (Run on WSL 2) - Kubernetes distribution, version: minikube version: v1.28.0 - Any relevant tool version: Docker Desktop v4.15.0 ### Additional context <!-- Add any other context about the problem here --> ### Exposure <!-- Is the bug intermittent, persistent? Is it widespread, local? -->
priority
unable to build adservice describe the bug build failed exit status docker build ran into internal error please retry to reproduce logs generating tags productcatalogservice productcatalogservice recommendationservice recommendationservice shippingservice shippingservice checkoutservice checkoutservice paymentservice paymentservice currencyservice currencyservice cartservice cartservice frontend frontend adservice adservice loadgenerator loadgenerator checking cache emailservice not found building productcatalogservice not found building recommendationservice not found building shippingservice found locally checkoutservice found locally paymentservice found locally currencyservice found locally cartservice found locally frontend found locally adservice not found building loadgenerator found locally starting build found context using local docker daemon building target platforms load build definition from dockerfile transferring dockerfile done done load dockerignore transferring context done done load metadata for docker io library eclipse temurin done load metadata for docker io library eclipse temurin jre alpine done from docker io library eclipse temurin done load build context transferring context done done copy gradle gradle cached workdir app cached copy cached run chmod x gradlew cached from docker io library eclipse temurin jre alpine resolve docker io library eclipse temurin jre alpine done done done canceled run gradlew downloadrepos bin sh gradlew not found error executor failed running exit code run gradlew downloadrepos executor failed running exit code building target platforms build was canceled building target platforms build was canceled building target platforms build was canceled build failed exit status docker build ran into internal error please retry if this keeps happening please open an issue screenshots environment os windows run on wsl kubernetes distribution version minikube version any relevant tool version docker desktop additional context exposure
1
486,771
14,014,097,590
IssuesEvent
2020-10-29 11:23:32
workcraft/workcraft
https://api.github.com/repos/workcraft/workcraft
opened
Thread-unsafe use of JTextArea.append for displaying log messages
bug priority:critical status:confirmed tag:core
Method `JTextArea.append(String s)` is used for for printing the logs in the Output and Problems tabs. Logging may be invoked from threads, which causes problems in some JDK implementations (e.g. in OpenJDK build 1.8.0_265 for Ubuntu 20.04). This method used to be officially thread-safe in Java 6, but in Java 7 the thread-safeness requirement was revoked (see the second answer in this thread): https://stackoverflow.com/questions/4869027/jtextarea-thread-safe A solution is to schedule those JTextArea.append(...) calls in EDT.
1.0
Thread-unsafe use of JTextArea.append for displaying log messages - Method `JTextArea.append(String s)` is used for for printing the logs in the Output and Problems tabs. Logging may be invoked from threads, which causes problems in some JDK implementations (e.g. in OpenJDK build 1.8.0_265 for Ubuntu 20.04). This method used to be officially thread-safe in Java 6, but in Java 7 the thread-safeness requirement was revoked (see the second answer in this thread): https://stackoverflow.com/questions/4869027/jtextarea-thread-safe A solution is to schedule those JTextArea.append(...) calls in EDT.
priority
thread unsafe use of jtextarea append for displaying log messages method jtextarea append string s is used for for printing the logs in the output and problems tabs logging may be invoked from threads which causes problems in some jdk implementations e g in openjdk build for ubuntu this method used to be officially thread safe in java but in java the thread safeness requirement was revoked see the second answer in this thread a solution is to schedule those jtextarea append calls in edt
1
368,718
10,883,689,815
IssuesEvent
2019-11-18 05:57:09
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Master: outside "scroll bar" not available for interaction.
Fixed Medium Priority
As i see, "window size" controller just over it :dagger: ![изображение](https://user-images.githubusercontent.com/4980243/66248144-f74adf80-e72b-11e9-839d-6543dab90427.png)
1.0
Master: outside "scroll bar" not available for interaction. - As i see, "window size" controller just over it :dagger: ![изображение](https://user-images.githubusercontent.com/4980243/66248144-f74adf80-e72b-11e9-839d-6543dab90427.png)
priority
master outside scroll bar not available for interaction as i see window size controller just over it dagger
1
189,669
6,800,552,181
IssuesEvent
2017-11-02 14:19:54
feelpp/feelpp
https://api.github.com/repos/feelpp/feelpp
opened
Add support for first order moment computation, element-wise
module:discr priority:Urgent project: eye2brain project: mso4sc status:in development type:feature
element wise first moment of a function f computes \int_K f(x) dx if we divide by |K| then it is the mean value of f
1.0
Add support for first order moment computation, element-wise - element wise first moment of a function f computes \int_K f(x) dx if we divide by |K| then it is the mean value of f
priority
add support for first order moment computation element wise element wise first moment of a function f computes int k f x dx if we divide by k then it is the mean value of f
1
55,751
13,671,968,566
IssuesEvent
2020-09-29 07:50:19
zeek/spicy
https://api.github.com/repos/zeek/spicy
opened
Ship a custom "filesystem" implementation for better supporting older platforms
Build system
This might help with #451. A candidate is https://github.com/gulrak/filesystem/
1.0
Ship a custom "filesystem" implementation for better supporting older platforms - This might help with #451. A candidate is https://github.com/gulrak/filesystem/
non_priority
ship a custom filesystem implementation for better supporting older platforms this might help with a candidate is
0
772,973
27,141,742,630
IssuesEvent
2023-02-16 16:49:00
googleapis/python-bigquery-sqlalchemy
https://api.github.com/repos/googleapis/python-bigquery-sqlalchemy
closed
tests.sqlalchemy_dialect_compliance.test_dialect_compliance.HasTableTest_bigquery+bigquery: test_has_table_view failed
type: bug priority: p1 flakybot: issue api: bigquery
Note: #622 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 074321ddaa10001773e7e6044f4a0df1bb530331 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/82580611-18ba-4d10-8e25-a7bf752e1da1), [Sponge](http://sponge2/82580611-18ba-4d10-8e25-a7bf752e1da1) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 311, in from_call result: Optional[TResult] = func() ^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 255, in <lambda> lambda: ihook(item=item, **kwds), when=when, reraise=reraise ^^^^^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_hooks.py", line 265, in __call__ return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_manager.py", line 80, in _hookexec return self._inner_hookexec(hook_name, methods, kwargs, firstresult) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 60, in _multicall return outcome.get_result() ^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_result.py", line 60, in get_result raise ex[1].with_traceback(ex[2]) File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 39, in _multicall res = hook_impl.function(*args) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 175, in pytest_runtest_teardown item.session._setupstate.teardown_exact(item, nextitem) File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 419, in teardown_exact self._teardown_towards(needed_collectors) File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 434, in _teardown_towards raise exc File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 427, in _teardown_towards self._pop_and_teardown() File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 387, in _pop_and_teardown self._teardown_with_finalization(colitem) File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 408, in _teardown_with_finalization assert colitem in self.stack ^^^^^^^^^^^^^^^^^^^^^ AssertionError</pre></details>
1.0
tests.sqlalchemy_dialect_compliance.test_dialect_compliance.HasTableTest_bigquery+bigquery: test_has_table_view failed - Note: #622 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 074321ddaa10001773e7e6044f4a0df1bb530331 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/82580611-18ba-4d10-8e25-a7bf752e1da1), [Sponge](http://sponge2/82580611-18ba-4d10-8e25-a7bf752e1da1) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 311, in from_call result: Optional[TResult] = func() ^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 255, in <lambda> lambda: ihook(item=item, **kwds), when=when, reraise=reraise ^^^^^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_hooks.py", line 265, in __call__ return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_manager.py", line 80, in _hookexec return self._inner_hookexec(hook_name, methods, kwargs, firstresult) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 60, in _multicall return outcome.get_result() ^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_result.py", line 60, in get_result raise ex[1].with_traceback(ex[2]) File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 39, in _multicall res = hook_impl.function(*args) ^^^^^^^^^^^^^^^^^^^^^^^^^ File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 175, in pytest_runtest_teardown item.session._setupstate.teardown_exact(item, nextitem) File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 419, in teardown_exact self._teardown_towards(needed_collectors) File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 434, in _teardown_towards raise exc File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 427, in _teardown_towards self._pop_and_teardown() File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 387, in _pop_and_teardown self._teardown_with_finalization(colitem) File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 408, in _teardown_with_finalization assert colitem in self.stack ^^^^^^^^^^^^^^^^^^^^^ AssertionError</pre></details>
priority
tests sqlalchemy dialect compliance test dialect compliance hastabletest bigquery bigquery test has table view failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in from call result optional func file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in lambda ihook item item kwds when when reraise reraise file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy hooks py line in call return self hookexec self name self get hookimpls kwargs firstresult file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy manager py line in hookexec return self inner hookexec hook name methods kwargs firstresult file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy callers py line in multicall return outcome get result file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy result py line in get result raise ex with traceback ex file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy callers py line in multicall res hook impl function args file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in pytest runtest teardown item session setupstate teardown exact item nextitem file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown exact self teardown towards needed collectors file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown towards raise exc file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown towards self pop and teardown file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in pop and teardown self teardown with finalization colitem file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown with finalization assert colitem in self stack assertionerror
1
31,342
5,939,857,565
IssuesEvent
2017-05-25 07:08:55
c3js/c3
https://api.github.com/repos/c3js/c3
closed
Document use of d3.locale to choose localisation settings
documentation needed
How do I change the labels on the donut chart to display X,Y% (instead of american X.Y%)? If I try to adjust the labels in D3 locale.js I break the labels. Seems like C3 is calculating the percentages independently of the D3 settings. Maybe this is just not documented. Can we document localization and i18n ?
1.0
Document use of d3.locale to choose localisation settings - How do I change the labels on the donut chart to display X,Y% (instead of american X.Y%)? If I try to adjust the labels in D3 locale.js I break the labels. Seems like C3 is calculating the percentages independently of the D3 settings. Maybe this is just not documented. Can we document localization and i18n ?
non_priority
document use of locale to choose localisation settings how do i change the labels on the donut chart to display x y instead of american x y if i try to adjust the labels in locale js i break the labels seems like is calculating the percentages independently of the settings maybe this is just not documented can we document localization and
0
324,682
9,907,486,991
IssuesEvent
2019-06-27 15:55:06
python/mypy
https://api.github.com/repos/python/mypy
closed
Avoid redundant extra errors after "Need type annotation for variable"
new-semantic-analyzer priority-0-high topic-type-variables topic-usability
Consider this (oversimplified) example: ```python from typing import TypeVar, Optional T = TypeVar('T') def f(x: Optional[T] = None) -> T: ... x = f() # Need type annotation for 'x' y = x # Cannot determine type of 'x' def g() -> None: x = f() # Need type annotation for 'x' y = x # Cannot determine type of 'x' ``` The subsequent errors look redundant (and maybe even annoying). This is more important with the new semantic analyzer because the old one doesn't show the error if the problematic variable was defined at the function scope. Potentially, we can replace `<nothing>` with `Any` after we show the error. Also the code that detects failed inference can be improved. For example, `is_valid_inferred_type()` should be replaced with a proper type visitor.
1.0
Avoid redundant extra errors after "Need type annotation for variable" - Consider this (oversimplified) example: ```python from typing import TypeVar, Optional T = TypeVar('T') def f(x: Optional[T] = None) -> T: ... x = f() # Need type annotation for 'x' y = x # Cannot determine type of 'x' def g() -> None: x = f() # Need type annotation for 'x' y = x # Cannot determine type of 'x' ``` The subsequent errors look redundant (and maybe even annoying). This is more important with the new semantic analyzer because the old one doesn't show the error if the problematic variable was defined at the function scope. Potentially, we can replace `<nothing>` with `Any` after we show the error. Also the code that detects failed inference can be improved. For example, `is_valid_inferred_type()` should be replaced with a proper type visitor.
priority
avoid redundant extra errors after need type annotation for variable consider this oversimplified example python from typing import typevar optional t typevar t def f x optional none t x f need type annotation for x y x cannot determine type of x def g none x f need type annotation for x y x cannot determine type of x the subsequent errors look redundant and maybe even annoying this is more important with the new semantic analyzer because the old one doesn t show the error if the problematic variable was defined at the function scope potentially we can replace with any after we show the error also the code that detects failed inference can be improved for example is valid inferred type should be replaced with a proper type visitor
1
10,009
26,002,092,469
IssuesEvent
2022-12-20 16:05:00
OWASP/raider
https://api.github.com/repos/OWASP/raider
opened
Create a raider REPL
enhancement architecture
To preserve a session across runs, raider would need to have a REPL, and the user should have the ability to inspect and manipulate all individual elements.
1.0
Create a raider REPL - To preserve a session across runs, raider would need to have a REPL, and the user should have the ability to inspect and manipulate all individual elements.
non_priority
create a raider repl to preserve a session across runs raider would need to have a repl and the user should have the ability to inspect and manipulate all individual elements
0
200,518
7,008,645,903
IssuesEvent
2017-12-19 16:20:35
metasfresh/metasfresh
https://api.github.com/repos/metasfresh/metasfresh
opened
Material Tracking ID in Invoice Candidates Filter shall be search
priority:high type:enhancement
### Is this a bug or feature request? Feature Request ### What is the current behavior? Currently, the Field Material Tracking in Invoice Candidates is a List. This makes it unusable as Filter criteria. #### Which are the steps to reproduce? Open, try and see. ### What is the expected or desired behavior? Change the Fields to search in invoice Candidates.
1.0
Material Tracking ID in Invoice Candidates Filter shall be search - ### Is this a bug or feature request? Feature Request ### What is the current behavior? Currently, the Field Material Tracking in Invoice Candidates is a List. This makes it unusable as Filter criteria. #### Which are the steps to reproduce? Open, try and see. ### What is the expected or desired behavior? Change the Fields to search in invoice Candidates.
priority
material tracking id in invoice candidates filter shall be search is this a bug or feature request feature request what is the current behavior currently the field material tracking in invoice candidates is a list this makes it unusable as filter criteria which are the steps to reproduce open try and see what is the expected or desired behavior change the fields to search in invoice candidates
1
9,074
3,253,451,565
IssuesEvent
2015-10-19 19:09:11
rancher/os
https://api.github.com/repos/rancher/os
closed
ros config fails on fresh v0.4.0-rc11 install
area/documentation release/v0.4.0
On a freshly-installed system, running the following, straight from the documentation, results in nothing being changed, and `ros config get` returns nothing: ``` $ sudo ros config set network.dns.domain myexampledomain.com $ sudo ros config get network.dns.domain ```
1.0
ros config fails on fresh v0.4.0-rc11 install - On a freshly-installed system, running the following, straight from the documentation, results in nothing being changed, and `ros config get` returns nothing: ``` $ sudo ros config set network.dns.domain myexampledomain.com $ sudo ros config get network.dns.domain ```
non_priority
ros config fails on fresh install on a freshly installed system running the following straight from the documentation results in nothing being changed and ros config get returns nothing sudo ros config set network dns domain myexampledomain com sudo ros config get network dns domain
0
502,091
14,539,988,527
IssuesEvent
2020-12-15 12:41:27
guardicore/monkey
https://api.github.com/repos/guardicore/monkey
opened
Remove support for manually uploaded config.
Complexity: Low Feature removal Priority: Low
## Describe the bug There is an unmaintained(and probably unused) feature which allows users to provide configuration file path via cmd arguments. I think this feature is heritage of days long gone and should be removed. ## Changes to code: Steps to reproduce the behavior: 1. Remove config argument and related parsing 2. Remove example config file
1.0
Remove support for manually uploaded config. - ## Describe the bug There is an unmaintained(and probably unused) feature which allows users to provide configuration file path via cmd arguments. I think this feature is heritage of days long gone and should be removed. ## Changes to code: Steps to reproduce the behavior: 1. Remove config argument and related parsing 2. Remove example config file
priority
remove support for manually uploaded config describe the bug there is an unmaintained and probably unused feature which allows users to provide configuration file path via cmd arguments i think this feature is heritage of days long gone and should be removed changes to code steps to reproduce the behavior remove config argument and related parsing remove example config file
1
653,793
21,626,825,373
IssuesEvent
2022-05-05 04:06:30
panel-attack/panel-attack
https://api.github.com/repos/panel-attack/panel-attack
opened
Menu buttons should be customizable
enhancement Client-side low-ish priority
You should be able to customize the background color and border color in the theme you should also be able to optionally set a background image and a focused and selected image
1.0
Menu buttons should be customizable - You should be able to customize the background color and border color in the theme you should also be able to optionally set a background image and a focused and selected image
priority
menu buttons should be customizable you should be able to customize the background color and border color in the theme you should also be able to optionally set a background image and a focused and selected image
1
54,604
6,829,865,785
IssuesEvent
2017-11-09 02:51:43
swaywm/wlroots-rs
https://api.github.com/repos/swaywm/wlroots-rs
opened
DataDevice
Safe-Lib-Design
This is one of the globals that the compositor needs to create at startup. Either it needs to be provided in the constructor (like the I/O managers) or it needs to be passed in later (like the extension protocols). I'm up for either, leaning towards the latter since this is sort of an "extension" and it's _possible_ that a compositor may not want to implement this interface (and we should take a page out of wlroots' book and let the user decide). Either way, it needs to live as long as the compositor and destroyed before it disconnects. Need to also ensure only one global is registered at a time of this type. # Functions - [ ] `send_selection` Needs to only be possible if the DnD global has been made (so if an instance was made of the struct). Needs [SeatClient](https://replace-this-with-link-to-seat-issue) to be made first. - [ ] `set_selection` # DataOffer ## Events - [ ] `destroy` ## Functions - [ ] `dnd_actions` Return bitfield, please wrap in a `bitfield!` - [ ] preferred_dnd_action` - [ ] `in_ask` - [ ] `resource` (unsafe) - [ ] `data_source` (Return `DataSource`) # DataSource ## Events - [ ] `destroy` ## Functions - [ ] `resource` (unsafe) - [ ] `offer` (Return `DataOffer` - [ ] `seat_client` (Return `SeatClient`) - [ ] `mime_types` (Wrap in `Vec`) - [ ] `accepted` - [ ] `current_dnd_action` - [ ] `dnd_actions` Return bitfield, like in `DataOffer` - [ ] `compositor_actions` Return bitfield, like in `DataOffer` - [ ] `actions_set` ## Global Callbacks They should be callable, setting them may be unsafe but probably not. They are already initialized in global creation. Double check because this could be safe. - [ ] `accept` - [ ] `send` - [ ] `cancel` # Global Cleanup * Not possible (yet?), see [this issue](https://github.com/swaywm/wlroots/issues/399)
1.0
DataDevice - This is one of the globals that the compositor needs to create at startup. Either it needs to be provided in the constructor (like the I/O managers) or it needs to be passed in later (like the extension protocols). I'm up for either, leaning towards the latter since this is sort of an "extension" and it's _possible_ that a compositor may not want to implement this interface (and we should take a page out of wlroots' book and let the user decide). Either way, it needs to live as long as the compositor and destroyed before it disconnects. Need to also ensure only one global is registered at a time of this type. # Functions - [ ] `send_selection` Needs to only be possible if the DnD global has been made (so if an instance was made of the struct). Needs [SeatClient](https://replace-this-with-link-to-seat-issue) to be made first. - [ ] `set_selection` # DataOffer ## Events - [ ] `destroy` ## Functions - [ ] `dnd_actions` Return bitfield, please wrap in a `bitfield!` - [ ] preferred_dnd_action` - [ ] `in_ask` - [ ] `resource` (unsafe) - [ ] `data_source` (Return `DataSource`) # DataSource ## Events - [ ] `destroy` ## Functions - [ ] `resource` (unsafe) - [ ] `offer` (Return `DataOffer` - [ ] `seat_client` (Return `SeatClient`) - [ ] `mime_types` (Wrap in `Vec`) - [ ] `accepted` - [ ] `current_dnd_action` - [ ] `dnd_actions` Return bitfield, like in `DataOffer` - [ ] `compositor_actions` Return bitfield, like in `DataOffer` - [ ] `actions_set` ## Global Callbacks They should be callable, setting them may be unsafe but probably not. They are already initialized in global creation. Double check because this could be safe. - [ ] `accept` - [ ] `send` - [ ] `cancel` # Global Cleanup * Not possible (yet?), see [this issue](https://github.com/swaywm/wlroots/issues/399)
non_priority
datadevice this is one of the globals that the compositor needs to create at startup either it needs to be provided in the constructor like the i o managers or it needs to be passed in later like the extension protocols i m up for either leaning towards the latter since this is sort of an extension and it s possible that a compositor may not want to implement this interface and we should take a page out of wlroots book and let the user decide either way it needs to live as long as the compositor and destroyed before it disconnects need to also ensure only one global is registered at a time of this type functions send selection needs to only be possible if the dnd global has been made so if an instance was made of the struct needs to be made first set selection dataoffer events destroy functions dnd actions return bitfield please wrap in a bitfield preferred dnd action in ask resource unsafe data source return datasource datasource events destroy functions resource unsafe offer return dataoffer seat client return seatclient mime types wrap in vec accepted current dnd action dnd actions return bitfield like in dataoffer compositor actions return bitfield like in dataoffer actions set global callbacks they should be callable setting them may be unsafe but probably not they are already initialized in global creation double check because this could be safe accept send cancel global cleanup not possible yet see
0
414,340
12,102,385,517
IssuesEvent
2020-04-20 16:36:43
qutebrowser/qutebrowser
https://api.github.com/repos/qutebrowser/qutebrowser
closed
Avoid encoding parameter in search engine parameter
priority: 2 - low
Hello. I would like to have a search engine for Internet Archive, as such: ``` ia = http://web.archive.org/web/*/{} ``` but I cannot use it with the special characters present in an URL, because `qutebrowser` encodes the parameter. For instance, doing so: ``` :open ia https://github.com/The-Compiler/qutebrowser ``` will go to the following URL: http://web.archive.org/web/*/https%3A//github.com/The-Compiler/qutebrowser which does not work. So, it would be nice to have a way to avoid encoding the parameter. Thanks.
1.0
Avoid encoding parameter in search engine parameter - Hello. I would like to have a search engine for Internet Archive, as such: ``` ia = http://web.archive.org/web/*/{} ``` but I cannot use it with the special characters present in an URL, because `qutebrowser` encodes the parameter. For instance, doing so: ``` :open ia https://github.com/The-Compiler/qutebrowser ``` will go to the following URL: http://web.archive.org/web/*/https%3A//github.com/The-Compiler/qutebrowser which does not work. So, it would be nice to have a way to avoid encoding the parameter. Thanks.
priority
avoid encoding parameter in search engine parameter hello i would like to have a search engine for internet archive as such ia but i cannot use it with the special characters present in an url because qutebrowser encodes the parameter for instance doing so open ia will go to the following url which does not work so it would be nice to have a way to avoid encoding the parameter thanks
1
151,342
19,648,810,994
IssuesEvent
2022-01-10 02:36:14
turkdevops/angular
https://api.github.com/repos/turkdevops/angular
closed
WS-2019-0318 (High) detected in handlebars-4.4.3.tgz, handlebars-4.4.2.tgz - autoclosed
security vulnerability
## WS-2019-0318 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>handlebars-4.4.3.tgz</b>, <b>handlebars-4.4.2.tgz</b></p></summary> <p> <details><summary><b>handlebars-4.4.3.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.3.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.3.tgz</a></p> <p>Path to dependency file: angular/integration/cli-hello-world-lazy/package.json</p> <p>Path to vulnerable library: angular/integration/cli-hello-world-lazy/node_modules/handlebars/package.json,angular/integration/cli-hello-world-lazy-rollup/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - karma-coverage-istanbul-reporter-2.1.0.tgz (Root Library) - istanbul-api-2.1.6.tgz - istanbul-reports-2.2.6.tgz - :x: **handlebars-4.4.3.tgz** (Vulnerable Library) </details> <details><summary><b>handlebars-4.4.2.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.2.tgz</a></p> <p>Path to dependency file: angular/integration/cli-hello-world-ivy-i18n/package.json</p> <p>Path to vulnerable library: angular/integration/cli-hello-world-ivy-i18n/node_modules/handlebars/package.json,angular/integration/ivy-i18n/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - karma-coverage-istanbul-reporter-2.1.0.tgz (Root Library) - istanbul-api-2.1.6.tgz - istanbul-reports-2.2.6.tgz - :x: **handlebars-4.4.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/angular/commit/c6aca37f442da8c55a02d7c53ccc58100ab004f3">c6aca37f442da8c55a02d7c53ccc58100ab004f3</a></p> <p>Found in base branch: <b>labs/router</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In "showdownjs/showdown", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates. <p>Publish Date: 2019-10-20 <p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p> <p>Release Date: 2019-10-20</p> <p>Fix Resolution: handlebars - 4.4.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0318 (High) detected in handlebars-4.4.3.tgz, handlebars-4.4.2.tgz - autoclosed - ## WS-2019-0318 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>handlebars-4.4.3.tgz</b>, <b>handlebars-4.4.2.tgz</b></p></summary> <p> <details><summary><b>handlebars-4.4.3.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.3.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.3.tgz</a></p> <p>Path to dependency file: angular/integration/cli-hello-world-lazy/package.json</p> <p>Path to vulnerable library: angular/integration/cli-hello-world-lazy/node_modules/handlebars/package.json,angular/integration/cli-hello-world-lazy-rollup/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - karma-coverage-istanbul-reporter-2.1.0.tgz (Root Library) - istanbul-api-2.1.6.tgz - istanbul-reports-2.2.6.tgz - :x: **handlebars-4.4.3.tgz** (Vulnerable Library) </details> <details><summary><b>handlebars-4.4.2.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.2.tgz</a></p> <p>Path to dependency file: angular/integration/cli-hello-world-ivy-i18n/package.json</p> <p>Path to vulnerable library: angular/integration/cli-hello-world-ivy-i18n/node_modules/handlebars/package.json,angular/integration/ivy-i18n/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - karma-coverage-istanbul-reporter-2.1.0.tgz (Root Library) - istanbul-api-2.1.6.tgz - istanbul-reports-2.2.6.tgz - :x: **handlebars-4.4.2.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/turkdevops/angular/commit/c6aca37f442da8c55a02d7c53ccc58100ab004f3">c6aca37f442da8c55a02d7c53ccc58100ab004f3</a></p> <p>Found in base branch: <b>labs/router</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In "showdownjs/showdown", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates. <p>Publish Date: 2019-10-20 <p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p> <p>Release Date: 2019-10-20</p> <p>Fix Resolution: handlebars - 4.4.5</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
ws high detected in handlebars tgz handlebars tgz autoclosed ws high severity vulnerability vulnerable libraries handlebars tgz handlebars tgz handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file angular integration cli hello world lazy package json path to vulnerable library angular integration cli hello world lazy node modules handlebars package json angular integration cli hello world lazy rollup node modules handlebars package json dependency hierarchy karma coverage istanbul reporter tgz root library istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file angular integration cli hello world ivy package json path to vulnerable library angular integration cli hello world ivy node modules handlebars package json angular integration ivy node modules handlebars package json dependency hierarchy karma coverage istanbul reporter tgz root library istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href found in base branch labs router vulnerability details in showdownjs showdown versions prior to are vulnerable against regular expression denial of service redos once receiving specially crafted templates publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource
0
20,410
10,516,884,790
IssuesEvent
2019-09-28 20:58:19
nekofar/gulp-fontiran
https://api.github.com/repos/nekofar/gulp-fontiran
closed
WS-2018-0085 (High) detected in http-proxy-agent-1.0.0.tgz
security vulnerability
## WS-2018-0085 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-proxy-agent-1.0.0.tgz</b></p></summary> <p>An HTTP(s) proxy `http.Agent` implementation for HTTP</p> <p>Library home page: <a href="https://registry.npmjs.org/http-proxy-agent/-/http-proxy-agent-1.0.0.tgz">https://registry.npmjs.org/http-proxy-agent/-/http-proxy-agent-1.0.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/gulp-fontiran/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/gulp-fontiran/node_modules/http-proxy-agent/package.json</p> <p> Dependency Hierarchy: - typings-2.1.1.tgz (Root Library) - typings-core-2.3.3.tgz - popsicle-proxy-agent-3.0.0.tgz - :x: **http-proxy-agent-1.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nekofar/gulp-fontiran/commit/c05e05e14b3eedeed5a142729f2bcc89c44438f4">c05e05e14b3eedeed5a142729f2bcc89c44438f4</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of http-proxy-agent before 2.1.0 are vulnerable to denial of service and uninitialized memory leak when unsanitized options are passed to Buffer. <p>Publish Date: 2018-04-25 <p>URL: <a href=https://github.com/TooTallNate/node-http-proxy-agent/blob/2.0.0/index.js#L80>WS-2018-0085</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/607">https://nodesecurity.io/advisories/607</a></p> <p>Release Date: 2018-01-27</p> <p>Fix Resolution: 2.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2018-0085 (High) detected in http-proxy-agent-1.0.0.tgz - ## WS-2018-0085 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-proxy-agent-1.0.0.tgz</b></p></summary> <p>An HTTP(s) proxy `http.Agent` implementation for HTTP</p> <p>Library home page: <a href="https://registry.npmjs.org/http-proxy-agent/-/http-proxy-agent-1.0.0.tgz">https://registry.npmjs.org/http-proxy-agent/-/http-proxy-agent-1.0.0.tgz</a></p> <p>Path to dependency file: /tmp/ws-scm/gulp-fontiran/package.json</p> <p>Path to vulnerable library: /tmp/ws-scm/gulp-fontiran/node_modules/http-proxy-agent/package.json</p> <p> Dependency Hierarchy: - typings-2.1.1.tgz (Root Library) - typings-core-2.3.3.tgz - popsicle-proxy-agent-3.0.0.tgz - :x: **http-proxy-agent-1.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/nekofar/gulp-fontiran/commit/c05e05e14b3eedeed5a142729f2bcc89c44438f4">c05e05e14b3eedeed5a142729f2bcc89c44438f4</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions of http-proxy-agent before 2.1.0 are vulnerable to denial of service and uninitialized memory leak when unsanitized options are passed to Buffer. <p>Publish Date: 2018-04-25 <p>URL: <a href=https://github.com/TooTallNate/node-http-proxy-agent/blob/2.0.0/index.js#L80>WS-2018-0085</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nodesecurity.io/advisories/607">https://nodesecurity.io/advisories/607</a></p> <p>Release Date: 2018-01-27</p> <p>Fix Resolution: 2.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
ws high detected in http proxy agent tgz ws high severity vulnerability vulnerable library http proxy agent tgz an http s proxy http agent implementation for http library home page a href path to dependency file tmp ws scm gulp fontiran package json path to vulnerable library tmp ws scm gulp fontiran node modules http proxy agent package json dependency hierarchy typings tgz root library typings core tgz popsicle proxy agent tgz x http proxy agent tgz vulnerable library found in head commit a href vulnerability details versions of http proxy agent before are vulnerable to denial of service and uninitialized memory leak when unsanitized options are passed to buffer publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
41,824
21,966,995,446
IssuesEvent
2022-05-24 21:28:19
flutter/flutter
https://api.github.com/repos/flutter/flutter
opened
Material3 Slow down
created via performance template
## Details <!-- When i use material 3 is true and press to icon button many times in list all view get lagged and drop frame happened to 28hz https://user-images.githubusercontent.com/30121703/170135403-3c6b312c-77c0-44ef-a8df-69f242fe1043.mp4 **ِAndroid:** **Android 12:** **Tested on mi9t and samsung A22:** ``` ``` Doctor summary (to see all details, run flutter doctor -v): [✓] Flutter (Channel stable, 3.0.1, on macOS 11.5.2 20G95 darwin-x64, locale en-EG) [✓] Android toolchain - develop for Android devices (Android SDK version 32.1.0-rc1) [✓] Xcode - develop for iOS and macOS (Xcode 13.2.1) [✓] Chrome - develop for the web [✓] Android Studio (version 2021.2) [!] Android Studio ✗ Unable to find bundled Java version. [✓] IntelliJ IDEA Community Edition (version 2021.3) [✓] VS Code (version 1.67.0) [✓] VS Code (version 1.53.2) [✓] Connected device (5 available) [✓] HTTP Host Availability ``` ``` </details>
True
Material3 Slow down - ## Details <!-- When i use material 3 is true and press to icon button many times in list all view get lagged and drop frame happened to 28hz https://user-images.githubusercontent.com/30121703/170135403-3c6b312c-77c0-44ef-a8df-69f242fe1043.mp4 **ِAndroid:** **Android 12:** **Tested on mi9t and samsung A22:** ``` ``` Doctor summary (to see all details, run flutter doctor -v): [✓] Flutter (Channel stable, 3.0.1, on macOS 11.5.2 20G95 darwin-x64, locale en-EG) [✓] Android toolchain - develop for Android devices (Android SDK version 32.1.0-rc1) [✓] Xcode - develop for iOS and macOS (Xcode 13.2.1) [✓] Chrome - develop for the web [✓] Android Studio (version 2021.2) [!] Android Studio ✗ Unable to find bundled Java version. [✓] IntelliJ IDEA Community Edition (version 2021.3) [✓] VS Code (version 1.67.0) [✓] VS Code (version 1.53.2) [✓] Connected device (5 available) [✓] HTTP Host Availability ``` ``` </details>
non_priority
slow down details when i use material is true and press to icon button many times in list all view get lagged and drop frame happened to ِandroid android tested on and samsung doctor summary to see all details run flutter doctor v flutter channel stable on macos darwin locale en eg android toolchain develop for android devices android sdk version xcode develop for ios and macos xcode chrome develop for the web android studio version android studio ✗ unable to find bundled java version intellij idea community edition version vs code version vs code version connected device available http host availability
0
239,243
7,787,761,447
IssuesEvent
2018-06-07 00:16:29
ngeraci/ucr_archivestools
https://api.github.com/repos/ngeraci/ucr_archivestools
closed
put processed file on shared drive
enhancement priority:medium
Rather than process in-place in the Downloads folder, put processed file in standard location on shared drive. potential issues: * what behavior do we want for overwriting existing files? (prompt?) * handling WRCA filenames
1.0
put processed file on shared drive - Rather than process in-place in the Downloads folder, put processed file in standard location on shared drive. potential issues: * what behavior do we want for overwriting existing files? (prompt?) * handling WRCA filenames
priority
put processed file on shared drive rather than process in place in the downloads folder put processed file in standard location on shared drive potential issues what behavior do we want for overwriting existing files prompt handling wrca filenames
1
10,674
3,134,322,967
IssuesEvent
2015-09-10 09:23:49
Leadera/ecoman_repo
https://api.github.com/repos/Leadera/ecoman_repo
closed
Open project: system should display the project information page
enhancement Portal Project Management waiting for testing
Hi, When the user open a project, the system display the screen with the large picture of Ankara (I guess). Then the user does not know what to do... Please change it so that the system display direcly the related project page. See image below. ![geneva_test_7](https://cloud.githubusercontent.com/assets/9587966/7721214/e8e5dcda-fed5-11e4-8890-bea553433862.png)
1.0
Open project: system should display the project information page - Hi, When the user open a project, the system display the screen with the large picture of Ankara (I guess). Then the user does not know what to do... Please change it so that the system display direcly the related project page. See image below. ![geneva_test_7](https://cloud.githubusercontent.com/assets/9587966/7721214/e8e5dcda-fed5-11e4-8890-bea553433862.png)
non_priority
open project system should display the project information page hi when the user open a project the system display the screen with the large picture of ankara i guess then the user does not know what to do please change it so that the system display direcly the related project page see image below
0
365,701
10,790,775,970
IssuesEvent
2019-11-05 15:33:00
kuzzleio/kuzzle
https://api.github.com/repos/kuzzleio/kuzzle
closed
Use Map/Set in HotelClerk and DSL in critical sections of code
enhancement priority-normal
Profiling analysis performed on real-time activites, including (un)subscriptions, reveal that a large part of the time consumed comes from `Object.keys`. This is not surprising considering the number of uses of this command in our code. But the HotelClerk and the DSL modules both use large objects as repositories, and large amounts of Object.keys are performed on these structures. Considering the following benchmarks, an analysis should be conducted on our code to track down objects primarily used as key repositories (meaning that Object.keys is far more used than random accesses), and change them to Map (or Set for objects containing only a set of key-`true` pairs) ``` Object.keys (slow obj) + walk x 192,417 ops/sec ±0.52% (95 runs sampled) Object.keys (fast obj) + walk x 7,355,807 ops/sec ±0.48% (92 runs sampled) Map.entries (slow obj) + walk x 2,225,232 ops/sec ±0.63% (90 runs sampled) Set.values (slow obj) + walk x 2,378,888 ops/sec ±0.48% (93 runs sampled) random access (slow obj) x 23,120,065 ops/sec ±0.48% (88 runs sampled) random access (fast obj) x 17,391,844 ops/sec ±0.73% (90 runs sampled) random access (map.get) x 15,694,340 ops/sec ±0.17% (90 runs sampled) random access (set.has) x 15,837,436 ops/sec ±0.44% (94 runs sampled) ```
1.0
Use Map/Set in HotelClerk and DSL in critical sections of code - Profiling analysis performed on real-time activites, including (un)subscriptions, reveal that a large part of the time consumed comes from `Object.keys`. This is not surprising considering the number of uses of this command in our code. But the HotelClerk and the DSL modules both use large objects as repositories, and large amounts of Object.keys are performed on these structures. Considering the following benchmarks, an analysis should be conducted on our code to track down objects primarily used as key repositories (meaning that Object.keys is far more used than random accesses), and change them to Map (or Set for objects containing only a set of key-`true` pairs) ``` Object.keys (slow obj) + walk x 192,417 ops/sec ±0.52% (95 runs sampled) Object.keys (fast obj) + walk x 7,355,807 ops/sec ±0.48% (92 runs sampled) Map.entries (slow obj) + walk x 2,225,232 ops/sec ±0.63% (90 runs sampled) Set.values (slow obj) + walk x 2,378,888 ops/sec ±0.48% (93 runs sampled) random access (slow obj) x 23,120,065 ops/sec ±0.48% (88 runs sampled) random access (fast obj) x 17,391,844 ops/sec ±0.73% (90 runs sampled) random access (map.get) x 15,694,340 ops/sec ±0.17% (90 runs sampled) random access (set.has) x 15,837,436 ops/sec ±0.44% (94 runs sampled) ```
priority
use map set in hotelclerk and dsl in critical sections of code profiling analysis performed on real time activites including un subscriptions reveal that a large part of the time consumed comes from object keys this is not surprising considering the number of uses of this command in our code but the hotelclerk and the dsl modules both use large objects as repositories and large amounts of object keys are performed on these structures considering the following benchmarks an analysis should be conducted on our code to track down objects primarily used as key repositories meaning that object keys is far more used than random accesses and change them to map or set for objects containing only a set of key true pairs object keys slow obj walk x ops sec ± runs sampled object keys fast obj walk x ops sec ± runs sampled map entries slow obj walk x ops sec ± runs sampled set values slow obj walk x ops sec ± runs sampled random access slow obj x ops sec ± runs sampled random access fast obj x ops sec ± runs sampled random access map get x ops sec ± runs sampled random access set has x ops sec ± runs sampled
1
218,078
16,938,104,411
IssuesEvent
2021-06-27 00:43:10
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Brasil] Java Developer @ Instituto Recôncavo de Tecnologia
CLT Docker Especialista Java PJ Pleno Remoto Rest SQL Stale Testes Unitários
<!-- ================================================== Caso a vaga for remoto durante a pandemia informar no texto "Remoto durante o covid" ================================================== --> <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA BACK-END! Não faça distinção de gênero no título da vaga. Use: "Back-End Developer" ao invés de "Desenvolvedor Back-End" \o/ [Brasil] Java Developer @ Instituto Recôncavo de Tecnologia ================================================== --> <!-- ================================================== Caso a vaga for remoto durante a pandemia deixar a linha abaixo ================================================== --> > Vaga Remota ## Nossa empresa Somos um dos maiores institutos privados do Brasil, com um corpo técnico de mais de 120 profissionais, e estamos entre os mais inovadores no desenvolvimento de soluções criativas que dinamizam e integram as tarefas e os negócios dos nossos clientes e parceiros. ## Descrição da vaga Analista Desenvolvimento JAVA ## Local Remota, Salvador-BA ## Requisitos **Obrigatórios:** - Experiência em desenvolvimento Java; - Conhecimento Modelagem de sistemas; - Conhecimento em desenvolvimento com SpringBoot; - Conhecimento em desenvolvimento com Angula 7+; - Conhecimento em banco de dados MS SQL Server; - Conhecimento em desenvolvimento de API REST **Desejáveis:** - Conhecimentos em testes unitários; - Conhecimento Docker. ## Benefícios Benefícios para CLT - Plano de saúde - Plano Odontológico - Vale Refeição - Bônus Semestral ## Contratação CLT ou PJ a combinar ## Como se candidatar Por favor envie um email para vagas.inovatech@gmail.com com seu CV anexado - enviar no assunto: Vaga Analista Desenvolvimento JAVA ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 07 dias após cada processo. E-mail para contato em caso de não haver resposta: vagas.inovatech@gmail.com #### Alocação - Remoto #### Regime - CLT - PJ #### Nível - Júnior - Pleno - Sênior - Especialista
1.0
[Brasil] Java Developer @ Instituto Recôncavo de Tecnologia - <!-- ================================================== Caso a vaga for remoto durante a pandemia informar no texto "Remoto durante o covid" ================================================== --> <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA BACK-END! Não faça distinção de gênero no título da vaga. Use: "Back-End Developer" ao invés de "Desenvolvedor Back-End" \o/ [Brasil] Java Developer @ Instituto Recôncavo de Tecnologia ================================================== --> <!-- ================================================== Caso a vaga for remoto durante a pandemia deixar a linha abaixo ================================================== --> > Vaga Remota ## Nossa empresa Somos um dos maiores institutos privados do Brasil, com um corpo técnico de mais de 120 profissionais, e estamos entre os mais inovadores no desenvolvimento de soluções criativas que dinamizam e integram as tarefas e os negócios dos nossos clientes e parceiros. ## Descrição da vaga Analista Desenvolvimento JAVA ## Local Remota, Salvador-BA ## Requisitos **Obrigatórios:** - Experiência em desenvolvimento Java; - Conhecimento Modelagem de sistemas; - Conhecimento em desenvolvimento com SpringBoot; - Conhecimento em desenvolvimento com Angula 7+; - Conhecimento em banco de dados MS SQL Server; - Conhecimento em desenvolvimento de API REST **Desejáveis:** - Conhecimentos em testes unitários; - Conhecimento Docker. ## Benefícios Benefícios para CLT - Plano de saúde - Plano Odontológico - Vale Refeição - Bônus Semestral ## Contratação CLT ou PJ a combinar ## Como se candidatar Por favor envie um email para vagas.inovatech@gmail.com com seu CV anexado - enviar no assunto: Vaga Analista Desenvolvimento JAVA ## Tempo médio de feedbacks Costumamos enviar feedbacks em até 07 dias após cada processo. E-mail para contato em caso de não haver resposta: vagas.inovatech@gmail.com #### Alocação - Remoto #### Regime - CLT - PJ #### Nível - Júnior - Pleno - Sênior - Especialista
non_priority
java developer instituto recôncavo de tecnologia caso a vaga for remoto durante a pandemia informar no texto remoto durante o covid por favor só poste se a vaga for para back end não faça distinção de gênero no título da vaga use back end developer ao invés de desenvolvedor back end o java developer instituto recôncavo de tecnologia caso a vaga for remoto durante a pandemia deixar a linha abaixo vaga remota nossa empresa somos um dos maiores institutos privados do brasil com um corpo técnico de mais de profissionais e estamos entre os mais inovadores no desenvolvimento de soluções criativas que dinamizam e integram as tarefas e os negócios dos nossos clientes e parceiros descrição da vaga analista desenvolvimento java local remota salvador ba requisitos obrigatórios experiência em desenvolvimento java conhecimento modelagem de sistemas conhecimento em desenvolvimento com springboot conhecimento em desenvolvimento com angula conhecimento em banco de dados ms sql server conhecimento em desenvolvimento de api rest desejáveis conhecimentos em testes unitários conhecimento docker benefícios benefícios para clt plano de saúde plano odontológico vale refeição bônus semestral contratação clt ou pj a combinar como se candidatar por favor envie um email para vagas inovatech gmail com com seu cv anexado enviar no assunto vaga analista desenvolvimento java tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta vagas inovatech gmail com alocação remoto regime clt pj nível júnior pleno sênior especialista
0
818,337
30,684,305,281
IssuesEvent
2023-07-26 11:16:21
wso2/product-is
https://api.github.com/repos/wso2/product-is
closed
Cannot disable Google onetap feature from carbon console.
Priority/Highest bug
**Describe the issue:** Once a Google IDP connection is created and google one tap is enabled from the carbon console, Getting an error message as below once we try to disable Google one-tap ``` Error while updating Identity Provider information. IdentityProviderMgtServiceIdentityProviderManagementExceptionException ``` <img width="1512" alt="Screenshot 2023-06-23 at 12 30 33" src="https://github.com/wso2/product-is/assets/41188245/bd4a92f6-0a2d-4fcc-a8b2-6966c05a60ac"> https://github.com/wso2/product-is/assets/41188245/0c3f956b-4e3d-450a-9520-feee34527fcc **How to reproduce:** 1. Create a Google IDP and enable one tap. https://is.docs.wso2.com/en/latest/guides/identity-federation/google/#sign-in-with-google-one-tap 3. Create a service provider and add Google as an authentication step 4. Go to the Google IDP from the carbon console and try to uncheck the Google one-tap check box. 5. An error message will pop up. **Expected behavior:** <!-- A clear and concise description of what you expected to happen. --> **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: [e.g., IS 5.10.0, IS 5.9.0] - IS 6.2.0-alpha - OS: [e.g., Windows, Linux, Mac] - Mac - Database: [e.g., MySQL, H2] - H2 / MSSQL - Userstore: [e.g., LDAP, JDBC] ---
1.0
Cannot disable Google onetap feature from carbon console. - **Describe the issue:** Once a Google IDP connection is created and google one tap is enabled from the carbon console, Getting an error message as below once we try to disable Google one-tap ``` Error while updating Identity Provider information. IdentityProviderMgtServiceIdentityProviderManagementExceptionException ``` <img width="1512" alt="Screenshot 2023-06-23 at 12 30 33" src="https://github.com/wso2/product-is/assets/41188245/bd4a92f6-0a2d-4fcc-a8b2-6966c05a60ac"> https://github.com/wso2/product-is/assets/41188245/0c3f956b-4e3d-450a-9520-feee34527fcc **How to reproduce:** 1. Create a Google IDP and enable one tap. https://is.docs.wso2.com/en/latest/guides/identity-federation/google/#sign-in-with-google-one-tap 3. Create a service provider and add Google as an authentication step 4. Go to the Google IDP from the carbon console and try to uncheck the Google one-tap check box. 5. An error message will pop up. **Expected behavior:** <!-- A clear and concise description of what you expected to happen. --> **Environment information** (_Please complete the following information; remove any unnecessary fields_) **:** - Product Version: [e.g., IS 5.10.0, IS 5.9.0] - IS 6.2.0-alpha - OS: [e.g., Windows, Linux, Mac] - Mac - Database: [e.g., MySQL, H2] - H2 / MSSQL - Userstore: [e.g., LDAP, JDBC] ---
priority
cannot disable google onetap feature from carbon console describe the issue once a google idp connection is created and google one tap is enabled from the carbon console getting an error message as below once we try to disable google one tap error while updating identity provider information identityprovidermgtserviceidentityprovidermanagementexceptionexception img width alt screenshot at src how to reproduce create a google idp and enable one tap create a service provider and add google as an authentication step go to the google idp from the carbon console and try to uncheck the google one tap check box an error message will pop up expected behavior environment information please complete the following information remove any unnecessary fields product version is alpha os mac database mssql userstore
1
714,092
24,550,336,929
IssuesEvent
2022-10-12 12:07:36
Sequel-Ace/Sequel-Ace
https://api.github.com/repos/Sequel-Ace/Sequel-Ace
closed
Allowing a query history > 100
Feature Request stale Low priority
The limit in the preferences is 100, whereas I would like to essentially save all queries I've ever run, which may exceed that. I would like to be able to set some large number instead, e.g. 1,000,000
1.0
Allowing a query history > 100 - The limit in the preferences is 100, whereas I would like to essentially save all queries I've ever run, which may exceed that. I would like to be able to set some large number instead, e.g. 1,000,000
priority
allowing a query history the limit in the preferences is whereas i would like to essentially save all queries i ve ever run which may exceed that i would like to be able to set some large number instead e g
1
292,340
21,963,483,546
IssuesEvent
2022-05-24 17:49:07
gravitational/teleport
https://api.github.com/repos/gravitational/teleport
closed
Reorganize Getting Started section.
documentation time-to-value
## Details The current **Getting Started** section houses two different types of doc. The Linux server guide is focused on installation and initial steps, while the docker-compose is more of a tutorial-style introduction to Teleport in a preconfigured environment. Move docker-compose to a separate tutorial (or Teleport lab) section and keep the Getting Started section focused on installation and initial steps. The final section would be: **Getting started** - [Linux Server](https://goteleport.com/docs/getting-started/linux-server/) (current one) - [Kubernetes Cluster](https://goteleport.com/docs/kubernetes-access/getting-started/cluster/) (move from Kubernetes Access as this is also related to installation and getting started on Kubernetes) **Tutorial (or Teleport lab)** - [docker-compose](https://goteleport.com/docs/getting-started/docker-compose/) - [minikube based teleport lab](https://github.com/gravitational/teleport/issues/9359) ### Category - Improve Existing
1.0
Reorganize Getting Started section. - ## Details The current **Getting Started** section houses two different types of doc. The Linux server guide is focused on installation and initial steps, while the docker-compose is more of a tutorial-style introduction to Teleport in a preconfigured environment. Move docker-compose to a separate tutorial (or Teleport lab) section and keep the Getting Started section focused on installation and initial steps. The final section would be: **Getting started** - [Linux Server](https://goteleport.com/docs/getting-started/linux-server/) (current one) - [Kubernetes Cluster](https://goteleport.com/docs/kubernetes-access/getting-started/cluster/) (move from Kubernetes Access as this is also related to installation and getting started on Kubernetes) **Tutorial (or Teleport lab)** - [docker-compose](https://goteleport.com/docs/getting-started/docker-compose/) - [minikube based teleport lab](https://github.com/gravitational/teleport/issues/9359) ### Category - Improve Existing
non_priority
reorganize getting started section details the current getting started section houses two different types of doc the linux server guide is focused on installation and initial steps while the docker compose is more of a tutorial style introduction to teleport in a preconfigured environment move docker compose to a separate tutorial or teleport lab section and keep the getting started section focused on installation and initial steps the final section would be getting started current one move from kubernetes access as this is also related to installation and getting started on kubernetes tutorial or teleport lab category improve existing
0
735,968
25,450,924,254
IssuesEvent
2022-11-24 10:22:43
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.msn.com - see bug description
browser-firefox priority-critical engine-gecko
<!-- @browser: Firefox 107.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:107.0) Gecko/20100101 Firefox/107.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/114556 --> **URL**: https://www.msn.com/en-us/news/politics/goodbye-cannon-experts-say-judge-s-controversial-pro-trump-order-doomed-after-hostile-hearing/ar-AA14sCt0?cvid=e8af6392f2a4436e812f45019ae71187 **Browser / Version**: Firefox 107.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: Ads overlapping page content **Steps to Reproduce**: After recent update to version 107.0 (64-bit), ads now overlap content making content unreadable. The same pages render without problems when using Chrome Version 107.0.5304.107 (Official Build) (64-bit). <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/11/0bf4bd5f-f9e4-4e90-8e68-210928520b70.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.msn.com - see bug description - <!-- @browser: Firefox 107.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:107.0) Gecko/20100101 Firefox/107.0 --> <!-- @reported_with: unknown --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/114556 --> **URL**: https://www.msn.com/en-us/news/politics/goodbye-cannon-experts-say-judge-s-controversial-pro-trump-order-doomed-after-hostile-hearing/ar-AA14sCt0?cvid=e8af6392f2a4436e812f45019ae71187 **Browser / Version**: Firefox 107.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: Ads overlapping page content **Steps to Reproduce**: After recent update to version 107.0 (64-bit), ads now overlap content making content unreadable. The same pages render without problems when using Chrome Version 107.0.5304.107 (Official Build) (64-bit). <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/11/0bf4bd5f-f9e4-4e90-8e68-210928520b70.jpg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
priority
see bug description url browser version firefox operating system windows tested another browser yes chrome problem type something else description ads overlapping page content steps to reproduce after recent update to version bit ads now overlap content making content unreadable the same pages render without problems when using chrome version official build bit view the screenshot img alt screenshot src browser configuration none from with ❤️
1
180,479
21,625,746,060
IssuesEvent
2022-05-05 01:43:40
michaeldotson/raevin
https://api.github.com/repos/michaeldotson/raevin
opened
CVE-2020-15169 (Medium) detected in actionview-5.2.3.gem
security vulnerability
## CVE-2020-15169 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionview-5.2.3.gem</b></p></summary> <p>Simple, battle-tested conventions and helpers for building web pages.</p> <p>Library home page: <a href="https://rubygems.org/gems/actionview-5.2.3.gem">https://rubygems.org/gems/actionview-5.2.3.gem</a></p> <p>Path to dependency file: /raevin/Gemfile.lock</p> <p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/actionview-5.2.3.gem</p> <p> Dependency Hierarchy: - sass-rails-5.0.7.gem (Root Library) - sprockets-rails-3.2.1.gem - actionpack-5.2.3.gem - :x: **actionview-5.2.3.gem** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Action View before versions 5.2.4.4 and 6.0.3.3 there is a potential Cross-Site Scripting (XSS) vulnerability in Action View's translation helpers. Views that allow the user to control the default (not found) value of the `t` and `translate` helpers could be susceptible to XSS attacks. When an HTML-unsafe string is passed as the default for a missing translation key named html or ending in _html, the default string is incorrectly marked as HTML-safe and not escaped. This is patched in versions 6.0.3.3 and 5.2.4.4. A workaround without upgrading is proposed in the source advisory. <p>Publish Date: 2020-09-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15169>CVE-2020-15169</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://rubygems.org/gems/actionview/versions/6.0.3.3">https://rubygems.org/gems/actionview/versions/6.0.3.3</a></p> <p>Release Date: 2020-09-11</p> <p>Fix Resolution: 6.0.3.3, 5.2.4.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15169 (Medium) detected in actionview-5.2.3.gem - ## CVE-2020-15169 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionview-5.2.3.gem</b></p></summary> <p>Simple, battle-tested conventions and helpers for building web pages.</p> <p>Library home page: <a href="https://rubygems.org/gems/actionview-5.2.3.gem">https://rubygems.org/gems/actionview-5.2.3.gem</a></p> <p>Path to dependency file: /raevin/Gemfile.lock</p> <p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/actionview-5.2.3.gem</p> <p> Dependency Hierarchy: - sass-rails-5.0.7.gem (Root Library) - sprockets-rails-3.2.1.gem - actionpack-5.2.3.gem - :x: **actionview-5.2.3.gem** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Action View before versions 5.2.4.4 and 6.0.3.3 there is a potential Cross-Site Scripting (XSS) vulnerability in Action View's translation helpers. Views that allow the user to control the default (not found) value of the `t` and `translate` helpers could be susceptible to XSS attacks. When an HTML-unsafe string is passed as the default for a missing translation key named html or ending in _html, the default string is incorrectly marked as HTML-safe and not escaped. This is patched in versions 6.0.3.3 and 5.2.4.4. A workaround without upgrading is proposed in the source advisory. <p>Publish Date: 2020-09-11 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15169>CVE-2020-15169</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://rubygems.org/gems/actionview/versions/6.0.3.3">https://rubygems.org/gems/actionview/versions/6.0.3.3</a></p> <p>Release Date: 2020-09-11</p> <p>Fix Resolution: 6.0.3.3, 5.2.4.4</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in actionview gem cve medium severity vulnerability vulnerable library actionview gem simple battle tested conventions and helpers for building web pages library home page a href path to dependency file raevin gemfile lock path to vulnerable library var lib gems cache actionview gem dependency hierarchy sass rails gem root library sprockets rails gem actionpack gem x actionview gem vulnerable library vulnerability details in action view before versions and there is a potential cross site scripting xss vulnerability in action view s translation helpers views that allow the user to control the default not found value of the t and translate helpers could be susceptible to xss attacks when an html unsafe string is passed as the default for a missing translation key named html or ending in html the default string is incorrectly marked as html safe and not escaped this is patched in versions and a workaround without upgrading is proposed in the source advisory publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
120,199
4,786,738,875
IssuesEvent
2016-10-29 16:08:00
NREL/OpenStudio-BEopt
https://api.github.com/repos/NREL/OpenStudio-BEopt
closed
Propane/oil heating coils
priority high
EnergyPlus only allows electricity and natural gas for HVAC heating coils. We'd like to accommodate other fuel types in E+ so as to avoid using EMS.
1.0
Propane/oil heating coils - EnergyPlus only allows electricity and natural gas for HVAC heating coils. We'd like to accommodate other fuel types in E+ so as to avoid using EMS.
priority
propane oil heating coils energyplus only allows electricity and natural gas for hvac heating coils we d like to accommodate other fuel types in e so as to avoid using ems
1
797,087
28,137,458,287
IssuesEvent
2023-04-01 14:54:51
SierraBay/SierraBay12
https://api.github.com/repos/SierraBay/SierraBay12
closed
Баг: у объекта flash shell отсутствует спрайт
:bug: Баг 🐂 Priority: Low
**Ckey**: `Atufancheg` **Шаги:** 1. распечатать в атолате flash shell 2. осмотреться 3. попытаться взять в руку через ПКМ 4. ткнуть по автолату **Реальное поведение:** в руку берётся(ПКМ), но слот отображается как пустой. при клике по атолату отображается "You dump the flash shell into the autolathe." в чате и автолат проигрывает анимацию поглощения **Ожидаемое поведение:** спрайт отображается и на автолате и в руке <hr> *Репорт сгенерирован автоматически* *Автор: `Atufancheg#7790` / `260139717121540107`*
1.0
Баг: у объекта flash shell отсутствует спрайт - **Ckey**: `Atufancheg` **Шаги:** 1. распечатать в атолате flash shell 2. осмотреться 3. попытаться взять в руку через ПКМ 4. ткнуть по автолату **Реальное поведение:** в руку берётся(ПКМ), но слот отображается как пустой. при клике по атолату отображается "You dump the flash shell into the autolathe." в чате и автолат проигрывает анимацию поглощения **Ожидаемое поведение:** спрайт отображается и на автолате и в руке <hr> *Репорт сгенерирован автоматически* *Автор: `Atufancheg#7790` / `260139717121540107`*
priority
баг у объекта flash shell отсутствует спрайт ckey atufancheg шаги распечатать в атолате flash shell осмотреться попытаться взять в руку через пкм ткнуть по автолату реальное поведение в руку берётся пкм но слот отображается как пустой при клике по атолату отображается you dump the flash shell into the autolathe в чате и автолат проигрывает анимацию поглощения ожидаемое поведение спрайт отображается и на автолате и в руке репорт сгенерирован автоматически автор atufancheg
1
225,492
24,848,197,144
IssuesEvent
2022-10-26 17:37:49
opensearch-project/performance-analyzer
https://api.github.com/repos/opensearch-project/performance-analyzer
closed
CVE-2022-42889 (High) detected in commons-text-1.9.jar - autoclosed
security vulnerability
## CVE-2022-42889 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-text-1.9.jar</b></p></summary> <p>Apache Commons Text is a library focused on algorithms working on strings.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-text">https://commons.apache.org/proper/commons-text</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-text/1.9/ba6ac8c2807490944a0a27f6f8e68fb5ed2e80e2/commons-text-1.9.jar</p> <p> Dependency Hierarchy: - spotbugs-4.5.0.jar (Root Library) - :x: **commons-text-1.9.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Commons Text performs variable interpolation, allowing properties to be dynamically evaluated and expanded. The standard format for interpolation is "${prefix:name}", where "prefix" is used to locate an instance of org.apache.commons.text.lookup.StringLookup that performs the interpolation. Starting with version 1.5 and continuing through 1.9, the set of default Lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers. These lookups are: - "script" - execute expressions using the JVM script execution engine (javax.script) - "dns" - resolve dns records - "url" - load values from urls, including from remote servers Applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used. Users are recommended to upgrade to Apache Commons Text 1.10.0, which disables the problematic interpolators by default. <p>Publish Date: Oct 13, 2022 1:15:00 PM <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-42889>CVE-2022-42889</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.openwall.com/lists/oss-security/2022/10/13/4">https://www.openwall.com/lists/oss-security/2022/10/13/4</a></p> <p>Release Date: Oct 13, 2022 1:15:00 PM</p> <p>Fix Resolution (org.apache.commons:commons-text): 1.10.0</p> <p>Direct dependency fix Resolution (com.github.spotbugs:spotbugs): 4.5.2</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2022-42889 (High) detected in commons-text-1.9.jar - autoclosed - ## CVE-2022-42889 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-text-1.9.jar</b></p></summary> <p>Apache Commons Text is a library focused on algorithms working on strings.</p> <p>Library home page: <a href="https://commons.apache.org/proper/commons-text">https://commons.apache.org/proper/commons-text</a></p> <p>Path to dependency file: /build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-text/1.9/ba6ac8c2807490944a0a27f6f8e68fb5ed2e80e2/commons-text-1.9.jar</p> <p> Dependency Hierarchy: - spotbugs-4.5.0.jar (Root Library) - :x: **commons-text-1.9.jar** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Commons Text performs variable interpolation, allowing properties to be dynamically evaluated and expanded. The standard format for interpolation is "${prefix:name}", where "prefix" is used to locate an instance of org.apache.commons.text.lookup.StringLookup that performs the interpolation. Starting with version 1.5 and continuing through 1.9, the set of default Lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers. These lookups are: - "script" - execute expressions using the JVM script execution engine (javax.script) - "dns" - resolve dns records - "url" - load values from urls, including from remote servers Applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used. Users are recommended to upgrade to Apache Commons Text 1.10.0, which disables the problematic interpolators by default. <p>Publish Date: Oct 13, 2022 1:15:00 PM <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-42889>CVE-2022-42889</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.openwall.com/lists/oss-security/2022/10/13/4">https://www.openwall.com/lists/oss-security/2022/10/13/4</a></p> <p>Release Date: Oct 13, 2022 1:15:00 PM</p> <p>Fix Resolution (org.apache.commons:commons-text): 1.10.0</p> <p>Direct dependency fix Resolution (com.github.spotbugs:spotbugs): 4.5.2</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_priority
cve high detected in commons text jar autoclosed cve high severity vulnerability vulnerable library commons text jar apache commons text is a library focused on algorithms working on strings library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files org apache commons commons text commons text jar dependency hierarchy spotbugs jar root library x commons text jar vulnerable library found in base branch main vulnerability details apache commons text performs variable interpolation allowing properties to be dynamically evaluated and expanded the standard format for interpolation is prefix name where prefix is used to locate an instance of org apache commons text lookup stringlookup that performs the interpolation starting with version and continuing through the set of default lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers these lookups are script execute expressions using the jvm script execution engine javax script dns resolve dns records url load values from urls including from remote servers applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used users are recommended to upgrade to apache commons text which disables the problematic interpolators by default publish date oct pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date oct pm fix resolution org apache commons commons text direct dependency fix resolution com github spotbugs spotbugs check this box to open an automated fix pr
0
157,301
5,997,124,168
IssuesEvent
2017-06-03 20:42:36
CAGoodman/CareWheelsCorp
https://api.github.com/repos/CAGoodman/CareWheelsCorp
opened
Need to update the email address in the lost credentials link
bug High Priority
We will use support@carewheels.com
1.0
Need to update the email address in the lost credentials link - We will use support@carewheels.com
priority
need to update the email address in the lost credentials link we will use support carewheels com
1
139,541
5,378,196,504
IssuesEvent
2017-02-23 14:23:38
juju/docs
https://api.github.com/repos/juju/docs
closed
juju gui command changes
2.1 high priority
The juju gui command has changed. no-browser is now the default behaviour. It also prints the login credential by default. There is a new --hide-credential option not to show the credential. The --no-browser option is supported but deprecated (it is effectively a no-op). To bring up a browser, use the --browser option. Snippet from juju gui --help Details: Print the Juju GUI URL and show admin credential to use to log into it: juju gui Print the Juju GUI URL only: juju gui --hide-credential Open the Juju GUI in the default browser and show admin credential to use to log into it: juju gui --browser Open the Juju GUI in the default browser without printing the login credential: juju gui --hide-credential --browser
1.0
juju gui command changes - The juju gui command has changed. no-browser is now the default behaviour. It also prints the login credential by default. There is a new --hide-credential option not to show the credential. The --no-browser option is supported but deprecated (it is effectively a no-op). To bring up a browser, use the --browser option. Snippet from juju gui --help Details: Print the Juju GUI URL and show admin credential to use to log into it: juju gui Print the Juju GUI URL only: juju gui --hide-credential Open the Juju GUI in the default browser and show admin credential to use to log into it: juju gui --browser Open the Juju GUI in the default browser without printing the login credential: juju gui --hide-credential --browser
priority
juju gui command changes the juju gui command has changed no browser is now the default behaviour it also prints the login credential by default there is a new hide credential option not to show the credential the no browser option is supported but deprecated it is effectively a no op to bring up a browser use the browser option snippet from juju gui help details print the juju gui url and show admin credential to use to log into it juju gui print the juju gui url only juju gui hide credential open the juju gui in the default browser and show admin credential to use to log into it juju gui browser open the juju gui in the default browser without printing the login credential juju gui hide credential browser
1
218,344
16,985,080,092
IssuesEvent
2021-06-30 13:34:49
mysteriumnetwork/node
https://api.github.com/repos/mysteriumnetwork/node
closed
Remove proposal listening on NATS
component:node testnet3
Currently, all consumers listen for proposals on nats. While it's a nice fallback to have, it creates a rather large load for limited gains.
1.0
Remove proposal listening on NATS - Currently, all consumers listen for proposals on nats. While it's a nice fallback to have, it creates a rather large load for limited gains.
non_priority
remove proposal listening on nats currently all consumers listen for proposals on nats while it s a nice fallback to have it creates a rather large load for limited gains
0
284,766
24,622,911,980
IssuesEvent
2022-10-16 06:04:23
nodejs/node
https://api.github.com/repos/nodejs/node
reopened
test_runner: update output TAP format to follow TAP 14 specs
test_runner
### What steps will reproduce the bug? - create an `index.js` file with a few test cases: ```javascript const test = require('node:test'); const assert = require('node:assert'); test('top-level test 1', async (t) => { await t.test('level 1.1', () => {}); }); test('top-level test 2', () => {}); ``` - run the test: `node index.js` (note: `node --test index.js` gives an incorrect result) ### How often does it reproduce? Is there a required condition? _No response_ ### What is the expected behavior? Even though the current output is syntactically valid TAP output. The expected output would be a TAP format that [follows specs 14.](http://testanything.org/tap-version-14-specification.html) ``` TAP version 14 # Subtest: top-level test 1 ok 1 - sub test level 1.1 1..1 ok 1 - top-level test 1 ok 2 - top-level test 2 1..2 ``` Note the version header `TAP version 14` (required by TAP14). ### What do you see instead? We currently output the following format (removed diagnostics for readability): ``` TAP version 13 # Subtest: top-level test 1 # Subtest: sub-test level 1.1 ok 1 - sub-test level 1.1 1..1 ok 1 - top-level test 1 1..1 ``` Top-level tests are incorrectly output as subtests: `# Subtest: top-level test 1` and `# Subtest: top-level test 2` ### Additional information As a reference, using node-tap (v16.3.0) gives the following output: ``` TAP version 13 # Subtest: top-level test 1 ok 1 - sub test level 1.1 1..1 ok 1 - top-level test 1 ok 2 - top-level test 2 1..2 ``` Here is the diff: ```diff TAP version 13 # Subtest: top-level test 1 - # Subtest: sub test level 1.1 ok 1 - sub test level 1.1 1..1 ok 1 - top-level test 1 + ok 2 - top-level test 2 -1..1 +1..2 ``` Related: https://github.com/nodejs/node/pull/43417 https://github.com/nodejs/node/pull/43525
1.0
test_runner: update output TAP format to follow TAP 14 specs - ### What steps will reproduce the bug? - create an `index.js` file with a few test cases: ```javascript const test = require('node:test'); const assert = require('node:assert'); test('top-level test 1', async (t) => { await t.test('level 1.1', () => {}); }); test('top-level test 2', () => {}); ``` - run the test: `node index.js` (note: `node --test index.js` gives an incorrect result) ### How often does it reproduce? Is there a required condition? _No response_ ### What is the expected behavior? Even though the current output is syntactically valid TAP output. The expected output would be a TAP format that [follows specs 14.](http://testanything.org/tap-version-14-specification.html) ``` TAP version 14 # Subtest: top-level test 1 ok 1 - sub test level 1.1 1..1 ok 1 - top-level test 1 ok 2 - top-level test 2 1..2 ``` Note the version header `TAP version 14` (required by TAP14). ### What do you see instead? We currently output the following format (removed diagnostics for readability): ``` TAP version 13 # Subtest: top-level test 1 # Subtest: sub-test level 1.1 ok 1 - sub-test level 1.1 1..1 ok 1 - top-level test 1 1..1 ``` Top-level tests are incorrectly output as subtests: `# Subtest: top-level test 1` and `# Subtest: top-level test 2` ### Additional information As a reference, using node-tap (v16.3.0) gives the following output: ``` TAP version 13 # Subtest: top-level test 1 ok 1 - sub test level 1.1 1..1 ok 1 - top-level test 1 ok 2 - top-level test 2 1..2 ``` Here is the diff: ```diff TAP version 13 # Subtest: top-level test 1 - # Subtest: sub test level 1.1 ok 1 - sub test level 1.1 1..1 ok 1 - top-level test 1 + ok 2 - top-level test 2 -1..1 +1..2 ``` Related: https://github.com/nodejs/node/pull/43417 https://github.com/nodejs/node/pull/43525
non_priority
test runner update output tap format to follow tap specs what steps will reproduce the bug create an index js file with a few test cases javascript const test require node test const assert require node assert test top level test async t await t test level test top level test run the test node index js note node test index js gives an incorrect result how often does it reproduce is there a required condition no response what is the expected behavior even though the current output is syntactically valid tap output the expected output would be a tap format that tap version subtest top level test ok sub test level ok top level test ok top level test note the version header tap version required by what do you see instead we currently output the following format removed diagnostics for readability tap version subtest top level test subtest sub test level ok sub test level ok top level test top level tests are incorrectly output as subtests subtest top level test and subtest top level test additional information as a reference using node tap gives the following output tap version subtest top level test ok sub test level ok top level test ok top level test here is the diff diff tap version subtest top level test subtest sub test level ok sub test level ok top level test ok top level test related
0
332,000
10,083,079,215
IssuesEvent
2019-07-25 12:54:04
ME-ICA/tedana
https://api.github.com/repos/ME-ICA/tedana
opened
Topics for Hackathon 2019
high-priority
<!-- This is a suggested issue template for tedana. If there is other information that would be helpful to include, please do not hesitate to add it! Before submitting, please check to make sure that the issue is not already addressed; if there is a related issue, then please cross-reference it by #. If this is a usage question, please check out NeuroStars here: https://neurostars.org/ and tag your topic with "multi-echo" --> <!-- Summarize the issue in 1-2 sentences, linking other issues if they are relevant Note: simply typing # will prompt you for open issues to select from --> ### The Story So Far We are planning a hackathon November 6-8 in Bethesda, Maryland. We will use this issue to discuss what possible items could be added to its agenda. This will be kept up to date below. Items with ✅ are approved by consensus and struck-out items were proposed but agreed not to be done. ### Proposed Items - Lecture/group discussion on math steps - Decision Tree rewrite <!-- If needed, add additional detail for: 1. Recreating a bug/problem 2. Any additional context necessary to understand the issue --> <!-- If desired, add suggested next steps. If you foresee them in a particular order or priority, please use numbering --> <!-- Thank you for submitting your issue! If you do not receive a response within a calendar week, please post a comment on this issue to catch our attention. Some issues may not be resolved right away due to the volunteer nature of the project; thank you for your patience! -->
1.0
Topics for Hackathon 2019 - <!-- This is a suggested issue template for tedana. If there is other information that would be helpful to include, please do not hesitate to add it! Before submitting, please check to make sure that the issue is not already addressed; if there is a related issue, then please cross-reference it by #. If this is a usage question, please check out NeuroStars here: https://neurostars.org/ and tag your topic with "multi-echo" --> <!-- Summarize the issue in 1-2 sentences, linking other issues if they are relevant Note: simply typing # will prompt you for open issues to select from --> ### The Story So Far We are planning a hackathon November 6-8 in Bethesda, Maryland. We will use this issue to discuss what possible items could be added to its agenda. This will be kept up to date below. Items with ✅ are approved by consensus and struck-out items were proposed but agreed not to be done. ### Proposed Items - Lecture/group discussion on math steps - Decision Tree rewrite <!-- If needed, add additional detail for: 1. Recreating a bug/problem 2. Any additional context necessary to understand the issue --> <!-- If desired, add suggested next steps. If you foresee them in a particular order or priority, please use numbering --> <!-- Thank you for submitting your issue! If you do not receive a response within a calendar week, please post a comment on this issue to catch our attention. Some issues may not be resolved right away due to the volunteer nature of the project; thank you for your patience! -->
priority
topics for hackathon this is a suggested issue template for tedana if there is other information that would be helpful to include please do not hesitate to add it before submitting please check to make sure that the issue is not already addressed if there is a related issue then please cross reference it by if this is a usage question please check out neurostars here and tag your topic with multi echo summarize the issue in sentences linking other issues if they are relevant note simply typing will prompt you for open issues to select from the story so far we are planning a hackathon november in bethesda maryland we will use this issue to discuss what possible items could be added to its agenda this will be kept up to date below items with ✅ are approved by consensus and struck out items were proposed but agreed not to be done proposed items lecture group discussion on math steps decision tree rewrite if needed add additional detail for recreating a bug problem any additional context necessary to understand the issue if desired add suggested next steps if you foresee them in a particular order or priority please use numbering thank you for submitting your issue if you do not receive a response within a calendar week please post a comment on this issue to catch our attention some issues may not be resolved right away due to the volunteer nature of the project thank you for your patience
1
519,842
15,058,036,207
IssuesEvent
2021-02-03 22:41:45
dietterc/SEO-ker
https://api.github.com/repos/dietterc/SEO-ker
opened
Feature 4: Integrate with Google Trends API
feature feature 4 high priority
The game relies on counting google searches, so it should be referring back to historic search data.
1.0
Feature 4: Integrate with Google Trends API - The game relies on counting google searches, so it should be referring back to historic search data.
priority
feature integrate with google trends api the game relies on counting google searches so it should be referring back to historic search data
1
12,304
19,604,036,580
IssuesEvent
2022-01-06 06:50:26
renovatebot/renovate
https://api.github.com/repos/renovatebot/renovate
opened
Support updating Terragrunt-generated `terraform.lock.hcl` files
type:feature status:requirements priority-5-triage
### What would you like Renovate to be able to do? Terraform lock files are generated when running Terragrunt commands (because Terragrunt is just a wrapper of Terraform). These lock files are exactly the same as lock files generated by using Terraform directly, but they don't currently seem to be updated as part of lock file maintenance ([minimal repo](https://github.com/tpansino/renovatebot-terragrunt-test) demonstrating this). I would like Renovate to update Terraform lock files, regardless of how they were generated. ### If you have any ideas on how this should be implemented, please tell us here. I'm a senior Terraform/Terragrunt dev, but not a skilled Javascript dev. I might be able to implement the feature if someone from the Renovate dev team can explain the Terraform manager code better to me. Maybe we could schedule a 20 minute chat? Otherwise, see #13375 , #13356 for my thoughts on how this could be implemented. ### Is this a feature you are interested in implementing yourself? Yes
1.0
Support updating Terragrunt-generated `terraform.lock.hcl` files - ### What would you like Renovate to be able to do? Terraform lock files are generated when running Terragrunt commands (because Terragrunt is just a wrapper of Terraform). These lock files are exactly the same as lock files generated by using Terraform directly, but they don't currently seem to be updated as part of lock file maintenance ([minimal repo](https://github.com/tpansino/renovatebot-terragrunt-test) demonstrating this). I would like Renovate to update Terraform lock files, regardless of how they were generated. ### If you have any ideas on how this should be implemented, please tell us here. I'm a senior Terraform/Terragrunt dev, but not a skilled Javascript dev. I might be able to implement the feature if someone from the Renovate dev team can explain the Terraform manager code better to me. Maybe we could schedule a 20 minute chat? Otherwise, see #13375 , #13356 for my thoughts on how this could be implemented. ### Is this a feature you are interested in implementing yourself? Yes
non_priority
support updating terragrunt generated terraform lock hcl files what would you like renovate to be able to do terraform lock files are generated when running terragrunt commands because terragrunt is just a wrapper of terraform these lock files are exactly the same as lock files generated by using terraform directly but they don t currently seem to be updated as part of lock file maintenance demonstrating this i would like renovate to update terraform lock files regardless of how they were generated if you have any ideas on how this should be implemented please tell us here i m a senior terraform terragrunt dev but not a skilled javascript dev i might be able to implement the feature if someone from the renovate dev team can explain the terraform manager code better to me maybe we could schedule a minute chat otherwise see for my thoughts on how this could be implemented is this a feature you are interested in implementing yourself yes
0
298,857
25,861,330,309
IssuesEvent
2022-12-13 17:09:20
ray-project/ray
https://api.github.com/repos/ray-project/ray
opened
[ml release] `air_benchmark_xgboost_cpu_10` is flaky due to memory issues
P1 triage air release-test
### What happened + What you expected to happen `air_benchmark_xgboost_cpu_10` fails with: ``` (raylet, ip=172.31.199.222) Spilled 3667 MiB, 16 objects, write throughput 1040 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.160.92) Spilled 2303 MiB, 10 objects, write throughput 737 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.247.167) Spilled 3753 MiB, 17 objects, write throughput 838 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.129.182) Spilled 2558 MiB, 12 objects, write throughput 643 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.205.18) Spilled 4520 MiB, 20 objects, write throughput 937 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.182.158) Spilled 2985 MiB, 13 objects, write throughput 613 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.199.222) Spilled 5373 MiB, 24 objects, write throughput 1151 MiB/s. (raylet, ip=172.31.182.205) Spilled 2132 MiB, 9 objects, write throughput 526 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.247.167) Spilled 5544 MiB, 25 objects, write throughput 906 MiB/s. (raylet, ip=172.31.182.158) Spilled 6141 MiB, 26 objects, write throughput 934 MiB/s. (raylet, ip=172.31.205.18) Spilled 6056 MiB, 27 objects, write throughput 819 MiB/s. (raylet, ip=172.31.162.203) Spilled 3753 MiB, 16 objects, write throughput 930 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.162.203) Spilled 4862 MiB, 21 objects, write throughput 954 MiB/s. (raylet, ip=172.31.129.182) Spilled 11430 MiB, 49 objects, write throughput 799 MiB/s. (raylet, ip=172.31.205.18) Spilled 13818 MiB, 58 objects, write throughput 885 MiB/s. (raylet, ip=172.31.160.92) Spilled 11345 MiB, 47 objects, write throughput 754 MiB/s. (raylet, ip=172.31.247.167) Spilled 14672 MiB, 62 objects, write throughput 847 MiB/s. (raylet, ip=172.31.227.81) Spilled 11601 MiB, 48 objects, write throughput 710 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.182.205) Spilled 12624 MiB, 54 objects, write throughput 718 MiB/s. (raylet, ip=172.31.227.230) Spilled 12795 MiB, 53 objects, write throughput 730 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.182.158) Spilled 15013 MiB, 62 objects, write throughput 789 MiB/s. (raylet, ip=172.31.162.203) Spilled 8274 MiB, 36 objects, write throughput 659 MiB/s. (raylet, ip=172.31.199.222) Spilled 15951 MiB, 68 objects, write throughput 808 MiB/s. (raylet, ip=172.31.162.203) Spilled 16804 MiB, 71 objects, write throughput 846 MiB/s. 2022-12-12 15:03:14,340 ERROR trial_runner.py:1095 -- Trial XGBoostTrainer_c44dd_00000: Error processing event. ray.exceptions.RayTaskError(MemoryError): ray::_Inner.train() (pid=469, ip=172.31.162.203, repr=XGBoostTrainer) File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/trainable.py", line 367, in train raise skipped from exception_cause(skipped) File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/function_trainable.py", line 338, in entrypoint self._status_reporter.get_checkpoint(), File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/base_trainer.py", line 480, in _trainable_func super()._trainable_func(self._merged_config, reporter, checkpoint_dir) File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/function_trainable.py", line 652, in _trainable_func output = fn() File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/base_trainer.py", line 390, in train_func trainer.training_loop() File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/gbdt_trainer.py", line 298, in training_loop **config, File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/xgboost/xgboost_trainer.py", line 84, in _train return xgboost_ray.train(**kwargs) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/main.py", line 1414, in train dtrain.load_data(ray_params.num_actors) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/matrix.py", line 819, in load_data self.num_actors, self.sharding, rank=rank) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/matrix.py", line 382, in load_data self.data, ignore=self.ignore, indices=None, **self.kwargs) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/data_sources/ray_dataset.py", line 68, in load_data return ObjectStore.load_data(obj_refs, ignore=ignore, indices=indices) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/data_sources/object_store.py", line 32, in load_data return Pandas.load_data(pd.concat(local_df, copy=False), ignore=ignore) File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/util/_decorators.py", line 311, in wrapper return func(*args, **kwargs) File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/reshape/concat.py", line 307, in concat return op.get_result() File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/reshape/concat.py", line 533, in get_result mgrs_indexers, self.new_axes, concat_axis=self.bm_axis, copy=self.copy File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/internals/concat.py", line 216, in concatenate_managers values = np.concatenate(vals, axis=blk.ndim - 1) File "<__array_function__ internals>", line 6, in concatenate numpy.core._exceptions._ArrayMemoryError: Unable to allocate 77.5 GiB for an array with shape (40, 260000000) and data type float64 ``` which suggests a memory issue. ### Versions / Dependencies master ### Reproduction script https://buildkite.com/ray-project/release-tests-branch/builds/1251#0185086b-9aec-429b-9455-13cf78e5c0db ### Issue Severity None
1.0
[ml release] `air_benchmark_xgboost_cpu_10` is flaky due to memory issues - ### What happened + What you expected to happen `air_benchmark_xgboost_cpu_10` fails with: ``` (raylet, ip=172.31.199.222) Spilled 3667 MiB, 16 objects, write throughput 1040 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.160.92) Spilled 2303 MiB, 10 objects, write throughput 737 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.247.167) Spilled 3753 MiB, 17 objects, write throughput 838 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.129.182) Spilled 2558 MiB, 12 objects, write throughput 643 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.205.18) Spilled 4520 MiB, 20 objects, write throughput 937 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.182.158) Spilled 2985 MiB, 13 objects, write throughput 613 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.199.222) Spilled 5373 MiB, 24 objects, write throughput 1151 MiB/s. (raylet, ip=172.31.182.205) Spilled 2132 MiB, 9 objects, write throughput 526 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.247.167) Spilled 5544 MiB, 25 objects, write throughput 906 MiB/s. (raylet, ip=172.31.182.158) Spilled 6141 MiB, 26 objects, write throughput 934 MiB/s. (raylet, ip=172.31.205.18) Spilled 6056 MiB, 27 objects, write throughput 819 MiB/s. (raylet, ip=172.31.162.203) Spilled 3753 MiB, 16 objects, write throughput 930 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.162.203) Spilled 4862 MiB, 21 objects, write throughput 954 MiB/s. (raylet, ip=172.31.129.182) Spilled 11430 MiB, 49 objects, write throughput 799 MiB/s. (raylet, ip=172.31.205.18) Spilled 13818 MiB, 58 objects, write throughput 885 MiB/s. (raylet, ip=172.31.160.92) Spilled 11345 MiB, 47 objects, write throughput 754 MiB/s. (raylet, ip=172.31.247.167) Spilled 14672 MiB, 62 objects, write throughput 847 MiB/s. (raylet, ip=172.31.227.81) Spilled 11601 MiB, 48 objects, write throughput 710 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.182.205) Spilled 12624 MiB, 54 objects, write throughput 718 MiB/s. (raylet, ip=172.31.227.230) Spilled 12795 MiB, 53 objects, write throughput 730 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message. (raylet, ip=172.31.182.158) Spilled 15013 MiB, 62 objects, write throughput 789 MiB/s. (raylet, ip=172.31.162.203) Spilled 8274 MiB, 36 objects, write throughput 659 MiB/s. (raylet, ip=172.31.199.222) Spilled 15951 MiB, 68 objects, write throughput 808 MiB/s. (raylet, ip=172.31.162.203) Spilled 16804 MiB, 71 objects, write throughput 846 MiB/s. 2022-12-12 15:03:14,340 ERROR trial_runner.py:1095 -- Trial XGBoostTrainer_c44dd_00000: Error processing event. ray.exceptions.RayTaskError(MemoryError): ray::_Inner.train() (pid=469, ip=172.31.162.203, repr=XGBoostTrainer) File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/trainable.py", line 367, in train raise skipped from exception_cause(skipped) File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/function_trainable.py", line 338, in entrypoint self._status_reporter.get_checkpoint(), File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/base_trainer.py", line 480, in _trainable_func super()._trainable_func(self._merged_config, reporter, checkpoint_dir) File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/function_trainable.py", line 652, in _trainable_func output = fn() File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/base_trainer.py", line 390, in train_func trainer.training_loop() File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/gbdt_trainer.py", line 298, in training_loop **config, File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/xgboost/xgboost_trainer.py", line 84, in _train return xgboost_ray.train(**kwargs) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/main.py", line 1414, in train dtrain.load_data(ray_params.num_actors) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/matrix.py", line 819, in load_data self.num_actors, self.sharding, rank=rank) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/matrix.py", line 382, in load_data self.data, ignore=self.ignore, indices=None, **self.kwargs) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/data_sources/ray_dataset.py", line 68, in load_data return ObjectStore.load_data(obj_refs, ignore=ignore, indices=indices) File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/data_sources/object_store.py", line 32, in load_data return Pandas.load_data(pd.concat(local_df, copy=False), ignore=ignore) File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/util/_decorators.py", line 311, in wrapper return func(*args, **kwargs) File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/reshape/concat.py", line 307, in concat return op.get_result() File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/reshape/concat.py", line 533, in get_result mgrs_indexers, self.new_axes, concat_axis=self.bm_axis, copy=self.copy File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/internals/concat.py", line 216, in concatenate_managers values = np.concatenate(vals, axis=blk.ndim - 1) File "<__array_function__ internals>", line 6, in concatenate numpy.core._exceptions._ArrayMemoryError: Unable to allocate 77.5 GiB for an array with shape (40, 260000000) and data type float64 ``` which suggests a memory issue. ### Versions / Dependencies master ### Reproduction script https://buildkite.com/ray-project/release-tests-branch/builds/1251#0185086b-9aec-429b-9455-13cf78e5c0db ### Issue Severity None
non_priority
air benchmark xgboost cpu is flaky due to memory issues what happened what you expected to happen air benchmark xgboost cpu fails with raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s error trial runner py trial xgboosttrainer error processing event ray exceptions raytaskerror memoryerror ray inner train pid ip repr xgboosttrainer file home ray lib site packages ray tune trainable trainable py line in train raise skipped from exception cause skipped file home ray lib site packages ray tune trainable function trainable py line in entrypoint self status reporter get checkpoint file home ray lib site packages ray train base trainer py line in trainable func super trainable func self merged config reporter checkpoint dir file home ray lib site packages ray tune trainable function trainable py line in trainable func output fn file home ray lib site packages ray train base trainer py line in train func trainer training loop file home ray lib site packages ray train gbdt trainer py line in training loop config file home ray lib site packages ray train xgboost xgboost trainer py line in train return xgboost ray train kwargs file home ray lib site packages xgboost ray main py line in train dtrain load data ray params num actors file home ray lib site packages xgboost ray matrix py line in load data self num actors self sharding rank rank file home ray lib site packages xgboost ray matrix py line in load data self data ignore self ignore indices none self kwargs file home ray lib site packages xgboost ray data sources ray dataset py line in load data return objectstore load data obj refs ignore ignore indices indices file home ray lib site packages xgboost ray data sources object store py line in load data return pandas load data pd concat local df copy false ignore ignore file home ray lib site packages pandas util decorators py line in wrapper return func args kwargs file home ray lib site packages pandas core reshape concat py line in concat return op get result file home ray lib site packages pandas core reshape concat py line in get result mgrs indexers self new axes concat axis self bm axis copy self copy file home ray lib site packages pandas core internals concat py line in concatenate managers values np concatenate vals axis blk ndim file line in concatenate numpy core exceptions arraymemoryerror unable to allocate gib for an array with shape and data type which suggests a memory issue versions dependencies master reproduction script issue severity none
0
85,324
10,436,242,287
IssuesEvent
2019-09-17 19:06:36
Programming-Engineering-Pmi-31/IntelligentCooking
https://api.github.com/repos/Programming-Engineering-Pmi-31/IntelligentCooking
closed
Describe all functional and non-functional requirements
documentation
- [x] Describe functional requirements - [x] Describe non-functional requirements
1.0
Describe all functional and non-functional requirements - - [x] Describe functional requirements - [x] Describe non-functional requirements
non_priority
describe all functional and non functional requirements describe functional requirements describe non functional requirements
0
213,125
7,246,172,225
IssuesEvent
2018-02-14 20:40:48
terascope/teraslice
https://api.github.com/repos/terascope/teraslice
closed
/ex/ID/slicer is broken
bug priority:high
This is on master with the most recent slicer related PR included. ``` curl localhost:5678/ex/6ea6e665-5549-42fc-bdb9-8403dc0dc54b/slicer { "error": 500, "message": "could not get slicer statistics, error: [object Object]" } ```
1.0
/ex/ID/slicer is broken - This is on master with the most recent slicer related PR included. ``` curl localhost:5678/ex/6ea6e665-5549-42fc-bdb9-8403dc0dc54b/slicer { "error": 500, "message": "could not get slicer statistics, error: [object Object]" } ```
priority
ex id slicer is broken this is on master with the most recent slicer related pr included curl localhost ex slicer error message could not get slicer statistics error
1
4,539
3,037,642,781
IssuesEvent
2015-08-06 18:09:26
Microsoft/PTVS
https://api.github.com/repos/Microsoft/PTVS
opened
NullReferenceException in DjangoAnalyzer.GetArg
bug Code Intelligence crash
From WER: ``` Level . . . Symbol - InstructionPointer - FunctionOffset - SourceLine - SourceFile 0 . . . microsoft_pythontools_django!Microsoft.PythonTools.Django.Project.DjangoAnalyzer.GetArg - 0x00000000 - 0x0000002C - - <<Failure related >> 1 . . . microsoft_pythontools_django!Microsoft.PythonTools.Django.Project.DjangoAnalyzer.RenderProcessor - 0x00000000 - 0x00000035 - - 2 . . . microsoft_pythontools_analysis!Microsoft.PythonTools.Analysis.Values.SpecializedCallable.Call - 0x00000000 - 0x00000038 - - 3 . . . microsoft_pythontools_analysis!Microsoft.PythonTools.Analysis.Analyzer.ExpressionEvaluator.EvaluateCall - 0x00000000 - 0x000000B4 - - ``` From my reading of the GetArg function, the most likely cause is that the call into `django.shortcuts.render` has an invalid `CallExpression` which results in `namedArgs` (see `EvaluateCall`) being an array with null entries. This occurs when an arg with null `.NameExpression` follows an arg with non-null `.Name`. We should probably assert in `GetNamedArguments` when `args[I].NameExpression` is null, and replace it with a safe non-null value to avoid crashing callers who assume that all elements are non-null.
1.0
NullReferenceException in DjangoAnalyzer.GetArg - From WER: ``` Level . . . Symbol - InstructionPointer - FunctionOffset - SourceLine - SourceFile 0 . . . microsoft_pythontools_django!Microsoft.PythonTools.Django.Project.DjangoAnalyzer.GetArg - 0x00000000 - 0x0000002C - - <<Failure related >> 1 . . . microsoft_pythontools_django!Microsoft.PythonTools.Django.Project.DjangoAnalyzer.RenderProcessor - 0x00000000 - 0x00000035 - - 2 . . . microsoft_pythontools_analysis!Microsoft.PythonTools.Analysis.Values.SpecializedCallable.Call - 0x00000000 - 0x00000038 - - 3 . . . microsoft_pythontools_analysis!Microsoft.PythonTools.Analysis.Analyzer.ExpressionEvaluator.EvaluateCall - 0x00000000 - 0x000000B4 - - ``` From my reading of the GetArg function, the most likely cause is that the call into `django.shortcuts.render` has an invalid `CallExpression` which results in `namedArgs` (see `EvaluateCall`) being an array with null entries. This occurs when an arg with null `.NameExpression` follows an arg with non-null `.Name`. We should probably assert in `GetNamedArguments` when `args[I].NameExpression` is null, and replace it with a safe non-null value to avoid crashing callers who assume that all elements are non-null.
non_priority
nullreferenceexception in djangoanalyzer getarg from wer level symbol instructionpointer functionoffset sourceline sourcefile microsoft pythontools django microsoft pythontools django project djangoanalyzer getarg microsoft pythontools django microsoft pythontools django project djangoanalyzer renderprocessor microsoft pythontools analysis microsoft pythontools analysis values specializedcallable call microsoft pythontools analysis microsoft pythontools analysis analyzer expressionevaluator evaluatecall from my reading of the getarg function the most likely cause is that the call into django shortcuts render has an invalid callexpression which results in namedargs see evaluatecall being an array with null entries this occurs when an arg with null nameexpression follows an arg with non null name we should probably assert in getnamedarguments when args nameexpression is null and replace it with a safe non null value to avoid crashing callers who assume that all elements are non null
0
56,818
3,081,204,954
IssuesEvent
2015-08-22 13:45:50
bitfighter/bitfighter
https://api.github.com/repos/bitfighter/bitfighter
closed
Bouncers bounce right through ship
015 bug imported Priority-Critical
_From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on January 20, 2011 02:51:29_ When armed with bouncers, if you are close to a wall, bouncers will pass right through your ship; instead, they should hit you as they used to. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=40_
1.0
Bouncers bounce right through ship - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on January 20, 2011 02:51:29_ When armed with bouncers, if you are close to a wall, bouncers will pass right through your ship; instead, they should hit you as they used to. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=40_
priority
bouncers bounce right through ship from on january when armed with bouncers if you are close to a wall bouncers will pass right through your ship instead they should hit you as they used to original issue
1
359,830
10,681,515,606
IssuesEvent
2019-10-22 01:09:17
SIGBlockchain/project_aurum
https://api.github.com/repos/SIGBlockchain/project_aurum
closed
Send Block Request
Priority: Critical Type: Enhancement
Branch Context: `send_block` Package context: `requests` Complete the function `SendBlockRequest` that given a block, creates a POST request to the incoming block endpoint with a JSON-block in the body. All unit tests must pass (look at `TestSendBlockRequest`)
1.0
Send Block Request - Branch Context: `send_block` Package context: `requests` Complete the function `SendBlockRequest` that given a block, creates a POST request to the incoming block endpoint with a JSON-block in the body. All unit tests must pass (look at `TestSendBlockRequest`)
priority
send block request branch context send block package context requests complete the function sendblockrequest that given a block creates a post request to the incoming block endpoint with a json block in the body all unit tests must pass look at testsendblockrequest
1
143,480
19,180,304,491
IssuesEvent
2021-12-04 09:05:13
AlexRogalskiy/java-patterns
https://api.github.com/repos/AlexRogalskiy/java-patterns
opened
CVE-2018-16487 (Medium) detected in lodash-2.4.2.tgz
security vulnerability
## CVE-2018-16487 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: java-patterns/package.json</p> <p>Path to vulnerable library: java-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - dockerfile_lint-0.3.4.tgz (Root Library) - :x: **lodash-2.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java-patterns/commit/850df196d336da76298c7b0f9c02003e236df715">850df196d336da76298c7b0f9c02003e236df715</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-16487 (Medium) detected in lodash-2.4.2.tgz - ## CVE-2018-16487 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: java-patterns/package.json</p> <p>Path to vulnerable library: java-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - dockerfile_lint-0.3.4.tgz (Root Library) - :x: **lodash-2.4.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java-patterns/commit/850df196d336da76298c7b0f9c02003e236df715">850df196d336da76298c7b0f9c02003e236df715</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype. <p>Publish Date: 2019-02-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p> <p>Release Date: 2019-02-01</p> <p>Fix Resolution: 4.17.11</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_priority
cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file java patterns package json path to vulnerable library java patterns node modules dockerfile lint node modules lodash package json dependency hierarchy dockerfile lint tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
821,811
30,837,542,644
IssuesEvent
2023-08-02 08:27:20
testomatio/app
https://api.github.com/repos/testomatio/app
closed
Tests counter shows 0 tests till switching Project on Tests Coverage
bug ui\ux jira priority medium
**Describe the bug** https://github.com/testomatio/app/assets/77803888/663528b2-0346-4f55-9df5-638d08b103d0 **To Reproduce** Steps to reproduce the behavior: 1. open Jira plugin 2. go to Tests Coverage page 3. search for an issue 4. see 0 tests shown 5. switch to some project 6. see tests counter updated 7. switch back to All projects 8. see tests counter updated shows the actual number **Expected behavior** Tests Coverage tests counters show actual tests number when users opens it for the first time
1.0
Tests counter shows 0 tests till switching Project on Tests Coverage - **Describe the bug** https://github.com/testomatio/app/assets/77803888/663528b2-0346-4f55-9df5-638d08b103d0 **To Reproduce** Steps to reproduce the behavior: 1. open Jira plugin 2. go to Tests Coverage page 3. search for an issue 4. see 0 tests shown 5. switch to some project 6. see tests counter updated 7. switch back to All projects 8. see tests counter updated shows the actual number **Expected behavior** Tests Coverage tests counters show actual tests number when users opens it for the first time
priority
tests counter shows tests till switching project on tests coverage describe the bug to reproduce steps to reproduce the behavior open jira plugin go to tests coverage page search for an issue see tests shown switch to some project see tests counter updated switch back to all projects see tests counter updated shows the actual number expected behavior tests coverage tests counters show actual tests number when users opens it for the first time
1
532,704
15,569,836,850
IssuesEvent
2021-03-17 01:06:24
Sage-Bionetworks/research-benchmarking-technology
https://api.github.com/repos/Sage-Bionetworks/research-benchmarking-technology
opened
Create Google OAuth 2.0 application for ROCC
Priority: Low
I've created Google OAuth 2.0 in the past for the SageBio Collaboration Portal.
1.0
Create Google OAuth 2.0 application for ROCC - I've created Google OAuth 2.0 in the past for the SageBio Collaboration Portal.
priority
create google oauth application for rocc i ve created google oauth in the past for the sagebio collaboration portal
1
691,154
23,684,943,668
IssuesEvent
2022-08-29 04:52:33
okTurtles/group-income
https://api.github.com/repos/okTurtles/group-income
opened
Dashboard on mobile website too wide
Kind:Bug App:Frontend Priority:High
### Problem On iOS Brave: ![b](https://user-images.githubusercontent.com/138706/187125442-5a3c933a-4c9e-47c7-814d-43c74a11d722.jpg) ### Solution Find out why this is suddenly happening and fix.
1.0
Dashboard on mobile website too wide - ### Problem On iOS Brave: ![b](https://user-images.githubusercontent.com/138706/187125442-5a3c933a-4c9e-47c7-814d-43c74a11d722.jpg) ### Solution Find out why this is suddenly happening and fix.
priority
dashboard on mobile website too wide problem on ios brave solution find out why this is suddenly happening and fix
1
411,384
12,017,556,015
IssuesEvent
2020-04-10 18:40:29
hikaya-io/activity
https://api.github.com/repos/hikaya-io/activity
reopened
ACT-712 Add description field (large text) to Stakeholder Type model
1 Priority
The stakeholder type model needs a large text field called `Description` so that users can provide details about the Stakeholder Type.
1.0
ACT-712 Add description field (large text) to Stakeholder Type model - The stakeholder type model needs a large text field called `Description` so that users can provide details about the Stakeholder Type.
priority
act add description field large text to stakeholder type model the stakeholder type model needs a large text field called description so that users can provide details about the stakeholder type
1
539,316
15,786,773,264
IssuesEvent
2021-04-01 18:14:10
microsoft/PowerToys
https://api.github.com/repos/microsoft/PowerToys
closed
Running as user causes installer to ask for the .msi
Area-Setup/Install Issue-Bug Priority-0 Resolution-Fix-Committed
<!-- **Important: When reporting BSODs or security issues, DO NOT attach memory dumps, logs, or traces to Github issues**. Instead, send dumps/traces to secure@microsoft.com, referencing this GitHub issue. --> ## ℹ Computer information - PowerToys version: 0.31.2.0 - PowerToy Utility: - Running PowerToys as Admin: no - Windows build number: [run "winver"] 18363.1377 ## 📝 Provide detailed reproduction steps (if any) 1. …INstalled PT using local admin account 2. …Tried to run the powertoys icon from the windows menu 3. … ### ✔️ Expected result Start the powertoys utility in the task bar ### ❌ Actual result I ambeing prompted for the network path to the installation files ## 📷 Screenshots _Are there any useful screenshots? WinKey+Shift+S and then just paste them directly into the form_ ![image](https://user-images.githubusercontent.com/79079188/107920807-73327a00-6f65-11eb-9f98-49679d0e7338.png)
1.0
Running as user causes installer to ask for the .msi - <!-- **Important: When reporting BSODs or security issues, DO NOT attach memory dumps, logs, or traces to Github issues**. Instead, send dumps/traces to secure@microsoft.com, referencing this GitHub issue. --> ## ℹ Computer information - PowerToys version: 0.31.2.0 - PowerToy Utility: - Running PowerToys as Admin: no - Windows build number: [run "winver"] 18363.1377 ## 📝 Provide detailed reproduction steps (if any) 1. …INstalled PT using local admin account 2. …Tried to run the powertoys icon from the windows menu 3. … ### ✔️ Expected result Start the powertoys utility in the task bar ### ❌ Actual result I ambeing prompted for the network path to the installation files ## 📷 Screenshots _Are there any useful screenshots? WinKey+Shift+S and then just paste them directly into the form_ ![image](https://user-images.githubusercontent.com/79079188/107920807-73327a00-6f65-11eb-9f98-49679d0e7338.png)
priority
running as user causes installer to ask for the msi important when reporting bsods or security issues do not attach memory dumps logs or traces to github issues instead send dumps traces to secure microsoft com referencing this github issue ℹ computer information powertoys version powertoy utility running powertoys as admin no windows build number 📝 provide detailed reproduction steps if any …installed pt using local admin account …tried to run the powertoys icon from the windows menu … ✔️ expected result start the powertoys utility in the task bar ❌ actual result i ambeing prompted for the network path to the installation files 📷 screenshots are there any useful screenshots winkey shift s and then just paste them directly into the form
1
522,897
15,169,197,374
IssuesEvent
2021-02-12 20:42:55
NCEAS/metacat
https://api.github.com/repos/NCEAS/metacat
closed
allow metacat to store files on multiple fs
Category: metacat Component: Bugzilla-Id Priority: Normal Status: In Progress Tracker: Bug wontfix
--- Author Name: **Matt Jones** (Matt Jones) Original Redmine Issue: 195, https://projects.ecoinformatics.org/ecoinfo/issues/195 Original Date: 2001-04-09 Original Assignee: Jing Tao --- Metacat currently stores files on a single file system. Need to changes this so that Metacat can be configured to store files on multiple file systems in case space management by the administrator requires this.
1.0
allow metacat to store files on multiple fs - --- Author Name: **Matt Jones** (Matt Jones) Original Redmine Issue: 195, https://projects.ecoinformatics.org/ecoinfo/issues/195 Original Date: 2001-04-09 Original Assignee: Jing Tao --- Metacat currently stores files on a single file system. Need to changes this so that Metacat can be configured to store files on multiple file systems in case space management by the administrator requires this.
priority
allow metacat to store files on multiple fs author name matt jones matt jones original redmine issue original date original assignee jing tao metacat currently stores files on a single file system need to changes this so that metacat can be configured to store files on multiple file systems in case space management by the administrator requires this
1
82,183
3,603,835,129
IssuesEvent
2016-02-03 20:33:17
ualbertalib/HydraNorth
https://api.github.com/repos/ualbertalib/HydraNorth
closed
ERA_Collections: Collection-specific deposit page
enhancement priority:medium size:large
As admin I want to be able to define an identifier in the deposit form to be used for a new collection.
1.0
ERA_Collections: Collection-specific deposit page - As admin I want to be able to define an identifier in the deposit form to be used for a new collection.
priority
era collections collection specific deposit page as admin i want to be able to define an identifier in the deposit form to be used for a new collection
1
151,931
5,830,082,174
IssuesEvent
2017-05-08 15:57:35
emfoundation/ce100-app
https://api.github.com/repos/emfoundation/ce100-app
opened
Improved tag selection flow
primary-user priority-2
This applies to creating/editing organisations/challenges. Rather than having a two step process, where the user first adds/edits other content to an org profile or challenge and then clicks next to add tags, the tag selection will be more integrated. The currently selected (if any) tags are shown in the edit form with a "Add/Edit tags" button below them. Clicking on that button opens up the tag selection screen. On that screen, clicking either "Select" or "Cancel" takes you back to the edit form. Only after clicking "Save" on that form are the changes to the tags saved. Mock-ups to be added.
1.0
Improved tag selection flow - This applies to creating/editing organisations/challenges. Rather than having a two step process, where the user first adds/edits other content to an org profile or challenge and then clicks next to add tags, the tag selection will be more integrated. The currently selected (if any) tags are shown in the edit form with a "Add/Edit tags" button below them. Clicking on that button opens up the tag selection screen. On that screen, clicking either "Select" or "Cancel" takes you back to the edit form. Only after clicking "Save" on that form are the changes to the tags saved. Mock-ups to be added.
priority
improved tag selection flow this applies to creating editing organisations challenges rather than having a two step process where the user first adds edits other content to an org profile or challenge and then clicks next to add tags the tag selection will be more integrated the currently selected if any tags are shown in the edit form with a add edit tags button below them clicking on that button opens up the tag selection screen on that screen clicking either select or cancel takes you back to the edit form only after clicking save on that form are the changes to the tags saved mock ups to be added
1
668,264
22,575,956,430
IssuesEvent
2022-06-28 07:21:55
OpenFunction/functions-framework-go
https://api.github.com/repos/OpenFunction/functions-framework-go
closed
[Feature]Support advanced http router and features
enhancement priority/medium
1. use [mux](https://github.com/gorilla/mux) as the default handler in ff-go which is more production-ready and we can bring in new features - https://github.com/OpenFunction/functions-framework/issues/21#issuecomment-1150587351, https://github.com/OpenFunction/functions-framework/issues/17 2. support defining variables in the path like `/functions/<funcname>`, and use it in the function by `ofnctx.Vars()` 3. support defining method matcher like `withFunctionMethods("GET", "POST")`
1.0
[Feature]Support advanced http router and features - 1. use [mux](https://github.com/gorilla/mux) as the default handler in ff-go which is more production-ready and we can bring in new features - https://github.com/OpenFunction/functions-framework/issues/21#issuecomment-1150587351, https://github.com/OpenFunction/functions-framework/issues/17 2. support defining variables in the path like `/functions/<funcname>`, and use it in the function by `ofnctx.Vars()` 3. support defining method matcher like `withFunctionMethods("GET", "POST")`
priority
support advanced http router and features use as the default handler in ff go which is more production ready and we can bring in new features support defining variables in the path like functions and use it in the function by ofnctx vars support defining method matcher like withfunctionmethods get post
1
246,816
7,895,650,547
IssuesEvent
2018-06-29 04:48:58
aowen87/BAR
https://api.github.com/repos/aowen87/BAR
closed
pillow snuck in to build_visit (bv_python) on the trunk
Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 4 - Crash / Wrong Results Support Group: Any bug version: 2.8.2
During a merge, I accidentally committed a change to build pillow instead of PIL in the trunk version of build_visit. We do want to use Pillow in the future, at this moment we need test suite changes to support Pillow instead of PIL. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Cyrus Harrison Original creation: 02/11/2015 04:44 pm Original update: 09/02/2015 04:19 pm Ticket number: 2144
1.0
pillow snuck in to build_visit (bv_python) on the trunk - During a merge, I accidentally committed a change to build pillow instead of PIL in the trunk version of build_visit. We do want to use Pillow in the future, at this moment we need test suite changes to support Pillow instead of PIL. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. The following information could not be accurately captured in the new ticket: Original author: Cyrus Harrison Original creation: 02/11/2015 04:44 pm Original update: 09/02/2015 04:19 pm Ticket number: 2144
priority
pillow snuck in to build visit bv python on the trunk during a merge i accidentally committed a change to build pillow instead of pil in the trunk version of build visit we do want to use pillow in the future at this moment we need test suite changes to support pillow instead of pil redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author cyrus harrison original creation pm original update pm ticket number
1
301,499
26,052,842,941
IssuesEvent
2022-12-22 20:43:14
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
closed
E2E Testing Tool: Storing metrics on client
type/enhancement e2e-testing-tool team/connector-ops
https://github.com/airbytehq/airbyte/issues/14915 We need to store/collect metrics somewhere on client side for reports/charts/analyzing/representation
1.0
E2E Testing Tool: Storing metrics on client - https://github.com/airbytehq/airbyte/issues/14915 We need to store/collect metrics somewhere on client side for reports/charts/analyzing/representation
non_priority
testing tool storing metrics on client we need to store collect metrics somewhere on client side for reports charts analyzing representation
0
89,320
3,792,633,480
IssuesEvent
2016-03-22 10:32:02
handsontable/handsontable
https://api.github.com/repos/handsontable/handsontable
closed
Add source maps to the repo
Priority: normal
As pointed in #1009, Chrome Developer Tools show 404 (Not Found) error message about missing source maps: ``` .../jquery-handsontable/lib/jquery-1.10.2.min.map .../jquery-handsontable/dist/json-patch-duplex.js.map ``` As solution I propose to add missing source maps and files to which they link (without altering the files that contain the source map reference) I'd rather keep the source maps, because this is also what jQuery does in their minified, production-ready files: http://jquery.com/download/ This issue only affects Chrome users with "Enable source maps" option enabled in the Developer Tools.
1.0
Add source maps to the repo - As pointed in #1009, Chrome Developer Tools show 404 (Not Found) error message about missing source maps: ``` .../jquery-handsontable/lib/jquery-1.10.2.min.map .../jquery-handsontable/dist/json-patch-duplex.js.map ``` As solution I propose to add missing source maps and files to which they link (without altering the files that contain the source map reference) I'd rather keep the source maps, because this is also what jQuery does in their minified, production-ready files: http://jquery.com/download/ This issue only affects Chrome users with "Enable source maps" option enabled in the Developer Tools.
priority
add source maps to the repo as pointed in chrome developer tools show not found error message about missing source maps jquery handsontable lib jquery min map jquery handsontable dist json patch duplex js map as solution i propose to add missing source maps and files to which they link without altering the files that contain the source map reference i d rather keep the source maps because this is also what jquery does in their minified production ready files this issue only affects chrome users with enable source maps option enabled in the developer tools
1
347,789
24,899,788,319
IssuesEvent
2022-10-28 19:32:16
xylan-org/mailspy
https://api.github.com/repos/xylan-org/mailspy
opened
Change default path to /devtools/mailspy in README after release 1.1.0
documentation
This includes the Spring Security related part and the config table.
1.0
Change default path to /devtools/mailspy in README after release 1.1.0 - This includes the Spring Security related part and the config table.
non_priority
change default path to devtools mailspy in readme after release this includes the spring security related part and the config table
0
134,864
10,933,139,693
IssuesEvent
2019-11-23 23:04:42
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
closed
df.groupby('symbol')['status'].ffill() results in Exception: cannot handle a non-unique multi-index!
Needs Tests good first issue
#### Code Sample, a copy-pastable example if possible ```python import pandas as pd df = pd.DataFrame([['2018-01-01','MSFT','short'], ['2018-01-01','MSFT','long'], ['2018-01-01','MSFT',pd.np.NaN], ['2018-01-01','AAPL','short'], ['2018-01-01','AAPL',pd.np.NaN], ['2018-01-02','AAPL','short'], ['2018-01-01','TSLA','neutral'], ['2018-01-02','TSLA',pd.np.NaN]], columns=['date','symbol','status']) df['date'] = pd.to_datetime(df['date']) df = df.set_index(['date','symbol']) df.groupby('symbol')['status'].ffill() ``` #### Raised exception <details> ``` --------------------------------------------------------------------------- Exception Traceback (most recent call last) <ipython-input-58-2fe0989150cf> in <module>() 13 df['date'] = pd.to_datetime(df['date']) 14 df = df.set_index(['date','symbol']) ---> 15 df.groupby('symbol')['status'].ffill() /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in pad(self, limit) 1362 DataFrame.fillna 1363 """ -> 1364 return self.apply(lambda x: x.ffill(limit=limit)) 1365 ffill = pad 1366 /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs) 3110 examples=_apply_docs['series_examples'])) 3111 def apply(self, func, *args, **kwargs): -> 3112 return super(SeriesGroupBy, self).apply(func, *args, **kwargs) 3113 3114 @Appender(_agg_doc) /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs) 803 # ignore SettingWithCopy here in case the user mutates 804 with option_context('mode.chained_assignment', None): --> 805 return self._python_apply_general(f) 806 807 def _python_apply_general(self, f): /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _python_apply_general(self, f) 812 keys, 813 values, --> 814 not_indexed_same=mutated or self.mutated) 815 816 def _iterate_slices(self): /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _wrap_applied_output(self, keys, values, not_indexed_same) 3250 if isinstance(values[0], (Series, dict)): 3251 return self._concat_objects(keys, values, -> 3252 not_indexed_same=not_indexed_same) 3253 elif isinstance(values[0], DataFrame): 3254 # possible that Series -> DataFrame by applied function /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _concat_objects(self, keys, values, not_indexed_same) 978 979 if isinstance(result, Series): --> 980 result = result.reindex(ax) 981 else: 982 /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/series.pyc in reindex(self, index, **kwargs) 2679 @Appender(generic._shared_docs['reindex'] % _shared_doc_kwargs) 2680 def reindex(self, index=None, **kwargs): -> 2681 return super(Series, self).reindex(index=index, **kwargs) 2682 2683 @Appender(generic._shared_docs['fillna'] % _shared_doc_kwargs) /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in reindex(self, *args, **kwargs) 3021 # perform the reindex on the axes 3022 return self._reindex_axes(axes, level, limit, tolerance, method, -> 3023 fill_value, copy).__finalize__(self) 3024 3025 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy) 3034 ax = self._get_axis(a) 3035 new_index, indexer = ax.reindex(labels, level=level, limit=limit, -> 3036 tolerance=tolerance, method=method) 3037 3038 axis = self._get_axis_number(a) /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/indexes/multi.pyc in reindex(self, target, method, level, limit, tolerance) 1901 tolerance=tolerance) 1902 else: -> 1903 raise Exception("cannot handle a non-unique multi-index!") 1904 1905 if not isinstance(target, MultiIndex): Exception: cannot handle a non-unique multi-index! ``` </details> #### Problem description Performing `ffill()` does not forward fill values as intended. The index has duplicate values, the expectation is that the values will be filled according to the order the rows are in already. Duplicate index values are possible when using high frequency time intervals. #### Expected Output ```python df = pd.DataFrame([['2018-01-01','MSFT','short'], ['2018-01-01','MSFT','long'], ['2018-01-01','MSFT','long'], ['2018-01-01','AAPL','short'], ['2018-01-01','AAPL','short'], ['2018-01-02','AAPL','short'], ['2018-01-01','TSLA','neutral'], ['2018-01-02','TSLA','neutral']], columns=['date','symbol','status']) df['date'] = pd.to_datetime(df['date']) ``` #### Output of ``pd.show_versions()`` <details> INSTALLED VERSIONS ------------------ commit: None python: 2.7.14.final.0 python-bits: 64 OS: Linux OS-release: 4.13.0-31-generic machine: x86_64 processor: x86_64 byteorder: little LC_ALL: None LANG: en_AU.UTF-8 LOCALE: None.None pandas: 0.22.0 pytest: 3.2.1 pip: 9.0.1 setuptools: 36.5.0.post20170921 Cython: 0.26.1 numpy: 1.14.0 scipy: 0.19.1 pyarrow: None xarray: None IPython: 5.4.1 sphinx: 1.6.3 patsy: 0.4.1 dateutil: 2.6.1 pytz: 2017.3 blosc: None bottleneck: 1.2.1 tables: 3.4.2 numexpr: 2.6.2 feather: None matplotlib: 2.1.0 openpyxl: 2.4.8 xlrd: 1.1.0 xlwt: 1.3.0 xlsxwriter: 1.0.2 lxml: 4.1.0 bs4: 4.6.0 html5lib: 0.999999999 sqlalchemy: 1.1.13 pymysql: None psycopg2: 2.7.3.2 (dt dec pq3 ext lo64) jinja2: 2.9.6 s3fs: None fastparquet: None pandas_gbq: None pandas_datareader: None </details> This also happens with `axis=1`. ```python import pandas as pd import numpy as np df = pd.DataFrame( np.ones([6, 4], dtype=int), columns=pd.MultiIndex.from_product([['A', 'B'], [1, 2]]) ) ( df .groupby(level=0, axis=1) .apply( lambda df: 2*df.xs(df.name, axis=1) ) ) ``` <details> ```pytb --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-12-e7c022399f89> in <module>() 3 .groupby(level=0, axis=1) 4 .apply( ----> 5 lambda df: 2*df.xs(df.name, axis=1) 6 ) 7 ) ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in apply(self, func, *args, **kwargs) 882 # ignore SettingWithCopy here in case the user mutates 883 with option_context('mode.chained_assignment', None): --> 884 return self._python_apply_general(f) 885 886 def _python_apply_general(self, f): ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _python_apply_general(self, f) 891 keys, 892 values, --> 893 not_indexed_same=mutated or self.mutated) 894 895 def _iterate_slices(self): ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _wrap_applied_output(self, keys, values, not_indexed_same) 3920 elif isinstance(v, DataFrame): 3921 return self._concat_objects(keys, values, -> 3922 not_indexed_same=not_indexed_same) 3923 elif self.grouper.groupings is not None: 3924 if len(self.grouper.groupings) > 1: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _concat_objects(self, keys, values, not_indexed_same) 1085 result = result.take(indexer, axis=self.axis) 1086 else: -> 1087 result = result.reindex(ax, axis=self.axis) 1088 1089 elif self.group_keys: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/util/_decorators.py in wrapper(*args, **kwargs) 145 @wraps(func) 146 def wrapper(*args, **kwargs): --> 147 return func(*args, **kwargs) 148 149 if not PY2: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in reindex(self, *args, **kwargs) 2979 kwargs.pop('axis', None) 2980 kwargs.pop('labels', None) -> 2981 return super(DataFrame, self).reindex(**kwargs) 2982 2983 @Appender(_shared_docs['reindex_axis'] % _shared_doc_kwargs) ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in reindex(self, *args, **kwargs) 3356 # perform the reindex on the axes 3357 return self._reindex_axes(axes, level, limit, tolerance, method, -> 3358 fill_value, copy).__finalize__(self) 3359 3360 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy) 2909 if columns is not None: 2910 frame = frame._reindex_columns(columns, method, copy, level, -> 2911 fill_value, limit, tolerance) 2912 2913 index = axes['index'] ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_columns(self, new_columns, method, copy, level, fill_value, limit, tolerance) 2934 return self._reindex_with_indexers({1: [new_columns, indexer]}, 2935 copy=copy, fill_value=fill_value, -> 2936 allow_dups=False) 2937 2938 def _reindex_multi(self, axes, copy, fill_value): ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in _reindex_with_indexers(self, reindexers, fill_value, copy, allow_dups) 3477 fill_value=fill_value, 3478 allow_dups=allow_dups, -> 3479 copy=copy) 3480 3481 if copy and new_data is self._data: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/internals.py in reindex_indexer(self, new_axis, indexer, axis, fill_value, allow_dups, copy) 4125 # some axes don't allow reindexing with dups 4126 if not allow_dups: -> 4127 self.axes[axis]._can_reindex(indexer) 4128 4129 if axis >= self.ndim: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/indexes/base.py in _can_reindex(self, indexer) 2940 # trying to reindex on an axis with duplicates 2941 if not self.is_unique and len(indexer): -> 2942 raise ValueError("cannot reindex from a duplicate axis") 2943 2944 def reindex(self, target, method=None, level=None, limit=None, ValueError: cannot reindex from a duplicate axis ``` </details>
1.0
df.groupby('symbol')['status'].ffill() results in Exception: cannot handle a non-unique multi-index! - #### Code Sample, a copy-pastable example if possible ```python import pandas as pd df = pd.DataFrame([['2018-01-01','MSFT','short'], ['2018-01-01','MSFT','long'], ['2018-01-01','MSFT',pd.np.NaN], ['2018-01-01','AAPL','short'], ['2018-01-01','AAPL',pd.np.NaN], ['2018-01-02','AAPL','short'], ['2018-01-01','TSLA','neutral'], ['2018-01-02','TSLA',pd.np.NaN]], columns=['date','symbol','status']) df['date'] = pd.to_datetime(df['date']) df = df.set_index(['date','symbol']) df.groupby('symbol')['status'].ffill() ``` #### Raised exception <details> ``` --------------------------------------------------------------------------- Exception Traceback (most recent call last) <ipython-input-58-2fe0989150cf> in <module>() 13 df['date'] = pd.to_datetime(df['date']) 14 df = df.set_index(['date','symbol']) ---> 15 df.groupby('symbol')['status'].ffill() /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in pad(self, limit) 1362 DataFrame.fillna 1363 """ -> 1364 return self.apply(lambda x: x.ffill(limit=limit)) 1365 ffill = pad 1366 /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs) 3110 examples=_apply_docs['series_examples'])) 3111 def apply(self, func, *args, **kwargs): -> 3112 return super(SeriesGroupBy, self).apply(func, *args, **kwargs) 3113 3114 @Appender(_agg_doc) /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs) 803 # ignore SettingWithCopy here in case the user mutates 804 with option_context('mode.chained_assignment', None): --> 805 return self._python_apply_general(f) 806 807 def _python_apply_general(self, f): /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _python_apply_general(self, f) 812 keys, 813 values, --> 814 not_indexed_same=mutated or self.mutated) 815 816 def _iterate_slices(self): /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _wrap_applied_output(self, keys, values, not_indexed_same) 3250 if isinstance(values[0], (Series, dict)): 3251 return self._concat_objects(keys, values, -> 3252 not_indexed_same=not_indexed_same) 3253 elif isinstance(values[0], DataFrame): 3254 # possible that Series -> DataFrame by applied function /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _concat_objects(self, keys, values, not_indexed_same) 978 979 if isinstance(result, Series): --> 980 result = result.reindex(ax) 981 else: 982 /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/series.pyc in reindex(self, index, **kwargs) 2679 @Appender(generic._shared_docs['reindex'] % _shared_doc_kwargs) 2680 def reindex(self, index=None, **kwargs): -> 2681 return super(Series, self).reindex(index=index, **kwargs) 2682 2683 @Appender(generic._shared_docs['fillna'] % _shared_doc_kwargs) /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in reindex(self, *args, **kwargs) 3021 # perform the reindex on the axes 3022 return self._reindex_axes(axes, level, limit, tolerance, method, -> 3023 fill_value, copy).__finalize__(self) 3024 3025 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy) 3034 ax = self._get_axis(a) 3035 new_index, indexer = ax.reindex(labels, level=level, limit=limit, -> 3036 tolerance=tolerance, method=method) 3037 3038 axis = self._get_axis_number(a) /home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/indexes/multi.pyc in reindex(self, target, method, level, limit, tolerance) 1901 tolerance=tolerance) 1902 else: -> 1903 raise Exception("cannot handle a non-unique multi-index!") 1904 1905 if not isinstance(target, MultiIndex): Exception: cannot handle a non-unique multi-index! ``` </details> #### Problem description Performing `ffill()` does not forward fill values as intended. The index has duplicate values, the expectation is that the values will be filled according to the order the rows are in already. Duplicate index values are possible when using high frequency time intervals. #### Expected Output ```python df = pd.DataFrame([['2018-01-01','MSFT','short'], ['2018-01-01','MSFT','long'], ['2018-01-01','MSFT','long'], ['2018-01-01','AAPL','short'], ['2018-01-01','AAPL','short'], ['2018-01-02','AAPL','short'], ['2018-01-01','TSLA','neutral'], ['2018-01-02','TSLA','neutral']], columns=['date','symbol','status']) df['date'] = pd.to_datetime(df['date']) ``` #### Output of ``pd.show_versions()`` <details> INSTALLED VERSIONS ------------------ commit: None python: 2.7.14.final.0 python-bits: 64 OS: Linux OS-release: 4.13.0-31-generic machine: x86_64 processor: x86_64 byteorder: little LC_ALL: None LANG: en_AU.UTF-8 LOCALE: None.None pandas: 0.22.0 pytest: 3.2.1 pip: 9.0.1 setuptools: 36.5.0.post20170921 Cython: 0.26.1 numpy: 1.14.0 scipy: 0.19.1 pyarrow: None xarray: None IPython: 5.4.1 sphinx: 1.6.3 patsy: 0.4.1 dateutil: 2.6.1 pytz: 2017.3 blosc: None bottleneck: 1.2.1 tables: 3.4.2 numexpr: 2.6.2 feather: None matplotlib: 2.1.0 openpyxl: 2.4.8 xlrd: 1.1.0 xlwt: 1.3.0 xlsxwriter: 1.0.2 lxml: 4.1.0 bs4: 4.6.0 html5lib: 0.999999999 sqlalchemy: 1.1.13 pymysql: None psycopg2: 2.7.3.2 (dt dec pq3 ext lo64) jinja2: 2.9.6 s3fs: None fastparquet: None pandas_gbq: None pandas_datareader: None </details> This also happens with `axis=1`. ```python import pandas as pd import numpy as np df = pd.DataFrame( np.ones([6, 4], dtype=int), columns=pd.MultiIndex.from_product([['A', 'B'], [1, 2]]) ) ( df .groupby(level=0, axis=1) .apply( lambda df: 2*df.xs(df.name, axis=1) ) ) ``` <details> ```pytb --------------------------------------------------------------------------- ValueError Traceback (most recent call last) <ipython-input-12-e7c022399f89> in <module>() 3 .groupby(level=0, axis=1) 4 .apply( ----> 5 lambda df: 2*df.xs(df.name, axis=1) 6 ) 7 ) ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in apply(self, func, *args, **kwargs) 882 # ignore SettingWithCopy here in case the user mutates 883 with option_context('mode.chained_assignment', None): --> 884 return self._python_apply_general(f) 885 886 def _python_apply_general(self, f): ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _python_apply_general(self, f) 891 keys, 892 values, --> 893 not_indexed_same=mutated or self.mutated) 894 895 def _iterate_slices(self): ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _wrap_applied_output(self, keys, values, not_indexed_same) 3920 elif isinstance(v, DataFrame): 3921 return self._concat_objects(keys, values, -> 3922 not_indexed_same=not_indexed_same) 3923 elif self.grouper.groupings is not None: 3924 if len(self.grouper.groupings) > 1: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _concat_objects(self, keys, values, not_indexed_same) 1085 result = result.take(indexer, axis=self.axis) 1086 else: -> 1087 result = result.reindex(ax, axis=self.axis) 1088 1089 elif self.group_keys: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/util/_decorators.py in wrapper(*args, **kwargs) 145 @wraps(func) 146 def wrapper(*args, **kwargs): --> 147 return func(*args, **kwargs) 148 149 if not PY2: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in reindex(self, *args, **kwargs) 2979 kwargs.pop('axis', None) 2980 kwargs.pop('labels', None) -> 2981 return super(DataFrame, self).reindex(**kwargs) 2982 2983 @Appender(_shared_docs['reindex_axis'] % _shared_doc_kwargs) ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in reindex(self, *args, **kwargs) 3356 # perform the reindex on the axes 3357 return self._reindex_axes(axes, level, limit, tolerance, method, -> 3358 fill_value, copy).__finalize__(self) 3359 3360 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy) 2909 if columns is not None: 2910 frame = frame._reindex_columns(columns, method, copy, level, -> 2911 fill_value, limit, tolerance) 2912 2913 index = axes['index'] ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_columns(self, new_columns, method, copy, level, fill_value, limit, tolerance) 2934 return self._reindex_with_indexers({1: [new_columns, indexer]}, 2935 copy=copy, fill_value=fill_value, -> 2936 allow_dups=False) 2937 2938 def _reindex_multi(self, axes, copy, fill_value): ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in _reindex_with_indexers(self, reindexers, fill_value, copy, allow_dups) 3477 fill_value=fill_value, 3478 allow_dups=allow_dups, -> 3479 copy=copy) 3480 3481 if copy and new_data is self._data: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/internals.py in reindex_indexer(self, new_axis, indexer, axis, fill_value, allow_dups, copy) 4125 # some axes don't allow reindexing with dups 4126 if not allow_dups: -> 4127 self.axes[axis]._can_reindex(indexer) 4128 4129 if axis >= self.ndim: ~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/indexes/base.py in _can_reindex(self, indexer) 2940 # trying to reindex on an axis with duplicates 2941 if not self.is_unique and len(indexer): -> 2942 raise ValueError("cannot reindex from a duplicate axis") 2943 2944 def reindex(self, target, method=None, level=None, limit=None, ValueError: cannot reindex from a duplicate axis ``` </details>
non_priority
df groupby symbol ffill results in exception cannot handle a non unique multi index code sample a copy pastable example if possible python import pandas as pd df pd dataframe columns df pd to datetime df df df set index df groupby symbol ffill raised exception exception traceback most recent call last in df pd to datetime df df df set index df groupby symbol ffill home greg lib site packages pandas core groupby pyc in pad self limit dataframe fillna return self apply lambda x x ffill limit limit ffill pad home greg lib site packages pandas core groupby pyc in apply self func args kwargs examples apply docs def apply self func args kwargs return super seriesgroupby self apply func args kwargs appender agg doc home greg lib site packages pandas core groupby pyc in apply self func args kwargs ignore settingwithcopy here in case the user mutates with option context mode chained assignment none return self python apply general f def python apply general self f home greg lib site packages pandas core groupby pyc in python apply general self f keys values not indexed same mutated or self mutated def iterate slices self home greg lib site packages pandas core groupby pyc in wrap applied output self keys values not indexed same if isinstance values series dict return self concat objects keys values not indexed same not indexed same elif isinstance values dataframe possible that series dataframe by applied function home greg lib site packages pandas core groupby pyc in concat objects self keys values not indexed same if isinstance result series result result reindex ax else home greg lib site packages pandas core series pyc in reindex self index kwargs appender generic shared docs shared doc kwargs def reindex self index none kwargs return super series self reindex index index kwargs appender generic shared docs shared doc kwargs home greg lib site packages pandas core generic pyc in reindex self args kwargs perform the reindex on the axes return self reindex axes axes level limit tolerance method fill value copy finalize self def reindex axes self axes level limit tolerance method fill value home greg lib site packages pandas core generic pyc in reindex axes self axes level limit tolerance method fill value copy ax self get axis a new index indexer ax reindex labels level level limit limit tolerance tolerance method method axis self get axis number a home greg lib site packages pandas core indexes multi pyc in reindex self target method level limit tolerance tolerance tolerance else raise exception cannot handle a non unique multi index if not isinstance target multiindex exception cannot handle a non unique multi index problem description performing ffill does not forward fill values as intended the index has duplicate values the expectation is that the values will be filled according to the order the rows are in already duplicate index values are possible when using high frequency time intervals expected output python df pd dataframe columns df pd to datetime df output of pd show versions installed versions commit none python final python bits os linux os release generic machine processor byteorder little lc all none lang en au utf locale none none pandas pytest pip setuptools cython numpy scipy pyarrow none xarray none ipython sphinx patsy dateutil pytz blosc none bottleneck tables numexpr feather none matplotlib openpyxl xlrd xlwt xlsxwriter lxml sqlalchemy pymysql none dt dec ext none fastparquet none pandas gbq none pandas datareader none this also happens with axis python import pandas as pd import numpy as np df pd dataframe np ones dtype int columns pd multiindex from product df groupby level axis apply lambda df df xs df name axis pytb valueerror traceback most recent call last in groupby level axis apply lambda df df xs df name axis envs pandas dev lib site packages pandas pandas core groupby py in apply self func args kwargs ignore settingwithcopy here in case the user mutates with option context mode chained assignment none return self python apply general f def python apply general self f envs pandas dev lib site packages pandas pandas core groupby py in python apply general self f keys values not indexed same mutated or self mutated def iterate slices self envs pandas dev lib site packages pandas pandas core groupby py in wrap applied output self keys values not indexed same elif isinstance v dataframe return self concat objects keys values not indexed same not indexed same elif self grouper groupings is not none if len self grouper groupings envs pandas dev lib site packages pandas pandas core groupby py in concat objects self keys values not indexed same result result take indexer axis self axis else result result reindex ax axis self axis elif self group keys envs pandas dev lib site packages pandas pandas util decorators py in wrapper args kwargs wraps func def wrapper args kwargs return func args kwargs if not envs pandas dev lib site packages pandas pandas core frame py in reindex self args kwargs kwargs pop axis none kwargs pop labels none return super dataframe self reindex kwargs appender shared docs shared doc kwargs envs pandas dev lib site packages pandas pandas core generic py in reindex self args kwargs perform the reindex on the axes return self reindex axes axes level limit tolerance method fill value copy finalize self def reindex axes self axes level limit tolerance method fill value envs pandas dev lib site packages pandas pandas core frame py in reindex axes self axes level limit tolerance method fill value copy if columns is not none frame frame reindex columns columns method copy level fill value limit tolerance index axes envs pandas dev lib site packages pandas pandas core frame py in reindex columns self new columns method copy level fill value limit tolerance return self reindex with indexers copy copy fill value fill value allow dups false def reindex multi self axes copy fill value envs pandas dev lib site packages pandas pandas core generic py in reindex with indexers self reindexers fill value copy allow dups fill value fill value allow dups allow dups copy copy if copy and new data is self data envs pandas dev lib site packages pandas pandas core internals py in reindex indexer self new axis indexer axis fill value allow dups copy some axes don t allow reindexing with dups if not allow dups self axes can reindex indexer if axis self ndim envs pandas dev lib site packages pandas pandas core indexes base py in can reindex self indexer trying to reindex on an axis with duplicates if not self is unique and len indexer raise valueerror cannot reindex from a duplicate axis def reindex self target method none level none limit none valueerror cannot reindex from a duplicate axis
0
39,550
10,352,152,114
IssuesEvent
2019-09-05 08:38:50
apollographql/apollo-ios
https://api.github.com/repos/apollographql/apollo-ios
closed
Unable to Compile Apollo version 0.15.0 with Xcode 10.2 no API.swift file is created.
build-issue
Using Apollo version 0.15.0 Npm version 6.10.3 Node version v12.9.1 **Not able to compile the project with Legacy Build System and New Build System** > > › Warning: apollo update available from 1.9.2 to 2.18.0. > ++ npx --no-install apollo codegen:generate --target=swift '--includes=./**/*.graphql' --localSchemaFile=schema.json API.swift > › Warning: apollo update available from 1.9.2 to 2.18.0. > › Error: Unexpected arguments: --localSchemaFile=schema.json, API.swift > › See more help with --help > Command /bin/sh failed with exit code 2
1.0
Unable to Compile Apollo version 0.15.0 with Xcode 10.2 no API.swift file is created. - Using Apollo version 0.15.0 Npm version 6.10.3 Node version v12.9.1 **Not able to compile the project with Legacy Build System and New Build System** > > › Warning: apollo update available from 1.9.2 to 2.18.0. > ++ npx --no-install apollo codegen:generate --target=swift '--includes=./**/*.graphql' --localSchemaFile=schema.json API.swift > › Warning: apollo update available from 1.9.2 to 2.18.0. > › Error: Unexpected arguments: --localSchemaFile=schema.json, API.swift > › See more help with --help > Command /bin/sh failed with exit code 2
non_priority
unable to compile apollo version with xcode no api swift file is created using apollo version npm version node version not able to compile the project with legacy build system and new build system › warning apollo update available from to npx no install apollo codegen generate target swift includes graphql localschemafile schema json api swift › warning apollo update available from to › error unexpected arguments localschemafile schema json api swift › see more help with help command bin sh failed with exit code
0
20,028
14,937,745,091
IssuesEvent
2021-01-25 14:59:25
andrewfstratton/quando
https://api.github.com/repos/andrewfstratton/quando
opened
Move delete away from load
usability
Easy to hit delete on narrow screen by accident - especially when viewing All, which is likely to delete the latest version.
True
Move delete away from load - Easy to hit delete on narrow screen by accident - especially when viewing All, which is likely to delete the latest version.
non_priority
move delete away from load easy to hit delete on narrow screen by accident especially when viewing all which is likely to delete the latest version
0
231,407
18,765,504,756
IssuesEvent
2021-11-05 23:04:26
numpy/numpy
https://api.github.com/repos/numpy/numpy
closed
4 test cases fail in test running with the latest master code base
05 - Testing
<!-- Please describe the issue in detail here, and fill in the fields below --> ### Reproducing code example: <!-- A short code example that reproduces the problem/missing feature. It should be self-contained, i.e., possible to run as-is via 'python myproblem.py' --> ```python python -c 'import numpy as np; np.test()' ``` <!-- Remove these sections for a feature request --> ### Error message: <!-- If you are reporting a segfault please include a GDB traceback, which you can generate by following https://github.com/numpy/numpy/blob/master/doc/source/dev/development_environment.rst#debugging --> <!-- Full error message, if any (starting from line Traceback: ...) --> ``` ======================================================= FAILURES =======================================================__________________________________________ TestLinspace.test_denormal_numbers __________________________________________ self = <numpy.core.tests.test_function_base.TestLinspace object at 0x7f70ac7be2e8> def test_denormal_numbers(self): # Regression test for gh-5437. Will probably fail when compiled # with ICC, which flushes denormals to zero for ftype in sctypes['float']: stop = nextafter(ftype(0), ftype(1)) * 5 # A denormal number > assert_(any(linspace(0, stop, 10, endpoint=False, dtype=ftype))) E AssertionError ftype = <class 'numpy.float128'> self = <numpy.core.tests.test_function_base.TestLinspace object at 0x7f70ac7be2e8> stop = 0.0 /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_function_base.py:308: AssertionError _____________________________________ TestFloatExceptions.test_floating_exceptions _____________________________________ self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550> def test_floating_exceptions(self): # Test basic arithmetic function errors with np.errstate(all='raise'): # Test for all real and complex float types for typecode in np.typecodes['AllFloat']: ftype = np.obj2sctype(typecode) if np.dtype(ftype).kind == 'f': # Get some extreme values for the type fi = np.finfo(ftype) ft_tiny = fi.tiny ft_max = fi.max ft_eps = fi.eps underflow = 'underflow' divbyzero = 'divide by zero' else: # 'c', complex, corresponding real dtype rtype = type(ftype(0).real) fi = np.finfo(rtype) ft_tiny = ftype(fi.tiny) ft_max = ftype(fi.max) ft_eps = ftype(fi.eps) # The complex types raise different exceptions underflow = '' divbyzero = '' overflow = 'overflow' invalid = 'invalid' self.assert_raises_fpe(underflow, > lambda a, b: a/b, ft_tiny, ft_max) divbyzero = 'divide by zero' fi = finfo(resolution=1.0000000000000002641e-18, min=-inf, max=inf, dtype=float128) ft_eps = 1.084202172485504434e-19 ft_max = inf ft_tiny = 3.3621031431120935063e-4932 ftype = <class 'numpy.float128'> invalid = 'invalid' overflow = 'overflow' self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550> typecode = 'g' underflow = 'underflow' /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_numeric.py:580: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>, fpeerr = 'underflow' flop = <function TestFloatExceptions.test_floating_exceptions.<locals>.<lambda> at 0x7f70aca5a730> x = 3.3621031431120935063e-4932, y = inf def assert_raises_fpe(self, fpeerr, flop, x, y): ftype = type(x) try: flop(x, y) assert_(False, > "Type %s did not raise fpe error '%s'." % (ftype, fpeerr)) E AssertionError: Type <class 'numpy.float128'> did not raise fpe error 'underflow'. flop = <function TestFloatExceptions.test_floating_exceptions.<locals>.<lambda> at 0x7f70aca5a730> fpeerr = 'underflow' ftype = <class 'numpy.float128'> self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550> x = 3.3621031431120935063e-4932 y = inf /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_numeric.py:535: AssertionError _______________________________ TestComplexFunctions.test_loss_of_precision_longcomplex ________________________________ self = <numpy.core.tests.test_umath.TestComplexFunctions object at 0x7f70ac0965c0> @pytest.mark.skipif(is_longdouble_finfo_bogus(), reason="Bogus long double finfo") def test_loss_of_precision_longcomplex(self): > self.check_loss_of_precision(np.longcomplex) self = <numpy.core.tests.test_umath.TestComplexFunctions object at 0x7f70ac0965c0> /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2642: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2592: in check_loss_of_precision check(x_series, 50*eps) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ x = array([1.00000000e-20, 1.21736864e-20, 1.48198641e-20, 1.80412378e-20, 2.19628372e-20, 2.67368693e-20, 3.254862...5156e-04, 4.54267386e-04, 5.53010871e-04, 6.73218092e-04, 8.19554595e-04, 9.97700064e-04], dtype=float128) rtol = 5.42101086242752217e-18 def check(x, rtol): x = x.astype(real_dtype) z = x.astype(dtype) d = np.absolute(np.arcsinh(x)/np.arcsinh(z).real - 1) assert_(np.all(d < rtol), (np.argmax(d), x[np.argmax(d)], d.max(), > 'arcsinh')) E AssertionError: (190, 0.00016990298344866943845, 4.4408920985006261617e-16, 'arcsinh') d = array([0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.000000...2302e-16, 0.00000000e+00, 3.33066907e-16, 0.00000000e+00, 2.22044605e-16, 1.11022302e-16], dtype=float128) dtype = <class 'numpy.complex256'> real_dtype = dtype('float128') rtol = 5.42101086242752217e-18 x = array([1.00000000e-20, 1.21736864e-20, 1.48198641e-20, 1.80412378e-20, 2.19628372e-20, 2.67368693e-20, 3.254862...5156e-04, 4.54267386e-04, 5.53010871e-04, 6.73218092e-04, 8.19554595e-04, 9.97700064e-04], dtype=float128) z = array([1.00000000e-20+0.j, 1.21736864e-20+0.j, 1.48198641e-20+0.j, 1.80412378e-20+0.j, 2.19628372e-20+0.j, 2.67...7386e-04+0.j, 5.53010871e-04+0.j, 6.73218092e-04+0.j, 8.19554595e-04+0.j, 9.97700064e-04+0.j], dtype=complex256) /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2565: AssertionError ___________________________________________________ test_nextafter_0 ___________________________________________________ def test_nextafter_0(): for t, direction in itertools.product(np.sctypes['float'], (1, -1)): tiny = np.finfo(t).tiny > assert_(0. < direction * np.nextafter(t(0), t(direction)) < tiny) E AssertionError direction = 1 t = <class 'numpy.float128'> tiny = 3.3621031431120935063e-4932 /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2776: AssertionError 4 failed, 5040 passed, 409 skipped, 167 deselected, 12 xfailed, 1 xpassed in 137.51 seconds ### Numpy/Python version information: <!-- Output from 'import sys, numpy; print(numpy.__version__, sys.version)' --> 1.16.0.dev0+db5f9d3 3.6.6 ```
1.0
4 test cases fail in test running with the latest master code base - <!-- Please describe the issue in detail here, and fill in the fields below --> ### Reproducing code example: <!-- A short code example that reproduces the problem/missing feature. It should be self-contained, i.e., possible to run as-is via 'python myproblem.py' --> ```python python -c 'import numpy as np; np.test()' ``` <!-- Remove these sections for a feature request --> ### Error message: <!-- If you are reporting a segfault please include a GDB traceback, which you can generate by following https://github.com/numpy/numpy/blob/master/doc/source/dev/development_environment.rst#debugging --> <!-- Full error message, if any (starting from line Traceback: ...) --> ``` ======================================================= FAILURES =======================================================__________________________________________ TestLinspace.test_denormal_numbers __________________________________________ self = <numpy.core.tests.test_function_base.TestLinspace object at 0x7f70ac7be2e8> def test_denormal_numbers(self): # Regression test for gh-5437. Will probably fail when compiled # with ICC, which flushes denormals to zero for ftype in sctypes['float']: stop = nextafter(ftype(0), ftype(1)) * 5 # A denormal number > assert_(any(linspace(0, stop, 10, endpoint=False, dtype=ftype))) E AssertionError ftype = <class 'numpy.float128'> self = <numpy.core.tests.test_function_base.TestLinspace object at 0x7f70ac7be2e8> stop = 0.0 /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_function_base.py:308: AssertionError _____________________________________ TestFloatExceptions.test_floating_exceptions _____________________________________ self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550> def test_floating_exceptions(self): # Test basic arithmetic function errors with np.errstate(all='raise'): # Test for all real and complex float types for typecode in np.typecodes['AllFloat']: ftype = np.obj2sctype(typecode) if np.dtype(ftype).kind == 'f': # Get some extreme values for the type fi = np.finfo(ftype) ft_tiny = fi.tiny ft_max = fi.max ft_eps = fi.eps underflow = 'underflow' divbyzero = 'divide by zero' else: # 'c', complex, corresponding real dtype rtype = type(ftype(0).real) fi = np.finfo(rtype) ft_tiny = ftype(fi.tiny) ft_max = ftype(fi.max) ft_eps = ftype(fi.eps) # The complex types raise different exceptions underflow = '' divbyzero = '' overflow = 'overflow' invalid = 'invalid' self.assert_raises_fpe(underflow, > lambda a, b: a/b, ft_tiny, ft_max) divbyzero = 'divide by zero' fi = finfo(resolution=1.0000000000000002641e-18, min=-inf, max=inf, dtype=float128) ft_eps = 1.084202172485504434e-19 ft_max = inf ft_tiny = 3.3621031431120935063e-4932 ftype = <class 'numpy.float128'> invalid = 'invalid' overflow = 'overflow' self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550> typecode = 'g' underflow = 'underflow' /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_numeric.py:580: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>, fpeerr = 'underflow' flop = <function TestFloatExceptions.test_floating_exceptions.<locals>.<lambda> at 0x7f70aca5a730> x = 3.3621031431120935063e-4932, y = inf def assert_raises_fpe(self, fpeerr, flop, x, y): ftype = type(x) try: flop(x, y) assert_(False, > "Type %s did not raise fpe error '%s'." % (ftype, fpeerr)) E AssertionError: Type <class 'numpy.float128'> did not raise fpe error 'underflow'. flop = <function TestFloatExceptions.test_floating_exceptions.<locals>.<lambda> at 0x7f70aca5a730> fpeerr = 'underflow' ftype = <class 'numpy.float128'> self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550> x = 3.3621031431120935063e-4932 y = inf /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_numeric.py:535: AssertionError _______________________________ TestComplexFunctions.test_loss_of_precision_longcomplex ________________________________ self = <numpy.core.tests.test_umath.TestComplexFunctions object at 0x7f70ac0965c0> @pytest.mark.skipif(is_longdouble_finfo_bogus(), reason="Bogus long double finfo") def test_loss_of_precision_longcomplex(self): > self.check_loss_of_precision(np.longcomplex) self = <numpy.core.tests.test_umath.TestComplexFunctions object at 0x7f70ac0965c0> /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2642: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2592: in check_loss_of_precision check(x_series, 50*eps) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ x = array([1.00000000e-20, 1.21736864e-20, 1.48198641e-20, 1.80412378e-20, 2.19628372e-20, 2.67368693e-20, 3.254862...5156e-04, 4.54267386e-04, 5.53010871e-04, 6.73218092e-04, 8.19554595e-04, 9.97700064e-04], dtype=float128) rtol = 5.42101086242752217e-18 def check(x, rtol): x = x.astype(real_dtype) z = x.astype(dtype) d = np.absolute(np.arcsinh(x)/np.arcsinh(z).real - 1) assert_(np.all(d < rtol), (np.argmax(d), x[np.argmax(d)], d.max(), > 'arcsinh')) E AssertionError: (190, 0.00016990298344866943845, 4.4408920985006261617e-16, 'arcsinh') d = array([0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.000000...2302e-16, 0.00000000e+00, 3.33066907e-16, 0.00000000e+00, 2.22044605e-16, 1.11022302e-16], dtype=float128) dtype = <class 'numpy.complex256'> real_dtype = dtype('float128') rtol = 5.42101086242752217e-18 x = array([1.00000000e-20, 1.21736864e-20, 1.48198641e-20, 1.80412378e-20, 2.19628372e-20, 2.67368693e-20, 3.254862...5156e-04, 4.54267386e-04, 5.53010871e-04, 6.73218092e-04, 8.19554595e-04, 9.97700064e-04], dtype=float128) z = array([1.00000000e-20+0.j, 1.21736864e-20+0.j, 1.48198641e-20+0.j, 1.80412378e-20+0.j, 2.19628372e-20+0.j, 2.67...7386e-04+0.j, 5.53010871e-04+0.j, 6.73218092e-04+0.j, 8.19554595e-04+0.j, 9.97700064e-04+0.j], dtype=complex256) /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2565: AssertionError ___________________________________________________ test_nextafter_0 ___________________________________________________ def test_nextafter_0(): for t, direction in itertools.product(np.sctypes['float'], (1, -1)): tiny = np.finfo(t).tiny > assert_(0. < direction * np.nextafter(t(0), t(direction)) < tiny) E AssertionError direction = 1 t = <class 'numpy.float128'> tiny = 3.3621031431120935063e-4932 /usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2776: AssertionError 4 failed, 5040 passed, 409 skipped, 167 deselected, 12 xfailed, 1 xpassed in 137.51 seconds ### Numpy/Python version information: <!-- Output from 'import sys, numpy; print(numpy.__version__, sys.version)' --> 1.16.0.dev0+db5f9d3 3.6.6 ```
non_priority
test cases fail in test running with the latest master code base reproducing code example a short code example that reproduces the problem missing feature it should be self contained i e possible to run as is via python myproblem py python python c import numpy as np np test error message if you are reporting a segfault please include a gdb traceback which you can generate by following failures testlinspace test denormal numbers self def test denormal numbers self regression test for gh will probably fail when compiled with icc which flushes denormals to zero for ftype in sctypes stop nextafter ftype ftype a denormal number assert any linspace stop endpoint false dtype ftype e assertionerror ftype self stop usr local lib dist packages numpy linux egg numpy core tests test function base py assertionerror testfloatexceptions test floating exceptions self def test floating exceptions self test basic arithmetic function errors with np errstate all raise test for all real and complex float types for typecode in np typecodes ftype np typecode if np dtype ftype kind f get some extreme values for the type fi np finfo ftype ft tiny fi tiny ft max fi max ft eps fi eps underflow underflow divbyzero divide by zero else c complex corresponding real dtype rtype type ftype real fi np finfo rtype ft tiny ftype fi tiny ft max ftype fi max ft eps ftype fi eps the complex types raise different exceptions underflow divbyzero overflow overflow invalid invalid self assert raises fpe underflow lambda a b a b ft tiny ft max divbyzero divide by zero fi finfo resolution min inf max inf dtype ft eps ft max inf ft tiny ftype invalid invalid overflow overflow self typecode g underflow underflow usr local lib dist packages numpy linux egg numpy core tests test numeric py self fpeerr underflow flop at x y inf def assert raises fpe self fpeerr flop x y ftype type x try flop x y assert false type s did not raise fpe error s ftype fpeerr e assertionerror type did not raise fpe error underflow flop at fpeerr underflow ftype self x y inf usr local lib dist packages numpy linux egg numpy core tests test numeric py assertionerror testcomplexfunctions test loss of precision longcomplex self pytest mark skipif is longdouble finfo bogus reason bogus long double finfo def test loss of precision longcomplex self self check loss of precision np longcomplex self usr local lib dist packages numpy linux egg numpy core tests test umath py usr local lib dist packages numpy linux egg numpy core tests test umath py in check loss of precision check x series eps x array dtype rtol def check x rtol x x astype real dtype z x astype dtype d np absolute np arcsinh x np arcsinh z real assert np all d rtol np argmax d x d max arcsinh e assertionerror arcsinh d array dtype dtype real dtype dtype rtol x array dtype z array j j j j j j j j j j dtype usr local lib dist packages numpy linux egg numpy core tests test umath py assertionerror test nextafter def test nextafter for t direction in itertools product np sctypes tiny np finfo t tiny assert direction np nextafter t t direction tiny e assertionerror direction t tiny usr local lib dist packages numpy linux egg numpy core tests test umath py assertionerror failed passed skipped deselected xfailed xpassed in seconds numpy python version information
0
35,384
17,037,982,218
IssuesEvent
2021-07-05 09:38:24
riboviz/riboviz
https://api.github.com/repos/riboviz/riboviz
opened
Parallelize count_reads.py
performance
The countReads processes (riboviz.count_reads.py) is running with only one core now, and is one of the most time consuming processes in the workflow. (For example, this process need 2h54m in Wallace dataset, while the execution time of the whole process is ~5h under 16 cores total, 8 core each sample) The core function of countReads is implemented in count_reads.py, function count_reads_df, which look like: ```python def count_reads_df(config_file, input_dir, tmp_dir, output_dir): df = pd.DataFrame(columns=HEADER) rows = [] rows.extend(input_fq(config_file, input_dir)) rows.append(cutadapt_fq(tmp_dir)) rows.extend(umi_tools_deplex_fq(tmp_dir)) tmp_samples = [f.name for f in os.scandir(tmp_dir) if f.is_dir()] tmp_samples.sort() for sample in tmp_samples: rows.append(cutadapt_fq(tmp_dir, sample)) rows.append(hisat2_fq(tmp_dir, sample, workflow_files.NON_RRNA_FQ, "rRNA or other contaminating reads removed by alignment to rRNA index files")) rows.append(hisat2_sam(tmp_dir, sample, workflow_files.RRNA_MAP_SAM, "Reads with rRNA and other contaminating reads removed by alignment to rRNA index files")) rows.append(hisat2_fq(tmp_dir, sample, workflow_files.UNALIGNED_FQ, "Unaligned reads removed by alignment of remaining reads to ORFs index files")) rows.append(hisat2_sam(tmp_dir, sample, workflow_files.ORF_MAP_SAM, "Reads aligned to ORFs index files")) rows.append(trim_5p_mismatch_sam(tmp_dir, sample)) rows.append(umi_tools_dedup_bam(tmp_dir, output_dir, sample)) rows = [row for row in rows if row is not None] df = df.append(rows) return df ``` We can see that the it runs different functions and add the result of each function into `rows`. This can be executed in parallel, by using the `multiprocessing` package, such as Pool and the apply_async to add tasks asynchronously. I tested it on my VM and it is working, but when I tried to run a full size dataset on eddie, I got an OOM error, which may be caused by #385 .
True
Parallelize count_reads.py - The countReads processes (riboviz.count_reads.py) is running with only one core now, and is one of the most time consuming processes in the workflow. (For example, this process need 2h54m in Wallace dataset, while the execution time of the whole process is ~5h under 16 cores total, 8 core each sample) The core function of countReads is implemented in count_reads.py, function count_reads_df, which look like: ```python def count_reads_df(config_file, input_dir, tmp_dir, output_dir): df = pd.DataFrame(columns=HEADER) rows = [] rows.extend(input_fq(config_file, input_dir)) rows.append(cutadapt_fq(tmp_dir)) rows.extend(umi_tools_deplex_fq(tmp_dir)) tmp_samples = [f.name for f in os.scandir(tmp_dir) if f.is_dir()] tmp_samples.sort() for sample in tmp_samples: rows.append(cutadapt_fq(tmp_dir, sample)) rows.append(hisat2_fq(tmp_dir, sample, workflow_files.NON_RRNA_FQ, "rRNA or other contaminating reads removed by alignment to rRNA index files")) rows.append(hisat2_sam(tmp_dir, sample, workflow_files.RRNA_MAP_SAM, "Reads with rRNA and other contaminating reads removed by alignment to rRNA index files")) rows.append(hisat2_fq(tmp_dir, sample, workflow_files.UNALIGNED_FQ, "Unaligned reads removed by alignment of remaining reads to ORFs index files")) rows.append(hisat2_sam(tmp_dir, sample, workflow_files.ORF_MAP_SAM, "Reads aligned to ORFs index files")) rows.append(trim_5p_mismatch_sam(tmp_dir, sample)) rows.append(umi_tools_dedup_bam(tmp_dir, output_dir, sample)) rows = [row for row in rows if row is not None] df = df.append(rows) return df ``` We can see that the it runs different functions and add the result of each function into `rows`. This can be executed in parallel, by using the `multiprocessing` package, such as Pool and the apply_async to add tasks asynchronously. I tested it on my VM and it is working, but when I tried to run a full size dataset on eddie, I got an OOM error, which may be caused by #385 .
non_priority
parallelize count reads py the countreads processes riboviz count reads py is running with only one core now and is one of the most time consuming processes in the workflow for example this process need in wallace dataset while the execution time of the whole process is under cores total core each sample the core function of countreads is implemented in count reads py function count reads df which look like python def count reads df config file input dir tmp dir output dir df pd dataframe columns header rows rows extend input fq config file input dir rows append cutadapt fq tmp dir rows extend umi tools deplex fq tmp dir tmp samples tmp samples sort for sample in tmp samples rows append cutadapt fq tmp dir sample rows append fq tmp dir sample workflow files non rrna fq rrna or other contaminating reads removed by alignment to rrna index files rows append sam tmp dir sample workflow files rrna map sam reads with rrna and other contaminating reads removed by alignment to rrna index files rows append fq tmp dir sample workflow files unaligned fq unaligned reads removed by alignment of remaining reads to orfs index files rows append sam tmp dir sample workflow files orf map sam reads aligned to orfs index files rows append trim mismatch sam tmp dir sample rows append umi tools dedup bam tmp dir output dir sample rows df df append rows return df we can see that the it runs different functions and add the result of each function into rows this can be executed in parallel by using the multiprocessing package such as pool and the apply async to add tasks asynchronously i tested it on my vm and it is working but when i tried to run a full size dataset on eddie i got an oom error which may be caused by
0
18,778
13,213,438,579
IssuesEvent
2020-08-16 12:48:03
textpattern/textpattern
https://api.github.com/repos/textpattern/textpattern
closed
'No styles recorded' message needed
usability
### Expected behaviour There should be a 'No styles recorded' message on the styles page panel like so when no styles are available for the current theme: ``` graf( span(null, array('class' => 'ui-icon ui-icon-info')).' '. gTxt('no_css_recorded'), array('class' => 'alert-block information') ); ``` ### Actual behaviour The left-hand column is blank, no message. This is inconsistent UI behaviour compared to the rest of the admin-side. ### Steps to reproduce 1. Select a theme. 2. Unassign styles from each section (i.e. style = none). 3. Go to styles panel and delete all styles. 4. See the empty left-hand column. #### Additional information Textpattern version: 4.8.3, 4.9 Once someone has coded this in, I will ensure the Textpacks are updated with this extra entry.
True
'No styles recorded' message needed - ### Expected behaviour There should be a 'No styles recorded' message on the styles page panel like so when no styles are available for the current theme: ``` graf( span(null, array('class' => 'ui-icon ui-icon-info')).' '. gTxt('no_css_recorded'), array('class' => 'alert-block information') ); ``` ### Actual behaviour The left-hand column is blank, no message. This is inconsistent UI behaviour compared to the rest of the admin-side. ### Steps to reproduce 1. Select a theme. 2. Unassign styles from each section (i.e. style = none). 3. Go to styles panel and delete all styles. 4. See the empty left-hand column. #### Additional information Textpattern version: 4.8.3, 4.9 Once someone has coded this in, I will ensure the Textpacks are updated with this extra entry.
non_priority
no styles recorded message needed expected behaviour there should be a no styles recorded message on the styles page panel like so when no styles are available for the current theme graf span null array class ui icon ui icon info gtxt no css recorded array class alert block information actual behaviour the left hand column is blank no message this is inconsistent ui behaviour compared to the rest of the admin side steps to reproduce select a theme unassign styles from each section i e style none go to styles panel and delete all styles see the empty left hand column additional information textpattern version once someone has coded this in i will ensure the textpacks are updated with this extra entry
0
290,155
21,867,854,876
IssuesEvent
2022-05-19 01:24:37
lukeed/cloudflare-docs
https://api.github.com/repos/lukeed/cloudflare-docs
opened
asdasd
documentation content:new
### Which Cloudflare product(s) does this pertain to? Analytics ### Subject Matter asd ### Content Location aasdasd ### Additional information _No response_
1.0
asdasd - ### Which Cloudflare product(s) does this pertain to? Analytics ### Subject Matter asd ### Content Location aasdasd ### Additional information _No response_
non_priority
asdasd which cloudflare product s does this pertain to analytics subject matter asd content location aasdasd additional information no response
0
426,112
29,509,836,653
IssuesEvent
2023-06-03 19:42:59
michaelNXT1/SoftwareEngineeringProject2023
https://api.github.com/repos/michaelNXT1/SoftwareEngineeringProject2023
closed
Architecture_update
documentation Priority: Low
הארכיטקטורה העדכנית של המערכת תכלול את הרכיבים הבאים: ● רכיב לקוח )client, presentation )עדכני. ● רכיב תקשורת )communication )ורכיב notifications עדכניים. ● רכיב שירות )service )עדכנית. ● רכיב לב-מערכת )domain )עדכני. ● שכבת קישור למסד נתונים )ה-ORM מהווה את את Layer Access Data.) ● מסד נתונים
1.0
Architecture_update - הארכיטקטורה העדכנית של המערכת תכלול את הרכיבים הבאים: ● רכיב לקוח )client, presentation )עדכני. ● רכיב תקשורת )communication )ורכיב notifications עדכניים. ● רכיב שירות )service )עדכנית. ● רכיב לב-מערכת )domain )עדכני. ● שכבת קישור למסד נתונים )ה-ORM מהווה את את Layer Access Data.) ● מסד נתונים
non_priority
architecture update הארכיטקטורה העדכנית של המערכת תכלול את הרכיבים הבאים ● רכיב לקוח client presentation עדכני ● רכיב תקשורת communication ורכיב notifications עדכניים ● רכיב שירות service עדכנית ● רכיב לב מערכת domain עדכני ● שכבת קישור למסד נתונים ה orm מהווה את את layer access data ● מסד נתונים
0
325,618
9,933,688,378
IssuesEvent
2019-07-02 12:56:12
juju/juju-gui
https://api.github.com/repos/juju/juju-gui
closed
Redirected to blank page when loading https://jujucharms.com while logged in
bug high priority
If I'm logged in to jujucharms.com and I try to open https://jujucharms.com in a new tab, I get redirected to https://jujucharms.com/u/cmars and then redirected to https://jujucharms.com/login/u/cmars, which is a blank page. I'm able to work around this by either: - Opening https://jujucharms.com in a private browsing window and logging in - Loading https://jujucharms.com/logout in a new tab and logging back in However, a short time after being logged in (10-15 minutes?) I get redirected to the same blank page. I am using Firefox 62.0.3 with uBlock disabled on jujucharms.com, so I don't think the issue is being caused by content blockers.
1.0
Redirected to blank page when loading https://jujucharms.com while logged in - If I'm logged in to jujucharms.com and I try to open https://jujucharms.com in a new tab, I get redirected to https://jujucharms.com/u/cmars and then redirected to https://jujucharms.com/login/u/cmars, which is a blank page. I'm able to work around this by either: - Opening https://jujucharms.com in a private browsing window and logging in - Loading https://jujucharms.com/logout in a new tab and logging back in However, a short time after being logged in (10-15 minutes?) I get redirected to the same blank page. I am using Firefox 62.0.3 with uBlock disabled on jujucharms.com, so I don't think the issue is being caused by content blockers.
priority
redirected to blank page when loading while logged in if i m logged in to jujucharms com and i try to open in a new tab i get redirected to and then redirected to which is a blank page i m able to work around this by either opening in a private browsing window and logging in loading in a new tab and logging back in however a short time after being logged in minutes i get redirected to the same blank page i am using firefox with ublock disabled on jujucharms com so i don t think the issue is being caused by content blockers
1
47,455
19,656,248,466
IssuesEvent
2022-01-10 12:50:22
PreMiD/Presences
https://api.github.com/repos/PreMiD/Presences
closed
EA Origin | origin.com
Service Request
### Discussed in https://github.com/PreMiD/Presences/discussions/4496 <div type='discussions-op-text'> <sup>Originally posted by **ghost** September 30, 2020</sup> **Prerequisites and essential questions** - [x] Is it a popular site? - [x] Is the website older than 2 months? - [ ] Is the site locked to a specific country/region? - [ ] Is the site a paid service? (e.g. Netflix, Hulu) - [ ] Does the website feature NSFW content? (e.g. porn, etc...) - [ ] Are you a donator/patron? - [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority? **What's your Discord username?** web#0385 **What's the name of the service?** www.origin.com I EA Origin **What should the Presence display?** What page the user is on, time elapsed. Nothing too complicated. **If possible, please provide a logo for the service (512x512 minimum)** <img src="https://i.imgur.com/M5aQ6b2.png" alt="presence" data-canonical-src="https://i.imgur.com/M5aQ6b2.png" style="max-width:100%;"> </div>
1.0
EA Origin | origin.com - ### Discussed in https://github.com/PreMiD/Presences/discussions/4496 <div type='discussions-op-text'> <sup>Originally posted by **ghost** September 30, 2020</sup> **Prerequisites and essential questions** - [x] Is it a popular site? - [x] Is the website older than 2 months? - [ ] Is the site locked to a specific country/region? - [ ] Is the site a paid service? (e.g. Netflix, Hulu) - [ ] Does the website feature NSFW content? (e.g. porn, etc...) - [ ] Are you a donator/patron? - [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority? **What's your Discord username?** web#0385 **What's the name of the service?** www.origin.com I EA Origin **What should the Presence display?** What page the user is on, time elapsed. Nothing too complicated. **If possible, please provide a logo for the service (512x512 minimum)** <img src="https://i.imgur.com/M5aQ6b2.png" alt="presence" data-canonical-src="https://i.imgur.com/M5aQ6b2.png" style="max-width:100%;"> </div>
non_priority
ea origin origin com discussed in originally posted by ghost september prerequisites and essential questions is it a popular site is the website older than months is the site locked to a specific country region is the site a paid service e g netflix hulu does the website feature nsfw content e g porn etc are you a donator patron do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority what s your discord username web what s the name of the service i ea origin what should the presence display what page the user is on time elapsed nothing too complicated if possible please provide a logo for the service minimum
0