Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1
value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3
values | title stringlengths 1 1k | labels stringlengths 4 1.38k | body stringlengths 1 262k | index stringclasses 16
values | text_combine stringlengths 96 262k | label stringclasses 2
values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
52,812 | 13,771,902,152 | IssuesEvent | 2020-10-07 23:04:32 | jgeraigery/please-del-this-repo | https://api.github.com/repos/jgeraigery/please-del-this-repo | opened | CVE-2017-16138 (High) detected in mime-1.3.4.tgz | security vulnerability | ## CVE-2017-16138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.3.4.tgz</b></p></summary>
<p>A comprehensive library for mime-type mapping</p>
<p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.3.4.tgz">https://registry.npmjs.org/mime/-/mime-1.3.4.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/mime/package.json</p>
<p>
Dependency Hierarchy:
- mailparser-0.6.2.tgz (Root Library)
- :x: **mime-1.3.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/please-del-this-repo/commit/e701d80cb459747be96e9c0e79c8640eca26c232">e701d80cb459747be96e9c0e79c8640eca26c232</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 1.4.1,2.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.3.4","isTransitiveDependency":true,"dependencyTree":"mailparser:0.6.2;mime:1.3.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-16138 (High) detected in mime-1.3.4.tgz - ## CVE-2017-16138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.3.4.tgz</b></p></summary>
<p>A comprehensive library for mime-type mapping</p>
<p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.3.4.tgz">https://registry.npmjs.org/mime/-/mime-1.3.4.tgz</a></p>
<p>Path to dependency file: please-del-this-repo/package.json</p>
<p>Path to vulnerable library: please-del-this-repo/node_modules/mime/package.json</p>
<p>
Dependency Hierarchy:
- mailparser-0.6.2.tgz (Root Library)
- :x: **mime-1.3.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/please-del-this-repo/commit/e701d80cb459747be96e9c0e79c8640eca26c232">e701d80cb459747be96e9c0e79c8640eca26c232</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 1.4.1,2.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.3.4","isTransitiveDependency":true,"dependencyTree":"mailparser:0.6.2;mime:1.3.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in mime tgz cve high severity vulnerability vulnerable library mime tgz a comprehensive library for mime type mapping library home page a href path to dependency file please del this repo package json path to vulnerable library please del this repo node modules mime package json dependency hierarchy mailparser tgz root library x mime tgz vulnerable library found in head commit a href found in base branch master vulnerability details the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input vulnerabilityurl | 0 |
71,198 | 18,518,853,908 | IssuesEvent | 2021-10-20 13:11:50 | astropy/astropy | https://api.github.com/repos/astropy/astropy | reopened | Add support to release aarch64 wheels | Feature Request Release build | Problem
---------
On aarch64, pip install astropy builds the wheels from source code and then install it. It requires user to have development environment installed on his system. also, it take some time to build the wheels than downloading and extracting the wheels from pypi.
Resolution
-----------
On aarch64, pip install astropy should download the wheels from pypi
@saimn, Please let me know your interest on releasing aarch64 wheels. I can help in this. | 1.0 | Add support to release aarch64 wheels - Problem
---------
On aarch64, pip install astropy builds the wheels from source code and then install it. It requires user to have development environment installed on his system. also, it take some time to build the wheels than downloading and extracting the wheels from pypi.
Resolution
-----------
On aarch64, pip install astropy should download the wheels from pypi
@saimn, Please let me know your interest on releasing aarch64 wheels. I can help in this. | non_priority | add support to release wheels problem on pip install astropy builds the wheels from source code and then install it it requires user to have development environment installed on his system also it take some time to build the wheels than downloading and extracting the wheels from pypi resolution on pip install astropy should download the wheels from pypi saimn please let me know your interest on releasing wheels i can help in this | 0 |
386,062 | 11,431,152,318 | IssuesEvent | 2020-02-04 11:31:31 | w3c/epubcheck | https://api.github.com/repos/w3c/epubcheck | closed | Testing Tools for EPUBCheck | priority: low status: in discussion type: feature | I have created a simple script for running a folder of files through EPUBCheck and outputting results in a single text file.
https://github.com/wareid/EPUB3-tests/tree/master/EPUBCheck%20Testing%20Tools
Would this be worthwhile including in the EPUBCheck repo for potential testers? Also, please let me know if there's other tools or changes I can make to improve this :). My personal version runs the same files against 2 versions of EPUBCheck for comparison. | 1.0 | Testing Tools for EPUBCheck - I have created a simple script for running a folder of files through EPUBCheck and outputting results in a single text file.
https://github.com/wareid/EPUB3-tests/tree/master/EPUBCheck%20Testing%20Tools
Would this be worthwhile including in the EPUBCheck repo for potential testers? Also, please let me know if there's other tools or changes I can make to improve this :). My personal version runs the same files against 2 versions of EPUBCheck for comparison. | priority | testing tools for epubcheck i have created a simple script for running a folder of files through epubcheck and outputting results in a single text file would this be worthwhile including in the epubcheck repo for potential testers also please let me know if there s other tools or changes i can make to improve this my personal version runs the same files against versions of epubcheck for comparison | 1 |
143,520 | 22,061,036,591 | IssuesEvent | 2022-05-30 17:52:44 | DXgovernance/dxvote | https://api.github.com/repos/DXgovernance/dxvote | closed | Add tenderly verification modal to add action flow | Enhancement Design | Guilds are able to do on-chain function calls which we call actions. We are going to use tenderly to simulate the defined actions to check if they are actually executable once a proposal has passed. Please go through the playground before you start working on this issue [HERE](https://www.figma.com/proto/cjLng0AbOKuDj9r5oZVDmR/Wireframe?page-id=4442%3A17030&node-id=4442%3A17031&viewport=325%2C48%2C0.1&scaling=min-zoom&starting-point-node-id=4442%3A17031). To make this step transparent we are gonna add a `Verify Action - Modal` to the `Add Action - Flow`:
### Verify Action

- Always appears once the user clicks on `Add Action - Button` in the add action process.
- Anytime a user edits, adds, removes an action we need to verify the whole package of actions via tenderly.
- The `Go to Tenderly - Button` is opening up a new tab with the actual tenderly simulation.
- [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5281%3A17505)
### Verified

- If tenderly's simulation passed successfully, the user can dismiss or close the modal with the close icon.
- User can still go to tenderly with the `Go to tenderly - Button`.
- After clicking dismiss user should see the defined action in the `Actions - Widget`
- [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5281%3A17836)
### Verification Failed

- If tenderly's simulation fails, the user can go back to the action modal, go to tenderly via the `Go to tenderly - Button` or close the modal with the close icon.
- We will not allow the user add an action which has not passed the simulation of tenderly.
- [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5353%3A17674)
| 1.0 | Add tenderly verification modal to add action flow - Guilds are able to do on-chain function calls which we call actions. We are going to use tenderly to simulate the defined actions to check if they are actually executable once a proposal has passed. Please go through the playground before you start working on this issue [HERE](https://www.figma.com/proto/cjLng0AbOKuDj9r5oZVDmR/Wireframe?page-id=4442%3A17030&node-id=4442%3A17031&viewport=325%2C48%2C0.1&scaling=min-zoom&starting-point-node-id=4442%3A17031). To make this step transparent we are gonna add a `Verify Action - Modal` to the `Add Action - Flow`:
### Verify Action

- Always appears once the user clicks on `Add Action - Button` in the add action process.
- Anytime a user edits, adds, removes an action we need to verify the whole package of actions via tenderly.
- The `Go to Tenderly - Button` is opening up a new tab with the actual tenderly simulation.
- [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5281%3A17505)
### Verified

- If tenderly's simulation passed successfully, the user can dismiss or close the modal with the close icon.
- User can still go to tenderly with the `Go to tenderly - Button`.
- After clicking dismiss user should see the defined action in the `Actions - Widget`
- [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5281%3A17836)
### Verification Failed

- If tenderly's simulation fails, the user can go back to the action modal, go to tenderly via the `Go to tenderly - Button` or close the modal with the close icon.
- We will not allow the user add an action which has not passed the simulation of tenderly.
- [Figma design specification](https://www.figma.com/file/cjLng0AbOKuDj9r5oZVDmR/Wireframe?node-id=5353%3A17674)
| non_priority | add tenderly verification modal to add action flow guilds are able to do on chain function calls which we call actions we are going to use tenderly to simulate the defined actions to check if they are actually executable once a proposal has passed please go through the playground before you start working on this issue to make this step transparent we are gonna add a verify action modal to the add action flow verify action always appears once the user clicks on add action button in the add action process anytime a user edits adds removes an action we need to verify the whole package of actions via tenderly the go to tenderly button is opening up a new tab with the actual tenderly simulation verified if tenderly s simulation passed successfully the user can dismiss or close the modal with the close icon user can still go to tenderly with the go to tenderly button after clicking dismiss user should see the defined action in the actions widget verification failed if tenderly s simulation fails the user can go back to the action modal go to tenderly via the go to tenderly button or close the modal with the close icon we will not allow the user add an action which has not passed the simulation of tenderly | 0 |
106,827 | 9,188,302,881 | IssuesEvent | 2019-03-06 06:55:06 | CARTAvis/carta-backend | https://api.github.com/repos/CARTAvis/carta-backend | closed | loading a file without specifying an HDU fails | bug ready to test | The frontend has a secondary method of loading files, via URL parameters. In this case, the HDU field will be left blank. However, this results in the file not loading, because the `casacore::string::fromString` call has the `chk` flag set to `true`, resulting in an exception being thrown.
The other instance of `fromString` being used (in `getFITShdu`) explicitly guards against empty strings. This should be done in `openFile` as well.
| 1.0 | loading a file without specifying an HDU fails - The frontend has a secondary method of loading files, via URL parameters. In this case, the HDU field will be left blank. However, this results in the file not loading, because the `casacore::string::fromString` call has the `chk` flag set to `true`, resulting in an exception being thrown.
The other instance of `fromString` being used (in `getFITShdu`) explicitly guards against empty strings. This should be done in `openFile` as well.
| non_priority | loading a file without specifying an hdu fails the frontend has a secondary method of loading files via url parameters in this case the hdu field will be left blank however this results in the file not loading because the casacore string fromstring call has the chk flag set to true resulting in an exception being thrown the other instance of fromstring being used in getfitshdu explicitly guards against empty strings this should be done in openfile as well | 0 |
485,837 | 14,000,209,192 | IssuesEvent | 2020-10-28 11:59:07 | ihhub/fheroes2 | https://api.github.com/repos/ihhub/fheroes2 | closed | Game screens, all "EVIL" windows are not centered vertically in the screen | Hacktoberfest good first issue low priority ui | In fheroes2, with the "EVIL" interface, all windows of this type are not centered vertically in the screen.
In fheroes2:









 | 1.0 | Game screens, all "EVIL" windows are not centered vertically in the screen - In fheroes2, with the "EVIL" interface, all windows of this type are not centered vertically in the screen.
In fheroes2:









 | priority | game screens all evil windows are not centered vertically in the screen in with the evil interface all windows of this type are not centered vertically in the screen in | 1 |
565,604 | 16,765,476,708 | IssuesEvent | 2021-06-14 08:17:48 | enso-org/ide | https://api.github.com/repos/enso-org/ide | opened | Undo Redo does not work with upload progress updates. | Category: Controllers Difficulty: Core Contributor Priority: High Type: Bug | ### Summary
During uploading file we update the node's metadata about progress to keep information that file has been uploading when IDE has been closed, and also the node's expression is updated when uploading finishes - because we use a special expression during upload to have correct error messages in dependent nodes (see also #
These operations should be transparent for the Undo Redo, because those are not user activities. However, with current UR architecture it is impossible, because each UR stack element have a full snapshot of AST - so even if the operations of uploading will be ignored, the other transactions committed during upload (user activity during upload) will have a trail of old metadata and node's expression.
### Value
<!--
- This section should describe the value of this task.
- This value can be for users, to the team, etc.
-->
### Specification
To be specified.
### Acceptance Criteria & Test Cases
<!--
- Any criteria that must be satisfied for the task to be accepted.
- The test plan for the feature, related to the acceptance criteria.
-->
| 1.0 | Undo Redo does not work with upload progress updates. - ### Summary
During uploading file we update the node's metadata about progress to keep information that file has been uploading when IDE has been closed, and also the node's expression is updated when uploading finishes - because we use a special expression during upload to have correct error messages in dependent nodes (see also #
These operations should be transparent for the Undo Redo, because those are not user activities. However, with current UR architecture it is impossible, because each UR stack element have a full snapshot of AST - so even if the operations of uploading will be ignored, the other transactions committed during upload (user activity during upload) will have a trail of old metadata and node's expression.
### Value
<!--
- This section should describe the value of this task.
- This value can be for users, to the team, etc.
-->
### Specification
To be specified.
### Acceptance Criteria & Test Cases
<!--
- Any criteria that must be satisfied for the task to be accepted.
- The test plan for the feature, related to the acceptance criteria.
-->
| priority | undo redo does not work with upload progress updates summary during uploading file we update the node s metadata about progress to keep information that file has been uploading when ide has been closed and also the node s expression is updated when uploading finishes because we use a special expression during upload to have correct error messages in dependent nodes see also these operations should be transparent for the undo redo because those are not user activities however with current ur architecture it is impossible because each ur stack element have a full snapshot of ast so even if the operations of uploading will be ignored the other transactions committed during upload user activity during upload will have a trail of old metadata and node s expression value this section should describe the value of this task this value can be for users to the team etc specification to be specified acceptance criteria test cases any criteria that must be satisfied for the task to be accepted the test plan for the feature related to the acceptance criteria | 1 |
396,598 | 11,711,190,704 | IssuesEvent | 2020-03-09 04:06:45 | aitorres/firelink | https://api.github.com/repos/aitorres/firelink | closed | Fix union access verification | medium priority | Union property access verification is being triggered on assignments (while unions are LVALs). They should ONLY be triggered on RVAL accesses, or particular LVAL accesses. Actually, a probable fix is to ONLY trigger them while parsing an Access expression. (maybe). | 1.0 | Fix union access verification - Union property access verification is being triggered on assignments (while unions are LVALs). They should ONLY be triggered on RVAL accesses, or particular LVAL accesses. Actually, a probable fix is to ONLY trigger them while parsing an Access expression. (maybe). | priority | fix union access verification union property access verification is being triggered on assignments while unions are lvals they should only be triggered on rval accesses or particular lval accesses actually a probable fix is to only trigger them while parsing an access expression maybe | 1 |
248,228 | 7,928,294,213 | IssuesEvent | 2018-07-06 11:05:49 | aharwood2/JBlock2D | https://api.github.com/repos/aharwood2/JBlock2D | closed | Error when trying to use larger batch file | bug high priority | Program won't run when a larger batch file of the same format is used | 1.0 | Error when trying to use larger batch file - Program won't run when a larger batch file of the same format is used | priority | error when trying to use larger batch file program won t run when a larger batch file of the same format is used | 1 |
65,128 | 14,708,442,645 | IssuesEvent | 2021-01-04 23:43:14 | SmartBear/readyapi-swagger-assertion-plugin | https://api.github.com/repos/SmartBear/readyapi-swagger-assertion-plugin | opened | CVE-2017-1000208 (High) detected in swagger-parser-1.0.8.jar | security vulnerability | ## CVE-2017-1000208 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>swagger-parser-1.0.8.jar</b></p></summary>
<p>Sonatype helps open source projects to set up Maven repositories on https://oss.sonatype.org/</p>
<p>Library home page: <a href="http://nexus.sonatype.org/oss-repository-hosting.html/swagger-parser-project/modules/swagger-parser">http://nexus.sonatype.org/oss-repository-hosting.html/swagger-parser-project/modules/swagger-parser</a></p>
<p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/swagger/swagger-parser/1.0.8/swagger-parser-1.0.8.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.7.0.jar (Root Library)
- :x: **swagger-parser-1.0.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability in Swagger-Parser's (version <= 1.0.30) yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml Open-API specification is parsed. This in particular, affects the 'generate' and 'validate' command in swagger-codegen (<= 2.2.2) and can lead to arbitrary code being executed when these commands are used on a well-crafted yaml specification.
<p>Publish Date: 2017-11-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000208>CVE-2017-1000208</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000208">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000208</a></p>
<p>Release Date: 2017-11-17</p>
<p>Fix Resolution: io.swagger:swagger-parser:1.0.31</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.swagger","packageName":"swagger-parser","packageVersion":"1.0.8","isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;io.swagger:swagger-parser:1.0.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.swagger:swagger-parser:1.0.31"}],"vulnerabilityIdentifier":"CVE-2017-1000208","vulnerabilityDetails":"A vulnerability in Swagger-Parser\u0027s (version \u003c\u003d 1.0.30) yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml Open-API specification is parsed. This in particular, affects the \u0027generate\u0027 and \u0027validate\u0027 command in swagger-codegen (\u003c\u003d 2.2.2) and can lead to arbitrary code being executed when these commands are used on a well-crafted yaml specification.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000208","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-1000208 (High) detected in swagger-parser-1.0.8.jar - ## CVE-2017-1000208 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>swagger-parser-1.0.8.jar</b></p></summary>
<p>Sonatype helps open source projects to set up Maven repositories on https://oss.sonatype.org/</p>
<p>Library home page: <a href="http://nexus.sonatype.org/oss-repository-hosting.html/swagger-parser-project/modules/swagger-parser">http://nexus.sonatype.org/oss-repository-hosting.html/swagger-parser-project/modules/swagger-parser</a></p>
<p>Path to dependency file: readyapi-swagger-assertion-plugin/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/io/swagger/swagger-parser/1.0.8/swagger-parser-1.0.8.jar</p>
<p>
Dependency Hierarchy:
- ready-api-soapui-pro-1.7.0.jar (Root Library)
- :x: **swagger-parser-1.0.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SmartBear/readyapi-swagger-assertion-plugin/commit/dc701947b2244c9b5dff8c9a75808fab2d107d60">dc701947b2244c9b5dff8c9a75808fab2d107d60</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability in Swagger-Parser's (version <= 1.0.30) yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml Open-API specification is parsed. This in particular, affects the 'generate' and 'validate' command in swagger-codegen (<= 2.2.2) and can lead to arbitrary code being executed when these commands are used on a well-crafted yaml specification.
<p>Publish Date: 2017-11-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000208>CVE-2017-1000208</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000208">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-1000208</a></p>
<p>Release Date: 2017-11-17</p>
<p>Fix Resolution: io.swagger:swagger-parser:1.0.31</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"io.swagger","packageName":"swagger-parser","packageVersion":"1.0.8","isTransitiveDependency":true,"dependencyTree":"com.smartbear:ready-api-soapui-pro:1.7.0;io.swagger:swagger-parser:1.0.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"io.swagger:swagger-parser:1.0.31"}],"vulnerabilityIdentifier":"CVE-2017-1000208","vulnerabilityDetails":"A vulnerability in Swagger-Parser\u0027s (version \u003c\u003d 1.0.30) yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml Open-API specification is parsed. This in particular, affects the \u0027generate\u0027 and \u0027validate\u0027 command in swagger-codegen (\u003c\u003d 2.2.2) and can lead to arbitrary code being executed when these commands are used on a well-crafted yaml specification.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-1000208","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in swagger parser jar cve high severity vulnerability vulnerable library swagger parser jar sonatype helps open source projects to set up maven repositories on library home page a href path to dependency file readyapi swagger assertion plugin pom xml path to vulnerable library home wss scanner repository io swagger swagger parser swagger parser jar dependency hierarchy ready api soapui pro jar root library x swagger parser jar vulnerable library found in head commit a href found in base branch master vulnerability details a vulnerability in swagger parser s version yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml open api specification is parsed this in particular affects the generate and validate command in swagger codegen and can lead to arbitrary code being executed when these commands are used on a well crafted yaml specification publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io swagger swagger parser isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a vulnerability in swagger parser version yaml parsing functionality results in arbitrary code being executed when a maliciously crafted yaml open api specification is parsed this in particular affects the and command in swagger codegen and can lead to arbitrary code being executed when these commands are used on a well crafted yaml specification vulnerabilityurl | 0 |
229,434 | 25,343,418,115 | IssuesEvent | 2022-11-19 01:02:28 | MidnightBSD/src | https://api.github.com/repos/MidnightBSD/src | closed | CVE-2022-41916 (High) detected in heimdalheimdal-7.7.0 - autoclosed | security vulnerability | ## CVE-2022-41916 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>heimdalheimdal-7.7.0</b></p></summary>
<p>
<p>Heimdal</p>
<p>Library home page: <a href=https://github.com/heimdal/heimdal.git>https://github.com/heimdal/heimdal.git</a></p>
<p>Found in base branches: <b>stable/2.1, stable/2.2, master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/heimdal/lib/wind/normalize.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Heimdal is an implementation of ASN.1/DER, PKIX, and Kerberos. Versions prior to 7.7.1 are vulnerable to a denial of service vulnerability in Heimdal's PKI certificate validation library, affecting the KDC (via PKINIT) and kinit (via PKINIT), as well as any third-party applications using Heimdal's libhx509. Users should upgrade to Heimdal 7.7.1 or 7.8. There are no known workarounds for this issue.
<p>Publish Date: 2022-11-15
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41916>CVE-2022-41916</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/heimdal/heimdal/security/advisories/GHSA-mgqr-gvh6-23cx">https://github.com/heimdal/heimdal/security/advisories/GHSA-mgqr-gvh6-23cx</a></p>
<p>Release Date: 2022-11-15</p>
<p>Fix Resolution: heimdal-7.7.1
</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-41916 (High) detected in heimdalheimdal-7.7.0 - autoclosed - ## CVE-2022-41916 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>heimdalheimdal-7.7.0</b></p></summary>
<p>
<p>Heimdal</p>
<p>Library home page: <a href=https://github.com/heimdal/heimdal.git>https://github.com/heimdal/heimdal.git</a></p>
<p>Found in base branches: <b>stable/2.1, stable/2.2, master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/crypto/heimdal/lib/wind/normalize.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Heimdal is an implementation of ASN.1/DER, PKIX, and Kerberos. Versions prior to 7.7.1 are vulnerable to a denial of service vulnerability in Heimdal's PKI certificate validation library, affecting the KDC (via PKINIT) and kinit (via PKINIT), as well as any third-party applications using Heimdal's libhx509. Users should upgrade to Heimdal 7.7.1 or 7.8. There are no known workarounds for this issue.
<p>Publish Date: 2022-11-15
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-41916>CVE-2022-41916</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/heimdal/heimdal/security/advisories/GHSA-mgqr-gvh6-23cx">https://github.com/heimdal/heimdal/security/advisories/GHSA-mgqr-gvh6-23cx</a></p>
<p>Release Date: 2022-11-15</p>
<p>Fix Resolution: heimdal-7.7.1
</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in heimdalheimdal autoclosed cve high severity vulnerability vulnerable library heimdalheimdal heimdal library home page a href found in base branches stable stable master vulnerable source files crypto heimdal lib wind normalize c vulnerability details heimdal is an implementation of asn der pkix and kerberos versions prior to are vulnerable to a denial of service vulnerability in heimdal s pki certificate validation library affecting the kdc via pkinit and kinit via pkinit as well as any third party applications using heimdal s users should upgrade to heimdal or there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution heimdal step up your open source security game with mend | 0 |
128,133 | 10,516,402,570 | IssuesEvent | 2019-09-28 17:17:26 | Vachok/ftpplus | https://api.github.com/repos/Vachok/ftpplus | closed | testSyncData [D172] | Medium TestQuality bug resolution_Dupe | Execute DBUploadUniversalTest::testSyncData\*\*testSyncData\*\* \*DBUploadUniversalTest\* \*expected [0 rows uploaded to test.test] but found [2 rows uploaded to test.test]\* \*java.lang.AssertionError\* | 1.0 | testSyncData [D172] - Execute DBUploadUniversalTest::testSyncData\*\*testSyncData\*\* \*DBUploadUniversalTest\* \*expected [0 rows uploaded to test.test] but found [2 rows uploaded to test.test]\* \*java.lang.AssertionError\* | non_priority | testsyncdata execute dbuploaduniversaltest testsyncdata testsyncdata dbuploaduniversaltest expected but found java lang assertionerror | 0 |
549 | 2,499,784,057 | IssuesEvent | 2015-01-08 05:48:51 | chessmasterhong/WaterEmblem | https://api.github.com/repos/chessmasterhong/WaterEmblem | closed | Bug with ending battle early | bug high priority | If the player kills the last attacking enemy (The enemy initiates the attack on the enemy phase, and is the last enemy on the map) and the attacking enemy had a droppable item (and if the map condition is to kill all enemies), the map will fade out before the player can choose what item to discard. | 1.0 | Bug with ending battle early - If the player kills the last attacking enemy (The enemy initiates the attack on the enemy phase, and is the last enemy on the map) and the attacking enemy had a droppable item (and if the map condition is to kill all enemies), the map will fade out before the player can choose what item to discard. | priority | bug with ending battle early if the player kills the last attacking enemy the enemy initiates the attack on the enemy phase and is the last enemy on the map and the attacking enemy had a droppable item and if the map condition is to kill all enemies the map will fade out before the player can choose what item to discard | 1 |
514,056 | 14,932,336,628 | IssuesEvent | 2021-01-25 07:34:22 | sct/overseerr | https://api.github.com/repos/sct/overseerr | closed | Use the Plex API to pre-populate the server information | enhancement priority:medium | **Is your feature request related to a problem? Please describe.**
Plex.tv has an API call that can be made to get a list of servers a user has access to. Pulling this list would make the initial setup for users much easier since it is the same method the players use to discover the server.
**Describe the solution you'd like**
Use the Plex API to pre-populate the Plex server information.
Additional details can be found here.
https://support.plex.tv/articles/206721658-using-plex-tv-resources-information-to-troubleshoot-app-connections/
| 1.0 | Use the Plex API to pre-populate the server information - **Is your feature request related to a problem? Please describe.**
Plex.tv has an API call that can be made to get a list of servers a user has access to. Pulling this list would make the initial setup for users much easier since it is the same method the players use to discover the server.
**Describe the solution you'd like**
Use the Plex API to pre-populate the Plex server information.
Additional details can be found here.
https://support.plex.tv/articles/206721658-using-plex-tv-resources-information-to-troubleshoot-app-connections/
| priority | use the plex api to pre populate the server information is your feature request related to a problem please describe plex tv has an api call that can be made to get a list of servers a user has access to pulling this list would make the initial setup for users much easier since it is the same method the players use to discover the server describe the solution you d like use the plex api to pre populate the plex server information additional details can be found here | 1 |
4,875 | 7,753,598,812 | IssuesEvent | 2018-05-31 01:38:07 | kubernetes-sigs/kubebuilder | https://api.github.com/repos/kubernetes-sigs/kubebuilder | closed | Create a SECURITY_CONTACTS file. | process | As per the email sent to kubernetes-dev[1], please create a SECURITY_CONTACTS
file.
The template for the file can be found in the kubernetes-template repository[2].
A description for the file is in the steering-committee docs[3], you might need
to search that page for "Security Contacts".
Please feel free to ping me on the PR when you make it, otherwise I will see when
you close this issue. :)
Thanks so much, let me know if you have any questions.
(This issue was generated from a tool, apologies for any weirdness.)
[1] https://groups.google.com/forum/#!topic/kubernetes-dev/codeiIoQ6QE
[2] https://github.com/kubernetes/kubernetes-template-project/blob/master/SECURITY_CONTACTS
[3] https://github.com/kubernetes/community/blob/master/committee-steering/governance/sig-governance-template-short.md
| 1.0 | Create a SECURITY_CONTACTS file. - As per the email sent to kubernetes-dev[1], please create a SECURITY_CONTACTS
file.
The template for the file can be found in the kubernetes-template repository[2].
A description for the file is in the steering-committee docs[3], you might need
to search that page for "Security Contacts".
Please feel free to ping me on the PR when you make it, otherwise I will see when
you close this issue. :)
Thanks so much, let me know if you have any questions.
(This issue was generated from a tool, apologies for any weirdness.)
[1] https://groups.google.com/forum/#!topic/kubernetes-dev/codeiIoQ6QE
[2] https://github.com/kubernetes/kubernetes-template-project/blob/master/SECURITY_CONTACTS
[3] https://github.com/kubernetes/community/blob/master/committee-steering/governance/sig-governance-template-short.md
| non_priority | create a security contacts file as per the email sent to kubernetes dev please create a security contacts file the template for the file can be found in the kubernetes template repository a description for the file is in the steering committee docs you might need to search that page for security contacts please feel free to ping me on the pr when you make it otherwise i will see when you close this issue thanks so much let me know if you have any questions this issue was generated from a tool apologies for any weirdness | 0 |
398,202 | 11,739,246,193 | IssuesEvent | 2020-03-11 17:22:41 | georchestra/mapstore2-georchestra | https://api.github.com/repos/georchestra/mapstore2-georchestra | closed | Documentation extension | Accepted Priority: Low | An extension must be provided to reach the online documentation related to the viewer and its installed extensions.
- The URL of the documentation must be parametrizable by the administrator in backoffice.

| 1.0 | Documentation extension - An extension must be provided to reach the online documentation related to the viewer and its installed extensions.
- The URL of the documentation must be parametrizable by the administrator in backoffice.

| priority | documentation extension an extension must be provided to reach the online documentation related to the viewer and its installed extensions the url of the documentation must be parametrizable by the administrator in backoffice | 1 |
235,963 | 19,474,869,774 | IssuesEvent | 2021-12-24 10:09:24 | optuna/optuna | https://api.github.com/repos/optuna/optuna | opened | Unify the tests between Plotly and matplotlib | test feature v3 | <!-- Please write a clear and concise description of the feature proposal. -->
## Motivation
This is a sub-issue of https://github.com/optuna/optuna/issues/2959.
Tests cases are inconsistent between two visualisation backends: Plotly or Matplotlib. Due to this inconsistency, this might cause an expected bug. In addition, we find it difficult to maintain asymmetric tests.
## Description
Please unify the test cases between both functions. For example, if either test file does not perform a test that has been done by the other, please add a test case. Note that the missing test case might be function level or fine-grained level.
<!-- Please write a detailed description of the new feature. -->
## Alternatives (optional)
<!-- Please write a clear and concise description of any alternative solutions or features you've considered. -->
## Additional context (optional)
<!-- Please add any other context or screenshots about the feature request here. -->
| 1.0 | Unify the tests between Plotly and matplotlib - <!-- Please write a clear and concise description of the feature proposal. -->
## Motivation
This is a sub-issue of https://github.com/optuna/optuna/issues/2959.
Tests cases are inconsistent between two visualisation backends: Plotly or Matplotlib. Due to this inconsistency, this might cause an expected bug. In addition, we find it difficult to maintain asymmetric tests.
## Description
Please unify the test cases between both functions. For example, if either test file does not perform a test that has been done by the other, please add a test case. Note that the missing test case might be function level or fine-grained level.
<!-- Please write a detailed description of the new feature. -->
## Alternatives (optional)
<!-- Please write a clear and concise description of any alternative solutions or features you've considered. -->
## Additional context (optional)
<!-- Please add any other context or screenshots about the feature request here. -->
| non_priority | unify the tests between plotly and matplotlib motivation this is a sub issue of tests cases are inconsistent between two visualisation backends plotly or matplotlib due to this inconsistency this might cause an expected bug in addition we find it difficult to maintain asymmetric tests description please unify the test cases between both functions for example if either test file does not perform a test that has been done by the other please add a test case note that the missing test case might be function level or fine grained level alternatives optional additional context optional | 0 |
125,768 | 10,352,818,652 | IssuesEvent | 2019-09-05 10:03:58 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | opened | Nothing happens after clicking 'Create Disk' button | :gear: disks 🧪 testing | **Storage Explorer Version**: 1.10.0
**Build:** 20190905.11
**Branch:** master
**Platform/OS:** Windows 10/Linux Ubuntu 18.04/macOS High Sierra
**Architecture:** ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Expand 'Disks' -> Open one resource group.
2. Click 'Create Disk' button on the toolbar.
**Expect Experience:**
The 'Create Disk' dialog displays.
**Actual Experience:**
Nothing happens.
**More Info:**
This issue also reproduces for 'Paste' button.
| 1.0 | Nothing happens after clicking 'Create Disk' button - **Storage Explorer Version**: 1.10.0
**Build:** 20190905.11
**Branch:** master
**Platform/OS:** Windows 10/Linux Ubuntu 18.04/macOS High Sierra
**Architecture:** ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Expand 'Disks' -> Open one resource group.
2. Click 'Create Disk' button on the toolbar.
**Expect Experience:**
The 'Create Disk' dialog displays.
**Actual Experience:**
Nothing happens.
**More Info:**
This issue also reproduces for 'Paste' button.
| non_priority | nothing happens after clicking create disk button storage explorer version build branch master platform os windows linux ubuntu macos high sierra architecture regression from not a regression steps to reproduce expand disks open one resource group click create disk button on the toolbar expect experience the create disk dialog displays actual experience nothing happens more info this issue also reproduces for paste button | 0 |
752,138 | 26,274,752,884 | IssuesEvent | 2023-01-06 20:41:54 | zowe/imperative | https://api.github.com/repos/zowe/imperative | closed | Plugin validation fails when using chained handlers | bug priority-low severity-high | Plugin validation fails for command with chainedHandlers defined with message "CmdError: Command name = '***' has no 'handler' property", resulting in plugin commands not available. | 1.0 | Plugin validation fails when using chained handlers - Plugin validation fails for command with chainedHandlers defined with message "CmdError: Command name = '***' has no 'handler' property", resulting in plugin commands not available. | priority | plugin validation fails when using chained handlers plugin validation fails for command with chainedhandlers defined with message cmderror command name has no handler property resulting in plugin commands not available | 1 |
154,751 | 19,758,348,713 | IssuesEvent | 2022-01-16 01:18:40 | lemurchop/pancake-frontend | https://api.github.com/repos/lemurchop/pancake-frontend | opened | CVE-2022-0155 (High) detected in follow-redirects-1.14.3.tgz, follow-redirects-1.14.1.tgz | security vulnerability | ## CVE-2022-0155 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>follow-redirects-1.14.3.tgz</b>, <b>follow-redirects-1.14.1.tgz</b></p></summary>
<p>
<details><summary><b>follow-redirects-1.14.3.tgz</b></p></summary>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.3.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/axios/node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- start-server-and-test-1.14.0.tgz (Root Library)
- wait-on-6.0.0.tgz
- axios-0.21.3.tgz
- :x: **follow-redirects-1.14.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>follow-redirects-1.14.1.tgz</b></p></summary>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.1.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- webpack-dev-server-3.11.1.tgz
- http-proxy-middleware-0.19.1.tgz
- http-proxy-1.18.1.tgz
- :x: **follow-redirects-1.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
follow-redirects is vulnerable to Exposure of Private Personal Information to an Unauthorized Actor
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155>CVE-2022-0155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/">https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/</a></p>
<p>Release Date: 2022-01-10</p>
<p>Fix Resolution: follow-redirects - v1.14.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-0155 (High) detected in follow-redirects-1.14.3.tgz, follow-redirects-1.14.1.tgz - ## CVE-2022-0155 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>follow-redirects-1.14.3.tgz</b>, <b>follow-redirects-1.14.1.tgz</b></p></summary>
<p>
<details><summary><b>follow-redirects-1.14.3.tgz</b></p></summary>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.3.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.3.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/axios/node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- start-server-and-test-1.14.0.tgz (Root Library)
- wait-on-6.0.0.tgz
- axios-0.21.3.tgz
- :x: **follow-redirects-1.14.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>follow-redirects-1.14.1.tgz</b></p></summary>
<p>HTTP and HTTPS modules that follow redirects.</p>
<p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.1.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.14.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/follow-redirects/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.3.tgz (Root Library)
- webpack-dev-server-3.11.1.tgz
- http-proxy-middleware-0.19.1.tgz
- http-proxy-1.18.1.tgz
- :x: **follow-redirects-1.14.1.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
follow-redirects is vulnerable to Exposure of Private Personal Information to an Unauthorized Actor
<p>Publish Date: 2022-01-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155>CVE-2022-0155</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/">https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/</a></p>
<p>Release Date: 2022-01-10</p>
<p>Fix Resolution: follow-redirects - v1.14.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in follow redirects tgz follow redirects tgz cve high severity vulnerability vulnerable libraries follow redirects tgz follow redirects tgz follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file package json path to vulnerable library node modules axios node modules follow redirects package json dependency hierarchy start server and test tgz root library wait on tgz axios tgz x follow redirects tgz vulnerable library follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file package json path to vulnerable library node modules follow redirects package json dependency hierarchy react scripts tgz root library webpack dev server tgz http proxy middleware tgz http proxy tgz x follow redirects tgz vulnerable library found in base branch develop vulnerability details follow redirects is vulnerable to exposure of private personal information to an unauthorized actor publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution follow redirects step up your open source security game with whitesource | 0 |
736,634 | 25,481,185,446 | IssuesEvent | 2022-11-25 21:19:38 | diffgram/diffgram | https://api.github.com/repos/diffgram/diffgram | closed | Helm timeout | lowpriority helm | Add timeout parameter documentation for cases where the index creation db migration is too big. | 1.0 | Helm timeout - Add timeout parameter documentation for cases where the index creation db migration is too big. | priority | helm timeout add timeout parameter documentation for cases where the index creation db migration is too big | 1 |
263,939 | 8,303,411,949 | IssuesEvent | 2018-09-21 17:29:11 | zephyrproject-rtos/zephyr | https://api.github.com/repos/zephyrproject-rtos/zephyr | closed | nRF52: MPU Fault issue | bug nRF priority: high | Hi,
This **zephyr/samples/boards/nrf52/mesh/onoff_level_lighting_vnd_app** App in latest master branch
works perfectly normal with Zephyr v1.12.99 (with last commit ba6763a187a347cfc825a2bece78e7d1ef28772d).
But with latest master branch or after v1.13 onward, I am facing issue of MPU FAULT intermittently.
If we set "LIGHT_CTL_TT" in publisher.c & configure buttons to publish Light CTL set (acknowledged) messages then it get easily encountered while playing with on-boards buttons on #nRF52840_PDK boards. | 1.0 | nRF52: MPU Fault issue - Hi,
This **zephyr/samples/boards/nrf52/mesh/onoff_level_lighting_vnd_app** App in latest master branch
works perfectly normal with Zephyr v1.12.99 (with last commit ba6763a187a347cfc825a2bece78e7d1ef28772d).
But with latest master branch or after v1.13 onward, I am facing issue of MPU FAULT intermittently.
If we set "LIGHT_CTL_TT" in publisher.c & configure buttons to publish Light CTL set (acknowledged) messages then it get easily encountered while playing with on-boards buttons on #nRF52840_PDK boards. | priority | mpu fault issue hi this zephyr samples boards mesh onoff level lighting vnd app app in latest master branch works perfectly normal with zephyr with last commit but with latest master branch or after onward i am facing issue of mpu fault intermittently if we set light ctl tt in publisher c configure buttons to publish light ctl set acknowledged messages then it get easily encountered while playing with on boards buttons on pdk boards | 1 |
331,900 | 29,170,244,980 | IssuesEvent | 2023-05-19 00:37:40 | ray-project/ray | https://api.github.com/repos/ray-project/ray | closed | [Ray release glue.py] `Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.` | P1 testing | Seeing this in the following release tests:
* dask_on_ray_1tb_sort https://buildkite.com/ray-project/release-tests-branch/builds/1304#0185d156-acdf-4b5c-b63f-9fc4f6b2f7c2
* rllib_stress_tests https://buildkite.com/ray-project/release-tests-branch/builds/1305#0185d81d-437e-49ff-9103-4bf92a4d11ea
* shuffle_1tb_5000_partitions https://buildkite.com/ray-project/release-tests-branch/builds/1304#0185d156-ace6-4f9a-8255-25901c91daa7
* map_batches_benchmark_single_node https://buildkite.com/ray-project/release-tests-branch/builds/1333#0186130c-2057-4ecf-8be2-25bc4159fade
```
[ERROR 2023-01-21 22:46:41,582] run_release_test.py: 157 Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.
--
| Traceback (most recent call last):
| File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 69, in prepare_remote_env
| self.file_manager.upload()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 122, in upload
| self._push_local_dir()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 107, in _push_local_dir
| f"Error uploading local dir to session "
| ray_release.exception.FileUploadError: Error uploading local dir to session rllib_stress_tests_1674368507.
|
| The above exception was the direct cause of the following exception:
|
| Traceback (most recent call last):
| File "ray_release/scripts/run_release_test.py", line 153, in main
| no_terminate=no_terminate,
| File "/tmp/release-4K7NOSGKrN/release/ray_release/glue.py", line 404, in run_release_test
| raise pipeline_exception
| File "/tmp/release-4K7NOSGKrN/release/ray_release/glue.py", line 275, in run_release_test
| command_runner.prepare_remote_env()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 73, in prepare_remote_env
| ) from e
| ray_release.exception.RemoteEnvSetupError: Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.
| ERROR:ray_release.logger:Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.
| Traceback (most recent call last):
| File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 69, in prepare_remote_env
| self.file_manager.upload()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 122, in upload
| self._push_local_dir()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 107, in _push_local_dir
| f"Error uploading local dir to session "
| ray_release.exception.FileUploadError: Error uploading local dir to session rllib_stress_tests_1674368507.
```
cc @krfricke | 1.0 | [Ray release glue.py] `Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.` - Seeing this in the following release tests:
* dask_on_ray_1tb_sort https://buildkite.com/ray-project/release-tests-branch/builds/1304#0185d156-acdf-4b5c-b63f-9fc4f6b2f7c2
* rllib_stress_tests https://buildkite.com/ray-project/release-tests-branch/builds/1305#0185d81d-437e-49ff-9103-4bf92a4d11ea
* shuffle_1tb_5000_partitions https://buildkite.com/ray-project/release-tests-branch/builds/1304#0185d156-ace6-4f9a-8255-25901c91daa7
* map_batches_benchmark_single_node https://buildkite.com/ray-project/release-tests-branch/builds/1333#0186130c-2057-4ecf-8be2-25bc4159fade
```
[ERROR 2023-01-21 22:46:41,582] run_release_test.py: 157 Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.
--
| Traceback (most recent call last):
| File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 69, in prepare_remote_env
| self.file_manager.upload()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 122, in upload
| self._push_local_dir()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 107, in _push_local_dir
| f"Error uploading local dir to session "
| ray_release.exception.FileUploadError: Error uploading local dir to session rllib_stress_tests_1674368507.
|
| The above exception was the direct cause of the following exception:
|
| Traceback (most recent call last):
| File "ray_release/scripts/run_release_test.py", line 153, in main
| no_terminate=no_terminate,
| File "/tmp/release-4K7NOSGKrN/release/ray_release/glue.py", line 404, in run_release_test
| raise pipeline_exception
| File "/tmp/release-4K7NOSGKrN/release/ray_release/glue.py", line 275, in run_release_test
| command_runner.prepare_remote_env()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 73, in prepare_remote_env
| ) from e
| ray_release.exception.RemoteEnvSetupError: Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.
| ERROR:ray_release.logger:Error setting up remote environment: Error uploading local dir to session rllib_stress_tests_1674368507.
| Traceback (most recent call last):
| File "/tmp/release-4K7NOSGKrN/release/ray_release/command_runner/sdk_runner.py", line 69, in prepare_remote_env
| self.file_manager.upload()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 122, in upload
| self._push_local_dir()
| File "/tmp/release-4K7NOSGKrN/release/ray_release/file_manager/job_file_manager.py", line 107, in _push_local_dir
| f"Error uploading local dir to session "
| ray_release.exception.FileUploadError: Error uploading local dir to session rllib_stress_tests_1674368507.
```
cc @krfricke | non_priority | error setting up remote environment error uploading local dir to session rllib stress tests seeing this in the following release tests dask on ray sort rllib stress tests shuffle partitions map batches benchmark single node run release test py error setting up remote environment error uploading local dir to session rllib stress tests traceback most recent call last file tmp release release ray release command runner sdk runner py line in prepare remote env self file manager upload file tmp release release ray release file manager job file manager py line in upload self push local dir file tmp release release ray release file manager job file manager py line in push local dir f error uploading local dir to session ray release exception fileuploaderror error uploading local dir to session rllib stress tests the above exception was the direct cause of the following exception traceback most recent call last file ray release scripts run release test py line in main no terminate no terminate file tmp release release ray release glue py line in run release test raise pipeline exception file tmp release release ray release glue py line in run release test command runner prepare remote env file tmp release release ray release command runner sdk runner py line in prepare remote env from e ray release exception remoteenvsetuperror error setting up remote environment error uploading local dir to session rllib stress tests error ray release logger error setting up remote environment error uploading local dir to session rllib stress tests traceback most recent call last file tmp release release ray release command runner sdk runner py line in prepare remote env self file manager upload file tmp release release ray release file manager job file manager py line in upload self push local dir file tmp release release ray release file manager job file manager py line in push local dir f error uploading local dir to session ray release exception fileuploaderror error uploading local dir to session rllib stress tests cc krfricke | 0 |
11,878 | 5,104,087,964 | IssuesEvent | 2017-01-04 23:38:16 | angular/material | https://api.github.com/repos/angular/material | closed | md-chips has aria-hidden='true' which makes it hidden from the screenreader | has: Pull Request P0 - Critical Fixes / Build or Deployment Issues ui: a11y | **Actual Behavior**:
- aria-hidden='true' is applied to the md-chips directive, which causes the the content inside md-chips does not get announced by the screenreader.
- Expect no aria-hidden='true' on the md-chips, so that the items inside it can be read by screenreader.
**Angular Versions**: *
- `Angular Version:` 1.5.8
- `Angular Material Version:` 1.5.8
**Additional Information**:
- `Browser Type: *` Chrome
| 1.0 | md-chips has aria-hidden='true' which makes it hidden from the screenreader - **Actual Behavior**:
- aria-hidden='true' is applied to the md-chips directive, which causes the the content inside md-chips does not get announced by the screenreader.
- Expect no aria-hidden='true' on the md-chips, so that the items inside it can be read by screenreader.
**Angular Versions**: *
- `Angular Version:` 1.5.8
- `Angular Material Version:` 1.5.8
**Additional Information**:
- `Browser Type: *` Chrome
| non_priority | md chips has aria hidden true which makes it hidden from the screenreader actual behavior aria hidden true is applied to the md chips directive which causes the the content inside md chips does not get announced by the screenreader expect no aria hidden true on the md chips so that the items inside it can be read by screenreader angular versions angular version angular material version additional information browser type chrome | 0 |
122,090 | 4,827,290,583 | IssuesEvent | 2016-11-07 13:09:16 | LuxoftAKutsan/RequirementsTest | https://api.github.com/repos/LuxoftAKutsan/RequirementsTest | opened | [GENIVI] SDL should respond "IGNORED" with correct result code for UnSubscribeVehicleData in case vi interface isn't available | bug high_priority | **Requirement:**
APPLINK-20043
**Occurrence:**
Always
**Preconditions:**
1. SubscribeVehicleData and UnSubscribeVehicleData are allowed in PT
**Steps to reproduce:**
1. RegisterApp in case HMI does not respond to VehicleInfo.IsReady.
2. Activate app
3. Send SubscribeVehicleData with gps = true -> SDL responds with SUCCESS.
4. Send UnSubscribeVehicleData with gps = true
**Expected result:**
SDL should respond "
{success = false, resultCode = "IGNORED", info = "Some provided VehicleData was not subscribed."}
" code to mobile app.
**Actual result:**
SDL respond "
{success = true, resultCode = "IGNORED", info = "Already subscribed on some provided VehicleData."}
".
**Problem can be reproduced with ATF script:**
1. Move IsReady_Template folder into sdl_atf\user_modules
2. Go to atf folder -> Run command: ./start.sh ../ATF_VehicleInfo_IsReady_available_false_SingleRPC.lua
**Attachments:**
| 1.0 | [GENIVI] SDL should respond "IGNORED" with correct result code for UnSubscribeVehicleData in case vi interface isn't available - **Requirement:**
APPLINK-20043
**Occurrence:**
Always
**Preconditions:**
1. SubscribeVehicleData and UnSubscribeVehicleData are allowed in PT
**Steps to reproduce:**
1. RegisterApp in case HMI does not respond to VehicleInfo.IsReady.
2. Activate app
3. Send SubscribeVehicleData with gps = true -> SDL responds with SUCCESS.
4. Send UnSubscribeVehicleData with gps = true
**Expected result:**
SDL should respond "
{success = false, resultCode = "IGNORED", info = "Some provided VehicleData was not subscribed."}
" code to mobile app.
**Actual result:**
SDL respond "
{success = true, resultCode = "IGNORED", info = "Already subscribed on some provided VehicleData."}
".
**Problem can be reproduced with ATF script:**
1. Move IsReady_Template folder into sdl_atf\user_modules
2. Go to atf folder -> Run command: ./start.sh ../ATF_VehicleInfo_IsReady_available_false_SingleRPC.lua
**Attachments:**
| priority | sdl should respond ignored with correct result code for unsubscribevehicledata in case vi interface isn t available requirement applink occurrence always preconditions subscribevehicledata and unsubscribevehicledata are allowed in pt steps to reproduce registerapp in case hmi does not respond to vehicleinfo isready activate app send subscribevehicledata with gps true sdl responds with success send unsubscribevehicledata with gps true expected result sdl should respond success false resultcode ignored info some provided vehicledata was not subscribed code to mobile app actual result sdl respond success true resultcode ignored info already subscribed on some provided vehicledata problem can be reproduced with atf script move isready template folder into sdl atf user modules go to atf folder run command start sh atf vehicleinfo isready available false singlerpc lua attachments | 1 |
130,330 | 18,155,766,998 | IssuesEvent | 2021-09-27 01:12:25 | benlazarine/cas-overlay | https://api.github.com/repos/benlazarine/cas-overlay | opened | CVE-2019-17531 (High) detected in jackson-databind-2.9.5.jar | security vulnerability | ## CVE-2019-17531 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: cas-overlay/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- cas-server-support-oauth-webflow-5.3.7.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload.
<p>Publish Date: 2019-10-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531>CVE-2019-17531</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531</a></p>
<p>Release Date: 2019-10-12</p>
<p>Fix Resolution: 2.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-17531 (High) detected in jackson-databind-2.9.5.jar - ## CVE-2019-17531 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: cas-overlay/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- cas-server-support-oauth-webflow-5.3.7.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.0.0 through 2.9.10. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has the apache-log4j-extra (version 1.2.x) jar in the classpath, and an attacker can provide a JNDI service to access, it is possible to make the service execute a malicious payload.
<p>Publish Date: 2019-10-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17531>CVE-2019-17531</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-17531</a></p>
<p>Release Date: 2019-10-12</p>
<p>Fix Resolution: 2.10</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file cas overlay pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy cas server support oauth webflow jar root library x jackson databind jar vulnerable library vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has the apache extra version x jar in the classpath and an attacker can provide a jndi service to access it is possible to make the service execute a malicious payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
334,570 | 10,142,506,143 | IssuesEvent | 2019-08-04 01:34:14 | jenkins-x/jx | https://api.github.com/repos/jenkins-x/jx | closed | jx install is skipped if (some undetermined) options specified to jx | area/install kind/bug lifecycle/rotten priority/important-soon | ### Summary
I attempted to install jx with `jx create terraform --environment-git-owner my-org --verbose -o testing --cluster goodclustername=gke --git-username=my-bot --git-api-token=somelonghash`
after terraform created the cluster I expected jx to be installed.
### Steps to reproduce the behavior
`jx create terraform --environment-git-owner my-org --verbose -o testing --cluster goodclustername=gke --git-username=my-bot --git-api-token=somelonghash`
### Jx version
The output of `jx version` is:
```
```
### Kubernetes cluster
What kind of Kubernetes cluster are you using & how did you create it?
### Operating system / Environment
### Expected behavior
the cluster is created using the options supplied **and jx is installed**.
### Actual behavior
The install of jx was "skipped" after the creation of the cluster
```
<snip>
google_container_node_pool.jx-node-pool: Still creating... (2m50s elapsed)
google_container_node_pool.jx-node-pool: Creation complete after 2m50s (ID: europe-west2-a/jamesntesting-jamesntesting/default-pool)
Apply complete! Resources: 2 added, 0 changed, 0 destroyed.
Outputs:
cluster_endpoint = 35.242.180.127
cluster_master_version = 1.9.7-gke.11
Fetching cluster endpoint and auth data.
kubeconfig entry generated for jamesntesting-jamesntesting.Skipping jx install
```
| 1.0 | jx install is skipped if (some undetermined) options specified to jx - ### Summary
I attempted to install jx with `jx create terraform --environment-git-owner my-org --verbose -o testing --cluster goodclustername=gke --git-username=my-bot --git-api-token=somelonghash`
after terraform created the cluster I expected jx to be installed.
### Steps to reproduce the behavior
`jx create terraform --environment-git-owner my-org --verbose -o testing --cluster goodclustername=gke --git-username=my-bot --git-api-token=somelonghash`
### Jx version
The output of `jx version` is:
```
```
### Kubernetes cluster
What kind of Kubernetes cluster are you using & how did you create it?
### Operating system / Environment
### Expected behavior
the cluster is created using the options supplied **and jx is installed**.
### Actual behavior
The install of jx was "skipped" after the creation of the cluster
```
<snip>
google_container_node_pool.jx-node-pool: Still creating... (2m50s elapsed)
google_container_node_pool.jx-node-pool: Creation complete after 2m50s (ID: europe-west2-a/jamesntesting-jamesntesting/default-pool)
Apply complete! Resources: 2 added, 0 changed, 0 destroyed.
Outputs:
cluster_endpoint = 35.242.180.127
cluster_master_version = 1.9.7-gke.11
Fetching cluster endpoint and auth data.
kubeconfig entry generated for jamesntesting-jamesntesting.Skipping jx install
```
| priority | jx install is skipped if some undetermined options specified to jx summary i attempted to install jx with jx create terraform environment git owner my org verbose o testing cluster goodclustername gke git username my bot git api token somelonghash after terraform created the cluster i expected jx to be installed steps to reproduce the behavior jx create terraform environment git owner my org verbose o testing cluster goodclustername gke git username my bot git api token somelonghash jx version the output of jx version is kubernetes cluster what kind of kubernetes cluster are you using how did you create it operating system environment expected behavior the cluster is created using the options supplied and jx is installed actual behavior the install of jx was skipped after the creation of the cluster google container node pool jx node pool still creating elapsed google container node pool jx node pool creation complete after id europe a jamesntesting jamesntesting default pool apply complete resources added changed destroyed outputs cluster endpoint cluster master version gke fetching cluster endpoint and auth data kubeconfig entry generated for jamesntesting jamesntesting skipping jx install | 1 |
827,562 | 31,778,715,985 | IssuesEvent | 2023-09-12 15:54:33 | opentensor/validators | https://api.github.com/repos/opentensor/validators | closed | Diversity model throwing exception | bug low priority | There are some runs logged in wandb that points to an error in the diversity model:
https://wandb.ai/opentensor-dev/openvalidators/runs/3em5sbqf/logs?workspace=user-opentensor-pedro
By the look of it, it seems to be the very same error that some of models in the reward/mask stack were facing: Whenever the text > context window of the model, the validator throws with this exception.
I believe it's worth investigating, even though the cases for this type of this exception is not that common
| 1.0 | Diversity model throwing exception - There are some runs logged in wandb that points to an error in the diversity model:
https://wandb.ai/opentensor-dev/openvalidators/runs/3em5sbqf/logs?workspace=user-opentensor-pedro
By the look of it, it seems to be the very same error that some of models in the reward/mask stack were facing: Whenever the text > context window of the model, the validator throws with this exception.
I believe it's worth investigating, even though the cases for this type of this exception is not that common
| priority | diversity model throwing exception there are some runs logged in wandb that points to an error in the diversity model by the look of it it seems to be the very same error that some of models in the reward mask stack were facing whenever the text context window of the model the validator throws with this exception i believe it s worth investigating even though the cases for this type of this exception is not that common | 1 |
825,635 | 31,465,145,649 | IssuesEvent | 2023-08-30 00:58:04 | War-Brokers/.github | https://api.github.com/repos/War-Brokers/.github | closed | FPS settings goes to 1 | type:bug priority:1 - high | When players close the game with vsync turned off and their fps set at 500, their FPS is set to 1 when they reopen the game | 1.0 | FPS settings goes to 1 - When players close the game with vsync turned off and their fps set at 500, their FPS is set to 1 when they reopen the game | priority | fps settings goes to when players close the game with vsync turned off and their fps set at their fps is set to when they reopen the game | 1 |
670,115 | 22,675,260,059 | IssuesEvent | 2022-07-04 03:22:27 | bitfoundation/bitplatform | https://api.github.com/repos/bitfoundation/bitplatform | closed | The `AutoInject` source generator ignores non-partial classes | high priority area / source generator | The `AutoInject` source generator should not ignore non-partial classes.
It should always generate the source code for all occurrences of the attribute. | 1.0 | The `AutoInject` source generator ignores non-partial classes - The `AutoInject` source generator should not ignore non-partial classes.
It should always generate the source code for all occurrences of the attribute. | priority | the autoinject source generator ignores non partial classes the autoinject source generator should not ignore non partial classes it should always generate the source code for all occurrences of the attribute | 1 |
110,491 | 11,704,451,503 | IssuesEvent | 2020-03-07 09:33:36 | bitfocus/companion-module-requests | https://api.github.com/repos/bitfocus/companion-module-requests | closed | SONOS | Hardware Missing documentation | I would like to user streamdeck for sonos volume up/down and
previous/next song in the playlist.
...maybe choose a playlist but that isn´t much important. | 1.0 | SONOS - I would like to user streamdeck for sonos volume up/down and
previous/next song in the playlist.
...maybe choose a playlist but that isn´t much important. | non_priority | sonos i would like to user streamdeck for sonos volume up down and previous next song in the playlist maybe choose a playlist but that isn´t much important | 0 |
375,354 | 11,103,100,011 | IssuesEvent | 2019-12-17 02:31:59 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | [0.33.7.1] Cannot create new users or send emails (pulses/alerts/invitations) | Administration/ Priority:P0 Reporting/Pulses Type:Bug | **Describe the bug**
Cannot create new users on 0.33.7.1, or sending pulse mails.
**Logs**
When trying to create a user:
```
12-15 11:43:04 ERROR middleware.log :: POST /api/user 500 247.0 ms (9 DB calls)
{:message "No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil",
:type java.lang.IllegalArgumentException,
:stacktrace
("clojure.core$_cache_protocol_fn.invokeStatic(core_deftype.clj:583)"
"clojure.core$_cache_protocol_fn.invoke(core_deftype.clj:575)"
"stencil.ast$fn__30303$G__30298__30312.invoke(ast.clj:19)"
"stencil.core$render.invokeStatic(core.clj:80)"
"stencil.core$render.invoke(core.clj:74)"
"stencil.core$render_file.invokeStatic(core.clj:87)"
"stencil.core$render_file.invoke(core.clj:83)"
"--> email.messages$send_new_user_email_BANG_.invokeStatic(messages.clj:67)"
"email.messages$send_new_user_email_BANG_.invoke(messages.clj:63)"
"models.user$send_welcome_email_BANG_.invokeStatic(user.clj:176)"
"models.user$send_welcome_email_BANG_.invoke(user.clj:172)"
"models.user$fn__32292$create_and_invite_user_BANG___32297$fn__32298.invoke(user.clj:208)"
"models.user$fn__32292$create_and_invite_user_BANG___32297.invoke(user.clj:203)"
"api.user$fn__63972$fn__63975.invoke(user.clj:101)"
"api.user$fn__63972.invokeStatic(user.clj:100)"
"api.user$fn__63972.invoke(user.clj:89)"
"middleware.auth$enforce_authentication$fn__64360.invoke(auth.clj:14)"
"routes$fn__65737$fn__65738.doInvoke(routes.clj:56)"
"middleware.exceptions$catch_uncaught_exceptions$fn__64459.invoke(exceptions.clj:104)"
"middleware.exceptions$catch_api_exceptions$fn__64456.invoke(exceptions.clj:92)"
"middleware.log$log_api_call$fn__66111$fn__66112.invoke(log.clj:170)"
"middleware.log$log_api_call$fn__66111.invoke(log.clj:164)"
"middleware.security$add_security_headers$fn__64422.invoke(security.clj:122)"
"middleware.json$wrap_json_body$fn__65816.invoke(json.clj:61)"
"middleware.json$wrap_streamed_json_response$fn__65834.invoke(json.clj:97)"
"middleware.session$bind_current_user$fn__60984$fn__60985.invoke(session.clj:209)"
"middleware.session$do_with_current_user.invokeStatic(session.clj:184)"
"middleware.session$do_with_current_user.invoke(session.clj:177)"
"middleware.session$bind_current_user$fn__60984.invoke(session.clj:208)"
"middleware.session$wrap_current_user_id$fn__60969.invoke(session.clj:161)"
"middleware.session$wrap_session_id$fn__60954.invoke(session.clj:123)"
"middleware.auth$wrap_api_key$fn__64368.invoke(auth.clj:27)"
"middleware.misc$maybe_set_site_url$fn__66138.invoke(misc.clj:56)"
"middleware.misc$bind_user_locale$fn__66141.invoke(misc.clj:72)"
"middleware.misc$add_content_type$fn__66126.invoke(misc.clj:28)"
"middleware.misc$disable_streaming_buffering$fn__66149.invoke(misc.clj:87)")}
```
When trying to send pulse test email:
```
12-15 12:49:41 ERROR middleware.log :: POST /api/pulse/test 500 187.6 ms (15 DB calls)
{:message "No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil",
:type java.lang.IllegalArgumentException,
:stacktrace
...
```
When sending scheduled pulse:
```
12-15 13:00:00 INFO task.send-pulses :: Sending scheduled pulses...
12-15 13:00:00 ERROR task.send-pulses :: Error sending Pulse 1
java.lang.IllegalArgumentException: No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil
...
```
**To Reproduce**
- Go to Admin > People > Add someone > fill in details and click Create
- Go to a pulse and click Send test email
- Go to a pulse and schedule a send, and check the log when pulse should have been sent
- ...perhaps there's other sections not working - will update as I find them...
**Information about your Metabase Installation:**
Metabase 0.33.7.1 Docker and JAR on various hosting
**Severity** P0
**Additional context**
https://discourse.metabase.com/t/cant-add-a-new-user-no-implementation-of-method-render-error/8350 | 1.0 | [0.33.7.1] Cannot create new users or send emails (pulses/alerts/invitations) - **Describe the bug**
Cannot create new users on 0.33.7.1, or sending pulse mails.
**Logs**
When trying to create a user:
```
12-15 11:43:04 ERROR middleware.log :: POST /api/user 500 247.0 ms (9 DB calls)
{:message "No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil",
:type java.lang.IllegalArgumentException,
:stacktrace
("clojure.core$_cache_protocol_fn.invokeStatic(core_deftype.clj:583)"
"clojure.core$_cache_protocol_fn.invoke(core_deftype.clj:575)"
"stencil.ast$fn__30303$G__30298__30312.invoke(ast.clj:19)"
"stencil.core$render.invokeStatic(core.clj:80)"
"stencil.core$render.invoke(core.clj:74)"
"stencil.core$render_file.invokeStatic(core.clj:87)"
"stencil.core$render_file.invoke(core.clj:83)"
"--> email.messages$send_new_user_email_BANG_.invokeStatic(messages.clj:67)"
"email.messages$send_new_user_email_BANG_.invoke(messages.clj:63)"
"models.user$send_welcome_email_BANG_.invokeStatic(user.clj:176)"
"models.user$send_welcome_email_BANG_.invoke(user.clj:172)"
"models.user$fn__32292$create_and_invite_user_BANG___32297$fn__32298.invoke(user.clj:208)"
"models.user$fn__32292$create_and_invite_user_BANG___32297.invoke(user.clj:203)"
"api.user$fn__63972$fn__63975.invoke(user.clj:101)"
"api.user$fn__63972.invokeStatic(user.clj:100)"
"api.user$fn__63972.invoke(user.clj:89)"
"middleware.auth$enforce_authentication$fn__64360.invoke(auth.clj:14)"
"routes$fn__65737$fn__65738.doInvoke(routes.clj:56)"
"middleware.exceptions$catch_uncaught_exceptions$fn__64459.invoke(exceptions.clj:104)"
"middleware.exceptions$catch_api_exceptions$fn__64456.invoke(exceptions.clj:92)"
"middleware.log$log_api_call$fn__66111$fn__66112.invoke(log.clj:170)"
"middleware.log$log_api_call$fn__66111.invoke(log.clj:164)"
"middleware.security$add_security_headers$fn__64422.invoke(security.clj:122)"
"middleware.json$wrap_json_body$fn__65816.invoke(json.clj:61)"
"middleware.json$wrap_streamed_json_response$fn__65834.invoke(json.clj:97)"
"middleware.session$bind_current_user$fn__60984$fn__60985.invoke(session.clj:209)"
"middleware.session$do_with_current_user.invokeStatic(session.clj:184)"
"middleware.session$do_with_current_user.invoke(session.clj:177)"
"middleware.session$bind_current_user$fn__60984.invoke(session.clj:208)"
"middleware.session$wrap_current_user_id$fn__60969.invoke(session.clj:161)"
"middleware.session$wrap_session_id$fn__60954.invoke(session.clj:123)"
"middleware.auth$wrap_api_key$fn__64368.invoke(auth.clj:27)"
"middleware.misc$maybe_set_site_url$fn__66138.invoke(misc.clj:56)"
"middleware.misc$bind_user_locale$fn__66141.invoke(misc.clj:72)"
"middleware.misc$add_content_type$fn__66126.invoke(misc.clj:28)"
"middleware.misc$disable_streaming_buffering$fn__66149.invoke(misc.clj:87)")}
```
When trying to send pulse test email:
```
12-15 12:49:41 ERROR middleware.log :: POST /api/pulse/test 500 187.6 ms (15 DB calls)
{:message "No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil",
:type java.lang.IllegalArgumentException,
:stacktrace
...
```
When sending scheduled pulse:
```
12-15 13:00:00 INFO task.send-pulses :: Sending scheduled pulses...
12-15 13:00:00 ERROR task.send-pulses :: Error sending Pulse 1
java.lang.IllegalArgumentException: No implementation of method: :render of protocol: #'stencil.ast/ASTNode found for class: nil
...
```
**To Reproduce**
- Go to Admin > People > Add someone > fill in details and click Create
- Go to a pulse and click Send test email
- Go to a pulse and schedule a send, and check the log when pulse should have been sent
- ...perhaps there's other sections not working - will update as I find them...
**Information about your Metabase Installation:**
Metabase 0.33.7.1 Docker and JAR on various hosting
**Severity** P0
**Additional context**
https://discourse.metabase.com/t/cant-add-a-new-user-no-implementation-of-method-render-error/8350 | priority | cannot create new users or send emails pulses alerts invitations describe the bug cannot create new users on or sending pulse mails logs when trying to create a user error middleware log post api user ms db calls message no implementation of method render of protocol stencil ast astnode found for class nil type java lang illegalargumentexception stacktrace clojure core cache protocol fn invokestatic core deftype clj clojure core cache protocol fn invoke core deftype clj stencil ast fn g invoke ast clj stencil core render invokestatic core clj stencil core render invoke core clj stencil core render file invokestatic core clj stencil core render file invoke core clj email messages send new user email bang invokestatic messages clj email messages send new user email bang invoke messages clj models user send welcome email bang invokestatic user clj models user send welcome email bang invoke user clj models user fn create and invite user bang fn invoke user clj models user fn create and invite user bang invoke user clj api user fn fn invoke user clj api user fn invokestatic user clj api user fn invoke user clj middleware auth enforce authentication fn invoke auth clj routes fn fn doinvoke routes clj middleware exceptions catch uncaught exceptions fn invoke exceptions clj middleware exceptions catch api exceptions fn invoke exceptions clj middleware log log api call fn fn invoke log clj middleware log log api call fn invoke log clj middleware security add security headers fn invoke security clj middleware json wrap json body fn invoke json clj middleware json wrap streamed json response fn invoke json clj middleware session bind current user fn fn invoke session clj middleware session do with current user invokestatic session clj middleware session do with current user invoke session clj middleware session bind current user fn invoke session clj middleware session wrap current user id fn invoke session clj middleware session wrap session id fn invoke session clj middleware auth wrap api key fn invoke auth clj middleware misc maybe set site url fn invoke misc clj middleware misc bind user locale fn invoke misc clj middleware misc add content type fn invoke misc clj middleware misc disable streaming buffering fn invoke misc clj when trying to send pulse test email error middleware log post api pulse test ms db calls message no implementation of method render of protocol stencil ast astnode found for class nil type java lang illegalargumentexception stacktrace when sending scheduled pulse info task send pulses sending scheduled pulses error task send pulses error sending pulse java lang illegalargumentexception no implementation of method render of protocol stencil ast astnode found for class nil to reproduce go to admin people add someone fill in details and click create go to a pulse and click send test email go to a pulse and schedule a send and check the log when pulse should have been sent perhaps there s other sections not working will update as i find them information about your metabase installation metabase docker and jar on various hosting severity additional context | 1 |
276,695 | 30,521,836,937 | IssuesEvent | 2023-07-19 08:36:57 | kseb49/my_blog | https://api.github.com/repos/kseb49/my_blog | closed | Handle the expired confirmed link | backend security | Generate a new link to confirmed an account
`// Expired link
if ($register->newToken() !== false) {
$mail = new Mail();
$message = $this->twig->render("templates/mail/validation-mail.twig", ["link" => $register->link,]);
$mail->mail($register->user['email'],$message,"Confirmez votre compte", $register->user['f_name']." ".$register->user['l_name'], 'Recopier ce lien pour valider votre compte : '.$register->link);
Flash::flash('success', 'Ce lien est expiré, vous avez reçu un nouveau mail pour confirmer votre compte');
$this->redirect();
}`
18/07/2023 | True | Handle the expired confirmed link - Generate a new link to confirmed an account
`// Expired link
if ($register->newToken() !== false) {
$mail = new Mail();
$message = $this->twig->render("templates/mail/validation-mail.twig", ["link" => $register->link,]);
$mail->mail($register->user['email'],$message,"Confirmez votre compte", $register->user['f_name']." ".$register->user['l_name'], 'Recopier ce lien pour valider votre compte : '.$register->link);
Flash::flash('success', 'Ce lien est expiré, vous avez reçu un nouveau mail pour confirmer votre compte');
$this->redirect();
}`
18/07/2023 | non_priority | handle the expired confirmed link generate a new link to confirmed an account expired link if register newtoken false mail new mail message this twig render templates mail validation mail twig mail mail register user message confirmez votre compte register user register user recopier ce lien pour valider votre compte register link flash flash success ce lien est expiré vous avez reçu un nouveau mail pour confirmer votre compte this redirect | 0 |
800,674 | 28,374,784,354 | IssuesEvent | 2023-04-12 19:54:53 | GoogleCloudPlatform/microservices-demo | https://api.github.com/repos/GoogleCloudPlatform/microservices-demo | closed | Unable to build adservice | type: bug priority: p2 | ### Describe the bug
build [adservice] failed: exit status 1. Docker build ran into internal error. Please retry.
### To Reproduce
<!-- Steps to reproduce the behavior: -->
<!-- 1. Built image '...' -->
<!-- 2. Ran command '....' -->
<!-- 3. See error -->
### Logs
Generating tags...
- productcatalogservice -> productcatalogservice:v0.5.0-30-gb2ef53f
- recommendationservice -> recommendationservice:v0.5.0-30-gb2ef53f
- shippingservice -> shippingservice:v0.5.0-30-gb2ef53f
- checkoutservice -> checkoutservice:v0.5.0-30-gb2ef53f
- paymentservice -> paymentservice:v0.5.0-30-gb2ef53f
- currencyservice -> currencyservice:v0.5.0-30-gb2ef53f
- cartservice -> cartservice:v0.5.0-30-gb2ef53f
- frontend -> frontend:v0.5.0-30-gb2ef53f
- adservice -> adservice:v0.5.0-30-gb2ef53f
- loadgenerator -> loadgenerator:v0.5.0-30-gb2ef53f
Checking cache...
- emailservice: Not found. Building
- productcatalogservice: Not found. Building
- recommendationservice: Not found. Building
- shippingservice: Found Locally
- checkoutservice: Found Locally
- paymentservice: Found Locally
- currencyservice: Found Locally
- cartservice: Found Locally
- frontend: Found Locally
- adservice: Not found. Building
- loadgenerator: Found Locally
Starting build...
Found [minikube] context, using local docker daemon.
Building [adservice]...
Target platforms: [linux/amd64]
#1 [internal] load build definition from Dockerfile
#1 sha256:4fd7c5b169637ade6dc46fa79f5ebe497faea8466f82e301c0fc5d1799fe1763
#1 transferring dockerfile: 38B done
#1 DONE 0.0s
#2 [internal] load .dockerignore
#2 sha256:3a3376a110bfc911c611e7778255bed1a6fcc5f12331b0f88bd3f8a17922905e
#2 transferring context: 2B done
#2 DONE 0.0s
#4 [internal] load metadata for docker.io/library/eclipse-temurin:19@sha256:17e3d3b61ca4a7606490f596feb77f69980939fecacf91a1f13ea7b17147058f
#4 sha256:fa4367848242f6486a2e5650a6e73e7b59134bc9cb3dc1bc2d5c9165b8260a0c
#4 DONE 0.0s
#3 [internal] load metadata for docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a
#3 sha256:095a08a5e87c0f08ca7fe08671a1bd8b85fa3840a51991c36774312eb5e5a0c7
#3 DONE 0.0s
#9 [builder 1/9] FROM docker.io/library/eclipse-temurin:19@sha256:17e3d3b61ca4a7606490f596feb77f69980939fecacf91a1f13ea7b17147058f
#9 sha256:e6598a80b4806e6f96c6639d605cf4c6eca4174e9d77b4bae443fffb1f2dcc2c
#9 DONE 0.0s
#11 [internal] load build context
#11 sha256:14ea7f78c196e3fc2a5c4cafe5ae47a5c9c8babae792578ef62a0fb5a60e3aaf
#11 transferring context: 840B done
#11 DONE 0.0s
#13 [builder 4/9] COPY gradle gradle
#13 sha256:fdaeb43ef35a139a2ecd1cead76d33dfdf08ec928d056377bb6d0e4fa7ba5adc
#13 CACHED
#10 [builder 2/9] WORKDIR /app
#10 sha256:c8376895fb2b0a3b0283686ae8db694693504e2b0084b35d3dad35064057b43b
#10 CACHED
#12 [builder 3/9] COPY [build.gradle, gradlew, ./]
#12 sha256:83c710fc667545608bfc816075122245e3b257bcc1277345bb9b3282cfe1fefe
#12 CACHED
#14 [builder 5/9] RUN chmod +x gradlew
#14 sha256:1a500f6f159b665ab926871cb66d0e20677a0a5f0b055537bcd781f6c6395e61
#14 CACHED
#5 [without-grpc-health-probe-bin 1/5] FROM docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a
3a
#5 sha256:3ec4fd675030cb4080679827d76012c9836cfa8e143f949ed6cf49b3d748d907
#5 resolve docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a done
#5 sha256:bfba53c59e931f5d57c41ba291b90b1ab7e3034d4df5a30ba2ff56f63583beef 0B / 161B 0.2s
#5 sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a 1.16kB / 1.16kB done
#5 sha256:0fdafb034be2a2cb5ed98aa1b8116b05184efb869e1d7026bb395dc7f7be8f50 4.22kB / 4.22kB done
#5 sha256:d8e5acd5897d762b9a83758d4ceae374df7b8b0367a48cc14b8a00e33998b3bf 0B / 12.02MB 0.2s
#5 sha256:b231e5a3c8c25582f722e10270be76c230ecdb668b26782fe0b4cd29037021c6 0B / 49.14MB 0.2s
#5 CANCELED
#15 [builder 6/9] RUN ./gradlew downloadRepos
#15 sha256:3997494df4830672a9af711404a7be3af2e05bafbda5e4154acd7d8958e0c28e
#15 0.366 /bin/sh: 1: ./gradlew: not found
#15 ERROR: executor failed running [/bin/sh -c ./gradlew downloadRepos]: exit code: 127
------
> [builder 6/9] RUN ./gradlew downloadRepos:
------
executor failed running [/bin/sh -c ./gradlew downloadRepos]: exit code: 127
Building [productcatalogservice]...
Target platforms: [linux/amd64]
Build [productcatalogservice] was canceled
Building [emailservice]...
Target platforms: [linux/amd64]
Build [emailservice] was canceled
Building [recommendationservice]...
Target platforms: [linux/amd64]
Build [recommendationservice] was canceled
build [adservice] failed: exit status 1. Docker build ran into internal error. Please retry.
If this keeps happening, please open an issue..
### Screenshots
<!-- If applicable, add screenshots to help explain your problem -->
### Environment
- OS: Windows 10 (Run on WSL 2)
- Kubernetes distribution, version: minikube version: v1.28.0
- Any relevant tool version: Docker Desktop v4.15.0
### Additional context
<!-- Add any other context about the problem here -->
### Exposure
<!-- Is the bug intermittent, persistent? Is it widespread, local? -->
| 1.0 | Unable to build adservice - ### Describe the bug
build [adservice] failed: exit status 1. Docker build ran into internal error. Please retry.
### To Reproduce
<!-- Steps to reproduce the behavior: -->
<!-- 1. Built image '...' -->
<!-- 2. Ran command '....' -->
<!-- 3. See error -->
### Logs
Generating tags...
- productcatalogservice -> productcatalogservice:v0.5.0-30-gb2ef53f
- recommendationservice -> recommendationservice:v0.5.0-30-gb2ef53f
- shippingservice -> shippingservice:v0.5.0-30-gb2ef53f
- checkoutservice -> checkoutservice:v0.5.0-30-gb2ef53f
- paymentservice -> paymentservice:v0.5.0-30-gb2ef53f
- currencyservice -> currencyservice:v0.5.0-30-gb2ef53f
- cartservice -> cartservice:v0.5.0-30-gb2ef53f
- frontend -> frontend:v0.5.0-30-gb2ef53f
- adservice -> adservice:v0.5.0-30-gb2ef53f
- loadgenerator -> loadgenerator:v0.5.0-30-gb2ef53f
Checking cache...
- emailservice: Not found. Building
- productcatalogservice: Not found. Building
- recommendationservice: Not found. Building
- shippingservice: Found Locally
- checkoutservice: Found Locally
- paymentservice: Found Locally
- currencyservice: Found Locally
- cartservice: Found Locally
- frontend: Found Locally
- adservice: Not found. Building
- loadgenerator: Found Locally
Starting build...
Found [minikube] context, using local docker daemon.
Building [adservice]...
Target platforms: [linux/amd64]
#1 [internal] load build definition from Dockerfile
#1 sha256:4fd7c5b169637ade6dc46fa79f5ebe497faea8466f82e301c0fc5d1799fe1763
#1 transferring dockerfile: 38B done
#1 DONE 0.0s
#2 [internal] load .dockerignore
#2 sha256:3a3376a110bfc911c611e7778255bed1a6fcc5f12331b0f88bd3f8a17922905e
#2 transferring context: 2B done
#2 DONE 0.0s
#4 [internal] load metadata for docker.io/library/eclipse-temurin:19@sha256:17e3d3b61ca4a7606490f596feb77f69980939fecacf91a1f13ea7b17147058f
#4 sha256:fa4367848242f6486a2e5650a6e73e7b59134bc9cb3dc1bc2d5c9165b8260a0c
#4 DONE 0.0s
#3 [internal] load metadata for docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a
#3 sha256:095a08a5e87c0f08ca7fe08671a1bd8b85fa3840a51991c36774312eb5e5a0c7
#3 DONE 0.0s
#9 [builder 1/9] FROM docker.io/library/eclipse-temurin:19@sha256:17e3d3b61ca4a7606490f596feb77f69980939fecacf91a1f13ea7b17147058f
#9 sha256:e6598a80b4806e6f96c6639d605cf4c6eca4174e9d77b4bae443fffb1f2dcc2c
#9 DONE 0.0s
#11 [internal] load build context
#11 sha256:14ea7f78c196e3fc2a5c4cafe5ae47a5c9c8babae792578ef62a0fb5a60e3aaf
#11 transferring context: 840B done
#11 DONE 0.0s
#13 [builder 4/9] COPY gradle gradle
#13 sha256:fdaeb43ef35a139a2ecd1cead76d33dfdf08ec928d056377bb6d0e4fa7ba5adc
#13 CACHED
#10 [builder 2/9] WORKDIR /app
#10 sha256:c8376895fb2b0a3b0283686ae8db694693504e2b0084b35d3dad35064057b43b
#10 CACHED
#12 [builder 3/9] COPY [build.gradle, gradlew, ./]
#12 sha256:83c710fc667545608bfc816075122245e3b257bcc1277345bb9b3282cfe1fefe
#12 CACHED
#14 [builder 5/9] RUN chmod +x gradlew
#14 sha256:1a500f6f159b665ab926871cb66d0e20677a0a5f0b055537bcd781f6c6395e61
#14 CACHED
#5 [without-grpc-health-probe-bin 1/5] FROM docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a
3a
#5 sha256:3ec4fd675030cb4080679827d76012c9836cfa8e143f949ed6cf49b3d748d907
#5 resolve docker.io/library/eclipse-temurin:19.0.1_10-jre-alpine@sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a done
#5 sha256:bfba53c59e931f5d57c41ba291b90b1ab7e3034d4df5a30ba2ff56f63583beef 0B / 161B 0.2s
#5 sha256:1aa167ab4f1498130e04bed5d6a83fed23c1fd8e3df8589723bc876770dd6a3a 1.16kB / 1.16kB done
#5 sha256:0fdafb034be2a2cb5ed98aa1b8116b05184efb869e1d7026bb395dc7f7be8f50 4.22kB / 4.22kB done
#5 sha256:d8e5acd5897d762b9a83758d4ceae374df7b8b0367a48cc14b8a00e33998b3bf 0B / 12.02MB 0.2s
#5 sha256:b231e5a3c8c25582f722e10270be76c230ecdb668b26782fe0b4cd29037021c6 0B / 49.14MB 0.2s
#5 CANCELED
#15 [builder 6/9] RUN ./gradlew downloadRepos
#15 sha256:3997494df4830672a9af711404a7be3af2e05bafbda5e4154acd7d8958e0c28e
#15 0.366 /bin/sh: 1: ./gradlew: not found
#15 ERROR: executor failed running [/bin/sh -c ./gradlew downloadRepos]: exit code: 127
------
> [builder 6/9] RUN ./gradlew downloadRepos:
------
executor failed running [/bin/sh -c ./gradlew downloadRepos]: exit code: 127
Building [productcatalogservice]...
Target platforms: [linux/amd64]
Build [productcatalogservice] was canceled
Building [emailservice]...
Target platforms: [linux/amd64]
Build [emailservice] was canceled
Building [recommendationservice]...
Target platforms: [linux/amd64]
Build [recommendationservice] was canceled
build [adservice] failed: exit status 1. Docker build ran into internal error. Please retry.
If this keeps happening, please open an issue..
### Screenshots
<!-- If applicable, add screenshots to help explain your problem -->
### Environment
- OS: Windows 10 (Run on WSL 2)
- Kubernetes distribution, version: minikube version: v1.28.0
- Any relevant tool version: Docker Desktop v4.15.0
### Additional context
<!-- Add any other context about the problem here -->
### Exposure
<!-- Is the bug intermittent, persistent? Is it widespread, local? -->
| priority | unable to build adservice describe the bug build failed exit status docker build ran into internal error please retry to reproduce logs generating tags productcatalogservice productcatalogservice recommendationservice recommendationservice shippingservice shippingservice checkoutservice checkoutservice paymentservice paymentservice currencyservice currencyservice cartservice cartservice frontend frontend adservice adservice loadgenerator loadgenerator checking cache emailservice not found building productcatalogservice not found building recommendationservice not found building shippingservice found locally checkoutservice found locally paymentservice found locally currencyservice found locally cartservice found locally frontend found locally adservice not found building loadgenerator found locally starting build found context using local docker daemon building target platforms load build definition from dockerfile transferring dockerfile done done load dockerignore transferring context done done load metadata for docker io library eclipse temurin done load metadata for docker io library eclipse temurin jre alpine done from docker io library eclipse temurin done load build context transferring context done done copy gradle gradle cached workdir app cached copy cached run chmod x gradlew cached from docker io library eclipse temurin jre alpine resolve docker io library eclipse temurin jre alpine done done done canceled run gradlew downloadrepos bin sh gradlew not found error executor failed running exit code run gradlew downloadrepos executor failed running exit code building target platforms build was canceled building target platforms build was canceled building target platforms build was canceled build failed exit status docker build ran into internal error please retry if this keeps happening please open an issue screenshots environment os windows run on wsl kubernetes distribution version minikube version any relevant tool version docker desktop additional context exposure | 1 |
486,771 | 14,014,097,590 | IssuesEvent | 2020-10-29 11:23:32 | workcraft/workcraft | https://api.github.com/repos/workcraft/workcraft | opened | Thread-unsafe use of JTextArea.append for displaying log messages | bug priority:critical status:confirmed tag:core | Method `JTextArea.append(String s)` is used for for printing the logs in the Output and Problems tabs. Logging may be invoked from threads, which causes problems in some JDK implementations (e.g. in OpenJDK build 1.8.0_265 for Ubuntu 20.04).
This method used to be officially thread-safe in Java 6, but in Java 7 the thread-safeness requirement was revoked (see the second answer in this thread): https://stackoverflow.com/questions/4869027/jtextarea-thread-safe
A solution is to schedule those JTextArea.append(...) calls in EDT.
| 1.0 | Thread-unsafe use of JTextArea.append for displaying log messages - Method `JTextArea.append(String s)` is used for for printing the logs in the Output and Problems tabs. Logging may be invoked from threads, which causes problems in some JDK implementations (e.g. in OpenJDK build 1.8.0_265 for Ubuntu 20.04).
This method used to be officially thread-safe in Java 6, but in Java 7 the thread-safeness requirement was revoked (see the second answer in this thread): https://stackoverflow.com/questions/4869027/jtextarea-thread-safe
A solution is to schedule those JTextArea.append(...) calls in EDT.
| priority | thread unsafe use of jtextarea append for displaying log messages method jtextarea append string s is used for for printing the logs in the output and problems tabs logging may be invoked from threads which causes problems in some jdk implementations e g in openjdk build for ubuntu this method used to be officially thread safe in java but in java the thread safeness requirement was revoked see the second answer in this thread a solution is to schedule those jtextarea append calls in edt | 1 |
368,718 | 10,883,689,815 | IssuesEvent | 2019-11-18 05:57:09 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Master: outside "scroll bar" not available for interaction. | Fixed Medium Priority | As i see, "window size" controller just over it :dagger:

| 1.0 | Master: outside "scroll bar" not available for interaction. - As i see, "window size" controller just over it :dagger:

| priority | master outside scroll bar not available for interaction as i see window size controller just over it dagger | 1 |
189,669 | 6,800,552,181 | IssuesEvent | 2017-11-02 14:19:54 | feelpp/feelpp | https://api.github.com/repos/feelpp/feelpp | opened | Add support for first order moment computation, element-wise | module:discr priority:Urgent project: eye2brain project: mso4sc status:in development type:feature | element wise first moment of a function f computes \int_K f(x) dx
if we divide by |K| then it is the mean value of f | 1.0 | Add support for first order moment computation, element-wise - element wise first moment of a function f computes \int_K f(x) dx
if we divide by |K| then it is the mean value of f | priority | add support for first order moment computation element wise element wise first moment of a function f computes int k f x dx if we divide by k then it is the mean value of f | 1 |
55,751 | 13,671,968,566 | IssuesEvent | 2020-09-29 07:50:19 | zeek/spicy | https://api.github.com/repos/zeek/spicy | opened | Ship a custom "filesystem" implementation for better supporting older platforms | Build system | This might help with #451. A candidate is https://github.com/gulrak/filesystem/ | 1.0 | Ship a custom "filesystem" implementation for better supporting older platforms - This might help with #451. A candidate is https://github.com/gulrak/filesystem/ | non_priority | ship a custom filesystem implementation for better supporting older platforms this might help with a candidate is | 0 |
772,973 | 27,141,742,630 | IssuesEvent | 2023-02-16 16:49:00 | googleapis/python-bigquery-sqlalchemy | https://api.github.com/repos/googleapis/python-bigquery-sqlalchemy | closed | tests.sqlalchemy_dialect_compliance.test_dialect_compliance.HasTableTest_bigquery+bigquery: test_has_table_view failed | type: bug priority: p1 flakybot: issue api: bigquery | Note: #622 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 074321ddaa10001773e7e6044f4a0df1bb530331
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/82580611-18ba-4d10-8e25-a7bf752e1da1), [Sponge](http://sponge2/82580611-18ba-4d10-8e25-a7bf752e1da1)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 311, in from_call
result: Optional[TResult] = func()
^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 255, in <lambda>
lambda: ihook(item=item, **kwds), when=when, reraise=reraise
^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_hooks.py", line 265, in __call__
return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_manager.py", line 80, in _hookexec
return self._inner_hookexec(hook_name, methods, kwargs, firstresult)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 60, in _multicall
return outcome.get_result()
^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_result.py", line 60, in get_result
raise ex[1].with_traceback(ex[2])
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 39, in _multicall
res = hook_impl.function(*args)
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 175, in pytest_runtest_teardown
item.session._setupstate.teardown_exact(item, nextitem)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 419, in teardown_exact
self._teardown_towards(needed_collectors)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 434, in _teardown_towards
raise exc
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 427, in _teardown_towards
self._pop_and_teardown()
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 387, in _pop_and_teardown
self._teardown_with_finalization(colitem)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 408, in _teardown_with_finalization
assert colitem in self.stack
^^^^^^^^^^^^^^^^^^^^^
AssertionError</pre></details> | 1.0 | tests.sqlalchemy_dialect_compliance.test_dialect_compliance.HasTableTest_bigquery+bigquery: test_has_table_view failed - Note: #622 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: 074321ddaa10001773e7e6044f4a0df1bb530331
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/82580611-18ba-4d10-8e25-a7bf752e1da1), [Sponge](http://sponge2/82580611-18ba-4d10-8e25-a7bf752e1da1)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 311, in from_call
result: Optional[TResult] = func()
^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 255, in <lambda>
lambda: ihook(item=item, **kwds), when=when, reraise=reraise
^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_hooks.py", line 265, in __call__
return self._hookexec(self.name, self.get_hookimpls(), kwargs, firstresult)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_manager.py", line 80, in _hookexec
return self._inner_hookexec(hook_name, methods, kwargs, firstresult)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 60, in _multicall
return outcome.get_result()
^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_result.py", line 60, in get_result
raise ex[1].with_traceback(ex[2])
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/pluggy/_callers.py", line 39, in _multicall
res = hook_impl.function(*args)
^^^^^^^^^^^^^^^^^^^^^^^^^
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 175, in pytest_runtest_teardown
item.session._setupstate.teardown_exact(item, nextitem)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 419, in teardown_exact
self._teardown_towards(needed_collectors)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 434, in _teardown_towards
raise exc
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 427, in _teardown_towards
self._pop_and_teardown()
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 387, in _pop_and_teardown
self._teardown_with_finalization(colitem)
File "/tmpfs/src/github/python-bigquery-sqlalchemy/.nox/compliance/lib/python3.11/site-packages/_pytest/runner.py", line 408, in _teardown_with_finalization
assert colitem in self.stack
^^^^^^^^^^^^^^^^^^^^^
AssertionError</pre></details> | priority | tests sqlalchemy dialect compliance test dialect compliance hastabletest bigquery bigquery test has table view failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in from call result optional func file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in lambda ihook item item kwds when when reraise reraise file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy hooks py line in call return self hookexec self name self get hookimpls kwargs firstresult file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy manager py line in hookexec return self inner hookexec hook name methods kwargs firstresult file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy callers py line in multicall return outcome get result file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy result py line in get result raise ex with traceback ex file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pluggy callers py line in multicall res hook impl function args file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in pytest runtest teardown item session setupstate teardown exact item nextitem file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown exact self teardown towards needed collectors file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown towards raise exc file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown towards self pop and teardown file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in pop and teardown self teardown with finalization colitem file tmpfs src github python bigquery sqlalchemy nox compliance lib site packages pytest runner py line in teardown with finalization assert colitem in self stack assertionerror | 1 |
31,342 | 5,939,857,565 | IssuesEvent | 2017-05-25 07:08:55 | c3js/c3 | https://api.github.com/repos/c3js/c3 | closed | Document use of d3.locale to choose localisation settings | documentation needed | How do I change the labels on the donut chart to display X,Y% (instead of american X.Y%)?
If I try to adjust the labels in D3 locale.js I break the labels. Seems like C3 is calculating the percentages independently of the D3 settings.
Maybe this is just not documented. Can we document localization and i18n ?
| 1.0 | Document use of d3.locale to choose localisation settings - How do I change the labels on the donut chart to display X,Y% (instead of american X.Y%)?
If I try to adjust the labels in D3 locale.js I break the labels. Seems like C3 is calculating the percentages independently of the D3 settings.
Maybe this is just not documented. Can we document localization and i18n ?
| non_priority | document use of locale to choose localisation settings how do i change the labels on the donut chart to display x y instead of american x y if i try to adjust the labels in locale js i break the labels seems like is calculating the percentages independently of the settings maybe this is just not documented can we document localization and | 0 |
324,682 | 9,907,486,991 | IssuesEvent | 2019-06-27 15:55:06 | python/mypy | https://api.github.com/repos/python/mypy | closed | Avoid redundant extra errors after "Need type annotation for variable" | new-semantic-analyzer priority-0-high topic-type-variables topic-usability | Consider this (oversimplified) example:
```python
from typing import TypeVar, Optional
T = TypeVar('T')
def f(x: Optional[T] = None) -> T: ...
x = f() # Need type annotation for 'x'
y = x # Cannot determine type of 'x'
def g() -> None:
x = f() # Need type annotation for 'x'
y = x # Cannot determine type of 'x'
```
The subsequent errors look redundant (and maybe even annoying). This is more important with the new semantic analyzer because the old one doesn't show the error if the problematic variable was defined at the function scope.
Potentially, we can replace `<nothing>` with `Any` after we show the error. Also the code that detects failed inference can be improved. For example, `is_valid_inferred_type()` should be replaced with a proper type visitor. | 1.0 | Avoid redundant extra errors after "Need type annotation for variable" - Consider this (oversimplified) example:
```python
from typing import TypeVar, Optional
T = TypeVar('T')
def f(x: Optional[T] = None) -> T: ...
x = f() # Need type annotation for 'x'
y = x # Cannot determine type of 'x'
def g() -> None:
x = f() # Need type annotation for 'x'
y = x # Cannot determine type of 'x'
```
The subsequent errors look redundant (and maybe even annoying). This is more important with the new semantic analyzer because the old one doesn't show the error if the problematic variable was defined at the function scope.
Potentially, we can replace `<nothing>` with `Any` after we show the error. Also the code that detects failed inference can be improved. For example, `is_valid_inferred_type()` should be replaced with a proper type visitor. | priority | avoid redundant extra errors after need type annotation for variable consider this oversimplified example python from typing import typevar optional t typevar t def f x optional none t x f need type annotation for x y x cannot determine type of x def g none x f need type annotation for x y x cannot determine type of x the subsequent errors look redundant and maybe even annoying this is more important with the new semantic analyzer because the old one doesn t show the error if the problematic variable was defined at the function scope potentially we can replace with any after we show the error also the code that detects failed inference can be improved for example is valid inferred type should be replaced with a proper type visitor | 1 |
10,009 | 26,002,092,469 | IssuesEvent | 2022-12-20 16:05:00 | OWASP/raider | https://api.github.com/repos/OWASP/raider | opened | Create a raider REPL | enhancement architecture | To preserve a session across runs, raider would need to have a REPL, and the user should have the ability to inspect and manipulate all individual elements. | 1.0 | Create a raider REPL - To preserve a session across runs, raider would need to have a REPL, and the user should have the ability to inspect and manipulate all individual elements. | non_priority | create a raider repl to preserve a session across runs raider would need to have a repl and the user should have the ability to inspect and manipulate all individual elements | 0 |
200,518 | 7,008,645,903 | IssuesEvent | 2017-12-19 16:20:35 | metasfresh/metasfresh | https://api.github.com/repos/metasfresh/metasfresh | opened | Material Tracking ID in Invoice Candidates Filter shall be search | priority:high type:enhancement | ### Is this a bug or feature request?
Feature Request
### What is the current behavior?
Currently, the Field Material Tracking in Invoice Candidates is a List. This makes it unusable as Filter criteria.
#### Which are the steps to reproduce?
Open, try and see.
### What is the expected or desired behavior?
Change the Fields to search in invoice Candidates. | 1.0 | Material Tracking ID in Invoice Candidates Filter shall be search - ### Is this a bug or feature request?
Feature Request
### What is the current behavior?
Currently, the Field Material Tracking in Invoice Candidates is a List. This makes it unusable as Filter criteria.
#### Which are the steps to reproduce?
Open, try and see.
### What is the expected or desired behavior?
Change the Fields to search in invoice Candidates. | priority | material tracking id in invoice candidates filter shall be search is this a bug or feature request feature request what is the current behavior currently the field material tracking in invoice candidates is a list this makes it unusable as filter criteria which are the steps to reproduce open try and see what is the expected or desired behavior change the fields to search in invoice candidates | 1 |
9,074 | 3,253,451,565 | IssuesEvent | 2015-10-19 19:09:11 | rancher/os | https://api.github.com/repos/rancher/os | closed | ros config fails on fresh v0.4.0-rc11 install | area/documentation release/v0.4.0 | On a freshly-installed system, running the following, straight from the documentation, results in nothing being changed, and `ros config get` returns nothing:
```
$ sudo ros config set network.dns.domain myexampledomain.com
$ sudo ros config get network.dns.domain
``` | 1.0 | ros config fails on fresh v0.4.0-rc11 install - On a freshly-installed system, running the following, straight from the documentation, results in nothing being changed, and `ros config get` returns nothing:
```
$ sudo ros config set network.dns.domain myexampledomain.com
$ sudo ros config get network.dns.domain
``` | non_priority | ros config fails on fresh install on a freshly installed system running the following straight from the documentation results in nothing being changed and ros config get returns nothing sudo ros config set network dns domain myexampledomain com sudo ros config get network dns domain | 0 |
502,091 | 14,539,988,527 | IssuesEvent | 2020-12-15 12:41:27 | guardicore/monkey | https://api.github.com/repos/guardicore/monkey | opened | Remove support for manually uploaded config. | Complexity: Low Feature removal Priority: Low | ## Describe the bug
There is an unmaintained(and probably unused) feature which allows users to provide configuration file path via cmd arguments.
I think this feature is heritage of days long gone and should be removed.
## Changes to code:
Steps to reproduce the behavior:
1. Remove config argument and related parsing
2. Remove example config file
| 1.0 | Remove support for manually uploaded config. - ## Describe the bug
There is an unmaintained(and probably unused) feature which allows users to provide configuration file path via cmd arguments.
I think this feature is heritage of days long gone and should be removed.
## Changes to code:
Steps to reproduce the behavior:
1. Remove config argument and related parsing
2. Remove example config file
| priority | remove support for manually uploaded config describe the bug there is an unmaintained and probably unused feature which allows users to provide configuration file path via cmd arguments i think this feature is heritage of days long gone and should be removed changes to code steps to reproduce the behavior remove config argument and related parsing remove example config file | 1 |
653,793 | 21,626,825,373 | IssuesEvent | 2022-05-05 04:06:30 | panel-attack/panel-attack | https://api.github.com/repos/panel-attack/panel-attack | opened | Menu buttons should be customizable | enhancement Client-side low-ish priority | You should be able to customize the background color and border color in the theme
you should also be able to optionally set a background image and a focused and selected image | 1.0 | Menu buttons should be customizable - You should be able to customize the background color and border color in the theme
you should also be able to optionally set a background image and a focused and selected image | priority | menu buttons should be customizable you should be able to customize the background color and border color in the theme you should also be able to optionally set a background image and a focused and selected image | 1 |
54,604 | 6,829,865,785 | IssuesEvent | 2017-11-09 02:51:43 | swaywm/wlroots-rs | https://api.github.com/repos/swaywm/wlroots-rs | opened | DataDevice | Safe-Lib-Design | This is one of the globals that the compositor needs to create at startup.
Either it needs to be provided in the constructor (like the I/O managers) or it needs to be passed in later (like the extension protocols).
I'm up for either, leaning towards the latter since this is sort of an "extension" and it's _possible_ that a compositor may not want to implement this interface (and we should take a page out of wlroots' book and let the user decide).
Either way, it needs to live as long as the compositor and destroyed before it disconnects.
Need to also ensure only one global is registered at a time of this type.
# Functions
- [ ] `send_selection` Needs to only be possible if the DnD global has been made (so if an instance was made of the struct). Needs [SeatClient](https://replace-this-with-link-to-seat-issue) to be made first.
- [ ] `set_selection`
# DataOffer
## Events
- [ ] `destroy`
## Functions
- [ ] `dnd_actions` Return bitfield, please wrap in a `bitfield!`
- [ ] preferred_dnd_action`
- [ ] `in_ask`
- [ ] `resource` (unsafe)
- [ ] `data_source` (Return `DataSource`)
# DataSource
## Events
- [ ] `destroy`
## Functions
- [ ] `resource` (unsafe)
- [ ] `offer` (Return `DataOffer`
- [ ] `seat_client` (Return `SeatClient`)
- [ ] `mime_types` (Wrap in `Vec`)
- [ ] `accepted`
- [ ] `current_dnd_action`
- [ ] `dnd_actions` Return bitfield, like in `DataOffer`
- [ ] `compositor_actions` Return bitfield, like in `DataOffer`
- [ ] `actions_set`
## Global Callbacks
They should be callable, setting them may be unsafe but probably not. They are already initialized in global creation. Double check because this could be safe.
- [ ] `accept`
- [ ] `send`
- [ ] `cancel`
# Global Cleanup
* Not possible (yet?), see [this issue](https://github.com/swaywm/wlroots/issues/399) | 1.0 | DataDevice - This is one of the globals that the compositor needs to create at startup.
Either it needs to be provided in the constructor (like the I/O managers) or it needs to be passed in later (like the extension protocols).
I'm up for either, leaning towards the latter since this is sort of an "extension" and it's _possible_ that a compositor may not want to implement this interface (and we should take a page out of wlroots' book and let the user decide).
Either way, it needs to live as long as the compositor and destroyed before it disconnects.
Need to also ensure only one global is registered at a time of this type.
# Functions
- [ ] `send_selection` Needs to only be possible if the DnD global has been made (so if an instance was made of the struct). Needs [SeatClient](https://replace-this-with-link-to-seat-issue) to be made first.
- [ ] `set_selection`
# DataOffer
## Events
- [ ] `destroy`
## Functions
- [ ] `dnd_actions` Return bitfield, please wrap in a `bitfield!`
- [ ] preferred_dnd_action`
- [ ] `in_ask`
- [ ] `resource` (unsafe)
- [ ] `data_source` (Return `DataSource`)
# DataSource
## Events
- [ ] `destroy`
## Functions
- [ ] `resource` (unsafe)
- [ ] `offer` (Return `DataOffer`
- [ ] `seat_client` (Return `SeatClient`)
- [ ] `mime_types` (Wrap in `Vec`)
- [ ] `accepted`
- [ ] `current_dnd_action`
- [ ] `dnd_actions` Return bitfield, like in `DataOffer`
- [ ] `compositor_actions` Return bitfield, like in `DataOffer`
- [ ] `actions_set`
## Global Callbacks
They should be callable, setting them may be unsafe but probably not. They are already initialized in global creation. Double check because this could be safe.
- [ ] `accept`
- [ ] `send`
- [ ] `cancel`
# Global Cleanup
* Not possible (yet?), see [this issue](https://github.com/swaywm/wlroots/issues/399) | non_priority | datadevice this is one of the globals that the compositor needs to create at startup either it needs to be provided in the constructor like the i o managers or it needs to be passed in later like the extension protocols i m up for either leaning towards the latter since this is sort of an extension and it s possible that a compositor may not want to implement this interface and we should take a page out of wlroots book and let the user decide either way it needs to live as long as the compositor and destroyed before it disconnects need to also ensure only one global is registered at a time of this type functions send selection needs to only be possible if the dnd global has been made so if an instance was made of the struct needs to be made first set selection dataoffer events destroy functions dnd actions return bitfield please wrap in a bitfield preferred dnd action in ask resource unsafe data source return datasource datasource events destroy functions resource unsafe offer return dataoffer seat client return seatclient mime types wrap in vec accepted current dnd action dnd actions return bitfield like in dataoffer compositor actions return bitfield like in dataoffer actions set global callbacks they should be callable setting them may be unsafe but probably not they are already initialized in global creation double check because this could be safe accept send cancel global cleanup not possible yet see | 0 |
414,340 | 12,102,385,517 | IssuesEvent | 2020-04-20 16:36:43 | qutebrowser/qutebrowser | https://api.github.com/repos/qutebrowser/qutebrowser | closed | Avoid encoding parameter in search engine parameter | priority: 2 - low | Hello.
I would like to have a search engine for Internet Archive, as such:
```
ia = http://web.archive.org/web/*/{}
```
but I cannot use it with the special characters present in an URL, because `qutebrowser` encodes the parameter.
For instance, doing so:
```
:open ia https://github.com/The-Compiler/qutebrowser
```
will go to the following URL: http://web.archive.org/web/*/https%3A//github.com/The-Compiler/qutebrowser
which does not work.
So, it would be nice to have a way to avoid encoding the parameter.
Thanks.
| 1.0 | Avoid encoding parameter in search engine parameter - Hello.
I would like to have a search engine for Internet Archive, as such:
```
ia = http://web.archive.org/web/*/{}
```
but I cannot use it with the special characters present in an URL, because `qutebrowser` encodes the parameter.
For instance, doing so:
```
:open ia https://github.com/The-Compiler/qutebrowser
```
will go to the following URL: http://web.archive.org/web/*/https%3A//github.com/The-Compiler/qutebrowser
which does not work.
So, it would be nice to have a way to avoid encoding the parameter.
Thanks.
| priority | avoid encoding parameter in search engine parameter hello i would like to have a search engine for internet archive as such ia but i cannot use it with the special characters present in an url because qutebrowser encodes the parameter for instance doing so open ia will go to the following url which does not work so it would be nice to have a way to avoid encoding the parameter thanks | 1 |
151,342 | 19,648,810,994 | IssuesEvent | 2022-01-10 02:36:14 | turkdevops/angular | https://api.github.com/repos/turkdevops/angular | closed | WS-2019-0318 (High) detected in handlebars-4.4.3.tgz, handlebars-4.4.2.tgz - autoclosed | security vulnerability | ## WS-2019-0318 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>handlebars-4.4.3.tgz</b>, <b>handlebars-4.4.2.tgz</b></p></summary>
<p>
<details><summary><b>handlebars-4.4.3.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.3.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.3.tgz</a></p>
<p>Path to dependency file: angular/integration/cli-hello-world-lazy/package.json</p>
<p>Path to vulnerable library: angular/integration/cli-hello-world-lazy/node_modules/handlebars/package.json,angular/integration/cli-hello-world-lazy-rollup/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-istanbul-reporter-2.1.0.tgz (Root Library)
- istanbul-api-2.1.6.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.4.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>handlebars-4.4.2.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.2.tgz</a></p>
<p>Path to dependency file: angular/integration/cli-hello-world-ivy-i18n/package.json</p>
<p>Path to vulnerable library: angular/integration/cli-hello-world-ivy-i18n/node_modules/handlebars/package.json,angular/integration/ivy-i18n/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-istanbul-reporter-2.1.0.tgz (Root Library)
- istanbul-api-2.1.6.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.4.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/angular/commit/c6aca37f442da8c55a02d7c53ccc58100ab004f3">c6aca37f442da8c55a02d7c53ccc58100ab004f3</a></p>
<p>Found in base branch: <b>labs/router</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In "showdownjs/showdown", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates.
<p>Publish Date: 2019-10-20
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p>
<p>Release Date: 2019-10-20</p>
<p>Fix Resolution: handlebars - 4.4.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0318 (High) detected in handlebars-4.4.3.tgz, handlebars-4.4.2.tgz - autoclosed - ## WS-2019-0318 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>handlebars-4.4.3.tgz</b>, <b>handlebars-4.4.2.tgz</b></p></summary>
<p>
<details><summary><b>handlebars-4.4.3.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.3.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.3.tgz</a></p>
<p>Path to dependency file: angular/integration/cli-hello-world-lazy/package.json</p>
<p>Path to vulnerable library: angular/integration/cli-hello-world-lazy/node_modules/handlebars/package.json,angular/integration/cli-hello-world-lazy-rollup/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-istanbul-reporter-2.1.0.tgz (Root Library)
- istanbul-api-2.1.6.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.4.3.tgz** (Vulnerable Library)
</details>
<details><summary><b>handlebars-4.4.2.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.4.2.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.4.2.tgz</a></p>
<p>Path to dependency file: angular/integration/cli-hello-world-ivy-i18n/package.json</p>
<p>Path to vulnerable library: angular/integration/cli-hello-world-ivy-i18n/node_modules/handlebars/package.json,angular/integration/ivy-i18n/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-istanbul-reporter-2.1.0.tgz (Root Library)
- istanbul-api-2.1.6.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.4.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/angular/commit/c6aca37f442da8c55a02d7c53ccc58100ab004f3">c6aca37f442da8c55a02d7c53ccc58100ab004f3</a></p>
<p>Found in base branch: <b>labs/router</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In "showdownjs/showdown", versions prior to v4.4.5 are vulnerable against Regular expression Denial of Service (ReDOS) once receiving specially-crafted templates.
<p>Publish Date: 2019-10-20
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b>WS-2019-0318</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1300">https://www.npmjs.com/advisories/1300</a></p>
<p>Release Date: 2019-10-20</p>
<p>Fix Resolution: handlebars - 4.4.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws high detected in handlebars tgz handlebars tgz autoclosed ws high severity vulnerability vulnerable libraries handlebars tgz handlebars tgz handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file angular integration cli hello world lazy package json path to vulnerable library angular integration cli hello world lazy node modules handlebars package json angular integration cli hello world lazy rollup node modules handlebars package json dependency hierarchy karma coverage istanbul reporter tgz root library istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file angular integration cli hello world ivy package json path to vulnerable library angular integration cli hello world ivy node modules handlebars package json angular integration ivy node modules handlebars package json dependency hierarchy karma coverage istanbul reporter tgz root library istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href found in base branch labs router vulnerability details in showdownjs showdown versions prior to are vulnerable against regular expression denial of service redos once receiving specially crafted templates publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource | 0 |
20,410 | 10,516,884,790 | IssuesEvent | 2019-09-28 20:58:19 | nekofar/gulp-fontiran | https://api.github.com/repos/nekofar/gulp-fontiran | closed | WS-2018-0085 (High) detected in http-proxy-agent-1.0.0.tgz | security vulnerability | ## WS-2018-0085 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-proxy-agent-1.0.0.tgz</b></p></summary>
<p>An HTTP(s) proxy `http.Agent` implementation for HTTP</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-proxy-agent/-/http-proxy-agent-1.0.0.tgz">https://registry.npmjs.org/http-proxy-agent/-/http-proxy-agent-1.0.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/gulp-fontiran/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/gulp-fontiran/node_modules/http-proxy-agent/package.json</p>
<p>
Dependency Hierarchy:
- typings-2.1.1.tgz (Root Library)
- typings-core-2.3.3.tgz
- popsicle-proxy-agent-3.0.0.tgz
- :x: **http-proxy-agent-1.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nekofar/gulp-fontiran/commit/c05e05e14b3eedeed5a142729f2bcc89c44438f4">c05e05e14b3eedeed5a142729f2bcc89c44438f4</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of http-proxy-agent before 2.1.0 are vulnerable to denial of service and uninitialized memory leak when unsanitized options are passed to Buffer.
<p>Publish Date: 2018-04-25
<p>URL: <a href=https://github.com/TooTallNate/node-http-proxy-agent/blob/2.0.0/index.js#L80>WS-2018-0085</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/607">https://nodesecurity.io/advisories/607</a></p>
<p>Release Date: 2018-01-27</p>
<p>Fix Resolution: 2.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2018-0085 (High) detected in http-proxy-agent-1.0.0.tgz - ## WS-2018-0085 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>http-proxy-agent-1.0.0.tgz</b></p></summary>
<p>An HTTP(s) proxy `http.Agent` implementation for HTTP</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-proxy-agent/-/http-proxy-agent-1.0.0.tgz">https://registry.npmjs.org/http-proxy-agent/-/http-proxy-agent-1.0.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/gulp-fontiran/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/gulp-fontiran/node_modules/http-proxy-agent/package.json</p>
<p>
Dependency Hierarchy:
- typings-2.1.1.tgz (Root Library)
- typings-core-2.3.3.tgz
- popsicle-proxy-agent-3.0.0.tgz
- :x: **http-proxy-agent-1.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/nekofar/gulp-fontiran/commit/c05e05e14b3eedeed5a142729f2bcc89c44438f4">c05e05e14b3eedeed5a142729f2bcc89c44438f4</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions of http-proxy-agent before 2.1.0 are vulnerable to denial of service and uninitialized memory leak when unsanitized options are passed to Buffer.
<p>Publish Date: 2018-04-25
<p>URL: <a href=https://github.com/TooTallNate/node-http-proxy-agent/blob/2.0.0/index.js#L80>WS-2018-0085</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>8.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/607">https://nodesecurity.io/advisories/607</a></p>
<p>Release Date: 2018-01-27</p>
<p>Fix Resolution: 2.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws high detected in http proxy agent tgz ws high severity vulnerability vulnerable library http proxy agent tgz an http s proxy http agent implementation for http library home page a href path to dependency file tmp ws scm gulp fontiran package json path to vulnerable library tmp ws scm gulp fontiran node modules http proxy agent package json dependency hierarchy typings tgz root library typings core tgz popsicle proxy agent tgz x http proxy agent tgz vulnerable library found in head commit a href vulnerability details versions of http proxy agent before are vulnerable to denial of service and uninitialized memory leak when unsanitized options are passed to buffer publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
41,824 | 21,966,995,446 | IssuesEvent | 2022-05-24 21:28:19 | flutter/flutter | https://api.github.com/repos/flutter/flutter | opened | Material3 Slow down | created via performance template | ## Details
<!--
When i use material 3 is true and press to icon button many times in list all view get lagged and drop frame happened to 28hz
https://user-images.githubusercontent.com/30121703/170135403-3c6b312c-77c0-44ef-a8df-69f242fe1043.mp4
**ِAndroid:**
**Android 12:**
**Tested on mi9t and samsung A22:**
```
```
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel stable, 3.0.1, on macOS 11.5.2 20G95 darwin-x64, locale
en-EG)
[✓] Android toolchain - develop for Android devices (Android SDK version
32.1.0-rc1)
[✓] Xcode - develop for iOS and macOS (Xcode 13.2.1)
[✓] Chrome - develop for the web
[✓] Android Studio (version 2021.2)
[!] Android Studio
✗ Unable to find bundled Java version.
[✓] IntelliJ IDEA Community Edition (version 2021.3)
[✓] VS Code (version 1.67.0)
[✓] VS Code (version 1.53.2)
[✓] Connected device (5 available)
[✓] HTTP Host Availability
```
```
</details>
| True | Material3 Slow down - ## Details
<!--
When i use material 3 is true and press to icon button many times in list all view get lagged and drop frame happened to 28hz
https://user-images.githubusercontent.com/30121703/170135403-3c6b312c-77c0-44ef-a8df-69f242fe1043.mp4
**ِAndroid:**
**Android 12:**
**Tested on mi9t and samsung A22:**
```
```
Doctor summary (to see all details, run flutter doctor -v):
[✓] Flutter (Channel stable, 3.0.1, on macOS 11.5.2 20G95 darwin-x64, locale
en-EG)
[✓] Android toolchain - develop for Android devices (Android SDK version
32.1.0-rc1)
[✓] Xcode - develop for iOS and macOS (Xcode 13.2.1)
[✓] Chrome - develop for the web
[✓] Android Studio (version 2021.2)
[!] Android Studio
✗ Unable to find bundled Java version.
[✓] IntelliJ IDEA Community Edition (version 2021.3)
[✓] VS Code (version 1.67.0)
[✓] VS Code (version 1.53.2)
[✓] Connected device (5 available)
[✓] HTTP Host Availability
```
```
</details>
| non_priority | slow down details when i use material is true and press to icon button many times in list all view get lagged and drop frame happened to ِandroid android tested on and samsung doctor summary to see all details run flutter doctor v flutter channel stable on macos darwin locale en eg android toolchain develop for android devices android sdk version xcode develop for ios and macos xcode chrome develop for the web android studio version android studio ✗ unable to find bundled java version intellij idea community edition version vs code version vs code version connected device available http host availability | 0 |
239,243 | 7,787,761,447 | IssuesEvent | 2018-06-07 00:16:29 | ngeraci/ucr_archivestools | https://api.github.com/repos/ngeraci/ucr_archivestools | closed | put processed file on shared drive | enhancement priority:medium | Rather than process in-place in the Downloads folder, put processed file in standard location on shared drive.
potential issues:
* what behavior do we want for overwriting existing files? (prompt?)
* handling WRCA filenames | 1.0 | put processed file on shared drive - Rather than process in-place in the Downloads folder, put processed file in standard location on shared drive.
potential issues:
* what behavior do we want for overwriting existing files? (prompt?)
* handling WRCA filenames | priority | put processed file on shared drive rather than process in place in the downloads folder put processed file in standard location on shared drive potential issues what behavior do we want for overwriting existing files prompt handling wrca filenames | 1 |
10,674 | 3,134,322,967 | IssuesEvent | 2015-09-10 09:23:49 | Leadera/ecoman_repo | https://api.github.com/repos/Leadera/ecoman_repo | closed | Open project: system should display the project information page | enhancement Portal Project Management waiting for testing | Hi, When the user open a project, the system display the screen with the large picture of Ankara (I guess). Then the user does not know what to do...
Please change it so that the system display direcly the related project page. See image below.

| 1.0 | Open project: system should display the project information page - Hi, When the user open a project, the system display the screen with the large picture of Ankara (I guess). Then the user does not know what to do...
Please change it so that the system display direcly the related project page. See image below.

| non_priority | open project system should display the project information page hi when the user open a project the system display the screen with the large picture of ankara i guess then the user does not know what to do please change it so that the system display direcly the related project page see image below | 0 |
365,701 | 10,790,775,970 | IssuesEvent | 2019-11-05 15:33:00 | kuzzleio/kuzzle | https://api.github.com/repos/kuzzleio/kuzzle | closed | Use Map/Set in HotelClerk and DSL in critical sections of code | enhancement priority-normal | Profiling analysis performed on real-time activites, including (un)subscriptions, reveal that a large part of the time consumed comes from `Object.keys`.
This is not surprising considering the number of uses of this command in our code. But the HotelClerk and the DSL modules both use large objects as repositories, and large amounts of Object.keys are performed on these structures.
Considering the following benchmarks, an analysis should be conducted on our code to track down objects primarily used as key repositories (meaning that Object.keys is far more used than random accesses), and change them to Map (or Set for objects containing only a set of key-`true` pairs)
```
Object.keys (slow obj) + walk x 192,417 ops/sec ±0.52% (95 runs sampled)
Object.keys (fast obj) + walk x 7,355,807 ops/sec ±0.48% (92 runs sampled)
Map.entries (slow obj) + walk x 2,225,232 ops/sec ±0.63% (90 runs sampled)
Set.values (slow obj) + walk x 2,378,888 ops/sec ±0.48% (93 runs sampled)
random access (slow obj) x 23,120,065 ops/sec ±0.48% (88 runs sampled)
random access (fast obj) x 17,391,844 ops/sec ±0.73% (90 runs sampled)
random access (map.get) x 15,694,340 ops/sec ±0.17% (90 runs sampled)
random access (set.has) x 15,837,436 ops/sec ±0.44% (94 runs sampled)
```
| 1.0 | Use Map/Set in HotelClerk and DSL in critical sections of code - Profiling analysis performed on real-time activites, including (un)subscriptions, reveal that a large part of the time consumed comes from `Object.keys`.
This is not surprising considering the number of uses of this command in our code. But the HotelClerk and the DSL modules both use large objects as repositories, and large amounts of Object.keys are performed on these structures.
Considering the following benchmarks, an analysis should be conducted on our code to track down objects primarily used as key repositories (meaning that Object.keys is far more used than random accesses), and change them to Map (or Set for objects containing only a set of key-`true` pairs)
```
Object.keys (slow obj) + walk x 192,417 ops/sec ±0.52% (95 runs sampled)
Object.keys (fast obj) + walk x 7,355,807 ops/sec ±0.48% (92 runs sampled)
Map.entries (slow obj) + walk x 2,225,232 ops/sec ±0.63% (90 runs sampled)
Set.values (slow obj) + walk x 2,378,888 ops/sec ±0.48% (93 runs sampled)
random access (slow obj) x 23,120,065 ops/sec ±0.48% (88 runs sampled)
random access (fast obj) x 17,391,844 ops/sec ±0.73% (90 runs sampled)
random access (map.get) x 15,694,340 ops/sec ±0.17% (90 runs sampled)
random access (set.has) x 15,837,436 ops/sec ±0.44% (94 runs sampled)
```
| priority | use map set in hotelclerk and dsl in critical sections of code profiling analysis performed on real time activites including un subscriptions reveal that a large part of the time consumed comes from object keys this is not surprising considering the number of uses of this command in our code but the hotelclerk and the dsl modules both use large objects as repositories and large amounts of object keys are performed on these structures considering the following benchmarks an analysis should be conducted on our code to track down objects primarily used as key repositories meaning that object keys is far more used than random accesses and change them to map or set for objects containing only a set of key true pairs object keys slow obj walk x ops sec ± runs sampled object keys fast obj walk x ops sec ± runs sampled map entries slow obj walk x ops sec ± runs sampled set values slow obj walk x ops sec ± runs sampled random access slow obj x ops sec ± runs sampled random access fast obj x ops sec ± runs sampled random access map get x ops sec ± runs sampled random access set has x ops sec ± runs sampled | 1 |
218,078 | 16,938,104,411 | IssuesEvent | 2021-06-27 00:43:10 | backend-br/vagas | https://api.github.com/repos/backend-br/vagas | closed | [Brasil] Java Developer @ Instituto Recôncavo de Tecnologia | CLT Docker Especialista Java PJ Pleno Remoto Rest SQL Stale Testes Unitários | <!--
==================================================
Caso a vaga for remoto durante a pandemia informar no texto "Remoto durante o covid"
==================================================
-->
<!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA BACK-END!
Não faça distinção de gênero no título da vaga.
Use: "Back-End Developer" ao invés de
"Desenvolvedor Back-End" \o/
[Brasil] Java Developer @ Instituto Recôncavo de Tecnologia
==================================================
-->
<!--
==================================================
Caso a vaga for remoto durante a pandemia deixar a linha abaixo
==================================================
-->
> Vaga Remota
## Nossa empresa
Somos um dos maiores institutos privados do Brasil, com um corpo técnico de mais de 120 profissionais, e estamos entre os mais inovadores no desenvolvimento de soluções criativas que dinamizam e integram as tarefas e os negócios dos nossos clientes e parceiros.
## Descrição da vaga
Analista Desenvolvimento JAVA
## Local
Remota, Salvador-BA
## Requisitos
**Obrigatórios:**
- Experiência em desenvolvimento Java;
- Conhecimento Modelagem de sistemas;
- Conhecimento em desenvolvimento com SpringBoot;
- Conhecimento em desenvolvimento com Angula 7+;
- Conhecimento em banco de dados MS SQL Server;
- Conhecimento em desenvolvimento de API REST
**Desejáveis:**
- Conhecimentos em testes unitários;
- Conhecimento Docker.
## Benefícios
Benefícios para CLT
- Plano de saúde
- Plano Odontológico
- Vale Refeição
- Bônus Semestral
## Contratação
CLT ou PJ a combinar
## Como se candidatar
Por favor envie um email para vagas.inovatech@gmail.com com seu CV anexado - enviar no assunto: Vaga Analista Desenvolvimento JAVA
## Tempo médio de feedbacks
Costumamos enviar feedbacks em até 07 dias após cada processo.
E-mail para contato em caso de não haver resposta: vagas.inovatech@gmail.com
#### Alocação
- Remoto
#### Regime
- CLT
- PJ
#### Nível
- Júnior
- Pleno
- Sênior
- Especialista
| 1.0 | [Brasil] Java Developer @ Instituto Recôncavo de Tecnologia - <!--
==================================================
Caso a vaga for remoto durante a pandemia informar no texto "Remoto durante o covid"
==================================================
-->
<!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA BACK-END!
Não faça distinção de gênero no título da vaga.
Use: "Back-End Developer" ao invés de
"Desenvolvedor Back-End" \o/
[Brasil] Java Developer @ Instituto Recôncavo de Tecnologia
==================================================
-->
<!--
==================================================
Caso a vaga for remoto durante a pandemia deixar a linha abaixo
==================================================
-->
> Vaga Remota
## Nossa empresa
Somos um dos maiores institutos privados do Brasil, com um corpo técnico de mais de 120 profissionais, e estamos entre os mais inovadores no desenvolvimento de soluções criativas que dinamizam e integram as tarefas e os negócios dos nossos clientes e parceiros.
## Descrição da vaga
Analista Desenvolvimento JAVA
## Local
Remota, Salvador-BA
## Requisitos
**Obrigatórios:**
- Experiência em desenvolvimento Java;
- Conhecimento Modelagem de sistemas;
- Conhecimento em desenvolvimento com SpringBoot;
- Conhecimento em desenvolvimento com Angula 7+;
- Conhecimento em banco de dados MS SQL Server;
- Conhecimento em desenvolvimento de API REST
**Desejáveis:**
- Conhecimentos em testes unitários;
- Conhecimento Docker.
## Benefícios
Benefícios para CLT
- Plano de saúde
- Plano Odontológico
- Vale Refeição
- Bônus Semestral
## Contratação
CLT ou PJ a combinar
## Como se candidatar
Por favor envie um email para vagas.inovatech@gmail.com com seu CV anexado - enviar no assunto: Vaga Analista Desenvolvimento JAVA
## Tempo médio de feedbacks
Costumamos enviar feedbacks em até 07 dias após cada processo.
E-mail para contato em caso de não haver resposta: vagas.inovatech@gmail.com
#### Alocação
- Remoto
#### Regime
- CLT
- PJ
#### Nível
- Júnior
- Pleno
- Sênior
- Especialista
| non_priority | java developer instituto recôncavo de tecnologia caso a vaga for remoto durante a pandemia informar no texto remoto durante o covid por favor só poste se a vaga for para back end não faça distinção de gênero no título da vaga use back end developer ao invés de desenvolvedor back end o java developer instituto recôncavo de tecnologia caso a vaga for remoto durante a pandemia deixar a linha abaixo vaga remota nossa empresa somos um dos maiores institutos privados do brasil com um corpo técnico de mais de profissionais e estamos entre os mais inovadores no desenvolvimento de soluções criativas que dinamizam e integram as tarefas e os negócios dos nossos clientes e parceiros descrição da vaga analista desenvolvimento java local remota salvador ba requisitos obrigatórios experiência em desenvolvimento java conhecimento modelagem de sistemas conhecimento em desenvolvimento com springboot conhecimento em desenvolvimento com angula conhecimento em banco de dados ms sql server conhecimento em desenvolvimento de api rest desejáveis conhecimentos em testes unitários conhecimento docker benefícios benefícios para clt plano de saúde plano odontológico vale refeição bônus semestral contratação clt ou pj a combinar como se candidatar por favor envie um email para vagas inovatech gmail com com seu cv anexado enviar no assunto vaga analista desenvolvimento java tempo médio de feedbacks costumamos enviar feedbacks em até dias após cada processo e mail para contato em caso de não haver resposta vagas inovatech gmail com alocação remoto regime clt pj nível júnior pleno sênior especialista | 0 |
818,337 | 30,684,305,281 | IssuesEvent | 2023-07-26 11:16:21 | wso2/product-is | https://api.github.com/repos/wso2/product-is | closed | Cannot disable Google onetap feature from carbon console. | Priority/Highest bug | **Describe the issue:**
Once a Google IDP connection is created and google one tap is enabled from the carbon console, Getting an error message as below once we try to disable Google one-tap
```
Error while updating Identity Provider information. IdentityProviderMgtServiceIdentityProviderManagementExceptionException
```
<img width="1512" alt="Screenshot 2023-06-23 at 12 30 33" src="https://github.com/wso2/product-is/assets/41188245/bd4a92f6-0a2d-4fcc-a8b2-6966c05a60ac">
https://github.com/wso2/product-is/assets/41188245/0c3f956b-4e3d-450a-9520-feee34527fcc
**How to reproduce:**
1. Create a Google IDP and enable one tap. https://is.docs.wso2.com/en/latest/guides/identity-federation/google/#sign-in-with-google-one-tap
3. Create a service provider and add Google as an authentication step
4. Go to the Google IDP from the carbon console and try to uncheck the Google one-tap check box.
5. An error message will pop up.
**Expected behavior:**
<!-- A clear and concise description of what you expected to happen. -->
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: [e.g., IS 5.10.0, IS 5.9.0] - IS 6.2.0-alpha
- OS: [e.g., Windows, Linux, Mac] - Mac
- Database: [e.g., MySQL, H2] - H2 / MSSQL
- Userstore: [e.g., LDAP, JDBC]
---
| 1.0 | Cannot disable Google onetap feature from carbon console. - **Describe the issue:**
Once a Google IDP connection is created and google one tap is enabled from the carbon console, Getting an error message as below once we try to disable Google one-tap
```
Error while updating Identity Provider information. IdentityProviderMgtServiceIdentityProviderManagementExceptionException
```
<img width="1512" alt="Screenshot 2023-06-23 at 12 30 33" src="https://github.com/wso2/product-is/assets/41188245/bd4a92f6-0a2d-4fcc-a8b2-6966c05a60ac">
https://github.com/wso2/product-is/assets/41188245/0c3f956b-4e3d-450a-9520-feee34527fcc
**How to reproduce:**
1. Create a Google IDP and enable one tap. https://is.docs.wso2.com/en/latest/guides/identity-federation/google/#sign-in-with-google-one-tap
3. Create a service provider and add Google as an authentication step
4. Go to the Google IDP from the carbon console and try to uncheck the Google one-tap check box.
5. An error message will pop up.
**Expected behavior:**
<!-- A clear and concise description of what you expected to happen. -->
**Environment information** (_Please complete the following information; remove any unnecessary fields_) **:**
- Product Version: [e.g., IS 5.10.0, IS 5.9.0] - IS 6.2.0-alpha
- OS: [e.g., Windows, Linux, Mac] - Mac
- Database: [e.g., MySQL, H2] - H2 / MSSQL
- Userstore: [e.g., LDAP, JDBC]
---
| priority | cannot disable google onetap feature from carbon console describe the issue once a google idp connection is created and google one tap is enabled from the carbon console getting an error message as below once we try to disable google one tap error while updating identity provider information identityprovidermgtserviceidentityprovidermanagementexceptionexception img width alt screenshot at src how to reproduce create a google idp and enable one tap create a service provider and add google as an authentication step go to the google idp from the carbon console and try to uncheck the google one tap check box an error message will pop up expected behavior environment information please complete the following information remove any unnecessary fields product version is alpha os mac database mssql userstore | 1 |
714,092 | 24,550,336,929 | IssuesEvent | 2022-10-12 12:07:36 | Sequel-Ace/Sequel-Ace | https://api.github.com/repos/Sequel-Ace/Sequel-Ace | closed | Allowing a query history > 100 | Feature Request stale Low priority | The limit in the preferences is 100, whereas I would like to essentially save all queries I've ever run, which may exceed that. I would like to be able to set some large number instead, e.g. 1,000,000 | 1.0 | Allowing a query history > 100 - The limit in the preferences is 100, whereas I would like to essentially save all queries I've ever run, which may exceed that. I would like to be able to set some large number instead, e.g. 1,000,000 | priority | allowing a query history the limit in the preferences is whereas i would like to essentially save all queries i ve ever run which may exceed that i would like to be able to set some large number instead e g | 1 |
292,340 | 21,963,483,546 | IssuesEvent | 2022-05-24 17:49:07 | gravitational/teleport | https://api.github.com/repos/gravitational/teleport | closed | Reorganize Getting Started section. | documentation time-to-value | ## Details
The current **Getting Started** section houses two different types of doc. The Linux server guide is focused on installation and initial steps, while the docker-compose is more of a tutorial-style introduction to Teleport in a preconfigured environment.
Move docker-compose to a separate tutorial (or Teleport lab) section and keep the Getting Started section focused on installation and initial steps.
The final section would be:
**Getting started**
- [Linux Server](https://goteleport.com/docs/getting-started/linux-server/) (current one)
- [Kubernetes Cluster](https://goteleport.com/docs/kubernetes-access/getting-started/cluster/) (move from Kubernetes Access as this is also related to installation and getting started on Kubernetes)
**Tutorial (or Teleport lab)**
- [docker-compose](https://goteleport.com/docs/getting-started/docker-compose/)
- [minikube based teleport lab](https://github.com/gravitational/teleport/issues/9359)
### Category
- Improve Existing
| 1.0 | Reorganize Getting Started section. - ## Details
The current **Getting Started** section houses two different types of doc. The Linux server guide is focused on installation and initial steps, while the docker-compose is more of a tutorial-style introduction to Teleport in a preconfigured environment.
Move docker-compose to a separate tutorial (or Teleport lab) section and keep the Getting Started section focused on installation and initial steps.
The final section would be:
**Getting started**
- [Linux Server](https://goteleport.com/docs/getting-started/linux-server/) (current one)
- [Kubernetes Cluster](https://goteleport.com/docs/kubernetes-access/getting-started/cluster/) (move from Kubernetes Access as this is also related to installation and getting started on Kubernetes)
**Tutorial (or Teleport lab)**
- [docker-compose](https://goteleport.com/docs/getting-started/docker-compose/)
- [minikube based teleport lab](https://github.com/gravitational/teleport/issues/9359)
### Category
- Improve Existing
| non_priority | reorganize getting started section details the current getting started section houses two different types of doc the linux server guide is focused on installation and initial steps while the docker compose is more of a tutorial style introduction to teleport in a preconfigured environment move docker compose to a separate tutorial or teleport lab section and keep the getting started section focused on installation and initial steps the final section would be getting started current one move from kubernetes access as this is also related to installation and getting started on kubernetes tutorial or teleport lab category improve existing | 0 |
735,968 | 25,450,924,254 | IssuesEvent | 2022-11-24 10:22:43 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.msn.com - see bug description | browser-firefox priority-critical engine-gecko | <!-- @browser: Firefox 107.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:107.0) Gecko/20100101 Firefox/107.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/114556 -->
**URL**: https://www.msn.com/en-us/news/politics/goodbye-cannon-experts-say-judge-s-controversial-pro-trump-order-doomed-after-hostile-hearing/ar-AA14sCt0?cvid=e8af6392f2a4436e812f45019ae71187
**Browser / Version**: Firefox 107.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: Ads overlapping page content
**Steps to Reproduce**:
After recent update to version 107.0 (64-bit), ads now overlap content making content unreadable. The same pages render without problems when using Chrome Version 107.0.5304.107 (Official Build) (64-bit).
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/11/0bf4bd5f-f9e4-4e90-8e68-210928520b70.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.msn.com - see bug description - <!-- @browser: Firefox 107.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:107.0) Gecko/20100101 Firefox/107.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/114556 -->
**URL**: https://www.msn.com/en-us/news/politics/goodbye-cannon-experts-say-judge-s-controversial-pro-trump-order-doomed-after-hostile-hearing/ar-AA14sCt0?cvid=e8af6392f2a4436e812f45019ae71187
**Browser / Version**: Firefox 107.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Something else
**Description**: Ads overlapping page content
**Steps to Reproduce**:
After recent update to version 107.0 (64-bit), ads now overlap content making content unreadable. The same pages render without problems when using Chrome Version 107.0.5304.107 (Official Build) (64-bit).
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2022/11/0bf4bd5f-f9e4-4e90-8e68-210928520b70.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | priority | see bug description url browser version firefox operating system windows tested another browser yes chrome problem type something else description ads overlapping page content steps to reproduce after recent update to version bit ads now overlap content making content unreadable the same pages render without problems when using chrome version official build bit view the screenshot img alt screenshot src browser configuration none from with ❤️ | 1 |
180,479 | 21,625,746,060 | IssuesEvent | 2022-05-05 01:43:40 | michaeldotson/raevin | https://api.github.com/repos/michaeldotson/raevin | opened | CVE-2020-15169 (Medium) detected in actionview-5.2.3.gem | security vulnerability | ## CVE-2020-15169 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionview-5.2.3.gem</b></p></summary>
<p>Simple, battle-tested conventions and helpers for building web pages.</p>
<p>Library home page: <a href="https://rubygems.org/gems/actionview-5.2.3.gem">https://rubygems.org/gems/actionview-5.2.3.gem</a></p>
<p>Path to dependency file: /raevin/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/actionview-5.2.3.gem</p>
<p>
Dependency Hierarchy:
- sass-rails-5.0.7.gem (Root Library)
- sprockets-rails-3.2.1.gem
- actionpack-5.2.3.gem
- :x: **actionview-5.2.3.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Action View before versions 5.2.4.4 and 6.0.3.3 there is a potential Cross-Site Scripting (XSS) vulnerability in Action View's translation helpers. Views that allow the user to control the default (not found) value of the `t` and `translate` helpers could be susceptible to XSS attacks. When an HTML-unsafe string is passed as the default for a missing translation key named html or ending in _html, the default string is incorrectly marked as HTML-safe and not escaped. This is patched in versions 6.0.3.3 and 5.2.4.4. A workaround without upgrading is proposed in the source advisory.
<p>Publish Date: 2020-09-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15169>CVE-2020-15169</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://rubygems.org/gems/actionview/versions/6.0.3.3">https://rubygems.org/gems/actionview/versions/6.0.3.3</a></p>
<p>Release Date: 2020-09-11</p>
<p>Fix Resolution: 6.0.3.3, 5.2.4.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-15169 (Medium) detected in actionview-5.2.3.gem - ## CVE-2020-15169 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>actionview-5.2.3.gem</b></p></summary>
<p>Simple, battle-tested conventions and helpers for building web pages.</p>
<p>Library home page: <a href="https://rubygems.org/gems/actionview-5.2.3.gem">https://rubygems.org/gems/actionview-5.2.3.gem</a></p>
<p>Path to dependency file: /raevin/Gemfile.lock</p>
<p>Path to vulnerable library: /var/lib/gems/2.3.0/cache/actionview-5.2.3.gem</p>
<p>
Dependency Hierarchy:
- sass-rails-5.0.7.gem (Root Library)
- sprockets-rails-3.2.1.gem
- actionpack-5.2.3.gem
- :x: **actionview-5.2.3.gem** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Action View before versions 5.2.4.4 and 6.0.3.3 there is a potential Cross-Site Scripting (XSS) vulnerability in Action View's translation helpers. Views that allow the user to control the default (not found) value of the `t` and `translate` helpers could be susceptible to XSS attacks. When an HTML-unsafe string is passed as the default for a missing translation key named html or ending in _html, the default string is incorrectly marked as HTML-safe and not escaped. This is patched in versions 6.0.3.3 and 5.2.4.4. A workaround without upgrading is proposed in the source advisory.
<p>Publish Date: 2020-09-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15169>CVE-2020-15169</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://rubygems.org/gems/actionview/versions/6.0.3.3">https://rubygems.org/gems/actionview/versions/6.0.3.3</a></p>
<p>Release Date: 2020-09-11</p>
<p>Fix Resolution: 6.0.3.3, 5.2.4.4</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in actionview gem cve medium severity vulnerability vulnerable library actionview gem simple battle tested conventions and helpers for building web pages library home page a href path to dependency file raevin gemfile lock path to vulnerable library var lib gems cache actionview gem dependency hierarchy sass rails gem root library sprockets rails gem actionpack gem x actionview gem vulnerable library vulnerability details in action view before versions and there is a potential cross site scripting xss vulnerability in action view s translation helpers views that allow the user to control the default not found value of the t and translate helpers could be susceptible to xss attacks when an html unsafe string is passed as the default for a missing translation key named html or ending in html the default string is incorrectly marked as html safe and not escaped this is patched in versions and a workaround without upgrading is proposed in the source advisory publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
120,199 | 4,786,738,875 | IssuesEvent | 2016-10-29 16:08:00 | NREL/OpenStudio-BEopt | https://api.github.com/repos/NREL/OpenStudio-BEopt | closed | Propane/oil heating coils | priority high | EnergyPlus only allows electricity and natural gas for HVAC heating coils. We'd like to accommodate other fuel types in E+ so as to avoid using EMS. | 1.0 | Propane/oil heating coils - EnergyPlus only allows electricity and natural gas for HVAC heating coils. We'd like to accommodate other fuel types in E+ so as to avoid using EMS. | priority | propane oil heating coils energyplus only allows electricity and natural gas for hvac heating coils we d like to accommodate other fuel types in e so as to avoid using ems | 1 |
797,087 | 28,137,458,287 | IssuesEvent | 2023-04-01 14:54:51 | SierraBay/SierraBay12 | https://api.github.com/repos/SierraBay/SierraBay12 | closed | Баг: у объекта flash shell отсутствует спрайт | :bug: Баг 🐂 Priority: Low | **Ckey**:
`Atufancheg`
**Шаги:**
1. распечатать в атолате flash shell
2. осмотреться
3. попытаться взять в руку через ПКМ
4. ткнуть по автолату
**Реальное поведение:**
в руку берётся(ПКМ), но слот отображается как пустой. при клике по атолату отображается "You dump the flash shell into the autolathe." в чате и автолат проигрывает анимацию поглощения
**Ожидаемое поведение:**
спрайт отображается и на автолате и в руке
<hr>
*Репорт сгенерирован автоматически*
*Автор: `Atufancheg#7790` / `260139717121540107`* | 1.0 | Баг: у объекта flash shell отсутствует спрайт - **Ckey**:
`Atufancheg`
**Шаги:**
1. распечатать в атолате flash shell
2. осмотреться
3. попытаться взять в руку через ПКМ
4. ткнуть по автолату
**Реальное поведение:**
в руку берётся(ПКМ), но слот отображается как пустой. при клике по атолату отображается "You dump the flash shell into the autolathe." в чате и автолат проигрывает анимацию поглощения
**Ожидаемое поведение:**
спрайт отображается и на автолате и в руке
<hr>
*Репорт сгенерирован автоматически*
*Автор: `Atufancheg#7790` / `260139717121540107`* | priority | баг у объекта flash shell отсутствует спрайт ckey atufancheg шаги распечатать в атолате flash shell осмотреться попытаться взять в руку через пкм ткнуть по автолату реальное поведение в руку берётся пкм но слот отображается как пустой при клике по атолату отображается you dump the flash shell into the autolathe в чате и автолат проигрывает анимацию поглощения ожидаемое поведение спрайт отображается и на автолате и в руке репорт сгенерирован автоматически автор atufancheg | 1 |
225,492 | 24,848,197,144 | IssuesEvent | 2022-10-26 17:37:49 | opensearch-project/performance-analyzer | https://api.github.com/repos/opensearch-project/performance-analyzer | closed | CVE-2022-42889 (High) detected in commons-text-1.9.jar - autoclosed | security vulnerability | ## CVE-2022-42889 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-text-1.9.jar</b></p></summary>
<p>Apache Commons Text is a library focused on algorithms working on strings.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-text">https://commons.apache.org/proper/commons-text</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-text/1.9/ba6ac8c2807490944a0a27f6f8e68fb5ed2e80e2/commons-text-1.9.jar</p>
<p>
Dependency Hierarchy:
- spotbugs-4.5.0.jar (Root Library)
- :x: **commons-text-1.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons Text performs variable interpolation, allowing properties to be dynamically evaluated and expanded. The standard format for interpolation is "${prefix:name}", where "prefix" is used to locate an instance of org.apache.commons.text.lookup.StringLookup that performs the interpolation. Starting with version 1.5 and continuing through 1.9, the set of default Lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers. These lookups are: - "script" - execute expressions using the JVM script execution engine (javax.script) - "dns" - resolve dns records - "url" - load values from urls, including from remote servers Applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used. Users are recommended to upgrade to Apache Commons Text 1.10.0, which disables the problematic interpolators by default.
<p>Publish Date: Oct 13, 2022 1:15:00 PM
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-42889>CVE-2022-42889</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2022/10/13/4">https://www.openwall.com/lists/oss-security/2022/10/13/4</a></p>
<p>Release Date: Oct 13, 2022 1:15:00 PM</p>
<p>Fix Resolution (org.apache.commons:commons-text): 1.10.0</p>
<p>Direct dependency fix Resolution (com.github.spotbugs:spotbugs): 4.5.2</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| True | CVE-2022-42889 (High) detected in commons-text-1.9.jar - autoclosed - ## CVE-2022-42889 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-text-1.9.jar</b></p></summary>
<p>Apache Commons Text is a library focused on algorithms working on strings.</p>
<p>Library home page: <a href="https://commons.apache.org/proper/commons-text">https://commons.apache.org/proper/commons-text</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.commons/commons-text/1.9/ba6ac8c2807490944a0a27f6f8e68fb5ed2e80e2/commons-text-1.9.jar</p>
<p>
Dependency Hierarchy:
- spotbugs-4.5.0.jar (Root Library)
- :x: **commons-text-1.9.jar** (Vulnerable Library)
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache Commons Text performs variable interpolation, allowing properties to be dynamically evaluated and expanded. The standard format for interpolation is "${prefix:name}", where "prefix" is used to locate an instance of org.apache.commons.text.lookup.StringLookup that performs the interpolation. Starting with version 1.5 and continuing through 1.9, the set of default Lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers. These lookups are: - "script" - execute expressions using the JVM script execution engine (javax.script) - "dns" - resolve dns records - "url" - load values from urls, including from remote servers Applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used. Users are recommended to upgrade to Apache Commons Text 1.10.0, which disables the problematic interpolators by default.
<p>Publish Date: Oct 13, 2022 1:15:00 PM
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-42889>CVE-2022-42889</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openwall.com/lists/oss-security/2022/10/13/4">https://www.openwall.com/lists/oss-security/2022/10/13/4</a></p>
<p>Release Date: Oct 13, 2022 1:15:00 PM</p>
<p>Fix Resolution (org.apache.commons:commons-text): 1.10.0</p>
<p>Direct dependency fix Resolution (com.github.spotbugs:spotbugs): 4.5.2</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| non_priority | cve high detected in commons text jar autoclosed cve high severity vulnerability vulnerable library commons text jar apache commons text is a library focused on algorithms working on strings library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files org apache commons commons text commons text jar dependency hierarchy spotbugs jar root library x commons text jar vulnerable library found in base branch main vulnerability details apache commons text performs variable interpolation allowing properties to be dynamically evaluated and expanded the standard format for interpolation is prefix name where prefix is used to locate an instance of org apache commons text lookup stringlookup that performs the interpolation starting with version and continuing through the set of default lookup instances included interpolators that could result in arbitrary code execution or contact with remote servers these lookups are script execute expressions using the jvm script execution engine javax script dns resolve dns records url load values from urls including from remote servers applications using the interpolation defaults in the affected versions may be vulnerable to remote code execution or unintentional contact with remote servers if untrusted configuration values are used users are recommended to upgrade to apache commons text which disables the problematic interpolators by default publish date oct pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date oct pm fix resolution org apache commons commons text direct dependency fix resolution com github spotbugs spotbugs check this box to open an automated fix pr | 0 |
157,301 | 5,997,124,168 | IssuesEvent | 2017-06-03 20:42:36 | CAGoodman/CareWheelsCorp | https://api.github.com/repos/CAGoodman/CareWheelsCorp | opened | Need to update the email address in the lost credentials link | bug High Priority | We will use support@carewheels.com | 1.0 | Need to update the email address in the lost credentials link - We will use support@carewheels.com | priority | need to update the email address in the lost credentials link we will use support carewheels com | 1 |
139,541 | 5,378,196,504 | IssuesEvent | 2017-02-23 14:23:38 | juju/docs | https://api.github.com/repos/juju/docs | closed | juju gui command changes | 2.1 high priority | The juju gui command has changed. no-browser is now the default behaviour. It also prints the login credential by default. There is a new --hide-credential option not to show the credential.
The --no-browser option is supported but deprecated (it is effectively a no-op). To bring up a browser, use the --browser option.
Snippet from juju gui --help
Details:
Print the Juju GUI URL and show admin credential to use to log into it:
juju gui
Print the Juju GUI URL only:
juju gui --hide-credential
Open the Juju GUI in the default browser and show admin credential to use to log into it:
juju gui --browser
Open the Juju GUI in the default browser without printing the login credential:
juju gui --hide-credential --browser
| 1.0 | juju gui command changes - The juju gui command has changed. no-browser is now the default behaviour. It also prints the login credential by default. There is a new --hide-credential option not to show the credential.
The --no-browser option is supported but deprecated (it is effectively a no-op). To bring up a browser, use the --browser option.
Snippet from juju gui --help
Details:
Print the Juju GUI URL and show admin credential to use to log into it:
juju gui
Print the Juju GUI URL only:
juju gui --hide-credential
Open the Juju GUI in the default browser and show admin credential to use to log into it:
juju gui --browser
Open the Juju GUI in the default browser without printing the login credential:
juju gui --hide-credential --browser
| priority | juju gui command changes the juju gui command has changed no browser is now the default behaviour it also prints the login credential by default there is a new hide credential option not to show the credential the no browser option is supported but deprecated it is effectively a no op to bring up a browser use the browser option snippet from juju gui help details print the juju gui url and show admin credential to use to log into it juju gui print the juju gui url only juju gui hide credential open the juju gui in the default browser and show admin credential to use to log into it juju gui browser open the juju gui in the default browser without printing the login credential juju gui hide credential browser | 1 |
218,344 | 16,985,080,092 | IssuesEvent | 2021-06-30 13:34:49 | mysteriumnetwork/node | https://api.github.com/repos/mysteriumnetwork/node | closed | Remove proposal listening on NATS | component:node testnet3 | Currently, all consumers listen for proposals on nats. While it's a nice fallback to have, it creates a rather large load for limited gains. | 1.0 | Remove proposal listening on NATS - Currently, all consumers listen for proposals on nats. While it's a nice fallback to have, it creates a rather large load for limited gains. | non_priority | remove proposal listening on nats currently all consumers listen for proposals on nats while it s a nice fallback to have it creates a rather large load for limited gains | 0 |
284,766 | 24,622,911,980 | IssuesEvent | 2022-10-16 06:04:23 | nodejs/node | https://api.github.com/repos/nodejs/node | reopened | test_runner: update output TAP format to follow TAP 14 specs | test_runner | ### What steps will reproduce the bug?
- create an `index.js` file with a few test cases:
```javascript
const test = require('node:test');
const assert = require('node:assert');
test('top-level test 1', async (t) => {
await t.test('level 1.1', () => {});
});
test('top-level test 2', () => {});
```
- run the test: `node index.js` (note: `node --test index.js` gives an incorrect result)
### How often does it reproduce? Is there a required condition?
_No response_
### What is the expected behavior?
Even though the current output is syntactically valid TAP output. The expected output would be a TAP format that [follows specs 14.](http://testanything.org/tap-version-14-specification.html)
```
TAP version 14
# Subtest: top-level test 1
ok 1 - sub test level 1.1
1..1
ok 1 - top-level test 1
ok 2 - top-level test 2
1..2
```
Note the version header `TAP version 14` (required by TAP14).
### What do you see instead?
We currently output the following format (removed diagnostics for readability):
```
TAP version 13
# Subtest: top-level test 1
# Subtest: sub-test level 1.1
ok 1 - sub-test level 1.1
1..1
ok 1 - top-level test 1
1..1
```
Top-level tests are incorrectly output as subtests: `# Subtest: top-level test 1` and `# Subtest: top-level test 2`
### Additional information
As a reference, using node-tap (v16.3.0) gives the following output:
```
TAP version 13
# Subtest: top-level test 1
ok 1 - sub test level 1.1
1..1
ok 1 - top-level test 1
ok 2 - top-level test 2
1..2
```
Here is the diff:
```diff
TAP version 13
# Subtest: top-level test 1
- # Subtest: sub test level 1.1
ok 1 - sub test level 1.1
1..1
ok 1 - top-level test 1
+ ok 2 - top-level test 2
-1..1
+1..2
```
Related: https://github.com/nodejs/node/pull/43417 https://github.com/nodejs/node/pull/43525 | 1.0 | test_runner: update output TAP format to follow TAP 14 specs - ### What steps will reproduce the bug?
- create an `index.js` file with a few test cases:
```javascript
const test = require('node:test');
const assert = require('node:assert');
test('top-level test 1', async (t) => {
await t.test('level 1.1', () => {});
});
test('top-level test 2', () => {});
```
- run the test: `node index.js` (note: `node --test index.js` gives an incorrect result)
### How often does it reproduce? Is there a required condition?
_No response_
### What is the expected behavior?
Even though the current output is syntactically valid TAP output. The expected output would be a TAP format that [follows specs 14.](http://testanything.org/tap-version-14-specification.html)
```
TAP version 14
# Subtest: top-level test 1
ok 1 - sub test level 1.1
1..1
ok 1 - top-level test 1
ok 2 - top-level test 2
1..2
```
Note the version header `TAP version 14` (required by TAP14).
### What do you see instead?
We currently output the following format (removed diagnostics for readability):
```
TAP version 13
# Subtest: top-level test 1
# Subtest: sub-test level 1.1
ok 1 - sub-test level 1.1
1..1
ok 1 - top-level test 1
1..1
```
Top-level tests are incorrectly output as subtests: `# Subtest: top-level test 1` and `# Subtest: top-level test 2`
### Additional information
As a reference, using node-tap (v16.3.0) gives the following output:
```
TAP version 13
# Subtest: top-level test 1
ok 1 - sub test level 1.1
1..1
ok 1 - top-level test 1
ok 2 - top-level test 2
1..2
```
Here is the diff:
```diff
TAP version 13
# Subtest: top-level test 1
- # Subtest: sub test level 1.1
ok 1 - sub test level 1.1
1..1
ok 1 - top-level test 1
+ ok 2 - top-level test 2
-1..1
+1..2
```
Related: https://github.com/nodejs/node/pull/43417 https://github.com/nodejs/node/pull/43525 | non_priority | test runner update output tap format to follow tap specs what steps will reproduce the bug create an index js file with a few test cases javascript const test require node test const assert require node assert test top level test async t await t test level test top level test run the test node index js note node test index js gives an incorrect result how often does it reproduce is there a required condition no response what is the expected behavior even though the current output is syntactically valid tap output the expected output would be a tap format that tap version subtest top level test ok sub test level ok top level test ok top level test note the version header tap version required by what do you see instead we currently output the following format removed diagnostics for readability tap version subtest top level test subtest sub test level ok sub test level ok top level test top level tests are incorrectly output as subtests subtest top level test and subtest top level test additional information as a reference using node tap gives the following output tap version subtest top level test ok sub test level ok top level test ok top level test here is the diff diff tap version subtest top level test subtest sub test level ok sub test level ok top level test ok top level test related | 0 |
332,000 | 10,083,079,215 | IssuesEvent | 2019-07-25 12:54:04 | ME-ICA/tedana | https://api.github.com/repos/ME-ICA/tedana | opened | Topics for Hackathon 2019 | high-priority | <!--
This is a suggested issue template for tedana.
If there is other information that would be helpful to include, please do not hesitate to add it!
Before submitting, please check to make sure that the issue is not already addressed; if there is a related issue, then please cross-reference it by #.
If this is a usage question, please check out NeuroStars here:
https://neurostars.org/
and tag your topic with "multi-echo"
-->
<!--
Summarize the issue in 1-2 sentences, linking other issues if they are relevant
Note: simply typing # will prompt you for open issues to select from
-->
### The Story So Far
We are planning a hackathon November 6-8 in Bethesda, Maryland. We will use this issue to discuss what possible items could be added to its agenda. This will be kept up to date below. Items with ✅ are approved by consensus and struck-out items were proposed but agreed not to be done.
### Proposed Items
- Lecture/group discussion on math steps
- Decision Tree rewrite
<!--
If needed, add additional detail for:
1. Recreating a bug/problem
2. Any additional context necessary to understand the issue
-->
<!--
If desired, add suggested next steps.
If you foresee them in a particular order or priority, please use numbering
-->
<!--
Thank you for submitting your issue!
If you do not receive a response within a calendar week, please post a comment on this issue to catch our attention.
Some issues may not be resolved right away due to the volunteer nature of the project; thank you for your patience!
-->
| 1.0 | Topics for Hackathon 2019 - <!--
This is a suggested issue template for tedana.
If there is other information that would be helpful to include, please do not hesitate to add it!
Before submitting, please check to make sure that the issue is not already addressed; if there is a related issue, then please cross-reference it by #.
If this is a usage question, please check out NeuroStars here:
https://neurostars.org/
and tag your topic with "multi-echo"
-->
<!--
Summarize the issue in 1-2 sentences, linking other issues if they are relevant
Note: simply typing # will prompt you for open issues to select from
-->
### The Story So Far
We are planning a hackathon November 6-8 in Bethesda, Maryland. We will use this issue to discuss what possible items could be added to its agenda. This will be kept up to date below. Items with ✅ are approved by consensus and struck-out items were proposed but agreed not to be done.
### Proposed Items
- Lecture/group discussion on math steps
- Decision Tree rewrite
<!--
If needed, add additional detail for:
1. Recreating a bug/problem
2. Any additional context necessary to understand the issue
-->
<!--
If desired, add suggested next steps.
If you foresee them in a particular order or priority, please use numbering
-->
<!--
Thank you for submitting your issue!
If you do not receive a response within a calendar week, please post a comment on this issue to catch our attention.
Some issues may not be resolved right away due to the volunteer nature of the project; thank you for your patience!
-->
| priority | topics for hackathon this is a suggested issue template for tedana if there is other information that would be helpful to include please do not hesitate to add it before submitting please check to make sure that the issue is not already addressed if there is a related issue then please cross reference it by if this is a usage question please check out neurostars here and tag your topic with multi echo summarize the issue in sentences linking other issues if they are relevant note simply typing will prompt you for open issues to select from the story so far we are planning a hackathon november in bethesda maryland we will use this issue to discuss what possible items could be added to its agenda this will be kept up to date below items with ✅ are approved by consensus and struck out items were proposed but agreed not to be done proposed items lecture group discussion on math steps decision tree rewrite if needed add additional detail for recreating a bug problem any additional context necessary to understand the issue if desired add suggested next steps if you foresee them in a particular order or priority please use numbering thank you for submitting your issue if you do not receive a response within a calendar week please post a comment on this issue to catch our attention some issues may not be resolved right away due to the volunteer nature of the project thank you for your patience | 1 |
519,842 | 15,058,036,207 | IssuesEvent | 2021-02-03 22:41:45 | dietterc/SEO-ker | https://api.github.com/repos/dietterc/SEO-ker | opened | Feature 4: Integrate with Google Trends API | feature feature 4 high priority | The game relies on counting google searches, so it should be referring back to historic search data. | 1.0 | Feature 4: Integrate with Google Trends API - The game relies on counting google searches, so it should be referring back to historic search data. | priority | feature integrate with google trends api the game relies on counting google searches so it should be referring back to historic search data | 1 |
12,304 | 19,604,036,580 | IssuesEvent | 2022-01-06 06:50:26 | renovatebot/renovate | https://api.github.com/repos/renovatebot/renovate | opened | Support updating Terragrunt-generated `terraform.lock.hcl` files | type:feature status:requirements priority-5-triage | ### What would you like Renovate to be able to do?
Terraform lock files are generated when running Terragrunt commands (because Terragrunt is just a wrapper of Terraform). These lock files are exactly the same as lock files generated by using Terraform directly, but they don't currently seem to be updated as part of lock file maintenance ([minimal repo](https://github.com/tpansino/renovatebot-terragrunt-test) demonstrating this).
I would like Renovate to update Terraform lock files, regardless of how they were generated.
### If you have any ideas on how this should be implemented, please tell us here.
I'm a senior Terraform/Terragrunt dev, but not a skilled Javascript dev. I might be able to implement the feature if someone from the Renovate dev team can explain the Terraform manager code better to me. Maybe we could schedule a 20 minute chat?
Otherwise, see #13375 , #13356 for my thoughts on how this could be implemented.
### Is this a feature you are interested in implementing yourself?
Yes | 1.0 | Support updating Terragrunt-generated `terraform.lock.hcl` files - ### What would you like Renovate to be able to do?
Terraform lock files are generated when running Terragrunt commands (because Terragrunt is just a wrapper of Terraform). These lock files are exactly the same as lock files generated by using Terraform directly, but they don't currently seem to be updated as part of lock file maintenance ([minimal repo](https://github.com/tpansino/renovatebot-terragrunt-test) demonstrating this).
I would like Renovate to update Terraform lock files, regardless of how they were generated.
### If you have any ideas on how this should be implemented, please tell us here.
I'm a senior Terraform/Terragrunt dev, but not a skilled Javascript dev. I might be able to implement the feature if someone from the Renovate dev team can explain the Terraform manager code better to me. Maybe we could schedule a 20 minute chat?
Otherwise, see #13375 , #13356 for my thoughts on how this could be implemented.
### Is this a feature you are interested in implementing yourself?
Yes | non_priority | support updating terragrunt generated terraform lock hcl files what would you like renovate to be able to do terraform lock files are generated when running terragrunt commands because terragrunt is just a wrapper of terraform these lock files are exactly the same as lock files generated by using terraform directly but they don t currently seem to be updated as part of lock file maintenance demonstrating this i would like renovate to update terraform lock files regardless of how they were generated if you have any ideas on how this should be implemented please tell us here i m a senior terraform terragrunt dev but not a skilled javascript dev i might be able to implement the feature if someone from the renovate dev team can explain the terraform manager code better to me maybe we could schedule a minute chat otherwise see for my thoughts on how this could be implemented is this a feature you are interested in implementing yourself yes | 0 |
298,857 | 25,861,330,309 | IssuesEvent | 2022-12-13 17:09:20 | ray-project/ray | https://api.github.com/repos/ray-project/ray | opened | [ml release] `air_benchmark_xgboost_cpu_10` is flaky due to memory issues | P1 triage air release-test | ### What happened + What you expected to happen
`air_benchmark_xgboost_cpu_10` fails with:
```
(raylet, ip=172.31.199.222) Spilled 3667 MiB, 16 objects, write throughput 1040 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.160.92) Spilled 2303 MiB, 10 objects, write throughput 737 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.247.167) Spilled 3753 MiB, 17 objects, write throughput 838 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.129.182) Spilled 2558 MiB, 12 objects, write throughput 643 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.205.18) Spilled 4520 MiB, 20 objects, write throughput 937 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.182.158) Spilled 2985 MiB, 13 objects, write throughput 613 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.199.222) Spilled 5373 MiB, 24 objects, write throughput 1151 MiB/s.
(raylet, ip=172.31.182.205) Spilled 2132 MiB, 9 objects, write throughput 526 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.247.167) Spilled 5544 MiB, 25 objects, write throughput 906 MiB/s.
(raylet, ip=172.31.182.158) Spilled 6141 MiB, 26 objects, write throughput 934 MiB/s.
(raylet, ip=172.31.205.18) Spilled 6056 MiB, 27 objects, write throughput 819 MiB/s.
(raylet, ip=172.31.162.203) Spilled 3753 MiB, 16 objects, write throughput 930 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.162.203) Spilled 4862 MiB, 21 objects, write throughput 954 MiB/s.
(raylet, ip=172.31.129.182) Spilled 11430 MiB, 49 objects, write throughput 799 MiB/s.
(raylet, ip=172.31.205.18) Spilled 13818 MiB, 58 objects, write throughput 885 MiB/s.
(raylet, ip=172.31.160.92) Spilled 11345 MiB, 47 objects, write throughput 754 MiB/s.
(raylet, ip=172.31.247.167) Spilled 14672 MiB, 62 objects, write throughput 847 MiB/s.
(raylet, ip=172.31.227.81) Spilled 11601 MiB, 48 objects, write throughput 710 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.182.205) Spilled 12624 MiB, 54 objects, write throughput 718 MiB/s.
(raylet, ip=172.31.227.230) Spilled 12795 MiB, 53 objects, write throughput 730 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.182.158) Spilled 15013 MiB, 62 objects, write throughput 789 MiB/s.
(raylet, ip=172.31.162.203) Spilled 8274 MiB, 36 objects, write throughput 659 MiB/s.
(raylet, ip=172.31.199.222) Spilled 15951 MiB, 68 objects, write throughput 808 MiB/s.
(raylet, ip=172.31.162.203) Spilled 16804 MiB, 71 objects, write throughput 846 MiB/s.
2022-12-12 15:03:14,340 ERROR trial_runner.py:1095 -- Trial XGBoostTrainer_c44dd_00000: Error processing event.
ray.exceptions.RayTaskError(MemoryError): ray::_Inner.train() (pid=469, ip=172.31.162.203, repr=XGBoostTrainer)
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/trainable.py", line 367, in train
raise skipped from exception_cause(skipped)
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/function_trainable.py", line 338, in entrypoint
self._status_reporter.get_checkpoint(),
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/base_trainer.py", line 480, in _trainable_func
super()._trainable_func(self._merged_config, reporter, checkpoint_dir)
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/function_trainable.py", line 652, in _trainable_func
output = fn()
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/base_trainer.py", line 390, in train_func
trainer.training_loop()
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/gbdt_trainer.py", line 298, in training_loop
**config,
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/xgboost/xgboost_trainer.py", line 84, in _train
return xgboost_ray.train(**kwargs)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/main.py", line 1414, in train
dtrain.load_data(ray_params.num_actors)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/matrix.py", line 819, in load_data
self.num_actors, self.sharding, rank=rank)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/matrix.py", line 382, in load_data
self.data, ignore=self.ignore, indices=None, **self.kwargs)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/data_sources/ray_dataset.py", line 68, in load_data
return ObjectStore.load_data(obj_refs, ignore=ignore, indices=indices)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/data_sources/object_store.py", line 32, in load_data
return Pandas.load_data(pd.concat(local_df, copy=False), ignore=ignore)
File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/util/_decorators.py", line 311, in wrapper
return func(*args, **kwargs)
File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/reshape/concat.py", line 307, in concat
return op.get_result()
File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/reshape/concat.py", line 533, in get_result
mgrs_indexers, self.new_axes, concat_axis=self.bm_axis, copy=self.copy
File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/internals/concat.py", line 216, in concatenate_managers
values = np.concatenate(vals, axis=blk.ndim - 1)
File "<__array_function__ internals>", line 6, in concatenate
numpy.core._exceptions._ArrayMemoryError: Unable to allocate 77.5 GiB for an array with shape (40, 260000000) and data type float64
```
which suggests a memory issue.
### Versions / Dependencies
master
### Reproduction script
https://buildkite.com/ray-project/release-tests-branch/builds/1251#0185086b-9aec-429b-9455-13cf78e5c0db
### Issue Severity
None | 1.0 | [ml release] `air_benchmark_xgboost_cpu_10` is flaky due to memory issues - ### What happened + What you expected to happen
`air_benchmark_xgboost_cpu_10` fails with:
```
(raylet, ip=172.31.199.222) Spilled 3667 MiB, 16 objects, write throughput 1040 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.160.92) Spilled 2303 MiB, 10 objects, write throughput 737 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.247.167) Spilled 3753 MiB, 17 objects, write throughput 838 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.129.182) Spilled 2558 MiB, 12 objects, write throughput 643 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.205.18) Spilled 4520 MiB, 20 objects, write throughput 937 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.182.158) Spilled 2985 MiB, 13 objects, write throughput 613 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.199.222) Spilled 5373 MiB, 24 objects, write throughput 1151 MiB/s.
(raylet, ip=172.31.182.205) Spilled 2132 MiB, 9 objects, write throughput 526 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.247.167) Spilled 5544 MiB, 25 objects, write throughput 906 MiB/s.
(raylet, ip=172.31.182.158) Spilled 6141 MiB, 26 objects, write throughput 934 MiB/s.
(raylet, ip=172.31.205.18) Spilled 6056 MiB, 27 objects, write throughput 819 MiB/s.
(raylet, ip=172.31.162.203) Spilled 3753 MiB, 16 objects, write throughput 930 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.162.203) Spilled 4862 MiB, 21 objects, write throughput 954 MiB/s.
(raylet, ip=172.31.129.182) Spilled 11430 MiB, 49 objects, write throughput 799 MiB/s.
(raylet, ip=172.31.205.18) Spilled 13818 MiB, 58 objects, write throughput 885 MiB/s.
(raylet, ip=172.31.160.92) Spilled 11345 MiB, 47 objects, write throughput 754 MiB/s.
(raylet, ip=172.31.247.167) Spilled 14672 MiB, 62 objects, write throughput 847 MiB/s.
(raylet, ip=172.31.227.81) Spilled 11601 MiB, 48 objects, write throughput 710 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.182.205) Spilled 12624 MiB, 54 objects, write throughput 718 MiB/s.
(raylet, ip=172.31.227.230) Spilled 12795 MiB, 53 objects, write throughput 730 MiB/s. Set RAY_verbose_spill_logs=0 to disable this message.
(raylet, ip=172.31.182.158) Spilled 15013 MiB, 62 objects, write throughput 789 MiB/s.
(raylet, ip=172.31.162.203) Spilled 8274 MiB, 36 objects, write throughput 659 MiB/s.
(raylet, ip=172.31.199.222) Spilled 15951 MiB, 68 objects, write throughput 808 MiB/s.
(raylet, ip=172.31.162.203) Spilled 16804 MiB, 71 objects, write throughput 846 MiB/s.
2022-12-12 15:03:14,340 ERROR trial_runner.py:1095 -- Trial XGBoostTrainer_c44dd_00000: Error processing event.
ray.exceptions.RayTaskError(MemoryError): ray::_Inner.train() (pid=469, ip=172.31.162.203, repr=XGBoostTrainer)
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/trainable.py", line 367, in train
raise skipped from exception_cause(skipped)
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/function_trainable.py", line 338, in entrypoint
self._status_reporter.get_checkpoint(),
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/base_trainer.py", line 480, in _trainable_func
super()._trainable_func(self._merged_config, reporter, checkpoint_dir)
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/tune/trainable/function_trainable.py", line 652, in _trainable_func
output = fn()
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/base_trainer.py", line 390, in train_func
trainer.training_loop()
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/gbdt_trainer.py", line 298, in training_loop
**config,
File "/home/ray/anaconda3/lib/python3.7/site-packages/ray/train/xgboost/xgboost_trainer.py", line 84, in _train
return xgboost_ray.train(**kwargs)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/main.py", line 1414, in train
dtrain.load_data(ray_params.num_actors)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/matrix.py", line 819, in load_data
self.num_actors, self.sharding, rank=rank)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/matrix.py", line 382, in load_data
self.data, ignore=self.ignore, indices=None, **self.kwargs)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/data_sources/ray_dataset.py", line 68, in load_data
return ObjectStore.load_data(obj_refs, ignore=ignore, indices=indices)
File "/home/ray/anaconda3/lib/python3.7/site-packages/xgboost_ray/data_sources/object_store.py", line 32, in load_data
return Pandas.load_data(pd.concat(local_df, copy=False), ignore=ignore)
File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/util/_decorators.py", line 311, in wrapper
return func(*args, **kwargs)
File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/reshape/concat.py", line 307, in concat
return op.get_result()
File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/reshape/concat.py", line 533, in get_result
mgrs_indexers, self.new_axes, concat_axis=self.bm_axis, copy=self.copy
File "/home/ray/anaconda3/lib/python3.7/site-packages/pandas/core/internals/concat.py", line 216, in concatenate_managers
values = np.concatenate(vals, axis=blk.ndim - 1)
File "<__array_function__ internals>", line 6, in concatenate
numpy.core._exceptions._ArrayMemoryError: Unable to allocate 77.5 GiB for an array with shape (40, 260000000) and data type float64
```
which suggests a memory issue.
### Versions / Dependencies
master
### Reproduction script
https://buildkite.com/ray-project/release-tests-branch/builds/1251#0185086b-9aec-429b-9455-13cf78e5c0db
### Issue Severity
None | non_priority | air benchmark xgboost cpu is flaky due to memory issues what happened what you expected to happen air benchmark xgboost cpu fails with raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s set ray verbose spill logs to disable this message raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s raylet ip spilled mib objects write throughput mib s error trial runner py trial xgboosttrainer error processing event ray exceptions raytaskerror memoryerror ray inner train pid ip repr xgboosttrainer file home ray lib site packages ray tune trainable trainable py line in train raise skipped from exception cause skipped file home ray lib site packages ray tune trainable function trainable py line in entrypoint self status reporter get checkpoint file home ray lib site packages ray train base trainer py line in trainable func super trainable func self merged config reporter checkpoint dir file home ray lib site packages ray tune trainable function trainable py line in trainable func output fn file home ray lib site packages ray train base trainer py line in train func trainer training loop file home ray lib site packages ray train gbdt trainer py line in training loop config file home ray lib site packages ray train xgboost xgboost trainer py line in train return xgboost ray train kwargs file home ray lib site packages xgboost ray main py line in train dtrain load data ray params num actors file home ray lib site packages xgboost ray matrix py line in load data self num actors self sharding rank rank file home ray lib site packages xgboost ray matrix py line in load data self data ignore self ignore indices none self kwargs file home ray lib site packages xgboost ray data sources ray dataset py line in load data return objectstore load data obj refs ignore ignore indices indices file home ray lib site packages xgboost ray data sources object store py line in load data return pandas load data pd concat local df copy false ignore ignore file home ray lib site packages pandas util decorators py line in wrapper return func args kwargs file home ray lib site packages pandas core reshape concat py line in concat return op get result file home ray lib site packages pandas core reshape concat py line in get result mgrs indexers self new axes concat axis self bm axis copy self copy file home ray lib site packages pandas core internals concat py line in concatenate managers values np concatenate vals axis blk ndim file line in concatenate numpy core exceptions arraymemoryerror unable to allocate gib for an array with shape and data type which suggests a memory issue versions dependencies master reproduction script issue severity none | 0 |
85,324 | 10,436,242,287 | IssuesEvent | 2019-09-17 19:06:36 | Programming-Engineering-Pmi-31/IntelligentCooking | https://api.github.com/repos/Programming-Engineering-Pmi-31/IntelligentCooking | closed | Describe all functional and non-functional requirements | documentation | - [x] Describe functional requirements
- [x] Describe non-functional requirements | 1.0 | Describe all functional and non-functional requirements - - [x] Describe functional requirements
- [x] Describe non-functional requirements | non_priority | describe all functional and non functional requirements describe functional requirements describe non functional requirements | 0 |
213,125 | 7,246,172,225 | IssuesEvent | 2018-02-14 20:40:48 | terascope/teraslice | https://api.github.com/repos/terascope/teraslice | closed | /ex/ID/slicer is broken | bug priority:high | This is on master with the most recent slicer related PR included.
```
curl localhost:5678/ex/6ea6e665-5549-42fc-bdb9-8403dc0dc54b/slicer
{
"error": 500,
"message": "could not get slicer statistics, error: [object Object]"
}
``` | 1.0 | /ex/ID/slicer is broken - This is on master with the most recent slicer related PR included.
```
curl localhost:5678/ex/6ea6e665-5549-42fc-bdb9-8403dc0dc54b/slicer
{
"error": 500,
"message": "could not get slicer statistics, error: [object Object]"
}
``` | priority | ex id slicer is broken this is on master with the most recent slicer related pr included curl localhost ex slicer error message could not get slicer statistics error | 1 |
4,539 | 3,037,642,781 | IssuesEvent | 2015-08-06 18:09:26 | Microsoft/PTVS | https://api.github.com/repos/Microsoft/PTVS | opened | NullReferenceException in DjangoAnalyzer.GetArg | bug Code Intelligence crash | From WER:
```
Level . . . Symbol - InstructionPointer - FunctionOffset - SourceLine - SourceFile
0 . . . microsoft_pythontools_django!Microsoft.PythonTools.Django.Project.DjangoAnalyzer.GetArg - 0x00000000 - 0x0000002C - - <<Failure related >>
1 . . . microsoft_pythontools_django!Microsoft.PythonTools.Django.Project.DjangoAnalyzer.RenderProcessor - 0x00000000 - 0x00000035 - -
2 . . . microsoft_pythontools_analysis!Microsoft.PythonTools.Analysis.Values.SpecializedCallable.Call - 0x00000000 - 0x00000038 - -
3 . . . microsoft_pythontools_analysis!Microsoft.PythonTools.Analysis.Analyzer.ExpressionEvaluator.EvaluateCall - 0x00000000 - 0x000000B4 - -
```
From my reading of the GetArg function, the most likely cause is that the call into `django.shortcuts.render` has an invalid `CallExpression` which results in `namedArgs` (see `EvaluateCall`) being an array with null entries. This occurs when an arg with null `.NameExpression` follows an arg with non-null `.Name`.
We should probably assert in `GetNamedArguments` when `args[I].NameExpression` is null, and replace it with a safe non-null value to avoid crashing callers who assume that all elements are non-null. | 1.0 | NullReferenceException in DjangoAnalyzer.GetArg - From WER:
```
Level . . . Symbol - InstructionPointer - FunctionOffset - SourceLine - SourceFile
0 . . . microsoft_pythontools_django!Microsoft.PythonTools.Django.Project.DjangoAnalyzer.GetArg - 0x00000000 - 0x0000002C - - <<Failure related >>
1 . . . microsoft_pythontools_django!Microsoft.PythonTools.Django.Project.DjangoAnalyzer.RenderProcessor - 0x00000000 - 0x00000035 - -
2 . . . microsoft_pythontools_analysis!Microsoft.PythonTools.Analysis.Values.SpecializedCallable.Call - 0x00000000 - 0x00000038 - -
3 . . . microsoft_pythontools_analysis!Microsoft.PythonTools.Analysis.Analyzer.ExpressionEvaluator.EvaluateCall - 0x00000000 - 0x000000B4 - -
```
From my reading of the GetArg function, the most likely cause is that the call into `django.shortcuts.render` has an invalid `CallExpression` which results in `namedArgs` (see `EvaluateCall`) being an array with null entries. This occurs when an arg with null `.NameExpression` follows an arg with non-null `.Name`.
We should probably assert in `GetNamedArguments` when `args[I].NameExpression` is null, and replace it with a safe non-null value to avoid crashing callers who assume that all elements are non-null. | non_priority | nullreferenceexception in djangoanalyzer getarg from wer level symbol instructionpointer functionoffset sourceline sourcefile microsoft pythontools django microsoft pythontools django project djangoanalyzer getarg microsoft pythontools django microsoft pythontools django project djangoanalyzer renderprocessor microsoft pythontools analysis microsoft pythontools analysis values specializedcallable call microsoft pythontools analysis microsoft pythontools analysis analyzer expressionevaluator evaluatecall from my reading of the getarg function the most likely cause is that the call into django shortcuts render has an invalid callexpression which results in namedargs see evaluatecall being an array with null entries this occurs when an arg with null nameexpression follows an arg with non null name we should probably assert in getnamedarguments when args nameexpression is null and replace it with a safe non null value to avoid crashing callers who assume that all elements are non null | 0 |
56,818 | 3,081,204,954 | IssuesEvent | 2015-08-22 13:45:50 | bitfighter/bitfighter | https://api.github.com/repos/bitfighter/bitfighter | closed | Bouncers bounce right through ship | 015 bug imported Priority-Critical | _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on January 20, 2011 02:51:29_
When armed with bouncers, if you are close to a wall, bouncers will pass right through your ship; instead, they should hit you as they used to.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=40_ | 1.0 | Bouncers bounce right through ship - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on January 20, 2011 02:51:29_
When armed with bouncers, if you are close to a wall, bouncers will pass right through your ship; instead, they should hit you as they used to.
_Original issue: http://code.google.com/p/bitfighter/issues/detail?id=40_ | priority | bouncers bounce right through ship from on january when armed with bouncers if you are close to a wall bouncers will pass right through your ship instead they should hit you as they used to original issue | 1 |
359,830 | 10,681,515,606 | IssuesEvent | 2019-10-22 01:09:17 | SIGBlockchain/project_aurum | https://api.github.com/repos/SIGBlockchain/project_aurum | closed | Send Block Request | Priority: Critical Type: Enhancement | Branch Context: `send_block`
Package context: `requests`
Complete the function `SendBlockRequest` that given a block, creates a POST request to the incoming block endpoint with a JSON-block in the body.
All unit tests must pass (look at `TestSendBlockRequest`)
| 1.0 | Send Block Request - Branch Context: `send_block`
Package context: `requests`
Complete the function `SendBlockRequest` that given a block, creates a POST request to the incoming block endpoint with a JSON-block in the body.
All unit tests must pass (look at `TestSendBlockRequest`)
| priority | send block request branch context send block package context requests complete the function sendblockrequest that given a block creates a post request to the incoming block endpoint with a json block in the body all unit tests must pass look at testsendblockrequest | 1 |
143,480 | 19,180,304,491 | IssuesEvent | 2021-12-04 09:05:13 | AlexRogalskiy/java-patterns | https://api.github.com/repos/AlexRogalskiy/java-patterns | opened | CVE-2018-16487 (Medium) detected in lodash-2.4.2.tgz | security vulnerability | ## CVE-2018-16487 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, & extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p>
<p>Path to dependency file: java-patterns/package.json</p>
<p>Path to vulnerable library: java-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- dockerfile_lint-0.3.4.tgz (Root Library)
- :x: **lodash-2.4.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java-patterns/commit/850df196d336da76298c7b0f9c02003e236df715">850df196d336da76298c7b0f9c02003e236df715</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.
<p>Publish Date: 2019-02-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p>
<p>Release Date: 2019-02-01</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-16487 (Medium) detected in lodash-2.4.2.tgz - ## CVE-2018-16487 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-2.4.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, & extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p>
<p>Path to dependency file: java-patterns/package.json</p>
<p>Path to vulnerable library: java-patterns/node_modules/dockerfile_lint/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- dockerfile_lint-0.3.4.tgz (Root Library)
- :x: **lodash-2.4.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/java-patterns/commit/850df196d336da76298c7b0f9c02003e236df715">850df196d336da76298c7b0f9c02003e236df715</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.
<p>Publish Date: 2019-02-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p>
<p>Release Date: 2019-02-01</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in lodash tgz cve medium severity vulnerability vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file java patterns package json path to vulnerable library java patterns node modules dockerfile lint node modules lodash package json dependency hierarchy dockerfile lint tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
821,811 | 30,837,542,644 | IssuesEvent | 2023-08-02 08:27:20 | testomatio/app | https://api.github.com/repos/testomatio/app | closed | Tests counter shows 0 tests till switching Project on Tests Coverage | bug ui\ux jira priority medium | **Describe the bug**
https://github.com/testomatio/app/assets/77803888/663528b2-0346-4f55-9df5-638d08b103d0
**To Reproduce**
Steps to reproduce the behavior:
1. open Jira plugin
2. go to Tests Coverage page
3. search for an issue
4. see 0 tests shown
5. switch to some project
6. see tests counter updated
7. switch back to All projects
8. see tests counter updated shows the actual number
**Expected behavior**
Tests Coverage tests counters show actual tests number when users opens it for the first time
| 1.0 | Tests counter shows 0 tests till switching Project on Tests Coverage - **Describe the bug**
https://github.com/testomatio/app/assets/77803888/663528b2-0346-4f55-9df5-638d08b103d0
**To Reproduce**
Steps to reproduce the behavior:
1. open Jira plugin
2. go to Tests Coverage page
3. search for an issue
4. see 0 tests shown
5. switch to some project
6. see tests counter updated
7. switch back to All projects
8. see tests counter updated shows the actual number
**Expected behavior**
Tests Coverage tests counters show actual tests number when users opens it for the first time
| priority | tests counter shows tests till switching project on tests coverage describe the bug to reproduce steps to reproduce the behavior open jira plugin go to tests coverage page search for an issue see tests shown switch to some project see tests counter updated switch back to all projects see tests counter updated shows the actual number expected behavior tests coverage tests counters show actual tests number when users opens it for the first time | 1 |
532,704 | 15,569,836,850 | IssuesEvent | 2021-03-17 01:06:24 | Sage-Bionetworks/research-benchmarking-technology | https://api.github.com/repos/Sage-Bionetworks/research-benchmarking-technology | opened | Create Google OAuth 2.0 application for ROCC | Priority: Low | I've created Google OAuth 2.0 in the past for the SageBio Collaboration Portal. | 1.0 | Create Google OAuth 2.0 application for ROCC - I've created Google OAuth 2.0 in the past for the SageBio Collaboration Portal. | priority | create google oauth application for rocc i ve created google oauth in the past for the sagebio collaboration portal | 1 |
691,154 | 23,684,943,668 | IssuesEvent | 2022-08-29 04:52:33 | okTurtles/group-income | https://api.github.com/repos/okTurtles/group-income | opened | Dashboard on mobile website too wide | Kind:Bug App:Frontend Priority:High | ### Problem
On iOS Brave:

### Solution
Find out why this is suddenly happening and fix. | 1.0 | Dashboard on mobile website too wide - ### Problem
On iOS Brave:

### Solution
Find out why this is suddenly happening and fix. | priority | dashboard on mobile website too wide problem on ios brave solution find out why this is suddenly happening and fix | 1 |
411,384 | 12,017,556,015 | IssuesEvent | 2020-04-10 18:40:29 | hikaya-io/activity | https://api.github.com/repos/hikaya-io/activity | reopened | ACT-712 Add description field (large text) to Stakeholder Type model | 1 Priority | The stakeholder type model needs a large text field called `Description` so that users can provide details about the Stakeholder Type. | 1.0 | ACT-712 Add description field (large text) to Stakeholder Type model - The stakeholder type model needs a large text field called `Description` so that users can provide details about the Stakeholder Type. | priority | act add description field large text to stakeholder type model the stakeholder type model needs a large text field called description so that users can provide details about the stakeholder type | 1 |
539,316 | 15,786,773,264 | IssuesEvent | 2021-04-01 18:14:10 | microsoft/PowerToys | https://api.github.com/repos/microsoft/PowerToys | closed | Running as user causes installer to ask for the .msi | Area-Setup/Install Issue-Bug Priority-0 Resolution-Fix-Committed | <!--
**Important: When reporting BSODs or security issues, DO NOT attach memory dumps, logs, or traces to Github issues**.
Instead, send dumps/traces to secure@microsoft.com, referencing this GitHub issue.
-->
## ℹ Computer information
- PowerToys version: 0.31.2.0
- PowerToy Utility:
- Running PowerToys as Admin: no
- Windows build number: [run "winver"] 18363.1377
## 📝 Provide detailed reproduction steps (if any)
1. …INstalled PT using local admin account
2. …Tried to run the powertoys icon from the windows menu
3. …
### ✔️ Expected result
Start the powertoys utility in the task bar
### ❌ Actual result
I ambeing prompted for the network path to the installation files
## 📷 Screenshots
_Are there any useful screenshots? WinKey+Shift+S and then just paste them directly into the form_

| 1.0 | Running as user causes installer to ask for the .msi - <!--
**Important: When reporting BSODs or security issues, DO NOT attach memory dumps, logs, or traces to Github issues**.
Instead, send dumps/traces to secure@microsoft.com, referencing this GitHub issue.
-->
## ℹ Computer information
- PowerToys version: 0.31.2.0
- PowerToy Utility:
- Running PowerToys as Admin: no
- Windows build number: [run "winver"] 18363.1377
## 📝 Provide detailed reproduction steps (if any)
1. …INstalled PT using local admin account
2. …Tried to run the powertoys icon from the windows menu
3. …
### ✔️ Expected result
Start the powertoys utility in the task bar
### ❌ Actual result
I ambeing prompted for the network path to the installation files
## 📷 Screenshots
_Are there any useful screenshots? WinKey+Shift+S and then just paste them directly into the form_

| priority | running as user causes installer to ask for the msi important when reporting bsods or security issues do not attach memory dumps logs or traces to github issues instead send dumps traces to secure microsoft com referencing this github issue ℹ computer information powertoys version powertoy utility running powertoys as admin no windows build number 📝 provide detailed reproduction steps if any …installed pt using local admin account …tried to run the powertoys icon from the windows menu … ✔️ expected result start the powertoys utility in the task bar ❌ actual result i ambeing prompted for the network path to the installation files 📷 screenshots are there any useful screenshots winkey shift s and then just paste them directly into the form | 1 |
522,897 | 15,169,197,374 | IssuesEvent | 2021-02-12 20:42:55 | NCEAS/metacat | https://api.github.com/repos/NCEAS/metacat | closed | allow metacat to store files on multiple fs | Category: metacat Component: Bugzilla-Id Priority: Normal Status: In Progress Tracker: Bug wontfix | ---
Author Name: **Matt Jones** (Matt Jones)
Original Redmine Issue: 195, https://projects.ecoinformatics.org/ecoinfo/issues/195
Original Date: 2001-04-09
Original Assignee: Jing Tao
---
Metacat currently stores files on a single file system. Need to changes this so
that Metacat can be configured to store files on multiple file systems in case
space management by the administrator requires this.
| 1.0 | allow metacat to store files on multiple fs - ---
Author Name: **Matt Jones** (Matt Jones)
Original Redmine Issue: 195, https://projects.ecoinformatics.org/ecoinfo/issues/195
Original Date: 2001-04-09
Original Assignee: Jing Tao
---
Metacat currently stores files on a single file system. Need to changes this so
that Metacat can be configured to store files on multiple file systems in case
space management by the administrator requires this.
| priority | allow metacat to store files on multiple fs author name matt jones matt jones original redmine issue original date original assignee jing tao metacat currently stores files on a single file system need to changes this so that metacat can be configured to store files on multiple file systems in case space management by the administrator requires this | 1 |
82,183 | 3,603,835,129 | IssuesEvent | 2016-02-03 20:33:17 | ualbertalib/HydraNorth | https://api.github.com/repos/ualbertalib/HydraNorth | closed | ERA_Collections: Collection-specific deposit page | enhancement priority:medium size:large | As admin I want to be able to define an identifier in the deposit form to be used for a new collection. | 1.0 | ERA_Collections: Collection-specific deposit page - As admin I want to be able to define an identifier in the deposit form to be used for a new collection. | priority | era collections collection specific deposit page as admin i want to be able to define an identifier in the deposit form to be used for a new collection | 1 |
151,931 | 5,830,082,174 | IssuesEvent | 2017-05-08 15:57:35 | emfoundation/ce100-app | https://api.github.com/repos/emfoundation/ce100-app | opened | Improved tag selection flow | primary-user priority-2 | This applies to creating/editing organisations/challenges.
Rather than having a two step process, where the user first adds/edits other content to an org profile or challenge and then clicks next to add tags, the tag selection will be more integrated.
The currently selected (if any) tags are shown in the edit form with a "Add/Edit tags" button below them. Clicking on that button opens up the tag selection screen. On that screen, clicking either "Select" or "Cancel" takes you back to the edit form. Only after clicking "Save" on that form are the changes to the tags saved.
Mock-ups to be added. | 1.0 | Improved tag selection flow - This applies to creating/editing organisations/challenges.
Rather than having a two step process, where the user first adds/edits other content to an org profile or challenge and then clicks next to add tags, the tag selection will be more integrated.
The currently selected (if any) tags are shown in the edit form with a "Add/Edit tags" button below them. Clicking on that button opens up the tag selection screen. On that screen, clicking either "Select" or "Cancel" takes you back to the edit form. Only after clicking "Save" on that form are the changes to the tags saved.
Mock-ups to be added. | priority | improved tag selection flow this applies to creating editing organisations challenges rather than having a two step process where the user first adds edits other content to an org profile or challenge and then clicks next to add tags the tag selection will be more integrated the currently selected if any tags are shown in the edit form with a add edit tags button below them clicking on that button opens up the tag selection screen on that screen clicking either select or cancel takes you back to the edit form only after clicking save on that form are the changes to the tags saved mock ups to be added | 1 |
668,264 | 22,575,956,430 | IssuesEvent | 2022-06-28 07:21:55 | OpenFunction/functions-framework-go | https://api.github.com/repos/OpenFunction/functions-framework-go | closed | [Feature]Support advanced http router and features | enhancement priority/medium |
1. use [mux](https://github.com/gorilla/mux) as the default handler in ff-go which is more production-ready and we can bring in new features - https://github.com/OpenFunction/functions-framework/issues/21#issuecomment-1150587351, https://github.com/OpenFunction/functions-framework/issues/17
2. support defining variables in the path like `/functions/<funcname>`, and use it in the function by `ofnctx.Vars()`
3. support defining method matcher like `withFunctionMethods("GET", "POST")`
| 1.0 | [Feature]Support advanced http router and features -
1. use [mux](https://github.com/gorilla/mux) as the default handler in ff-go which is more production-ready and we can bring in new features - https://github.com/OpenFunction/functions-framework/issues/21#issuecomment-1150587351, https://github.com/OpenFunction/functions-framework/issues/17
2. support defining variables in the path like `/functions/<funcname>`, and use it in the function by `ofnctx.Vars()`
3. support defining method matcher like `withFunctionMethods("GET", "POST")`
| priority | support advanced http router and features use as the default handler in ff go which is more production ready and we can bring in new features support defining variables in the path like functions and use it in the function by ofnctx vars support defining method matcher like withfunctionmethods get post | 1 |
246,816 | 7,895,650,547 | IssuesEvent | 2018-06-29 04:48:58 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | pillow snuck in to build_visit (bv_python) on the trunk | Likelihood: 3 - Occasional OS: All Priority: Normal Severity: 4 - Crash / Wrong Results Support Group: Any bug version: 2.8.2 | During a merge, I accidentally committed a change to build pillow instead of PIL in the trunk version of build_visit.
We do want to use Pillow in the future, at this moment we need test suite changes to support Pillow instead of PIL.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Cyrus Harrison
Original creation: 02/11/2015 04:44 pm
Original update: 09/02/2015 04:19 pm
Ticket number: 2144 | 1.0 | pillow snuck in to build_visit (bv_python) on the trunk - During a merge, I accidentally committed a change to build pillow instead of PIL in the trunk version of build_visit.
We do want to use Pillow in the future, at this moment we need test suite changes to support Pillow instead of PIL.
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. The following information
could not be accurately captured in the new ticket:
Original author: Cyrus Harrison
Original creation: 02/11/2015 04:44 pm
Original update: 09/02/2015 04:19 pm
Ticket number: 2144 | priority | pillow snuck in to build visit bv python on the trunk during a merge i accidentally committed a change to build pillow instead of pil in the trunk version of build visit we do want to use pillow in the future at this moment we need test suite changes to support pillow instead of pil redmine migration this ticket was migrated from redmine the following information could not be accurately captured in the new ticket original author cyrus harrison original creation pm original update pm ticket number | 1 |
301,499 | 26,052,842,941 | IssuesEvent | 2022-12-22 20:43:14 | airbytehq/airbyte | https://api.github.com/repos/airbytehq/airbyte | closed | E2E Testing Tool: Storing metrics on client | type/enhancement e2e-testing-tool team/connector-ops | https://github.com/airbytehq/airbyte/issues/14915
We need to store/collect metrics somewhere on client side for reports/charts/analyzing/representation | 1.0 | E2E Testing Tool: Storing metrics on client - https://github.com/airbytehq/airbyte/issues/14915
We need to store/collect metrics somewhere on client side for reports/charts/analyzing/representation | non_priority | testing tool storing metrics on client we need to store collect metrics somewhere on client side for reports charts analyzing representation | 0 |
89,320 | 3,792,633,480 | IssuesEvent | 2016-03-22 10:32:02 | handsontable/handsontable | https://api.github.com/repos/handsontable/handsontable | closed | Add source maps to the repo | Priority: normal | As pointed in #1009, Chrome Developer Tools show 404 (Not Found) error message about missing source maps:
```
.../jquery-handsontable/lib/jquery-1.10.2.min.map
.../jquery-handsontable/dist/json-patch-duplex.js.map
```
As solution I propose to add missing source maps and files to which they link (without altering the files that contain the source map reference)
I'd rather keep the source maps, because this is also what jQuery does in their minified, production-ready files: http://jquery.com/download/
This issue only affects Chrome users with "Enable source maps" option enabled in the Developer Tools. | 1.0 | Add source maps to the repo - As pointed in #1009, Chrome Developer Tools show 404 (Not Found) error message about missing source maps:
```
.../jquery-handsontable/lib/jquery-1.10.2.min.map
.../jquery-handsontable/dist/json-patch-duplex.js.map
```
As solution I propose to add missing source maps and files to which they link (without altering the files that contain the source map reference)
I'd rather keep the source maps, because this is also what jQuery does in their minified, production-ready files: http://jquery.com/download/
This issue only affects Chrome users with "Enable source maps" option enabled in the Developer Tools. | priority | add source maps to the repo as pointed in chrome developer tools show not found error message about missing source maps jquery handsontable lib jquery min map jquery handsontable dist json patch duplex js map as solution i propose to add missing source maps and files to which they link without altering the files that contain the source map reference i d rather keep the source maps because this is also what jquery does in their minified production ready files this issue only affects chrome users with enable source maps option enabled in the developer tools | 1 |
347,789 | 24,899,788,319 | IssuesEvent | 2022-10-28 19:32:16 | xylan-org/mailspy | https://api.github.com/repos/xylan-org/mailspy | opened | Change default path to /devtools/mailspy in README after release 1.1.0 | documentation | This includes the Spring Security related part and the config table. | 1.0 | Change default path to /devtools/mailspy in README after release 1.1.0 - This includes the Spring Security related part and the config table. | non_priority | change default path to devtools mailspy in readme after release this includes the spring security related part and the config table | 0 |
134,864 | 10,933,139,693 | IssuesEvent | 2019-11-23 23:04:42 | pandas-dev/pandas | https://api.github.com/repos/pandas-dev/pandas | closed | df.groupby('symbol')['status'].ffill() results in Exception: cannot handle a non-unique multi-index! | Needs Tests good first issue | #### Code Sample, a copy-pastable example if possible
```python
import pandas as pd
df = pd.DataFrame([['2018-01-01','MSFT','short'],
['2018-01-01','MSFT','long'],
['2018-01-01','MSFT',pd.np.NaN],
['2018-01-01','AAPL','short'],
['2018-01-01','AAPL',pd.np.NaN],
['2018-01-02','AAPL','short'],
['2018-01-01','TSLA','neutral'],
['2018-01-02','TSLA',pd.np.NaN]],
columns=['date','symbol','status'])
df['date'] = pd.to_datetime(df['date'])
df = df.set_index(['date','symbol'])
df.groupby('symbol')['status'].ffill()
```
#### Raised exception
<details>
```
---------------------------------------------------------------------------
Exception Traceback (most recent call last)
<ipython-input-58-2fe0989150cf> in <module>()
13 df['date'] = pd.to_datetime(df['date'])
14 df = df.set_index(['date','symbol'])
---> 15 df.groupby('symbol')['status'].ffill()
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in pad(self, limit)
1362 DataFrame.fillna
1363 """
-> 1364 return self.apply(lambda x: x.ffill(limit=limit))
1365 ffill = pad
1366
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs)
3110 examples=_apply_docs['series_examples']))
3111 def apply(self, func, *args, **kwargs):
-> 3112 return super(SeriesGroupBy, self).apply(func, *args, **kwargs)
3113
3114 @Appender(_agg_doc)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs)
803 # ignore SettingWithCopy here in case the user mutates
804 with option_context('mode.chained_assignment', None):
--> 805 return self._python_apply_general(f)
806
807 def _python_apply_general(self, f):
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _python_apply_general(self, f)
812 keys,
813 values,
--> 814 not_indexed_same=mutated or self.mutated)
815
816 def _iterate_slices(self):
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _wrap_applied_output(self, keys, values, not_indexed_same)
3250 if isinstance(values[0], (Series, dict)):
3251 return self._concat_objects(keys, values,
-> 3252 not_indexed_same=not_indexed_same)
3253 elif isinstance(values[0], DataFrame):
3254 # possible that Series -> DataFrame by applied function
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _concat_objects(self, keys, values, not_indexed_same)
978
979 if isinstance(result, Series):
--> 980 result = result.reindex(ax)
981 else:
982
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/series.pyc in reindex(self, index, **kwargs)
2679 @Appender(generic._shared_docs['reindex'] % _shared_doc_kwargs)
2680 def reindex(self, index=None, **kwargs):
-> 2681 return super(Series, self).reindex(index=index, **kwargs)
2682
2683 @Appender(generic._shared_docs['fillna'] % _shared_doc_kwargs)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in reindex(self, *args, **kwargs)
3021 # perform the reindex on the axes
3022 return self._reindex_axes(axes, level, limit, tolerance, method,
-> 3023 fill_value, copy).__finalize__(self)
3024
3025 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value,
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy)
3034 ax = self._get_axis(a)
3035 new_index, indexer = ax.reindex(labels, level=level, limit=limit,
-> 3036 tolerance=tolerance, method=method)
3037
3038 axis = self._get_axis_number(a)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/indexes/multi.pyc in reindex(self, target, method, level, limit, tolerance)
1901 tolerance=tolerance)
1902 else:
-> 1903 raise Exception("cannot handle a non-unique multi-index!")
1904
1905 if not isinstance(target, MultiIndex):
Exception: cannot handle a non-unique multi-index!
```
</details>
#### Problem description
Performing `ffill()` does not forward fill values as intended. The index has duplicate values, the expectation is that the values will be filled according to the order the rows are in already.
Duplicate index values are possible when using high frequency time intervals.
#### Expected Output
```python
df = pd.DataFrame([['2018-01-01','MSFT','short'],
['2018-01-01','MSFT','long'],
['2018-01-01','MSFT','long'],
['2018-01-01','AAPL','short'],
['2018-01-01','AAPL','short'],
['2018-01-02','AAPL','short'],
['2018-01-01','TSLA','neutral'],
['2018-01-02','TSLA','neutral']],
columns=['date','symbol','status'])
df['date'] = pd.to_datetime(df['date'])
```
#### Output of ``pd.show_versions()``
<details>
INSTALLED VERSIONS
------------------
commit: None
python: 2.7.14.final.0
python-bits: 64
OS: Linux
OS-release: 4.13.0-31-generic
machine: x86_64
processor: x86_64
byteorder: little
LC_ALL: None
LANG: en_AU.UTF-8
LOCALE: None.None
pandas: 0.22.0
pytest: 3.2.1
pip: 9.0.1
setuptools: 36.5.0.post20170921
Cython: 0.26.1
numpy: 1.14.0
scipy: 0.19.1
pyarrow: None
xarray: None
IPython: 5.4.1
sphinx: 1.6.3
patsy: 0.4.1
dateutil: 2.6.1
pytz: 2017.3
blosc: None
bottleneck: 1.2.1
tables: 3.4.2
numexpr: 2.6.2
feather: None
matplotlib: 2.1.0
openpyxl: 2.4.8
xlrd: 1.1.0
xlwt: 1.3.0
xlsxwriter: 1.0.2
lxml: 4.1.0
bs4: 4.6.0
html5lib: 0.999999999
sqlalchemy: 1.1.13
pymysql: None
psycopg2: 2.7.3.2 (dt dec pq3 ext lo64)
jinja2: 2.9.6
s3fs: None
fastparquet: None
pandas_gbq: None
pandas_datareader: None
</details>
This also happens with `axis=1`.
```python
import pandas as pd
import numpy as np
df = pd.DataFrame(
np.ones([6, 4], dtype=int),
columns=pd.MultiIndex.from_product([['A', 'B'], [1, 2]])
)
(
df
.groupby(level=0, axis=1)
.apply(
lambda df: 2*df.xs(df.name, axis=1)
)
)
```
<details>
```pytb
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-12-e7c022399f89> in <module>()
3 .groupby(level=0, axis=1)
4 .apply(
----> 5 lambda df: 2*df.xs(df.name, axis=1)
6 )
7 )
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in apply(self, func, *args, **kwargs)
882 # ignore SettingWithCopy here in case the user mutates
883 with option_context('mode.chained_assignment', None):
--> 884 return self._python_apply_general(f)
885
886 def _python_apply_general(self, f):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _python_apply_general(self, f)
891 keys,
892 values,
--> 893 not_indexed_same=mutated or self.mutated)
894
895 def _iterate_slices(self):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _wrap_applied_output(self, keys, values, not_indexed_same)
3920 elif isinstance(v, DataFrame):
3921 return self._concat_objects(keys, values,
-> 3922 not_indexed_same=not_indexed_same)
3923 elif self.grouper.groupings is not None:
3924 if len(self.grouper.groupings) > 1:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _concat_objects(self, keys, values, not_indexed_same)
1085 result = result.take(indexer, axis=self.axis)
1086 else:
-> 1087 result = result.reindex(ax, axis=self.axis)
1088
1089 elif self.group_keys:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/util/_decorators.py in wrapper(*args, **kwargs)
145 @wraps(func)
146 def wrapper(*args, **kwargs):
--> 147 return func(*args, **kwargs)
148
149 if not PY2:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in reindex(self, *args, **kwargs)
2979 kwargs.pop('axis', None)
2980 kwargs.pop('labels', None)
-> 2981 return super(DataFrame, self).reindex(**kwargs)
2982
2983 @Appender(_shared_docs['reindex_axis'] % _shared_doc_kwargs)
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in reindex(self, *args, **kwargs)
3356 # perform the reindex on the axes
3357 return self._reindex_axes(axes, level, limit, tolerance, method,
-> 3358 fill_value, copy).__finalize__(self)
3359
3360 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value,
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy)
2909 if columns is not None:
2910 frame = frame._reindex_columns(columns, method, copy, level,
-> 2911 fill_value, limit, tolerance)
2912
2913 index = axes['index']
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_columns(self, new_columns, method, copy, level, fill_value, limit, tolerance)
2934 return self._reindex_with_indexers({1: [new_columns, indexer]},
2935 copy=copy, fill_value=fill_value,
-> 2936 allow_dups=False)
2937
2938 def _reindex_multi(self, axes, copy, fill_value):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in _reindex_with_indexers(self, reindexers, fill_value, copy, allow_dups)
3477 fill_value=fill_value,
3478 allow_dups=allow_dups,
-> 3479 copy=copy)
3480
3481 if copy and new_data is self._data:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/internals.py in reindex_indexer(self, new_axis, indexer, axis, fill_value, allow_dups, copy)
4125 # some axes don't allow reindexing with dups
4126 if not allow_dups:
-> 4127 self.axes[axis]._can_reindex(indexer)
4128
4129 if axis >= self.ndim:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/indexes/base.py in _can_reindex(self, indexer)
2940 # trying to reindex on an axis with duplicates
2941 if not self.is_unique and len(indexer):
-> 2942 raise ValueError("cannot reindex from a duplicate axis")
2943
2944 def reindex(self, target, method=None, level=None, limit=None,
ValueError: cannot reindex from a duplicate axis
```
</details> | 1.0 | df.groupby('symbol')['status'].ffill() results in Exception: cannot handle a non-unique multi-index! - #### Code Sample, a copy-pastable example if possible
```python
import pandas as pd
df = pd.DataFrame([['2018-01-01','MSFT','short'],
['2018-01-01','MSFT','long'],
['2018-01-01','MSFT',pd.np.NaN],
['2018-01-01','AAPL','short'],
['2018-01-01','AAPL',pd.np.NaN],
['2018-01-02','AAPL','short'],
['2018-01-01','TSLA','neutral'],
['2018-01-02','TSLA',pd.np.NaN]],
columns=['date','symbol','status'])
df['date'] = pd.to_datetime(df['date'])
df = df.set_index(['date','symbol'])
df.groupby('symbol')['status'].ffill()
```
#### Raised exception
<details>
```
---------------------------------------------------------------------------
Exception Traceback (most recent call last)
<ipython-input-58-2fe0989150cf> in <module>()
13 df['date'] = pd.to_datetime(df['date'])
14 df = df.set_index(['date','symbol'])
---> 15 df.groupby('symbol')['status'].ffill()
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in pad(self, limit)
1362 DataFrame.fillna
1363 """
-> 1364 return self.apply(lambda x: x.ffill(limit=limit))
1365 ffill = pad
1366
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs)
3110 examples=_apply_docs['series_examples']))
3111 def apply(self, func, *args, **kwargs):
-> 3112 return super(SeriesGroupBy, self).apply(func, *args, **kwargs)
3113
3114 @Appender(_agg_doc)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in apply(self, func, *args, **kwargs)
803 # ignore SettingWithCopy here in case the user mutates
804 with option_context('mode.chained_assignment', None):
--> 805 return self._python_apply_general(f)
806
807 def _python_apply_general(self, f):
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _python_apply_general(self, f)
812 keys,
813 values,
--> 814 not_indexed_same=mutated or self.mutated)
815
816 def _iterate_slices(self):
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _wrap_applied_output(self, keys, values, not_indexed_same)
3250 if isinstance(values[0], (Series, dict)):
3251 return self._concat_objects(keys, values,
-> 3252 not_indexed_same=not_indexed_same)
3253 elif isinstance(values[0], DataFrame):
3254 # possible that Series -> DataFrame by applied function
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/groupby.pyc in _concat_objects(self, keys, values, not_indexed_same)
978
979 if isinstance(result, Series):
--> 980 result = result.reindex(ax)
981 else:
982
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/series.pyc in reindex(self, index, **kwargs)
2679 @Appender(generic._shared_docs['reindex'] % _shared_doc_kwargs)
2680 def reindex(self, index=None, **kwargs):
-> 2681 return super(Series, self).reindex(index=index, **kwargs)
2682
2683 @Appender(generic._shared_docs['fillna'] % _shared_doc_kwargs)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in reindex(self, *args, **kwargs)
3021 # perform the reindex on the axes
3022 return self._reindex_axes(axes, level, limit, tolerance, method,
-> 3023 fill_value, copy).__finalize__(self)
3024
3025 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value,
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/generic.pyc in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy)
3034 ax = self._get_axis(a)
3035 new_index, indexer = ax.reindex(labels, level=level, limit=limit,
-> 3036 tolerance=tolerance, method=method)
3037
3038 axis = self._get_axis_number(a)
/home/greg/anaconda2/lib/python2.7/site-packages/pandas/core/indexes/multi.pyc in reindex(self, target, method, level, limit, tolerance)
1901 tolerance=tolerance)
1902 else:
-> 1903 raise Exception("cannot handle a non-unique multi-index!")
1904
1905 if not isinstance(target, MultiIndex):
Exception: cannot handle a non-unique multi-index!
```
</details>
#### Problem description
Performing `ffill()` does not forward fill values as intended. The index has duplicate values, the expectation is that the values will be filled according to the order the rows are in already.
Duplicate index values are possible when using high frequency time intervals.
#### Expected Output
```python
df = pd.DataFrame([['2018-01-01','MSFT','short'],
['2018-01-01','MSFT','long'],
['2018-01-01','MSFT','long'],
['2018-01-01','AAPL','short'],
['2018-01-01','AAPL','short'],
['2018-01-02','AAPL','short'],
['2018-01-01','TSLA','neutral'],
['2018-01-02','TSLA','neutral']],
columns=['date','symbol','status'])
df['date'] = pd.to_datetime(df['date'])
```
#### Output of ``pd.show_versions()``
<details>
INSTALLED VERSIONS
------------------
commit: None
python: 2.7.14.final.0
python-bits: 64
OS: Linux
OS-release: 4.13.0-31-generic
machine: x86_64
processor: x86_64
byteorder: little
LC_ALL: None
LANG: en_AU.UTF-8
LOCALE: None.None
pandas: 0.22.0
pytest: 3.2.1
pip: 9.0.1
setuptools: 36.5.0.post20170921
Cython: 0.26.1
numpy: 1.14.0
scipy: 0.19.1
pyarrow: None
xarray: None
IPython: 5.4.1
sphinx: 1.6.3
patsy: 0.4.1
dateutil: 2.6.1
pytz: 2017.3
blosc: None
bottleneck: 1.2.1
tables: 3.4.2
numexpr: 2.6.2
feather: None
matplotlib: 2.1.0
openpyxl: 2.4.8
xlrd: 1.1.0
xlwt: 1.3.0
xlsxwriter: 1.0.2
lxml: 4.1.0
bs4: 4.6.0
html5lib: 0.999999999
sqlalchemy: 1.1.13
pymysql: None
psycopg2: 2.7.3.2 (dt dec pq3 ext lo64)
jinja2: 2.9.6
s3fs: None
fastparquet: None
pandas_gbq: None
pandas_datareader: None
</details>
This also happens with `axis=1`.
```python
import pandas as pd
import numpy as np
df = pd.DataFrame(
np.ones([6, 4], dtype=int),
columns=pd.MultiIndex.from_product([['A', 'B'], [1, 2]])
)
(
df
.groupby(level=0, axis=1)
.apply(
lambda df: 2*df.xs(df.name, axis=1)
)
)
```
<details>
```pytb
---------------------------------------------------------------------------
ValueError Traceback (most recent call last)
<ipython-input-12-e7c022399f89> in <module>()
3 .groupby(level=0, axis=1)
4 .apply(
----> 5 lambda df: 2*df.xs(df.name, axis=1)
6 )
7 )
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in apply(self, func, *args, **kwargs)
882 # ignore SettingWithCopy here in case the user mutates
883 with option_context('mode.chained_assignment', None):
--> 884 return self._python_apply_general(f)
885
886 def _python_apply_general(self, f):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _python_apply_general(self, f)
891 keys,
892 values,
--> 893 not_indexed_same=mutated or self.mutated)
894
895 def _iterate_slices(self):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _wrap_applied_output(self, keys, values, not_indexed_same)
3920 elif isinstance(v, DataFrame):
3921 return self._concat_objects(keys, values,
-> 3922 not_indexed_same=not_indexed_same)
3923 elif self.grouper.groupings is not None:
3924 if len(self.grouper.groupings) > 1:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/groupby.py in _concat_objects(self, keys, values, not_indexed_same)
1085 result = result.take(indexer, axis=self.axis)
1086 else:
-> 1087 result = result.reindex(ax, axis=self.axis)
1088
1089 elif self.group_keys:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/util/_decorators.py in wrapper(*args, **kwargs)
145 @wraps(func)
146 def wrapper(*args, **kwargs):
--> 147 return func(*args, **kwargs)
148
149 if not PY2:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in reindex(self, *args, **kwargs)
2979 kwargs.pop('axis', None)
2980 kwargs.pop('labels', None)
-> 2981 return super(DataFrame, self).reindex(**kwargs)
2982
2983 @Appender(_shared_docs['reindex_axis'] % _shared_doc_kwargs)
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in reindex(self, *args, **kwargs)
3356 # perform the reindex on the axes
3357 return self._reindex_axes(axes, level, limit, tolerance, method,
-> 3358 fill_value, copy).__finalize__(self)
3359
3360 def _reindex_axes(self, axes, level, limit, tolerance, method, fill_value,
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_axes(self, axes, level, limit, tolerance, method, fill_value, copy)
2909 if columns is not None:
2910 frame = frame._reindex_columns(columns, method, copy, level,
-> 2911 fill_value, limit, tolerance)
2912
2913 index = axes['index']
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/frame.py in _reindex_columns(self, new_columns, method, copy, level, fill_value, limit, tolerance)
2934 return self._reindex_with_indexers({1: [new_columns, indexer]},
2935 copy=copy, fill_value=fill_value,
-> 2936 allow_dups=False)
2937
2938 def _reindex_multi(self, axes, copy, fill_value):
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/generic.py in _reindex_with_indexers(self, reindexers, fill_value, copy, allow_dups)
3477 fill_value=fill_value,
3478 allow_dups=allow_dups,
-> 3479 copy=copy)
3480
3481 if copy and new_data is self._data:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/internals.py in reindex_indexer(self, new_axis, indexer, axis, fill_value, allow_dups, copy)
4125 # some axes don't allow reindexing with dups
4126 if not allow_dups:
-> 4127 self.axes[axis]._can_reindex(indexer)
4128
4129 if axis >= self.ndim:
~/Envs/pandas-dev/lib/python3.6/site-packages/pandas/pandas/core/indexes/base.py in _can_reindex(self, indexer)
2940 # trying to reindex on an axis with duplicates
2941 if not self.is_unique and len(indexer):
-> 2942 raise ValueError("cannot reindex from a duplicate axis")
2943
2944 def reindex(self, target, method=None, level=None, limit=None,
ValueError: cannot reindex from a duplicate axis
```
</details> | non_priority | df groupby symbol ffill results in exception cannot handle a non unique multi index code sample a copy pastable example if possible python import pandas as pd df pd dataframe columns df pd to datetime df df df set index df groupby symbol ffill raised exception exception traceback most recent call last in df pd to datetime df df df set index df groupby symbol ffill home greg lib site packages pandas core groupby pyc in pad self limit dataframe fillna return self apply lambda x x ffill limit limit ffill pad home greg lib site packages pandas core groupby pyc in apply self func args kwargs examples apply docs def apply self func args kwargs return super seriesgroupby self apply func args kwargs appender agg doc home greg lib site packages pandas core groupby pyc in apply self func args kwargs ignore settingwithcopy here in case the user mutates with option context mode chained assignment none return self python apply general f def python apply general self f home greg lib site packages pandas core groupby pyc in python apply general self f keys values not indexed same mutated or self mutated def iterate slices self home greg lib site packages pandas core groupby pyc in wrap applied output self keys values not indexed same if isinstance values series dict return self concat objects keys values not indexed same not indexed same elif isinstance values dataframe possible that series dataframe by applied function home greg lib site packages pandas core groupby pyc in concat objects self keys values not indexed same if isinstance result series result result reindex ax else home greg lib site packages pandas core series pyc in reindex self index kwargs appender generic shared docs shared doc kwargs def reindex self index none kwargs return super series self reindex index index kwargs appender generic shared docs shared doc kwargs home greg lib site packages pandas core generic pyc in reindex self args kwargs perform the reindex on the axes return self reindex axes axes level limit tolerance method fill value copy finalize self def reindex axes self axes level limit tolerance method fill value home greg lib site packages pandas core generic pyc in reindex axes self axes level limit tolerance method fill value copy ax self get axis a new index indexer ax reindex labels level level limit limit tolerance tolerance method method axis self get axis number a home greg lib site packages pandas core indexes multi pyc in reindex self target method level limit tolerance tolerance tolerance else raise exception cannot handle a non unique multi index if not isinstance target multiindex exception cannot handle a non unique multi index problem description performing ffill does not forward fill values as intended the index has duplicate values the expectation is that the values will be filled according to the order the rows are in already duplicate index values are possible when using high frequency time intervals expected output python df pd dataframe columns df pd to datetime df output of pd show versions installed versions commit none python final python bits os linux os release generic machine processor byteorder little lc all none lang en au utf locale none none pandas pytest pip setuptools cython numpy scipy pyarrow none xarray none ipython sphinx patsy dateutil pytz blosc none bottleneck tables numexpr feather none matplotlib openpyxl xlrd xlwt xlsxwriter lxml sqlalchemy pymysql none dt dec ext none fastparquet none pandas gbq none pandas datareader none this also happens with axis python import pandas as pd import numpy as np df pd dataframe np ones dtype int columns pd multiindex from product df groupby level axis apply lambda df df xs df name axis pytb valueerror traceback most recent call last in groupby level axis apply lambda df df xs df name axis envs pandas dev lib site packages pandas pandas core groupby py in apply self func args kwargs ignore settingwithcopy here in case the user mutates with option context mode chained assignment none return self python apply general f def python apply general self f envs pandas dev lib site packages pandas pandas core groupby py in python apply general self f keys values not indexed same mutated or self mutated def iterate slices self envs pandas dev lib site packages pandas pandas core groupby py in wrap applied output self keys values not indexed same elif isinstance v dataframe return self concat objects keys values not indexed same not indexed same elif self grouper groupings is not none if len self grouper groupings envs pandas dev lib site packages pandas pandas core groupby py in concat objects self keys values not indexed same result result take indexer axis self axis else result result reindex ax axis self axis elif self group keys envs pandas dev lib site packages pandas pandas util decorators py in wrapper args kwargs wraps func def wrapper args kwargs return func args kwargs if not envs pandas dev lib site packages pandas pandas core frame py in reindex self args kwargs kwargs pop axis none kwargs pop labels none return super dataframe self reindex kwargs appender shared docs shared doc kwargs envs pandas dev lib site packages pandas pandas core generic py in reindex self args kwargs perform the reindex on the axes return self reindex axes axes level limit tolerance method fill value copy finalize self def reindex axes self axes level limit tolerance method fill value envs pandas dev lib site packages pandas pandas core frame py in reindex axes self axes level limit tolerance method fill value copy if columns is not none frame frame reindex columns columns method copy level fill value limit tolerance index axes envs pandas dev lib site packages pandas pandas core frame py in reindex columns self new columns method copy level fill value limit tolerance return self reindex with indexers copy copy fill value fill value allow dups false def reindex multi self axes copy fill value envs pandas dev lib site packages pandas pandas core generic py in reindex with indexers self reindexers fill value copy allow dups fill value fill value allow dups allow dups copy copy if copy and new data is self data envs pandas dev lib site packages pandas pandas core internals py in reindex indexer self new axis indexer axis fill value allow dups copy some axes don t allow reindexing with dups if not allow dups self axes can reindex indexer if axis self ndim envs pandas dev lib site packages pandas pandas core indexes base py in can reindex self indexer trying to reindex on an axis with duplicates if not self is unique and len indexer raise valueerror cannot reindex from a duplicate axis def reindex self target method none level none limit none valueerror cannot reindex from a duplicate axis | 0 |
39,550 | 10,352,152,114 | IssuesEvent | 2019-09-05 08:38:50 | apollographql/apollo-ios | https://api.github.com/repos/apollographql/apollo-ios | closed | Unable to Compile Apollo version 0.15.0 with Xcode 10.2 no API.swift file is created. | build-issue | Using Apollo version 0.15.0
Npm version 6.10.3
Node version v12.9.1
**Not able to compile the project with Legacy Build System and New Build System**
> > › Warning: apollo update available from 1.9.2 to 2.18.0.
> ++ npx --no-install apollo codegen:generate --target=swift '--includes=./**/*.graphql' --localSchemaFile=schema.json API.swift
> › Warning: apollo update available from 1.9.2 to 2.18.0.
> › Error: Unexpected arguments: --localSchemaFile=schema.json, API.swift
> › See more help with --help
> Command /bin/sh failed with exit code 2 | 1.0 | Unable to Compile Apollo version 0.15.0 with Xcode 10.2 no API.swift file is created. - Using Apollo version 0.15.0
Npm version 6.10.3
Node version v12.9.1
**Not able to compile the project with Legacy Build System and New Build System**
> > › Warning: apollo update available from 1.9.2 to 2.18.0.
> ++ npx --no-install apollo codegen:generate --target=swift '--includes=./**/*.graphql' --localSchemaFile=schema.json API.swift
> › Warning: apollo update available from 1.9.2 to 2.18.0.
> › Error: Unexpected arguments: --localSchemaFile=schema.json, API.swift
> › See more help with --help
> Command /bin/sh failed with exit code 2 | non_priority | unable to compile apollo version with xcode no api swift file is created using apollo version npm version node version not able to compile the project with legacy build system and new build system › warning apollo update available from to npx no install apollo codegen generate target swift includes graphql localschemafile schema json api swift › warning apollo update available from to › error unexpected arguments localschemafile schema json api swift › see more help with help command bin sh failed with exit code | 0 |
20,028 | 14,937,745,091 | IssuesEvent | 2021-01-25 14:59:25 | andrewfstratton/quando | https://api.github.com/repos/andrewfstratton/quando | opened | Move delete away from load | usability | Easy to hit delete on narrow screen by accident - especially when viewing All, which is likely to delete the latest version. | True | Move delete away from load - Easy to hit delete on narrow screen by accident - especially when viewing All, which is likely to delete the latest version. | non_priority | move delete away from load easy to hit delete on narrow screen by accident especially when viewing all which is likely to delete the latest version | 0 |
231,407 | 18,765,504,756 | IssuesEvent | 2021-11-05 23:04:26 | numpy/numpy | https://api.github.com/repos/numpy/numpy | closed | 4 test cases fail in test running with the latest master code base | 05 - Testing | <!-- Please describe the issue in detail here, and fill in the fields below -->
### Reproducing code example:
<!-- A short code example that reproduces the problem/missing feature. It should be
self-contained, i.e., possible to run as-is via 'python myproblem.py' -->
```python
python -c 'import numpy as np; np.test()'
```
<!-- Remove these sections for a feature request -->
### Error message:
<!-- If you are reporting a segfault please include a GDB traceback, which you
can generate by following
https://github.com/numpy/numpy/blob/master/doc/source/dev/development_environment.rst#debugging -->
<!-- Full error message, if any (starting from line Traceback: ...) -->
```
======================================================= FAILURES =======================================================__________________________________________ TestLinspace.test_denormal_numbers __________________________________________
self = <numpy.core.tests.test_function_base.TestLinspace object at 0x7f70ac7be2e8>
def test_denormal_numbers(self):
# Regression test for gh-5437. Will probably fail when compiled
# with ICC, which flushes denormals to zero
for ftype in sctypes['float']:
stop = nextafter(ftype(0), ftype(1)) * 5 # A denormal number
> assert_(any(linspace(0, stop, 10, endpoint=False, dtype=ftype)))
E AssertionError
ftype = <class 'numpy.float128'>
self = <numpy.core.tests.test_function_base.TestLinspace object at 0x7f70ac7be2e8>
stop = 0.0
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_function_base.py:308: AssertionError
_____________________________________ TestFloatExceptions.test_floating_exceptions _____________________________________
self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>
def test_floating_exceptions(self):
# Test basic arithmetic function errors
with np.errstate(all='raise'):
# Test for all real and complex float types
for typecode in np.typecodes['AllFloat']:
ftype = np.obj2sctype(typecode)
if np.dtype(ftype).kind == 'f':
# Get some extreme values for the type
fi = np.finfo(ftype)
ft_tiny = fi.tiny
ft_max = fi.max
ft_eps = fi.eps
underflow = 'underflow'
divbyzero = 'divide by zero'
else:
# 'c', complex, corresponding real dtype
rtype = type(ftype(0).real)
fi = np.finfo(rtype)
ft_tiny = ftype(fi.tiny)
ft_max = ftype(fi.max)
ft_eps = ftype(fi.eps)
# The complex types raise different exceptions
underflow = ''
divbyzero = ''
overflow = 'overflow'
invalid = 'invalid'
self.assert_raises_fpe(underflow,
> lambda a, b: a/b, ft_tiny, ft_max)
divbyzero = 'divide by zero'
fi = finfo(resolution=1.0000000000000002641e-18, min=-inf, max=inf, dtype=float128)
ft_eps = 1.084202172485504434e-19
ft_max = inf
ft_tiny = 3.3621031431120935063e-4932
ftype = <class 'numpy.float128'>
invalid = 'invalid'
overflow = 'overflow'
self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>
typecode = 'g'
underflow = 'underflow'
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_numeric.py:580:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>, fpeerr = 'underflow'
flop = <function TestFloatExceptions.test_floating_exceptions.<locals>.<lambda> at 0x7f70aca5a730>
x = 3.3621031431120935063e-4932, y = inf
def assert_raises_fpe(self, fpeerr, flop, x, y):
ftype = type(x)
try:
flop(x, y)
assert_(False,
> "Type %s did not raise fpe error '%s'." % (ftype, fpeerr))
E AssertionError: Type <class 'numpy.float128'> did not raise fpe error 'underflow'.
flop = <function TestFloatExceptions.test_floating_exceptions.<locals>.<lambda> at 0x7f70aca5a730>
fpeerr = 'underflow'
ftype = <class 'numpy.float128'>
self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>
x = 3.3621031431120935063e-4932
y = inf
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_numeric.py:535: AssertionError
_______________________________ TestComplexFunctions.test_loss_of_precision_longcomplex ________________________________
self = <numpy.core.tests.test_umath.TestComplexFunctions object at 0x7f70ac0965c0>
@pytest.mark.skipif(is_longdouble_finfo_bogus(),
reason="Bogus long double finfo")
def test_loss_of_precision_longcomplex(self):
> self.check_loss_of_precision(np.longcomplex)
self = <numpy.core.tests.test_umath.TestComplexFunctions object at 0x7f70ac0965c0>
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2642:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2592: in check_loss_of_precision
check(x_series, 50*eps)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
x = array([1.00000000e-20, 1.21736864e-20, 1.48198641e-20, 1.80412378e-20,
2.19628372e-20, 2.67368693e-20, 3.254862...5156e-04, 4.54267386e-04,
5.53010871e-04, 6.73218092e-04, 8.19554595e-04, 9.97700064e-04],
dtype=float128)
rtol = 5.42101086242752217e-18
def check(x, rtol):
x = x.astype(real_dtype)
z = x.astype(dtype)
d = np.absolute(np.arcsinh(x)/np.arcsinh(z).real - 1)
assert_(np.all(d < rtol), (np.argmax(d), x[np.argmax(d)], d.max(),
> 'arcsinh'))
E AssertionError: (190, 0.00016990298344866943845, 4.4408920985006261617e-16, 'arcsinh')
d = array([0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.00000000e+00,
0.00000000e+00, 0.00000000e+00, 0.000000...2302e-16, 0.00000000e+00,
3.33066907e-16, 0.00000000e+00, 2.22044605e-16, 1.11022302e-16],
dtype=float128)
dtype = <class 'numpy.complex256'>
real_dtype = dtype('float128')
rtol = 5.42101086242752217e-18
x = array([1.00000000e-20, 1.21736864e-20, 1.48198641e-20, 1.80412378e-20,
2.19628372e-20, 2.67368693e-20, 3.254862...5156e-04, 4.54267386e-04,
5.53010871e-04, 6.73218092e-04, 8.19554595e-04, 9.97700064e-04],
dtype=float128)
z = array([1.00000000e-20+0.j, 1.21736864e-20+0.j, 1.48198641e-20+0.j,
1.80412378e-20+0.j, 2.19628372e-20+0.j, 2.67...7386e-04+0.j, 5.53010871e-04+0.j, 6.73218092e-04+0.j,
8.19554595e-04+0.j, 9.97700064e-04+0.j], dtype=complex256)
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2565: AssertionError
___________________________________________________ test_nextafter_0 ___________________________________________________
def test_nextafter_0():
for t, direction in itertools.product(np.sctypes['float'], (1, -1)):
tiny = np.finfo(t).tiny
> assert_(0. < direction * np.nextafter(t(0), t(direction)) < tiny)
E AssertionError
direction = 1
t = <class 'numpy.float128'>
tiny = 3.3621031431120935063e-4932
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2776: AssertionError
4 failed, 5040 passed, 409 skipped, 167 deselected, 12 xfailed, 1 xpassed in 137.51 seconds
### Numpy/Python version information:
<!-- Output from 'import sys, numpy; print(numpy.__version__, sys.version)' -->
1.16.0.dev0+db5f9d3 3.6.6
``` | 1.0 | 4 test cases fail in test running with the latest master code base - <!-- Please describe the issue in detail here, and fill in the fields below -->
### Reproducing code example:
<!-- A short code example that reproduces the problem/missing feature. It should be
self-contained, i.e., possible to run as-is via 'python myproblem.py' -->
```python
python -c 'import numpy as np; np.test()'
```
<!-- Remove these sections for a feature request -->
### Error message:
<!-- If you are reporting a segfault please include a GDB traceback, which you
can generate by following
https://github.com/numpy/numpy/blob/master/doc/source/dev/development_environment.rst#debugging -->
<!-- Full error message, if any (starting from line Traceback: ...) -->
```
======================================================= FAILURES =======================================================__________________________________________ TestLinspace.test_denormal_numbers __________________________________________
self = <numpy.core.tests.test_function_base.TestLinspace object at 0x7f70ac7be2e8>
def test_denormal_numbers(self):
# Regression test for gh-5437. Will probably fail when compiled
# with ICC, which flushes denormals to zero
for ftype in sctypes['float']:
stop = nextafter(ftype(0), ftype(1)) * 5 # A denormal number
> assert_(any(linspace(0, stop, 10, endpoint=False, dtype=ftype)))
E AssertionError
ftype = <class 'numpy.float128'>
self = <numpy.core.tests.test_function_base.TestLinspace object at 0x7f70ac7be2e8>
stop = 0.0
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_function_base.py:308: AssertionError
_____________________________________ TestFloatExceptions.test_floating_exceptions _____________________________________
self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>
def test_floating_exceptions(self):
# Test basic arithmetic function errors
with np.errstate(all='raise'):
# Test for all real and complex float types
for typecode in np.typecodes['AllFloat']:
ftype = np.obj2sctype(typecode)
if np.dtype(ftype).kind == 'f':
# Get some extreme values for the type
fi = np.finfo(ftype)
ft_tiny = fi.tiny
ft_max = fi.max
ft_eps = fi.eps
underflow = 'underflow'
divbyzero = 'divide by zero'
else:
# 'c', complex, corresponding real dtype
rtype = type(ftype(0).real)
fi = np.finfo(rtype)
ft_tiny = ftype(fi.tiny)
ft_max = ftype(fi.max)
ft_eps = ftype(fi.eps)
# The complex types raise different exceptions
underflow = ''
divbyzero = ''
overflow = 'overflow'
invalid = 'invalid'
self.assert_raises_fpe(underflow,
> lambda a, b: a/b, ft_tiny, ft_max)
divbyzero = 'divide by zero'
fi = finfo(resolution=1.0000000000000002641e-18, min=-inf, max=inf, dtype=float128)
ft_eps = 1.084202172485504434e-19
ft_max = inf
ft_tiny = 3.3621031431120935063e-4932
ftype = <class 'numpy.float128'>
invalid = 'invalid'
overflow = 'overflow'
self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>
typecode = 'g'
underflow = 'underflow'
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_numeric.py:580:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>, fpeerr = 'underflow'
flop = <function TestFloatExceptions.test_floating_exceptions.<locals>.<lambda> at 0x7f70aca5a730>
x = 3.3621031431120935063e-4932, y = inf
def assert_raises_fpe(self, fpeerr, flop, x, y):
ftype = type(x)
try:
flop(x, y)
assert_(False,
> "Type %s did not raise fpe error '%s'." % (ftype, fpeerr))
E AssertionError: Type <class 'numpy.float128'> did not raise fpe error 'underflow'.
flop = <function TestFloatExceptions.test_floating_exceptions.<locals>.<lambda> at 0x7f70aca5a730>
fpeerr = 'underflow'
ftype = <class 'numpy.float128'>
self = <numpy.core.tests.test_numeric.TestFloatExceptions object at 0x7f70acbb4550>
x = 3.3621031431120935063e-4932
y = inf
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_numeric.py:535: AssertionError
_______________________________ TestComplexFunctions.test_loss_of_precision_longcomplex ________________________________
self = <numpy.core.tests.test_umath.TestComplexFunctions object at 0x7f70ac0965c0>
@pytest.mark.skipif(is_longdouble_finfo_bogus(),
reason="Bogus long double finfo")
def test_loss_of_precision_longcomplex(self):
> self.check_loss_of_precision(np.longcomplex)
self = <numpy.core.tests.test_umath.TestComplexFunctions object at 0x7f70ac0965c0>
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2642:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2592: in check_loss_of_precision
check(x_series, 50*eps)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
x = array([1.00000000e-20, 1.21736864e-20, 1.48198641e-20, 1.80412378e-20,
2.19628372e-20, 2.67368693e-20, 3.254862...5156e-04, 4.54267386e-04,
5.53010871e-04, 6.73218092e-04, 8.19554595e-04, 9.97700064e-04],
dtype=float128)
rtol = 5.42101086242752217e-18
def check(x, rtol):
x = x.astype(real_dtype)
z = x.astype(dtype)
d = np.absolute(np.arcsinh(x)/np.arcsinh(z).real - 1)
assert_(np.all(d < rtol), (np.argmax(d), x[np.argmax(d)], d.max(),
> 'arcsinh'))
E AssertionError: (190, 0.00016990298344866943845, 4.4408920985006261617e-16, 'arcsinh')
d = array([0.00000000e+00, 0.00000000e+00, 0.00000000e+00, 0.00000000e+00,
0.00000000e+00, 0.00000000e+00, 0.000000...2302e-16, 0.00000000e+00,
3.33066907e-16, 0.00000000e+00, 2.22044605e-16, 1.11022302e-16],
dtype=float128)
dtype = <class 'numpy.complex256'>
real_dtype = dtype('float128')
rtol = 5.42101086242752217e-18
x = array([1.00000000e-20, 1.21736864e-20, 1.48198641e-20, 1.80412378e-20,
2.19628372e-20, 2.67368693e-20, 3.254862...5156e-04, 4.54267386e-04,
5.53010871e-04, 6.73218092e-04, 8.19554595e-04, 9.97700064e-04],
dtype=float128)
z = array([1.00000000e-20+0.j, 1.21736864e-20+0.j, 1.48198641e-20+0.j,
1.80412378e-20+0.j, 2.19628372e-20+0.j, 2.67...7386e-04+0.j, 5.53010871e-04+0.j, 6.73218092e-04+0.j,
8.19554595e-04+0.j, 9.97700064e-04+0.j], dtype=complex256)
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2565: AssertionError
___________________________________________________ test_nextafter_0 ___________________________________________________
def test_nextafter_0():
for t, direction in itertools.product(np.sctypes['float'], (1, -1)):
tiny = np.finfo(t).tiny
> assert_(0. < direction * np.nextafter(t(0), t(direction)) < tiny)
E AssertionError
direction = 1
t = <class 'numpy.float128'>
tiny = 3.3621031431120935063e-4932
/usr/local/lib/python3.6/dist-packages/numpy-1.16.0.dev0+db5f9d3-py3.6-linux-x86_64.egg/numpy/core/tests/test_umath.py:2776: AssertionError
4 failed, 5040 passed, 409 skipped, 167 deselected, 12 xfailed, 1 xpassed in 137.51 seconds
### Numpy/Python version information:
<!-- Output from 'import sys, numpy; print(numpy.__version__, sys.version)' -->
1.16.0.dev0+db5f9d3 3.6.6
``` | non_priority | test cases fail in test running with the latest master code base reproducing code example a short code example that reproduces the problem missing feature it should be self contained i e possible to run as is via python myproblem py python python c import numpy as np np test error message if you are reporting a segfault please include a gdb traceback which you can generate by following failures testlinspace test denormal numbers self def test denormal numbers self regression test for gh will probably fail when compiled with icc which flushes denormals to zero for ftype in sctypes stop nextafter ftype ftype a denormal number assert any linspace stop endpoint false dtype ftype e assertionerror ftype self stop usr local lib dist packages numpy linux egg numpy core tests test function base py assertionerror testfloatexceptions test floating exceptions self def test floating exceptions self test basic arithmetic function errors with np errstate all raise test for all real and complex float types for typecode in np typecodes ftype np typecode if np dtype ftype kind f get some extreme values for the type fi np finfo ftype ft tiny fi tiny ft max fi max ft eps fi eps underflow underflow divbyzero divide by zero else c complex corresponding real dtype rtype type ftype real fi np finfo rtype ft tiny ftype fi tiny ft max ftype fi max ft eps ftype fi eps the complex types raise different exceptions underflow divbyzero overflow overflow invalid invalid self assert raises fpe underflow lambda a b a b ft tiny ft max divbyzero divide by zero fi finfo resolution min inf max inf dtype ft eps ft max inf ft tiny ftype invalid invalid overflow overflow self typecode g underflow underflow usr local lib dist packages numpy linux egg numpy core tests test numeric py self fpeerr underflow flop at x y inf def assert raises fpe self fpeerr flop x y ftype type x try flop x y assert false type s did not raise fpe error s ftype fpeerr e assertionerror type did not raise fpe error underflow flop at fpeerr underflow ftype self x y inf usr local lib dist packages numpy linux egg numpy core tests test numeric py assertionerror testcomplexfunctions test loss of precision longcomplex self pytest mark skipif is longdouble finfo bogus reason bogus long double finfo def test loss of precision longcomplex self self check loss of precision np longcomplex self usr local lib dist packages numpy linux egg numpy core tests test umath py usr local lib dist packages numpy linux egg numpy core tests test umath py in check loss of precision check x series eps x array dtype rtol def check x rtol x x astype real dtype z x astype dtype d np absolute np arcsinh x np arcsinh z real assert np all d rtol np argmax d x d max arcsinh e assertionerror arcsinh d array dtype dtype real dtype dtype rtol x array dtype z array j j j j j j j j j j dtype usr local lib dist packages numpy linux egg numpy core tests test umath py assertionerror test nextafter def test nextafter for t direction in itertools product np sctypes tiny np finfo t tiny assert direction np nextafter t t direction tiny e assertionerror direction t tiny usr local lib dist packages numpy linux egg numpy core tests test umath py assertionerror failed passed skipped deselected xfailed xpassed in seconds numpy python version information | 0 |
35,384 | 17,037,982,218 | IssuesEvent | 2021-07-05 09:38:24 | riboviz/riboviz | https://api.github.com/repos/riboviz/riboviz | opened | Parallelize count_reads.py | performance | The countReads processes (riboviz.count_reads.py) is running with only one core now, and is one of the most time consuming processes in the workflow. (For example, this process need 2h54m in Wallace dataset, while the execution time of the whole process is ~5h under 16 cores total, 8 core each sample)
The core function of countReads is implemented in count_reads.py, function count_reads_df, which look like:
```python
def count_reads_df(config_file, input_dir, tmp_dir, output_dir):
df = pd.DataFrame(columns=HEADER)
rows = []
rows.extend(input_fq(config_file, input_dir))
rows.append(cutadapt_fq(tmp_dir))
rows.extend(umi_tools_deplex_fq(tmp_dir))
tmp_samples = [f.name for f in os.scandir(tmp_dir) if f.is_dir()]
tmp_samples.sort()
for sample in tmp_samples:
rows.append(cutadapt_fq(tmp_dir, sample))
rows.append(hisat2_fq(tmp_dir, sample, workflow_files.NON_RRNA_FQ,
"rRNA or other contaminating reads removed by alignment to rRNA index files"))
rows.append(hisat2_sam(tmp_dir, sample, workflow_files.RRNA_MAP_SAM,
"Reads with rRNA and other contaminating reads removed by alignment to rRNA index files"))
rows.append(hisat2_fq(tmp_dir, sample, workflow_files.UNALIGNED_FQ,
"Unaligned reads removed by alignment of remaining reads to ORFs index files"))
rows.append(hisat2_sam(tmp_dir, sample, workflow_files.ORF_MAP_SAM,
"Reads aligned to ORFs index files"))
rows.append(trim_5p_mismatch_sam(tmp_dir, sample))
rows.append(umi_tools_dedup_bam(tmp_dir, output_dir, sample))
rows = [row for row in rows if row is not None]
df = df.append(rows)
return df
```
We can see that the it runs different functions and add the result of each function into `rows`. This can be executed in parallel, by using the `multiprocessing` package, such as Pool and the apply_async to add tasks asynchronously.
I tested it on my VM and it is working, but when I tried to run a full size dataset on eddie, I got an OOM error, which may be caused by #385 . | True | Parallelize count_reads.py - The countReads processes (riboviz.count_reads.py) is running with only one core now, and is one of the most time consuming processes in the workflow. (For example, this process need 2h54m in Wallace dataset, while the execution time of the whole process is ~5h under 16 cores total, 8 core each sample)
The core function of countReads is implemented in count_reads.py, function count_reads_df, which look like:
```python
def count_reads_df(config_file, input_dir, tmp_dir, output_dir):
df = pd.DataFrame(columns=HEADER)
rows = []
rows.extend(input_fq(config_file, input_dir))
rows.append(cutadapt_fq(tmp_dir))
rows.extend(umi_tools_deplex_fq(tmp_dir))
tmp_samples = [f.name for f in os.scandir(tmp_dir) if f.is_dir()]
tmp_samples.sort()
for sample in tmp_samples:
rows.append(cutadapt_fq(tmp_dir, sample))
rows.append(hisat2_fq(tmp_dir, sample, workflow_files.NON_RRNA_FQ,
"rRNA or other contaminating reads removed by alignment to rRNA index files"))
rows.append(hisat2_sam(tmp_dir, sample, workflow_files.RRNA_MAP_SAM,
"Reads with rRNA and other contaminating reads removed by alignment to rRNA index files"))
rows.append(hisat2_fq(tmp_dir, sample, workflow_files.UNALIGNED_FQ,
"Unaligned reads removed by alignment of remaining reads to ORFs index files"))
rows.append(hisat2_sam(tmp_dir, sample, workflow_files.ORF_MAP_SAM,
"Reads aligned to ORFs index files"))
rows.append(trim_5p_mismatch_sam(tmp_dir, sample))
rows.append(umi_tools_dedup_bam(tmp_dir, output_dir, sample))
rows = [row for row in rows if row is not None]
df = df.append(rows)
return df
```
We can see that the it runs different functions and add the result of each function into `rows`. This can be executed in parallel, by using the `multiprocessing` package, such as Pool and the apply_async to add tasks asynchronously.
I tested it on my VM and it is working, but when I tried to run a full size dataset on eddie, I got an OOM error, which may be caused by #385 . | non_priority | parallelize count reads py the countreads processes riboviz count reads py is running with only one core now and is one of the most time consuming processes in the workflow for example this process need in wallace dataset while the execution time of the whole process is under cores total core each sample the core function of countreads is implemented in count reads py function count reads df which look like python def count reads df config file input dir tmp dir output dir df pd dataframe columns header rows rows extend input fq config file input dir rows append cutadapt fq tmp dir rows extend umi tools deplex fq tmp dir tmp samples tmp samples sort for sample in tmp samples rows append cutadapt fq tmp dir sample rows append fq tmp dir sample workflow files non rrna fq rrna or other contaminating reads removed by alignment to rrna index files rows append sam tmp dir sample workflow files rrna map sam reads with rrna and other contaminating reads removed by alignment to rrna index files rows append fq tmp dir sample workflow files unaligned fq unaligned reads removed by alignment of remaining reads to orfs index files rows append sam tmp dir sample workflow files orf map sam reads aligned to orfs index files rows append trim mismatch sam tmp dir sample rows append umi tools dedup bam tmp dir output dir sample rows df df append rows return df we can see that the it runs different functions and add the result of each function into rows this can be executed in parallel by using the multiprocessing package such as pool and the apply async to add tasks asynchronously i tested it on my vm and it is working but when i tried to run a full size dataset on eddie i got an oom error which may be caused by | 0 |
18,778 | 13,213,438,579 | IssuesEvent | 2020-08-16 12:48:03 | textpattern/textpattern | https://api.github.com/repos/textpattern/textpattern | closed | 'No styles recorded' message needed | usability | ### Expected behaviour
There should be a 'No styles recorded' message on the styles page panel like so when no styles are available for the current theme:
```
graf(
span(null, array('class' => 'ui-icon ui-icon-info')).' '.
gTxt('no_css_recorded'),
array('class' => 'alert-block information')
);
```
### Actual behaviour
The left-hand column is blank, no message. This is inconsistent UI behaviour compared to the rest of the admin-side.
### Steps to reproduce
1. Select a theme.
2. Unassign styles from each section (i.e. style = none).
3. Go to styles panel and delete all styles.
4. See the empty left-hand column.
#### Additional information
Textpattern version: 4.8.3, 4.9
Once someone has coded this in, I will ensure the Textpacks are updated with this extra entry. | True | 'No styles recorded' message needed - ### Expected behaviour
There should be a 'No styles recorded' message on the styles page panel like so when no styles are available for the current theme:
```
graf(
span(null, array('class' => 'ui-icon ui-icon-info')).' '.
gTxt('no_css_recorded'),
array('class' => 'alert-block information')
);
```
### Actual behaviour
The left-hand column is blank, no message. This is inconsistent UI behaviour compared to the rest of the admin-side.
### Steps to reproduce
1. Select a theme.
2. Unassign styles from each section (i.e. style = none).
3. Go to styles panel and delete all styles.
4. See the empty left-hand column.
#### Additional information
Textpattern version: 4.8.3, 4.9
Once someone has coded this in, I will ensure the Textpacks are updated with this extra entry. | non_priority | no styles recorded message needed expected behaviour there should be a no styles recorded message on the styles page panel like so when no styles are available for the current theme graf span null array class ui icon ui icon info gtxt no css recorded array class alert block information actual behaviour the left hand column is blank no message this is inconsistent ui behaviour compared to the rest of the admin side steps to reproduce select a theme unassign styles from each section i e style none go to styles panel and delete all styles see the empty left hand column additional information textpattern version once someone has coded this in i will ensure the textpacks are updated with this extra entry | 0 |
290,155 | 21,867,854,876 | IssuesEvent | 2022-05-19 01:24:37 | lukeed/cloudflare-docs | https://api.github.com/repos/lukeed/cloudflare-docs | opened | asdasd | documentation content:new | ### Which Cloudflare product(s) does this pertain to?
Analytics
### Subject Matter
asd
### Content Location
aasdasd
### Additional information
_No response_ | 1.0 | asdasd - ### Which Cloudflare product(s) does this pertain to?
Analytics
### Subject Matter
asd
### Content Location
aasdasd
### Additional information
_No response_ | non_priority | asdasd which cloudflare product s does this pertain to analytics subject matter asd content location aasdasd additional information no response | 0 |
426,112 | 29,509,836,653 | IssuesEvent | 2023-06-03 19:42:59 | michaelNXT1/SoftwareEngineeringProject2023 | https://api.github.com/repos/michaelNXT1/SoftwareEngineeringProject2023 | closed | Architecture_update | documentation Priority: Low | הארכיטקטורה העדכנית של המערכת תכלול את הרכיבים הבאים:
● רכיב לקוח )client, presentation )עדכני.
● רכיב תקשורת )communication )ורכיב notifications עדכניים.
● רכיב שירות )service )עדכנית.
● רכיב לב-מערכת )domain )עדכני.
● שכבת קישור למסד נתונים )ה-ORM מהווה את את Layer Access Data.)
● מסד נתונים | 1.0 | Architecture_update - הארכיטקטורה העדכנית של המערכת תכלול את הרכיבים הבאים:
● רכיב לקוח )client, presentation )עדכני.
● רכיב תקשורת )communication )ורכיב notifications עדכניים.
● רכיב שירות )service )עדכנית.
● רכיב לב-מערכת )domain )עדכני.
● שכבת קישור למסד נתונים )ה-ORM מהווה את את Layer Access Data.)
● מסד נתונים | non_priority | architecture update הארכיטקטורה העדכנית של המערכת תכלול את הרכיבים הבאים ● רכיב לקוח client presentation עדכני ● רכיב תקשורת communication ורכיב notifications עדכניים ● רכיב שירות service עדכנית ● רכיב לב מערכת domain עדכני ● שכבת קישור למסד נתונים ה orm מהווה את את layer access data ● מסד נתונים | 0 |
325,618 | 9,933,688,378 | IssuesEvent | 2019-07-02 12:56:12 | juju/juju-gui | https://api.github.com/repos/juju/juju-gui | closed | Redirected to blank page when loading https://jujucharms.com while logged in | bug high priority | If I'm logged in to jujucharms.com and I try to open https://jujucharms.com in a new tab, I get redirected to https://jujucharms.com/u/cmars and then redirected to https://jujucharms.com/login/u/cmars, which is a blank page.
I'm able to work around this by either:
- Opening https://jujucharms.com in a private browsing window and logging in
- Loading https://jujucharms.com/logout in a new tab and logging back in
However, a short time after being logged in (10-15 minutes?) I get redirected to the same blank page.
I am using Firefox 62.0.3 with uBlock disabled on jujucharms.com, so I don't think the issue is being caused by content blockers. | 1.0 | Redirected to blank page when loading https://jujucharms.com while logged in - If I'm logged in to jujucharms.com and I try to open https://jujucharms.com in a new tab, I get redirected to https://jujucharms.com/u/cmars and then redirected to https://jujucharms.com/login/u/cmars, which is a blank page.
I'm able to work around this by either:
- Opening https://jujucharms.com in a private browsing window and logging in
- Loading https://jujucharms.com/logout in a new tab and logging back in
However, a short time after being logged in (10-15 minutes?) I get redirected to the same blank page.
I am using Firefox 62.0.3 with uBlock disabled on jujucharms.com, so I don't think the issue is being caused by content blockers. | priority | redirected to blank page when loading while logged in if i m logged in to jujucharms com and i try to open in a new tab i get redirected to and then redirected to which is a blank page i m able to work around this by either opening in a private browsing window and logging in loading in a new tab and logging back in however a short time after being logged in minutes i get redirected to the same blank page i am using firefox with ublock disabled on jujucharms com so i don t think the issue is being caused by content blockers | 1 |
47,455 | 19,656,248,466 | IssuesEvent | 2022-01-10 12:50:22 | PreMiD/Presences | https://api.github.com/repos/PreMiD/Presences | closed | EA Origin | origin.com | Service Request | ### Discussed in https://github.com/PreMiD/Presences/discussions/4496
<div type='discussions-op-text'>
<sup>Originally posted by **ghost** September 30, 2020</sup>
**Prerequisites and essential questions**
- [x] Is it a popular site?
- [x] Is the website older than 2 months?
- [ ] Is the site locked to a specific country/region?
- [ ] Is the site a paid service? (e.g. Netflix, Hulu)
- [ ] Does the website feature NSFW content? (e.g. porn, etc...)
- [ ] Are you a donator/patron?
- [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority?
**What's your Discord username?**
web#0385
**What's the name of the service?**
www.origin.com I EA Origin
**What should the Presence display?**
What page the user is on, time elapsed. Nothing too complicated.
**If possible, please provide a logo for the service (512x512 minimum)**
<img src="https://i.imgur.com/M5aQ6b2.png" alt="presence" data-canonical-src="https://i.imgur.com/M5aQ6b2.png" style="max-width:100%;">
</div> | 1.0 | EA Origin | origin.com - ### Discussed in https://github.com/PreMiD/Presences/discussions/4496
<div type='discussions-op-text'>
<sup>Originally posted by **ghost** September 30, 2020</sup>
**Prerequisites and essential questions**
- [x] Is it a popular site?
- [x] Is the website older than 2 months?
- [ ] Is the site locked to a specific country/region?
- [ ] Is the site a paid service? (e.g. Netflix, Hulu)
- [ ] Does the website feature NSFW content? (e.g. porn, etc...)
- [ ] Are you a donator/patron?
- [x] Do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority?
**What's your Discord username?**
web#0385
**What's the name of the service?**
www.origin.com I EA Origin
**What should the Presence display?**
What page the user is on, time elapsed. Nothing too complicated.
**If possible, please provide a logo for the service (512x512 minimum)**
<img src="https://i.imgur.com/M5aQ6b2.png" alt="presence" data-canonical-src="https://i.imgur.com/M5aQ6b2.png" style="max-width:100%;">
</div> | non_priority | ea origin origin com discussed in originally posted by ghost september prerequisites and essential questions is it a popular site is the website older than months is the site locked to a specific country region is the site a paid service e g netflix hulu does the website feature nsfw content e g porn etc are you a donator patron do you acknowledge that coding presences is completely voluntary and may take time for your service to be added regardless of priority what s your discord username web what s the name of the service i ea origin what should the presence display what page the user is on time elapsed nothing too complicated if possible please provide a logo for the service minimum | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.