Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 844 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 12 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 248k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,028 | 3,315,181,195 | IssuesEvent | 2015-11-06 10:34:19 | xdoo/vaadin-demo | https://api.github.com/repos/xdoo/vaadin-demo | closed | DateValidation auf MicroService Seite funktioniert nicht | bug ready SERVICE | Die Validation des Datums auf MicroService Seite funktioniert nicht. Dies bitte beheben. | 1.0 | DateValidation auf MicroService Seite funktioniert nicht - Die Validation des Datums auf MicroService Seite funktioniert nicht. Dies bitte beheben. | non_priority | datevalidation auf microservice seite funktioniert nicht die validation des datums auf microservice seite funktioniert nicht dies bitte beheben | 0 |
48,298 | 12,192,341,815 | IssuesEvent | 2020-04-29 12:47:45 | alan-turing-institute/ptype-dmkd | https://api.github.com/repos/alan-turing-institute/ptype-dmkd | closed | Publish package to PyPI | type:feature what:build | Automate publishing to PyPI using `develop` and `master` pattern.
- [x] test upload to TestPyPI
- [x] `publish.sh` script that takes PyPI password as argument
- [x] add`pypi-password` secret to repo
- [x] GitHub job to run `publish.sh` using `pypi-password`
- [x] new `release` branch; Publish job should only run on `release`
- [x] new `develop` branch as default
- <s>add `upload-artifact` step to store `dist` folder as artifact</s>
- <s>add `download-artifact` step to publish job</s>
- <s>missing project description</s>
- `HTTPError: 400 Client Error: File already exists`
- [x] bump version number in `setup.py` (probably using [bump2version](https://pypi.org/project/bump2version/))
- `HTTPError: 400 Client Error: Only one sdist may be uploaded per release`
- [x] clean `dist` directory otherwise will end up with multiple source distributions
- [x] `git tag` with version number
- [x] `git push` to push changes to version number in `setup.py` and `.bumpversion.cfg`
- <s>only publish if [latest commit is tagged](https://packaging.python.org/guides/publishing-package-distribution-releases-using-github-actions-ci-cd-workflows/)</s>
- switch from TestPyPI to PyPI
- [x] PyPI secret in repo
- [x] parameterise `publish.sh` on `pypi` or `testpypi`
- [x] `build-test-publish.yml` to use `pypi` by default
- [x] top-level Bash script to emulate GitHub workflow (with appropriate password)
- [x] extract common Python `virtualenv` gubbins | 1.0 | Publish package to PyPI - Automate publishing to PyPI using `develop` and `master` pattern.
- [x] test upload to TestPyPI
- [x] `publish.sh` script that takes PyPI password as argument
- [x] add`pypi-password` secret to repo
- [x] GitHub job to run `publish.sh` using `pypi-password`
- [x] new `release` branch; Publish job should only run on `release`
- [x] new `develop` branch as default
- <s>add `upload-artifact` step to store `dist` folder as artifact</s>
- <s>add `download-artifact` step to publish job</s>
- <s>missing project description</s>
- `HTTPError: 400 Client Error: File already exists`
- [x] bump version number in `setup.py` (probably using [bump2version](https://pypi.org/project/bump2version/))
- `HTTPError: 400 Client Error: Only one sdist may be uploaded per release`
- [x] clean `dist` directory otherwise will end up with multiple source distributions
- [x] `git tag` with version number
- [x] `git push` to push changes to version number in `setup.py` and `.bumpversion.cfg`
- <s>only publish if [latest commit is tagged](https://packaging.python.org/guides/publishing-package-distribution-releases-using-github-actions-ci-cd-workflows/)</s>
- switch from TestPyPI to PyPI
- [x] PyPI secret in repo
- [x] parameterise `publish.sh` on `pypi` or `testpypi`
- [x] `build-test-publish.yml` to use `pypi` by default
- [x] top-level Bash script to emulate GitHub workflow (with appropriate password)
- [x] extract common Python `virtualenv` gubbins | non_priority | publish package to pypi automate publishing to pypi using develop and master pattern test upload to testpypi publish sh script that takes pypi password as argument add pypi password secret to repo github job to run publish sh using pypi password new release branch publish job should only run on release new develop branch as default add upload artifact step to store dist folder as artifact add download artifact step to publish job missing project description httperror client error file already exists bump version number in setup py probably using httperror client error only one sdist may be uploaded per release clean dist directory otherwise will end up with multiple source distributions git tag with version number git push to push changes to version number in setup py and bumpversion cfg only publish if switch from testpypi to pypi pypi secret in repo parameterise publish sh on pypi or testpypi build test publish yml to use pypi by default top level bash script to emulate github workflow with appropriate password extract common python virtualenv gubbins | 0 |
173,950 | 21,188,233,758 | IssuesEvent | 2022-04-08 14:44:40 | jgeraigery/experian-java | https://api.github.com/repos/jgeraigery/experian-java | closed | CVE-2020-36518 (High) detected in jackson-databind-2.9.2.jar - autoclosed | security vulnerability | ## CVE-2020-36518 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p>
<p>Path to vulnerable library: /ository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/d7e1610d67401ff2f3dce6bee3ce5e8f20172c2e">d7e1610d67401ff2f3dce6bee3ce5e8f20172c2e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.
WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.
<p>Publish Date: 2022-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2816">https://github.com/FasterXML/jackson-databind/issues/2816</a></p>
<p>Release Date: 2022-03-11</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.12.6.1,2.13.2.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.12.6.1,2.13.2.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36518","vulnerabilityDetails":"jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.\n WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-36518 (High) detected in jackson-databind-2.9.2.jar - autoclosed - ## CVE-2020-36518 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /MavenWorkspace/bis-services-lib/bis-services-base/pom.xml</p>
<p>Path to vulnerable library: /ository/com/fasterxml/jackson/core/jackson-databind/2.9.2/jackson-databind-2.9.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.2.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/experian-java/commit/d7e1610d67401ff2f3dce6bee3ce5e8f20172c2e">d7e1610d67401ff2f3dce6bee3ce5e8f20172c2e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.
WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.
<p>Publish Date: 2022-03-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518>CVE-2020-36518</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2816">https://github.com/FasterXML/jackson-databind/issues/2816</a></p>
<p>Release Date: 2022-03-11</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.12.6.1,2.13.2.1</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.2","packageFilePaths":["/MavenWorkspace/bis-services-lib/bis-services-base/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.12.6.1,2.13.2.1","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36518","vulnerabilityDetails":"jackson-databind before 2.13.0 allows a Java StackOverflow exception and denial of service via a large depth of nested objects.\n WhiteSource Note: After conducting further research, WhiteSource has determined that all versions of com.fasterxml.jackson.core:jackson-databind up to version 2.13.2 are vulnerable to CVE-2020-36518.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36518","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file mavenworkspace bis services lib bis services base pom xml path to vulnerable library ository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects whitesource note after conducting further research whitesource has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails jackson databind before allows a java stackoverflow exception and denial of service via a large depth of nested objects n whitesource note after conducting further research whitesource has determined that all versions of com fasterxml jackson core jackson databind up to version are vulnerable to cve vulnerabilityurl | 0 |
314,184 | 23,510,258,081 | IssuesEvent | 2022-08-18 15:52:33 | honeycombio/beeline-nodejs | https://api.github.com/repos/honeycombio/beeline-nodejs | closed | Add docs for running serverless apps locally | type: documentation | Hi there!
This isn't really an issue, more of an experience running serverless offline with the honeycomb mock implementation. It looks like you use `async_hooks.executionAsyncId()` for spanId / traceIds. In sls offline there is no guarantee that the same process will be running the request when running multiple lambdas, therefore you can get intermittent errors around missing traceIds / spanIds / stacks (as the event can be passed through but when it attempts the ctx map lookup it won't find anything).
To get around this you can add this flag in your sls offline run script in package.json `--useChildProcesses` which will ensure the context. (e.g. `sls offline --host 0.0.0.0 --useChildProcesses --lambdaPort 3004`)
Hopefully this might help someone else in the same situation :) | 1.0 | Add docs for running serverless apps locally - Hi there!
This isn't really an issue, more of an experience running serverless offline with the honeycomb mock implementation. It looks like you use `async_hooks.executionAsyncId()` for spanId / traceIds. In sls offline there is no guarantee that the same process will be running the request when running multiple lambdas, therefore you can get intermittent errors around missing traceIds / spanIds / stacks (as the event can be passed through but when it attempts the ctx map lookup it won't find anything).
To get around this you can add this flag in your sls offline run script in package.json `--useChildProcesses` which will ensure the context. (e.g. `sls offline --host 0.0.0.0 --useChildProcesses --lambdaPort 3004`)
Hopefully this might help someone else in the same situation :) | non_priority | add docs for running serverless apps locally hi there this isn t really an issue more of an experience running serverless offline with the honeycomb mock implementation it looks like you use async hooks executionasyncid for spanid traceids in sls offline there is no guarantee that the same process will be running the request when running multiple lambdas therefore you can get intermittent errors around missing traceids spanids stacks as the event can be passed through but when it attempts the ctx map lookup it won t find anything to get around this you can add this flag in your sls offline run script in package json usechildprocesses which will ensure the context e g sls offline host usechildprocesses lambdaport hopefully this might help someone else in the same situation | 0 |
8,629 | 5,877,727,059 | IssuesEvent | 2017-05-16 01:01:11 | piwik/piwik | https://api.github.com/repos/piwik/piwik | opened | Ask for confirmation before granting Anonymous user a "view" permission on any website | c: Usability Enhancement | Currently, granting "View" permission on the "anonymous" user is done without asking for permission. In one click, one may by mistake grant anyone access to view the reports for a given website.
-> In order to prevent any misclick or unexpected click, let's ask for a confirmation before granting anonymous user permission view permission on any website (via a yes/no model and a clear warning message).
Proposed message "You are about to grant the anonymous user the 'view' access to this website. This means your analytics reports and your visitors information will be publicly viewable by anyone even without a login. Are you sure you want to proceed?" | True | Ask for confirmation before granting Anonymous user a "view" permission on any website - Currently, granting "View" permission on the "anonymous" user is done without asking for permission. In one click, one may by mistake grant anyone access to view the reports for a given website.
-> In order to prevent any misclick or unexpected click, let's ask for a confirmation before granting anonymous user permission view permission on any website (via a yes/no model and a clear warning message).
Proposed message "You are about to grant the anonymous user the 'view' access to this website. This means your analytics reports and your visitors information will be publicly viewable by anyone even without a login. Are you sure you want to proceed?" | non_priority | ask for confirmation before granting anonymous user a view permission on any website currently granting view permission on the anonymous user is done without asking for permission in one click one may by mistake grant anyone access to view the reports for a given website in order to prevent any misclick or unexpected click let s ask for a confirmation before granting anonymous user permission view permission on any website via a yes no model and a clear warning message proposed message you are about to grant the anonymous user the view access to this website this means your analytics reports and your visitors information will be publicly viewable by anyone even without a login are you sure you want to proceed | 0 |
216,059 | 16,627,360,332 | IssuesEvent | 2021-06-03 11:21:27 | wultra/powerauth-apple-extensions | https://api.github.com/repos/wultra/powerauth-apple-extensions | opened | Migrate documentation | documentation | Migrate documentation from PowerAuth mobile SDK and split chapters into separate files. Keep content as-is, we'll update it later with each implementation ticket. | 1.0 | Migrate documentation - Migrate documentation from PowerAuth mobile SDK and split chapters into separate files. Keep content as-is, we'll update it later with each implementation ticket. | non_priority | migrate documentation migrate documentation from powerauth mobile sdk and split chapters into separate files keep content as is we ll update it later with each implementation ticket | 0 |
155,999 | 19,803,188,050 | IssuesEvent | 2022-01-19 01:38:29 | Squ3D/Trident_MMO | https://api.github.com/repos/Squ3D/Trident_MMO | opened | CVE-2020-9488 (Low) detected in log4j-1.2.15.jar | security vulnerability | ## CVE-2020-9488 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.15.jar</b></p></summary>
<p>Java logging framework</p>
<p>Library home page: <a href="http://logging.apache.org/">http://logging.apache.org/</a></p>
<p>Path to vulnerable library: /Trident_MMO/Source/lib/log4j-1.2.15.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.15.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-9488 (Low) detected in log4j-1.2.15.jar - ## CVE-2020-9488 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.15.jar</b></p></summary>
<p>Java logging framework</p>
<p>Library home page: <a href="http://logging.apache.org/">http://logging.apache.org/</a></p>
<p>Path to vulnerable library: /Trident_MMO/Source/lib/log4j-1.2.15.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-1.2.15.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve low detected in jar cve low severity vulnerability vulnerable library jar java logging framework library home page a href path to vulnerable library trident mmo source lib jar dependency hierarchy x jar vulnerable library vulnerability details improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core step up your open source security game with whitesource | 0 |
16,609 | 6,246,082,939 | IssuesEvent | 2017-07-13 02:25:16 | Frannsoft/FrannHammer | https://api.github.com/repos/Frannsoft/FrannHammer | closed | Deploy database to MongoDb hosting provider | Build enhancement Refactoring | mLab seems like a solid choice for this so far. This Issue includes:
- Creating an automatic backup process
- Automatically running health tests on the database
- Seeding the database automatically.
This issue will likely also involve the creation of a service that monitors the KH site files on the current VM for changes. When a file changes, flush the Redis cache and reseed to maintain updated data. Once implemented this will drastically decrease feature disparity between the existing static site and API. | 1.0 | Deploy database to MongoDb hosting provider - mLab seems like a solid choice for this so far. This Issue includes:
- Creating an automatic backup process
- Automatically running health tests on the database
- Seeding the database automatically.
This issue will likely also involve the creation of a service that monitors the KH site files on the current VM for changes. When a file changes, flush the Redis cache and reseed to maintain updated data. Once implemented this will drastically decrease feature disparity between the existing static site and API. | non_priority | deploy database to mongodb hosting provider mlab seems like a solid choice for this so far this issue includes creating an automatic backup process automatically running health tests on the database seeding the database automatically this issue will likely also involve the creation of a service that monitors the kh site files on the current vm for changes when a file changes flush the redis cache and reseed to maintain updated data once implemented this will drastically decrease feature disparity between the existing static site and api | 0 |
152,798 | 19,697,089,776 | IssuesEvent | 2022-01-12 13:17:45 | Shai-Demo-Org/SecurityShepherd | https://api.github.com/repos/Shai-Demo-Org/SecurityShepherd | opened | CVE-2020-2934 (Medium) detected in mysql-connector-java-5.1.24.jar | security vulnerability | ## CVE-2020-2934 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.24.jar</b></p></summary>
<p>MySQL JDBC Type 4 driver</p>
<p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /tory/mysql/mysql-connector-java/5.1.24/mysql-connector-java-5.1.24.jar</p>
<p>
Dependency Hierarchy:
- :x: **mysql-connector-java-5.1.24.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Shai-Demo-Org/SecurityShepherd/commit/7831d8d3dee27b4f3484226e8279dee4e185d3f2">7831d8d3dee27b4f3484226e8279dee4e185d3f2</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 8.0.19 and prior and 5.1.48 and prior. Difficult to exploit vulnerability allows unauthenticated attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks require human interaction from a person other than the attacker. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data as well as unauthorized read access to a subset of MySQL Connectors accessible data and unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 5.0 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:N/UI:R/S:U/C:L/I:L/A:L).
<p>Publish Date: 2020-04-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2934>CVE-2020-2934</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.oracle.com/security-alerts/cpuapr2020.html">https://www.oracle.com/security-alerts/cpuapr2020.html</a></p>
<p>Release Date: 2020-04-15</p>
<p>Fix Resolution: mysql:mysql-connector-java:5.1.49,8.0.20</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.24","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.24","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:5.1.49,8.0.20","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-2934","vulnerabilityDetails":"Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 8.0.19 and prior and 5.1.48 and prior. Difficult to exploit vulnerability allows unauthenticated attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks require human interaction from a person other than the attacker. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data as well as unauthorized read access to a subset of MySQL Connectors accessible data and unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 5.0 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:N/UI:R/S:U/C:L/I:L/A:L).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2934","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-2934 (Medium) detected in mysql-connector-java-5.1.24.jar - ## CVE-2020-2934 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-connector-java-5.1.24.jar</b></p></summary>
<p>MySQL JDBC Type 4 driver</p>
<p>Library home page: <a href="http://dev.mysql.com/doc/connector-j/en/">http://dev.mysql.com/doc/connector-j/en/</a></p>
<p>Path to dependency file: /pom.xml</p>
<p>Path to vulnerable library: /tory/mysql/mysql-connector-java/5.1.24/mysql-connector-java-5.1.24.jar</p>
<p>
Dependency Hierarchy:
- :x: **mysql-connector-java-5.1.24.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Shai-Demo-Org/SecurityShepherd/commit/7831d8d3dee27b4f3484226e8279dee4e185d3f2">7831d8d3dee27b4f3484226e8279dee4e185d3f2</a></p>
<p>Found in base branch: <b>dev</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 8.0.19 and prior and 5.1.48 and prior. Difficult to exploit vulnerability allows unauthenticated attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks require human interaction from a person other than the attacker. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data as well as unauthorized read access to a subset of MySQL Connectors accessible data and unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 5.0 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:N/UI:R/S:U/C:L/I:L/A:L).
<p>Publish Date: 2020-04-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2934>CVE-2020-2934</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.oracle.com/security-alerts/cpuapr2020.html">https://www.oracle.com/security-alerts/cpuapr2020.html</a></p>
<p>Release Date: 2020-04-15</p>
<p>Fix Resolution: mysql:mysql-connector-java:5.1.49,8.0.20</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"mysql","packageName":"mysql-connector-java","packageVersion":"5.1.24","packageFilePaths":["/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"mysql:mysql-connector-java:5.1.24","isMinimumFixVersionAvailable":true,"minimumFixVersion":"mysql:mysql-connector-java:5.1.49,8.0.20","isBinary":false}],"baseBranches":["dev"],"vulnerabilityIdentifier":"CVE-2020-2934","vulnerabilityDetails":"Vulnerability in the MySQL Connectors product of Oracle MySQL (component: Connector/J). Supported versions that are affected are 8.0.19 and prior and 5.1.48 and prior. Difficult to exploit vulnerability allows unauthenticated attacker with network access via multiple protocols to compromise MySQL Connectors. Successful attacks require human interaction from a person other than the attacker. Successful attacks of this vulnerability can result in unauthorized update, insert or delete access to some of MySQL Connectors accessible data as well as unauthorized read access to a subset of MySQL Connectors accessible data and unauthorized ability to cause a partial denial of service (partial DOS) of MySQL Connectors. CVSS 3.0 Base Score 5.0 (Confidentiality, Integrity and Availability impacts). CVSS Vector: (CVSS:3.0/AV:N/AC:H/PR:N/UI:R/S:U/C:L/I:L/A:L).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-2934","cvss3Severity":"medium","cvss3Score":"5.0","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in mysql connector java jar cve medium severity vulnerability vulnerable library mysql connector java jar mysql jdbc type driver library home page a href path to dependency file pom xml path to vulnerable library tory mysql mysql connector java mysql connector java jar dependency hierarchy x mysql connector java jar vulnerable library found in head commit a href found in base branch dev vulnerability details vulnerability in the mysql connectors product of oracle mysql component connector j supported versions that are affected are and prior and and prior difficult to exploit vulnerability allows unauthenticated attacker with network access via multiple protocols to compromise mysql connectors successful attacks require human interaction from a person other than the attacker successful attacks of this vulnerability can result in unauthorized update insert or delete access to some of mysql connectors accessible data as well as unauthorized read access to a subset of mysql connectors accessible data and unauthorized ability to cause a partial denial of service partial dos of mysql connectors cvss base score confidentiality integrity and availability impacts cvss vector cvss av n ac h pr n ui r s u c l i l a l publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution mysql mysql connector java rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree mysql mysql connector java isminimumfixversionavailable true minimumfixversion mysql mysql connector java isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails vulnerability in the mysql connectors product of oracle mysql component connector j supported versions that are affected are and prior and and prior difficult to exploit vulnerability allows unauthenticated attacker with network access via multiple protocols to compromise mysql connectors successful attacks require human interaction from a person other than the attacker successful attacks of this vulnerability can result in unauthorized update insert or delete access to some of mysql connectors accessible data as well as unauthorized read access to a subset of mysql connectors accessible data and unauthorized ability to cause a partial denial of service partial dos of mysql connectors cvss base score confidentiality integrity and availability impacts cvss vector cvss av n ac h pr n ui r s u c l i l a l vulnerabilityurl | 0 |
253,452 | 19,101,796,557 | IssuesEvent | 2021-11-29 23:50:14 | aws-amplify/amplify-cli | https://api.github.com/repos/aws-amplify/amplify-cli | closed | Labels on PR's | feature-request documentation | Given there are >50 PR's open right now, the majority over 15 days old...it might be helpful to do some triaging/management with labels...which ones are critical? which ones are just complicated or waiting on other functionality? Which ones are new features, fixes, security? | 1.0 | Labels on PR's - Given there are >50 PR's open right now, the majority over 15 days old...it might be helpful to do some triaging/management with labels...which ones are critical? which ones are just complicated or waiting on other functionality? Which ones are new features, fixes, security? | non_priority | labels on pr s given there are pr s open right now the majority over days old it might be helpful to do some triaging management with labels which ones are critical which ones are just complicated or waiting on other functionality which ones are new features fixes security | 0 |
138,811 | 11,217,681,948 | IssuesEvent | 2020-01-07 09:48:02 | microsoft/azure-pipelines-tasks | https://api.github.com/repos/microsoft/azure-pipelines-tasks | closed | PublishTestResultsV2: Introduce option to reduce logging verbosity | Area: Test Area: TestManagement enhancement | ## Required Information
**Question, Bug, or Feature?**
*Type*: Feature
**Enter Task Name**: PublishTestResultsV2
## Issue Description
As you can see in [this example](https://dev.azure.com/rmetzger/Flink/_build/results?buildId=3585&view=logs&j=821169f2-253d-577e-fcd2-77ed3c3937d9&t=428c25e8-0840-52a2-5aac-e4f460b57078), our builds are "polluted" with ~800 log messages from the `PublishTestResultsV2` task.
It would be nice to have a config flag to reduce the verbosity.
Example log messages:
```
Only single test suite found, parsing its information.
Timestamp is not available for one or more testsuites. Total run duration is being calculated as the sum of time durations of detected testsuites.
```
The problem is that we are publishing the test results after the main build tasks. Engineers trying to debug build failures always need to reverse-scroll over the `PublishTestResultsV2` log messages.
| 2.0 | PublishTestResultsV2: Introduce option to reduce logging verbosity - ## Required Information
**Question, Bug, or Feature?**
*Type*: Feature
**Enter Task Name**: PublishTestResultsV2
## Issue Description
As you can see in [this example](https://dev.azure.com/rmetzger/Flink/_build/results?buildId=3585&view=logs&j=821169f2-253d-577e-fcd2-77ed3c3937d9&t=428c25e8-0840-52a2-5aac-e4f460b57078), our builds are "polluted" with ~800 log messages from the `PublishTestResultsV2` task.
It would be nice to have a config flag to reduce the verbosity.
Example log messages:
```
Only single test suite found, parsing its information.
Timestamp is not available for one or more testsuites. Total run duration is being calculated as the sum of time durations of detected testsuites.
```
The problem is that we are publishing the test results after the main build tasks. Engineers trying to debug build failures always need to reverse-scroll over the `PublishTestResultsV2` log messages.
| non_priority | introduce option to reduce logging verbosity required information question bug or feature type feature enter task name issue description as you can see in our builds are polluted with log messages from the task it would be nice to have a config flag to reduce the verbosity example log messages only single test suite found parsing its information timestamp is not available for one or more testsuites total run duration is being calculated as the sum of time durations of detected testsuites the problem is that we are publishing the test results after the main build tasks engineers trying to debug build failures always need to reverse scroll over the log messages | 0 |
94,591 | 11,888,013,396 | IssuesEvent | 2020-03-28 05:34:35 | azl397985856/fe-interview | https://api.github.com/repos/azl397985856/fe-interview | closed | 【每日一题】- 2020-01-20 - node cli如何实现只显示两行信息的功能? | Daily Question Design Node stale | 如图是显示一行:

我们不希望显示超过两行内容,我们超过两行之后需要将之前的内容进行清除
| 1.0 | 【每日一题】- 2020-01-20 - node cli如何实现只显示两行信息的功能? - 如图是显示一行:

我们不希望显示超过两行内容,我们超过两行之后需要将之前的内容进行清除
| non_priority | 【每日一题】 node cli如何实现只显示两行信息的功能? 如图是显示一行: 我们不希望显示超过两行内容,我们超过两行之后需要将之前的内容进行清除 | 0 |
243,581 | 20,507,208,012 | IssuesEvent | 2022-03-01 00:02:14 | vmware-tanzu/community-edition | https://api.github.com/repos/vmware-tanzu/community-edition | closed | Figure out what to do with Grafana (new licensing) | area/packages kind/test-release | ## Feedback
Grafana has switched their licensing to AGPL. Which raises a question for whether we can repackage and distribute Grafana in TCE. | 1.0 | Figure out what to do with Grafana (new licensing) - ## Feedback
Grafana has switched their licensing to AGPL. Which raises a question for whether we can repackage and distribute Grafana in TCE. | non_priority | figure out what to do with grafana new licensing feedback grafana has switched their licensing to agpl which raises a question for whether we can repackage and distribute grafana in tce | 0 |
197,434 | 15,680,291,532 | IssuesEvent | 2021-03-25 02:32:13 | fga-eps-mds/2020.2-Anunbis | https://api.github.com/repos/fga-eps-mds/2020.2-Anunbis | closed | Revisar documento de arquitetura | documentation | ### Descrição:
<!-- Descrever de maneira clara e objetiva o propósito da issue. -->
Revisar documento de arquitetura para o momento onde estamos
### Tarefas:
<!-- Checklist de ações que devem ser realizadas. -->
- [x] Adição do modelo MVC do backend
- [x] Adição do diagrama de pacotes do frontend
- [x] Adição do diagrama de pacotes do backend
- [x] Diagrama logico dos dados
- [x] Diagrama entidade relacionamento
- [x] Adicionar novo atributo de Professor "id_professor"
- [x] Adicionar novo atributo de Post "is_anonymous"
- [x] Atributo de professor 'rating' foi removido
### Critérios de aceitação:
<!-- Descrever os requisitos necessários para que a issue possar ser finalizada. -->
- [ ] O documento deve estar atualizado.
| 1.0 | Revisar documento de arquitetura - ### Descrição:
<!-- Descrever de maneira clara e objetiva o propósito da issue. -->
Revisar documento de arquitetura para o momento onde estamos
### Tarefas:
<!-- Checklist de ações que devem ser realizadas. -->
- [x] Adição do modelo MVC do backend
- [x] Adição do diagrama de pacotes do frontend
- [x] Adição do diagrama de pacotes do backend
- [x] Diagrama logico dos dados
- [x] Diagrama entidade relacionamento
- [x] Adicionar novo atributo de Professor "id_professor"
- [x] Adicionar novo atributo de Post "is_anonymous"
- [x] Atributo de professor 'rating' foi removido
### Critérios de aceitação:
<!-- Descrever os requisitos necessários para que a issue possar ser finalizada. -->
- [ ] O documento deve estar atualizado.
| non_priority | revisar documento de arquitetura descrição revisar documento de arquitetura para o momento onde estamos tarefas adição do modelo mvc do backend adição do diagrama de pacotes do frontend adição do diagrama de pacotes do backend diagrama logico dos dados diagrama entidade relacionamento adicionar novo atributo de professor id professor adicionar novo atributo de post is anonymous atributo de professor rating foi removido critérios de aceitação o documento deve estar atualizado | 0 |
229,708 | 18,416,903,064 | IssuesEvent | 2021-10-13 12:28:46 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | opened | [CI] MlDeprecationIT testMlDeprecationChecks failing | >test-failure Team:ML | **Build scan:**
https://gradle-enterprise.elastic.co/s/u2qkxilapasye/tests/:x-pack:plugin:deprecation:qa:rest:javaRestTest/org.elasticsearch.xpack.deprecation.MlDeprecationIT/testMlDeprecationChecks
**Reproduction line:**
`./gradlew ':x-pack:plugin:deprecation:qa:rest:javaRestTest' --tests "org.elasticsearch.xpack.deprecation.MlDeprecationIT.testMlDeprecationChecks" -Dtests.seed=9D1455189E3BE57E -Dtests.locale=sk-SK -Dtests.timezone=Etc/GMT+12 -Druntime.java=11`
**Applicable branches:**
master
**Reproduces locally?:**
Didn't try
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.deprecation.MlDeprecationIT&tests.test=testMlDeprecationChecks
**Failure excerpt:**
```
java.lang.RuntimeException: failed to delete policy: .deprecation-indexing-ilm-policy
at __randomizedtesting.SeedInfo.seed([9D1455189E3BE57E:B20D02CBADA83A35]:0)
at org.elasticsearch.test.rest.ESRestTestCase.lambda$deleteAllILMPolicies$17(ESRestTestCase.java:1015)
at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183)
at java.util.stream.ReferencePipeline$2$1.accept(ReferencePipeline.java:177)
at java.util.HashMap$KeySpliterator.forEachRemaining(HashMap.java:1603)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:150)
at java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:173)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:497)
at org.elasticsearch.test.rest.ESRestTestCase.deleteAllILMPolicies(ESRestTestCase.java:1011)
at org.elasticsearch.test.rest.ESRestTestCase.wipeCluster(ESRestTestCase.java:748)
at org.elasticsearch.test.rest.ESRestTestCase.cleanUpCluster(ESRestTestCase.java:371)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:566)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:1004)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:824)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:475)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:831)
at java.lang.Thread.run(Thread.java:834)
Caused by: org.elasticsearch.client.ResponseException: method [DELETE], host [http://127.0.0.1:35261], URI [/_ilm/policy/.deprecation-indexing-ilm-policy], status line [HTTP/1.1 400 Bad Request]
{"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"Cannot delete policy [.deprecation-indexing-ilm-policy]. It is in use by one or more indices: [.ds-.logs-deprecation.elasticsearch-default-2021.10.13-000001]"}],"type":"illegal_argument_exception","reason":"Cannot delete policy [.deprecation-indexing-ilm-policy]. It is in use by one or more indices: [.ds-.logs-deprecation.elasticsearch-default-2021.10.13-000001]"},"status":400}
at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:329)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:295)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:269)
at org.elasticsearch.test.rest.ESRestTestCase.lambda$deleteAllILMPolicies$17(ESRestTestCase.java:1013)
at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183)
at java.util.stream.ReferencePipeline$2$1.accept(ReferencePipeline.java:177)
at java.util.HashMap$KeySpliterator.forEachRemaining(HashMap.java:1603)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:150)
at java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:173)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:497)
at org.elasticsearch.test.rest.ESRestTestCase.deleteAllILMPolicies(ESRestTestCase.java:1011)
at org.elasticsearch.test.rest.ESRestTestCase.wipeCluster(ESRestTestCase.java:748)
at org.elasticsearch.test.rest.ESRestTestCase.cleanUpCluster(ESRestTestCase.java:371)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:566)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:1004)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:824)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:475)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:831)
at java.lang.Thread.run(Thread.java:834)
``` | 1.0 | [CI] MlDeprecationIT testMlDeprecationChecks failing - **Build scan:**
https://gradle-enterprise.elastic.co/s/u2qkxilapasye/tests/:x-pack:plugin:deprecation:qa:rest:javaRestTest/org.elasticsearch.xpack.deprecation.MlDeprecationIT/testMlDeprecationChecks
**Reproduction line:**
`./gradlew ':x-pack:plugin:deprecation:qa:rest:javaRestTest' --tests "org.elasticsearch.xpack.deprecation.MlDeprecationIT.testMlDeprecationChecks" -Dtests.seed=9D1455189E3BE57E -Dtests.locale=sk-SK -Dtests.timezone=Etc/GMT+12 -Druntime.java=11`
**Applicable branches:**
master
**Reproduces locally?:**
Didn't try
**Failure history:**
https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.deprecation.MlDeprecationIT&tests.test=testMlDeprecationChecks
**Failure excerpt:**
```
java.lang.RuntimeException: failed to delete policy: .deprecation-indexing-ilm-policy
at __randomizedtesting.SeedInfo.seed([9D1455189E3BE57E:B20D02CBADA83A35]:0)
at org.elasticsearch.test.rest.ESRestTestCase.lambda$deleteAllILMPolicies$17(ESRestTestCase.java:1015)
at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183)
at java.util.stream.ReferencePipeline$2$1.accept(ReferencePipeline.java:177)
at java.util.HashMap$KeySpliterator.forEachRemaining(HashMap.java:1603)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:150)
at java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:173)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:497)
at org.elasticsearch.test.rest.ESRestTestCase.deleteAllILMPolicies(ESRestTestCase.java:1011)
at org.elasticsearch.test.rest.ESRestTestCase.wipeCluster(ESRestTestCase.java:748)
at org.elasticsearch.test.rest.ESRestTestCase.cleanUpCluster(ESRestTestCase.java:371)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:566)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:1004)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:824)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:475)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:831)
at java.lang.Thread.run(Thread.java:834)
Caused by: org.elasticsearch.client.ResponseException: method [DELETE], host [http://127.0.0.1:35261], URI [/_ilm/policy/.deprecation-indexing-ilm-policy], status line [HTTP/1.1 400 Bad Request]
{"error":{"root_cause":[{"type":"illegal_argument_exception","reason":"Cannot delete policy [.deprecation-indexing-ilm-policy]. It is in use by one or more indices: [.ds-.logs-deprecation.elasticsearch-default-2021.10.13-000001]"}],"type":"illegal_argument_exception","reason":"Cannot delete policy [.deprecation-indexing-ilm-policy]. It is in use by one or more indices: [.ds-.logs-deprecation.elasticsearch-default-2021.10.13-000001]"},"status":400}
at org.elasticsearch.client.RestClient.convertResponse(RestClient.java:329)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:295)
at org.elasticsearch.client.RestClient.performRequest(RestClient.java:269)
at org.elasticsearch.test.rest.ESRestTestCase.lambda$deleteAllILMPolicies$17(ESRestTestCase.java:1013)
at java.util.stream.ForEachOps$ForEachOp$OfRef.accept(ForEachOps.java:183)
at java.util.stream.ReferencePipeline$2$1.accept(ReferencePipeline.java:177)
at java.util.HashMap$KeySpliterator.forEachRemaining(HashMap.java:1603)
at java.util.stream.AbstractPipeline.copyInto(AbstractPipeline.java:484)
at java.util.stream.AbstractPipeline.wrapAndCopyInto(AbstractPipeline.java:474)
at java.util.stream.ForEachOps$ForEachOp.evaluateSequential(ForEachOps.java:150)
at java.util.stream.ForEachOps$ForEachOp$OfRef.evaluateSequential(ForEachOps.java:173)
at java.util.stream.AbstractPipeline.evaluate(AbstractPipeline.java:234)
at java.util.stream.ReferencePipeline.forEach(ReferencePipeline.java:497)
at org.elasticsearch.test.rest.ESRestTestCase.deleteAllILMPolicies(ESRestTestCase.java:1011)
at org.elasticsearch.test.rest.ESRestTestCase.wipeCluster(ESRestTestCase.java:748)
at org.elasticsearch.test.rest.ESRestTestCase.cleanUpCluster(ESRestTestCase.java:371)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2)
at jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:566)
at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758)
at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:1004)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:44)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:45)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:824)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:475)
at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955)
at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840)
at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891)
at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:38)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53)
at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:43)
at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:44)
at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:60)
at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:47)
at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36)
at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:375)
at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:831)
at java.lang.Thread.run(Thread.java:834)
``` | non_priority | mldeprecationit testmldeprecationchecks failing build scan reproduction line gradlew x pack plugin deprecation qa rest javaresttest tests org elasticsearch xpack deprecation mldeprecationit testmldeprecationchecks dtests seed dtests locale sk sk dtests timezone etc gmt druntime java applicable branches master reproduces locally didn t try failure history failure excerpt java lang runtimeexception failed to delete policy deprecation indexing ilm policy at randomizedtesting seedinfo seed at org elasticsearch test rest esresttestcase lambda deleteallilmpolicies esresttestcase java at java util stream foreachops foreachop ofref accept foreachops java at java util stream referencepipeline accept referencepipeline java at java util hashmap keyspliterator foreachremaining hashmap java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream foreachops foreachop evaluatesequential foreachops java at java util stream foreachops foreachop ofref evaluatesequential foreachops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline foreach referencepipeline java at org elasticsearch test rest esresttestcase deleteallilmpolicies esresttestcase java at org elasticsearch test rest esresttestcase wipecluster esresttestcase java at org elasticsearch test rest esresttestcase cleanupcluster esresttestcase java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java caused by org elasticsearch client responseexception method host uri status line error root cause it is in use by one or more indices type illegal argument exception reason cannot delete policy it is in use by one or more indices status at org elasticsearch client restclient convertresponse restclient java at org elasticsearch client restclient performrequest restclient java at org elasticsearch client restclient performrequest restclient java at org elasticsearch test rest esresttestcase lambda deleteallilmpolicies esresttestcase java at java util stream foreachops foreachop ofref accept foreachops java at java util stream referencepipeline accept referencepipeline java at java util hashmap keyspliterator foreachremaining hashmap java at java util stream abstractpipeline copyinto abstractpipeline java at java util stream abstractpipeline wrapandcopyinto abstractpipeline java at java util stream foreachops foreachop evaluatesequential foreachops java at java util stream foreachops foreachop ofref evaluatesequential foreachops java at java util stream abstractpipeline evaluate abstractpipeline java at java util stream referencepipeline foreach referencepipeline java at org elasticsearch test rest esresttestcase deleteallilmpolicies esresttestcase java at org elasticsearch test rest esresttestcase wipecluster esresttestcase java at org elasticsearch test rest esresttestcase cleanupcluster esresttestcase java at jdk internal reflect nativemethodaccessorimpl nativemethodaccessorimpl java at jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java | 0 |
180,242 | 30,473,090,056 | IssuesEvent | 2023-07-17 14:46:26 | Tonomy-Foundation/Tonomy-ID-Management | https://api.github.com/repos/Tonomy-Foundation/Tonomy-ID-Management | opened | Design VC sign animation v3: mockup | design | Acceptance criteria
- [ ] new design of sign VC animation (when user presses sign button)
- [ ] should show the steps taken
- [ ] good to have (not required): also show how fast it happens
- [ ] reviewed and accepted by 2x BD team
STEPS:
1. Fetching sovereign signer and check key is still valid(1000ms)
2. Creating W3C Verifiable Credential data structure (100ms)
3. Signing data (100ms) | 1.0 | Design VC sign animation v3: mockup - Acceptance criteria
- [ ] new design of sign VC animation (when user presses sign button)
- [ ] should show the steps taken
- [ ] good to have (not required): also show how fast it happens
- [ ] reviewed and accepted by 2x BD team
STEPS:
1. Fetching sovereign signer and check key is still valid(1000ms)
2. Creating W3C Verifiable Credential data structure (100ms)
3. Signing data (100ms) | non_priority | design vc sign animation mockup acceptance criteria new design of sign vc animation when user presses sign button should show the steps taken good to have not required also show how fast it happens reviewed and accepted by bd team steps fetching sovereign signer and check key is still valid creating verifiable credential data structure signing data | 0 |
16,366 | 9,378,536,927 | IssuesEvent | 2019-04-04 13:07:31 | kyma-project/console | https://api.github.com/repos/kyma-project/console | closed | Load Test Lambda Implementation | Epic area/serverless enhancement quality/performance | **Description**
As a DevOps I need to know the performance limits of my lambda function. To prove the service quality of the lambda functionality frequently running load tests should be integrated into the prow pipeline.
AC:
- Determine the max number of requests for each Function size with three different performing functions.
- Have a report including graphs to show the performance over time (including the HPA)
- (Think about and maybe implement) Develop a metrics to test if the performance has changed between two runs.
**Reasons**
Serving requests is the key feature of a lambda functions. We need to ensure to know the limits and find performance problems upfront.
| True | Load Test Lambda Implementation - **Description**
As a DevOps I need to know the performance limits of my lambda function. To prove the service quality of the lambda functionality frequently running load tests should be integrated into the prow pipeline.
AC:
- Determine the max number of requests for each Function size with three different performing functions.
- Have a report including graphs to show the performance over time (including the HPA)
- (Think about and maybe implement) Develop a metrics to test if the performance has changed between two runs.
**Reasons**
Serving requests is the key feature of a lambda functions. We need to ensure to know the limits and find performance problems upfront.
| non_priority | load test lambda implementation description as a devops i need to know the performance limits of my lambda function to prove the service quality of the lambda functionality frequently running load tests should be integrated into the prow pipeline ac determine the max number of requests for each function size with three different performing functions have a report including graphs to show the performance over time including the hpa think about and maybe implement develop a metrics to test if the performance has changed between two runs reasons serving requests is the key feature of a lambda functions we need to ensure to know the limits and find performance problems upfront | 0 |
253,716 | 19,162,246,489 | IssuesEvent | 2021-12-03 02:32:45 | santiagov916/jest-another-rpg | https://api.github.com/repos/santiagov916/jest-another-rpg | opened | Consolidate the code. | documentation | We don't yet know exactly what this step entails, but we do know that we'll need to refactor our code after the tests pass, so we'll create this issue as a placeholder. | 1.0 | Consolidate the code. - We don't yet know exactly what this step entails, but we do know that we'll need to refactor our code after the tests pass, so we'll create this issue as a placeholder. | non_priority | consolidate the code we don t yet know exactly what this step entails but we do know that we ll need to refactor our code after the tests pass so we ll create this issue as a placeholder | 0 |
309,480 | 23,297,027,207 | IssuesEvent | 2022-08-06 18:43:49 | Pocco81/auto-save.nvim | https://api.github.com/repos/Pocco81/auto-save.nvim | closed | auto-save.nvim rewrite: everything changed | documentation enhancement | Heya! I decided to rewrite auto-save because I kind of disliked the previous state of the code-base.
### Notes:
1. Backwards compatibility with previous auto-save confs won't be provided
2. This plugin is completely different from the original, so please update your settings!
3. Previous issues won't be attended because `2`. If your issue still exists please open a new one.
4. I added previous feature requests as I saw fit.
Enjoy :partying_face: :tada: | 1.0 | auto-save.nvim rewrite: everything changed - Heya! I decided to rewrite auto-save because I kind of disliked the previous state of the code-base.
### Notes:
1. Backwards compatibility with previous auto-save confs won't be provided
2. This plugin is completely different from the original, so please update your settings!
3. Previous issues won't be attended because `2`. If your issue still exists please open a new one.
4. I added previous feature requests as I saw fit.
Enjoy :partying_face: :tada: | non_priority | auto save nvim rewrite everything changed heya i decided to rewrite auto save because i kind of disliked the previous state of the code base notes backwards compatibility with previous auto save confs won t be provided this plugin is completely different from the original so please update your settings previous issues won t be attended because if your issue still exists please open a new one i added previous feature requests as i saw fit enjoy partying face tada | 0 |
37,707 | 10,060,599,016 | IssuesEvent | 2019-07-22 19:14:10 | VowpalWabbit/vowpal_wabbit | https://api.github.com/repos/VowpalWabbit/vowpal_wabbit | closed | python wrapper cannot compile on python3.7 | Build Issue Lang: Python | ```
cd python; make things
make[1]: Entering directory '/tmp/pip-install-5vj7bcv7/vowpalwabbit/src/python'
Using Python 3.7
/usr/bin/g++ -std=c++0x -I/home/wenjian/anaconda3/include/python3.7m -I/home/wenjian/anaconda3/include/python3.7m -I /usr/local/include/boost -I /usr/include -I ../rapidjson/include -fPIC -c pylibvw.cc -o pylibvw.o
/usr/bin/g++ -shared -Wl,--export-dynamic pylibvw.o -L /usr/local/lib -L /usr/lib -L /usr/lib/x86_64-linux-gnu -L/home/wenjian/anaconda3/lib/python3.7/config-3.7m-x86_64-linux-gnu -L/home/wenjian/anaconda3/lib -lpython3.7m -lpthread -ldl -lutil -lrt -lm -Xlinker -export-dynamic ../vowpalwabbit/libvw.a ../vowpalwabbit/liballreduce.a -L /usr/local/lib -L /usr/lib -L /usr/lib/x86_64-linux-gnu -l boost_program_options -l pthread -l z -l boost_python3 -o pylibvw.so
lto1: internal compiler error: in lto_tag_to_tree_code, at lto-streamer.h:1005
Please submit a full bug report,
with preprocessed source if appropriate.
See <file:///usr/share/doc/gcc-7/README.Bugs> for instructions.
lto-wrapper: fatal error: /usr/bin/g++ returned 1 exit status
compilation terminated.
/usr/bin/ld: error: lto-wrapper failed
collect2: error: ld returned 1 exit status
Makefile:55: recipe for target 'pylibvw.so' failed
make[1]: *** [pylibvw.so] Error 1
make[1]: Leaving directory '/tmp/pip-install-5vj7bcv7/vowpalwabbit/src/python'
Makefile:117: recipe for target 'python' failed
make: *** [python] Error 2
``` | 1.0 | python wrapper cannot compile on python3.7 - ```
cd python; make things
make[1]: Entering directory '/tmp/pip-install-5vj7bcv7/vowpalwabbit/src/python'
Using Python 3.7
/usr/bin/g++ -std=c++0x -I/home/wenjian/anaconda3/include/python3.7m -I/home/wenjian/anaconda3/include/python3.7m -I /usr/local/include/boost -I /usr/include -I ../rapidjson/include -fPIC -c pylibvw.cc -o pylibvw.o
/usr/bin/g++ -shared -Wl,--export-dynamic pylibvw.o -L /usr/local/lib -L /usr/lib -L /usr/lib/x86_64-linux-gnu -L/home/wenjian/anaconda3/lib/python3.7/config-3.7m-x86_64-linux-gnu -L/home/wenjian/anaconda3/lib -lpython3.7m -lpthread -ldl -lutil -lrt -lm -Xlinker -export-dynamic ../vowpalwabbit/libvw.a ../vowpalwabbit/liballreduce.a -L /usr/local/lib -L /usr/lib -L /usr/lib/x86_64-linux-gnu -l boost_program_options -l pthread -l z -l boost_python3 -o pylibvw.so
lto1: internal compiler error: in lto_tag_to_tree_code, at lto-streamer.h:1005
Please submit a full bug report,
with preprocessed source if appropriate.
See <file:///usr/share/doc/gcc-7/README.Bugs> for instructions.
lto-wrapper: fatal error: /usr/bin/g++ returned 1 exit status
compilation terminated.
/usr/bin/ld: error: lto-wrapper failed
collect2: error: ld returned 1 exit status
Makefile:55: recipe for target 'pylibvw.so' failed
make[1]: *** [pylibvw.so] Error 1
make[1]: Leaving directory '/tmp/pip-install-5vj7bcv7/vowpalwabbit/src/python'
Makefile:117: recipe for target 'python' failed
make: *** [python] Error 2
``` | non_priority | python wrapper cannot compile on cd python make things make entering directory tmp pip install vowpalwabbit src python using python usr bin g std c i home wenjian include i home wenjian include i usr local include boost i usr include i rapidjson include fpic c pylibvw cc o pylibvw o usr bin g shared wl export dynamic pylibvw o l usr local lib l usr lib l usr lib linux gnu l home wenjian lib config linux gnu l home wenjian lib lpthread ldl lutil lrt lm xlinker export dynamic vowpalwabbit libvw a vowpalwabbit liballreduce a l usr local lib l usr lib l usr lib linux gnu l boost program options l pthread l z l boost o pylibvw so internal compiler error in lto tag to tree code at lto streamer h please submit a full bug report with preprocessed source if appropriate see for instructions lto wrapper fatal error usr bin g returned exit status compilation terminated usr bin ld error lto wrapper failed error ld returned exit status makefile recipe for target pylibvw so failed make error make leaving directory tmp pip install vowpalwabbit src python makefile recipe for target python failed make error | 0 |
105,745 | 13,211,834,625 | IssuesEvent | 2020-08-16 02:24:05 | microsoft/pyright | https://api.github.com/repos/microsoft/pyright | closed | Type "str" cannot be assigned to type "str | None" | as designed | ```python
from typing import List, Optional
a = list(["a", "b", "c"])
aopt1: List[Optional[str]] = list(["a", "b", "c"])
aopt2: List[Optional[str]] = a
```
pyright 1.1.62 reports a type error on `aopt2`:
```
6:30 - error: Expression of type "list[str]" cannot be assigned to declared type "List[str | None]"
TypeVar "_T" is invariant
Type "str" cannot be assigned to type "str | None" (reportGeneralTypeIssues)
```
I wasn't expecting this because `aopt1` is equivalent and doesn't have the type error | 1.0 | Type "str" cannot be assigned to type "str | None" - ```python
from typing import List, Optional
a = list(["a", "b", "c"])
aopt1: List[Optional[str]] = list(["a", "b", "c"])
aopt2: List[Optional[str]] = a
```
pyright 1.1.62 reports a type error on `aopt2`:
```
6:30 - error: Expression of type "list[str]" cannot be assigned to declared type "List[str | None]"
TypeVar "_T" is invariant
Type "str" cannot be assigned to type "str | None" (reportGeneralTypeIssues)
```
I wasn't expecting this because `aopt1` is equivalent and doesn't have the type error | non_priority | type str cannot be assigned to type str none python from typing import list optional a list list list list a pyright reports a type error on error expression of type list cannot be assigned to declared type list typevar t is invariant type str cannot be assigned to type str none reportgeneraltypeissues i wasn t expecting this because is equivalent and doesn t have the type error | 0 |
445,023 | 31,161,368,957 | IssuesEvent | 2023-08-16 16:13:50 | cooperative-computing-lab/cctools | https://api.github.com/repos/cooperative-computing-lab/cctools | opened | Vine-Parsl: Improve Docs | documentation TaskVine | We have a very brief section on Vine-Parsl integration in our manual here:
https://cctools.readthedocs.io/en/stable/taskvine/#workflow-integration
- [ ] Extend this example so that (a) it shows a large number of tasks to run and (b) explicitly sets up the taskvine config objects, so that someone can see how to deploy multiple workers on batch systems, with project names, etc.
- [ ] Update the [Parsl documentation](https://parsl.readthedocs.io/en/stable/userguide/configuring.html#ccl-notre-dame-with-work-queue) in a similar way. It currently shows how to do Parsl+WQ, but needs a similar detailed section on Parsl+TaskVine. | 1.0 | Vine-Parsl: Improve Docs - We have a very brief section on Vine-Parsl integration in our manual here:
https://cctools.readthedocs.io/en/stable/taskvine/#workflow-integration
- [ ] Extend this example so that (a) it shows a large number of tasks to run and (b) explicitly sets up the taskvine config objects, so that someone can see how to deploy multiple workers on batch systems, with project names, etc.
- [ ] Update the [Parsl documentation](https://parsl.readthedocs.io/en/stable/userguide/configuring.html#ccl-notre-dame-with-work-queue) in a similar way. It currently shows how to do Parsl+WQ, but needs a similar detailed section on Parsl+TaskVine. | non_priority | vine parsl improve docs we have a very brief section on vine parsl integration in our manual here extend this example so that a it shows a large number of tasks to run and b explicitly sets up the taskvine config objects so that someone can see how to deploy multiple workers on batch systems with project names etc update the in a similar way it currently shows how to do parsl wq but needs a similar detailed section on parsl taskvine | 0 |
125,744 | 12,268,027,136 | IssuesEvent | 2020-05-07 11:47:50 | crate/crate-docs-theme | https://api.github.com/repos/crate/crate-docs-theme | closed | testing source file addition | documentation | <!--Please do not edit or remove the following information -->
- **Page title**: Crate Docs Theme
- **Page URL**: https://crate.io/docs/fake/en/latest/index.rst
- **Source file**: https://github.com/crate/crate-docs-theme/blob/master/docs/index.rst
- **DocID**: 6a992d55
<!-- Please add your comments below -->
**Comments**:
test | 1.0 | testing source file addition - <!--Please do not edit or remove the following information -->
- **Page title**: Crate Docs Theme
- **Page URL**: https://crate.io/docs/fake/en/latest/index.rst
- **Source file**: https://github.com/crate/crate-docs-theme/blob/master/docs/index.rst
- **DocID**: 6a992d55
<!-- Please add your comments below -->
**Comments**:
test | non_priority | testing source file addition page title crate docs theme page url source file docid comments test | 0 |
20,136 | 6,822,540,577 | IssuesEvent | 2017-11-07 20:26:44 | ros2/build_cop | https://api.github.com/repos/ros2/build_cop | closed | Update java on Windows and Mac OS machines | buildfarm | Java8u144 is available and updates pop-ups show up on most machines. Looks like some windows ones updated automatically. We should update a few, test and if all goes well update the remaining ones | 1.0 | Update java on Windows and Mac OS machines - Java8u144 is available and updates pop-ups show up on most machines. Looks like some windows ones updated automatically. We should update a few, test and if all goes well update the remaining ones | non_priority | update java on windows and mac os machines is available and updates pop ups show up on most machines looks like some windows ones updated automatically we should update a few test and if all goes well update the remaining ones | 0 |
294,432 | 25,371,216,316 | IssuesEvent | 2022-11-21 10:41:47 | lowRISC/opentitan | https://api.github.com/repos/lowRISC/opentitan | opened | [test-triage] chip_csr_mem_rw_with_rand_reset timeout | Component:TestTriage | ### Hierarchy of regression failure
Chip Level
### Failure Description
```
UVM_FATAL @ 13120.065123 us: (csr_utils_pkg.sv:564) [csr_utils::csr_spinwait] timeout chip_reg_block.kmac.cfg_regwen (addr=0x41120010) == 0x0
UVM_INFO @ 13120.065123 us: (uvm_report_catcher.svh:705) [UVM/REPORT/CATCHER]
--- UVM Report catcher Summary ---
```
### Steps to Reproduce
- GitHub Revision: HASH_VALUE
- dvsim invocation command to reproduce the failure, inclusive of build and run seeds:
./util/dvsim/dvsim.py hw/top_earlgrey/dv/chip_sim_cfg.hjson -i chip_csr_mem_rw_with_rand_reset --build-seed 863917349 --fixed-seed 1477319714 --waves fsdb
### Tests with similar or related failures
- [ ] chip_csr_mem_rw_with_rand_reset
| 1.0 | [test-triage] chip_csr_mem_rw_with_rand_reset timeout - ### Hierarchy of regression failure
Chip Level
### Failure Description
```
UVM_FATAL @ 13120.065123 us: (csr_utils_pkg.sv:564) [csr_utils::csr_spinwait] timeout chip_reg_block.kmac.cfg_regwen (addr=0x41120010) == 0x0
UVM_INFO @ 13120.065123 us: (uvm_report_catcher.svh:705) [UVM/REPORT/CATCHER]
--- UVM Report catcher Summary ---
```
### Steps to Reproduce
- GitHub Revision: HASH_VALUE
- dvsim invocation command to reproduce the failure, inclusive of build and run seeds:
./util/dvsim/dvsim.py hw/top_earlgrey/dv/chip_sim_cfg.hjson -i chip_csr_mem_rw_with_rand_reset --build-seed 863917349 --fixed-seed 1477319714 --waves fsdb
### Tests with similar or related failures
- [ ] chip_csr_mem_rw_with_rand_reset
| non_priority | chip csr mem rw with rand reset timeout hierarchy of regression failure chip level failure description uvm fatal us csr utils pkg sv timeout chip reg block kmac cfg regwen addr uvm info us uvm report catcher svh uvm report catcher summary steps to reproduce github revision hash value dvsim invocation command to reproduce the failure inclusive of build and run seeds util dvsim dvsim py hw top earlgrey dv chip sim cfg hjson i chip csr mem rw with rand reset build seed fixed seed waves fsdb tests with similar or related failures chip csr mem rw with rand reset | 0 |
33,154 | 15,802,971,194 | IssuesEvent | 2021-04-03 12:17:47 | gnuless/ncc | https://api.github.com/repos/gnuless/ncc | opened | global constant propagation implementation is naive and inefficient | performance | It works properly, but its efficiency is godawful. It needs to be rewritten to avoid using associative containers during the data-flow analysis: the meet functions are quadratic at best, ick. | True | global constant propagation implementation is naive and inefficient - It works properly, but its efficiency is godawful. It needs to be rewritten to avoid using associative containers during the data-flow analysis: the meet functions are quadratic at best, ick. | non_priority | global constant propagation implementation is naive and inefficient it works properly but its efficiency is godawful it needs to be rewritten to avoid using associative containers during the data flow analysis the meet functions are quadratic at best ick | 0 |
46,112 | 24,369,059,669 | IssuesEvent | 2022-10-03 17:35:24 | statamic/cms | https://api.github.com/repos/statamic/cms | closed | Performance issues with 3.3.41 | performance | ### Bug description
I updated Statamic from 3.3.12 to 3.3.41 and experienced a major peformance drop from ~1.04s to ~2.5s (that was a random subpage of the local installation).
As I also updated 2 addons, I downgraded them again which showed no effect. Then I updated Statamic one version after the other. Up until 3.3.40 everything is fine, in fact it's a bit faster (~890ms), but 3.3.41 slows it down.
Did anyone else experience such a slowdown?
### How to reproduce
I tried the same thing on another installation and couldn't reproduce it. One difference is that the problematic installation uses S3 as a filesystem, maybe it's related to https://github.com/statamic/cms/pull/6769?
### Logs
_No response_
### Environment
```yaml
Statamic 3.3.40 Pro
Laravel 8.83.25
PHP 7.4.30
edalzell/variable 1.0.4
mia/statamic-image-renderer 1.0.20
withcandour/aardvark-seo 2.0.28
```
### Installation
Fresh statamic/statamic site via CLI
### Antlers Parser
runtime (new)
### Additional details
_No response_ | True | Performance issues with 3.3.41 - ### Bug description
I updated Statamic from 3.3.12 to 3.3.41 and experienced a major peformance drop from ~1.04s to ~2.5s (that was a random subpage of the local installation).
As I also updated 2 addons, I downgraded them again which showed no effect. Then I updated Statamic one version after the other. Up until 3.3.40 everything is fine, in fact it's a bit faster (~890ms), but 3.3.41 slows it down.
Did anyone else experience such a slowdown?
### How to reproduce
I tried the same thing on another installation and couldn't reproduce it. One difference is that the problematic installation uses S3 as a filesystem, maybe it's related to https://github.com/statamic/cms/pull/6769?
### Logs
_No response_
### Environment
```yaml
Statamic 3.3.40 Pro
Laravel 8.83.25
PHP 7.4.30
edalzell/variable 1.0.4
mia/statamic-image-renderer 1.0.20
withcandour/aardvark-seo 2.0.28
```
### Installation
Fresh statamic/statamic site via CLI
### Antlers Parser
runtime (new)
### Additional details
_No response_ | non_priority | performance issues with bug description i updated statamic from to and experienced a major peformance drop from to that was a random subpage of the local installation as i also updated addons i downgraded them again which showed no effect then i updated statamic one version after the other up until everything is fine in fact it s a bit faster but slows it down did anyone else experience such a slowdown how to reproduce i tried the same thing on another installation and couldn t reproduce it one difference is that the problematic installation uses as a filesystem maybe it s related to logs no response environment yaml statamic pro laravel php edalzell variable mia statamic image renderer withcandour aardvark seo installation fresh statamic statamic site via cli antlers parser runtime new additional details no response | 0 |
154,952 | 19,765,606,179 | IssuesEvent | 2022-01-17 01:33:39 | tuanducdesign/reactjs-mern | https://api.github.com/repos/tuanducdesign/reactjs-mern | opened | CVE-2021-3807 (High) detected in ansi-regex-4.1.0.tgz, ansi-regex-5.0.0.tgz | security vulnerability | ## CVE-2021-3807 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-4.1.0.tgz</b>, <b>ansi-regex-5.0.0.tgz</b></p></summary>
<p>
<details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/webpack-dev-server/node_modules/wrap-ansi/node_modules/ansi-regex/package.json,/client/node_modules/webpack-dev-server/node_modules/string-width/node_modules/ansi-regex/package.json,/server/node_modules/ansi-align/node_modules/ansi-regex/package.json,/client/node_modules/webpack-dev-server/node_modules/cliui/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-2.0.7.tgz (Root Library)
- update-notifier-4.1.3.tgz
- boxen-4.2.0.tgz
- ansi-align-3.0.0.tgz
- string-width-3.1.0.tgz
- strip-ansi-5.2.0.tgz
- :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>ansi-regex-5.0.0.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz</a></p>
<p>Path to dependency file: /server/package.json</p>
<p>Path to vulnerable library: /server/node_modules/boxen/node_modules/ansi-regex/package.json,/client/node_modules/ansi-regex/package.json,/server/node_modules/widest-line/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-2.0.7.tgz (Root Library)
- update-notifier-4.1.3.tgz
- boxen-4.2.0.tgz
- widest-line-3.1.0.tgz
- string-width-4.2.2.tgz
- strip-ansi-6.0.0.tgz
- :x: **ansi-regex-5.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/tuanducdesign/reactjs-mern/commit/128d5a54aa10a5b42852c9d7020b9920bc1f823c">128d5a54aa10a5b42852c9d7020b9920bc1f823c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ansi-regex is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3807 (High) detected in ansi-regex-4.1.0.tgz, ansi-regex-5.0.0.tgz - ## CVE-2021-3807 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-4.1.0.tgz</b>, <b>ansi-regex-5.0.0.tgz</b></p></summary>
<p>
<details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/webpack-dev-server/node_modules/wrap-ansi/node_modules/ansi-regex/package.json,/client/node_modules/webpack-dev-server/node_modules/string-width/node_modules/ansi-regex/package.json,/server/node_modules/ansi-align/node_modules/ansi-regex/package.json,/client/node_modules/webpack-dev-server/node_modules/cliui/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-2.0.7.tgz (Root Library)
- update-notifier-4.1.3.tgz
- boxen-4.2.0.tgz
- ansi-align-3.0.0.tgz
- string-width-3.1.0.tgz
- strip-ansi-5.2.0.tgz
- :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>ansi-regex-5.0.0.tgz</b></p></summary>
<p>Regular expression for matching ANSI escape codes</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-5.0.0.tgz</a></p>
<p>Path to dependency file: /server/package.json</p>
<p>Path to vulnerable library: /server/node_modules/boxen/node_modules/ansi-regex/package.json,/client/node_modules/ansi-regex/package.json,/server/node_modules/widest-line/node_modules/ansi-regex/package.json</p>
<p>
Dependency Hierarchy:
- nodemon-2.0.7.tgz (Root Library)
- update-notifier-4.1.3.tgz
- boxen-4.2.0.tgz
- widest-line-3.1.0.tgz
- string-width-4.2.2.tgz
- strip-ansi-6.0.0.tgz
- :x: **ansi-regex-5.0.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/tuanducdesign/reactjs-mern/commit/128d5a54aa10a5b42852c9d7020b9920bc1f823c">128d5a54aa10a5b42852c9d7020b9920bc1f823c</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ansi-regex is vulnerable to Inefficient Regular Expression Complexity
<p>Publish Date: 2021-09-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p>
<p>Release Date: 2021-09-17</p>
<p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in ansi regex tgz ansi regex tgz cve high severity vulnerability vulnerable libraries ansi regex tgz ansi regex tgz ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file client package json path to vulnerable library client node modules webpack dev server node modules wrap ansi node modules ansi regex package json client node modules webpack dev server node modules string width node modules ansi regex package json server node modules ansi align node modules ansi regex package json client node modules webpack dev server node modules cliui node modules ansi regex package json dependency hierarchy nodemon tgz root library update notifier tgz boxen tgz ansi align tgz string width tgz strip ansi tgz x ansi regex tgz vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file server package json path to vulnerable library server node modules boxen node modules ansi regex package json client node modules ansi regex package json server node modules widest line node modules ansi regex package json dependency hierarchy nodemon tgz root library update notifier tgz boxen tgz widest line tgz string width tgz strip ansi tgz x ansi regex tgz vulnerable library found in head commit a href found in base branch master vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex step up your open source security game with whitesource | 0 |
177,729 | 29,045,294,595 | IssuesEvent | 2023-05-13 13:27:37 | Leafwing-Studios/Emergence | https://api.github.com/repos/Leafwing-Studios/Emergence | opened | Add a renewable source and sink of soil | gameplay game design | Dehydrating water / dissolving soil into the water is the most promising generic solution.
Important for building up larger structures.
| 1.0 | Add a renewable source and sink of soil - Dehydrating water / dissolving soil into the water is the most promising generic solution.
Important for building up larger structures.
| non_priority | add a renewable source and sink of soil dehydrating water dissolving soil into the water is the most promising generic solution important for building up larger structures | 0 |
41,182 | 12,831,582,486 | IssuesEvent | 2020-07-07 05:46:18 | rvvergara/haiku-android | https://api.github.com/repos/rvvergara/haiku-android | closed | WS-2020-0070 (High) detected in lodash-4.17.15.tgz | security vulnerability | ## WS-2020-0070 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/haiku-android/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/haiku-android/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- core-7.8.4.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/haiku-android/commit/0560eaf1946befb898a1ac9da2b3c8e0ab81ceb4">0560eaf1946befb898a1ac9da2b3c8e0ab81ceb4</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of lodash are vulnerable to Prototype Pollution. The function zipObjectDeep allows a malicious user to modify the prototype of Object if the property identifiers are user-supplied. Being affected by this issue requires zipping objects based on user-provided property arrays. This vulnerability may lead to Denial of Service or Code Execution.
<p>Publish Date: 2020-04-28
<p>URL: <a href=https://hackerone.com/reports/712065>WS-2020-0070</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2020-0070 (High) detected in lodash-4.17.15.tgz - ## WS-2020-0070 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary>
<p>Lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/haiku-android/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/haiku-android/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- core-7.8.4.tgz (Root Library)
- :x: **lodash-4.17.15.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rvvergara/haiku-android/commit/0560eaf1946befb898a1ac9da2b3c8e0ab81ceb4">0560eaf1946befb898a1ac9da2b3c8e0ab81ceb4</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
All versions of lodash are vulnerable to Prototype Pollution. The function zipObjectDeep allows a malicious user to modify the prototype of Object if the property identifiers are user-supplied. Being affected by this issue requires zipping objects based on user-provided property arrays. This vulnerability may lead to Denial of Service or Code Execution.
<p>Publish Date: 2020-04-28
<p>URL: <a href=https://hackerone.com/reports/712065>WS-2020-0070</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws high detected in lodash tgz ws high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file tmp ws scm haiku android package json path to vulnerable library tmp ws scm haiku android node modules lodash package json dependency hierarchy core tgz root library x lodash tgz vulnerable library found in head commit a href vulnerability details all versions of lodash are vulnerable to prototype pollution the function zipobjectdeep allows a malicious user to modify the prototype of object if the property identifiers are user supplied being affected by this issue requires zipping objects based on user provided property arrays this vulnerability may lead to denial of service or code execution publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
136,564 | 30,550,432,873 | IssuesEvent | 2023-07-20 08:09:06 | jOOQ/jOOQ | https://api.github.com/repos/jOOQ/jOOQ | opened | Add support for CockroachDB 21 MATERIALIZED VIEWS | T: Enhancement C: Functionality C: Code Generation P: Medium E: Professional Edition E: Enterprise Edition C: DB: CockroachDB | Starting with CockroachDB 20.2, `MATERIALIZED VIEWS` are now supported, see:
https://www.cockroachlabs.com/docs/v20.2/views#materialized-views
We're not currently integration testing things for these, meaning, there might be bugs in the code generator or at runtime, see e.g.:
- https://github.com/jOOQ/jOOQ/issues/15414 | 1.0 | Add support for CockroachDB 21 MATERIALIZED VIEWS - Starting with CockroachDB 20.2, `MATERIALIZED VIEWS` are now supported, see:
https://www.cockroachlabs.com/docs/v20.2/views#materialized-views
We're not currently integration testing things for these, meaning, there might be bugs in the code generator or at runtime, see e.g.:
- https://github.com/jOOQ/jOOQ/issues/15414 | non_priority | add support for cockroachdb materialized views starting with cockroachdb materialized views are now supported see we re not currently integration testing things for these meaning there might be bugs in the code generator or at runtime see e g | 0 |
130,328 | 12,426,423,782 | IssuesEvent | 2020-05-24 21:06:42 | WiillyWonka/RecipesFinder | https://api.github.com/repos/WiillyWonka/RecipesFinder | closed | Подготовить речь по презентации. | documentation | Речью занимается тот человек, что представляет продукт на демо. | 1.0 | Подготовить речь по презентации. - Речью занимается тот человек, что представляет продукт на демо. | non_priority | подготовить речь по презентации речью занимается тот человек что представляет продукт на демо | 0 |
435,108 | 30,485,894,551 | IssuesEvent | 2023-07-18 02:08:17 | casdoor/casdoor-website | https://api.github.com/repos/casdoor/casdoor-website | closed | Could we have a comparison table for casdoor and keycloak? | documentation | As some ones are familiar with keycloak and learning to use casdoor, it's better to know the differences on the same features.
here is an example for comparing opa and casbin - https://gist.github.com/StevenACoffman/1644ec1157a793eb7d868aa22b260e91 | 1.0 | Could we have a comparison table for casdoor and keycloak? - As some ones are familiar with keycloak and learning to use casdoor, it's better to know the differences on the same features.
here is an example for comparing opa and casbin - https://gist.github.com/StevenACoffman/1644ec1157a793eb7d868aa22b260e91 | non_priority | could we have a comparison table for casdoor and keycloak as some ones are familiar with keycloak and learning to use casdoor it s better to know the differences on the same features here is an example for comparing opa and casbin | 0 |
29,595 | 13,136,106,891 | IssuesEvent | 2020-08-07 05:05:10 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Command markdowns have sample output data too | Pri2 container-service/svc cxp doc-enhancement triaged |
Please remove sample output data from command markdowns, this makes command copy experience useless.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6afe8c14-efa4-31a6-9fd9-e68f2d878226
* Version Independent ID: c67857a5-ee1b-3ed2-f281-dd62c4c4c6f3
* Content: [Kubernetes on Azure tutorial - Create a container registry - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/tutorial-kubernetes-prepare-acr)
* Content Source: [articles/aks/tutorial-kubernetes-prepare-acr.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/tutorial-kubernetes-prepare-acr.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | 1.0 | Command markdowns have sample output data too -
Please remove sample output data from command markdowns, this makes command copy experience useless.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 6afe8c14-efa4-31a6-9fd9-e68f2d878226
* Version Independent ID: c67857a5-ee1b-3ed2-f281-dd62c4c4c6f3
* Content: [Kubernetes on Azure tutorial - Create a container registry - Azure Kubernetes Service](https://docs.microsoft.com/en-us/azure/aks/tutorial-kubernetes-prepare-acr)
* Content Source: [articles/aks/tutorial-kubernetes-prepare-acr.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/aks/tutorial-kubernetes-prepare-acr.md)
* Service: **container-service**
* GitHub Login: @mlearned
* Microsoft Alias: **mlearned** | non_priority | command markdowns have sample output data too please remove sample output data from command markdowns this makes command copy experience useless document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service container service github login mlearned microsoft alias mlearned | 0 |
75,982 | 26,192,876,336 | IssuesEvent | 2023-01-03 10:42:21 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Bring back search room like in 1.07 | T-Defect S-Minor X-Needs-Product O-Uncommon A-New-Search-Experience | **Is your feature request related to a problem? Please describe.**
In 1.07 I can use Search field like filter of my rooms. So, I typed "element" and have seen all my rooms about Element
| 1.0 | Bring back search room like in 1.07 - **Is your feature request related to a problem? Please describe.**
In 1.07 I can use Search field like filter of my rooms. So, I typed "element" and have seen all my rooms about Element
| non_priority | bring back search room like in is your feature request related to a problem please describe in i can use search field like filter of my rooms so i typed element and have seen all my rooms about element | 0 |
324,782 | 24,016,594,293 | IssuesEvent | 2022-09-15 01:45:37 | tacosontitan/Mauve.Framework | https://api.github.com/repos/tacosontitan/Mauve.Framework | closed | Create `README.md` file. | documentation help wanted good first issue | The project needs a readme file. It can simply say "Hello World!" for now. | 1.0 | Create `README.md` file. - The project needs a readme file. It can simply say "Hello World!" for now. | non_priority | create readme md file the project needs a readme file it can simply say hello world for now | 0 |
133,067 | 18,279,016,328 | IssuesEvent | 2021-10-04 23:07:03 | occmundial/occ-atomic | https://api.github.com/repos/occmundial/occ-atomic | closed | CVE-2020-7774 (High) detected in y18n-4.0.0.tgz - autoclosed | security vulnerability | ## CVE-2020-7774 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-4.0.0.tgz</b></p></summary>
<p>the bare-bones internationalization library used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz">https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz</a></p>
<p>Path to dependency file: occ-atomic/package.json</p>
<p>Path to vulnerable library: occ-atomic/node_modules/y18n/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.7.2.tgz (Root Library)
- yargs-12.0.5.tgz
- :x: **y18n-4.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/occmundial/occ-atomic/commit/974392d6d4c4f898f918b85898ffa84da5def955">974392d6d4c4f898f918b85898ffa84da5def955</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package y18n before 3.2.2, 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true
<p>Publish Date: 2020-11-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1654">https://www.npmjs.com/advisories/1654</a></p>
<p>Release Date: 2020-11-17</p>
<p>Fix Resolution: 3.2.2, 4.0.1, 5.0.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-7774 (High) detected in y18n-4.0.0.tgz - autoclosed - ## CVE-2020-7774 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>y18n-4.0.0.tgz</b></p></summary>
<p>the bare-bones internationalization library used by yargs</p>
<p>Library home page: <a href="https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz">https://registry.npmjs.org/y18n/-/y18n-4.0.0.tgz</a></p>
<p>Path to dependency file: occ-atomic/package.json</p>
<p>Path to vulnerable library: occ-atomic/node_modules/y18n/package.json</p>
<p>
Dependency Hierarchy:
- webpack-dev-server-3.7.2.tgz (Root Library)
- yargs-12.0.5.tgz
- :x: **y18n-4.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/occmundial/occ-atomic/commit/974392d6d4c4f898f918b85898ffa84da5def955">974392d6d4c4f898f918b85898ffa84da5def955</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects the package y18n before 3.2.2, 4.0.1 and 5.0.5. PoC by po6ix: const y18n = require('y18n')(); y18n.setLocale('__proto__'); y18n.updateLocale({polluted: true}); console.log(polluted); // true
<p>Publish Date: 2020-11-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774>CVE-2020-7774</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1654">https://www.npmjs.com/advisories/1654</a></p>
<p>Release Date: 2020-11-17</p>
<p>Fix Resolution: 3.2.2, 4.0.1, 5.0.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in tgz autoclosed cve high severity vulnerability vulnerable library tgz the bare bones internationalization library used by yargs library home page a href path to dependency file occ atomic package json path to vulnerable library occ atomic node modules package json dependency hierarchy webpack dev server tgz root library yargs tgz x tgz vulnerable library found in head commit a href found in base branch main vulnerability details this affects the package before and poc by const require setlocale proto updatelocale polluted true console log polluted true publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
10,964 | 8,231,076,437 | IssuesEvent | 2018-09-07 14:50:42 | hashblock/hashblock-exchange | https://api.github.com/repos/hashblock/hashblock-exchange | opened | Enable encryption again | security | Shifting to different lib which implies:
1. Changes to config key management
2. Changes to state
| True | Enable encryption again - Shifting to different lib which implies:
1. Changes to config key management
2. Changes to state
| non_priority | enable encryption again shifting to different lib which implies changes to config key management changes to state | 0 |
130,331 | 18,065,289,024 | IssuesEvent | 2021-09-20 18:22:24 | penumbra-zone/penumbra | https://api.github.com/repos/penumbra-zone/penumbra | opened | Specify address encodings | A-shielded-crypto E-easy C-design | Followup from #61
Related reading:
* p. 112 of Zcash protocol spec
* ZIP-173 defining Bech32 encoding https://zips.z.cash/zip-0173
* ZIP-316 defining unified addresses https://zips.z.cash/zip-0316 | 1.0 | Specify address encodings - Followup from #61
Related reading:
* p. 112 of Zcash protocol spec
* ZIP-173 defining Bech32 encoding https://zips.z.cash/zip-0173
* ZIP-316 defining unified addresses https://zips.z.cash/zip-0316 | non_priority | specify address encodings followup from related reading p of zcash protocol spec zip defining encoding zip defining unified addresses | 0 |
24,592 | 7,528,147,710 | IssuesEvent | 2018-04-13 19:38:54 | iotile/coretools | https://api.github.com/repos/iotile/coretools | opened | Better error message when qemu emulator is not installed | enhancement iotile-build self-contained | Currently it just says 'file not found' and its not clear what file it was searching for | 1.0 | Better error message when qemu emulator is not installed - Currently it just says 'file not found' and its not clear what file it was searching for | non_priority | better error message when qemu emulator is not installed currently it just says file not found and its not clear what file it was searching for | 0 |
164,897 | 20,509,288,142 | IssuesEvent | 2022-03-01 03:28:10 | ZcashFoundation/zebra | https://api.github.com/repos/ZcashFoundation/zebra | closed | Clearly identify 32-bit and 64-bit times in serialization and proptests | A-consensus P-Medium :zap: C-security I-consensus 💥 I-panic I-bad-data A-network | **Scheduling**
This risk is acceptable for the stable release, but we need to fix it before we support lightwalletd.
**Is your feature request related to a problem? Please describe.**
Zcash has two serialized time fields: 32-bit seconds, and 64-bit seconds. Zebra also keeps internal 64-bit times with additional 32-bit nanosecond precision.
In general, it's hard to identify, convert, serialize and proptest these different time types correctly.
This is a follow-up to #1849.
**Describe the solution you'd like**
Make wrapper types for serialized times:
- [x] `DateTime32` (#2210)
- [ ] `DateTime64`
Tasks for each wrapper type:
- [ ] make arbitrary impls
- the arbitrary impls should replace the existing `datetime_full` and `datetime_u32` proptest strategies
- we can use `datetime_full` for `DateTime64` by removing nanoseconds
- [ ] make serialization impls
Replace `chrono::DateTime<Utc>` with the appropriate types:
- [ ] 32-bit serialized: `DateTime32` #2211
- [ ] 64-bit serialized: `DateTime64`
- [ ] internal monotonic: `std::time::Instant`
- [ ] Remove unused dependencies on `chrono`
**Describe alternatives you've considered**
It would be nice if Zcash just used 64-bit times throughout the protocol. | True | Clearly identify 32-bit and 64-bit times in serialization and proptests - **Scheduling**
This risk is acceptable for the stable release, but we need to fix it before we support lightwalletd.
**Is your feature request related to a problem? Please describe.**
Zcash has two serialized time fields: 32-bit seconds, and 64-bit seconds. Zebra also keeps internal 64-bit times with additional 32-bit nanosecond precision.
In general, it's hard to identify, convert, serialize and proptest these different time types correctly.
This is a follow-up to #1849.
**Describe the solution you'd like**
Make wrapper types for serialized times:
- [x] `DateTime32` (#2210)
- [ ] `DateTime64`
Tasks for each wrapper type:
- [ ] make arbitrary impls
- the arbitrary impls should replace the existing `datetime_full` and `datetime_u32` proptest strategies
- we can use `datetime_full` for `DateTime64` by removing nanoseconds
- [ ] make serialization impls
Replace `chrono::DateTime<Utc>` with the appropriate types:
- [ ] 32-bit serialized: `DateTime32` #2211
- [ ] 64-bit serialized: `DateTime64`
- [ ] internal monotonic: `std::time::Instant`
- [ ] Remove unused dependencies on `chrono`
**Describe alternatives you've considered**
It would be nice if Zcash just used 64-bit times throughout the protocol. | non_priority | clearly identify bit and bit times in serialization and proptests scheduling this risk is acceptable for the stable release but we need to fix it before we support lightwalletd is your feature request related to a problem please describe zcash has two serialized time fields bit seconds and bit seconds zebra also keeps internal bit times with additional bit nanosecond precision in general it s hard to identify convert serialize and proptest these different time types correctly this is a follow up to describe the solution you d like make wrapper types for serialized times tasks for each wrapper type make arbitrary impls the arbitrary impls should replace the existing datetime full and datetime proptest strategies we can use datetime full for by removing nanoseconds make serialization impls replace chrono datetime with the appropriate types bit serialized bit serialized internal monotonic std time instant remove unused dependencies on chrono describe alternatives you ve considered it would be nice if zcash just used bit times throughout the protocol | 0 |
78,937 | 10,096,392,372 | IssuesEvent | 2019-07-27 17:39:26 | processing/p5.js | https://api.github.com/repos/processing/p5.js | closed | Changing Stroke Mid-Curve Drawing Changes Entire Curve Colour | area:documentation | #### Nature of issue?
- [x] Found a bug
#### Most appropriate sub-area of p5.js?
- [x] Color
- [X] Shape
#### Which platform were you using when you encountered this?
- [X] Desktop/Laptop
<!-- If you found a bug, the following information might prove to be helpful for us. Simply remove whatever you can't determine/don't know. -->
#### Details about the bug:
I've been following the Coding Challenge videos, recreating the most interesting ones in P5; tonight I recreated the Lorenz Attractor, and I discovered that changing 'stroke' while adding curveVertex() to a shape causes the entire curve to be coloured with the stroke's argument (i.e., instead of drawing a rainbow, the whole curve changes from one solid colour to another).
- p5.js version: https://cdnjs.cloudflare.com/ajax/libs/p5.js/0.5.16/p5.js
- Web browser and version: Chrome 61.0.3163.100
- Operating System: Windows 10
- Steps to reproduce this:
The code I'm playing with is [here](https://github.com/SethGreylyn/Experiments-with-P5/blob/c938d9604319b0cd971e1656085dc51b70d1f7b6/lorenzAttractor/lorenzAttractor.js#L43), and the code in question is
` translate(width / 2, height / 2);`
` scale(5);`
` noFill();`
` beginShape(points);`
` `
` var rSeed = 50;`
` var gSeed = 100;`
` var bSeed = 198;`
` `
` points.forEach((pnt) => {`
` print(rSeed, gSeed, bSeed);`
` stroke((rSeed++)%256, (gSeed++)%256, (bSeed++)%256);`
` curveVertex(pnt.x,pnt.y);`
` });`
` endShape(); `
Changing 'curveVertex()' to 'point()' results in differently-coloured points, so the basic logic is correct. (It also matches the Java logic in the Lorenz video.) | 1.0 | Changing Stroke Mid-Curve Drawing Changes Entire Curve Colour - #### Nature of issue?
- [x] Found a bug
#### Most appropriate sub-area of p5.js?
- [x] Color
- [X] Shape
#### Which platform were you using when you encountered this?
- [X] Desktop/Laptop
<!-- If you found a bug, the following information might prove to be helpful for us. Simply remove whatever you can't determine/don't know. -->
#### Details about the bug:
I've been following the Coding Challenge videos, recreating the most interesting ones in P5; tonight I recreated the Lorenz Attractor, and I discovered that changing 'stroke' while adding curveVertex() to a shape causes the entire curve to be coloured with the stroke's argument (i.e., instead of drawing a rainbow, the whole curve changes from one solid colour to another).
- p5.js version: https://cdnjs.cloudflare.com/ajax/libs/p5.js/0.5.16/p5.js
- Web browser and version: Chrome 61.0.3163.100
- Operating System: Windows 10
- Steps to reproduce this:
The code I'm playing with is [here](https://github.com/SethGreylyn/Experiments-with-P5/blob/c938d9604319b0cd971e1656085dc51b70d1f7b6/lorenzAttractor/lorenzAttractor.js#L43), and the code in question is
` translate(width / 2, height / 2);`
` scale(5);`
` noFill();`
` beginShape(points);`
` `
` var rSeed = 50;`
` var gSeed = 100;`
` var bSeed = 198;`
` `
` points.forEach((pnt) => {`
` print(rSeed, gSeed, bSeed);`
` stroke((rSeed++)%256, (gSeed++)%256, (bSeed++)%256);`
` curveVertex(pnt.x,pnt.y);`
` });`
` endShape(); `
Changing 'curveVertex()' to 'point()' results in differently-coloured points, so the basic logic is correct. (It also matches the Java logic in the Lorenz video.) | non_priority | changing stroke mid curve drawing changes entire curve colour nature of issue found a bug most appropriate sub area of js color shape which platform were you using when you encountered this desktop laptop details about the bug i ve been following the coding challenge videos recreating the most interesting ones in tonight i recreated the lorenz attractor and i discovered that changing stroke while adding curvevertex to a shape causes the entire curve to be coloured with the stroke s argument i e instead of drawing a rainbow the whole curve changes from one solid colour to another js version web browser and version chrome operating system windows steps to reproduce this the code i m playing with is and the code in question is translate width height scale nofill beginshape points var rseed var gseed var bseed points foreach pnt print rseed gseed bseed stroke rseed gseed bseed curvevertex pnt x pnt y endshape changing curvevertex to point results in differently coloured points so the basic logic is correct it also matches the java logic in the lorenz video | 0 |
243,388 | 18,685,204,831 | IssuesEvent | 2021-11-01 11:31:18 | perun-network/perun-eth-demo | https://api.github.com/repos/perun-network/perun-eth-demo | opened | README: Close the CLI with `Ctrl+D` | documentation | It is not `Ctrl+C` any longer but `Ctrl+D`, therefore the `README` should be updated. | 1.0 | README: Close the CLI with `Ctrl+D` - It is not `Ctrl+C` any longer but `Ctrl+D`, therefore the `README` should be updated. | non_priority | readme close the cli with ctrl d it is not ctrl c any longer but ctrl d therefore the readme should be updated | 0 |
64,825 | 7,842,874,030 | IssuesEvent | 2018-06-19 02:12:31 | skycoin/skycoin-web | https://api.github.com/repos/skycoin/skycoin-web | closed | Block navigation while creating a wallet in the wizard | UI/UX design | While creating a wallet in the wizard, the `DoubleButtonComponent` with the "New" and "Load" buttons must be disabled, to prevent the user from navigating. While disabled, the `DoubleButtonComponent` should be transparent. If there is an error, the `DoubleButtonComponent` must be enabled again.
Also, while creating a wallet, the "skip" button, besides being disabled, should be invisible. | 1.0 | Block navigation while creating a wallet in the wizard - While creating a wallet in the wizard, the `DoubleButtonComponent` with the "New" and "Load" buttons must be disabled, to prevent the user from navigating. While disabled, the `DoubleButtonComponent` should be transparent. If there is an error, the `DoubleButtonComponent` must be enabled again.
Also, while creating a wallet, the "skip" button, besides being disabled, should be invisible. | non_priority | block navigation while creating a wallet in the wizard while creating a wallet in the wizard the doublebuttoncomponent with the new and load buttons must be disabled to prevent the user from navigating while disabled the doublebuttoncomponent should be transparent if there is an error the doublebuttoncomponent must be enabled again also while creating a wallet the skip button besides being disabled should be invisible | 0 |
64,140 | 7,767,838,616 | IssuesEvent | 2018-06-03 11:34:17 | mawww/kakoune | https://api.github.com/repos/mawww/kakoune | closed | [RFC] command line parsing overhaul | RFC design | Hello,
As a follow up to #2029, and in order to fix the long standing bugs we have regarding parsing, such as #1049 or #1612. I would like to clean up Kakoune's command line parsing.
# The current state
Our current command line parsing is very ad-hoc, and works as follow:
* Non quoted words support escaping of `%`, whitespaces and `;`, which must be preceded by a backslash to avoid their special meaning. `%` only has a special meaning if it appears as the first character of the word (in which case we try to parse it as a `%...{...}` string), but escaping it anywhere in the word works. backslash themselves are not escaped, so `a\b` is a word composed of three character: `a`, `\` and `b`. That means it is not possible to end a word with a backslash, as it will interpreted as escaping the whitespace that should end the word.
* single quoted strings support escaping of the `'` delimiter, similarly to non quoted words, we cannot end such a string with a backslash.
* double quoted strings support escaping the `"` delimiter and cannot end with backslash. They are then reparsed and backslash-escaped `%` are replaced with `%`, non-escaped ones are treated as a `%...{...}` string and expanded. `:echo "%{a}%{b}\%"` therefore outputs `ab%`
* %...{...} strings support backslash-escaping their delimiter if its not nestable (not `{`, `[`, `<` or `(`). Nestable delimiters are not quotable.
All those string types can only happen at the beginning of a word, a %, " or ' appearing in the middle of a word is considered literally.
The obvious solution to #1049 would be to support escaping backslash. But this lead to escaping hell. We have many regexes in rc/ files that already refers to backslash using `\\`, that means we would have to write a literal backslash for regex as `\\\\`.
# What we would want
1. We want familiar, simple and easy to write interactive commands, so we want a whitespace separated command syntax (`:command <param> <param> <param>`)
2. We want to be able to easily write regular expressions and Kakoune keys they are very common (`:add-highlighter global regex \bword\b`). This means we want a nice way to write backslash heavy strings without having to escape them.
3. We want to be able to write arbitrary strings as single parameters, we want a simple algorithm generate a string that we know is correctly escaped for any content.
* This means in particular that we cannot rely on checking for delimiters that are not in the string, as we can always imagine a string containing all possible delimiters.
4. We want to easily write nested strings, as Kakoune follows tcl model of list of commands as strings. This is pretty much solved with `%...{...}` strings.
5. We would prefer not to have to change every existing Kakoune script to fix their strings. Keeping close to existing practice would avoid having to change too much.
6. We would prefer to be close to existing well established practice in the unix world.
7. We would prefer to have a consistent quoting behaviour with the different strings.
# Possible directions
* Single quoted strings could be made to work as they do in the shell: no escaping supported, no way to have a single quote inside a single quoted string. This would solve 2. for any string not containing single quotes.
* We could make use of doubling up as a quoting way, for example, single quoted strings could contain single quotes with a `''` representing a single quote inside a single quoted strings. This would solve 2. and 3. but would violate 6. and 7.
* For double quoted strings, we could use doubling up as well, but while it would work fine for the `"` character, ~escaping % by doubling up would interact badly with the explicit reparsing syntax `%%`. The explicit reparsing syntax is not yet in master, so it could be changed to something else (what ?).~
* For single words, we could not use doubling up to escape, as there are no delimiters, and doubling up whitespaces would be confusing. We could remove support for escaping whitespaces and require use of quoting for words containing whitespace. We would still need to support escaping `%` and `;` which could be done by doubling up, we also need a way to escaped end of line, which cannot.
* Nestable `%` strings need not change, I think they proved to be pretty robust.
* Non nestable `%` string could use doubling up as well.
This means we would generally use doubling up as an escape mechanism in Kakoune command line. This would be on purpose inconsistent with the shell and regular expressions, as it would avoid bad quoting interactions between the two.
For end of line escaping (line continuation), we could make an exception and use a backslash there.
# Additional questions
* Should we try to follow the Shell and support `non_quoted="Quoted Part"` ? I think it is not necessary and breaks use of " and ' as keys.
* Should we support escaping % in non quoted words elsewhere than at the first character ?
* Should we mandate that `;` as an end of command separator be separated as its own word (require a whitespace before it ?). This would avoid needing to escape it.
* ~What syntax should we use for explicit reparsing if we want to keep doubling up for escaping ?~
What are your opinions on that ?
| 1.0 | [RFC] command line parsing overhaul - Hello,
As a follow up to #2029, and in order to fix the long standing bugs we have regarding parsing, such as #1049 or #1612. I would like to clean up Kakoune's command line parsing.
# The current state
Our current command line parsing is very ad-hoc, and works as follow:
* Non quoted words support escaping of `%`, whitespaces and `;`, which must be preceded by a backslash to avoid their special meaning. `%` only has a special meaning if it appears as the first character of the word (in which case we try to parse it as a `%...{...}` string), but escaping it anywhere in the word works. backslash themselves are not escaped, so `a\b` is a word composed of three character: `a`, `\` and `b`. That means it is not possible to end a word with a backslash, as it will interpreted as escaping the whitespace that should end the word.
* single quoted strings support escaping of the `'` delimiter, similarly to non quoted words, we cannot end such a string with a backslash.
* double quoted strings support escaping the `"` delimiter and cannot end with backslash. They are then reparsed and backslash-escaped `%` are replaced with `%`, non-escaped ones are treated as a `%...{...}` string and expanded. `:echo "%{a}%{b}\%"` therefore outputs `ab%`
* %...{...} strings support backslash-escaping their delimiter if its not nestable (not `{`, `[`, `<` or `(`). Nestable delimiters are not quotable.
All those string types can only happen at the beginning of a word, a %, " or ' appearing in the middle of a word is considered literally.
The obvious solution to #1049 would be to support escaping backslash. But this lead to escaping hell. We have many regexes in rc/ files that already refers to backslash using `\\`, that means we would have to write a literal backslash for regex as `\\\\`.
# What we would want
1. We want familiar, simple and easy to write interactive commands, so we want a whitespace separated command syntax (`:command <param> <param> <param>`)
2. We want to be able to easily write regular expressions and Kakoune keys they are very common (`:add-highlighter global regex \bword\b`). This means we want a nice way to write backslash heavy strings without having to escape them.
3. We want to be able to write arbitrary strings as single parameters, we want a simple algorithm generate a string that we know is correctly escaped for any content.
* This means in particular that we cannot rely on checking for delimiters that are not in the string, as we can always imagine a string containing all possible delimiters.
4. We want to easily write nested strings, as Kakoune follows tcl model of list of commands as strings. This is pretty much solved with `%...{...}` strings.
5. We would prefer not to have to change every existing Kakoune script to fix their strings. Keeping close to existing practice would avoid having to change too much.
6. We would prefer to be close to existing well established practice in the unix world.
7. We would prefer to have a consistent quoting behaviour with the different strings.
# Possible directions
* Single quoted strings could be made to work as they do in the shell: no escaping supported, no way to have a single quote inside a single quoted string. This would solve 2. for any string not containing single quotes.
* We could make use of doubling up as a quoting way, for example, single quoted strings could contain single quotes with a `''` representing a single quote inside a single quoted strings. This would solve 2. and 3. but would violate 6. and 7.
* For double quoted strings, we could use doubling up as well, but while it would work fine for the `"` character, ~escaping % by doubling up would interact badly with the explicit reparsing syntax `%%`. The explicit reparsing syntax is not yet in master, so it could be changed to something else (what ?).~
* For single words, we could not use doubling up to escape, as there are no delimiters, and doubling up whitespaces would be confusing. We could remove support for escaping whitespaces and require use of quoting for words containing whitespace. We would still need to support escaping `%` and `;` which could be done by doubling up, we also need a way to escaped end of line, which cannot.
* Nestable `%` strings need not change, I think they proved to be pretty robust.
* Non nestable `%` string could use doubling up as well.
This means we would generally use doubling up as an escape mechanism in Kakoune command line. This would be on purpose inconsistent with the shell and regular expressions, as it would avoid bad quoting interactions between the two.
For end of line escaping (line continuation), we could make an exception and use a backslash there.
# Additional questions
* Should we try to follow the Shell and support `non_quoted="Quoted Part"` ? I think it is not necessary and breaks use of " and ' as keys.
* Should we support escaping % in non quoted words elsewhere than at the first character ?
* Should we mandate that `;` as an end of command separator be separated as its own word (require a whitespace before it ?). This would avoid needing to escape it.
* ~What syntax should we use for explicit reparsing if we want to keep doubling up for escaping ?~
What are your opinions on that ?
| non_priority | command line parsing overhaul hello as a follow up to and in order to fix the long standing bugs we have regarding parsing such as or i would like to clean up kakoune s command line parsing the current state our current command line parsing is very ad hoc and works as follow non quoted words support escaping of whitespaces and which must be preceded by a backslash to avoid their special meaning only has a special meaning if it appears as the first character of the word in which case we try to parse it as a string but escaping it anywhere in the word works backslash themselves are not escaped so a b is a word composed of three character a and b that means it is not possible to end a word with a backslash as it will interpreted as escaping the whitespace that should end the word single quoted strings support escaping of the delimiter similarly to non quoted words we cannot end such a string with a backslash double quoted strings support escaping the delimiter and cannot end with backslash they are then reparsed and backslash escaped are replaced with non escaped ones are treated as a string and expanded echo a b therefore outputs ab strings support backslash escaping their delimiter if its not nestable not or nestable delimiters are not quotable all those string types can only happen at the beginning of a word a or appearing in the middle of a word is considered literally the obvious solution to would be to support escaping backslash but this lead to escaping hell we have many regexes in rc files that already refers to backslash using that means we would have to write a literal backslash for regex as what we would want we want familiar simple and easy to write interactive commands so we want a whitespace separated command syntax command we want to be able to easily write regular expressions and kakoune keys they are very common add highlighter global regex bword b this means we want a nice way to write backslash heavy strings without having to escape them we want to be able to write arbitrary strings as single parameters we want a simple algorithm generate a string that we know is correctly escaped for any content this means in particular that we cannot rely on checking for delimiters that are not in the string as we can always imagine a string containing all possible delimiters we want to easily write nested strings as kakoune follows tcl model of list of commands as strings this is pretty much solved with strings we would prefer not to have to change every existing kakoune script to fix their strings keeping close to existing practice would avoid having to change too much we would prefer to be close to existing well established practice in the unix world we would prefer to have a consistent quoting behaviour with the different strings possible directions single quoted strings could be made to work as they do in the shell no escaping supported no way to have a single quote inside a single quoted string this would solve for any string not containing single quotes we could make use of doubling up as a quoting way for example single quoted strings could contain single quotes with a representing a single quote inside a single quoted strings this would solve and but would violate and for double quoted strings we could use doubling up as well but while it would work fine for the character escaping by doubling up would interact badly with the explicit reparsing syntax the explicit reparsing syntax is not yet in master so it could be changed to something else what for single words we could not use doubling up to escape as there are no delimiters and doubling up whitespaces would be confusing we could remove support for escaping whitespaces and require use of quoting for words containing whitespace we would still need to support escaping and which could be done by doubling up we also need a way to escaped end of line which cannot nestable strings need not change i think they proved to be pretty robust non nestable string could use doubling up as well this means we would generally use doubling up as an escape mechanism in kakoune command line this would be on purpose inconsistent with the shell and regular expressions as it would avoid bad quoting interactions between the two for end of line escaping line continuation we could make an exception and use a backslash there additional questions should we try to follow the shell and support non quoted quoted part i think it is not necessary and breaks use of and as keys should we support escaping in non quoted words elsewhere than at the first character should we mandate that as an end of command separator be separated as its own word require a whitespace before it this would avoid needing to escape it what syntax should we use for explicit reparsing if we want to keep doubling up for escaping what are your opinions on that | 0 |
30,214 | 11,801,270,821 | IssuesEvent | 2020-03-18 19:07:07 | jgeraigery/blueocean-environments | https://api.github.com/repos/jgeraigery/blueocean-environments | opened | WS-2019-0336 (Medium) detected in react-0.8.0.tgz | security vulnerability | ## WS-2019-0336 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>react-0.8.0.tgz</b></p></summary>
<p>An npm package to get you immediate access to [React](http://facebook.github.io/react/), without also requiring the JSX transformer. This is especially useful for cases where you want to [`browserify`](https://github.com/substack/node-browserify) your mod</p>
<p>Library home page: <a href="https://registry.npmjs.org/react/-/react-0.8.0.tgz">https://registry.npmjs.org/react/-/react-0.8.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/blueocean-environments/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/blueocean-environments/node_modules/react-test-utils/node_modules/react/package.json</p>
<p>
Dependency Hierarchy:
- react-test-utils-0.0.1.tgz (Root Library)
- :x: **react-0.8.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/blueocean-environments/commit/906df1e2c2b1353a7f809ef51960f8980d6cec13">906df1e2c2b1353a7f809ef51960f8980d6cec13</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-Site Scripting vulnerability found in react before 0.14.0. The package's createElement function fails to properly validate its input object, allowing attackers to execute arbitrary JavaScript in a victim's browser.
<p>Publish Date: 2019-12-17
<p>URL: <a href=http://danlec.com/blog/xss-via-a-spoofed-react-element>WS-2019-0336</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://danlec.com/blog/xss-via-a-spoofed-react-element">http://danlec.com/blog/xss-via-a-spoofed-react-element</a></p>
<p>Release Date: 2019-12-17</p>
<p>Fix Resolution: react - 0.14.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"react","packageVersion":"0.8.0","isTransitiveDependency":true,"dependencyTree":"react-test-utils:0.0.1;react:0.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"react - 0.14.0"}],"vulnerabilityIdentifier":"WS-2019-0336","vulnerabilityDetails":"Cross-Site Scripting vulnerability found in react before 0.14.0. The package\u0027s createElement function fails to properly validate its input object, allowing attackers to execute arbitrary JavaScript in a victim\u0027s browser.","vulnerabilityUrl":"http://danlec.com/blog/xss-via-a-spoofed-react-element","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> --> | True | WS-2019-0336 (Medium) detected in react-0.8.0.tgz - ## WS-2019-0336 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>react-0.8.0.tgz</b></p></summary>
<p>An npm package to get you immediate access to [React](http://facebook.github.io/react/), without also requiring the JSX transformer. This is especially useful for cases where you want to [`browserify`](https://github.com/substack/node-browserify) your mod</p>
<p>Library home page: <a href="https://registry.npmjs.org/react/-/react-0.8.0.tgz">https://registry.npmjs.org/react/-/react-0.8.0.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/blueocean-environments/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/blueocean-environments/node_modules/react-test-utils/node_modules/react/package.json</p>
<p>
Dependency Hierarchy:
- react-test-utils-0.0.1.tgz (Root Library)
- :x: **react-0.8.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/blueocean-environments/commit/906df1e2c2b1353a7f809ef51960f8980d6cec13">906df1e2c2b1353a7f809ef51960f8980d6cec13</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Cross-Site Scripting vulnerability found in react before 0.14.0. The package's createElement function fails to properly validate its input object, allowing attackers to execute arbitrary JavaScript in a victim's browser.
<p>Publish Date: 2019-12-17
<p>URL: <a href=http://danlec.com/blog/xss-via-a-spoofed-react-element>WS-2019-0336</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://danlec.com/blog/xss-via-a-spoofed-react-element">http://danlec.com/blog/xss-via-a-spoofed-react-element</a></p>
<p>Release Date: 2019-12-17</p>
<p>Fix Resolution: react - 0.14.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"react","packageVersion":"0.8.0","isTransitiveDependency":true,"dependencyTree":"react-test-utils:0.0.1;react:0.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"react - 0.14.0"}],"vulnerabilityIdentifier":"WS-2019-0336","vulnerabilityDetails":"Cross-Site Scripting vulnerability found in react before 0.14.0. The package\u0027s createElement function fails to properly validate its input object, allowing attackers to execute arbitrary JavaScript in a victim\u0027s browser.","vulnerabilityUrl":"http://danlec.com/blog/xss-via-a-spoofed-react-element","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> --> | non_priority | ws medium detected in react tgz ws medium severity vulnerability vulnerable library react tgz an npm package to get you immediate access to without also requiring the jsx transformer this is especially useful for cases where you want to your mod library home page a href path to dependency file tmp ws scm blueocean environments package json path to vulnerable library tmp ws scm blueocean environments node modules react test utils node modules react package json dependency hierarchy react test utils tgz root library x react tgz vulnerable library found in head commit a href vulnerability details cross site scripting vulnerability found in react before the package s createelement function fails to properly validate its input object allowing attackers to execute arbitrary javascript in a victim s browser publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution react isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier ws vulnerabilitydetails cross site scripting vulnerability found in react before the package createelement function fails to properly validate its input object allowing attackers to execute arbitrary javascript in a victim browser vulnerabilityurl | 0 |
70,811 | 7,201,197,935 | IssuesEvent | 2018-02-05 21:42:20 | QubesOS/updates-status | https://api.github.com/repos/QubesOS/updates-status | closed | installer-qubes-os v25.20.9-9-anaconda (r4.0) | r4.0-dom0-cur-test | Update of installer-qubes-os to v25.20.9-9-anaconda for Qubes r4.0, see comments below for details.
Built from: https://github.com/QubesOS/qubes-installer-qubes-os/commit/3c71ef6c1e4d4f8442962eeab4585158d6f640c2
[Changes since previous version](https://github.com/QubesOS/qubes-installer-qubes-os/compare/v25.20.9-8-anaconda...v25.20.9-9-anaconda):
QubesOS/qubes-installer-qubes-os@3c71ef6 qubes-anaconda-addon 4.0.4, anaconda 25.20.9-9
QubesOS/qubes-installer-qubes-os@d2c98e2 Revert "qubes-anaconda-addon: initialize default-fw-netvm"
QubesOS/qubes-installer-qubes-os@ad15922 conf, qubes-anaconda-addon: switch templates to fedora-26 and debian-9
QubesOS/qubes-installer-qubes-os@3b4a02c anaconda: fix default scheme in custom partitioning
Referenced issues:
QubesOS/qubes-issues#3225
If you're release manager, you can issue GPG-inline signed command:
* `Upload installer-qubes-os 3c71ef6c1e4d4f8442962eeab4585158d6f640c2 r4.0 current repo` (available 7 days from now)
* `Upload installer-qubes-os 3c71ef6c1e4d4f8442962eeab4585158d6f640c2 r4.0 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload installer-qubes-os 3c71ef6c1e4d4f8442962eeab4585158d6f640c2 r4.0 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
| 1.0 | installer-qubes-os v25.20.9-9-anaconda (r4.0) - Update of installer-qubes-os to v25.20.9-9-anaconda for Qubes r4.0, see comments below for details.
Built from: https://github.com/QubesOS/qubes-installer-qubes-os/commit/3c71ef6c1e4d4f8442962eeab4585158d6f640c2
[Changes since previous version](https://github.com/QubesOS/qubes-installer-qubes-os/compare/v25.20.9-8-anaconda...v25.20.9-9-anaconda):
QubesOS/qubes-installer-qubes-os@3c71ef6 qubes-anaconda-addon 4.0.4, anaconda 25.20.9-9
QubesOS/qubes-installer-qubes-os@d2c98e2 Revert "qubes-anaconda-addon: initialize default-fw-netvm"
QubesOS/qubes-installer-qubes-os@ad15922 conf, qubes-anaconda-addon: switch templates to fedora-26 and debian-9
QubesOS/qubes-installer-qubes-os@3b4a02c anaconda: fix default scheme in custom partitioning
Referenced issues:
QubesOS/qubes-issues#3225
If you're release manager, you can issue GPG-inline signed command:
* `Upload installer-qubes-os 3c71ef6c1e4d4f8442962eeab4585158d6f640c2 r4.0 current repo` (available 7 days from now)
* `Upload installer-qubes-os 3c71ef6c1e4d4f8442962eeab4585158d6f640c2 r4.0 current (dists) repo`, you can choose subset of distributions, like `vm-fc24 vm-fc25` (available 7 days from now)
* `Upload installer-qubes-os 3c71ef6c1e4d4f8442962eeab4585158d6f640c2 r4.0 security-testing repo`
Above commands will work only if packages in current-testing repository were built from given commit (i.e. no new version superseded it).
| non_priority | installer qubes os anaconda update of installer qubes os to anaconda for qubes see comments below for details built from qubesos qubes installer qubes os qubes anaconda addon anaconda qubesos qubes installer qubes os revert qubes anaconda addon initialize default fw netvm qubesos qubes installer qubes os conf qubes anaconda addon switch templates to fedora and debian qubesos qubes installer qubes os anaconda fix default scheme in custom partitioning referenced issues qubesos qubes issues if you re release manager you can issue gpg inline signed command upload installer qubes os current repo available days from now upload installer qubes os current dists repo you can choose subset of distributions like vm vm available days from now upload installer qubes os security testing repo above commands will work only if packages in current testing repository were built from given commit i e no new version superseded it | 0 |
38,237 | 5,169,959,185 | IssuesEvent | 2017-01-18 03:22:37 | bitshares/bitshares-core | https://api.github.com/repos/bitshares/bitshares-core | closed | Implement unit test for whitelist_markets and blacklist_markets | needs testing question | _From @theoreticalbts on July 22, 2015 15:23_
The `whitelist_markets` and `blacklist_markets` are properties on an asset. But first we must ask, what are the intended semantics of these fields?
Assigning to @bytemaster to answer this question.
_Copied from original issue: cryptonomex/graphene#186_ | 1.0 | Implement unit test for whitelist_markets and blacklist_markets - _From @theoreticalbts on July 22, 2015 15:23_
The `whitelist_markets` and `blacklist_markets` are properties on an asset. But first we must ask, what are the intended semantics of these fields?
Assigning to @bytemaster to answer this question.
_Copied from original issue: cryptonomex/graphene#186_ | non_priority | implement unit test for whitelist markets and blacklist markets from theoreticalbts on july the whitelist markets and blacklist markets are properties on an asset but first we must ask what are the intended semantics of these fields assigning to bytemaster to answer this question copied from original issue cryptonomex graphene | 0 |
38,496 | 8,489,542,569 | IssuesEvent | 2018-10-26 20:17:11 | phetsims/john-travoltage | https://api.github.com/repos/phetsims/john-travoltage | closed | Getting the audio context | dev:code-review dev:sonification | Currently, this code is essentially duplicated in a few places:
```js
var audioContext = new ( window.AudioContext || window.webkitAudioContext )();
```
It would be good to have this available somewhere global that could be accessed without the duplicated code. Scenery's Features.js would potentially be appropriate.
Additionally, PHET_CORE/detectPrefix would generally be recommended for grabbing prefixed objects, e.g.:
```
var AudioContext = detectPrefix( window, 'AudioContext' );
var audioContext = new AudioContext();
```
Code review for https://github.com/phetsims/john-travoltage/issues/128 | 1.0 | Getting the audio context - Currently, this code is essentially duplicated in a few places:
```js
var audioContext = new ( window.AudioContext || window.webkitAudioContext )();
```
It would be good to have this available somewhere global that could be accessed without the duplicated code. Scenery's Features.js would potentially be appropriate.
Additionally, PHET_CORE/detectPrefix would generally be recommended for grabbing prefixed objects, e.g.:
```
var AudioContext = detectPrefix( window, 'AudioContext' );
var audioContext = new AudioContext();
```
Code review for https://github.com/phetsims/john-travoltage/issues/128 | non_priority | getting the audio context currently this code is essentially duplicated in a few places js var audiocontext new window audiocontext window webkitaudiocontext it would be good to have this available somewhere global that could be accessed without the duplicated code scenery s features js would potentially be appropriate additionally phet core detectprefix would generally be recommended for grabbing prefixed objects e g var audiocontext detectprefix window audiocontext var audiocontext new audiocontext code review for | 0 |
63,261 | 15,529,693,271 | IssuesEvent | 2021-03-13 16:11:14 | NixOS/nixpkgs | https://api.github.com/repos/NixOS/nixpkgs | closed | Python: determinism of the interpreters and bytecode (pyc) files | 6.topic: python 6.topic: reproducible builds | ## Issue description
Bytecode is created during builds of the interpreter and packages. The bytecode records a timestamp, which we cannot set. We could patch the interpreter, but we do want it to use the actual timestamp when used outside of Nix.
A fix for this issue is available in https://github.com/NixOS/nixpkgs/pull/2281 where a `useFakeTime` option is added to the generic builder. | 1.0 | Python: determinism of the interpreters and bytecode (pyc) files - ## Issue description
Bytecode is created during builds of the interpreter and packages. The bytecode records a timestamp, which we cannot set. We could patch the interpreter, but we do want it to use the actual timestamp when used outside of Nix.
A fix for this issue is available in https://github.com/NixOS/nixpkgs/pull/2281 where a `useFakeTime` option is added to the generic builder. | non_priority | python determinism of the interpreters and bytecode pyc files issue description bytecode is created during builds of the interpreter and packages the bytecode records a timestamp which we cannot set we could patch the interpreter but we do want it to use the actual timestamp when used outside of nix a fix for this issue is available in where a usefaketime option is added to the generic builder | 0 |
1,230 | 2,535,454,465 | IssuesEvent | 2015-01-26 01:06:18 | pydata/pandas | https://api.github.com/repos/pydata/pandas | closed | API: bool dtype casts to object arrays with np.array | API Design Dtypes | Bug? Did this used to work?
```
np.array(pandas.DataFrame.from_dict(dict(a=np.random.rand(50), b=np.random.randint(2, size=50).astype(bool))))
```
We would expect that this casts to a plain float ndarray.
See https://github.com/statsmodels/statsmodels/issues/880 | 1.0 | API: bool dtype casts to object arrays with np.array - Bug? Did this used to work?
```
np.array(pandas.DataFrame.from_dict(dict(a=np.random.rand(50), b=np.random.randint(2, size=50).astype(bool))))
```
We would expect that this casts to a plain float ndarray.
See https://github.com/statsmodels/statsmodels/issues/880 | non_priority | api bool dtype casts to object arrays with np array bug did this used to work np array pandas dataframe from dict dict a np random rand b np random randint size astype bool we would expect that this casts to a plain float ndarray see | 0 |
35,028 | 12,308,660,238 | IssuesEvent | 2020-05-12 07:38:27 | benchabot/joplin | https://api.github.com/repos/benchabot/joplin | opened | CVE-2020-8124 (Medium) detected in url-parse-1.4.4.tgz | security vulnerability | ## CVE-2020-8124 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.4.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.4.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.4.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/joplin/ElectronClient/package.json</p>
<p>Path to vulnerable library: /joplin/ElectronClient/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- :x: **url-parse-1.4.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/benchabot/joplin/commit/c73e3ff9ac9ce99322e28b08589d0cde405067a8">c73e3ff9ac9ce99322e28b08589d0cde405067a8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Insufficient validation and sanitization of user input exists in url-parse npm package version 1.4.4 and earlier may allow attacker to bypass security checks.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8124>CVE-2020-8124</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8124">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8124</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: url-parse - 1.4.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-8124 (Medium) detected in url-parse-1.4.4.tgz - ## CVE-2020-8124 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.4.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.4.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.4.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/joplin/ElectronClient/package.json</p>
<p>Path to vulnerable library: /joplin/ElectronClient/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- :x: **url-parse-1.4.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/benchabot/joplin/commit/c73e3ff9ac9ce99322e28b08589d0cde405067a8">c73e3ff9ac9ce99322e28b08589d0cde405067a8</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Insufficient validation and sanitization of user input exists in url-parse npm package version 1.4.4 and earlier may allow attacker to bypass security checks.
<p>Publish Date: 2020-02-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8124>CVE-2020-8124</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8124">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-8124</a></p>
<p>Release Date: 2020-02-04</p>
<p>Fix Resolution: url-parse - 1.4.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file tmp ws scm joplin electronclient package json path to vulnerable library joplin electronclient node modules url parse package json dependency hierarchy x url parse tgz vulnerable library found in head commit a href vulnerability details insufficient validation and sanitization of user input exists in url parse npm package version and earlier may allow attacker to bypass security checks publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource | 0 |
81,233 | 10,113,554,243 | IssuesEvent | 2019-07-30 17:00:25 | material-components/material-components-ios | https://api.github.com/repos/material-components/material-components-ios | closed | [FeatureHighlight] Should body be renamed message? | [FeatureHighlight] skill:API design skill:Client migrations type:API change | To be consistent with MDCAlertController & UIKit, should we rename body* properties to message* properties?
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/117179016](http://b/117179016) | 1.0 | [FeatureHighlight] Should body be renamed message? - To be consistent with MDCAlertController & UIKit, should we rename body* properties to message* properties?
<!-- Auto-generated content below, do not modify -->
---
#### Internal data
- Associated internal bug: [b/117179016](http://b/117179016) | non_priority | should body be renamed message to be consistent with mdcalertcontroller uikit should we rename body properties to message properties internal data associated internal bug | 0 |
259,252 | 27,621,778,012 | IssuesEvent | 2023-03-10 01:10:57 | artsking/linux-4.1.15 | https://api.github.com/repos/artsking/linux-4.1.15 | opened | CVE-2023-1074 (Medium) detected in linux-stable-rtv4.1.33 | Mend: dependency security vulnerability | ## CVE-2023-1074 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry.
This page will reflect the classification results once they are available through NVD
<p>Publish Date: 2023-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1074>CVE-2023-1074</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1074">https://www.linuxkernelcves.com/cves/CVE-2023-1074</a></p>
<p>Release Date: 2023-02-28</p>
<p>Fix Resolution: v4.14.305,v4.19.272,v5.4.231,v5.10.166,v5.15.91,v6.1.9,v6.2-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-1074 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2023-1074 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/artsking/linux-4.1.15/commit/b1c15f7dc4cfe553aeed8332e46f285ee92b5756">b1c15f7dc4cfe553aeed8332e46f285ee92b5756</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/sctp/bind_addr.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This CVE was marked as RESERVED by NVD. This means that the CVE-ID is reserved for future use by the CVE Numbering Authority (CNA) or a security researcher. Still, the CVE details were not included in the CVE entry. This is because the original requester of the CVE ID assignment has not sent an update to MITRE with the information needed to populate the CVE entry.
This page will reflect the classification results once they are available through NVD
<p>Publish Date: 2023-02-28
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-1074>CVE-2023-1074</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-1074">https://www.linuxkernelcves.com/cves/CVE-2023-1074</a></p>
<p>Release Date: 2023-02-28</p>
<p>Fix Resolution: v4.14.305,v4.19.272,v5.4.231,v5.10.166,v5.15.91,v6.1.9,v6.2-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files net sctp bind addr c net sctp bind addr c vulnerability details this cve was marked as reserved by nvd this means that the cve id is reserved for future use by the cve numbering authority cna or a security researcher still the cve details were not included in the cve entry this is because the original requester of the cve id assignment has not sent an update to mitre with the information needed to populate the cve entry this page will reflect the classification results once they are available through nvd publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
160,867 | 12,520,854,296 | IssuesEvent | 2020-06-03 16:31:30 | aliasrobotics/RVD | https://api.github.com/repos/aliasrobotics/RVD | opened | Possible binding to all interfaces, ./src/ros/roslib/src/roslib/network.py:241 | bandit bug components software robot component: ROS static analysis testing triage version: melodic | ```yaml
{
"id": 1,
"title": "Possible binding to all interfaces, ./src/ros/roslib/src/roslib/network.py:241",
"type": "bug",
"description": "MEDIUM confidence of MEDIUM severity bug. Possible binding to all interfaces. ./src/ros/roslib/src/roslib/network.py:241. See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug",
"version: melodic",
"robot component: ROS",
"components software"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-06-03 (16:31)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-06-03 (16:31)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "./src/ros/roslib/src/roslib/network.py:241",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
``` | 1.0 | Possible binding to all interfaces, ./src/ros/roslib/src/roslib/network.py:241 - ```yaml
{
"id": 1,
"title": "Possible binding to all interfaces, ./src/ros/roslib/src/roslib/network.py:241",
"type": "bug",
"description": "MEDIUM confidence of MEDIUM severity bug. Possible binding to all interfaces. ./src/ros/roslib/src/roslib/network.py:241. See links for more info on the bug.",
"cwe": "None",
"cve": "None",
"keywords": [
"bandit",
"bug",
"static analysis",
"testing",
"triage",
"bug",
"version: melodic",
"robot component: ROS",
"components software"
],
"system": "",
"vendor": null,
"severity": {
"rvss-score": 0,
"rvss-vector": "",
"severity-description": "",
"cvss-score": 0,
"cvss-vector": ""
},
"links": "",
"flaw": {
"phase": "testing",
"specificity": "subject-specific",
"architectural-location": "application-specific",
"application": "N/A",
"subsystem": "N/A",
"package": "N/A",
"languages": "None",
"date-detected": "2020-06-03 (16:31)",
"detected-by": "Alias Robotics",
"detected-by-method": "testing static",
"date-reported": "2020-06-03 (16:31)",
"reported-by": "Alias Robotics",
"reported-by-relationship": "automatic",
"issue": "",
"reproducibility": "always",
"trace": "./src/ros/roslib/src/roslib/network.py:241",
"reproduction": "See artifacts below (if available)",
"reproduction-image": ""
},
"exploitation": {
"description": "",
"exploitation-image": "",
"exploitation-vector": ""
},
"mitigation": {
"description": "",
"pull-request": "",
"date-mitigation": ""
}
}
``` | non_priority | possible binding to all interfaces src ros roslib src roslib network py yaml id title possible binding to all interfaces src ros roslib src roslib network py type bug description medium confidence of medium severity bug possible binding to all interfaces src ros roslib src roslib network py see links for more info on the bug cwe none cve none keywords bandit bug static analysis testing triage bug version melodic robot component ros components software system vendor null severity rvss score rvss vector severity description cvss score cvss vector links flaw phase testing specificity subject specific architectural location application specific application n a subsystem n a package n a languages none date detected detected by alias robotics detected by method testing static date reported reported by alias robotics reported by relationship automatic issue reproducibility always trace src ros roslib src roslib network py reproduction see artifacts below if available reproduction image exploitation description exploitation image exploitation vector mitigation description pull request date mitigation | 0 |
171,264 | 27,089,864,246 | IssuesEvent | 2023-02-14 20:03:35 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | opened | Add aspect ratio support to image block placeholder state | [Type] Enhancement [Block] Image [Feature] Design Tools | The `core/image` block does not currently allow for dimensions or aspect-ratio (https://github.com/WordPress/gutenberg/issues/38990), before an image is added to the block.
I propose we remove the need to set an image in place, before assigning any dimensions/aspect ratio, to further simplify designing within the editor. One shouldn't be required to have an image in mind before laying out the contents of a page.
Note that this issue is blocked by first adding support for aspect ratio within the Image block (https://github.com/WordPress/gutenberg/issues/38990).
| 1.0 | Add aspect ratio support to image block placeholder state - The `core/image` block does not currently allow for dimensions or aspect-ratio (https://github.com/WordPress/gutenberg/issues/38990), before an image is added to the block.
I propose we remove the need to set an image in place, before assigning any dimensions/aspect ratio, to further simplify designing within the editor. One shouldn't be required to have an image in mind before laying out the contents of a page.
Note that this issue is blocked by first adding support for aspect ratio within the Image block (https://github.com/WordPress/gutenberg/issues/38990).
| non_priority | add aspect ratio support to image block placeholder state the core image block does not currently allow for dimensions or aspect ratio before an image is added to the block i propose we remove the need to set an image in place before assigning any dimensions aspect ratio to further simplify designing within the editor one shouldn t be required to have an image in mind before laying out the contents of a page note that this issue is blocked by first adding support for aspect ratio within the image block | 0 |
341,943 | 30,606,989,788 | IssuesEvent | 2023-07-23 05:49:15 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix jax_numpy_creation.test_jax_full | JAX Frontend Sub Task Failing Test | | | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix jax_numpy_creation.test_jax_full - | | |
|---|---|
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5524115179/jobs/10075966997"><img src=https://img.shields.io/badge/-success-success></a>
| non_priority | fix jax numpy creation test jax full jax a href src numpy a href src tensorflow a href src torch a href src paddle a href src | 0 |
193,666 | 22,216,253,658 | IssuesEvent | 2022-06-08 02:11:25 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | reopened | CVE-2016-2549 (Medium) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2016-2549 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
sound/core/hrtimer.c in the Linux kernel before 4.4.1 does not prevent recursive callback access, which allows local users to cause a denial of service (deadlock) via a crafted ioctl call.
<p>Publish Date: 2016-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2549>CVE-2016-2549</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2549">https://nvd.nist.gov/vuln/detail/CVE-2016-2549</a></p>
<p>Release Date: 2016-04-27</p>
<p>Fix Resolution: 4.4.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2016-2549 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2016-2549 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
sound/core/hrtimer.c in the Linux kernel before 4.4.1 does not prevent recursive callback access, which allows local users to cause a denial of service (deadlock) via a crafted ioctl call.
<p>Publish Date: 2016-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2549>CVE-2016-2549</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2016-2549">https://nvd.nist.gov/vuln/detail/CVE-2016-2549</a></p>
<p>Release Date: 2016-04-27</p>
<p>Fix Resolution: 4.4.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details sound core hrtimer c in the linux kernel before does not prevent recursive callback access which allows local users to cause a denial of service deadlock via a crafted ioctl call publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
382,645 | 26,508,224,356 | IssuesEvent | 2023-01-18 15:16:49 | girlscript/winter-of-contributing | https://api.github.com/repos/girlscript/winter-of-contributing | closed | JS: Error-Handling | documentation GWOC21 JavaScript Video Audio Assigned | ### Description
Discussion about error-handling
- What is a try-catch block?
- How to use it to deal with exceptions in code?
- Example
Directory:
### Domain
JavaScript
### Type of Contribution
- [x] Documentation
- [ ] Video
- [ ] Audio
### Code of Conduct
- [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project. | 1.0 | JS: Error-Handling - ### Description
Discussion about error-handling
- What is a try-catch block?
- How to use it to deal with exceptions in code?
- Example
Directory:
### Domain
JavaScript
### Type of Contribution
- [x] Documentation
- [ ] Video
- [ ] Audio
### Code of Conduct
- [X] I follow [Contributing Guidelines](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CONTRIBUTING.md) & [Code of conduct](https://github.com/girlscript/winter-of-contributing/blob/main/.github/CODE_OF_CONDUCT.md) of this project. | non_priority | js error handling description discussion about error handling what is a try catch block how to use it to deal with exceptions in code example directory domain javascript type of contribution documentation video audio code of conduct i follow of this project | 0 |
347,471 | 24,888,591,294 | IssuesEvent | 2022-10-28 09:54:25 | isanidiot/ped | https://api.github.com/repos/isanidiot/ped | opened | Inaccurate examples in UG | type.DocumentationBug severity.Low | Some examples given in UG doesn't follow the correct format.


<!--session: 1666945844337-67510e57-6d90-4772-9916-3787fce01316-->
<!--Version: Web v3.4.4--> | 1.0 | Inaccurate examples in UG - Some examples given in UG doesn't follow the correct format.


<!--session: 1666945844337-67510e57-6d90-4772-9916-3787fce01316-->
<!--Version: Web v3.4.4--> | non_priority | inaccurate examples in ug some examples given in ug doesn t follow the correct format | 0 |
78,890 | 10,095,527,028 | IssuesEvent | 2019-07-27 09:43:43 | zkovari/gradle-changelog-automation-plugin | https://api.github.com/repos/zkovari/gradle-changelog-automation-plugin | opened | Provide contribution guides | contributing documentation | Rather put it on readthedocs, while CONTRIBUTING.md could just refer it. | 1.0 | Provide contribution guides - Rather put it on readthedocs, while CONTRIBUTING.md could just refer it. | non_priority | provide contribution guides rather put it on readthedocs while contributing md could just refer it | 0 |
224,237 | 17,169,608,077 | IssuesEvent | 2021-07-15 01:02:52 | FirebaseExtended/flutterfire | https://api.github.com/repos/FirebaseExtended/flutterfire | closed | Firebase Crashlytics 'getting started' guide omits platform specific required steps. | Stale blocked: customer-response good first issue in triage type: documentation | If you following the Flutter Firebase Crashlytics getting started guide [here](https://firebase.flutter.dev/docs/crashlytics/usage/), you will end up with a runtime crash on android.
The guide fails to mention android specific setup steps in [this guide](https://firebase.google.com/docs/crashlytics/upgrade-sdk?platform=android), and fails to link to the guide, or mention it as a pre-requisite.
This is a common theme with the flutterfire 'getting started' guides. They omit required steps for getting started, and assume that the developer following the guide somehow knows about them.
| 1.0 | Firebase Crashlytics 'getting started' guide omits platform specific required steps. - If you following the Flutter Firebase Crashlytics getting started guide [here](https://firebase.flutter.dev/docs/crashlytics/usage/), you will end up with a runtime crash on android.
The guide fails to mention android specific setup steps in [this guide](https://firebase.google.com/docs/crashlytics/upgrade-sdk?platform=android), and fails to link to the guide, or mention it as a pre-requisite.
This is a common theme with the flutterfire 'getting started' guides. They omit required steps for getting started, and assume that the developer following the guide somehow knows about them.
| non_priority | firebase crashlytics getting started guide omits platform specific required steps if you following the flutter firebase crashlytics getting started guide you will end up with a runtime crash on android the guide fails to mention android specific setup steps in and fails to link to the guide or mention it as a pre requisite this is a common theme with the flutterfire getting started guides they omit required steps for getting started and assume that the developer following the guide somehow knows about them | 0 |
1,773 | 3,455,392,989 | IssuesEvent | 2015-12-17 19:55:35 | tomascharad/cause-management | https://api.github.com/repos/tomascharad/cause-management | closed | Protección CSRF | security | Para evitar requests que no surjan desde un formulario, en `app/controllers/application_controller.rb` descomentar la linea: `# protect_from_forgery with: :exception` | True | Protección CSRF - Para evitar requests que no surjan desde un formulario, en `app/controllers/application_controller.rb` descomentar la linea: `# protect_from_forgery with: :exception` | non_priority | protección csrf para evitar requests que no surjan desde un formulario en app controllers application controller rb descomentar la linea protect from forgery with exception | 0 |
100,724 | 16,490,229,454 | IssuesEvent | 2021-05-25 01:54:03 | rvvergara/fazebuk-next-app | https://api.github.com/repos/rvvergara/fazebuk-next-app | opened | CVE-2021-23383 (High) detected in handlebars-4.3.0.tgz | security vulnerability | ## CVE-2021-23383 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.3.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.3.0.tgz</a></p>
<p>Path to dependency file: fazebuk-next-app/package.json</p>
<p>Path to vulnerable library: fazebuk-next-app/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- jest-24.9.0.tgz (Root Library)
- jest-cli-24.9.0.tgz
- core-24.9.0.tgz
- reporters-24.9.0.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.3.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>development</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package handlebars before 4.7.7 are vulnerable to Prototype Pollution when selecting certain compiling options to compile templates coming from an untrusted source.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23383>CVE-2021-23383</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: handlebars - v4.7.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23383 (High) detected in handlebars-4.3.0.tgz - ## CVE-2021-23383 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.3.0.tgz</b></p></summary>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.3.0.tgz</a></p>
<p>Path to dependency file: fazebuk-next-app/package.json</p>
<p>Path to vulnerable library: fazebuk-next-app/node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- jest-24.9.0.tgz (Root Library)
- jest-cli-24.9.0.tgz
- core-24.9.0.tgz
- reporters-24.9.0.tgz
- istanbul-reports-2.2.6.tgz
- :x: **handlebars-4.3.0.tgz** (Vulnerable Library)
<p>Found in base branch: <b>development</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package handlebars before 4.7.7 are vulnerable to Prototype Pollution when selecting certain compiling options to compile templates coming from an untrusted source.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23383>CVE-2021-23383</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution: handlebars - v4.7.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in handlebars tgz cve high severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file fazebuk next app package json path to vulnerable library fazebuk next app node modules handlebars package json dependency hierarchy jest tgz root library jest cli tgz core tgz reporters tgz istanbul reports tgz x handlebars tgz vulnerable library found in base branch development vulnerability details the package handlebars before are vulnerable to prototype pollution when selecting certain compiling options to compile templates coming from an untrusted source publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars step up your open source security game with whitesource | 0 |
77,887 | 27,215,742,882 | IssuesEvent | 2023-02-20 21:32:32 | amyjko/bookish | https://api.github.com/repos/amyjko/bookish | opened | Footnotes in lists in callouts have incorrect vertical positions | defect reading | ```
Paragraph
=
Use the pseudocode below to answer the question that follows.
`python
i = 1
n = 1
while (i <= 4)
n = n * i
i = i + 1
print(n)
`
What will this pseudocode print?
1. 4
2. 12
3. 16
4. 24{footnote}
=
The footnote is offset incorrectly. | 1.0 | Footnotes in lists in callouts have incorrect vertical positions - ```
Paragraph
=
Use the pseudocode below to answer the question that follows.
`python
i = 1
n = 1
while (i <= 4)
n = n * i
i = i + 1
print(n)
`
What will this pseudocode print?
1. 4
2. 12
3. 16
4. 24{footnote}
=
The footnote is offset incorrectly. | non_priority | footnotes in lists in callouts have incorrect vertical positions paragraph use the pseudocode below to answer the question that follows python i n while i n n i i i print n what will this pseudocode print footnote the footnote is offset incorrectly | 0 |
1,150 | 2,870,524,710 | IssuesEvent | 2015-06-07 07:49:51 | pynac/pynac | https://api.github.com/repos/pynac/pynac | opened | cache expairseq::info computations | enhancement performance | The original behaviour is to compute `add`/`mul``::info` every time anew when it's called. | True | cache expairseq::info computations - The original behaviour is to compute `add`/`mul``::info` every time anew when it's called. | non_priority | cache expairseq info computations the original behaviour is to compute add mul info every time anew when it s called | 0 |
45,548 | 11,696,282,984 | IssuesEvent | 2020-03-06 09:29:18 | tensorflow/tensorflow | https://api.github.com/repos/tensorflow/tensorflow | closed | Which NVIDIA GPU to buy for TF with Windows? | TF 2.1 stat:awaiting response subtype:windows type:build/install | <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 8.1 & 10 dual-boot
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): binary
- TensorFlow version: the =last one
- Python version:3.7
- Installed using virtualenv? pip? conda?:pip conda
- Bazel version (if compiling from source):
- GCC/Compiler version (if compiling from source):
- CUDA/cuDNN version:10
- GPU model and memory: That's the question!
**Describe the problem**
Hello, I've tried since some days to use TF with my GPU GeForce GTX 760. It doesn't work because its computing capabilities are 3.0.
I've even tried to recompile the 1.8 version unsuccessfully.
Well, I've decided to change it.
The NVIDIA MSI GTX 1050 Ti 4GT OC - 4 Go seems to fit with the requisites but I can read it's not more available on their site.
Is it recommended anywhere?
I will mainly use TF for fractal art.
Below some infos of my PC.
Thanks for your help and patience.
Best regards,
François
Type de processeur QuadCore Intel Core i5-4670K, 4200 MHz (42 x 100)
Nom de la carte mère Asus Z87-A (2 PCI, 2 PCI-E x1, 3 PCI-E x16, 4 DDR3 DIMM, Audio, Video, Gigabit LAN) 3.0 (Gen 3) slots https://www.asus.com/fr/Motherboards/Z87A/
DIMM1: G Skill RipjawsX F3-12800CL9-4GBXL 4 Go DDR3-1600 DDR3 SDRAM (11-11-11-28 @ 800 MHz) (10-10-10-27 @ 761 MHz) (9-9-9-24 @ 685 MHz) (8-8-8-22 @ 609 MHz) (7-7-7-19 @ 533 MHz) (6-6-6-16 @ 457 MHz)
Moniteur Samsung SyncMaster S24F350 [24" LCD] (H4ZM101320)
**Provide the exact sequence of commands / steps that you executed before running into the problem**
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
| 1.0 | Which NVIDIA GPU to buy for TF with Windows? - <em>Please make sure that this is a build/installation issue. As per our [GitHub Policy](https://github.com/tensorflow/tensorflow/blob/master/ISSUES.md), we only address code/doc bugs, performance issues, feature requests and build/installation issues on GitHub. tag:build_template</em>
**System information**
- OS Platform and Distribution (e.g., Linux Ubuntu 16.04): Windows 8.1 & 10 dual-boot
- Mobile device (e.g. iPhone 8, Pixel 2, Samsung Galaxy) if the issue happens on mobile device:
- TensorFlow installed from (source or binary): binary
- TensorFlow version: the =last one
- Python version:3.7
- Installed using virtualenv? pip? conda?:pip conda
- Bazel version (if compiling from source):
- GCC/Compiler version (if compiling from source):
- CUDA/cuDNN version:10
- GPU model and memory: That's the question!
**Describe the problem**
Hello, I've tried since some days to use TF with my GPU GeForce GTX 760. It doesn't work because its computing capabilities are 3.0.
I've even tried to recompile the 1.8 version unsuccessfully.
Well, I've decided to change it.
The NVIDIA MSI GTX 1050 Ti 4GT OC - 4 Go seems to fit with the requisites but I can read it's not more available on their site.
Is it recommended anywhere?
I will mainly use TF for fractal art.
Below some infos of my PC.
Thanks for your help and patience.
Best regards,
François
Type de processeur QuadCore Intel Core i5-4670K, 4200 MHz (42 x 100)
Nom de la carte mère Asus Z87-A (2 PCI, 2 PCI-E x1, 3 PCI-E x16, 4 DDR3 DIMM, Audio, Video, Gigabit LAN) 3.0 (Gen 3) slots https://www.asus.com/fr/Motherboards/Z87A/
DIMM1: G Skill RipjawsX F3-12800CL9-4GBXL 4 Go DDR3-1600 DDR3 SDRAM (11-11-11-28 @ 800 MHz) (10-10-10-27 @ 761 MHz) (9-9-9-24 @ 685 MHz) (8-8-8-22 @ 609 MHz) (7-7-7-19 @ 533 MHz) (6-6-6-16 @ 457 MHz)
Moniteur Samsung SyncMaster S24F350 [24" LCD] (H4ZM101320)
**Provide the exact sequence of commands / steps that you executed before running into the problem**
**Any other info / logs**
Include any logs or source code that would be helpful to diagnose the problem. If including tracebacks, please include the full traceback. Large logs and files should be attached.
| non_priority | which nvidia gpu to buy for tf with windows please make sure that this is a build installation issue as per our we only address code doc bugs performance issues feature requests and build installation issues on github tag build template system information os platform and distribution e g linux ubuntu windows dual boot mobile device e g iphone pixel samsung galaxy if the issue happens on mobile device tensorflow installed from source or binary binary tensorflow version the last one python version installed using virtualenv pip conda pip conda bazel version if compiling from source gcc compiler version if compiling from source cuda cudnn version gpu model and memory that s the question describe the problem hello i ve tried since some days to use tf with my gpu geforce gtx it doesn t work because its computing capabilities are i ve even tried to recompile the version unsuccessfully well i ve decided to change it the nvidia msi gtx ti oc go seems to fit with the requisites but i can read it s not more available on their site is it recommended anywhere i will mainly use tf for fractal art below some infos of my pc thanks for your help and patience best regards françois type de processeur quadcore intel core mhz x nom de la carte mère asus a pci pci e pci e dimm audio video gigabit lan gen slots g skill ripjawsx go sdram mhz mhz mhz mhz mhz mhz moniteur samsung syncmaster provide the exact sequence of commands steps that you executed before running into the problem any other info logs include any logs or source code that would be helpful to diagnose the problem if including tracebacks please include the full traceback large logs and files should be attached | 0 |
9,216 | 6,808,221,618 | IssuesEvent | 2017-11-04 00:07:41 | eriq-augustine/psl | https://api.github.com/repos/eriq-augustine/psl | opened | Batches in GroundRuleStore | Performance | Occasionally, we will have to regenerate optimization terms (like in LazyMPEInference and LazyMaxLikelihoodMPE) because there are some changes (all additions at this point) in the ground rules.
It would be nice if we only had to re-generate a subset of the rules.
Perhaps we can have something like a batch in the GRS and then we can look for rules that came in during that batch.
(But, tracking deletes is much harder).
| True | Batches in GroundRuleStore - Occasionally, we will have to regenerate optimization terms (like in LazyMPEInference and LazyMaxLikelihoodMPE) because there are some changes (all additions at this point) in the ground rules.
It would be nice if we only had to re-generate a subset of the rules.
Perhaps we can have something like a batch in the GRS and then we can look for rules that came in during that batch.
(But, tracking deletes is much harder).
| non_priority | batches in groundrulestore occasionally we will have to regenerate optimization terms like in lazympeinference and lazymaxlikelihoodmpe because there are some changes all additions at this point in the ground rules it would be nice if we only had to re generate a subset of the rules perhaps we can have something like a batch in the grs and then we can look for rules that came in during that batch but tracking deletes is much harder | 0 |
194,523 | 15,434,477,392 | IssuesEvent | 2021-03-07 03:34:24 | kintoproj/kintohub | https://api.github.com/repos/kintoproj/kintohub | opened | Onboarding documentation | documentation | **Is your feature request related to a problem? Please describe.**
New contributors cannot start development easily unless they reach out to core contributors.
**Describe the solution you'd like**
- An onboarding page on how to setup basic development environment
- On each component we should set up a CONTRIBUTION/DEVELOPMENT readme indicate how to develop
| 1.0 | Onboarding documentation - **Is your feature request related to a problem? Please describe.**
New contributors cannot start development easily unless they reach out to core contributors.
**Describe the solution you'd like**
- An onboarding page on how to setup basic development environment
- On each component we should set up a CONTRIBUTION/DEVELOPMENT readme indicate how to develop
| non_priority | onboarding documentation is your feature request related to a problem please describe new contributors cannot start development easily unless they reach out to core contributors describe the solution you d like an onboarding page on how to setup basic development environment on each component we should set up a contribution development readme indicate how to develop | 0 |
284,491 | 24,603,743,572 | IssuesEvent | 2022-10-14 14:30:35 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | sql/importer: TestImportPgDump failed | C-test-failure O-robot branch-master T-disaster-recovery | sql/importer.TestImportPgDump [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4481194&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4481194&tab=artifacts#/) on master @ [bf13dc6061ab080c8c4306ebabb5065331877508](https://github.com/cockroachdb/cockroach/commits/bf13dc6061ab080c8c4306ebabb5065331877508):
```
=== RUN TestImportPgDump
test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/3e94a3d804ab82b4843bb87c9154ed0f/logTestImportPgDump565274107
test_log_scope.go:80: use -show-logs to present logs inline
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss
</p>
</details>
/cc @cockroachdb/bulk-io
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestImportPgDump.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-13463 | 1.0 | sql/importer: TestImportPgDump failed - sql/importer.TestImportPgDump [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4481194&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4481194&tab=artifacts#/) on master @ [bf13dc6061ab080c8c4306ebabb5065331877508](https://github.com/cockroachdb/cockroach/commits/bf13dc6061ab080c8c4306ebabb5065331877508):
```
=== RUN TestImportPgDump
test_log_scope.go:79: test logs captured to: /artifacts/tmp/_tmp/3e94a3d804ab82b4843bb87c9154ed0f/logTestImportPgDump565274107
test_log_scope.go:80: use -show-logs to present logs inline
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
Parameters in this failure:
- TAGS=bazel,gss
</p>
</details>
/cc @cockroachdb/bulk-io
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestImportPgDump.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-13463 | non_priority | sql importer testimportpgdump failed sql importer testimportpgdump with on master run testimportpgdump test log scope go test logs captured to artifacts tmp tmp test log scope go use show logs to present logs inline help see also parameters in this failure tags bazel gss cc cockroachdb bulk io jira issue crdb | 0 |
75,154 | 15,394,106,048 | IssuesEvent | 2021-03-03 17:29:46 | jgeraigery/kar-apps | https://api.github.com/repos/jgeraigery/kar-apps | opened | CVE-2020-36049 (High) detected in socket.io-parser-3.3.0.tgz | security vulnerability | ## CVE-2020-36049 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-parser-3.3.0.tgz</b></p></summary>
<p>socket.io protocol parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/socket.io-parser/-/socket.io-parser-3.3.0.tgz">https://registry.npmjs.org/socket.io-parser/-/socket.io-parser-3.3.0.tgz</a></p>
<p>Path to dependency file: kar-apps/reefer/frontend/package.json</p>
<p>Path to vulnerable library: kar-apps/reefer/frontend/node_modules/socket.io-client/node_modules/socket.io-parser/package.json</p>
<p>
Dependency Hierarchy:
- ngx-socket-io-3.2.0.tgz (Root Library)
- socket.io-client-2.3.0.tgz
- :x: **socket.io-parser-3.3.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/kar-apps/commit/ae1167158057c306dd8c0900bdc8dc0dae80c4eb">ae1167158057c306dd8c0900bdc8dc0dae80c4eb</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
socket.io-parser before 3.4.1 allows attackers to cause a denial of service (memory consumption) via a large packet because a concatenation approach is used.
<p>Publish Date: 2021-01-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36049>CVE-2020-36049</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36049">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36049</a></p>
<p>Release Date: 2021-01-08</p>
<p>Fix Resolution: socket.io-parser - 3.4.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"socket.io-parser","packageVersion":"3.3.0","packageFilePaths":["/reefer/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"ngx-socket-io:3.2.0;socket.io-client:2.3.0;socket.io-parser:3.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"socket.io-parser - 3.4.1"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-36049","vulnerabilityDetails":"socket.io-parser before 3.4.1 allows attackers to cause a denial of service (memory consumption) via a large packet because a concatenation approach is used.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36049","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-36049 (High) detected in socket.io-parser-3.3.0.tgz - ## CVE-2020-36049 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>socket.io-parser-3.3.0.tgz</b></p></summary>
<p>socket.io protocol parser</p>
<p>Library home page: <a href="https://registry.npmjs.org/socket.io-parser/-/socket.io-parser-3.3.0.tgz">https://registry.npmjs.org/socket.io-parser/-/socket.io-parser-3.3.0.tgz</a></p>
<p>Path to dependency file: kar-apps/reefer/frontend/package.json</p>
<p>Path to vulnerable library: kar-apps/reefer/frontend/node_modules/socket.io-client/node_modules/socket.io-parser/package.json</p>
<p>
Dependency Hierarchy:
- ngx-socket-io-3.2.0.tgz (Root Library)
- socket.io-client-2.3.0.tgz
- :x: **socket.io-parser-3.3.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/kar-apps/commit/ae1167158057c306dd8c0900bdc8dc0dae80c4eb">ae1167158057c306dd8c0900bdc8dc0dae80c4eb</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
socket.io-parser before 3.4.1 allows attackers to cause a denial of service (memory consumption) via a large packet because a concatenation approach is used.
<p>Publish Date: 2021-01-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36049>CVE-2020-36049</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36049">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36049</a></p>
<p>Release Date: 2021-01-08</p>
<p>Fix Resolution: socket.io-parser - 3.4.1</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"socket.io-parser","packageVersion":"3.3.0","packageFilePaths":["/reefer/frontend/package.json"],"isTransitiveDependency":true,"dependencyTree":"ngx-socket-io:3.2.0;socket.io-client:2.3.0;socket.io-parser:3.3.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"socket.io-parser - 3.4.1"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2020-36049","vulnerabilityDetails":"socket.io-parser before 3.4.1 allows attackers to cause a denial of service (memory consumption) via a large packet because a concatenation approach is used.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36049","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve high detected in socket io parser tgz cve high severity vulnerability vulnerable library socket io parser tgz socket io protocol parser library home page a href path to dependency file kar apps reefer frontend package json path to vulnerable library kar apps reefer frontend node modules socket io client node modules socket io parser package json dependency hierarchy ngx socket io tgz root library socket io client tgz x socket io parser tgz vulnerable library found in head commit a href found in base branch main vulnerability details socket io parser before allows attackers to cause a denial of service memory consumption via a large packet because a concatenation approach is used publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution socket io parser isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree ngx socket io socket io client socket io parser isminimumfixversionavailable true minimumfixversion socket io parser basebranches vulnerabilityidentifier cve vulnerabilitydetails socket io parser before allows attackers to cause a denial of service memory consumption via a large packet because a concatenation approach is used vulnerabilityurl | 0 |
35,959 | 14,910,275,505 | IssuesEvent | 2021-01-22 09:21:15 | terraform-providers/terraform-provider-azurerm | https://api.github.com/repos/terraform-providers/terraform-provider-azurerm | closed | Error destroying running Streaming Endpoint | bug services/media | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform (and AzureRM Provider) Version
version: 2.42.0
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* `azurerm_media_streaming_endpoint`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_media_streaming_endpoint" "endpoint" {
name = "endpoint"
resource_group_name = azurerm_resource_group.resource_group.name
media_services_account_name = azurerm_media_services_account.media_account.name
location = azurerm_resource_group.resource_group.location
scale_units = 1
auto_start_enabled = true
cdn_enabled = false
tags = var.tags
}
```
### Expected Behaviour
Run "Terraform destroy" should destroy Streaming Endpoint. I tested with stopped Streaming Endpoint and works fine, so the problem is only with running Streaming Endpoints.
### Actual Behaviour
Raise an error:
Error: Error deleting Streaming Endpoint "endpoint" in Media Services Account "medianonprod" (Resource Group "rg_media_streaming"): media.StreamingEndpointsClient#Delete: Failure sending request: StatusCode=412 -- Original Error: Code="InvalidOperation" Message="The server cannot execute the operation in its current state."
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. `terraform apply`
2. `terraform destroy`
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? For example: Running in a Azure China/Germany/Government? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Such as vendor documentation?
--->
* #0000
| 1.0 | Error destroying running Streaming Endpoint - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform (and AzureRM Provider) Version
version: 2.42.0
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* `azurerm_media_streaming_endpoint`
### Terraform Configuration Files
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
resource "azurerm_media_streaming_endpoint" "endpoint" {
name = "endpoint"
resource_group_name = azurerm_resource_group.resource_group.name
media_services_account_name = azurerm_media_services_account.media_account.name
location = azurerm_resource_group.resource_group.location
scale_units = 1
auto_start_enabled = true
cdn_enabled = false
tags = var.tags
}
```
### Expected Behaviour
Run "Terraform destroy" should destroy Streaming Endpoint. I tested with stopped Streaming Endpoint and works fine, so the problem is only with running Streaming Endpoints.
### Actual Behaviour
Raise an error:
Error: Error deleting Streaming Endpoint "endpoint" in Media Services Account "medianonprod" (Resource Group "rg_media_streaming"): media.StreamingEndpointsClient#Delete: Failure sending request: StatusCode=412 -- Original Error: Code="InvalidOperation" Message="The server cannot execute the operation in its current state."
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. `terraform apply`
2. `terraform destroy`
### Important Factoids
<!--- Are there anything atypical about your accounts that we should know? For example: Running in a Azure China/Germany/Government? --->
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Such as vendor documentation?
--->
* #0000
| non_priority | error destroying running streaming endpoint community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform and azurerm provider version version affected resource s azurerm media streaming endpoint terraform configuration files hcl resource azurerm media streaming endpoint endpoint name endpoint resource group name azurerm resource group resource group name media services account name azurerm media services account media account name location azurerm resource group resource group location scale units auto start enabled true cdn enabled false tags var tags expected behaviour run terraform destroy should destroy streaming endpoint i tested with stopped streaming endpoint and works fine so the problem is only with running streaming endpoints actual behaviour raise an error error error deleting streaming endpoint endpoint in media services account medianonprod resource group rg media streaming media streamingendpointsclient delete failure sending request statuscode original error code invalidoperation message the server cannot execute the operation in its current state steps to reproduce terraform apply terraform destroy important factoids references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here such as vendor documentation | 0 |
16,659 | 5,267,353,567 | IssuesEvent | 2017-02-04 21:38:06 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | [3.7.0b1] No items in Components menu on clean install | No Code Attached Yet | ### Steps to reproduce the issue
Just install beta 1
### Expected result
Menu items in the components menu such as 'contacts...
### Actual result

### System information (as much as possible)
PHP 5.6.25 MySql 5.7.14 Apache 2.4.23
### Additional comments
| 1.0 | [3.7.0b1] No items in Components menu on clean install - ### Steps to reproduce the issue
Just install beta 1
### Expected result
Menu items in the components menu such as 'contacts...
### Actual result

### System information (as much as possible)
PHP 5.6.25 MySql 5.7.14 Apache 2.4.23
### Additional comments
| non_priority | no items in components menu on clean install steps to reproduce the issue just install beta expected result menu items in the components menu such as contacts actual result system information as much as possible php mysql apache additional comments | 0 |
33,212 | 14,015,260,183 | IssuesEvent | 2020-10-29 13:05:57 | terraform-providers/terraform-provider-aws | https://api.github.com/repos/terraform-providers/terraform-provider-aws | closed | Significant slowdowns running terraform for WAF resources on AWS provider v2.69.0 | bug service/wafv2 upstream-terraform | <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
0.12.28
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* aws_wafv2_web_acl
### Terraform Configuration Files
```hcl
resource "aws_wafv2_web_acl" "demo-waf" {
name = "Demo-WAF"
description = "Demo-WAF"
scope = "REGIONAL"
default_action {
block {}
}
rule {
name = "RateLimit"
priority = 200
action {
block {}
}
statement {
rate_based_statement {
limit = 1000
aggregate_key_type = "IP"
}
}
visibility_config {
cloudwatch_metrics_enabled = false
metric_name = "demo_RateLimit"
sampled_requests_enabled = false
}
}
rule {
name = "AWSManagedRulesCommonRuleSet"
priority = 998
override_action {
none {}
}
statement {
managed_rule_group_statement {
name = "AWSManagedRulesCommonRuleSet"
vendor_name = "AWS"
}
}
visibility_config {
cloudwatch_metrics_enabled = false
metric_name = "AzureAD-AWSManagedRulesCommonRuleSet"
sampled_requests_enabled = false
}
}
rule {
name = "AWSManagedRulesKnownBadInputsRuleSet"
priority = 999
override_action {
none {}
}
statement {
managed_rule_group_statement {
name = "AWSManagedRulesKnownBadInputsRuleSet"
vendor_name = "AWS"
}
}
visibility_config {
cloudwatch_metrics_enabled = false
metric_name = "AzureAD-AWSManagedRulesKnownBadInputsRuleSet"
sampled_requests_enabled = false
}
}
visibility_config {
cloudwatch_metrics_enabled = true
metric_name = "Demo-WAF"
sampled_requests_enabled = false
}
}
```
### Debug Output
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
### Expected Behavior
This should create a plan within a few seconds after running `terraform plan`, and validate the code after a `terraform validate`
It should also only take up to a couple of minutes to run a `terraform plan`
### Actual Behavior
The plan and validate take a very long time to run - it works eventually, but it's taking upwards of 3 minutes for the validate, and five minutes, normally around ten for the plan. This is just for the one resource.
Apply takes even longer than this, presumably because it's running a plan on-top of doing other things.
If I downgrade my provider version to v 2.67.0, all of the actions are completed within a few seconds, as expected.
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. Set provider version to v2.69.0
2. `terraform plan`
3. Set provider version to v2.67.0
4. `terraform plan`
### Important Factoids
This only seems to affect WAF resources, I've tried the provider in other projects and haven't seen any issues.
I'm unsure if it's limited to the web acl resource specifically, but that's the only one I've been able to reproduce it in
### References
N/A
| 1.0 | Significant slowdowns running terraform for WAF resources on AWS provider v2.69.0 - <!---
Please note the following potential times when an issue might be in Terraform core:
* [Configuration Language](https://www.terraform.io/docs/configuration/index.html) or resource ordering issues
* [State](https://www.terraform.io/docs/state/index.html) and [State Backend](https://www.terraform.io/docs/backends/index.html) issues
* [Provisioner](https://www.terraform.io/docs/provisioners/index.html) issues
* [Registry](https://registry.terraform.io/) issues
* Spans resources across multiple providers
If you are running into one of these scenarios, we recommend opening an issue in the [Terraform core repository](https://github.com/hashicorp/terraform/) instead.
--->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
0.12.28
<!--- Please run `terraform -v` to show the Terraform core version and provider version(s). If you are not running the latest version of Terraform or the provider, please upgrade because your issue may have already been fixed. [Terraform documentation on provider versioning](https://www.terraform.io/docs/configuration/providers.html#provider-versions). --->
### Affected Resource(s)
<!--- Please list the affected resources and data sources. --->
* aws_wafv2_web_acl
### Terraform Configuration Files
```hcl
resource "aws_wafv2_web_acl" "demo-waf" {
name = "Demo-WAF"
description = "Demo-WAF"
scope = "REGIONAL"
default_action {
block {}
}
rule {
name = "RateLimit"
priority = 200
action {
block {}
}
statement {
rate_based_statement {
limit = 1000
aggregate_key_type = "IP"
}
}
visibility_config {
cloudwatch_metrics_enabled = false
metric_name = "demo_RateLimit"
sampled_requests_enabled = false
}
}
rule {
name = "AWSManagedRulesCommonRuleSet"
priority = 998
override_action {
none {}
}
statement {
managed_rule_group_statement {
name = "AWSManagedRulesCommonRuleSet"
vendor_name = "AWS"
}
}
visibility_config {
cloudwatch_metrics_enabled = false
metric_name = "AzureAD-AWSManagedRulesCommonRuleSet"
sampled_requests_enabled = false
}
}
rule {
name = "AWSManagedRulesKnownBadInputsRuleSet"
priority = 999
override_action {
none {}
}
statement {
managed_rule_group_statement {
name = "AWSManagedRulesKnownBadInputsRuleSet"
vendor_name = "AWS"
}
}
visibility_config {
cloudwatch_metrics_enabled = false
metric_name = "AzureAD-AWSManagedRulesKnownBadInputsRuleSet"
sampled_requests_enabled = false
}
}
visibility_config {
cloudwatch_metrics_enabled = true
metric_name = "Demo-WAF"
sampled_requests_enabled = false
}
}
```
### Debug Output
<!---
Please provide a link to a GitHub Gist containing the complete debug output. Please do NOT paste the debug output in the issue; just paste a link to the Gist.
To obtain the debug output, see the [Terraform documentation on debugging](https://www.terraform.io/docs/internals/debugging.html).
--->
### Expected Behavior
This should create a plan within a few seconds after running `terraform plan`, and validate the code after a `terraform validate`
It should also only take up to a couple of minutes to run a `terraform plan`
### Actual Behavior
The plan and validate take a very long time to run - it works eventually, but it's taking upwards of 3 minutes for the validate, and five minutes, normally around ten for the plan. This is just for the one resource.
Apply takes even longer than this, presumably because it's running a plan on-top of doing other things.
If I downgrade my provider version to v 2.67.0, all of the actions are completed within a few seconds, as expected.
### Steps to Reproduce
<!--- Please list the steps required to reproduce the issue. --->
1. Set provider version to v2.69.0
2. `terraform plan`
3. Set provider version to v2.67.0
4. `terraform plan`
### Important Factoids
This only seems to affect WAF resources, I've tried the provider in other projects and haven't seen any issues.
I'm unsure if it's limited to the web acl resource specifically, but that's the only one I've been able to reproduce it in
### References
N/A
| non_priority | significant slowdowns running terraform for waf resources on aws provider please note the following potential times when an issue might be in terraform core or resource ordering issues and issues issues issues spans resources across multiple providers if you are running into one of these scenarios we recommend opening an issue in the instead community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version affected resource s aws web acl terraform configuration files hcl resource aws web acl demo waf name demo waf description demo waf scope regional default action block rule name ratelimit priority action block statement rate based statement limit aggregate key type ip visibility config cloudwatch metrics enabled false metric name demo ratelimit sampled requests enabled false rule name awsmanagedrulescommonruleset priority override action none statement managed rule group statement name awsmanagedrulescommonruleset vendor name aws visibility config cloudwatch metrics enabled false metric name azuread awsmanagedrulescommonruleset sampled requests enabled false rule name awsmanagedrulesknownbadinputsruleset priority override action none statement managed rule group statement name awsmanagedrulesknownbadinputsruleset vendor name aws visibility config cloudwatch metrics enabled false metric name azuread awsmanagedrulesknownbadinputsruleset sampled requests enabled false visibility config cloudwatch metrics enabled true metric name demo waf sampled requests enabled false debug output please provide a link to a github gist containing the complete debug output please do not paste the debug output in the issue just paste a link to the gist to obtain the debug output see the expected behavior this should create a plan within a few seconds after running terraform plan and validate the code after a terraform validate it should also only take up to a couple of minutes to run a terraform plan actual behavior the plan and validate take a very long time to run it works eventually but it s taking upwards of minutes for the validate and five minutes normally around ten for the plan this is just for the one resource apply takes even longer than this presumably because it s running a plan on top of doing other things if i downgrade my provider version to v all of the actions are completed within a few seconds as expected steps to reproduce set provider version to terraform plan set provider version to terraform plan important factoids this only seems to affect waf resources i ve tried the provider in other projects and haven t seen any issues i m unsure if it s limited to the web acl resource specifically but that s the only one i ve been able to reproduce it in references n a | 0 |
87,319 | 25,085,725,091 | IssuesEvent | 2022-11-07 23:40:29 | opensearch-project/OpenSearch-Dashboards | https://api.github.com/repos/opensearch-project/OpenSearch-Dashboards | closed | [Vis Builder] hookup table visualization in vis buidler | vis builder v2.4.0 tableVis | **Description**
We would like to hook up table vis in the vis builder to allow customer to explore the new table vis.
**Expected behavior**
After enabling vis builder, go to `Visualize` and click `Create Visualization`. Click `VisBuilder` then choose `table` as the visualization type.
**OpenSearch Version**
2.4
**Dashboards Version**
2.4
| 1.0 | [Vis Builder] hookup table visualization in vis buidler - **Description**
We would like to hook up table vis in the vis builder to allow customer to explore the new table vis.
**Expected behavior**
After enabling vis builder, go to `Visualize` and click `Create Visualization`. Click `VisBuilder` then choose `table` as the visualization type.
**OpenSearch Version**
2.4
**Dashboards Version**
2.4
| non_priority | hookup table visualization in vis buidler description we would like to hook up table vis in the vis builder to allow customer to explore the new table vis expected behavior after enabling vis builder go to visualize and click create visualization click visbuilder then choose table as the visualization type opensearch version dashboards version | 0 |
27,269 | 27,979,737,755 | IssuesEvent | 2023-03-26 01:50:09 | PurpleI2P/i2pd | https://api.github.com/repos/PurpleI2P/i2pd | closed | how to run and use i2pd ? (client - home users) | question docs & usability | plz create this in wiki and put in the first line
| True | how to run and use i2pd ? (client - home users) - plz create this in wiki and put in the first line
| non_priority | how to run and use client home users plz create this in wiki and put in the first line | 0 |
6,606 | 7,700,605,395 | IssuesEvent | 2018-05-20 03:49:52 | ssube/isolex | https://api.github.com/repos/ssube/isolex | opened | trigger commands on reactions | service/parser status/confirmed type/feature | ### Summary
Reactions should appear in the listener's stream as messages from the user, with the reaction emoji as the body. The message should otherwise be normal.
This _should_ allow reactions (or any use of the emoji) to trigger a normal command for the bot.
### Scope
v1:
- discord listener should emit message for new reactions
- differentiate reaction emojis from normal message body emojis
- reaction will be entire body?
v2:
- slack listener also
### Use Case
Send increment/decrement commands to the counter handler.
### Issues
Need to make sure emojis mid-sentence don't trigger rep changes.
### Details
Requires #10 | 1.0 | trigger commands on reactions - ### Summary
Reactions should appear in the listener's stream as messages from the user, with the reaction emoji as the body. The message should otherwise be normal.
This _should_ allow reactions (or any use of the emoji) to trigger a normal command for the bot.
### Scope
v1:
- discord listener should emit message for new reactions
- differentiate reaction emojis from normal message body emojis
- reaction will be entire body?
v2:
- slack listener also
### Use Case
Send increment/decrement commands to the counter handler.
### Issues
Need to make sure emojis mid-sentence don't trigger rep changes.
### Details
Requires #10 | non_priority | trigger commands on reactions summary reactions should appear in the listener s stream as messages from the user with the reaction emoji as the body the message should otherwise be normal this should allow reactions or any use of the emoji to trigger a normal command for the bot scope discord listener should emit message for new reactions differentiate reaction emojis from normal message body emojis reaction will be entire body slack listener also use case send increment decrement commands to the counter handler issues need to make sure emojis mid sentence don t trigger rep changes details requires | 0 |
21,459 | 3,899,500,958 | IssuesEvent | 2016-04-17 19:39:14 | A3ReallifeRPG/ReallifeRPG5.0 | https://api.github.com/repos/A3ReallifeRPG/ReallifeRPG5.0 | closed | Medic HQ kaufen | bug test unconfirmed | Es ist möglich das Medic HQ kaufen mit einer Eigentumslizenz. Bitte dringend ändern | 1.0 | Medic HQ kaufen - Es ist möglich das Medic HQ kaufen mit einer Eigentumslizenz. Bitte dringend ändern | non_priority | medic hq kaufen es ist möglich das medic hq kaufen mit einer eigentumslizenz bitte dringend ändern | 0 |
59,820 | 8,380,415,452 | IssuesEvent | 2018-10-07 14:33:13 | keras-team/keras | https://api.github.com/repos/keras-team/keras | closed | Keras documentation of multi_gpu_model: example 2 can be misleading | Documentation good first issue help wanted | In the Keras documentation for `multi_gpu_model`, it is stated:
> To save the multi-gpu model, use .save(fname) or .save_weights(fname) with the template model (the argument you passed to multi_gpu_model), rather than the model returned by multi_gpu_model.
However in example 2 the template model is overwritten by the multi-gpu model:
```python
..
# Not needed to change the device scope for model definition:
model = Xception(weights=None, ..)
try:
model = multi_gpu_model(model, cpu_relocation=True)
print("Training using multiple GPUs..")
except:
print("Training using single GPU or CPU..")
model.compile(..)
..
```
This means that in this example it would not be possible to save the weights of the template model. I suggest rewritting to something like:
```python
..
# Not needed to change the device scope for model definition:
model = Xception(weights=None, ..)
try:
parallel_model = multi_gpu_model(model, cpu_relocation=True)
print("Training using multiple GPUs..")
except ValueError:
parallel_model = model
print("Training using single GPU or CPU..")
parallel_model.compile(..)
..
```
(I take this opportunity to except only a specific error)
| 1.0 | Keras documentation of multi_gpu_model: example 2 can be misleading - In the Keras documentation for `multi_gpu_model`, it is stated:
> To save the multi-gpu model, use .save(fname) or .save_weights(fname) with the template model (the argument you passed to multi_gpu_model), rather than the model returned by multi_gpu_model.
However in example 2 the template model is overwritten by the multi-gpu model:
```python
..
# Not needed to change the device scope for model definition:
model = Xception(weights=None, ..)
try:
model = multi_gpu_model(model, cpu_relocation=True)
print("Training using multiple GPUs..")
except:
print("Training using single GPU or CPU..")
model.compile(..)
..
```
This means that in this example it would not be possible to save the weights of the template model. I suggest rewritting to something like:
```python
..
# Not needed to change the device scope for model definition:
model = Xception(weights=None, ..)
try:
parallel_model = multi_gpu_model(model, cpu_relocation=True)
print("Training using multiple GPUs..")
except ValueError:
parallel_model = model
print("Training using single GPU or CPU..")
parallel_model.compile(..)
..
```
(I take this opportunity to except only a specific error)
| non_priority | keras documentation of multi gpu model example can be misleading in the keras documentation for multi gpu model it is stated to save the multi gpu model use save fname or save weights fname with the template model the argument you passed to multi gpu model rather than the model returned by multi gpu model however in example the template model is overwritten by the multi gpu model python not needed to change the device scope for model definition model xception weights none try model multi gpu model model cpu relocation true print training using multiple gpus except print training using single gpu or cpu model compile this means that in this example it would not be possible to save the weights of the template model i suggest rewritting to something like python not needed to change the device scope for model definition model xception weights none try parallel model multi gpu model model cpu relocation true print training using multiple gpus except valueerror parallel model model print training using single gpu or cpu parallel model compile i take this opportunity to except only a specific error | 0 |
205,337 | 23,333,091,223 | IssuesEvent | 2022-08-09 07:35:04 | Security-Phoenix-demo/Damn-Vulnerable-Source-Code | https://api.github.com/repos/Security-Phoenix-demo/Damn-Vulnerable-Source-Code | opened | ASPHX-WEB-MEDIUM-Use of JavaScript Library with Known Vulnerability | Security AppSec_Phoenix ASP_app3 ASP_588 ASP_Appsec_Phoenix_Website | # Info
**Application:** app3
**Component:** Appsec Phoenix Website
**Sub component / Asset:** https://appsecphoenix.com/
# Details
Vulnerable javascript library: Bootstrap
version: 4.0.0
Details:
CVE-2018-14041: The data-target property of scrollspy in bootstrap versions on or above 4.0.0-alpha and before 4.1.2 are vulnerable to Cross-Site Scripting(XSS) attacks. Please refer to vendor documentation (https://github.com/twbs/bootstrap/issues/20184) for the latest security updates.
----------------------------------------------
CVE-2018-14040: Bootstrap versions on or above 4.0.0-alpha and before 4.1.2 are vulnerable to Cross-Site Scripting(XSS) in collapse data-parent attribute. Please refer to vendor documentation (https://github.com/twbs/bootstrap/issues/20184) for latest security updates.
----------------------------------------------
CVE-2018-14042: Bootstrap versions on or above 4.0.0-alpha and before 4.1.2 are vulnerable to Cross-Site Scripting(XSS) in data-container property of tooltip. Please refer to vendor documentation (https://github.com/twbs/bootstrap/issues/20184) for latest security updates.
Found on the following pages (only first 10 pages are reported):
https://appsecphoenix.com/
https://appsecphoenix.com/platform/
https://appsecphoenix.com/pricing-benefits/
https://appsecphoenix.com/integration/
https://appsecphoenix.com/resources/
https://appsecphoenix.com/blog/
https://appsecphoenix.com/company/
https://appsecphoenix.com/contact/
https://appsecphoenix.com/log4j-log4shell-overview/
https://appsecphoenix.com/request-a-demo/
# Risk Context
This vulnerability's risk is Medium because the base severity is High
with a CVSS value of 7, the probability of exploitationin the wild is High and it isn't visible externally.
This vulnerability is selected to fix because the application is outside risk tolerance.
[Link to vulnerability](http://localhost:9000/app/vulnerability/09432782-6f59-448a-9065-0b277c56434b)
Created by AppSec Phoenix
| True | ASPHX-WEB-MEDIUM-Use of JavaScript Library with Known Vulnerability - # Info
**Application:** app3
**Component:** Appsec Phoenix Website
**Sub component / Asset:** https://appsecphoenix.com/
# Details
Vulnerable javascript library: Bootstrap
version: 4.0.0
Details:
CVE-2018-14041: The data-target property of scrollspy in bootstrap versions on or above 4.0.0-alpha and before 4.1.2 are vulnerable to Cross-Site Scripting(XSS) attacks. Please refer to vendor documentation (https://github.com/twbs/bootstrap/issues/20184) for the latest security updates.
----------------------------------------------
CVE-2018-14040: Bootstrap versions on or above 4.0.0-alpha and before 4.1.2 are vulnerable to Cross-Site Scripting(XSS) in collapse data-parent attribute. Please refer to vendor documentation (https://github.com/twbs/bootstrap/issues/20184) for latest security updates.
----------------------------------------------
CVE-2018-14042: Bootstrap versions on or above 4.0.0-alpha and before 4.1.2 are vulnerable to Cross-Site Scripting(XSS) in data-container property of tooltip. Please refer to vendor documentation (https://github.com/twbs/bootstrap/issues/20184) for latest security updates.
Found on the following pages (only first 10 pages are reported):
https://appsecphoenix.com/
https://appsecphoenix.com/platform/
https://appsecphoenix.com/pricing-benefits/
https://appsecphoenix.com/integration/
https://appsecphoenix.com/resources/
https://appsecphoenix.com/blog/
https://appsecphoenix.com/company/
https://appsecphoenix.com/contact/
https://appsecphoenix.com/log4j-log4shell-overview/
https://appsecphoenix.com/request-a-demo/
# Risk Context
This vulnerability's risk is Medium because the base severity is High
with a CVSS value of 7, the probability of exploitationin the wild is High and it isn't visible externally.
This vulnerability is selected to fix because the application is outside risk tolerance.
[Link to vulnerability](http://localhost:9000/app/vulnerability/09432782-6f59-448a-9065-0b277c56434b)
Created by AppSec Phoenix
| non_priority | asphx web medium use of javascript library with known vulnerability info application component appsec phoenix website sub component asset details vulnerable javascript library bootstrap version details cve the data target property of scrollspy in bootstrap versions on or above alpha and before are vulnerable to cross site scripting xss attacks please refer to vendor documentation for the latest security updates cve bootstrap versions on or above alpha and before are vulnerable to cross site scripting xss in collapse data parent attribute please refer to vendor documentation for latest security updates cve bootstrap versions on or above alpha and before are vulnerable to cross site scripting xss in data container property of tooltip please refer to vendor documentation for latest security updates found on the following pages only first pages are reported risk context this vulnerability s risk is medium because the base severity is high with a cvss value of the probability of exploitationin the wild is high and it isn t visible externally this vulnerability is selected to fix because the application is outside risk tolerance created by appsec phoenix | 0 |
6,973 | 16,041,383,397 | IssuesEvent | 2021-04-22 08:20:11 | dotnet/docs | https://api.github.com/repos/dotnet/docs | closed | This not working when class derived from enumeration is in composite key | Pri1 dotnet-architecture/prod microservices/tech needs-more-info |
```
public class FakeEntity: BaseEntity
{
public string Name { get; set; }
public int ContractId { get; set; }
}
public class FakeEntityLink
{
public FakeEntity FakeEntity { get; set; }
public int FakeEntityId { get; set; }
public int BillParamTypeEnumId { get; set; }
public BillParamTypeEnum BillParamTypeEnum { get; private set; } = BillParamTypeEnum.PriceCategory;
private FakeEntityLink() { }
public FakeEntityLink(FakeEntity fakeEntity, BillParamTypeEnum billParamTypeEnum)
{
FakeEntity = fakeEntity;
BillParamTypeEnum = billParamTypeEnum;
}
}
public class BillParamTypeEnum
: Enumeration
{
public static BillParamTypeEnum PriceCategory => new BillParamTypeEnum(1,"PriceCategory ");
public static BillParamTypeEnum VoltageTarifLevel => new BillParamTypeEnum(2, "VoltageTarifLevel ");
public static BillParamTypeEnum Sign => new BillParamTypeEnum(3,"Sign ");
public static BillParamTypeEnum VolumeCategory => new BillParamTypeEnum(4,"VolumeCategory");
public BillParamTypeEnum(int id, string name)
: base(id, name)
{
}
}
//and in dbContext
modelBuilder.Entity<FakeEntityLink>(entity => {
entity.HasKey(c => new { c.FakeEntityId, c.BillParamTypeEnumId });
entity.HasOne(link => link.FakeEntity).WithMany();
entity.HasOne(link => link.BillParamTypeEnum).WithMany().HasForeignKey(p=>p.BillParamTypeEnumId);
// entity.Property(p => p.BillParamTypeEnum).HasConversion<int>();
}
);
// code in controller
var fe =_db.FakeEntities.Find(1);
_db.FakeEntityLinks.Add(new FakeEntityLink(fe, BillParamTypeEnum.VoltageTarifLevel));
_db.SaveChanges(); // throws SqlException: Cannot insert explicit value for identity column in table 'BillParamTypeEnum' when IDENTITY_INSERT is set to OFF.
```
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d118f070-8743-e905-2d08-dc725a4e8b5f
* Version Independent ID: fae35cbd-75ae-4448-f066-7c91ccc35779
* Content: [Using Enumeration classes instead of enum types](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/microservice-ddd-cqrs-patterns/enumeration-classes-over-enum-types)
* Content Source: [docs/architecture/microservices/microservice-ddd-cqrs-patterns/enumeration-classes-over-enum-types.md](https://github.com/dotnet/docs/blob/main/docs/architecture/microservices/microservice-ddd-cqrs-patterns/enumeration-classes-over-enum-types.md)
* Product: **dotnet-architecture**
* Technology: **microservices**
* GitHub Login: @nishanil
* Microsoft Alias: **nanil** | 1.0 | This not working when class derived from enumeration is in composite key -
```
public class FakeEntity: BaseEntity
{
public string Name { get; set; }
public int ContractId { get; set; }
}
public class FakeEntityLink
{
public FakeEntity FakeEntity { get; set; }
public int FakeEntityId { get; set; }
public int BillParamTypeEnumId { get; set; }
public BillParamTypeEnum BillParamTypeEnum { get; private set; } = BillParamTypeEnum.PriceCategory;
private FakeEntityLink() { }
public FakeEntityLink(FakeEntity fakeEntity, BillParamTypeEnum billParamTypeEnum)
{
FakeEntity = fakeEntity;
BillParamTypeEnum = billParamTypeEnum;
}
}
public class BillParamTypeEnum
: Enumeration
{
public static BillParamTypeEnum PriceCategory => new BillParamTypeEnum(1,"PriceCategory ");
public static BillParamTypeEnum VoltageTarifLevel => new BillParamTypeEnum(2, "VoltageTarifLevel ");
public static BillParamTypeEnum Sign => new BillParamTypeEnum(3,"Sign ");
public static BillParamTypeEnum VolumeCategory => new BillParamTypeEnum(4,"VolumeCategory");
public BillParamTypeEnum(int id, string name)
: base(id, name)
{
}
}
//and in dbContext
modelBuilder.Entity<FakeEntityLink>(entity => {
entity.HasKey(c => new { c.FakeEntityId, c.BillParamTypeEnumId });
entity.HasOne(link => link.FakeEntity).WithMany();
entity.HasOne(link => link.BillParamTypeEnum).WithMany().HasForeignKey(p=>p.BillParamTypeEnumId);
// entity.Property(p => p.BillParamTypeEnum).HasConversion<int>();
}
);
// code in controller
var fe =_db.FakeEntities.Find(1);
_db.FakeEntityLinks.Add(new FakeEntityLink(fe, BillParamTypeEnum.VoltageTarifLevel));
_db.SaveChanges(); // throws SqlException: Cannot insert explicit value for identity column in table 'BillParamTypeEnum' when IDENTITY_INSERT is set to OFF.
```
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: d118f070-8743-e905-2d08-dc725a4e8b5f
* Version Independent ID: fae35cbd-75ae-4448-f066-7c91ccc35779
* Content: [Using Enumeration classes instead of enum types](https://docs.microsoft.com/en-us/dotnet/architecture/microservices/microservice-ddd-cqrs-patterns/enumeration-classes-over-enum-types)
* Content Source: [docs/architecture/microservices/microservice-ddd-cqrs-patterns/enumeration-classes-over-enum-types.md](https://github.com/dotnet/docs/blob/main/docs/architecture/microservices/microservice-ddd-cqrs-patterns/enumeration-classes-over-enum-types.md)
* Product: **dotnet-architecture**
* Technology: **microservices**
* GitHub Login: @nishanil
* Microsoft Alias: **nanil** | non_priority | this not working when class derived from enumeration is in composite key public class fakeentity baseentity public string name get set public int contractid get set public class fakeentitylink public fakeentity fakeentity get set public int fakeentityid get set public int billparamtypeenumid get set public billparamtypeenum billparamtypeenum get private set billparamtypeenum pricecategory private fakeentitylink public fakeentitylink fakeentity fakeentity billparamtypeenum billparamtypeenum fakeentity fakeentity billparamtypeenum billparamtypeenum public class billparamtypeenum enumeration public static billparamtypeenum pricecategory new billparamtypeenum pricecategory public static billparamtypeenum voltagetariflevel new billparamtypeenum voltagetariflevel public static billparamtypeenum sign new billparamtypeenum sign public static billparamtypeenum volumecategory new billparamtypeenum volumecategory public billparamtypeenum int id string name base id name and in dbcontext modelbuilder entity entity entity haskey c new c fakeentityid c billparamtypeenumid entity hasone link link fakeentity withmany entity hasone link link billparamtypeenum withmany hasforeignkey p p billparamtypeenumid entity property p p billparamtypeenum hasconversion code in controller var fe db fakeentities find db fakeentitylinks add new fakeentitylink fe billparamtypeenum voltagetariflevel db savechanges throws sqlexception cannot insert explicit value for identity column in table billparamtypeenum when identity insert is set to off document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product dotnet architecture technology microservices github login nishanil microsoft alias nanil | 0 |
34,388 | 29,796,407,746 | IssuesEvent | 2023-06-16 03:01:31 | flutter/website | https://api.github.com/repos/flutter/website | closed | Archive page no longer defaults to your platform | fd.infrastructure e1-hours | The archive page (https://flutter.io/docs/development/tools/sdk/archive) used to default to your OS ([based on user agent sniffing](https://github.com/flutter/website/blob/2f8781a59dd87b8bfc10dc132256d53f439cd478/js/tabs.js#L61)) before the update but it doesn't seem to anymore. It does persist your selected tab but if I clear all site data and hit the URL above from Chrome macOS it always seems to show the Windows section.
| 1.0 | Archive page no longer defaults to your platform - The archive page (https://flutter.io/docs/development/tools/sdk/archive) used to default to your OS ([based on user agent sniffing](https://github.com/flutter/website/blob/2f8781a59dd87b8bfc10dc132256d53f439cd478/js/tabs.js#L61)) before the update but it doesn't seem to anymore. It does persist your selected tab but if I clear all site data and hit the URL above from Chrome macOS it always seems to show the Windows section.
| non_priority | archive page no longer defaults to your platform the archive page used to default to your os before the update but it doesn t seem to anymore it does persist your selected tab but if i clear all site data and hit the url above from chrome macos it always seems to show the windows section | 0 |
161,621 | 13,864,255,118 | IssuesEvent | 2020-10-16 00:54:31 | ignitionrobotics/ign-physics | https://api.github.com/repos/ignitionrobotics/ign-physics | closed | ign-physics getting-started tutorials | documentation 🔮 dome | Some planning for getting-started tutorials:
- Introduction:
- Overview and motivation of ign-physics
- What are ign-physics purposes and where it is in the Ignition Robotics framework (optionally draw a framework graph highlighting ign-physics)
- Currently supported physics engines and corresponding features overview (link to [plugin description](https://github.com/ignitionrobotics/ign-physics/blob/master/tutorials/03_physics_plugins.md))
- Roadmap of future plans of ign-physics, a potential timeline of certain features.
- Installation: Tweaking and improving [current installation guide](https://github.com/ignitionrobotics/ign-physics/blob/master/tutorials/02_installation.md) for Dome release.
- Switching physics plugins tutorial
- Describe ways to switch physics engine from ign-gazebo
| 1.0 | ign-physics getting-started tutorials - Some planning for getting-started tutorials:
- Introduction:
- Overview and motivation of ign-physics
- What are ign-physics purposes and where it is in the Ignition Robotics framework (optionally draw a framework graph highlighting ign-physics)
- Currently supported physics engines and corresponding features overview (link to [plugin description](https://github.com/ignitionrobotics/ign-physics/blob/master/tutorials/03_physics_plugins.md))
- Roadmap of future plans of ign-physics, a potential timeline of certain features.
- Installation: Tweaking and improving [current installation guide](https://github.com/ignitionrobotics/ign-physics/blob/master/tutorials/02_installation.md) for Dome release.
- Switching physics plugins tutorial
- Describe ways to switch physics engine from ign-gazebo
| non_priority | ign physics getting started tutorials some planning for getting started tutorials introduction overview and motivation of ign physics what are ign physics purposes and where it is in the ignition robotics framework optionally draw a framework graph highlighting ign physics currently supported physics engines and corresponding features overview link to roadmap of future plans of ign physics a potential timeline of certain features installation tweaking and improving for dome release switching physics plugins tutorial describe ways to switch physics engine from ign gazebo | 0 |
76,375 | 15,495,954,864 | IssuesEvent | 2021-03-11 01:47:59 | kadirselcuk/wp-graphql | https://api.github.com/repos/kadirselcuk/wp-graphql | opened | CVE-2020-11022 (Medium) detected in jquery-2.1.4.min.js, jquery-1.7.1.min.js | security vulnerability | ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: wp-graphql/docs/node_modules/deep-diff/test/tests.html</p>
<p>Path to vulnerable library: wp-graphql/docs/node_modules/deep-diff/test/tests.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/js-base64/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/echo/index.html</p>
<p>Path to vulnerable library: wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/echo/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/multiplex/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/express-3.x/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/hapi/html/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/vm-browserify/example/run/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/express/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-11022 (Medium) detected in jquery-2.1.4.min.js, jquery-1.7.1.min.js - ## CVE-2020-11022 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.7.1.min.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: wp-graphql/docs/node_modules/deep-diff/test/tests.html</p>
<p>Path to vulnerable library: wp-graphql/docs/node_modules/deep-diff/test/tests.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/js-base64/test/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/echo/index.html</p>
<p>Path to vulnerable library: wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/echo/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/multiplex/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/express-3.x/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/hapi/html/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/vm-browserify/example/run/index.html,wp-graphql/src/Admin/GraphiQL/app/node_modules/sockjs/examples/express/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In jQuery versions greater than or equal to 1.2 and before 3.5.0, passing HTML from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.
<p>Publish Date: 2020-04-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11022>CVE-2020-11022</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/">https://blog.jquery.com/2020/04/10/jquery-3-5-0-released/</a></p>
<p>Release Date: 2020-04-29</p>
<p>Fix Resolution: jQuery - 3.5.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in jquery min js jquery min js cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file wp graphql docs node modules deep diff test tests html path to vulnerable library wp graphql docs node modules deep diff test tests html wp graphql src admin graphiql app node modules js test index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file wp graphql src admin graphiql app node modules sockjs examples echo index html path to vulnerable library wp graphql src admin graphiql app node modules sockjs examples echo index html wp graphql src admin graphiql app node modules sockjs examples multiplex index html wp graphql src admin graphiql app node modules sockjs examples express x index html wp graphql src admin graphiql app node modules sockjs examples hapi html index html wp graphql src admin graphiql app node modules vm browserify example run index html wp graphql src admin graphiql app node modules sockjs examples express index html dependency hierarchy x jquery min js vulnerable library found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
13,018 | 10,072,829,310 | IssuesEvent | 2019-07-24 08:13:50 | dotnet/roslyn | https://api.github.com/repos/dotnet/roslyn | closed | Cannot build latest Roslyn | Area-Infrastructure Discussion | **Version Used**: master as of earlier today, VS 16.20 preview 4
**Steps to Reproduce**:
cannot build roslyn. I've tried building in VS, as well as shutting down VS, deleting .vs folder, running Restore.cmd, running build.cmd, but it fails.
The error message is:
```
error CS1617: Invalid option '8.0' for /langversion. Use '/langversion:?' to list supported values.
``` | 1.0 | Cannot build latest Roslyn - **Version Used**: master as of earlier today, VS 16.20 preview 4
**Steps to Reproduce**:
cannot build roslyn. I've tried building in VS, as well as shutting down VS, deleting .vs folder, running Restore.cmd, running build.cmd, but it fails.
The error message is:
```
error CS1617: Invalid option '8.0' for /langversion. Use '/langversion:?' to list supported values.
``` | non_priority | cannot build latest roslyn version used master as of earlier today vs preview steps to reproduce cannot build roslyn i ve tried building in vs as well as shutting down vs deleting vs folder running restore cmd running build cmd but it fails the error message is error invalid option for langversion use langversion to list supported values | 0 |
86,099 | 10,719,656,452 | IssuesEvent | 2019-10-26 11:51:48 | gladki24/po-project | https://api.github.com/repos/gladki24/po-project | reopened | [BH] - analiza | designing discussion documentation | Analiza dotycząca proponowanego projektu Bounty Hunter ([BH]). Wszystko będzie później spisane w WIKI pod odpowiednim folderem. | 1.0 | [BH] - analiza - Analiza dotycząca proponowanego projektu Bounty Hunter ([BH]). Wszystko będzie później spisane w WIKI pod odpowiednim folderem. | non_priority | analiza analiza dotycząca proponowanego projektu bounty hunter wszystko będzie później spisane w wiki pod odpowiednim folderem | 0 |
9,060 | 4,390,157,329 | IssuesEvent | 2016-08-09 01:37:24 | jeff1evesque/machine-learning | https://api.github.com/repos/jeff1evesque/machine-learning | closed | Ensure 'build_docker' runs for OSX | build enhancement | We need to adjust the following from the `build_docker` bash script, to include the OSX case:
```bash
...
# run containers within common network
if [ "$(uname -s)" == 'Linux' ]; then
docker network create -d bridge app_nw && \
docker run --name base --net=app_nw -d container-default && \
docker run --name redis --net=app_nw -d container-redis && \
docker run --name webserver --net=app_nw -d container-webserver run && \
docker run --name database --net=app_nw -d container-database && \
docker run --name webserver-pytest --net=app_nw -it container-webserver test
else
winpty docker network create -d bridge app_nw && \
winpty docker run --name base --net=app_nw -d container-default && \
winpty docker run --name redis --net=app_nw -d container-redis && \
winpty docker run --name webserver --net=app_nw -d container-webserver run && \
winpty docker run --name database --net=app_nw -d container-database && \
winpty docker run --name webserver-pytest --net=app_nw -it container-webserver test
fi
...
``` | 1.0 | Ensure 'build_docker' runs for OSX - We need to adjust the following from the `build_docker` bash script, to include the OSX case:
```bash
...
# run containers within common network
if [ "$(uname -s)" == 'Linux' ]; then
docker network create -d bridge app_nw && \
docker run --name base --net=app_nw -d container-default && \
docker run --name redis --net=app_nw -d container-redis && \
docker run --name webserver --net=app_nw -d container-webserver run && \
docker run --name database --net=app_nw -d container-database && \
docker run --name webserver-pytest --net=app_nw -it container-webserver test
else
winpty docker network create -d bridge app_nw && \
winpty docker run --name base --net=app_nw -d container-default && \
winpty docker run --name redis --net=app_nw -d container-redis && \
winpty docker run --name webserver --net=app_nw -d container-webserver run && \
winpty docker run --name database --net=app_nw -d container-database && \
winpty docker run --name webserver-pytest --net=app_nw -it container-webserver test
fi
...
``` | non_priority | ensure build docker runs for osx we need to adjust the following from the build docker bash script to include the osx case bash run containers within common network if then docker network create d bridge app nw docker run name base net app nw d container default docker run name redis net app nw d container redis docker run name webserver net app nw d container webserver run docker run name database net app nw d container database docker run name webserver pytest net app nw it container webserver test else winpty docker network create d bridge app nw winpty docker run name base net app nw d container default winpty docker run name redis net app nw d container redis winpty docker run name webserver net app nw d container webserver run winpty docker run name database net app nw d container database winpty docker run name webserver pytest net app nw it container webserver test fi | 0 |
244,457 | 26,403,925,983 | IssuesEvent | 2023-01-13 05:35:56 | switcherapi/switcher-api | https://api.github.com/repos/switcherapi/switcher-api | closed | Security Patch: Vulnerability introduced by jsonwebtoken@8.5.1 | security patch | **Describe the patch**
Vulnerable module: jsonwebtoken
Introduced through: jsonwebtoken@8.5.1
**Optional - Remediation**
- **Name**: jsonwebtoken
- **Version**: 9.0.0
**Additional context**
[Links or documentation to the vulnerability or dependency update.](https://security.snyk.io/vuln/SNYK-JS-JSONWEBTOKEN-3180022)
| True | Security Patch: Vulnerability introduced by jsonwebtoken@8.5.1 - **Describe the patch**
Vulnerable module: jsonwebtoken
Introduced through: jsonwebtoken@8.5.1
**Optional - Remediation**
- **Name**: jsonwebtoken
- **Version**: 9.0.0
**Additional context**
[Links or documentation to the vulnerability or dependency update.](https://security.snyk.io/vuln/SNYK-JS-JSONWEBTOKEN-3180022)
| non_priority | security patch vulnerability introduced by jsonwebtoken describe the patch vulnerable module jsonwebtoken introduced through jsonwebtoken optional remediation name jsonwebtoken version additional context | 0 |
258,166 | 27,563,862,253 | IssuesEvent | 2023-03-08 01:11:47 | jtimberlake/pacbot | https://api.github.com/repos/jtimberlake/pacbot | opened | CVE-2020-15522 (Medium) detected in bcprov-jdk15on-1.56.jar | security vulnerability | ## CVE-2020-15522 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.56.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p>
<p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p>
<p>Path to dependency file: /api/pacman-api-auth/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RELEASE.jar (Root Library)
- spring-cloud-starter-2.0.0.RELEASE.jar
- spring-security-rsa-1.0.5.RELEASE.jar
- bcpkix-jdk15on-1.56.jar
- :x: **bcprov-jdk15on-1.56.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Bouncy Castle BC Java before 1.66, BC C# .NET before 1.8.7, BC-FJA before 1.0.1.2, 1.0.2.1, and BC-FNA before 1.0.1.1 have a timing issue within the EC math library that can expose information about the private key when an attacker is able to observe timing information for the generation of multiple deterministic ECDSA signatures.
<p>Publish Date: 2021-05-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15522>CVE-2020-15522</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-15522">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-15522</a></p>
<p>Release Date: 2021-05-20</p>
<p>Fix Resolution (org.bouncycastle:bcprov-jdk15on): 1.66</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 3.0.3</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| True | CVE-2020-15522 (Medium) detected in bcprov-jdk15on-1.56.jar - ## CVE-2020-15522 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>bcprov-jdk15on-1.56.jar</b></p></summary>
<p>The Bouncy Castle Crypto package is a Java implementation of cryptographic algorithms. This jar contains JCE provider and lightweight API for the Bouncy Castle Cryptography APIs for JDK 1.5 to JDK 1.8.</p>
<p>Library home page: <a href="http://www.bouncycastle.org/java.html">http://www.bouncycastle.org/java.html</a></p>
<p>Path to dependency file: /api/pacman-api-auth/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar,/home/wss-scanner/.m2/repository/org/bouncycastle/bcprov-jdk15on/1.56/bcprov-jdk15on-1.56.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-openfeign-2.0.0.RELEASE.jar (Root Library)
- spring-cloud-starter-2.0.0.RELEASE.jar
- spring-security-rsa-1.0.5.RELEASE.jar
- bcpkix-jdk15on-1.56.jar
- :x: **bcprov-jdk15on-1.56.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Bouncy Castle BC Java before 1.66, BC C# .NET before 1.8.7, BC-FJA before 1.0.1.2, 1.0.2.1, and BC-FNA before 1.0.1.1 have a timing issue within the EC math library that can expose information about the private key when an attacker is able to observe timing information for the generation of multiple deterministic ECDSA signatures.
<p>Publish Date: 2021-05-20
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15522>CVE-2020-15522</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-15522">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-15522</a></p>
<p>Release Date: 2021-05-20</p>
<p>Fix Resolution (org.bouncycastle:bcprov-jdk15on): 1.66</p>
<p>Direct dependency fix Resolution (org.springframework.cloud:spring-cloud-starter-openfeign): 3.0.3</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| non_priority | cve medium detected in bcprov jar cve medium severity vulnerability vulnerable library bcprov jar the bouncy castle crypto package is a java implementation of cryptographic algorithms this jar contains jce provider and lightweight api for the bouncy castle cryptography apis for jdk to jdk library home page a href path to dependency file api pacman api auth pom xml path to vulnerable library home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar home wss scanner repository org bouncycastle bcprov bcprov jar dependency hierarchy spring cloud starter openfeign release jar root library spring cloud starter release jar spring security rsa release jar bcpkix jar x bcprov jar vulnerable library found in base branch master vulnerability details bouncy castle bc java before bc c net before bc fja before and bc fna before have a timing issue within the ec math library that can expose information about the private key when an attacker is able to observe timing information for the generation of multiple deterministic ecdsa signatures publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org bouncycastle bcprov direct dependency fix resolution org springframework cloud spring cloud starter openfeign check this box to open an automated fix pr | 0 |
17,256 | 2,993,217,628 | IssuesEvent | 2015-07-22 01:08:20 | googlei18n/noto-fonts | https://api.github.com/repos/googlei18n/noto-fonts | opened | Svarita preceeding visarga not rendered correctly | Script-Devanagari Type-Defect | Moved from googlei18n/noto-alpha/issues/284
Imported from Google Code issue #284 created by vvasuki@google.com on 2015-06-02T18:14:16.000Z:
Some lines in the Rigveda text http://www.detlef108.de/RV-D-UTF8.html are not rendered correctly. In particular, consider: [स न॑ः पि॒तेव॑]
This is what it looks like with Noto Sans: http://i.imgur.com/jocRriY.png
This is what it looks like with Chandas : http://i.imgur.com/PqyVQ1i.png
-------------------------------------------------------------------
Comment #1 originally posted by cibu@google.com on 2015-06-03T16:06:51.000Z:
I don't think Noto Devanagari is really designed for vedic. However, there are some low hanging fruits like this could be corrected. | 1.0 | Svarita preceeding visarga not rendered correctly - Moved from googlei18n/noto-alpha/issues/284
Imported from Google Code issue #284 created by vvasuki@google.com on 2015-06-02T18:14:16.000Z:
Some lines in the Rigveda text http://www.detlef108.de/RV-D-UTF8.html are not rendered correctly. In particular, consider: [स न॑ः पि॒तेव॑]
This is what it looks like with Noto Sans: http://i.imgur.com/jocRriY.png
This is what it looks like with Chandas : http://i.imgur.com/PqyVQ1i.png
-------------------------------------------------------------------
Comment #1 originally posted by cibu@google.com on 2015-06-03T16:06:51.000Z:
I don't think Noto Devanagari is really designed for vedic. However, there are some low hanging fruits like this could be corrected. | non_priority | svarita preceeding visarga not rendered correctly moved from noto alpha issues imported from google code issue created by vvasuki google com on some lines in the rigveda text are not rendered correctly in particular consider this is what it looks like with noto sans this is what it looks like with chandas comment originally posted by cibu google com on i don t think noto devanagari is really designed for vedic however there are some low hanging fruits like this could be corrected | 0 |
130,163 | 10,599,684,735 | IssuesEvent | 2019-10-10 08:30:42 | microsoft/AzureStorageExplorer | https://api.github.com/repos/microsoft/AzureStorageExplorer | closed | 'Set Default Access Tier…' is not localized for one GPV2 storage account | 🌐 localization 🧪 testing | **Storage Explorer Version:** 1.10.1
**Build:** [20191010.3](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3128931)
**Branch:** master
**Platform/OS:** Windows 10/ Linux Ubuntu 19.04/macOS High Sierra
**Language:** Chinese (zh-CN)
**Architecture:** ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Launch Storage Explorer using the command 'StorageExplorer.exe --lang=zh-CN'.
2. Expand Storage Accounts -> Select one GPV2 storage account.
3. Check its context menu.
**Expect Experience:**
All strings on the context menu are localized.
**Actual Experience:**
'Set Default Access Tier…' is not localized.
**More Info:**
Only verified this issue in Chinese (zh-CN). | 1.0 | 'Set Default Access Tier…' is not localized for one GPV2 storage account - **Storage Explorer Version:** 1.10.1
**Build:** [20191010.3](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3128931)
**Branch:** master
**Platform/OS:** Windows 10/ Linux Ubuntu 19.04/macOS High Sierra
**Language:** Chinese (zh-CN)
**Architecture:** ia32/x64
**Regression From:** Not a regression
**Steps to reproduce:**
1. Launch Storage Explorer using the command 'StorageExplorer.exe --lang=zh-CN'.
2. Expand Storage Accounts -> Select one GPV2 storage account.
3. Check its context menu.
**Expect Experience:**
All strings on the context menu are localized.
**Actual Experience:**
'Set Default Access Tier…' is not localized.
**More Info:**
Only verified this issue in Chinese (zh-CN). | non_priority | set default access tier… is not localized for one storage account storage explorer version build branch master platform os windows linux ubuntu macos high sierra language chinese zh cn architecture regression from not a regression steps to reproduce launch storage explorer using the command storageexplorer exe lang zh cn expand storage accounts select one storage account check its context menu expect experience all strings on the context menu are localized actual experience set default access tier… is not localized more info only verified this issue in chinese zh cn | 0 |
27,137 | 4,282,466,365 | IssuesEvent | 2016-07-15 09:16:15 | imixs/imixs-workflow | https://api.github.com/repos/imixs/imixs-workflow | closed | WorkflowSevice - standard queries - remove useless joins | enhancement testing | Some of the standard queries of the WorklowService include usells joins for the $workitemid | 1.0 | WorkflowSevice - standard queries - remove useless joins - Some of the standard queries of the WorklowService include usells joins for the $workitemid | non_priority | workflowsevice standard queries remove useless joins some of the standard queries of the worklowservice include usells joins for the workitemid | 0 |
55,185 | 13,538,136,791 | IssuesEvent | 2020-09-16 11:37:57 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | opened | Mac 10.14.4 can not build the clickhouse | build |
**Operating system**
Mac 10.14.4
**Cmake version**
`cmake --version
3.18.2`
**Ninja version**
`ninja --version
1.10.1`
**Compiler name and version**
`Apple LLVM version 10.0.1 (clang-1001.0.46.4)
Target: x86_64-apple-darwin18.5.0
Thread model: posix
InstalledDir: /Library/Developer/CommandLineTools/usr/bin
`
**Full cmake and/or ninja output**
```-- Looking for pthread.h
-- Looking for pthread.h - not found
CMake Error at /usr/local/Cellar/cmake/3.18.2/share/cmake/Modules/FindPackageHandleStandardArgs.cmake:165 (message):
Could NOT find Threads (missing: Threads_FOUND)
Call Stack (most recent call first):
/usr/local/Cellar/cmake/3.18.2/share/cmake/Modules/FindPackageHandleStandardArgs.cmake:458 (_FPHSA_FAILURE_MESSAGE)
/usr/local/Cellar/cmake/3.18.2/share/cmake/Modules/FindThreads.cmake:234 (FIND_PACKAGE_HANDLE_STANDARD_ARGS)
cmake/default_libs.cmake:30 (find_package)
CMakeLists.txt:234 (include)
-- Configuring incomplete, errors occurred!
See also "/Users/doude/CLionProjects/ClickHouse/build/CMakeFiles/CMakeOutput.log".
See also "/Users/doude/CLionProjects/ClickHouse/build/CMakeFiles/CMakeError.log".
```
**CMakeLists.txt**
```
foreach(policy
CMP0023
CMP0048 # CMake 3.0
CMP0074 # CMake 3.12
CMP0077
CMP0079
)
if(POLICY ${policy})
cmake_policy(SET ${policy} NEW)
endif()
endforeach()
project(ClickHouse)
cmake_minimum_required(VERSION 3.3)
# Ignore export() since we don't use it,
# but it gets broken with a global targets via link_libraries()
macro (export)
endmacro ()
set(CMAKE_MODULE_PATH ${CMAKE_MODULE_PATH} "${CMAKE_CURRENT_SOURCE_DIR}/cmake/Modules/")
set(CMAKE_EXPORT_COMPILE_COMMANDS 1) # Write compile_commands.json
set(CMAKE_LINK_DEPENDS_NO_SHARED 1) # Do not relink all depended targets on .so
set(CMAKE_CONFIGURATION_TYPES "RelWithDebInfo;Debug;Release;MinSizeRel" CACHE STRING "" FORCE)
set(CMAKE_DEBUG_POSTFIX "d" CACHE STRING "Generate debug library name with a postfix.") # To be consistent with CMakeLists from contrib libs.
include (cmake/arch.cmake)
option(ENABLE_IPO "Enable inter-procedural optimization (aka LTO)" OFF) # need cmake 3.9+
if(ENABLE_IPO)
cmake_policy(SET CMP0069 NEW)
include(CheckIPOSupported)
check_ipo_supported(RESULT IPO_SUPPORTED OUTPUT IPO_NOT_SUPPORTED)
if(IPO_SUPPORTED)
message(STATUS "IPO/LTO is supported, enabling")
set(CMAKE_INTERPROCEDURAL_OPTIMIZATION TRUE)
else()
message(STATUS "IPO/LTO is not supported: <${IPO_NOT_SUPPORTED}>")
endif()
else()
message(STATUS "IPO/LTO not enabled.")
endif()
if (COMPILER_GCC)
# Require minimum version of gcc
set (GCC_MINIMUM_VERSION 8)
if (CMAKE_CXX_COMPILER_VERSION VERSION_LESS ${GCC_MINIMUM_VERSION} AND NOT CMAKE_VERSION VERSION_LESS 2.8.9)
message (FATAL_ERROR "GCC version must be at least ${GCC_MINIMUM_VERSION}. For example, if GCC ${GCC_MINIMUM_VERSION} is available under gcc-${GCC_MINIMUM_VERSION}, g++-${GCC_MINIMUM_VERSION} names, do the following: export CC=gcc-${GCC_MINIMUM_VERSION} CXX=g++-${GCC_MINIMUM_VERSION}; rm -rf CMakeCache.txt CMakeFiles; and re run cmake or ./release.")
endif ()
elseif (CMAKE_CXX_COMPILER_ID STREQUAL "Clang")
# Require minimum version of clang
set (CLANG_MINIMUM_VERSION 7)
if (CMAKE_CXX_COMPILER_VERSION VERSION_LESS ${CLANG_MINIMUM_VERSION})
message (FATAL_ERROR "Clang version must be at least ${CLANG_MINIMUM_VERSION}.")
endif ()
else ()
message (WARNING "You are using an unsupported compiler. Compilation has only been tested with Clang 6+ and GCC 7+.")
endif ()
# Check that submodules are present only if source was downloaded with git
if (EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/.git" AND NOT EXISTS "${ClickHouse_SOURCE_DIR}/contrib/boost/boost")
message (FATAL_ERROR "Submodules are not initialized. Run\n\tgit submodule update --init --recursive")
endif ()
include (cmake/find_ccache.cmake)
if (NOT CMAKE_BUILD_TYPE OR CMAKE_BUILD_TYPE STREQUAL "None")
message (STATUS "CMAKE_BUILD_TYPE is not set, set to default = RELWITHDEBINFO")
set (CMAKE_BUILD_TYPE "RELWITHDEBINFO")
endif ()
string(TOUPPER ${CMAKE_BUILD_TYPE} CMAKE_BUILD_TYPE_UC)
message (STATUS "CMAKE_BUILD_TYPE: ${CMAKE_BUILD_TYPE}")
option (USE_STATIC_LIBRARIES "Set to FALSE to use shared libraries" ON)
option (MAKE_STATIC_LIBRARIES "Set to FALSE to make shared libraries" ${USE_STATIC_LIBRARIES})
if (NOT MAKE_STATIC_LIBRARIES)
option (SPLIT_SHARED_LIBRARIES "DEV ONLY. Keep all internal libs as separate .so for faster linking" OFF)
option (CLICKHOUSE_SPLIT_BINARY "Make several binaries instead one bundled (clickhouse-server, clickhouse-client, ... )" OFF)
endif ()
if (MAKE_STATIC_LIBRARIES AND SPLIT_SHARED_LIBRARIES)
message(FATAL_ERROR "Defining SPLIT_SHARED_LIBRARIES=1 without MAKE_STATIC_LIBRARIES=0 has no effect.")
endif()
if (NOT MAKE_STATIC_LIBRARIES AND SPLIT_SHARED_LIBRARIES)
set(BUILD_SHARED_LIBS 1 CACHE INTERNAL "")
endif ()
if (USE_STATIC_LIBRARIES)
list(REVERSE CMAKE_FIND_LIBRARY_SUFFIXES)
endif ()
include (cmake/sanitize.cmake)
if (CMAKE_GENERATOR STREQUAL "Ninja")
# Turn on colored output. https://github.com/ninja-build/ninja/wiki/FAQ
set (CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -fdiagnostics-color=always")
set (CMAKE_C_FLAGS "${CMAKE_C_FLAGS} -fdiagnostics-color=always")
endif ()
if (NOT MSVC)
set (COMMON_WARNING_FLAGS "${COMMON_WARNING_FLAGS} -Wall") # -Werror is also added inside directories with our own code.
endif ()
if (COMPILER_GCC OR COMPILER_CLANG)
set (CXX_WARNING_FLAGS "${CXX_WARNING_FLAGS} -Wnon-virtual-dtor")
endif ()
if (COMPILER_GCC AND CMAKE_CXX_COMPILER_VERSION VERSION_GREATER "8.3.0")
# Warnings in protobuf generating
set (CXX_WARNING_FLAGS "${CXX_WARNING_FLAGS} -Wno-array-bounds")
endif ()
if (COMPILER_CLANG)
# clang: warning: argument unused during compilation: '-specs=/usr/share/dpkg/no-pie-compile.specs' [-Wunused-command-line-argument]
set (COMMON_WARNING_FLAGS "${COMMON_WARNING_FLAGS} -Wno-unused-command-line-argument")
# generate ranges for fast "addr2line" search
if (NOT CMAKE_BUILD_TYPE_UC STREQUAL "RELEASE")
set(COMPILER_FLAGS "${COMPILER_FLAGS} -gdwarf-aranges")
endif ()
endif ()
option (ENABLE_TESTS "Enables tests" ON)
if (CMAKE_SYSTEM_PROCESSOR MATCHES "amd64|x86_64")
option (USE_INTERNAL_MEMCPY "Use internal implementation of 'memcpy' function instead of provided by libc. Only for x86_64." ON)
if (OS_LINUX AND NOT UNBUNDLED AND MAKE_STATIC_LIBRARIES AND NOT SPLIT_SHARED_LIBRARIES AND CMAKE_VERSION VERSION_GREATER "3.9.0")
option (GLIBC_COMPATIBILITY "Set to TRUE to enable compatibility with older glibc libraries. Only for x86_64, Linux. Implies USE_INTERNAL_MEMCPY." ON)
endif ()
if (NOT CMAKE_VERSION VERSION_GREATER "3.9.0")
message (WARNING "CMake version must be greater than 3.9.0 for production builds.")
endif ()
endif ()
string(REGEX MATCH "-?[0-9]+(.[0-9]+)?$" COMPILER_POSTFIX ${CMAKE_CXX_COMPILER})
find_program (LLD_PATH NAMES "lld${COMPILER_POSTFIX}" "lld")
find_program (GOLD_PATH NAMES "gold")
if (COMPILER_CLANG AND LLD_PATH AND NOT LINKER_NAME)
set (LINKER_NAME "lld")
elseif (GOLD_PATH)
set (LINKER_NAME "gold")
endif ()
if (LINKER_NAME)
message(STATUS "Using linker: ${LINKER_NAME} (selected from: LLD_PATH=${LLD_PATH}; GOLD_PATH=${GOLD_PATH}; COMPILER_POSTFIX=${COMPILER_POSTFIX})")
set (CMAKE_EXE_LINKER_FLAGS "${CMAKE_EXE_LINKER_FLAGS} -fuse-ld=${LINKER_NAME}")
endif ()
cmake_host_system_information(RESULT AVAILABLE_PHYSICAL_MEMORY QUERY AVAILABLE_PHYSICAL_MEMORY) # Not available under freebsd
if(NOT AVAILABLE_PHYSICAL_MEMORY OR AVAILABLE_PHYSICAL_MEMORY GREATER 8000)
option(COMPILER_PIPE "-pipe compiler option [less /tmp usage, more ram usage]" ON)
endif()
if(COMPILER_PIPE)
set(COMPILER_FLAGS "${COMPILER_FLAGS} -pipe")
else()
message(STATUS "Disabling compiler -pipe option (have only ${AVAILABLE_PHYSICAL_MEMORY} mb of memory)")
endif()
if(NOT DISABLE_CPU_OPTIMIZE)
include(cmake/test_cpu.cmake)
endif()
if(NOT COMPILER_CLANG) # clang: error: the clang compiler does not support '-march=native'
option(ARCH_NATIVE "Enable -march=native compiler flag" ${ARCH_ARM})
endif()
if (ARCH_NATIVE)
set (COMPILER_FLAGS "${COMPILER_FLAGS} -march=native")
endif ()
if (CMAKE_VERSION VERSION_LESS "3.8.0")
if (NOT MSVC)
set (CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17")
endif ()
else ()
set (CMAKE_CXX_STANDARD 17)
set (CMAKE_CXX_EXTENSIONS 0) # https://cmake.org/cmake/help/latest/prop_tgt/CXX_EXTENSIONS.html#prop_tgt:CXX_EXTENSIONS
set (CMAKE_CXX_STANDARD_REQUIRED ON)
set (CXX_FLAGS_INTERNAL_COMPILER "-std=c++17")
endif ()
if (COMPILER_GCC OR COMPILER_CLANG)
# Enable C++14 sized global deallocation functions. It should be enabled by setting -std=c++14 but I'm not sure.
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -fsized-deallocation")
endif ()
option(WITH_COVERAGE "Build with coverage." 0)
if(WITH_COVERAGE AND COMPILER_CLANG)
set(COMPILER_FLAGS "${COMPILER_FLAGS} -fprofile-instr-generate -fcoverage-mapping")
# If we want to disable coverage for specific translation units
set(WITHOUT_COVERAGE "-fno-profile-instr-generate -fno-coverage-mapping")
endif()
if(WITH_COVERAGE AND COMPILER_GCC)
set(COMPILER_FLAGS "${COMPILER_FLAGS} -fprofile-arcs -ftest-coverage")
set(COVERAGE_OPTION "-lgcov")
set(WITHOUT_COVERAGE "-fno-profile-arcs -fno-test-coverage")
endif()
set (CMAKE_BUILD_COLOR_MAKEFILE ON)
set (CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} ${COMPILER_FLAGS} ${PLATFORM_EXTRA_CXX_FLAG} -fno-omit-frame-pointer ${COMMON_WARNING_FLAGS} ${CXX_WARNING_FLAGS}")
set (CMAKE_CXX_FLAGS_RELWITHDEBINFO "${CMAKE_CXX_FLAGS_RELWITHDEBINFO} -O3 ${CMAKE_CXX_FLAGS_ADD}")
set (CMAKE_CXX_FLAGS_DEBUG "${CMAKE_CXX_FLAGS_DEBUG} -O0 -g3 -ggdb3 -fno-inline ${CMAKE_CXX_FLAGS_ADD}")
set (CMAKE_C_FLAGS "${CMAKE_C_FLAGS} ${COMPILER_FLAGS} -fno-omit-frame-pointer ${COMMON_WARNING_FLAGS} ${CMAKE_C_FLAGS_ADD}")
set (CMAKE_C_FLAGS_RELWITHDEBINFO "${CMAKE_C_FLAGS_RELWITHDEBINFO} -O3 ${CMAKE_C_FLAGS_ADD}")
set (CMAKE_C_FLAGS_DEBUG "${CMAKE_C_FLAGS_DEBUG} -O0 -g3 -ggdb3 -fno-inline ${CMAKE_C_FLAGS_ADD}")
option (UNBUNDLED "Try find all libraries in system. We recommend to avoid this mode for production builds, because we cannot guarantee exact versions and variants of libraries your system has installed. This mode exists for enthusiastic developers who search for trouble. Also it is useful for maintainers of OS packages." OFF)
if (UNBUNDLED)
set(NOT_UNBUNDLED 0)
else ()
set(NOT_UNBUNDLED 1)
endif ()
# Using system libs can cause lot of warnings in includes.
if (UNBUNDLED OR NOT (OS_LINUX OR APPLE) OR ARCH_32)
option (NO_WERROR "Disable -Werror compiler option" ON)
endif ()
# Make this extra-checks for correct library dependencies.
if (NOT SANITIZE)
set (CMAKE_EXE_LINKER_FLAGS "${CMAKE_EXE_LINKER_FLAGS} -Wl,--no-undefined")
set (CMAKE_SHARED_LINKER_FLAGS "${CMAKE_SHARED_LINKER_FLAGS} -Wl,--no-undefined")
endif ()
include(cmake/dbms_glob_sources.cmake)
include(cmake/default_libs.cmake)
######################################
### Add targets below this comment ###
######################################
string (TOUPPER ${CMAKE_BUILD_TYPE} CMAKE_BUILD_TYPE_UC)
set (CMAKE_POSTFIX_VARIABLE "CMAKE_${CMAKE_BUILD_TYPE_UC}_POSTFIX")
if (NOT MAKE_STATIC_LIBRARIES)
set(CMAKE_POSITION_INDEPENDENT_CODE ON)
endif ()
# Using "include-what-you-use" tool.
option (USE_INCLUDE_WHAT_YOU_USE "Use 'include-what-you-use' tool" OFF)
if (USE_INCLUDE_WHAT_YOU_USE)
find_program(IWYU_PATH NAMES include-what-you-use iwyu)
if (NOT IWYU_PATH)
message(FATAL_ERROR "Could not find the program include-what-you-use")
endif()
if (${CMAKE_VERSION} VERSION_LESS "3.3.0")
message(FATAL_ERROR "include-what-you-use requires CMake version at least 3.3.")
endif()
endif ()
# Using clang-tidy static analyzer http://mariobadr.com/using-clang-tidy-with-cmake-36.html https://cmake.org/cmake/help/v3.6/prop_tgt/LANG_CLANG_TIDY.html
option (ENABLE_CLANG_TIDY "Use 'clang-tidy' static analyzer" OFF)
if (ENABLE_CLANG_TIDY)
if (${CMAKE_VERSION} VERSION_LESS "3.6.0")
message(FATAL_ERROR "clang-tidy requires CMake version at least 3.6.")
endif()
find_program (CLANG_TIDY_EXE NAMES "clang-tidy" DOC "Path to clang-tidy executable")
if (NOT CLANG_TIDY_EXE)
set (USE_CLANG_TIDY 0)
message (STATUS "clang-tidy not found.")
else ()
set (USE_CLANG_TIDY 1)
message (STATUS "clang-tidy found: ${CLANG_TIDY_EXE}")
set (DO_CLANG_TIDY "${CLANG_TIDY_EXE}" "-checks=*,-clang-analyzer-alpha.*")
# You can enable it within a directory by: set (CMAKE_CXX_CLANG_TIDY "${DO_CLANG_TIDY}")
endif ()
endif ()
if (ENABLE_TESTS)
message (STATUS "Tests are enabled")
endif ()
enable_testing() # Enable for tests without binary
# when installing to /usr - place configs to /etc but for /usr/local place to /usr/local/etc
if (CMAKE_INSTALL_PREFIX STREQUAL "/usr")
set (CLICKHOUSE_ETC_DIR "/etc")
else ()
set (CLICKHOUSE_ETC_DIR "${CMAKE_INSTALL_PREFIX}/etc")
endif ()
option (UNBUNDLED "Try find all libraries in system. We recommend to avoid this mode for production builds, because we cannot guarantee exact versions and variants of libraries your system has installed. This mode exists for enthusiastic developers who search for trouble. Also it is useful for maintainers of OS packages." OFF)
if (UNBUNDLED)
set(NOT_UNBUNDLED 0)
else ()
set(NOT_UNBUNDLED 1)
endif ()
# Using system libs can cause lot of warnings in includes.
if (UNBUNDLED OR NOT (OS_LINUX OR APPLE) OR ARCH_32)
option (NO_WERROR "Disable -Werror compiler option" ON)
endif ()
message (STATUS "Building for: ${CMAKE_SYSTEM} ${CMAKE_SYSTEM_PROCESSOR} ${CMAKE_LIBRARY_ARCHITECTURE} ; USE_STATIC_LIBRARIES=${USE_STATIC_LIBRARIES} MAKE_STATIC_LIBRARIES=${MAKE_STATIC_LIBRARIES} SPLIT_SHARED=${SPLIT_SHARED_LIBRARIES} UNBUNDLED=${UNBUNDLED} CCACHE=${CCACHE_FOUND} ${CCACHE_VERSION}")
include(GNUInstallDirs)
include (cmake/find_contrib_lib.cmake)
include (cmake/lib_name.cmake)
find_contrib_lib(double-conversion) # Must be before parquet
include (cmake/find_ssl.cmake)
include (cmake/find_icu.cmake)
include (cmake/find_boost.cmake)
include (cmake/find_zlib.cmake)
include (cmake/find_zstd.cmake)
include (cmake/find_ltdl.cmake) # for odbc
include (cmake/find_termcap.cmake)
include (cmake/find_odbc.cmake)
# openssl, zlib, odbc before poco
include (cmake/find_poco.cmake)
include (cmake/find_lz4.cmake)
include (cmake/find_xxhash.cmake)
include (cmake/find_sparsehash.cmake)
include (cmake/find_rt.cmake)
include (cmake/find_execinfo.cmake)
include (cmake/find_readline_edit.cmake)
include (cmake/find_re2.cmake)
include (cmake/find_libgsasl.cmake)
include (cmake/find_rdkafka.cmake)
include (cmake/find_capnp.cmake)
include (cmake/find_llvm.cmake)
include (cmake/find_h3.cmake)
include (cmake/find_cpuid.cmake) # Freebsd, bundled
if (NOT USE_CPUID)
include (cmake/find_cpuinfo.cmake) # Debian
endif()
include (cmake/find_libxml2.cmake)
include (cmake/find_brotli.cmake)
include (cmake/find_protobuf.cmake)
include (cmake/find_pdqsort.cmake)
include (cmake/find_hdfs3.cmake) # uses protobuf
include (cmake/find_consistent-hashing.cmake)
include (cmake/find_base64.cmake)
include (cmake/find_parquet.cmake)
include (cmake/find_hyperscan.cmake)
include (cmake/find_simdjson.cmake)
include (cmake/find_rapidjson.cmake)
include (cmake/find_fastops.cmake)
#include (cmake/find_orc.cmake)
find_contrib_lib(cityhash)
find_contrib_lib(farmhash)
find_contrib_lib(metrohash)
find_contrib_lib(btrie)
if (ENABLE_TESTS)
include (cmake/find_gtest.cmake)
endif ()
# Need to process before "contrib" dir:
include (libs/libcommon/cmake/find_gperftools.cmake)
include (libs/libcommon/cmake/find_jemalloc.cmake)
include (libs/libcommon/cmake/find_cctz.cmake)
include (libs/libmysqlxx/cmake/find_mysqlclient.cmake)
include (cmake/print_flags.cmake)
install (EXPORT global DESTINATION cmake)
add_subdirectory (contrib EXCLUDE_FROM_ALL)
add_subdirectory (libs)
add_subdirectory (utils)
add_subdirectory (dbms)
include (cmake/print_include_directories.cmake)
```
| 1.0 | Mac 10.14.4 can not build the clickhouse -
**Operating system**
Mac 10.14.4
**Cmake version**
`cmake --version
3.18.2`
**Ninja version**
`ninja --version
1.10.1`
**Compiler name and version**
`Apple LLVM version 10.0.1 (clang-1001.0.46.4)
Target: x86_64-apple-darwin18.5.0
Thread model: posix
InstalledDir: /Library/Developer/CommandLineTools/usr/bin
`
**Full cmake and/or ninja output**
```-- Looking for pthread.h
-- Looking for pthread.h - not found
CMake Error at /usr/local/Cellar/cmake/3.18.2/share/cmake/Modules/FindPackageHandleStandardArgs.cmake:165 (message):
Could NOT find Threads (missing: Threads_FOUND)
Call Stack (most recent call first):
/usr/local/Cellar/cmake/3.18.2/share/cmake/Modules/FindPackageHandleStandardArgs.cmake:458 (_FPHSA_FAILURE_MESSAGE)
/usr/local/Cellar/cmake/3.18.2/share/cmake/Modules/FindThreads.cmake:234 (FIND_PACKAGE_HANDLE_STANDARD_ARGS)
cmake/default_libs.cmake:30 (find_package)
CMakeLists.txt:234 (include)
-- Configuring incomplete, errors occurred!
See also "/Users/doude/CLionProjects/ClickHouse/build/CMakeFiles/CMakeOutput.log".
See also "/Users/doude/CLionProjects/ClickHouse/build/CMakeFiles/CMakeError.log".
```
**CMakeLists.txt**
```
foreach(policy
CMP0023
CMP0048 # CMake 3.0
CMP0074 # CMake 3.12
CMP0077
CMP0079
)
if(POLICY ${policy})
cmake_policy(SET ${policy} NEW)
endif()
endforeach()
project(ClickHouse)
cmake_minimum_required(VERSION 3.3)
# Ignore export() since we don't use it,
# but it gets broken with a global targets via link_libraries()
macro (export)
endmacro ()
set(CMAKE_MODULE_PATH ${CMAKE_MODULE_PATH} "${CMAKE_CURRENT_SOURCE_DIR}/cmake/Modules/")
set(CMAKE_EXPORT_COMPILE_COMMANDS 1) # Write compile_commands.json
set(CMAKE_LINK_DEPENDS_NO_SHARED 1) # Do not relink all depended targets on .so
set(CMAKE_CONFIGURATION_TYPES "RelWithDebInfo;Debug;Release;MinSizeRel" CACHE STRING "" FORCE)
set(CMAKE_DEBUG_POSTFIX "d" CACHE STRING "Generate debug library name with a postfix.") # To be consistent with CMakeLists from contrib libs.
include (cmake/arch.cmake)
option(ENABLE_IPO "Enable inter-procedural optimization (aka LTO)" OFF) # need cmake 3.9+
if(ENABLE_IPO)
cmake_policy(SET CMP0069 NEW)
include(CheckIPOSupported)
check_ipo_supported(RESULT IPO_SUPPORTED OUTPUT IPO_NOT_SUPPORTED)
if(IPO_SUPPORTED)
message(STATUS "IPO/LTO is supported, enabling")
set(CMAKE_INTERPROCEDURAL_OPTIMIZATION TRUE)
else()
message(STATUS "IPO/LTO is not supported: <${IPO_NOT_SUPPORTED}>")
endif()
else()
message(STATUS "IPO/LTO not enabled.")
endif()
if (COMPILER_GCC)
# Require minimum version of gcc
set (GCC_MINIMUM_VERSION 8)
if (CMAKE_CXX_COMPILER_VERSION VERSION_LESS ${GCC_MINIMUM_VERSION} AND NOT CMAKE_VERSION VERSION_LESS 2.8.9)
message (FATAL_ERROR "GCC version must be at least ${GCC_MINIMUM_VERSION}. For example, if GCC ${GCC_MINIMUM_VERSION} is available under gcc-${GCC_MINIMUM_VERSION}, g++-${GCC_MINIMUM_VERSION} names, do the following: export CC=gcc-${GCC_MINIMUM_VERSION} CXX=g++-${GCC_MINIMUM_VERSION}; rm -rf CMakeCache.txt CMakeFiles; and re run cmake or ./release.")
endif ()
elseif (CMAKE_CXX_COMPILER_ID STREQUAL "Clang")
# Require minimum version of clang
set (CLANG_MINIMUM_VERSION 7)
if (CMAKE_CXX_COMPILER_VERSION VERSION_LESS ${CLANG_MINIMUM_VERSION})
message (FATAL_ERROR "Clang version must be at least ${CLANG_MINIMUM_VERSION}.")
endif ()
else ()
message (WARNING "You are using an unsupported compiler. Compilation has only been tested with Clang 6+ and GCC 7+.")
endif ()
# Check that submodules are present only if source was downloaded with git
if (EXISTS "${CMAKE_CURRENT_SOURCE_DIR}/.git" AND NOT EXISTS "${ClickHouse_SOURCE_DIR}/contrib/boost/boost")
message (FATAL_ERROR "Submodules are not initialized. Run\n\tgit submodule update --init --recursive")
endif ()
include (cmake/find_ccache.cmake)
if (NOT CMAKE_BUILD_TYPE OR CMAKE_BUILD_TYPE STREQUAL "None")
message (STATUS "CMAKE_BUILD_TYPE is not set, set to default = RELWITHDEBINFO")
set (CMAKE_BUILD_TYPE "RELWITHDEBINFO")
endif ()
string(TOUPPER ${CMAKE_BUILD_TYPE} CMAKE_BUILD_TYPE_UC)
message (STATUS "CMAKE_BUILD_TYPE: ${CMAKE_BUILD_TYPE}")
option (USE_STATIC_LIBRARIES "Set to FALSE to use shared libraries" ON)
option (MAKE_STATIC_LIBRARIES "Set to FALSE to make shared libraries" ${USE_STATIC_LIBRARIES})
if (NOT MAKE_STATIC_LIBRARIES)
option (SPLIT_SHARED_LIBRARIES "DEV ONLY. Keep all internal libs as separate .so for faster linking" OFF)
option (CLICKHOUSE_SPLIT_BINARY "Make several binaries instead one bundled (clickhouse-server, clickhouse-client, ... )" OFF)
endif ()
if (MAKE_STATIC_LIBRARIES AND SPLIT_SHARED_LIBRARIES)
message(FATAL_ERROR "Defining SPLIT_SHARED_LIBRARIES=1 without MAKE_STATIC_LIBRARIES=0 has no effect.")
endif()
if (NOT MAKE_STATIC_LIBRARIES AND SPLIT_SHARED_LIBRARIES)
set(BUILD_SHARED_LIBS 1 CACHE INTERNAL "")
endif ()
if (USE_STATIC_LIBRARIES)
list(REVERSE CMAKE_FIND_LIBRARY_SUFFIXES)
endif ()
include (cmake/sanitize.cmake)
if (CMAKE_GENERATOR STREQUAL "Ninja")
# Turn on colored output. https://github.com/ninja-build/ninja/wiki/FAQ
set (CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -fdiagnostics-color=always")
set (CMAKE_C_FLAGS "${CMAKE_C_FLAGS} -fdiagnostics-color=always")
endif ()
if (NOT MSVC)
set (COMMON_WARNING_FLAGS "${COMMON_WARNING_FLAGS} -Wall") # -Werror is also added inside directories with our own code.
endif ()
if (COMPILER_GCC OR COMPILER_CLANG)
set (CXX_WARNING_FLAGS "${CXX_WARNING_FLAGS} -Wnon-virtual-dtor")
endif ()
if (COMPILER_GCC AND CMAKE_CXX_COMPILER_VERSION VERSION_GREATER "8.3.0")
# Warnings in protobuf generating
set (CXX_WARNING_FLAGS "${CXX_WARNING_FLAGS} -Wno-array-bounds")
endif ()
if (COMPILER_CLANG)
# clang: warning: argument unused during compilation: '-specs=/usr/share/dpkg/no-pie-compile.specs' [-Wunused-command-line-argument]
set (COMMON_WARNING_FLAGS "${COMMON_WARNING_FLAGS} -Wno-unused-command-line-argument")
# generate ranges for fast "addr2line" search
if (NOT CMAKE_BUILD_TYPE_UC STREQUAL "RELEASE")
set(COMPILER_FLAGS "${COMPILER_FLAGS} -gdwarf-aranges")
endif ()
endif ()
option (ENABLE_TESTS "Enables tests" ON)
if (CMAKE_SYSTEM_PROCESSOR MATCHES "amd64|x86_64")
option (USE_INTERNAL_MEMCPY "Use internal implementation of 'memcpy' function instead of provided by libc. Only for x86_64." ON)
if (OS_LINUX AND NOT UNBUNDLED AND MAKE_STATIC_LIBRARIES AND NOT SPLIT_SHARED_LIBRARIES AND CMAKE_VERSION VERSION_GREATER "3.9.0")
option (GLIBC_COMPATIBILITY "Set to TRUE to enable compatibility with older glibc libraries. Only for x86_64, Linux. Implies USE_INTERNAL_MEMCPY." ON)
endif ()
if (NOT CMAKE_VERSION VERSION_GREATER "3.9.0")
message (WARNING "CMake version must be greater than 3.9.0 for production builds.")
endif ()
endif ()
string(REGEX MATCH "-?[0-9]+(.[0-9]+)?$" COMPILER_POSTFIX ${CMAKE_CXX_COMPILER})
find_program (LLD_PATH NAMES "lld${COMPILER_POSTFIX}" "lld")
find_program (GOLD_PATH NAMES "gold")
if (COMPILER_CLANG AND LLD_PATH AND NOT LINKER_NAME)
set (LINKER_NAME "lld")
elseif (GOLD_PATH)
set (LINKER_NAME "gold")
endif ()
if (LINKER_NAME)
message(STATUS "Using linker: ${LINKER_NAME} (selected from: LLD_PATH=${LLD_PATH}; GOLD_PATH=${GOLD_PATH}; COMPILER_POSTFIX=${COMPILER_POSTFIX})")
set (CMAKE_EXE_LINKER_FLAGS "${CMAKE_EXE_LINKER_FLAGS} -fuse-ld=${LINKER_NAME}")
endif ()
cmake_host_system_information(RESULT AVAILABLE_PHYSICAL_MEMORY QUERY AVAILABLE_PHYSICAL_MEMORY) # Not available under freebsd
if(NOT AVAILABLE_PHYSICAL_MEMORY OR AVAILABLE_PHYSICAL_MEMORY GREATER 8000)
option(COMPILER_PIPE "-pipe compiler option [less /tmp usage, more ram usage]" ON)
endif()
if(COMPILER_PIPE)
set(COMPILER_FLAGS "${COMPILER_FLAGS} -pipe")
else()
message(STATUS "Disabling compiler -pipe option (have only ${AVAILABLE_PHYSICAL_MEMORY} mb of memory)")
endif()
if(NOT DISABLE_CPU_OPTIMIZE)
include(cmake/test_cpu.cmake)
endif()
if(NOT COMPILER_CLANG) # clang: error: the clang compiler does not support '-march=native'
option(ARCH_NATIVE "Enable -march=native compiler flag" ${ARCH_ARM})
endif()
if (ARCH_NATIVE)
set (COMPILER_FLAGS "${COMPILER_FLAGS} -march=native")
endif ()
if (CMAKE_VERSION VERSION_LESS "3.8.0")
if (NOT MSVC)
set (CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -std=c++17")
endif ()
else ()
set (CMAKE_CXX_STANDARD 17)
set (CMAKE_CXX_EXTENSIONS 0) # https://cmake.org/cmake/help/latest/prop_tgt/CXX_EXTENSIONS.html#prop_tgt:CXX_EXTENSIONS
set (CMAKE_CXX_STANDARD_REQUIRED ON)
set (CXX_FLAGS_INTERNAL_COMPILER "-std=c++17")
endif ()
if (COMPILER_GCC OR COMPILER_CLANG)
# Enable C++14 sized global deallocation functions. It should be enabled by setting -std=c++14 but I'm not sure.
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -fsized-deallocation")
endif ()
option(WITH_COVERAGE "Build with coverage." 0)
if(WITH_COVERAGE AND COMPILER_CLANG)
set(COMPILER_FLAGS "${COMPILER_FLAGS} -fprofile-instr-generate -fcoverage-mapping")
# If we want to disable coverage for specific translation units
set(WITHOUT_COVERAGE "-fno-profile-instr-generate -fno-coverage-mapping")
endif()
if(WITH_COVERAGE AND COMPILER_GCC)
set(COMPILER_FLAGS "${COMPILER_FLAGS} -fprofile-arcs -ftest-coverage")
set(COVERAGE_OPTION "-lgcov")
set(WITHOUT_COVERAGE "-fno-profile-arcs -fno-test-coverage")
endif()
set (CMAKE_BUILD_COLOR_MAKEFILE ON)
set (CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} ${COMPILER_FLAGS} ${PLATFORM_EXTRA_CXX_FLAG} -fno-omit-frame-pointer ${COMMON_WARNING_FLAGS} ${CXX_WARNING_FLAGS}")
set (CMAKE_CXX_FLAGS_RELWITHDEBINFO "${CMAKE_CXX_FLAGS_RELWITHDEBINFO} -O3 ${CMAKE_CXX_FLAGS_ADD}")
set (CMAKE_CXX_FLAGS_DEBUG "${CMAKE_CXX_FLAGS_DEBUG} -O0 -g3 -ggdb3 -fno-inline ${CMAKE_CXX_FLAGS_ADD}")
set (CMAKE_C_FLAGS "${CMAKE_C_FLAGS} ${COMPILER_FLAGS} -fno-omit-frame-pointer ${COMMON_WARNING_FLAGS} ${CMAKE_C_FLAGS_ADD}")
set (CMAKE_C_FLAGS_RELWITHDEBINFO "${CMAKE_C_FLAGS_RELWITHDEBINFO} -O3 ${CMAKE_C_FLAGS_ADD}")
set (CMAKE_C_FLAGS_DEBUG "${CMAKE_C_FLAGS_DEBUG} -O0 -g3 -ggdb3 -fno-inline ${CMAKE_C_FLAGS_ADD}")
option (UNBUNDLED "Try find all libraries in system. We recommend to avoid this mode for production builds, because we cannot guarantee exact versions and variants of libraries your system has installed. This mode exists for enthusiastic developers who search for trouble. Also it is useful for maintainers of OS packages." OFF)
if (UNBUNDLED)
set(NOT_UNBUNDLED 0)
else ()
set(NOT_UNBUNDLED 1)
endif ()
# Using system libs can cause lot of warnings in includes.
if (UNBUNDLED OR NOT (OS_LINUX OR APPLE) OR ARCH_32)
option (NO_WERROR "Disable -Werror compiler option" ON)
endif ()
# Make this extra-checks for correct library dependencies.
if (NOT SANITIZE)
set (CMAKE_EXE_LINKER_FLAGS "${CMAKE_EXE_LINKER_FLAGS} -Wl,--no-undefined")
set (CMAKE_SHARED_LINKER_FLAGS "${CMAKE_SHARED_LINKER_FLAGS} -Wl,--no-undefined")
endif ()
include(cmake/dbms_glob_sources.cmake)
include(cmake/default_libs.cmake)
######################################
### Add targets below this comment ###
######################################
string (TOUPPER ${CMAKE_BUILD_TYPE} CMAKE_BUILD_TYPE_UC)
set (CMAKE_POSTFIX_VARIABLE "CMAKE_${CMAKE_BUILD_TYPE_UC}_POSTFIX")
if (NOT MAKE_STATIC_LIBRARIES)
set(CMAKE_POSITION_INDEPENDENT_CODE ON)
endif ()
# Using "include-what-you-use" tool.
option (USE_INCLUDE_WHAT_YOU_USE "Use 'include-what-you-use' tool" OFF)
if (USE_INCLUDE_WHAT_YOU_USE)
find_program(IWYU_PATH NAMES include-what-you-use iwyu)
if (NOT IWYU_PATH)
message(FATAL_ERROR "Could not find the program include-what-you-use")
endif()
if (${CMAKE_VERSION} VERSION_LESS "3.3.0")
message(FATAL_ERROR "include-what-you-use requires CMake version at least 3.3.")
endif()
endif ()
# Using clang-tidy static analyzer http://mariobadr.com/using-clang-tidy-with-cmake-36.html https://cmake.org/cmake/help/v3.6/prop_tgt/LANG_CLANG_TIDY.html
option (ENABLE_CLANG_TIDY "Use 'clang-tidy' static analyzer" OFF)
if (ENABLE_CLANG_TIDY)
if (${CMAKE_VERSION} VERSION_LESS "3.6.0")
message(FATAL_ERROR "clang-tidy requires CMake version at least 3.6.")
endif()
find_program (CLANG_TIDY_EXE NAMES "clang-tidy" DOC "Path to clang-tidy executable")
if (NOT CLANG_TIDY_EXE)
set (USE_CLANG_TIDY 0)
message (STATUS "clang-tidy not found.")
else ()
set (USE_CLANG_TIDY 1)
message (STATUS "clang-tidy found: ${CLANG_TIDY_EXE}")
set (DO_CLANG_TIDY "${CLANG_TIDY_EXE}" "-checks=*,-clang-analyzer-alpha.*")
# You can enable it within a directory by: set (CMAKE_CXX_CLANG_TIDY "${DO_CLANG_TIDY}")
endif ()
endif ()
if (ENABLE_TESTS)
message (STATUS "Tests are enabled")
endif ()
enable_testing() # Enable for tests without binary
# when installing to /usr - place configs to /etc but for /usr/local place to /usr/local/etc
if (CMAKE_INSTALL_PREFIX STREQUAL "/usr")
set (CLICKHOUSE_ETC_DIR "/etc")
else ()
set (CLICKHOUSE_ETC_DIR "${CMAKE_INSTALL_PREFIX}/etc")
endif ()
option (UNBUNDLED "Try find all libraries in system. We recommend to avoid this mode for production builds, because we cannot guarantee exact versions and variants of libraries your system has installed. This mode exists for enthusiastic developers who search for trouble. Also it is useful for maintainers of OS packages." OFF)
if (UNBUNDLED)
set(NOT_UNBUNDLED 0)
else ()
set(NOT_UNBUNDLED 1)
endif ()
# Using system libs can cause lot of warnings in includes.
if (UNBUNDLED OR NOT (OS_LINUX OR APPLE) OR ARCH_32)
option (NO_WERROR "Disable -Werror compiler option" ON)
endif ()
message (STATUS "Building for: ${CMAKE_SYSTEM} ${CMAKE_SYSTEM_PROCESSOR} ${CMAKE_LIBRARY_ARCHITECTURE} ; USE_STATIC_LIBRARIES=${USE_STATIC_LIBRARIES} MAKE_STATIC_LIBRARIES=${MAKE_STATIC_LIBRARIES} SPLIT_SHARED=${SPLIT_SHARED_LIBRARIES} UNBUNDLED=${UNBUNDLED} CCACHE=${CCACHE_FOUND} ${CCACHE_VERSION}")
include(GNUInstallDirs)
include (cmake/find_contrib_lib.cmake)
include (cmake/lib_name.cmake)
find_contrib_lib(double-conversion) # Must be before parquet
include (cmake/find_ssl.cmake)
include (cmake/find_icu.cmake)
include (cmake/find_boost.cmake)
include (cmake/find_zlib.cmake)
include (cmake/find_zstd.cmake)
include (cmake/find_ltdl.cmake) # for odbc
include (cmake/find_termcap.cmake)
include (cmake/find_odbc.cmake)
# openssl, zlib, odbc before poco
include (cmake/find_poco.cmake)
include (cmake/find_lz4.cmake)
include (cmake/find_xxhash.cmake)
include (cmake/find_sparsehash.cmake)
include (cmake/find_rt.cmake)
include (cmake/find_execinfo.cmake)
include (cmake/find_readline_edit.cmake)
include (cmake/find_re2.cmake)
include (cmake/find_libgsasl.cmake)
include (cmake/find_rdkafka.cmake)
include (cmake/find_capnp.cmake)
include (cmake/find_llvm.cmake)
include (cmake/find_h3.cmake)
include (cmake/find_cpuid.cmake) # Freebsd, bundled
if (NOT USE_CPUID)
include (cmake/find_cpuinfo.cmake) # Debian
endif()
include (cmake/find_libxml2.cmake)
include (cmake/find_brotli.cmake)
include (cmake/find_protobuf.cmake)
include (cmake/find_pdqsort.cmake)
include (cmake/find_hdfs3.cmake) # uses protobuf
include (cmake/find_consistent-hashing.cmake)
include (cmake/find_base64.cmake)
include (cmake/find_parquet.cmake)
include (cmake/find_hyperscan.cmake)
include (cmake/find_simdjson.cmake)
include (cmake/find_rapidjson.cmake)
include (cmake/find_fastops.cmake)
#include (cmake/find_orc.cmake)
find_contrib_lib(cityhash)
find_contrib_lib(farmhash)
find_contrib_lib(metrohash)
find_contrib_lib(btrie)
if (ENABLE_TESTS)
include (cmake/find_gtest.cmake)
endif ()
# Need to process before "contrib" dir:
include (libs/libcommon/cmake/find_gperftools.cmake)
include (libs/libcommon/cmake/find_jemalloc.cmake)
include (libs/libcommon/cmake/find_cctz.cmake)
include (libs/libmysqlxx/cmake/find_mysqlclient.cmake)
include (cmake/print_flags.cmake)
install (EXPORT global DESTINATION cmake)
add_subdirectory (contrib EXCLUDE_FROM_ALL)
add_subdirectory (libs)
add_subdirectory (utils)
add_subdirectory (dbms)
include (cmake/print_include_directories.cmake)
```
| non_priority | mac can not build the clickhouse operating system mac cmake version cmake version ninja version ninja version compiler name and version apple llvm version clang target apple thread model posix installeddir library developer commandlinetools usr bin full cmake and or ninja output looking for pthread h looking for pthread h not found cmake error at usr local cellar cmake share cmake modules findpackagehandlestandardargs cmake message could not find threads missing threads found call stack most recent call first usr local cellar cmake share cmake modules findpackagehandlestandardargs cmake fphsa failure message usr local cellar cmake share cmake modules findthreads cmake find package handle standard args cmake default libs cmake find package cmakelists txt include configuring incomplete errors occurred see also users doude clionprojects clickhouse build cmakefiles cmakeoutput log see also users doude clionprojects clickhouse build cmakefiles cmakeerror log cmakelists txt foreach policy cmake cmake if policy policy cmake policy set policy new endif endforeach project clickhouse cmake minimum required version ignore export since we don t use it but it gets broken with a global targets via link libraries macro export endmacro set cmake module path cmake module path cmake current source dir cmake modules set cmake export compile commands write compile commands json set cmake link depends no shared do not relink all depended targets on so set cmake configuration types relwithdebinfo debug release minsizerel cache string force set cmake debug postfix d cache string generate debug library name with a postfix to be consistent with cmakelists from contrib libs include cmake arch cmake option enable ipo enable inter procedural optimization aka lto off need cmake if enable ipo cmake policy set new include checkiposupported check ipo supported result ipo supported output ipo not supported if ipo supported message status ipo lto is supported enabling set cmake interprocedural optimization true else message status ipo lto is not supported endif else message status ipo lto not enabled endif if compiler gcc require minimum version of gcc set gcc minimum version if cmake cxx compiler version version less gcc minimum version and not cmake version version less message fatal error gcc version must be at least gcc minimum version for example if gcc gcc minimum version is available under gcc gcc minimum version g gcc minimum version names do the following export cc gcc gcc minimum version cxx g gcc minimum version rm rf cmakecache txt cmakefiles and re run cmake or release endif elseif cmake cxx compiler id strequal clang require minimum version of clang set clang minimum version if cmake cxx compiler version version less clang minimum version message fatal error clang version must be at least clang minimum version endif else message warning you are using an unsupported compiler compilation has only been tested with clang and gcc endif check that submodules are present only if source was downloaded with git if exists cmake current source dir git and not exists clickhouse source dir contrib boost boost message fatal error submodules are not initialized run n tgit submodule update init recursive endif include cmake find ccache cmake if not cmake build type or cmake build type strequal none message status cmake build type is not set set to default relwithdebinfo set cmake build type relwithdebinfo endif string toupper cmake build type cmake build type uc message status cmake build type cmake build type option use static libraries set to false to use shared libraries on option make static libraries set to false to make shared libraries use static libraries if not make static libraries option split shared libraries dev only keep all internal libs as separate so for faster linking off option clickhouse split binary make several binaries instead one bundled clickhouse server clickhouse client off endif if make static libraries and split shared libraries message fatal error defining split shared libraries without make static libraries has no effect endif if not make static libraries and split shared libraries set build shared libs cache internal endif if use static libraries list reverse cmake find library suffixes endif include cmake sanitize cmake if cmake generator strequal ninja turn on colored output set cmake cxx flags cmake cxx flags fdiagnostics color always set cmake c flags cmake c flags fdiagnostics color always endif if not msvc set common warning flags common warning flags wall werror is also added inside directories with our own code endif if compiler gcc or compiler clang set cxx warning flags cxx warning flags wnon virtual dtor endif if compiler gcc and cmake cxx compiler version version greater warnings in protobuf generating set cxx warning flags cxx warning flags wno array bounds endif if compiler clang clang warning argument unused during compilation specs usr share dpkg no pie compile specs set common warning flags common warning flags wno unused command line argument generate ranges for fast search if not cmake build type uc strequal release set compiler flags compiler flags gdwarf aranges endif endif option enable tests enables tests on if cmake system processor matches option use internal memcpy use internal implementation of memcpy function instead of provided by libc only for on if os linux and not unbundled and make static libraries and not split shared libraries and cmake version version greater option glibc compatibility set to true to enable compatibility with older glibc libraries only for linux implies use internal memcpy on endif if not cmake version version greater message warning cmake version must be greater than for production builds endif endif string regex match compiler postfix cmake cxx compiler find program lld path names lld compiler postfix lld find program gold path names gold if compiler clang and lld path and not linker name set linker name lld elseif gold path set linker name gold endif if linker name message status using linker linker name selected from lld path lld path gold path gold path compiler postfix compiler postfix set cmake exe linker flags cmake exe linker flags fuse ld linker name endif cmake host system information result available physical memory query available physical memory not available under freebsd if not available physical memory or available physical memory greater option compiler pipe pipe compiler option on endif if compiler pipe set compiler flags compiler flags pipe else message status disabling compiler pipe option have only available physical memory mb of memory endif if not disable cpu optimize include cmake test cpu cmake endif if not compiler clang clang error the clang compiler does not support march native option arch native enable march native compiler flag arch arm endif if arch native set compiler flags compiler flags march native endif if cmake version version less if not msvc set cmake cxx flags cmake cxx flags std c endif else set cmake cxx standard set cmake cxx extensions set cmake cxx standard required on set cxx flags internal compiler std c endif if compiler gcc or compiler clang enable c sized global deallocation functions it should be enabled by setting std c but i m not sure set cmake cxx flags cmake cxx flags fsized deallocation endif option with coverage build with coverage if with coverage and compiler clang set compiler flags compiler flags fprofile instr generate fcoverage mapping if we want to disable coverage for specific translation units set without coverage fno profile instr generate fno coverage mapping endif if with coverage and compiler gcc set compiler flags compiler flags fprofile arcs ftest coverage set coverage option lgcov set without coverage fno profile arcs fno test coverage endif set cmake build color makefile on set cmake cxx flags cmake cxx flags compiler flags platform extra cxx flag fno omit frame pointer common warning flags cxx warning flags set cmake cxx flags relwithdebinfo cmake cxx flags relwithdebinfo cmake cxx flags add set cmake cxx flags debug cmake cxx flags debug fno inline cmake cxx flags add set cmake c flags cmake c flags compiler flags fno omit frame pointer common warning flags cmake c flags add set cmake c flags relwithdebinfo cmake c flags relwithdebinfo cmake c flags add set cmake c flags debug cmake c flags debug fno inline cmake c flags add option unbundled try find all libraries in system we recommend to avoid this mode for production builds because we cannot guarantee exact versions and variants of libraries your system has installed this mode exists for enthusiastic developers who search for trouble also it is useful for maintainers of os packages off if unbundled set not unbundled else set not unbundled endif using system libs can cause lot of warnings in includes if unbundled or not os linux or apple or arch option no werror disable werror compiler option on endif make this extra checks for correct library dependencies if not sanitize set cmake exe linker flags cmake exe linker flags wl no undefined set cmake shared linker flags cmake shared linker flags wl no undefined endif include cmake dbms glob sources cmake include cmake default libs cmake add targets below this comment string toupper cmake build type cmake build type uc set cmake postfix variable cmake cmake build type uc postfix if not make static libraries set cmake position independent code on endif using include what you use tool option use include what you use use include what you use tool off if use include what you use find program iwyu path names include what you use iwyu if not iwyu path message fatal error could not find the program include what you use endif if cmake version version less message fatal error include what you use requires cmake version at least endif endif using clang tidy static analyzer option enable clang tidy use clang tidy static analyzer off if enable clang tidy if cmake version version less message fatal error clang tidy requires cmake version at least endif find program clang tidy exe names clang tidy doc path to clang tidy executable if not clang tidy exe set use clang tidy message status clang tidy not found else set use clang tidy message status clang tidy found clang tidy exe set do clang tidy clang tidy exe checks clang analyzer alpha you can enable it within a directory by set cmake cxx clang tidy do clang tidy endif endif if enable tests message status tests are enabled endif enable testing enable for tests without binary when installing to usr place configs to etc but for usr local place to usr local etc if cmake install prefix strequal usr set clickhouse etc dir etc else set clickhouse etc dir cmake install prefix etc endif option unbundled try find all libraries in system we recommend to avoid this mode for production builds because we cannot guarantee exact versions and variants of libraries your system has installed this mode exists for enthusiastic developers who search for trouble also it is useful for maintainers of os packages off if unbundled set not unbundled else set not unbundled endif using system libs can cause lot of warnings in includes if unbundled or not os linux or apple or arch option no werror disable werror compiler option on endif message status building for cmake system cmake system processor cmake library architecture use static libraries use static libraries make static libraries make static libraries split shared split shared libraries unbundled unbundled ccache ccache found ccache version include gnuinstalldirs include cmake find contrib lib cmake include cmake lib name cmake find contrib lib double conversion must be before parquet include cmake find ssl cmake include cmake find icu cmake include cmake find boost cmake include cmake find zlib cmake include cmake find zstd cmake include cmake find ltdl cmake for odbc include cmake find termcap cmake include cmake find odbc cmake openssl zlib odbc before poco include cmake find poco cmake include cmake find cmake include cmake find xxhash cmake include cmake find sparsehash cmake include cmake find rt cmake include cmake find execinfo cmake include cmake find readline edit cmake include cmake find cmake include cmake find libgsasl cmake include cmake find rdkafka cmake include cmake find capnp cmake include cmake find llvm cmake include cmake find cmake include cmake find cpuid cmake freebsd bundled if not use cpuid include cmake find cpuinfo cmake debian endif include cmake find cmake include cmake find brotli cmake include cmake find protobuf cmake include cmake find pdqsort cmake include cmake find cmake uses protobuf include cmake find consistent hashing cmake include cmake find cmake include cmake find parquet cmake include cmake find hyperscan cmake include cmake find simdjson cmake include cmake find rapidjson cmake include cmake find fastops cmake include cmake find orc cmake find contrib lib cityhash find contrib lib farmhash find contrib lib metrohash find contrib lib btrie if enable tests include cmake find gtest cmake endif need to process before contrib dir include libs libcommon cmake find gperftools cmake include libs libcommon cmake find jemalloc cmake include libs libcommon cmake find cctz cmake include libs libmysqlxx cmake find mysqlclient cmake include cmake print flags cmake install export global destination cmake add subdirectory contrib exclude from all add subdirectory libs add subdirectory utils add subdirectory dbms include cmake print include directories cmake | 0 |
108,867 | 13,674,770,178 | IssuesEvent | 2020-09-29 11:45:43 | nextcloud/text | https://api.github.com/repos/nextcloud/text | closed | Issue with Link Tooltop in Firefox | 1. to develop bug design good first issue | NextCloud version: 17.0.2
Text version: 1.1.1
Firefox: 72.0.1
I can't add links to text in Firefox. This is what it looks like. When I try to click the tooltip it just goes away.
<img width="631" alt="Screen Shot 2020-01-21 at 1 11 10 PM" src="https://user-images.githubusercontent.com/590678/72839538-b251a700-3c4f-11ea-9718-783f1a862180.png">
As a side note, it would also be nice to have a source editing option in the browser, not just wysiwyg.
| 1.0 | Issue with Link Tooltop in Firefox - NextCloud version: 17.0.2
Text version: 1.1.1
Firefox: 72.0.1
I can't add links to text in Firefox. This is what it looks like. When I try to click the tooltip it just goes away.
<img width="631" alt="Screen Shot 2020-01-21 at 1 11 10 PM" src="https://user-images.githubusercontent.com/590678/72839538-b251a700-3c4f-11ea-9718-783f1a862180.png">
As a side note, it would also be nice to have a source editing option in the browser, not just wysiwyg.
| non_priority | issue with link tooltop in firefox nextcloud version text version firefox i can t add links to text in firefox this is what it looks like when i try to click the tooltip it just goes away img width alt screen shot at pm src as a side note it would also be nice to have a source editing option in the browser not just wysiwyg | 0 |
366,354 | 25,578,994,368 | IssuesEvent | 2022-12-01 01:42:05 | containerd/nerdctl | https://api.github.com/repos/containerd/nerdctl | opened | Set up VEX to eliminates false-positives from vulnerability scanning tools | documentation area/project expert | e.g., https://github.com/CycloneDX/bom-examples/tree/master/VEX/Use-Cases/Case-4
For false positives like:
- #1571 | 1.0 | Set up VEX to eliminates false-positives from vulnerability scanning tools - e.g., https://github.com/CycloneDX/bom-examples/tree/master/VEX/Use-Cases/Case-4
For false positives like:
- #1571 | non_priority | set up vex to eliminates false positives from vulnerability scanning tools e g for false positives like | 0 |
13,634 | 3,172,589,191 | IssuesEvent | 2015-09-23 09:13:18 | owncloud/core | https://api.github.com/repos/owncloud/core | opened | Preview in sidebare hides close button | bug design feature:sidebar | The preview in the sidebar hides th (X) at the top right to close the sidebar which makes it impossible to close the sidebar.

cc @jancborchardt @PVince81 | 1.0 | Preview in sidebare hides close button - The preview in the sidebar hides th (X) at the top right to close the sidebar which makes it impossible to close the sidebar.

cc @jancborchardt @PVince81 | non_priority | preview in sidebare hides close button the preview in the sidebar hides th x at the top right to close the sidebar which makes it impossible to close the sidebar cc jancborchardt | 0 |
425,521 | 29,484,296,096 | IssuesEvent | 2023-06-02 08:33:39 | aisland-dao/dex | https://api.github.com/repos/aisland-dao/dex | closed | Smart Contracts | documentation | Publish the smart contracts used from the protocols and the auditing documents. | 1.0 | Smart Contracts - Publish the smart contracts used from the protocols and the auditing documents. | non_priority | smart contracts publish the smart contracts used from the protocols and the auditing documents | 0 |
70,750 | 15,105,867,480 | IssuesEvent | 2021-02-08 13:36:53 | rammatzkvosky/epsagon-node | https://api.github.com/repos/rammatzkvosky/epsagon-node | opened | CVE-2019-14939 (Medium) detected in mysql-2.16.0.tgz | security vulnerability | ## CVE-2019-14939 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-2.16.0.tgz</b></p></summary>
<p>A node.js driver for mysql. It is written in JavaScript, does not require compiling, and is 100% MIT licensed.</p>
<p>Library home page: <a href="https://registry.npmjs.org/mysql/-/mysql-2.16.0.tgz">https://registry.npmjs.org/mysql/-/mysql-2.16.0.tgz</a></p>
<p>Path to dependency file: epsagon-node/package.json</p>
<p>Path to vulnerable library: epsagon-node/node_modules/mysql/package.json</p>
<p>
Dependency Hierarchy:
- :x: **mysql-2.16.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/epsagon-node/commit/7bad2838d8ab86c5c41e847c8eee2086c7d69f83">7bad2838d8ab86c5c41e847c8eee2086c7d69f83</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the mysql (aka mysqljs) module 2.17.1 for Node.js. The LOAD DATA LOCAL INFILE option is open by default.
<p>Publish Date: 2019-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14939>CVE-2019-14939</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mysql","packageVersion":"2.16.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"mysql:2.16.0","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14939","vulnerabilityDetails":"An issue was discovered in the mysql (aka mysqljs) module 2.17.1 for Node.js. The LOAD DATA LOCAL INFILE option is open by default.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14939","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-14939 (Medium) detected in mysql-2.16.0.tgz - ## CVE-2019-14939 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mysql-2.16.0.tgz</b></p></summary>
<p>A node.js driver for mysql. It is written in JavaScript, does not require compiling, and is 100% MIT licensed.</p>
<p>Library home page: <a href="https://registry.npmjs.org/mysql/-/mysql-2.16.0.tgz">https://registry.npmjs.org/mysql/-/mysql-2.16.0.tgz</a></p>
<p>Path to dependency file: epsagon-node/package.json</p>
<p>Path to vulnerable library: epsagon-node/node_modules/mysql/package.json</p>
<p>
Dependency Hierarchy:
- :x: **mysql-2.16.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/rammatzkvosky/epsagon-node/commit/7bad2838d8ab86c5c41e847c8eee2086c7d69f83">7bad2838d8ab86c5c41e847c8eee2086c7d69f83</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the mysql (aka mysqljs) module 2.17.1 for Node.js. The LOAD DATA LOCAL INFILE option is open by default.
<p>Publish Date: 2019-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14939>CVE-2019-14939</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mysql","packageVersion":"2.16.0","packageFilePaths":["/package.json"],"isTransitiveDependency":false,"dependencyTree":"mysql:2.16.0","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14939","vulnerabilityDetails":"An issue was discovered in the mysql (aka mysqljs) module 2.17.1 for Node.js. The LOAD DATA LOCAL INFILE option is open by default.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14939","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"None","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in mysql tgz cve medium severity vulnerability vulnerable library mysql tgz a node js driver for mysql it is written in javascript does not require compiling and is mit licensed library home page a href path to dependency file epsagon node package json path to vulnerable library epsagon node node modules mysql package json dependency hierarchy x mysql tgz vulnerable library found in head commit a href found in base branch master vulnerability details an issue was discovered in the mysql aka mysqljs module for node js the load data local infile option is open by default publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree mysql isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in the mysql aka mysqljs module for node js the load data local infile option is open by default vulnerabilityurl | 0 |
14,102 | 10,605,426,652 | IssuesEvent | 2019-10-10 20:26:53 | earthlab/abc-classroom | https://api.github.com/repos/earthlab/abc-classroom | reopened | Add Changelog to abc-classroom | infrastructure | Currently we don't have a good way to track updates and versions. add a changelog following earthpy and matplotcheck to do this!!
Also update the contributing file to document protocol!! | 1.0 | Add Changelog to abc-classroom - Currently we don't have a good way to track updates and versions. add a changelog following earthpy and matplotcheck to do this!!
Also update the contributing file to document protocol!! | non_priority | add changelog to abc classroom currently we don t have a good way to track updates and versions add a changelog following earthpy and matplotcheck to do this also update the contributing file to document protocol | 0 |
245,373 | 18,781,118,781 | IssuesEvent | 2021-11-08 06:50:03 | Azubis-Dataport-Magdeburg/dp_ausb_md | https://api.github.com/repos/Azubis-Dataport-Magdeburg/dp_ausb_md | closed | Webhooks für restliche GitHub-Repositories einrichten - Discord | documentation good first issue | Kanäle + Webooks in Discord einrichten als extra Kanäle | 1.0 | Webhooks für restliche GitHub-Repositories einrichten - Discord - Kanäle + Webooks in Discord einrichten als extra Kanäle | non_priority | webhooks für restliche github repositories einrichten discord kanäle webooks in discord einrichten als extra kanäle | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.