Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 1 744 | labels stringlengths 4 574 | body stringlengths 9 211k | index stringclasses 10 values | text_combine stringlengths 96 211k | label stringclasses 2 values | text stringlengths 96 188k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
21,435 | 3,508,924,060 | IssuesEvent | 2016-01-08 20:08:23 | aleph7/Upsurge | https://api.github.com/repos/aleph7/Upsurge | closed | LinearType should be a SequenceType | t1: defect | There should be a test case for iterating through elements:
```swift
let elements: LinearType = RealArray([0, 1, 2, 3])
for (index, element) in elements.enumerate() {
XCTAssertEqual(index, element)
}
``` | 1.0 | LinearType should be a SequenceType - There should be a test case for iterating through elements:
```swift
let elements: LinearType = RealArray([0, 1, 2, 3])
for (index, element) in elements.enumerate() {
XCTAssertEqual(index, element)
}
``` | non_process | lineartype should be a sequencetype there should be a test case for iterating through elements swift let elements lineartype realarray for index element in elements enumerate xctassertequal index element | 0 |
690,054 | 23,644,410,986 | IssuesEvent | 2022-08-25 20:24:31 | WarwickAI/wai-platform-v2 | https://api.github.com/repos/WarwickAI/wai-platform-v2 | closed | Add date field to events | enhancement high-priority | For Tutorials and Talks, this will be the date of the event.
For Courses and Project, this will be the starting date of the event.
Can be shown in the `extraInfo` section on the cards, and beneath the heading on the page. | 1.0 | Add date field to events - For Tutorials and Talks, this will be the date of the event.
For Courses and Project, this will be the starting date of the event.
Can be shown in the `extraInfo` section on the cards, and beneath the heading on the page. | non_process | add date field to events for tutorials and talks this will be the date of the event for courses and project this will be the starting date of the event can be shown in the extrainfo section on the cards and beneath the heading on the page | 0 |
30,908 | 4,668,582,161 | IssuesEvent | 2016-10-06 03:24:05 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | opened | kubernetes-e2e-gci-gce-alpha-features-release-1.4: broken test run | kind/flake priority/P2 team/test-infra | Failed: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gci-gce-alpha-features-release-1.4/-1/
Run so broken it didn't make JUnit output! | 1.0 | kubernetes-e2e-gci-gce-alpha-features-release-1.4: broken test run - Failed: https://k8s-gubernator.appspot.com/build/kubernetes-jenkins/logs/kubernetes-e2e-gci-gce-alpha-features-release-1.4/-1/
Run so broken it didn't make JUnit output! | non_process | kubernetes gci gce alpha features release broken test run failed run so broken it didn t make junit output | 0 |
16,322 | 20,976,637,976 | IssuesEvent | 2022-03-28 15:45:53 | Automattic/woocommerce-payments | https://api.github.com/repos/Automattic/woocommerce-payments | closed | Documents view: document download | category: projects component: invoices for processing fees | ### Description
<!-- A clear and concise description of what the new feature or improvement is. -->
As part of the VAT invoices for processing fees project, merchants should be able to download any documents they see on the Documents list.
### Acceptance criteria
<!-- A list of predefined requirements that must be met in order to mark the issue as complete. -->
* Clicking the download button in the documents list downloads the document from WCPay's server
### Designs
<!-- If applicable, add screenshots or links to the UI design for this feature. -->
Designs can be checked in paJDYF-3mi-p2 and more details around scope are available in paJDYF-3sv-p2.
### Dev notes
<!-- If applicable, additional technical or implementation details that will help when developing this feature or improvement. -->
Since files will be served directly, we might need to create a new request helper that will actually redirect to the WCPay's server API instead of making an API request.
| 1.0 | Documents view: document download - ### Description
<!-- A clear and concise description of what the new feature or improvement is. -->
As part of the VAT invoices for processing fees project, merchants should be able to download any documents they see on the Documents list.
### Acceptance criteria
<!-- A list of predefined requirements that must be met in order to mark the issue as complete. -->
* Clicking the download button in the documents list downloads the document from WCPay's server
### Designs
<!-- If applicable, add screenshots or links to the UI design for this feature. -->
Designs can be checked in paJDYF-3mi-p2 and more details around scope are available in paJDYF-3sv-p2.
### Dev notes
<!-- If applicable, additional technical or implementation details that will help when developing this feature or improvement. -->
Since files will be served directly, we might need to create a new request helper that will actually redirect to the WCPay's server API instead of making an API request.
| process | documents view document download description as part of the vat invoices for processing fees project merchants should be able to download any documents they see on the documents list acceptance criteria clicking the download button in the documents list downloads the document from wcpay s server designs designs can be checked in pajdyf and more details around scope are available in pajdyf dev notes since files will be served directly we might need to create a new request helper that will actually redirect to the wcpay s server api instead of making an api request | 1 |
62,705 | 15,343,148,402 | IssuesEvent | 2021-02-27 19:04:31 | nunit/nunit-console | https://api.github.com/repos/nunit/nunit-console | closed | Implement Package tests for engine and console runner | is:build pri:high | This will initially be a set of tests implemented in the cake script itself. Since the script is c#, we can port to a test assembly later if we want to test more deeply. However, I've been testing this way in three or four projects and haven't seen any big limitations so far. | 1.0 | Implement Package tests for engine and console runner - This will initially be a set of tests implemented in the cake script itself. Since the script is c#, we can port to a test assembly later if we want to test more deeply. However, I've been testing this way in three or four projects and haven't seen any big limitations so far. | non_process | implement package tests for engine and console runner this will initially be a set of tests implemented in the cake script itself since the script is c we can port to a test assembly later if we want to test more deeply however i ve been testing this way in three or four projects and haven t seen any big limitations so far | 0 |
211,435 | 23,825,600,641 | IssuesEvent | 2022-09-05 14:37:35 | sast-automation-dev/soundcloud-redux-11 | https://api.github.com/repos/sast-automation-dev/soundcloud-redux-11 | opened | react-scripts-1.0.10.tgz: 77 vulnerabilities (highest severity is: 10.0) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>react-scripts-1.0.10.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2018-3774](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3774) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 10.0 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [WS-2018-0113](https://hackerone.com/reports/319467) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 10.0 | macaddress-0.2.8.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-23369](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2019-19919](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19919) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-13797](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-13797) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | macaddress-0.2.8.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-1000620](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | cryptiles-2.0.5.tgz | Transitive | 1.1.1 | ❌ |
| [CVE-2020-28499](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | merge-1.2.0.tgz | Transitive | 3.0.0 | ✅ |
| [WS-2020-0344](https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | is-my-json-valid-2.16.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-23383](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23383) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-3750](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3750) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | deep-extend-0.4.2.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-16492](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | extend-3.0.1.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-6342](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-6342) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | react-dev-utils-3.0.2.tgz | Transitive | N/A | ❌ |
| [CVE-2019-10744](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | lodash.template-4.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-3728](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.8 | hoek-2.16.3.tgz | Transitive | 1.1.1 | ❌ |
| [WS-2019-0333](https://github.com/wycats/handlebars.js/commit/f7f05d7558e674856686b62a00cde5758f3b7a08) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2019-20920](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20920) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0063](https://github.com/nodeca/js-yaml/pull/480) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | js-yaml-3.7.0.tgz | Transitive | 2.0.0 | ✅ |
| [WS-2018-0084](https://github.com/joyent/node-sshpk/blob/v1.13.1/lib/formats/ssh.js#L17) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.0 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2020-13822](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13822) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.7 | elliptic-6.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2018-0588](https://github.com/unshiftio/querystringify/pull/19) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.6 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2017-15010](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-15010) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | tough-cookie-2.3.2.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0492](https://github.com/handlebars-lang/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-3737](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2018-16469](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16469) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | merge-1.2.0.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0493](https://github.com/handlebars-lang/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2017-16099](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16099) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | no-case-2.3.1.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0318](https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-33587](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | css-what-2.1.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-14732](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14732) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | webpack-dev-server-2.5.0.tgz | Transitive | 2.0.0 | ✅ |
| [CVE-2021-23343](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | path-parse-1.0.5.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7662](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7662) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | websocket-extensions-0.1.1.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2019-20922](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20922) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-20834](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20834) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | tar-2.2.1.tgz | Transitive | 1.0.11 | ❌ |
| [CVE-2021-33502](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | normalize-url-1.9.1.tgz | Transitive | 5.0.0 | ✅ |
| [CVE-2021-33623](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-newlines-1.0.0.tgz | Transitive | 2.0.1 | ✅ |
| [WS-2018-0069](https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | is-my-json-valid-2.16.0.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2021-0152](https://github.com/Qix-/color-string/commit/0789e21284c33d89ebc4ab4ca6f759b9375ac9d3) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | color-string-0.3.0.tgz | Transitive | 2.0.0 | ✅ |
| [WS-2020-0218](https://github.com/yeikos/js.merge/pull/38) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | merge-1.2.0.tgz | Transitive | 3.0.0 | ✅ |
| [WS-2019-0032](https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | js-yaml-3.7.0.tgz | Transitive | 2.0.0 | ✅ |
| [CVE-2021-27516](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27516) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | urijs-1.18.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-28092](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | is-svg-2.1.0.tgz | Transitive | 2.0.0 | ✅ |
| [WS-2020-0091](https://github.com/http-party/node-http-proxy/pull/1447) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | http-proxy-1.16.2.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0541](https://github.com/scravy/node-macaddress/commit/ca9e24df906c9066d49fba658e35ce44584552c7) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | macaddress-0.2.8.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2019-13173](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13173) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | fstream-1.0.11.tgz | Transitive | 1.0.11 | ❌ |
| [CVE-2020-7774](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | y18n-3.2.1.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7788](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | ini-1.3.4.tgz | Transitive | 1.0.11 | ❌ |
| [CVE-2020-8116](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | dot-prop-3.0.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7720](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | node-forge-0.6.33.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0064](https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2018-0590](https://bugzilla.redhat.com/show_bug.cgi?id=1552148) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.0 | diff-3.3.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-28498](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.8 | elliptic-6.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-21270](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-21270) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | stringstream-0.0.5.tgz | Transitive | 1.0.11 | ❌ |
| [CVE-2021-23386](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23386) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | dns-packet-1.1.1.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-26291](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26291) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | urijs-1.18.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0424](https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | elliptic-6.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7789](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7789) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | node-notifier-5.1.2.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-24033](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24033) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | react-dev-utils-3.0.2.tgz | Transitive | N/A | ❌ |
| [CVE-2020-7598](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [WS-2019-0103](https://github.com/wycats/handlebars.js/issues/1267#issue-187151586) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-15366](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | detected in multiple dependencies | Transitive | 2.0.0 | ✅ |
| [CVE-2020-7693](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | sockjs-0.3.18.tgz | Transitive | 3.4.2 | ✅ |
| [WS-2019-0017](https://github.com/jakubpawlowicz/clean-css/commit/2929bafbf8cdf7dccb24e0949c70833764fa87e3) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | clean-css-4.1.7.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2017-3757](https://github.com/jsdom/whatwg-mimetype/commit/26c539a699778f8743b8319c298b5fb28a4328d0) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | content-type-parser-1.0.1.tgz | Transitive | N/A | ❌ |
| [WS-2020-0342](https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | is-my-json-valid-2.16.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-23382](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | 3.0.0 | ✅ |
| [WS-2021-0154](https://github.com/gulpjs/glob-parent/commit/f9231168b0041fea3f8f954b3cceb56269fc6366) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | glob-parent-2.0.0.tgz | Transitive | N/A | ❌ |
| [CVE-2021-27515](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2020-8124](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8124) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7608](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | 2.0.0 | ✅ |
| [CVE-2021-23362](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | hosted-git-info-2.5.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-1107](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | is-my-json-valid-2.16.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2017-16028](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16028) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | randomatic-1.1.7.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0332](https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0331](https://github.com/wycats/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0427](https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | elliptic-6.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2018-0103](https://hackerone.com/reports/321670) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | stringstream-0.0.5.tgz | Transitive | 1.0.11 | ❌ |
| [WS-2018-0589](https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.0 | nwmatcher-1.4.1.tgz | Transitive | 1.0.11 | ✅ |
## Details
> Partial details (18 vulnerabilities) are displayed below due to a content size limitation in GitHub. To view information on the remaining vulnerabilities, navigate to the Mend Application.<br>
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-3774</summary>
### Vulnerable Libraries - <b>url-parse-1.1.9.tgz</b>, <b>url-parse-1.0.5.tgz</b></p>
<p>
### <b>url-parse-1.1.9.tgz</b></p>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.1.9.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.1.9.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- react-dev-utils-3.0.2.tgz
- sockjs-client-1.1.4.tgz
- :x: **url-parse-1.1.9.tgz** (Vulnerable Library)
### <b>url-parse-1.0.5.tgz</b></p>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.0.5.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.0.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/original/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- react-dev-utils-3.0.2.tgz
- sockjs-client-1.1.4.tgz
- eventsource-0.1.6.tgz
- original-1.0.0.tgz
- :x: **url-parse-1.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Incorrect parsing in url-parse <1.4.3 returns wrong hostname which leads to multiple vulnerabilities such as SSRF, Open Redirect, Bypass Authentication Protocol.
<p>Publish Date: 2018-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3774>CVE-2018-3774</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>10.0</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3774">https://nvd.nist.gov/vuln/detail/CVE-2018-3774</a></p>
<p>Release Date: 2018-08-12</p>
<p>Fix Resolution (url-parse): 1.4.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p><p>Fix Resolution (url-parse): 1.4.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0113</summary>
### Vulnerable Library - <b>macaddress-0.2.8.tgz</b></p>
<p>Get the MAC addresses (hardware addresses) of the hosts network interfaces.</p>
<p>Library home page: <a href="https://registry.npmjs.org/macaddress/-/macaddress-0.2.8.tgz">https://registry.npmjs.org/macaddress/-/macaddress-0.2.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/macaddress/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- css-loader-0.28.4.tgz
- cssnano-3.10.0.tgz
- postcss-filter-plugins-2.0.2.tgz
- uniqid-4.1.1.tgz
- :x: **macaddress-0.2.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of macaddress are vulnerable to command injection. For this vulnerability to be exploited an attacker needs to control the iface argument to the one method.
<p>Publish Date: 2018-05-16
<p>URL: <a href=https://hackerone.com/reports/319467>WS-2018-0113</a></p>
</p>
<p></p>
### CVSS 2 Score Details (<b>10.0</b>)
<p>
Base Score Metrics not available</p>
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="2.8.2">2.8.2</a></p>
<p>Release Date: 2018-01-27</p>
<p>Fix Resolution (macaddress): 0.2.9</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23369</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source.
<p>Publish Date: 2021-04-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369>CVE-2021-23369</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369</a></p>
<p>Release Date: 2021-04-12</p>
<p>Fix Resolution (handlebars): 4.1.2-0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-19919</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Versions of handlebars prior to 4.3.0 are vulnerable to Prototype Pollution leading to Remote Code Execution. Templates may alter an Object's __proto__ and __defineGetter__ properties, which may allow an attacker to execute arbitrary code through crafted payloads.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19919>CVE-2019-19919</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1164">https://www.npmjs.com/advisories/1164</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution (handlebars): 4.3.0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-13797</summary>
### Vulnerable Library - <b>macaddress-0.2.8.tgz</b></p>
<p>Get the MAC addresses (hardware addresses) of the hosts network interfaces.</p>
<p>Library home page: <a href="https://registry.npmjs.org/macaddress/-/macaddress-0.2.8.tgz">https://registry.npmjs.org/macaddress/-/macaddress-0.2.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/macaddress/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- css-loader-0.28.4.tgz
- cssnano-3.10.0.tgz
- postcss-filter-plugins-2.0.2.tgz
- uniqid-4.1.1.tgz
- :x: **macaddress-0.2.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The macaddress module before 0.2.9 for Node.js is prone to an arbitrary command injection flaw, due to allowing unsanitized input to an exec (rather than execFile) call.
<p>Publish Date: 2018-07-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-13797>CVE-2018-13797</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-13797">https://nvd.nist.gov/vuln/detail/CVE-2018-13797</a></p>
<p>Release Date: 2018-07-10</p>
<p>Fix Resolution (macaddress): 0.2.9</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-1000620</summary>
### Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p>
<p>General purpose crypto utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- hawk-3.1.3.tgz
- :x: **cryptiles-2.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.
<p>Publish Date: 2018-07-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620>CVE-2018-1000620</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620</a></p>
<p>Release Date: 2018-07-09</p>
<p>Fix Resolution (cryptiles): 4.1.2</p>
<p>Direct dependency fix Resolution (react-scripts): 1.1.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28499</summary>
### Vulnerable Library - <b>merge-1.2.0.tgz</b></p>
<p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p>
<p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.0.tgz">https://registry.npmjs.org/merge/-/merge-1.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/merge/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- jest-haste-map-20.0.4.tgz
- sane-1.6.0.tgz
- exec-sh-0.2.0.tgz
- :x: **merge-1.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1666">https://www.npmjs.com/advisories/1666</a></p>
<p>Release Date: 2021-02-18</p>
<p>Fix Resolution (merge): 2.1.1</p>
<p>Direct dependency fix Resolution (react-scripts): 3.0.0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0344</summary>
### Vulnerable Library - <b>is-my-json-valid-2.16.0.tgz</b></p>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- eslint-3.19.0.tgz
- :x: **is-my-json-valid-2.16.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.
<p>Publish Date: 2020-06-09
<p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d>WS-2020-0344</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-06-09</p>
<p>Fix Resolution (is-my-json-valid): 2.20.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23383</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package handlebars before 4.7.7 are vulnerable to Prototype Pollution when selecting certain compiling options to compile templates coming from an untrusted source.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23383>CVE-2021-23383</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution (handlebars): 4.1.2-0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-3750</summary>
### Vulnerable Library - <b>deep-extend-0.4.2.tgz</b></p>
<p>Recursive object extending</p>
<p>Library home page: <a href="https://registry.npmjs.org/deep-extend/-/deep-extend-0.4.2.tgz">https://registry.npmjs.org/deep-extend/-/deep-extend-0.4.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/deep-extend/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- sw-precache-webpack-plugin-0.11.3.tgz
- sw-precache-5.2.0.tgz
- update-notifier-1.0.3.tgz
- latest-version-2.0.0.tgz
- package-json-2.4.0.tgz
- registry-auth-token-3.3.1.tgz
- rc-1.2.1.tgz
- :x: **deep-extend-0.4.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The utilities function in all versions <= 0.5.0 of the deep-extend node module can be tricked into modifying the prototype of Object when the attacker can control part of the structure passed to this function. This can let an attacker add or modify existing properties that will exist on all objects.
<p>Publish Date: 2018-07-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3750>CVE-2018-3750</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750</a></p>
<p>Release Date: 2018-07-03</p>
<p>Fix Resolution (deep-extend): 0.5.1</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-16492</summary>
### Vulnerable Library - <b>extend-3.0.1.tgz</b></p>
<p>Port of jQuery.extend for node.js and the browser</p>
<p>Library home page: <a href="https://registry.npmjs.org/extend/-/extend-3.0.1.tgz">https://registry.npmjs.org/extend/-/extend-3.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/extend/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- :x: **extend-3.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A prototype pollution vulnerability was found in module extend <2.0.2, ~<3.0.2 that allows an attacker to inject arbitrary properties onto Object.prototype.
<p>Publish Date: 2019-02-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492>CVE-2018-16492</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/381185">https://hackerone.com/reports/381185</a></p>
<p>Release Date: 2019-02-01</p>
<p>Fix Resolution (extend): 3.0.2</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-6342</summary>
### Vulnerable Library - <b>react-dev-utils-3.0.2.tgz</b></p>
<p>Webpack utilities used by Create React App</p>
<p>Library home page: <a href="https://registry.npmjs.org/react-dev-utils/-/react-dev-utils-3.0.2.tgz">https://registry.npmjs.org/react-dev-utils/-/react-dev-utils-3.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/react-dev-utils/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- :x: **react-dev-utils-3.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
react-dev-utils on Windows allows developers to run a local webserver for accepting various commands, including a command to launch an editor. The input to that command was not properly sanitized, allowing an attacker who can make a network request to the server (either via CSRF or by direct request) to execute arbitrary commands on the targeted system. This issue affects multiple branches: 1.x.x prior to 1.0.4, 2.x.x prior to 2.0.2, 3.x.x prior to 3.1.2, 4.x.x prior to 4.2.2, and 5.x.x prior to 5.0.2.
<p>Publish Date: 2018-12-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-6342>CVE-2018-6342</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-6342">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-6342</a></p>
<p>Release Date: 2018-12-31</p>
<p>Fix Resolution: 1.0.4,2.0.2,3.1.2,4.2.2,5.0.2</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10744</summary>
### Vulnerable Library - <b>lodash.template-4.4.0.tgz</b></p>
<p>The lodash method `_.template` exported as a module.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash.template/-/lodash.template-4.4.0.tgz">https://registry.npmjs.org/lodash.template/-/lodash.template-4.4.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/lodash.template/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- sw-precache-webpack-plugin-0.11.3.tgz
- sw-precache-5.2.0.tgz
- :x: **lodash.template-4.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.
<p>Publish Date: 2019-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744>CVE-2019-10744</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-jf85-cpcp-j695">https://github.com/advisories/GHSA-jf85-cpcp-j695</a></p>
<p>Release Date: 2019-07-26</p>
<p>Fix Resolution (lodash.template): 4.5.0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-3728</summary>
### Vulnerable Library - <b>hoek-2.16.3.tgz</b></p>
<p>General purpose node utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- hawk-3.1.3.tgz
- :x: **hoek-2.16.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3728">https://nvd.nist.gov/vuln/detail/CVE-2018-3728</a></p>
<p>Release Date: 2018-03-30</p>
<p>Fix Resolution (hoek): 4.2.0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.1.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2019-0333</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In handlebars, versions prior to v4.5.3 are vulnerable to prototype pollution. Using a malicious template it's possbile to add or modify properties to the Object prototype. This can also lead to DOS and RCE in certain conditions.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/f7f05d7558e674856686b62a00cde5758f3b7a08>WS-2019-0333</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1325">https://www.npmjs.com/advisories/1325</a></p>
<p>Release Date: 2019-11-18</p>
<p>Fix Resolution (handlebars): 4.5.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-20920</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS).
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20920>CVE-2019-20920</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p>
<p>Release Date: 2020-10-15</p>
<p>Fix Resolution (handlebars): 4.5.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2019-0063</summary>
### Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- eslint-3.19.0.tgz
- :x: **js-yaml-3.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Js-yaml prior to 3.13.1 are vulnerable to Code Injection. The load() function may execute arbitrary code injected through a malicious YAML file.
<p>Publish Date: 2019-04-05
<p>URL: <a href=https://github.com/nodeca/js-yaml/pull/480>WS-2019-0063</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/813">https://www.npmjs.com/advisories/813</a></p>
<p>Release Date: 2019-04-05</p>
<p>Fix Resolution (js-yaml): 3.13.1</p>
<p>Direct dependency fix Resolution (react-scripts): 2.0.0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0084</summary>
### Vulnerable Libraries - <b>sshpk-1.13.0.tgz</b>, <b>sshpk-1.13.1.tgz</b></p>
<p>
### <b>sshpk-1.13.0.tgz</b></p>
<p>A library for finding and using SSH public keys</p>
<p>Library home page: <a href="https://registry.npmjs.org/sshpk/-/sshpk-1.13.0.tgz">https://registry.npmjs.org/sshpk/-/sshpk-1.13.0.tgz</a></p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- http-signature-1.1.1.tgz
- :x: **sshpk-1.13.0.tgz** (Vulnerable Library)
### <b>sshpk-1.13.1.tgz</b></p>
<p>A library for finding and using SSH public keys</p>
<p>Library home page: <a href="https://registry.npmjs.org/sshpk/-/sshpk-1.13.1.tgz">https://registry.npmjs.org/sshpk/-/sshpk-1.13.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/sshpk/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- jest-environment-jsdom-20.0.3.tgz
- jsdom-9.12.0.tgz
- request-2.81.0.tgz
- http-signature-1.1.1.tgz
- :x: **sshpk-1.13.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Versions of sshpk before 1.14.1 are vulnerable to regular expression denial of service when parsing crafted invalid public keys.
<p>Publish Date: 2018-04-25
<p>URL: <a href=https://github.com/joyent/node-sshpk/blob/v1.13.1/lib/formats/ssh.js#L17>WS-2018-0084</a></p>
</p>
<p></p>
### CVSS 2 Score Details (<b>8.0</b>)
<p>
Base Score Metrics not available</p>
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/606">https://nodesecurity.io/advisories/606</a></p>
<p>Release Date: 2018-01-27</p>
<p>Fix Resolution (sshpk): 1.14.1</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p><p>Fix Resolution (sshpk): 1.14.1</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | react-scripts-1.0.10.tgz: 77 vulnerabilities (highest severity is: 10.0) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>react-scripts-1.0.10.tgz</b></p></summary>
<p></p>
<p>Path to dependency file: /package.json</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2018-3774](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3774) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 10.0 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [WS-2018-0113](https://hackerone.com/reports/319467) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 10.0 | macaddress-0.2.8.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-23369](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2019-19919](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19919) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-13797](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-13797) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | macaddress-0.2.8.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-1000620](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | cryptiles-2.0.5.tgz | Transitive | 1.1.1 | ❌ |
| [CVE-2020-28499](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | merge-1.2.0.tgz | Transitive | 3.0.0 | ✅ |
| [WS-2020-0344](https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | is-my-json-valid-2.16.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-23383](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23383) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-3750](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3750) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | deep-extend-0.4.2.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-16492](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | extend-3.0.1.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-6342](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-6342) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | react-dev-utils-3.0.2.tgz | Transitive | N/A | ❌ |
| [CVE-2019-10744](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.1 | lodash.template-4.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-3728](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.8 | hoek-2.16.3.tgz | Transitive | 1.1.1 | ❌ |
| [WS-2019-0333](https://github.com/wycats/handlebars.js/commit/f7f05d7558e674856686b62a00cde5758f3b7a08) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2019-20920](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20920) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0063](https://github.com/nodeca/js-yaml/pull/480) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.1 | js-yaml-3.7.0.tgz | Transitive | 2.0.0 | ✅ |
| [WS-2018-0084](https://github.com/joyent/node-sshpk/blob/v1.13.1/lib/formats/ssh.js#L17) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 8.0 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2020-13822](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13822) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.7 | elliptic-6.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2018-0588](https://github.com/unshiftio/querystringify/pull/19) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.6 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2017-15010](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-15010) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | tough-cookie-2.3.2.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0492](https://github.com/handlebars-lang/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-3737](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3737) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2018-16469](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16469) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | merge-1.2.0.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0493](https://github.com/handlebars-lang/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2017-16099](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16099) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | no-case-2.3.1.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0318](https://github.com/wycats/handlebars.js/commit/8d5530ee2c3ea9f0aee3fde310b9f36887d00b8b) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-33587](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33587) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | css-what-2.1.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-14732](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-14732) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | webpack-dev-server-2.5.0.tgz | Transitive | 2.0.0 | ✅ |
| [CVE-2021-23343](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23343) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | path-parse-1.0.5.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7662](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7662) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | websocket-extensions-0.1.1.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2019-20922](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20922) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-20834](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-20834) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | tar-2.2.1.tgz | Transitive | 1.0.11 | ❌ |
| [CVE-2021-33502](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33502) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | normalize-url-1.9.1.tgz | Transitive | 5.0.0 | ✅ |
| [CVE-2021-33623](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-33623) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | trim-newlines-1.0.0.tgz | Transitive | 2.0.1 | ✅ |
| [WS-2018-0069](https://github.com/mafintosh/is-my-json-valid/commit/b3051b277f7caa08cd2edc6f74f50aeda65d2976) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | is-my-json-valid-2.16.0.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2021-0152](https://github.com/Qix-/color-string/commit/0789e21284c33d89ebc4ab4ca6f759b9375ac9d3) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | color-string-0.3.0.tgz | Transitive | 2.0.0 | ✅ |
| [WS-2020-0218](https://github.com/yeikos/js.merge/pull/38) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | merge-1.2.0.tgz | Transitive | 3.0.0 | ✅ |
| [WS-2019-0032](https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | js-yaml-3.7.0.tgz | Transitive | 2.0.0 | ✅ |
| [CVE-2021-27516](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27516) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | urijs-1.18.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-28092](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28092) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | is-svg-2.1.0.tgz | Transitive | 2.0.0 | ✅ |
| [WS-2020-0091](https://github.com/http-party/node-http-proxy/pull/1447) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | http-proxy-1.16.2.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0541](https://github.com/scravy/node-macaddress/commit/ca9e24df906c9066d49fba658e35ce44584552c7) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | macaddress-0.2.8.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2019-13173](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-13173) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | fstream-1.0.11.tgz | Transitive | 1.0.11 | ❌ |
| [CVE-2020-7774](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7774) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | y18n-3.2.1.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7788](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7788) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | ini-1.3.4.tgz | Transitive | 1.0.11 | ❌ |
| [CVE-2020-8116](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8116) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | dot-prop-3.0.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7720](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7720) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | node-forge-0.6.33.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0064](https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2018-0590](https://bugzilla.redhat.com/show_bug.cgi?id=1552148) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.0 | diff-3.3.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-28498](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.8 | elliptic-6.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-21270](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-21270) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | stringstream-0.0.5.tgz | Transitive | 1.0.11 | ❌ |
| [CVE-2021-23386](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23386) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | dns-packet-1.1.1.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-26291](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-26291) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | urijs-1.18.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0424](https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | elliptic-6.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7789](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7789) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | node-notifier-5.1.2.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-24033](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-24033) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | react-dev-utils-3.0.2.tgz | Transitive | N/A | ❌ |
| [CVE-2020-7598](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7598) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [WS-2019-0103](https://github.com/wycats/handlebars.js/issues/1267#issue-187151586) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2020-15366](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15366) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.6 | detected in multiple dependencies | Transitive | 2.0.0 | ✅ |
| [CVE-2020-7693](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7693) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | sockjs-0.3.18.tgz | Transitive | 3.4.2 | ✅ |
| [WS-2019-0017](https://github.com/jakubpawlowicz/clean-css/commit/2929bafbf8cdf7dccb24e0949c70833764fa87e3) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | clean-css-4.1.7.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2017-3757](https://github.com/jsdom/whatwg-mimetype/commit/26c539a699778f8743b8319c298b5fb28a4328d0) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | content-type-parser-1.0.1.tgz | Transitive | N/A | ❌ |
| [WS-2020-0342](https://github.com/mafintosh/is-my-json-valid/commit/c3fc04fc455d40e9b29537f8e2c73a28ce106edb) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | is-my-json-valid-2.16.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2021-23382](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | 3.0.0 | ✅ |
| [WS-2021-0154](https://github.com/gulpjs/glob-parent/commit/f9231168b0041fea3f8f954b3cceb56269fc6366) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | glob-parent-2.0.0.tgz | Transitive | N/A | ❌ |
| [CVE-2021-27515](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-27515) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2020-8124](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8124) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | 1.0.11 | ✅ |
| [CVE-2020-7608](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-7608) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | 2.0.0 | ✅ |
| [CVE-2021-23362](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23362) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | hosted-git-info-2.5.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2018-1107](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1107) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | is-my-json-valid-2.16.0.tgz | Transitive | 1.0.11 | ✅ |
| [CVE-2017-16028](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16028) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | randomatic-1.1.7.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0332](https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0331](https://github.com/wycats/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | handlebars-4.0.10.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2019-0427](https://github.com/indutny/elliptic/commit/ec735edde187a43693197f6fa3667ceade751a3a) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | elliptic-6.4.0.tgz | Transitive | 1.0.11 | ✅ |
| [WS-2018-0103](https://hackerone.com/reports/321670) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.8 | stringstream-0.0.5.tgz | Transitive | 1.0.11 | ❌ |
| [WS-2018-0589](https://github.com/dperini/nwmatcher/commit/9dcc2b039beeabd18327a5ebaa537625872e16f0) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.0 | nwmatcher-1.4.1.tgz | Transitive | 1.0.11 | ✅ |
## Details
> Partial details (18 vulnerabilities) are displayed below due to a content size limitation in GitHub. To view information on the remaining vulnerabilities, navigate to the Mend Application.<br>
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-3774</summary>
### Vulnerable Libraries - <b>url-parse-1.1.9.tgz</b>, <b>url-parse-1.0.5.tgz</b></p>
<p>
### <b>url-parse-1.1.9.tgz</b></p>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.1.9.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.1.9.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- react-dev-utils-3.0.2.tgz
- sockjs-client-1.1.4.tgz
- :x: **url-parse-1.1.9.tgz** (Vulnerable Library)
### <b>url-parse-1.0.5.tgz</b></p>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.0.5.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.0.5.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/original/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- react-dev-utils-3.0.2.tgz
- sockjs-client-1.1.4.tgz
- eventsource-0.1.6.tgz
- original-1.0.0.tgz
- :x: **url-parse-1.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Incorrect parsing in url-parse <1.4.3 returns wrong hostname which leads to multiple vulnerabilities such as SSRF, Open Redirect, Bypass Authentication Protocol.
<p>Publish Date: 2018-08-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3774>CVE-2018-3774</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>10.0</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3774">https://nvd.nist.gov/vuln/detail/CVE-2018-3774</a></p>
<p>Release Date: 2018-08-12</p>
<p>Fix Resolution (url-parse): 1.4.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p><p>Fix Resolution (url-parse): 1.4.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0113</summary>
### Vulnerable Library - <b>macaddress-0.2.8.tgz</b></p>
<p>Get the MAC addresses (hardware addresses) of the hosts network interfaces.</p>
<p>Library home page: <a href="https://registry.npmjs.org/macaddress/-/macaddress-0.2.8.tgz">https://registry.npmjs.org/macaddress/-/macaddress-0.2.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/macaddress/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- css-loader-0.28.4.tgz
- cssnano-3.10.0.tgz
- postcss-filter-plugins-2.0.2.tgz
- uniqid-4.1.1.tgz
- :x: **macaddress-0.2.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of macaddress are vulnerable to command injection. For this vulnerability to be exploited an attacker needs to control the iface argument to the one method.
<p>Publish Date: 2018-05-16
<p>URL: <a href=https://hackerone.com/reports/319467>WS-2018-0113</a></p>
</p>
<p></p>
### CVSS 2 Score Details (<b>10.0</b>)
<p>
Base Score Metrics not available</p>
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="2.8.2">2.8.2</a></p>
<p>Release Date: 2018-01-27</p>
<p>Fix Resolution (macaddress): 0.2.9</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23369</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package handlebars before 4.7.7 are vulnerable to Remote Code Execution (RCE) when selecting certain compiling options to compile templates coming from an untrusted source.
<p>Publish Date: 2021-04-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23369>CVE-2021-23369</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23369</a></p>
<p>Release Date: 2021-04-12</p>
<p>Fix Resolution (handlebars): 4.1.2-0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-19919</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Versions of handlebars prior to 4.3.0 are vulnerable to Prototype Pollution leading to Remote Code Execution. Templates may alter an Object's __proto__ and __defineGetter__ properties, which may allow an attacker to execute arbitrary code through crafted payloads.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-19919>CVE-2019-19919</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1164">https://www.npmjs.com/advisories/1164</a></p>
<p>Release Date: 2019-12-20</p>
<p>Fix Resolution (handlebars): 4.3.0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-13797</summary>
### Vulnerable Library - <b>macaddress-0.2.8.tgz</b></p>
<p>Get the MAC addresses (hardware addresses) of the hosts network interfaces.</p>
<p>Library home page: <a href="https://registry.npmjs.org/macaddress/-/macaddress-0.2.8.tgz">https://registry.npmjs.org/macaddress/-/macaddress-0.2.8.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/macaddress/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- css-loader-0.28.4.tgz
- cssnano-3.10.0.tgz
- postcss-filter-plugins-2.0.2.tgz
- uniqid-4.1.1.tgz
- :x: **macaddress-0.2.8.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The macaddress module before 0.2.9 for Node.js is prone to an arbitrary command injection flaw, due to allowing unsanitized input to an exec (rather than execFile) call.
<p>Publish Date: 2018-07-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-13797>CVE-2018-13797</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-13797">https://nvd.nist.gov/vuln/detail/CVE-2018-13797</a></p>
<p>Release Date: 2018-07-10</p>
<p>Fix Resolution (macaddress): 0.2.9</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-1000620</summary>
### Vulnerable Library - <b>cryptiles-2.0.5.tgz</b></p>
<p>General purpose crypto utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz">https://registry.npmjs.org/cryptiles/-/cryptiles-2.0.5.tgz</a></p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- hawk-3.1.3.tgz
- :x: **cryptiles-2.0.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Eran Hammer cryptiles version 4.1.1 earlier contains a CWE-331: Insufficient Entropy vulnerability in randomDigits() method that can result in An attacker is more likely to be able to brute force something that was supposed to be random.. This attack appear to be exploitable via Depends upon the calling application.. This vulnerability appears to have been fixed in 4.1.2.
<p>Publish Date: 2018-07-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-1000620>CVE-2018-1000620</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-1000620</a></p>
<p>Release Date: 2018-07-09</p>
<p>Fix Resolution (cryptiles): 4.1.2</p>
<p>Direct dependency fix Resolution (react-scripts): 1.1.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-28499</summary>
### Vulnerable Library - <b>merge-1.2.0.tgz</b></p>
<p>Merge multiple objects into one, optionally creating a new cloned object. Similar to the jQuery.extend but more flexible. Works in Node.js and the browser.</p>
<p>Library home page: <a href="https://registry.npmjs.org/merge/-/merge-1.2.0.tgz">https://registry.npmjs.org/merge/-/merge-1.2.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/merge/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- jest-haste-map-20.0.4.tgz
- sane-1.6.0.tgz
- exec-sh-0.2.0.tgz
- :x: **merge-1.2.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
All versions of package merge are vulnerable to Prototype Pollution via _recursiveMerge .
<p>Publish Date: 2021-02-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28499>CVE-2020-28499</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1666">https://www.npmjs.com/advisories/1666</a></p>
<p>Release Date: 2021-02-18</p>
<p>Fix Resolution (merge): 2.1.1</p>
<p>Direct dependency fix Resolution (react-scripts): 3.0.0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2020-0344</summary>
### Vulnerable Library - <b>is-my-json-valid-2.16.0.tgz</b></p>
<p>A JSONSchema validator that uses code generation to be extremely fast</p>
<p>Library home page: <a href="https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.0.tgz">https://registry.npmjs.org/is-my-json-valid/-/is-my-json-valid-2.16.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/is-my-json-valid/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- eslint-3.19.0.tgz
- :x: **is-my-json-valid-2.16.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Arbitrary Code Execution vulnerability was found in is-my-json-valid before 2.20.3 via the fromatName function.
<p>Publish Date: 2020-06-09
<p>URL: <a href=https://github.com/mafintosh/is-my-json-valid/commit/3419563687df463b4ca709a2b46be8e15d6a2b3d>WS-2020-0344</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2020-06-09</p>
<p>Fix Resolution (is-my-json-valid): 2.20.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-23383</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The package handlebars before 4.7.7 are vulnerable to Prototype Pollution when selecting certain compiling options to compile templates coming from an untrusted source.
<p>Publish Date: 2021-05-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23383>CVE-2021-23383</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23383</a></p>
<p>Release Date: 2021-05-04</p>
<p>Fix Resolution (handlebars): 4.1.2-0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-3750</summary>
### Vulnerable Library - <b>deep-extend-0.4.2.tgz</b></p>
<p>Recursive object extending</p>
<p>Library home page: <a href="https://registry.npmjs.org/deep-extend/-/deep-extend-0.4.2.tgz">https://registry.npmjs.org/deep-extend/-/deep-extend-0.4.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/deep-extend/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- sw-precache-webpack-plugin-0.11.3.tgz
- sw-precache-5.2.0.tgz
- update-notifier-1.0.3.tgz
- latest-version-2.0.0.tgz
- package-json-2.4.0.tgz
- registry-auth-token-3.3.1.tgz
- rc-1.2.1.tgz
- :x: **deep-extend-0.4.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The utilities function in all versions <= 0.5.0 of the deep-extend node module can be tricked into modifying the prototype of Object when the attacker can control part of the structure passed to this function. This can let an attacker add or modify existing properties that will exist on all objects.
<p>Publish Date: 2018-07-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3750>CVE-2018-3750</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-3750</a></p>
<p>Release Date: 2018-07-03</p>
<p>Fix Resolution (deep-extend): 0.5.1</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-16492</summary>
### Vulnerable Library - <b>extend-3.0.1.tgz</b></p>
<p>Port of jQuery.extend for node.js and the browser</p>
<p>Library home page: <a href="https://registry.npmjs.org/extend/-/extend-3.0.1.tgz">https://registry.npmjs.org/extend/-/extend-3.0.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/extend/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- :x: **extend-3.0.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A prototype pollution vulnerability was found in module extend <2.0.2, ~<3.0.2 that allows an attacker to inject arbitrary properties onto Object.prototype.
<p>Publish Date: 2019-02-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16492>CVE-2018-16492</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://hackerone.com/reports/381185">https://hackerone.com/reports/381185</a></p>
<p>Release Date: 2019-02-01</p>
<p>Fix Resolution (extend): 3.0.2</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-6342</summary>
### Vulnerable Library - <b>react-dev-utils-3.0.2.tgz</b></p>
<p>Webpack utilities used by Create React App</p>
<p>Library home page: <a href="https://registry.npmjs.org/react-dev-utils/-/react-dev-utils-3.0.2.tgz">https://registry.npmjs.org/react-dev-utils/-/react-dev-utils-3.0.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/react-dev-utils/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- :x: **react-dev-utils-3.0.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
react-dev-utils on Windows allows developers to run a local webserver for accepting various commands, including a command to launch an editor. The input to that command was not properly sanitized, allowing an attacker who can make a network request to the server (either via CSRF or by direct request) to execute arbitrary commands on the targeted system. This issue affects multiple branches: 1.x.x prior to 1.0.4, 2.x.x prior to 2.0.2, 3.x.x prior to 3.1.2, 4.x.x prior to 4.2.2, and 5.x.x prior to 5.0.2.
<p>Publish Date: 2018-12-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-6342>CVE-2018-6342</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-6342">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-6342</a></p>
<p>Release Date: 2018-12-31</p>
<p>Fix Resolution: 1.0.4,2.0.2,3.1.2,4.2.2,5.0.2</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-10744</summary>
### Vulnerable Library - <b>lodash.template-4.4.0.tgz</b></p>
<p>The lodash method `_.template` exported as a module.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash.template/-/lodash.template-4.4.0.tgz">https://registry.npmjs.org/lodash.template/-/lodash.template-4.4.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/lodash.template/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- sw-precache-webpack-plugin-0.11.3.tgz
- sw-precache-5.2.0.tgz
- :x: **lodash.template-4.4.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Versions of lodash lower than 4.17.12 are vulnerable to Prototype Pollution. The function defaultsDeep could be tricked into adding or modifying properties of Object.prototype using a constructor payload.
<p>Publish Date: 2019-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10744>CVE-2019-10744</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-jf85-cpcp-j695">https://github.com/advisories/GHSA-jf85-cpcp-j695</a></p>
<p>Release Date: 2019-07-26</p>
<p>Fix Resolution (lodash.template): 4.5.0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2018-3728</summary>
### Vulnerable Library - <b>hoek-2.16.3.tgz</b></p>
<p>General purpose node utilities</p>
<p>Library home page: <a href="https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz">https://registry.npmjs.org/hoek/-/hoek-2.16.3.tgz</a></p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- hawk-3.1.3.tgz
- :x: **hoek-2.16.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
hoek node module before 4.2.0 and 5.0.x before 5.0.3 suffers from a Modification of Assumed-Immutable Data (MAID) vulnerability via 'merge' and 'applyToDefaults' functions, which allows a malicious user to modify the prototype of "Object" via __proto__, causing the addition or modification of an existing property that will exist on all objects.
<p>Publish Date: 2018-03-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-3728>CVE-2018-3728</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-3728">https://nvd.nist.gov/vuln/detail/CVE-2018-3728</a></p>
<p>Release Date: 2018-03-30</p>
<p>Fix Resolution (hoek): 4.2.0</p>
<p>Direct dependency fix Resolution (react-scripts): 1.1.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2019-0333</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In handlebars, versions prior to v4.5.3 are vulnerable to prototype pollution. Using a malicious template it's possbile to add or modify properties to the Object prototype. This can also lead to DOS and RCE in certain conditions.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://github.com/wycats/handlebars.js/commit/f7f05d7558e674856686b62a00cde5758f3b7a08>WS-2019-0333</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1325">https://www.npmjs.com/advisories/1325</a></p>
<p>Release Date: 2019-11-18</p>
<p>Fix Resolution (handlebars): 4.5.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-20920</summary>
### Vulnerable Library - <b>handlebars-4.0.10.tgz</b></p>
<p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p>
<p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz">https://registry.npmjs.org/handlebars/-/handlebars-4.0.10.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/handlebars/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- istanbul-api-1.1.11.tgz
- istanbul-reports-1.1.1.tgz
- :x: **handlebars-4.0.10.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Handlebars before 3.0.8 and 4.x before 4.5.3 is vulnerable to Arbitrary Code Execution. The lookup helper fails to properly validate templates, allowing attackers to submit templates that execute arbitrary JavaScript. This can be used to run arbitrary code on a server processing Handlebars templates or in a victim's browser (effectively serving as XSS).
<p>Publish Date: 2020-09-30
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-20920>CVE-2019-20920</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p>
<p>Release Date: 2020-10-15</p>
<p>Fix Resolution (handlebars): 4.5.3</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2019-0063</summary>
### Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- eslint-3.19.0.tgz
- :x: **js-yaml-3.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Js-yaml prior to 3.13.1 are vulnerable to Code Injection. The load() function may execute arbitrary code injected through a malicious YAML file.
<p>Publish Date: 2019-04-05
<p>URL: <a href=https://github.com/nodeca/js-yaml/pull/480>WS-2019-0063</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>8.1</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/813">https://www.npmjs.com/advisories/813</a></p>
<p>Release Date: 2019-04-05</p>
<p>Fix Resolution (js-yaml): 3.13.1</p>
<p>Direct dependency fix Resolution (react-scripts): 2.0.0</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2018-0084</summary>
### Vulnerable Libraries - <b>sshpk-1.13.0.tgz</b>, <b>sshpk-1.13.1.tgz</b></p>
<p>
### <b>sshpk-1.13.0.tgz</b></p>
<p>A library for finding and using SSH public keys</p>
<p>Library home page: <a href="https://registry.npmjs.org/sshpk/-/sshpk-1.13.0.tgz">https://registry.npmjs.org/sshpk/-/sshpk-1.13.0.tgz</a></p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- fsevents-1.1.2.tgz
- node-pre-gyp-0.6.36.tgz
- request-2.81.0.tgz
- http-signature-1.1.1.tgz
- :x: **sshpk-1.13.0.tgz** (Vulnerable Library)
### <b>sshpk-1.13.1.tgz</b></p>
<p>A library for finding and using SSH public keys</p>
<p>Library home page: <a href="https://registry.npmjs.org/sshpk/-/sshpk-1.13.1.tgz">https://registry.npmjs.org/sshpk/-/sshpk-1.13.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/sshpk/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-1.0.10.tgz (Root Library)
- jest-20.0.4.tgz
- jest-cli-20.0.4.tgz
- jest-environment-jsdom-20.0.3.tgz
- jsdom-9.12.0.tgz
- request-2.81.0.tgz
- http-signature-1.1.1.tgz
- :x: **sshpk-1.13.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/soundcloud-redux-11/commit/4adf157937b5eb9532c8c763a3671bdc84d1791b">4adf157937b5eb9532c8c763a3671bdc84d1791b</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Versions of sshpk before 1.14.1 are vulnerable to regular expression denial of service when parsing crafted invalid public keys.
<p>Publish Date: 2018-04-25
<p>URL: <a href=https://github.com/joyent/node-sshpk/blob/v1.13.1/lib/formats/ssh.js#L17>WS-2018-0084</a></p>
</p>
<p></p>
### CVSS 2 Score Details (<b>8.0</b>)
<p>
Base Score Metrics not available</p>
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nodesecurity.io/advisories/606">https://nodesecurity.io/advisories/606</a></p>
<p>Release Date: 2018-01-27</p>
<p>Fix Resolution (sshpk): 1.14.1</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p><p>Fix Resolution (sshpk): 1.14.1</p>
<p>Direct dependency fix Resolution (react-scripts): 1.0.11</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_process | react scripts tgz vulnerabilities highest severity is vulnerable library react scripts tgz path to dependency file package json found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high detected in multiple dependencies transitive high macaddress tgz transitive high handlebars tgz transitive high handlebars tgz transitive high macaddress tgz transitive high cryptiles tgz transitive high merge tgz transitive high is my json valid tgz transitive high handlebars tgz transitive high deep extend tgz transitive high extend tgz transitive high react dev utils tgz transitive n a high lodash template tgz transitive high hoek tgz transitive high handlebars tgz transitive high handlebars tgz transitive high js yaml tgz transitive high detected in multiple dependencies transitive high elliptic tgz transitive high detected in multiple dependencies transitive high tough cookie tgz transitive high handlebars tgz transitive high detected in multiple dependencies transitive high merge tgz transitive high handlebars tgz transitive high no case tgz transitive high handlebars tgz transitive high css what tgz transitive high webpack dev server tgz transitive high path parse tgz transitive high websocket extensions tgz transitive high handlebars tgz transitive high tar tgz transitive high normalize url tgz transitive high trim newlines tgz transitive high is my json valid tgz transitive high color string tgz transitive high merge tgz transitive high js yaml tgz transitive high urijs tgz transitive high is svg tgz transitive high http proxy tgz transitive high macaddress tgz transitive high fstream tgz transitive high tgz transitive high ini tgz transitive high dot prop tgz transitive high node forge tgz transitive high handlebars tgz transitive high diff tgz transitive medium elliptic tgz transitive medium stringstream tgz transitive medium dns packet tgz transitive medium urijs tgz transitive medium elliptic tgz transitive medium node notifier tgz transitive medium react dev utils tgz transitive n a medium detected in multiple dependencies transitive medium handlebars tgz transitive medium detected in multiple dependencies transitive medium sockjs tgz transitive medium clean css tgz transitive medium content type parser tgz transitive n a medium is my json valid tgz transitive medium detected in multiple dependencies transitive medium glob parent tgz transitive n a medium detected in multiple dependencies transitive medium detected in multiple dependencies transitive medium detected in multiple dependencies transitive medium hosted git info tgz transitive medium is my json valid tgz transitive medium randomatic tgz transitive medium handlebars tgz transitive medium handlebars tgz transitive medium elliptic tgz transitive medium stringstream tgz transitive medium nwmatcher tgz transitive details partial details vulnerabilities are displayed below due to a content size limitation in github to view information on the remaining vulnerabilities navigate to the mend application cve vulnerable libraries url parse tgz url parse tgz url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file package json path to vulnerable library node modules url parse package json dependency hierarchy react scripts tgz root library react dev utils tgz sockjs client tgz x url parse tgz vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file package json path to vulnerable library node modules original node modules url parse package json dependency hierarchy react scripts tgz root library react dev utils tgz sockjs client tgz eventsource tgz original tgz x url parse tgz vulnerable library found in head commit a href found in base branch master vulnerability details incorrect parsing in url parse returns wrong hostname which leads to multiple vulnerabilities such as ssrf open redirect bypass authentication protocol publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution react scripts fix resolution url parse direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue ws vulnerable library macaddress tgz get the mac addresses hardware addresses of the hosts network interfaces library home page a href path to dependency file package json path to vulnerable library node modules macaddress package json dependency hierarchy react scripts tgz root library css loader tgz cssnano tgz postcss filter plugins tgz uniqid tgz x macaddress tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of macaddress are vulnerable to command injection for this vulnerability to be exploited an attacker needs to control the iface argument to the one method publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin release date fix resolution macaddress direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file package json path to vulnerable library node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package handlebars before are vulnerable to remote code execution rce when selecting certain compiling options to compile templates coming from an untrusted source publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file package json path to vulnerable library node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of handlebars prior to are vulnerable to prototype pollution leading to remote code execution templates may alter an object s proto and definegetter properties which may allow an attacker to execute arbitrary code through crafted payloads publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library macaddress tgz get the mac addresses hardware addresses of the hosts network interfaces library home page a href path to dependency file package json path to vulnerable library node modules macaddress package json dependency hierarchy react scripts tgz root library css loader tgz cssnano tgz postcss filter plugins tgz uniqid tgz x macaddress tgz vulnerable library found in head commit a href found in base branch master vulnerability details the macaddress module before for node js is prone to an arbitrary command injection flaw due to allowing unsanitized input to an exec rather than execfile call publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution macaddress direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library cryptiles tgz general purpose crypto utilities library home page a href dependency hierarchy react scripts tgz root library fsevents tgz node pre gyp tgz request tgz hawk tgz x cryptiles tgz vulnerable library found in head commit a href found in base branch master vulnerability details eran hammer cryptiles version earlier contains a cwe insufficient entropy vulnerability in randomdigits method that can result in an attacker is more likely to be able to brute force something that was supposed to be random this attack appear to be exploitable via depends upon the calling application this vulnerability appears to have been fixed in publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cryptiles direct dependency fix resolution react scripts cve vulnerable library merge tgz merge multiple objects into one optionally creating a new cloned object similar to the jquery extend but more flexible works in node js and the browser library home page a href path to dependency file package json path to vulnerable library node modules merge package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz jest haste map tgz sane tgz exec sh tgz x merge tgz vulnerable library found in head commit a href found in base branch master vulnerability details all versions of package merge are vulnerable to prototype pollution via recursivemerge publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution merge direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue ws vulnerable library is my json valid tgz a jsonschema validator that uses code generation to be extremely fast library home page a href path to dependency file package json path to vulnerable library node modules is my json valid package json dependency hierarchy react scripts tgz root library eslint tgz x is my json valid tgz vulnerable library found in head commit a href found in base branch master vulnerability details arbitrary code execution vulnerability was found in is my json valid before via the fromatname function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution is my json valid direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file package json path to vulnerable library node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package handlebars before are vulnerable to prototype pollution when selecting certain compiling options to compile templates coming from an untrusted source publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library deep extend tgz recursive object extending library home page a href path to dependency file package json path to vulnerable library node modules deep extend package json dependency hierarchy react scripts tgz root library sw precache webpack plugin tgz sw precache tgz update notifier tgz latest version tgz package json tgz registry auth token tgz rc tgz x deep extend tgz vulnerable library found in head commit a href found in base branch master vulnerability details the utilities function in all versions of the deep extend node module can be tricked into modifying the prototype of object when the attacker can control part of the structure passed to this function this can let an attacker add or modify existing properties that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution deep extend direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library extend tgz port of jquery extend for node js and the browser library home page a href path to dependency file package json path to vulnerable library node modules extend package json dependency hierarchy react scripts tgz root library fsevents tgz node pre gyp tgz request tgz x extend tgz vulnerable library found in head commit a href found in base branch master vulnerability details a prototype pollution vulnerability was found in module extend that allows an attacker to inject arbitrary properties onto object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution extend direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library react dev utils tgz webpack utilities used by create react app library home page a href path to dependency file package json path to vulnerable library node modules react dev utils package json dependency hierarchy react scripts tgz root library x react dev utils tgz vulnerable library found in head commit a href found in base branch master vulnerability details react dev utils on windows allows developers to run a local webserver for accepting various commands including a command to launch an editor the input to that command was not properly sanitized allowing an attacker who can make a network request to the server either via csrf or by direct request to execute arbitrary commands on the targeted system this issue affects multiple branches x x prior to x x prior to x x prior to x x prior to and x x prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cve vulnerable library lodash template tgz the lodash method template exported as a module library home page a href path to dependency file package json path to vulnerable library node modules lodash template package json dependency hierarchy react scripts tgz root library sw precache webpack plugin tgz sw precache tgz x lodash template tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of lodash lower than are vulnerable to prototype pollution the function defaultsdeep could be tricked into adding or modifying properties of object prototype using a constructor payload publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash template direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library hoek tgz general purpose node utilities library home page a href dependency hierarchy react scripts tgz root library fsevents tgz node pre gyp tgz request tgz hawk tgz x hoek tgz vulnerable library found in head commit a href found in base branch master vulnerability details hoek node module before and x before suffers from a modification of assumed immutable data maid vulnerability via merge and applytodefaults functions which allows a malicious user to modify the prototype of object via proto causing the addition or modification of an existing property that will exist on all objects publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution hoek direct dependency fix resolution react scripts ws vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file package json path to vulnerable library node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details in handlebars versions prior to are vulnerable to prototype pollution using a malicious template it s possbile to add or modify properties to the object prototype this can also lead to dos and rce in certain conditions publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue cve vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file package json path to vulnerable library node modules handlebars package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz istanbul api tgz istanbul reports tgz x handlebars tgz vulnerable library found in head commit a href found in base branch master vulnerability details handlebars before and x before is vulnerable to arbitrary code execution the lookup helper fails to properly validate templates allowing attackers to submit templates that execute arbitrary javascript this can be used to run arbitrary code on a server processing handlebars templates or in a victim s browser effectively serving as xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution handlebars direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue ws vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file package json path to vulnerable library node modules js yaml package json dependency hierarchy react scripts tgz root library eslint tgz x js yaml tgz vulnerable library found in head commit a href found in base branch master vulnerability details js yaml prior to are vulnerable to code injection the load function may execute arbitrary code injected through a malicious yaml file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution js yaml direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue ws vulnerable libraries sshpk tgz sshpk tgz sshpk tgz a library for finding and using ssh public keys library home page a href dependency hierarchy react scripts tgz root library fsevents tgz node pre gyp tgz request tgz http signature tgz x sshpk tgz vulnerable library sshpk tgz a library for finding and using ssh public keys library home page a href path to dependency file package json path to vulnerable library node modules sshpk package json dependency hierarchy react scripts tgz root library jest tgz jest cli tgz jest environment jsdom tgz jsdom tgz request tgz http signature tgz x sshpk tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions of sshpk before are vulnerable to regular expression denial of service when parsing crafted invalid public keys publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution sshpk direct dependency fix resolution react scripts fix resolution sshpk direct dependency fix resolution react scripts rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
21,971 | 30,464,658,255 | IssuesEvent | 2023-07-17 09:30:28 | MicrosoftDocs/azure-devops-docs | https://api.github.com/repos/MicrosoftDocs/azure-devops-docs | closed | Typo: "...and works out its way." should be "...and works its way out." | doc-bug Pri1 azure-devops-pipelines/svc azure-devops-pipelines-process/subsvc |
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 21e5cee4-eaae-3a96-db91-540ac759e83a
* Version Independent ID: 9bdc837c-ffe0-d999-f922-f3a5debc7f92
* Content: [Conditions - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml%2Cstages)
* Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/conditions.md)
* Service: **azure-devops-pipelines**
* Sub-service: **azure-devops-pipelines-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | 1.0 | Typo: "...and works out its way." should be "...and works its way out." -
[Enter feedback here]
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: 21e5cee4-eaae-3a96-db91-540ac759e83a
* Version Independent ID: 9bdc837c-ffe0-d999-f922-f3a5debc7f92
* Content: [Conditions - Azure Pipelines](https://learn.microsoft.com/en-us/azure/devops/pipelines/process/conditions?view=azure-devops&tabs=yaml%2Cstages)
* Content Source: [docs/pipelines/process/conditions.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/main/docs/pipelines/process/conditions.md)
* Service: **azure-devops-pipelines**
* Sub-service: **azure-devops-pipelines-process**
* GitHub Login: @juliakm
* Microsoft Alias: **jukullam** | process | typo and works out its way should be and works its way out document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id eaae version independent id content content source service azure devops pipelines sub service azure devops pipelines process github login juliakm microsoft alias jukullam | 1 |
4,871 | 7,753,030,038 | IssuesEvent | 2018-05-30 22:27:02 | GoogleCloudPlatform/google-cloud-python | https://api.github.com/repos/GoogleCloudPlatform/google-cloud-python | closed | BigQuery 1.2.0 release | api: bigquery packaging type: process | The big new feature is https://github.com/GoogleCloudPlatform/google-cloud-python/pull/5267 which adds support for column-based time partitioning. | 1.0 | BigQuery 1.2.0 release - The big new feature is https://github.com/GoogleCloudPlatform/google-cloud-python/pull/5267 which adds support for column-based time partitioning. | process | bigquery release the big new feature is which adds support for column based time partitioning | 1 |
29 | 2,497,604,967 | IssuesEvent | 2015-01-07 09:07:59 | sysown/proxysql-0.2 | https://api.github.com/repos/sysown/proxysql-0.2 | closed | implement query processing in Standard_Query_Processor::process_mysql_query() | bug development QUERY PROCESSOR | It seems that right now only regex matching is enabled.
username, schemaname, flagIN/flagOUT and apply are still not implemented.
Query rewrite (replace_pattern) is not implemented either. | 1.0 | implement query processing in Standard_Query_Processor::process_mysql_query() - It seems that right now only regex matching is enabled.
username, schemaname, flagIN/flagOUT and apply are still not implemented.
Query rewrite (replace_pattern) is not implemented either. | process | implement query processing in standard query processor process mysql query it seems that right now only regex matching is enabled username schemaname flagin flagout and apply are still not implemented query rewrite replace pattern is not implemented either | 1 |
1,823 | 4,581,159,709 | IssuesEvent | 2016-09-19 02:52:29 | VietOpenCPS/opencps | https://api.github.com/repos/VietOpenCPS/opencps | closed | Hiển thị dịch vụ công chưa được gán quy trình xử lý trên màn hình tạo mới hồ sơ | bug service_process | 1. Đăng nhập tài khoản doanh nghiệp
2. Vào quản lý hồ sơ, nhấn nút Tạo mới hồ sơ
Lỗi: Hiển thị dịch vụ công chưa được gán quy trình xử lý trên màn hình tạo mới hồ sơ | 1.0 | Hiển thị dịch vụ công chưa được gán quy trình xử lý trên màn hình tạo mới hồ sơ - 1. Đăng nhập tài khoản doanh nghiệp
2. Vào quản lý hồ sơ, nhấn nút Tạo mới hồ sơ
Lỗi: Hiển thị dịch vụ công chưa được gán quy trình xử lý trên màn hình tạo mới hồ sơ | process | hiển thị dịch vụ công chưa được gán quy trình xử lý trên màn hình tạo mới hồ sơ đăng nhập tài khoản doanh nghiệp vào quản lý hồ sơ nhấn nút tạo mới hồ sơ lỗi hiển thị dịch vụ công chưa được gán quy trình xử lý trên màn hình tạo mới hồ sơ | 1 |
20,266 | 26,889,164,214 | IssuesEvent | 2023-02-06 07:21:48 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Area Simplify drops Z and M values | Processing Bug | ### What is the bug or the crash?
If you run the Simplify Algo with Area(Visvalingam) the Z and M values will become null if any vertexes are removed.
### Steps to reproduce the issue
1. Create a line layer with Z at 100 and M values set and enough vertices to be able to run simplify.
2. Run the simplify algo and set Simplification method to Area(Visvalingam).
3. Set the tolerance to a value that will remove at least one vertex.
4. Check that the M and Z values are now null.
Test data:
[Area_Simplify_ZMdrop.zip](https://github.com/qgis/QGIS/files/9896735/Area_Simplify_ZMdrop.zip)
Starting layer with all Z set to 100

Simplified with some vertex's removed all Z now null or 0

Simplified with no vertex's removed all Z's still 100 (one vertex was set to 200 did the same thing without that)

### Versions
<!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css">
p, li { white-space: pre-wrap; }
</style></head><body>
QGIS version | 3.29.0-Master | QGIS code revision | ca130dd49c
-- | -- | -- | --
Qt version | 5.15.3
Python version | 3.9.5
GDAL/OGR version | 3.6.0dev-5df18be65c
PROJ version | 9.1.0
EPSG Registry database version | v10.074 (2022-08-01)
GEOS version | 3.10.3-CAPI-1.16.1
SQLite version | 3.39.4
PDAL version | 2.4.3
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.6
QScintilla2 version | 2.13.1
OS version | Windows 10 Version 2009
| | |
This copy of QGIS writes debugging output.
| | |
Active Python plugins
db_manager | 0.1.20
grassprovider | 2.12.99
MetaSearch | 0.3.6
otbprovider | 2.12.99
processing | 2.12.99
sagaprovider | 2.12.99
</body></html><!--EndFragment-->QGIS version
3.29.0-Master
QGIS code revision
[ca130dd49c](https://github.com/qgis/QGIS/commit/ca130dd49c)
Qt version
5.15.3
Python version
3.9.5
GDAL/OGR version
3.6.0dev-5df18be65c
PROJ version
9.1.0
EPSG Registry database version
v10.074 (2022-08-01)
GEOS version
3.10.3-CAPI-1.16.1
SQLite version
3.39.4
PDAL version
2.4.3
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.6
QScintilla2 version
2.13.1
OS version
Windows 10 Version 2009
This copy of QGIS writes debugging output.
Active Python plugins
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.6
otbprovider
2.12.99
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
The Distance (Douglas-Puecker) stetting does not appear to be affected but it also doesn't seem to take into account Z values in simplification.
I was excited by the prospect of being able to simplify draped lines and have them take into account the Z dimension which the Area setting does.
I did not try this on polygons. | 1.0 | Area Simplify drops Z and M values - ### What is the bug or the crash?
If you run the Simplify Algo with Area(Visvalingam) the Z and M values will become null if any vertexes are removed.
### Steps to reproduce the issue
1. Create a line layer with Z at 100 and M values set and enough vertices to be able to run simplify.
2. Run the simplify algo and set Simplification method to Area(Visvalingam).
3. Set the tolerance to a value that will remove at least one vertex.
4. Check that the M and Z values are now null.
Test data:
[Area_Simplify_ZMdrop.zip](https://github.com/qgis/QGIS/files/9896735/Area_Simplify_ZMdrop.zip)
Starting layer with all Z set to 100

Simplified with some vertex's removed all Z now null or 0

Simplified with no vertex's removed all Z's still 100 (one vertex was set to 200 did the same thing without that)

### Versions
<!--StartFragment--><!DOCTYPE HTML PUBLIC "-//W3C//DTD HTML 4.0//EN" "http://www.w3.org/TR/REC-html40/strict.dtd">
<html><head><meta http-equiv="Content-Type" content="text/html; charset=utf-8" /><style type="text/css">
p, li { white-space: pre-wrap; }
</style></head><body>
QGIS version | 3.29.0-Master | QGIS code revision | ca130dd49c
-- | -- | -- | --
Qt version | 5.15.3
Python version | 3.9.5
GDAL/OGR version | 3.6.0dev-5df18be65c
PROJ version | 9.1.0
EPSG Registry database version | v10.074 (2022-08-01)
GEOS version | 3.10.3-CAPI-1.16.1
SQLite version | 3.39.4
PDAL version | 2.4.3
PostgreSQL client version | unknown
SpatiaLite version | 5.0.1
QWT version | 6.1.6
QScintilla2 version | 2.13.1
OS version | Windows 10 Version 2009
| | |
This copy of QGIS writes debugging output.
| | |
Active Python plugins
db_manager | 0.1.20
grassprovider | 2.12.99
MetaSearch | 0.3.6
otbprovider | 2.12.99
processing | 2.12.99
sagaprovider | 2.12.99
</body></html><!--EndFragment-->QGIS version
3.29.0-Master
QGIS code revision
[ca130dd49c](https://github.com/qgis/QGIS/commit/ca130dd49c)
Qt version
5.15.3
Python version
3.9.5
GDAL/OGR version
3.6.0dev-5df18be65c
PROJ version
9.1.0
EPSG Registry database version
v10.074 (2022-08-01)
GEOS version
3.10.3-CAPI-1.16.1
SQLite version
3.39.4
PDAL version
2.4.3
PostgreSQL client version
unknown
SpatiaLite version
5.0.1
QWT version
6.1.6
QScintilla2 version
2.13.1
OS version
Windows 10 Version 2009
This copy of QGIS writes debugging output.
Active Python plugins
db_manager
0.1.20
grassprovider
2.12.99
MetaSearch
0.3.6
otbprovider
2.12.99
processing
2.12.99
sagaprovider
2.12.99
### Supported QGIS version
- [X] I'm running a supported QGIS version according to the roadmap.
### New profile
- [X] I tried with a new QGIS profile
### Additional context
The Distance (Douglas-Puecker) stetting does not appear to be affected but it also doesn't seem to take into account Z values in simplification.
I was excited by the prospect of being able to simplify draped lines and have them take into account the Z dimension which the Area setting does.
I did not try this on polygons. | process | area simplify drops z and m values what is the bug or the crash if you run the simplify algo with area visvalingam the z and m values will become null if any vertexes are removed steps to reproduce the issue create a line layer with z at and m values set and enough vertices to be able to run simplify run the simplify algo and set simplification method to area visvalingam set the tolerance to a value that will remove at least one vertex check that the m and z values are now null test data starting layer with all z set to simplified with some vertex s removed all z now null or simplified with no vertex s removed all z s still one vertex was set to did the same thing without that versions doctype html public dtd html en p li white space pre wrap qgis version master qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version windows version this copy of qgis writes debugging output active python plugins db manager grassprovider metasearch otbprovider processing sagaprovider qgis version master qgis code revision qt version python version gdal ogr version proj version epsg registry database version geos version capi sqlite version pdal version postgresql client version unknown spatialite version qwt version version os version windows version this copy of qgis writes debugging output active python plugins db manager grassprovider metasearch otbprovider processing sagaprovider supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context the distance douglas puecker stetting does not appear to be affected but it also doesn t seem to take into account z values in simplification i was excited by the prospect of being able to simplify draped lines and have them take into account the z dimension which the area setting does i did not try this on polygons | 1 |
35,692 | 9,645,533,883 | IssuesEvent | 2019-05-17 08:54:50 | yandex/ClickHouse | https://api.github.com/repos/yandex/ClickHouse | opened | build failed, by clang 8.0.0 | build | hi, Thank you for your work.
I compiled in stable version (19.5.3.8 stable) wrong.
first, Download the code from LLVM and copy the code from the driver directory to compiler_8.0.0 and compiler_8.0.0svn.
second, The build code.
failed message:
CMake Error at dbms/programs/clang/Compiler-8.0.0svn/CMakeLists.txt:32 (add_clang_tool):
Unknown CMake command "add_clang_tool".
| 1.0 | build failed, by clang 8.0.0 - hi, Thank you for your work.
I compiled in stable version (19.5.3.8 stable) wrong.
first, Download the code from LLVM and copy the code from the driver directory to compiler_8.0.0 and compiler_8.0.0svn.
second, The build code.
failed message:
CMake Error at dbms/programs/clang/Compiler-8.0.0svn/CMakeLists.txt:32 (add_clang_tool):
Unknown CMake command "add_clang_tool".
| non_process | build failed by clang hi thank you for your work i compiled in stable version stable wrong first download the code from llvm and copy the code from the driver directory to compiler and compiler second the build code failed message cmake error at dbms programs clang compiler cmakelists txt add clang tool unknown cmake command add clang tool | 0 |
4,235 | 7,186,868,053 | IssuesEvent | 2018-02-02 01:33:03 | Great-Hill-Corporation/quickBlocks | https://api.github.com/repos/Great-Hill-Corporation/quickBlocks | closed | Transactions may fail without sending "out of gas". Internal calls does not use gas in parent (maybe) | monitors-all status-inprocess type-bug | An internal call in this transaction (0xd20a95a9f051d6961ca766a9b2d0f18b6afdf6d8465931a3f70f2c4b1c023688) ran out of gas, but this transaction did not run out of gas, so I don't identify them as inError. The gas!=gasUsed, therefore, I don't trace it. | 1.0 | Transactions may fail without sending "out of gas". Internal calls does not use gas in parent (maybe) - An internal call in this transaction (0xd20a95a9f051d6961ca766a9b2d0f18b6afdf6d8465931a3f70f2c4b1c023688) ran out of gas, but this transaction did not run out of gas, so I don't identify them as inError. The gas!=gasUsed, therefore, I don't trace it. | process | transactions may fail without sending out of gas internal calls does not use gas in parent maybe an internal call in this transaction ran out of gas but this transaction did not run out of gas so i don t identify them as inerror the gas gasused therefore i don t trace it | 1 |
71,670 | 7,254,797,535 | IssuesEvent | 2018-02-16 12:31:42 | edenlabllc/ehealth.api | https://api.github.com/repos/edenlabllc/ehealth.api | closed | Cancelled declaration_request with person w/o tax_id | epic/declaration kind/bug kind/support status/test | original request https://github.com/ehealth-ua/api-spec/issues/213
## Steps to Reproduce (Repro steps)
- [ ] POST https://api.ehealth.world/api/declaration_requests
- [ ] Створюємо декл.реквест для користувача без ІПН 63a3128c-cbbf-4398-a5ba-2598cf83f80e , не підтверджуємо його і потім зразу ж створюємо для іншого також без ІПН 6e2c79f5-b699-472e-969a-2ebee0700aa8. У них різні ПІБ, дати народження і стать. В результаті створення другої переша стає CANCELLED employee_id a1d9ddf1-8cab-4671-80e0-811f5a63df8d
[Запит і відповідь по 63a3128c-cbbf-4398-a5ba-2598cf83f80e.txt](https://github.com/ehealth-ua/api-spec/files/1635036/63a3128c-cbbf-4398-a5ba-2598cf83f80e.txt)
[Запит і відповідь по 6e2c79f5-b699-472e-969a-2ebee0700aa8.txt](https://github.com/ehealth-ua/api-spec/files/1635037/6e2c79f5-b699-472e-969a-2ebee0700aa8.txt)
## Actual result
- [ ] 63a3128c-cbbf-4398-a5ba-2598cf83f80e - перейшла в статус CANCELLED
## Expected result
- [ ] 63a3128c-cbbf-4398-a5ba-2598cf83f80e - залишилась в статусі NEW
------------------------
Can reproduce on DEV
Should add new validation for persons w/o tax_id
```
WHERE
IL_DB.declaration_requests.data.declaration_request.person.first_name = :($.declaration_request.person.first_name)
AND IL_DB.declaration_requests.data.declaration_request.person.last_name = :($.declaration_request.person.last_name)
AND IL_DB.declaration_requests.data.declaration_request.person.birth_date = :($.declaration_request.person.birth_date)
AND IL_DB.declaration_requests.data.declaration_request.legal_entity_id = :($.declaration_request.legal_entity_id)
AND IL_DB.declaration_requests.data.declaration_request.employee_id = :($.declaration_request.employee_id)
AND IL_DB.declaration_requests.status IN ('NEW', 'APPROVED'))
```
[spec](https://edenlab.atlassian.net/wiki/spaces/EH/pages/983335/IL.Create+declaration+request) | 1.0 | Cancelled declaration_request with person w/o tax_id - original request https://github.com/ehealth-ua/api-spec/issues/213
## Steps to Reproduce (Repro steps)
- [ ] POST https://api.ehealth.world/api/declaration_requests
- [ ] Створюємо декл.реквест для користувача без ІПН 63a3128c-cbbf-4398-a5ba-2598cf83f80e , не підтверджуємо його і потім зразу ж створюємо для іншого також без ІПН 6e2c79f5-b699-472e-969a-2ebee0700aa8. У них різні ПІБ, дати народження і стать. В результаті створення другої переша стає CANCELLED employee_id a1d9ddf1-8cab-4671-80e0-811f5a63df8d
[Запит і відповідь по 63a3128c-cbbf-4398-a5ba-2598cf83f80e.txt](https://github.com/ehealth-ua/api-spec/files/1635036/63a3128c-cbbf-4398-a5ba-2598cf83f80e.txt)
[Запит і відповідь по 6e2c79f5-b699-472e-969a-2ebee0700aa8.txt](https://github.com/ehealth-ua/api-spec/files/1635037/6e2c79f5-b699-472e-969a-2ebee0700aa8.txt)
## Actual result
- [ ] 63a3128c-cbbf-4398-a5ba-2598cf83f80e - перейшла в статус CANCELLED
## Expected result
- [ ] 63a3128c-cbbf-4398-a5ba-2598cf83f80e - залишилась в статусі NEW
------------------------
Can reproduce on DEV
Should add new validation for persons w/o tax_id
```
WHERE
IL_DB.declaration_requests.data.declaration_request.person.first_name = :($.declaration_request.person.first_name)
AND IL_DB.declaration_requests.data.declaration_request.person.last_name = :($.declaration_request.person.last_name)
AND IL_DB.declaration_requests.data.declaration_request.person.birth_date = :($.declaration_request.person.birth_date)
AND IL_DB.declaration_requests.data.declaration_request.legal_entity_id = :($.declaration_request.legal_entity_id)
AND IL_DB.declaration_requests.data.declaration_request.employee_id = :($.declaration_request.employee_id)
AND IL_DB.declaration_requests.status IN ('NEW', 'APPROVED'))
```
[spec](https://edenlab.atlassian.net/wiki/spaces/EH/pages/983335/IL.Create+declaration+request) | non_process | cancelled declaration request with person w o tax id original request steps to reproduce repro steps post створюємо декл реквест для користувача без іпн cbbf не підтверджуємо його і потім зразу ж створюємо для іншого також без іпн у них різні піб дати народження і стать в результаті створення другої переша стає cancelled employee id actual result cbbf перейшла в статус cancelled expected result cbbf залишилась в статусі new can reproduce on dev should add new validation for persons w o tax id where il db declaration requests data declaration request person first name declaration request person first name and il db declaration requests data declaration request person last name declaration request person last name and il db declaration requests data declaration request person birth date declaration request person birth date and il db declaration requests data declaration request legal entity id declaration request legal entity id and il db declaration requests data declaration request employee id declaration request employee id and il db declaration requests status in new approved | 0 |
38,778 | 15,793,582,500 | IssuesEvent | 2021-04-02 09:16:17 | AutoPacker-OSS/autopacker | https://api.github.com/repos/AutoPacker-OSS/autopacker | opened | "undefined Projects" when organization has no projects | Priority: Low Service: Web App Status: Pending Type: Bug | When an organization has no projects, its dashboard shows a wrong message:

Should say "No projects", or something like that. No "undefined" | 1.0 | "undefined Projects" when organization has no projects - When an organization has no projects, its dashboard shows a wrong message:

Should say "No projects", or something like that. No "undefined" | non_process | undefined projects when organization has no projects when an organization has no projects its dashboard shows a wrong message should say no projects or something like that no undefined | 0 |
88,777 | 17,663,608,706 | IssuesEvent | 2021-08-22 02:12:31 | adventuregamestudio/ags | https://api.github.com/repos/adventuregamestudio/ags | closed | INFO: a variant of utf8 support for translations | type: information context: unicode | A while ago @mgambrell has pointed me his work on hacking in utf-8 string support for translations. It's based on modifying alfont source to use parts of the **musl** library to convert strings.
Here's the commit (and maybe there are more commits around): https://github.com/RatalaikaGames/ags/commit/a00e7d933458d8360c17cf84c8cfbf5125111c3b
I am leaving this here only for the reference if someone would like to look into this at some point, because afaik the solution was not applied universally to whole engine, and it's hard to tell which nuances may be met when fully moving to utf8 support in both editor and engine (especially if you care about loading old games into new interpreter). | 1.0 | INFO: a variant of utf8 support for translations - A while ago @mgambrell has pointed me his work on hacking in utf-8 string support for translations. It's based on modifying alfont source to use parts of the **musl** library to convert strings.
Here's the commit (and maybe there are more commits around): https://github.com/RatalaikaGames/ags/commit/a00e7d933458d8360c17cf84c8cfbf5125111c3b
I am leaving this here only for the reference if someone would like to look into this at some point, because afaik the solution was not applied universally to whole engine, and it's hard to tell which nuances may be met when fully moving to utf8 support in both editor and engine (especially if you care about loading old games into new interpreter). | non_process | info a variant of support for translations a while ago mgambrell has pointed me his work on hacking in utf string support for translations it s based on modifying alfont source to use parts of the musl library to convert strings here s the commit and maybe there are more commits around i am leaving this here only for the reference if someone would like to look into this at some point because afaik the solution was not applied universally to whole engine and it s hard to tell which nuances may be met when fully moving to support in both editor and engine especially if you care about loading old games into new interpreter | 0 |
5,822 | 8,657,197,579 | IssuesEvent | 2018-11-27 20:36:28 | brucemiller/LaTeXML | https://api.github.com/repos/brucemiller/LaTeXML | closed | Page break compatibility? | enhancement postprocessing | Hi there, this is a great program!
In case the authors are not aware, LaTeXML can also be used as part of the process to produce ebooks! Since it produces HTML, it is just a matter of converting the HTML into EPUB or other format using tools like Calibre.
Anyway, my question relates to commands like `\pagebreak` and `\newpage`. It seems LaTeXML takes no notice of these commands, as they do not produce HTML commands related to them (no surprise, since a page break in an HTML document makes little sense!). An example of such command would be the one presented in [this page](https://blog.epubandebookhelp.com/2011/12/21/epub-formatting-how-to-add-a-page-break-to-an-epub-document-using-html-or-css/). What LaTeXML can do is to produce separate files in terms of sections (e.g. based on `\chapter{}`). However, when producing an ebook, you just would like to add a `\pagebreak` command rather than adding a section for the title page, copyright page, acknowledgment and other minor content. Adding a section introduces them into the table of contents of the ebook, which is not desired.
At the moment I'm using sections for everything and modifying the TOC manually in Calibre, so this is not a big issue. But I wonder if there is a plan for LaTeXML to allow for page breaks (perhaps as suggested in the linked page)?
Best regards.
| 1.0 | Page break compatibility? - Hi there, this is a great program!
In case the authors are not aware, LaTeXML can also be used as part of the process to produce ebooks! Since it produces HTML, it is just a matter of converting the HTML into EPUB or other format using tools like Calibre.
Anyway, my question relates to commands like `\pagebreak` and `\newpage`. It seems LaTeXML takes no notice of these commands, as they do not produce HTML commands related to them (no surprise, since a page break in an HTML document makes little sense!). An example of such command would be the one presented in [this page](https://blog.epubandebookhelp.com/2011/12/21/epub-formatting-how-to-add-a-page-break-to-an-epub-document-using-html-or-css/). What LaTeXML can do is to produce separate files in terms of sections (e.g. based on `\chapter{}`). However, when producing an ebook, you just would like to add a `\pagebreak` command rather than adding a section for the title page, copyright page, acknowledgment and other minor content. Adding a section introduces them into the table of contents of the ebook, which is not desired.
At the moment I'm using sections for everything and modifying the TOC manually in Calibre, so this is not a big issue. But I wonder if there is a plan for LaTeXML to allow for page breaks (perhaps as suggested in the linked page)?
Best regards.
| process | page break compatibility hi there this is a great program in case the authors are not aware latexml can also be used as part of the process to produce ebooks since it produces html it is just a matter of converting the html into epub or other format using tools like calibre anyway my question relates to commands like pagebreak and newpage it seems latexml takes no notice of these commands as they do not produce html commands related to them no surprise since a page break in an html document makes little sense an example of such command would be the one presented in what latexml can do is to produce separate files in terms of sections e g based on chapter however when producing an ebook you just would like to add a pagebreak command rather than adding a section for the title page copyright page acknowledgment and other minor content adding a section introduces them into the table of contents of the ebook which is not desired at the moment i m using sections for everything and modifying the toc manually in calibre so this is not a big issue but i wonder if there is a plan for latexml to allow for page breaks perhaps as suggested in the linked page best regards | 1 |
17,342 | 23,165,852,310 | IssuesEvent | 2022-07-30 00:57:46 | hashgraph/hedera-json-rpc-relay | https://api.github.com/repos/hashgraph/hedera-json-rpc-relay | closed | Add consensus node calling endpoints to postman tests | enhancement P2 process | ### Problem
Currently the postman tests are missing calls that query the consensus nodes.
This would be useful in quick sanity checks to verify relay operation as they do not require knowledge of accounts and blocks details
### Solution
Add consensus node calling checks
- `eth_gasPrice`
- `eth_feeHistory` - use 3 last blocks
Fix `base_url` vs `baseUrl` inconsistency between the `README.md` and `postman.js` files
### Alternatives
_No response_ | 1.0 | Add consensus node calling endpoints to postman tests - ### Problem
Currently the postman tests are missing calls that query the consensus nodes.
This would be useful in quick sanity checks to verify relay operation as they do not require knowledge of accounts and blocks details
### Solution
Add consensus node calling checks
- `eth_gasPrice`
- `eth_feeHistory` - use 3 last blocks
Fix `base_url` vs `baseUrl` inconsistency between the `README.md` and `postman.js` files
### Alternatives
_No response_ | process | add consensus node calling endpoints to postman tests problem currently the postman tests are missing calls that query the consensus nodes this would be useful in quick sanity checks to verify relay operation as they do not require knowledge of accounts and blocks details solution add consensus node calling checks eth gasprice eth feehistory use last blocks fix base url vs baseurl inconsistency between the readme md and postman js files alternatives no response | 1 |
13,681 | 2,775,782,979 | IssuesEvent | 2015-05-04 18:03:27 | douglasdrumond/macvim | https://api.github.com/repos/douglasdrumond/macvim | closed | Auto-constraints warning under Yosemite | auto-migrated Priority-Medium Type-Defect | ```
Got this error when running MacVim under Yosemite:
Unable to simultaneously satisfy constraints:
(
"<NSAutoresizingMaskLayoutConstraint:0x7fe499dcde70 h=--& v=--& V:[FI_TBrowserBackgroundView:0x7fe49c5793a0(0)]>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe49c4146e0 h=-&- v=-&- V:|-(0)-[FI_TListScrollView:0x7fe499df0480] (Names: '|':FI_TBrowserBackgroundView:0x7fe49c5793a0 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe49c4070c0 h=-&- v=-&- V:|-(0)-[NSClipView:0x7fe499df00d0] (Names: '|':FI_TListScrollView:0x7fe499df0480 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe499e0d2b0 h=-&- v=-&- V:[NSClipView:0x7fe499df00d0]-(0)-| (Names: '|':FI_TListScrollView:0x7fe499df0480 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe49c414690 h=-&- v=-&- V:[FI_TListScrollView:0x7fe499df0480]-(0)-| (Names: '|':FI_TBrowserBackgroundView:0x7fe49c5793a0 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe499ed8290 h=-&- v=-&- V:[FI_TListView:0x7fe499defad0]-(17)-| (Names: '|':NSClipView:0x7fe499df00d0 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe499ed8240 h=-&- v=-&- V:|-(0)-[FI_TListView:0x7fe499defad0] (Names: '|':NSClipView:0x7fe499df00d0 )>"
)
Will attempt to recover by breaking constraint
<NSAutoresizingMaskLayoutConstraint:0x7fe499ed8290 h=-&- v=-&-
V:[FI_TListView:0x7fe499defad0]-(17)-| (Names: '|':NSClipView:0x7fe499df00d0
)>
Set the NSUserDefault
NSConstraintBasedLayoutVisualizeMutuallyExclusiveConstraints to YES to have
-[NSWindow visualizeConstraints:] automatically called when this happens.
And/or, break on objc_exception_throw to catch this in the debugger.
```
Original issue reported on code.google.com by `fushunpoon@googlemail.com` on 14 Dec 2014 at 1:19 | 1.0 | Auto-constraints warning under Yosemite - ```
Got this error when running MacVim under Yosemite:
Unable to simultaneously satisfy constraints:
(
"<NSAutoresizingMaskLayoutConstraint:0x7fe499dcde70 h=--& v=--& V:[FI_TBrowserBackgroundView:0x7fe49c5793a0(0)]>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe49c4146e0 h=-&- v=-&- V:|-(0)-[FI_TListScrollView:0x7fe499df0480] (Names: '|':FI_TBrowserBackgroundView:0x7fe49c5793a0 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe49c4070c0 h=-&- v=-&- V:|-(0)-[NSClipView:0x7fe499df00d0] (Names: '|':FI_TListScrollView:0x7fe499df0480 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe499e0d2b0 h=-&- v=-&- V:[NSClipView:0x7fe499df00d0]-(0)-| (Names: '|':FI_TListScrollView:0x7fe499df0480 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe49c414690 h=-&- v=-&- V:[FI_TListScrollView:0x7fe499df0480]-(0)-| (Names: '|':FI_TBrowserBackgroundView:0x7fe49c5793a0 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe499ed8290 h=-&- v=-&- V:[FI_TListView:0x7fe499defad0]-(17)-| (Names: '|':NSClipView:0x7fe499df00d0 )>",
"<NSAutoresizingMaskLayoutConstraint:0x7fe499ed8240 h=-&- v=-&- V:|-(0)-[FI_TListView:0x7fe499defad0] (Names: '|':NSClipView:0x7fe499df00d0 )>"
)
Will attempt to recover by breaking constraint
<NSAutoresizingMaskLayoutConstraint:0x7fe499ed8290 h=-&- v=-&-
V:[FI_TListView:0x7fe499defad0]-(17)-| (Names: '|':NSClipView:0x7fe499df00d0
)>
Set the NSUserDefault
NSConstraintBasedLayoutVisualizeMutuallyExclusiveConstraints to YES to have
-[NSWindow visualizeConstraints:] automatically called when this happens.
And/or, break on objc_exception_throw to catch this in the debugger.
```
Original issue reported on code.google.com by `fushunpoon@googlemail.com` on 14 Dec 2014 at 1:19 | non_process | auto constraints warning under yosemite got this error when running macvim under yosemite unable to simultaneously satisfy constraints will attempt to recover by breaking constraint nsautoresizingmasklayoutconstraint h v v names nsclipview set the nsuserdefault nsconstraintbasedlayoutvisualizemutuallyexclusiveconstraints to yes to have automatically called when this happens and or break on objc exception throw to catch this in the debugger original issue reported on code google com by fushunpoon googlemail com on dec at | 0 |
22,105 | 30,635,736,103 | IssuesEvent | 2023-07-24 17:36:32 | metabase/metabase | https://api.github.com/repos/metabase/metabase | closed | [MLv2] [Bug] Column name calculation is wrong for columns in `:breakout` from joins | Type:Bug .Backend .metabase-lib .Team/QueryProcessor :hammer_and_wrench: | Note: uses new mock metadata from #29906
Failing test:
```clj
(deftest ^:parallel metadata-for-join-in-previous-stage-test
(let [query {:lib/type :mbql/query
:lib/metadata meta/metadata-provider
:database (meta/id)
:type :pipeline
:stages [{:lib/type :mbql.stage/mbql
:source-table (meta/id :orders)
:joins [{:lib/type :mbql/join
:lib/options {:lib/uuid (str (random-uuid))}
:stages [{:lib/type :mbql.stage/mbql, :source-table (meta/id :products)}]
:alias "P1"
:conditions [[:=
{:lib/uuid (str (random-uuid))}
[:field
{:lib/uuid (str (random-uuid))}
(meta/id :orders :product-id)]
[:field
{:join-alias "P1"
:lib/uuid (str (random-uuid))}
(meta/id :products :id)]]]
:fields :all}
{:lib/type :mbql/join
:lib/options {:lib/uuid (str (random-uuid))}
:stages [{:lib/type :mbql.stage/mbql, :source-table (meta/id :people)}]
:alias "People"
:conditions [[:=
{:lib/uuid (str (random-uuid))}
[:field
{:lib/uuid (str (random-uuid))}
(meta/id :orders :user-id)]
[:field
{:join-alias "People"
:lib/uuid (str (random-uuid))}
(meta/id :people :id)]]]
:fields :all}]
:breakout [[:field
{:join-alias "P1", :lib/uuid (str (random-uuid))}
(meta/id :products :category)]
[:field
{:join-alias "People", :lib/uuid (str (random-uuid))}
(meta/id :people :source)]]
:aggregation [[:count {:lib/uuid (str (random-uuid))}]]}]}]
(is (= ["P1__CATEGORY" "People__SOURCE" "count"]
(mapv :lib/desired-column-alias (lib.metadata.calculation/returned-columns query))))
(let [query' (update query :stages conj {:lib/type :mbql.stage/mbql
:joins [{:lib/type :mbql/join
:lib/options {:lib/uuid (str (random-uuid))}
:stages [{:lib/type :mbql.stage/mbql
:source-table (meta/id :reviews)
:joins [{:lib/type :mbql/join
:lib/options {:lib/uuid (str (random-uuid))}
:stages [{:lib/type :mbql.stage/mbql
:source-table (meta/id :products)}]
:alias "P2"
:conditions [[:=
{:lib/uuid (str (random-uuid))}
[:field
{:lib/uuid (str (random-uuid))}
(meta/id :reviews :product-id)]
[:field
{:join-alias "P2"
:lib/uuid (str (random-uuid))}
(meta/id :products :id)]]]
:fields :all}]
:breakout [[:field
{:join-alias "P2"
:lib/uuid (str (random-uuid))}
(meta/id :products :category)]]
:aggregation [[:avg
{:lib/uuid (str (random-uuid))}
[:field
{:lib/uuid (str (random-uuid))}
(meta/id :reviews :rating)]]]}
{:lib/type :mbql.stage/mbql}]
:alias "Q2"
:conditions [[:=
{:lib/uuid (str (random-uuid))}
[:field
{:join-alias "P1", :lib/uuid (str (random-uuid))}
(meta/id :products :category)]
[:field
{:join-alias "Q2", :lib/uuid (str (random-uuid))}
(meta/id :products :category)]]]
:fields :all}]
:order-by [[:asc
{:lib/uuid (str (random-uuid))}
[:field
{:join-alias "P1", :lib/uuid (str (random-uuid))}
(meta/id :products :category)]]
[:asc
{:lib/uuid (str (random-uuid))}
[:field
{:join-alias "People", :lib/uuid (str (random-uuid))}
(meta/id :people :source)]]]
:limit 2})]
(is (= ["P1__CATEGORY" "People__SOURCE" "count" "Q2__P2__CATEGORY" "Q2__avg"]
(mapv :lib/desired-column-alias (lib.metadata.calculation/returned-columns query')))))))
```
Failures:
```clj
Fail in metadata-for-join-in-previous-stage-test
expected: ["P1__CATEGORY" "People__SOURCE" "count"]
actual: ["CATEGORY" "SOURCE" "count"]
diff: - ["P1__CATEGORY" "People__SOURCE"]
+ ["CATEGORY" "SOURCE"]
Fail in metadata-for-join-in-previous-stage-test
expected: ["P1__CATEGORY" "People__SOURCE" "count" "Q2__P2__CATEGORY" "Q2__avg"]
actual: ["CATEGORY" "SOURCE" "count" "Q2__CATEGORY" "Q2__avg"]
diff: - ["P1__CATEGORY" "People__SOURCE" nil "Q2__P2__CATEGORY"]
+ ["CATEGORY" "SOURCE" nil "Q2__CATEGORY"]
``` | 1.0 | [MLv2] [Bug] Column name calculation is wrong for columns in `:breakout` from joins - Note: uses new mock metadata from #29906
Failing test:
```clj
(deftest ^:parallel metadata-for-join-in-previous-stage-test
(let [query {:lib/type :mbql/query
:lib/metadata meta/metadata-provider
:database (meta/id)
:type :pipeline
:stages [{:lib/type :mbql.stage/mbql
:source-table (meta/id :orders)
:joins [{:lib/type :mbql/join
:lib/options {:lib/uuid (str (random-uuid))}
:stages [{:lib/type :mbql.stage/mbql, :source-table (meta/id :products)}]
:alias "P1"
:conditions [[:=
{:lib/uuid (str (random-uuid))}
[:field
{:lib/uuid (str (random-uuid))}
(meta/id :orders :product-id)]
[:field
{:join-alias "P1"
:lib/uuid (str (random-uuid))}
(meta/id :products :id)]]]
:fields :all}
{:lib/type :mbql/join
:lib/options {:lib/uuid (str (random-uuid))}
:stages [{:lib/type :mbql.stage/mbql, :source-table (meta/id :people)}]
:alias "People"
:conditions [[:=
{:lib/uuid (str (random-uuid))}
[:field
{:lib/uuid (str (random-uuid))}
(meta/id :orders :user-id)]
[:field
{:join-alias "People"
:lib/uuid (str (random-uuid))}
(meta/id :people :id)]]]
:fields :all}]
:breakout [[:field
{:join-alias "P1", :lib/uuid (str (random-uuid))}
(meta/id :products :category)]
[:field
{:join-alias "People", :lib/uuid (str (random-uuid))}
(meta/id :people :source)]]
:aggregation [[:count {:lib/uuid (str (random-uuid))}]]}]}]
(is (= ["P1__CATEGORY" "People__SOURCE" "count"]
(mapv :lib/desired-column-alias (lib.metadata.calculation/returned-columns query))))
(let [query' (update query :stages conj {:lib/type :mbql.stage/mbql
:joins [{:lib/type :mbql/join
:lib/options {:lib/uuid (str (random-uuid))}
:stages [{:lib/type :mbql.stage/mbql
:source-table (meta/id :reviews)
:joins [{:lib/type :mbql/join
:lib/options {:lib/uuid (str (random-uuid))}
:stages [{:lib/type :mbql.stage/mbql
:source-table (meta/id :products)}]
:alias "P2"
:conditions [[:=
{:lib/uuid (str (random-uuid))}
[:field
{:lib/uuid (str (random-uuid))}
(meta/id :reviews :product-id)]
[:field
{:join-alias "P2"
:lib/uuid (str (random-uuid))}
(meta/id :products :id)]]]
:fields :all}]
:breakout [[:field
{:join-alias "P2"
:lib/uuid (str (random-uuid))}
(meta/id :products :category)]]
:aggregation [[:avg
{:lib/uuid (str (random-uuid))}
[:field
{:lib/uuid (str (random-uuid))}
(meta/id :reviews :rating)]]]}
{:lib/type :mbql.stage/mbql}]
:alias "Q2"
:conditions [[:=
{:lib/uuid (str (random-uuid))}
[:field
{:join-alias "P1", :lib/uuid (str (random-uuid))}
(meta/id :products :category)]
[:field
{:join-alias "Q2", :lib/uuid (str (random-uuid))}
(meta/id :products :category)]]]
:fields :all}]
:order-by [[:asc
{:lib/uuid (str (random-uuid))}
[:field
{:join-alias "P1", :lib/uuid (str (random-uuid))}
(meta/id :products :category)]]
[:asc
{:lib/uuid (str (random-uuid))}
[:field
{:join-alias "People", :lib/uuid (str (random-uuid))}
(meta/id :people :source)]]]
:limit 2})]
(is (= ["P1__CATEGORY" "People__SOURCE" "count" "Q2__P2__CATEGORY" "Q2__avg"]
(mapv :lib/desired-column-alias (lib.metadata.calculation/returned-columns query')))))))
```
Failures:
```clj
Fail in metadata-for-join-in-previous-stage-test
expected: ["P1__CATEGORY" "People__SOURCE" "count"]
actual: ["CATEGORY" "SOURCE" "count"]
diff: - ["P1__CATEGORY" "People__SOURCE"]
+ ["CATEGORY" "SOURCE"]
Fail in metadata-for-join-in-previous-stage-test
expected: ["P1__CATEGORY" "People__SOURCE" "count" "Q2__P2__CATEGORY" "Q2__avg"]
actual: ["CATEGORY" "SOURCE" "count" "Q2__CATEGORY" "Q2__avg"]
diff: - ["P1__CATEGORY" "People__SOURCE" nil "Q2__P2__CATEGORY"]
+ ["CATEGORY" "SOURCE" nil "Q2__CATEGORY"]
``` | process | column name calculation is wrong for columns in breakout from joins note uses new mock metadata from failing test clj deftest parallel metadata for join in previous stage test let query lib type mbql query lib metadata meta metadata provider database meta id type pipeline stages lib type mbql stage mbql source table meta id orders joins lib type mbql join lib options lib uuid str random uuid stages alias conditions lib uuid str random uuid field lib uuid str random uuid meta id orders product id field join alias lib uuid str random uuid meta id products id fields all lib type mbql join lib options lib uuid str random uuid stages alias people conditions lib uuid str random uuid field lib uuid str random uuid meta id orders user id field join alias people lib uuid str random uuid meta id people id fields all breakout field join alias lib uuid str random uuid meta id products category field join alias people lib uuid str random uuid meta id people source aggregation is mapv lib desired column alias lib metadata calculation returned columns query let query update query stages conj lib type mbql stage mbql joins lib type mbql join lib options lib uuid str random uuid stages lib type mbql stage mbql source table meta id reviews joins lib type mbql join lib options lib uuid str random uuid stages lib type mbql stage mbql source table meta id products alias conditions lib uuid str random uuid field lib uuid str random uuid meta id reviews product id field join alias lib uuid str random uuid meta id products id fields all breakout field join alias lib uuid str random uuid meta id products category aggregation avg lib uuid str random uuid field lib uuid str random uuid meta id reviews rating lib type mbql stage mbql alias conditions lib uuid str random uuid field join alias lib uuid str random uuid meta id products category field join alias lib uuid str random uuid meta id products category fields all order by asc lib uuid str random uuid field join alias lib uuid str random uuid meta id products category asc lib uuid str random uuid field join alias people lib uuid str random uuid meta id people source limit is mapv lib desired column alias lib metadata calculation returned columns query failures clj fail in metadata for join in previous stage test expected actual diff fail in metadata for join in previous stage test expected actual diff | 1 |
18,676 | 24,594,078,527 | IssuesEvent | 2022-10-14 06:40:06 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [FHIR] FHIR viewer > Questionnaire > 'Spatial span memory' activity is not getting displayed | Bug P0 Response datastore Process: Fixed Process: Tested dev | **Pre-condition:** Add 'Spatial span memory' activity in the study builder and publish the study
**AR:** FHIR viewer > Questionnaire > Spatial span memory activity is not getting displayed
**ER:** FHIR viewer > Questionnaire > Spatial span memory activity should get displayed | 2.0 | [FHIR] FHIR viewer > Questionnaire > 'Spatial span memory' activity is not getting displayed - **Pre-condition:** Add 'Spatial span memory' activity in the study builder and publish the study
**AR:** FHIR viewer > Questionnaire > Spatial span memory activity is not getting displayed
**ER:** FHIR viewer > Questionnaire > Spatial span memory activity should get displayed | process | fhir viewer questionnaire spatial span memory activity is not getting displayed pre condition add spatial span memory activity in the study builder and publish the study ar fhir viewer questionnaire spatial span memory activity is not getting displayed er fhir viewer questionnaire spatial span memory activity should get displayed | 1 |
16,485 | 21,443,608,013 | IssuesEvent | 2022-04-25 02:12:06 | huutho77/CNPMNC_ThayAi | https://api.github.com/repos/huutho77/CNPMNC_ThayAi | opened | [API] Coding feature filter products based on Category | dev/thnguyen processing | - Parameters need to be input:
- [ ] categoryID
- Output expect:
- [ ] Products list related has been found | 1.0 | [API] Coding feature filter products based on Category - - Parameters need to be input:
- [ ] categoryID
- Output expect:
- [ ] Products list related has been found | process | coding feature filter products based on category parameters need to be input categoryid output expect products list related has been found | 1 |
17,681 | 23,515,865,320 | IssuesEvent | 2022-08-18 21:17:26 | open-telemetry/opentelemetry-collector-contrib | https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib | closed | [processor/metricstransform] aggregate_labels inconsistent results | bug priority:p2 data:metrics processor/metricstransform | **Describe the bug**
Trying to aggregate a histogram using metricstransform (aggregate_labels) produce inconsistent results.
**Steps to reproduce**
Use a prometheus recivier to scrape a histogram that have multiple values like this:
<details><summary>METRICS</summary>
<p>
```
# HELP Response_Time Operation Duration
# TYPE Response_Time histogram
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="400"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="500"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="600"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow"} 350
Response_Time_count{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="400"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="500"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="600"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow"} 378
Response_Time_count{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="400"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="500"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="600"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow"} 603
Response_Time_count{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="400"} 0
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="500"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="600"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow"} 454
Response_Time_count{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="400"} 0
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="500"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="600"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow"} 456
Response_Time_count{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow"} 1
```
</p>
</details>
and after that export it to prometheus using metricstransform processors (aggregate_labels, label_set: [operationName], aggregation_type: sum)
**What did you expect to see?**
I expect to see 1 histogram with all metrics aggregated like:
<details><summary>METRICS</summary>
<p>
```
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="100"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="200"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="300"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="400"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="500"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="600"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="700"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="800"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="900"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="1000"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="1500"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="2000"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="5000"} 5
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="+Inf"} 5
Response_Time_sum_sum{instance="foo",job="bar",operationName="flow"} 2241
Response_Time_sum_count{instance="foo",job="bar",operationName="flow"} 5
```
</p>
</details>
**What did you see instead?**
It produces 1 histogram but values are always wrong, sometimes its aggregates (sum) just some values, and other times it doesn't aggregate at all.
Like this:
<details><summary>METRICS</summary>
<p>
Here it aggregates just 2 and drop others...
```
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="100"} 0
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="200"} 0
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="300"} 0
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="400"} 0
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="500"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="600"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="700"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="800"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="900"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="1000"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="1500"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="2000"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="5000"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="+Inf"} 2
Response_Time_sum_sum{instance="foo",job="bar",operationName="flow"} 910
Response_Time_sum_count{instance="foo",job="bar",operationName="flow"} 2
```
</p>
</details>
<details><summary>Step by Step</summary>
<p>
Receiver: 0 Values:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066733-52aca324-ca1c-4f36-a047-d0818d70dbb5.png">
Exporter Result:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066769-c49bf66d-ed98-4a02-a6ca-70e6c820692d.png">
Receiver: 1 Histogram Value
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066840-d0a1f2d5-274f-4a00-941c-c7cabd8022c1.png">
Exporter Result:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066895-62361850-894b-41dd-aa00-800b9c9ba2f0.png">
Receiver: 3 Histogram Value
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066967-fdf7fe7d-7586-438c-8d28-f54b19182117.png">
Exporter Result:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180067042-83f48eb9-db2c-46fd-be41-27400bacb2d2.png">
Receiver: 5 Histogram Value
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180067199-56619062-ba7d-4548-94ee-45993555b24f.png">
Exporter Result:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180067275-57aafa88-f7a6-4f8e-bf20-cd7ff9c83ff2.png">
scrape_samples_scraped = 80 but just 1 histogram is aggregated
</p>
</details>
**What version did you use?**
Version: 0.55.0
**What config did you use?**
```yaml
receivers:
prometheus/local:
config:
scrape_configs:
- job_name: "local_metrics"
scrape_interval: 10s
static_configs:
- targets: ["xxxx:8888"]
exporters:
prometheus/public:
endpoint: "0.0.0.0:9464"
logging:
loglevel: debug
processors:
batch:
metricstransform:
transforms:
- include: Response_Time
action: update
operations:
- action: aggregate_labels
label_set: [operationName]
aggregation_type: sum
service:
telemetry:
logs:
level: debug
pipelines:
metrics/:
receivers: [prometheus/local]
exporters: [prometheus/public]
processors: [metricstransform]
```
**Environment**
OS: macOS 12.4
Running on docker, image: otel/opentelemetry-collector-contrib:0.55.0
docker ver: 4.9.0
| 1.0 | [processor/metricstransform] aggregate_labels inconsistent results - **Describe the bug**
Trying to aggregate a histogram using metricstransform (aggregate_labels) produce inconsistent results.
**Steps to reproduce**
Use a prometheus recivier to scrape a histogram that have multiple values like this:
<details><summary>METRICS</summary>
<p>
```
# HELP Response_Time Operation Duration
# TYPE Response_Time histogram
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="400"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="500"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="600"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow"} 350
Response_Time_count{hostname="10b2d401-03c0-4136-937f-96afeac251ad",job="my_job",operationName="flow"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="400"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="500"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="600"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow"} 378
Response_Time_count{hostname="4ddc318f-613c-473c-a8a8-b1690f61bc1d",job="my_job",operationName="flow"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="400"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="500"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="600"} 0
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow"} 603
Response_Time_count{hostname="7a2d8927-3894-4a8c-b4eb-ddaaca058862",job="my_job",operationName="flow"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="400"} 0
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="500"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="600"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow"} 454
Response_Time_count{hostname="a8eafe88-804b-4e49-a489-d69df1736ba0",job="my_job",operationName="flow"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="100"} 0
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="200"} 0
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="300"} 0
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="400"} 0
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="500"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="600"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="700"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="800"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="900"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="1000"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="1500"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="2000"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="5000"} 1
Response_Time_bucket{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow",le="+Inf"} 1
Response_Time_sum{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow"} 456
Response_Time_count{hostname="d58086b4-6728-40ba-ac39-2f6782d05abb",job="my_job",operationName="flow"} 1
```
</p>
</details>
and after that export it to prometheus using metricstransform processors (aggregate_labels, label_set: [operationName], aggregation_type: sum)
**What did you expect to see?**
I expect to see 1 histogram with all metrics aggregated like:
<details><summary>METRICS</summary>
<p>
```
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="100"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="200"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="300"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="400"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="500"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="600"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="700"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="800"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="900"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="1000"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="1500"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="2000"} x
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="5000"} 5
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="+Inf"} 5
Response_Time_sum_sum{instance="foo",job="bar",operationName="flow"} 2241
Response_Time_sum_count{instance="foo",job="bar",operationName="flow"} 5
```
</p>
</details>
**What did you see instead?**
It produces 1 histogram but values are always wrong, sometimes its aggregates (sum) just some values, and other times it doesn't aggregate at all.
Like this:
<details><summary>METRICS</summary>
<p>
Here it aggregates just 2 and drop others...
```
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="100"} 0
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="200"} 0
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="300"} 0
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="400"} 0
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="500"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="600"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="700"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="800"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="900"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="1000"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="1500"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="2000"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="5000"} 2
Response_Time_sum_bucket{instance="foo",job="bar",operationName="flow",le="+Inf"} 2
Response_Time_sum_sum{instance="foo",job="bar",operationName="flow"} 910
Response_Time_sum_count{instance="foo",job="bar",operationName="flow"} 2
```
</p>
</details>
<details><summary>Step by Step</summary>
<p>
Receiver: 0 Values:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066733-52aca324-ca1c-4f36-a047-d0818d70dbb5.png">
Exporter Result:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066769-c49bf66d-ed98-4a02-a6ca-70e6c820692d.png">
Receiver: 1 Histogram Value
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066840-d0a1f2d5-274f-4a00-941c-c7cabd8022c1.png">
Exporter Result:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066895-62361850-894b-41dd-aa00-800b9c9ba2f0.png">
Receiver: 3 Histogram Value
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180066967-fdf7fe7d-7586-438c-8d28-f54b19182117.png">
Exporter Result:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180067042-83f48eb9-db2c-46fd-be41-27400bacb2d2.png">
Receiver: 5 Histogram Value
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180067199-56619062-ba7d-4548-94ee-45993555b24f.png">
Exporter Result:
<img width="1136" alt="image" src="https://user-images.githubusercontent.com/8527149/180067275-57aafa88-f7a6-4f8e-bf20-cd7ff9c83ff2.png">
scrape_samples_scraped = 80 but just 1 histogram is aggregated
</p>
</details>
**What version did you use?**
Version: 0.55.0
**What config did you use?**
```yaml
receivers:
prometheus/local:
config:
scrape_configs:
- job_name: "local_metrics"
scrape_interval: 10s
static_configs:
- targets: ["xxxx:8888"]
exporters:
prometheus/public:
endpoint: "0.0.0.0:9464"
logging:
loglevel: debug
processors:
batch:
metricstransform:
transforms:
- include: Response_Time
action: update
operations:
- action: aggregate_labels
label_set: [operationName]
aggregation_type: sum
service:
telemetry:
logs:
level: debug
pipelines:
metrics/:
receivers: [prometheus/local]
exporters: [prometheus/public]
processors: [metricstransform]
```
**Environment**
OS: macOS 12.4
Running on docker, image: otel/opentelemetry-collector-contrib:0.55.0
docker ver: 4.9.0
| process | aggregate labels inconsistent results describe the bug trying to aggregate a histogram using metricstransform aggregate labels produce inconsistent results steps to reproduce use a prometheus recivier to scrape a histogram that have multiple values like this metrics help response time operation duration type response time histogram response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le inf response time sum hostname job my job operationname flow response time count hostname job my job operationname flow response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le inf response time sum hostname job my job operationname flow response time count hostname job my job operationname flow response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le inf response time sum hostname job my job operationname flow response time count hostname job my job operationname flow response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le inf response time sum hostname job my job operationname flow response time count hostname job my job operationname flow response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le response time bucket hostname job my job operationname flow le inf response time sum hostname job my job operationname flow response time count hostname job my job operationname flow and after that export it to prometheus using metricstransform processors aggregate labels label set aggregation type sum what did you expect to see i expect to see histogram with all metrics aggregated like metrics response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le x response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le inf response time sum sum instance foo job bar operationname flow response time sum count instance foo job bar operationname flow what did you see instead it produces histogram but values are always wrong sometimes its aggregates sum just some values and other times it doesn t aggregate at all like this metrics here it aggregates just and drop others response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le response time sum bucket instance foo job bar operationname flow le inf response time sum sum instance foo job bar operationname flow response time sum count instance foo job bar operationname flow step by step receiver values img width alt image src exporter result img width alt image src receiver histogram value img width alt image src exporter result img width alt image src receiver histogram value img width alt image src exporter result img width alt image src receiver histogram value img width alt image src exporter result img width alt image src scrape samples scraped but just histogram is aggregated what version did you use version what config did you use yaml receivers prometheus local config scrape configs job name local metrics scrape interval static configs targets exporters prometheus public endpoint logging loglevel debug processors batch metricstransform transforms include response time action update operations action aggregate labels label set aggregation type sum service telemetry logs level debug pipelines metrics receivers exporters processors environment os macos running on docker image otel opentelemetry collector contrib docker ver | 1 |
2,672 | 5,476,133,327 | IssuesEvent | 2017-03-11 17:53:51 | paulkornikov/Pragonas | https://api.github.com/repos/paulkornikov/Pragonas | closed | Fonction d'export des opés déductibles impôts | a-enhancement processus workload V | Filtrer les opérations déductibles puis produire le csv.
Sauvegarder le csv sur le serveur.
Mettre la location de mapping dans le XML de config. | 1.0 | Fonction d'export des opés déductibles impôts - Filtrer les opérations déductibles puis produire le csv.
Sauvegarder le csv sur le serveur.
Mettre la location de mapping dans le XML de config. | process | fonction d export des opés déductibles impôts filtrer les opérations déductibles puis produire le csv sauvegarder le csv sur le serveur mettre la location de mapping dans le xml de config | 1 |
16,863 | 22,143,037,764 | IssuesEvent | 2022-06-03 08:58:05 | camunda/zeebe | https://api.github.com/repos/camunda/zeebe | closed | [EPIC] Refactor Elasticsearch Exporter Tests | kind/toil scope/broker kind/epic team/process-automation area/test | **Description**
The `elasticsearch-exporter` integration tests are currently quite convoluted. Historically, the `elasticsearch-exporter` tests used to depend on `zeebe-test` which was removed in https://github.com/camunda-cloud/zeebe/pull/8608. During the removal, some of the `zeebe-test` utilities ended up in the `elasticsearch-exporter` module where we also needed to add a lot of test scoped dependencies.
The main goals here were:
- Cover as much as possible via only unit tests. This gives us a faster feedback cycle when changing something - developers can focus on first fixing unit tests, which have smaller blast radius and faster execution, before even looking at integration tests.
- Narrow the scope of the integration tests as much as possible. This meant getting rid of the broker, workflow engine semantics, etc., and just testing the integration of the exporter instance with Elasticsearch.
In general, I wanted to keep the same coverage as before. I think the coverage is now higher, but that was not the original motivation.
**Breakdown**
- [x] #9162
- [x] #9315
- [x] #9316
- [x] #9317
- [x] #9318
- [x] #9319
- [x] #9320
- [x] #9321
- [x] Clean up ElasticsearchExporter test refactoring
| 1.0 | [EPIC] Refactor Elasticsearch Exporter Tests - **Description**
The `elasticsearch-exporter` integration tests are currently quite convoluted. Historically, the `elasticsearch-exporter` tests used to depend on `zeebe-test` which was removed in https://github.com/camunda-cloud/zeebe/pull/8608. During the removal, some of the `zeebe-test` utilities ended up in the `elasticsearch-exporter` module where we also needed to add a lot of test scoped dependencies.
The main goals here were:
- Cover as much as possible via only unit tests. This gives us a faster feedback cycle when changing something - developers can focus on first fixing unit tests, which have smaller blast radius and faster execution, before even looking at integration tests.
- Narrow the scope of the integration tests as much as possible. This meant getting rid of the broker, workflow engine semantics, etc., and just testing the integration of the exporter instance with Elasticsearch.
In general, I wanted to keep the same coverage as before. I think the coverage is now higher, but that was not the original motivation.
**Breakdown**
- [x] #9162
- [x] #9315
- [x] #9316
- [x] #9317
- [x] #9318
- [x] #9319
- [x] #9320
- [x] #9321
- [x] Clean up ElasticsearchExporter test refactoring
| process | refactor elasticsearch exporter tests description the elasticsearch exporter integration tests are currently quite convoluted historically the elasticsearch exporter tests used to depend on zeebe test which was removed in during the removal some of the zeebe test utilities ended up in the elasticsearch exporter module where we also needed to add a lot of test scoped dependencies the main goals here were cover as much as possible via only unit tests this gives us a faster feedback cycle when changing something developers can focus on first fixing unit tests which have smaller blast radius and faster execution before even looking at integration tests narrow the scope of the integration tests as much as possible this meant getting rid of the broker workflow engine semantics etc and just testing the integration of the exporter instance with elasticsearch in general i wanted to keep the same coverage as before i think the coverage is now higher but that was not the original motivation breakdown clean up elasticsearchexporter test refactoring | 1 |
220,688 | 17,227,450,969 | IssuesEvent | 2021-07-20 05:18:15 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | closed | Dashicon improperly aligned in ToolbarButton without label | Needs Testing [Feature] UI Components | ## Description
Since 5.7-RC3 a dashicon in ToolbarButton without any label text has right margin. This wasn't the case before and is also different from how a svg icon works.
## Screenshots or screen recording (optional)


## Code snippet (optional)
```js
const CONTROLS = [
{
icon: 'editor-alignleft',
title: __('Left aligned'),
value: 'left',
},
{
icon: 'editor-aligncenter',
title: __('Center aligned'),
value: 'center',
},
{
icon: 'editor-alignright',
title: __('Right aligned'),
value: 'right',
},
];
...
<wp.blockEditor.BlockControls>
<wp.components.ToolbarGroup>
{CONTROLS.map(control => {
const {value} = control;
const isActive = (textAlign === value);
return (
<wp.components.ToolbarButton
{...control}
isActive={isActive}
onClick={() => setAttributes({textAlign: isActive ? undefined : value})}
/>
);
})}
</wp.components.ToolbarGroup>
</wp.blockEditor.BlockControls>
```
## WordPress information
- WordPress version: 5.7-RC3
- Gutenberg version: Not installed
- Are all plugins except Gutenberg deactivated? No
- Are you using a default theme (e.g. Twenty Twenty-One)? No
## Device information
- Device: Desktop
- Operating system: Windows 10
- Browser: Chrome 88
| 1.0 | Dashicon improperly aligned in ToolbarButton without label - ## Description
Since 5.7-RC3 a dashicon in ToolbarButton without any label text has right margin. This wasn't the case before and is also different from how a svg icon works.
## Screenshots or screen recording (optional)


## Code snippet (optional)
```js
const CONTROLS = [
{
icon: 'editor-alignleft',
title: __('Left aligned'),
value: 'left',
},
{
icon: 'editor-aligncenter',
title: __('Center aligned'),
value: 'center',
},
{
icon: 'editor-alignright',
title: __('Right aligned'),
value: 'right',
},
];
...
<wp.blockEditor.BlockControls>
<wp.components.ToolbarGroup>
{CONTROLS.map(control => {
const {value} = control;
const isActive = (textAlign === value);
return (
<wp.components.ToolbarButton
{...control}
isActive={isActive}
onClick={() => setAttributes({textAlign: isActive ? undefined : value})}
/>
);
})}
</wp.components.ToolbarGroup>
</wp.blockEditor.BlockControls>
```
## WordPress information
- WordPress version: 5.7-RC3
- Gutenberg version: Not installed
- Are all plugins except Gutenberg deactivated? No
- Are you using a default theme (e.g. Twenty Twenty-One)? No
## Device information
- Device: Desktop
- Operating system: Windows 10
- Browser: Chrome 88
| non_process | dashicon improperly aligned in toolbarbutton without label description since a dashicon in toolbarbutton without any label text has right margin this wasn t the case before and is also different from how a svg icon works screenshots or screen recording optional code snippet optional js const controls icon editor alignleft title left aligned value left icon editor aligncenter title center aligned value center icon editor alignright title right aligned value right controls map control const value control const isactive textalign value return wp components toolbarbutton control isactive isactive onclick setattributes textalign isactive undefined value wordpress information wordpress version gutenberg version not installed are all plugins except gutenberg deactivated no are you using a default theme e g twenty twenty one no device information device desktop operating system windows browser chrome | 0 |
6,598 | 9,670,782,698 | IssuesEvent | 2019-05-21 20:47:35 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Using the Marketplace Deployment Method described Forces Japan East as Automation Account | assigned-to-author automation/svc process-automation/subsvc product-issue triaged | If you try and deploy the Start/Stop from the Marketplace - after creating an OMS Workspace in Central US and then create an Automation Account - the Location is set to Japan East - and you cannot change it. And, when created it does create it in Japan East. This is broken. If you try and create an Automation Account ahead of time to select it during setup - it does not show up as available.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @georgewallace
* Microsoft Alias: **gwallace** | 1.0 | Using the Marketplace Deployment Method described Forces Japan East as Automation Account - If you try and deploy the Start/Stop from the Marketplace - after creating an OMS Workspace in Central US and then create an Automation Account - the Location is set to Japan East - and you cannot change it. And, when created it does create it in Japan East. This is broken. If you try and create an Automation Account ahead of time to select it during setup - it does not show up as available.
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 225c9d05-83dd-b006-0025-3753f5ab25bf
* Version Independent ID: 9eecef0c-b1cb-1136-faf7-542214492096
* Content: [Start/Stop VMs during off-hours solution](https://docs.microsoft.com/en-us/azure/automation/automation-solution-vm-management)
* Content Source: [articles/automation/automation-solution-vm-management.md](https://github.com/Microsoft/azure-docs/blob/master/articles/automation/automation-solution-vm-management.md)
* Service: **automation**
* Sub-service: **process-automation**
* GitHub Login: @georgewallace
* Microsoft Alias: **gwallace** | process | using the marketplace deployment method described forces japan east as automation account if you try and deploy the start stop from the marketplace after creating an oms workspace in central us and then create an automation account the location is set to japan east and you cannot change it and when created it does create it in japan east this is broken if you try and create an automation account ahead of time to select it during setup it does not show up as available document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login georgewallace microsoft alias gwallace | 1 |
307,218 | 26,519,183,508 | IssuesEvent | 2023-01-19 00:09:32 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | closed | Fix raw_ops.test_tensorflow_FloorDiv | TensorFlow Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3941620207/jobs/6744307327" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/3928568571/jobs/6716331188" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3928568571/jobs/6716331188" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3941620207/jobs/6744291469" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_FloorDiv[cpu-ivy.functional.backends.numpy-False-False]</summary>
2023-01-16T09:01:53.8846943Z E AssertionError: [[ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8847414Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8848944Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 -1.000000e+00 9.007199e+15]] != [[ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8849590Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8850211Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 -0.000000e+00 9.007199e+15]]
2023-01-16T09:01:53.8850661Z E Falsifying example: test_tensorflow_FloorDiv(
2023-01-16T09:01:53.8851208Z E dtype_and_x=(['float32', 'float32'],
2023-01-16T09:01:53.8851788Z E [array([[-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8852271Z E -1.000000e+00],
2023-01-16T09:01:53.8852802Z E [-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8853266Z E -1.000000e+00],
2023-01-16T09:01:53.8854131Z E [-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8854645Z E -9.007199e+15]], dtype=float32),
2023-01-16T09:01:53.8855191Z E array([[-1.00000e+00, -1.00000e+00, -1.00000e+00, -1.00000e+00,
2023-01-16T09:01:53.8855891Z E -1.00000e+00],
2023-01-16T09:01:53.8856434Z E [-1.00000e+00, -1.00000e+00, -1.00000e+00, -1.00000e+00,
2023-01-16T09:01:53.8857163Z E -1.00000e+00],
2023-01-16T09:01:53.8857687Z E [-1.00000e+00, -1.00000e+00, -1.00000e+00, 8.50706e+37,
2023-01-16T09:01:53.8858195Z E -1.00000e+00]], dtype=float32)]),
2023-01-16T09:01:53.8858603Z E num_positional_args=0,
2023-01-16T09:01:53.8858983Z E native_array=[False],
2023-01-16T09:01:53.8859358Z E as_variable=[False],
2023-01-16T09:01:53.8859948Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.FloorDiv',
2023-01-16T09:01:53.8860496Z E frontend='tensorflow',
2023-01-16T09:01:53.8861071Z E on_device='cpu',
2023-01-16T09:01:53.8861426Z E )
2023-01-16T09:01:53.8861722Z E
2023-01-16T09:01:53.8862573Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2JkYGJkYGEAAkYGCGAc9Gz5/0DwL3nQuAcLG8Rof8DAIICphgEAYpEHOQ==') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_FloorDiv[cpu-ivy.functional.backends.numpy-False-False]</summary>
2023-01-16T09:01:53.8846943Z E AssertionError: [[ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8847414Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8848944Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 -1.000000e+00 9.007199e+15]] != [[ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8849590Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8850211Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 -0.000000e+00 9.007199e+15]]
2023-01-16T09:01:53.8850661Z E Falsifying example: test_tensorflow_FloorDiv(
2023-01-16T09:01:53.8851208Z E dtype_and_x=(['float32', 'float32'],
2023-01-16T09:01:53.8851788Z E [array([[-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8852271Z E -1.000000e+00],
2023-01-16T09:01:53.8852802Z E [-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8853266Z E -1.000000e+00],
2023-01-16T09:01:53.8854131Z E [-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8854645Z E -9.007199e+15]], dtype=float32),
2023-01-16T09:01:53.8855191Z E array([[-1.00000e+00, -1.00000e+00, -1.00000e+00, -1.00000e+00,
2023-01-16T09:01:53.8855891Z E -1.00000e+00],
2023-01-16T09:01:53.8856434Z E [-1.00000e+00, -1.00000e+00, -1.00000e+00, -1.00000e+00,
2023-01-16T09:01:53.8857163Z E -1.00000e+00],
2023-01-16T09:01:53.8857687Z E [-1.00000e+00, -1.00000e+00, -1.00000e+00, 8.50706e+37,
2023-01-16T09:01:53.8858195Z E -1.00000e+00]], dtype=float32)]),
2023-01-16T09:01:53.8858603Z E num_positional_args=0,
2023-01-16T09:01:53.8858983Z E native_array=[False],
2023-01-16T09:01:53.8859358Z E as_variable=[False],
2023-01-16T09:01:53.8859948Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.FloorDiv',
2023-01-16T09:01:53.8860496Z E frontend='tensorflow',
2023-01-16T09:01:53.8861071Z E on_device='cpu',
2023-01-16T09:01:53.8861426Z E )
2023-01-16T09:01:53.8861722Z E
2023-01-16T09:01:53.8862573Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2JkYGJkYGEAAkYGCGAc9Gz5/0DwL3nQuAcLG8Rof8DAIICphgEAYpEHOQ==') as a decorator on your test case
</details>
| 1.0 | Fix raw_ops.test_tensorflow_FloorDiv - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/3941620207/jobs/6744307327" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/3928568571/jobs/6716331188" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/3928568571/jobs/6716331188" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/3941620207/jobs/6744291469" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_FloorDiv[cpu-ivy.functional.backends.numpy-False-False]</summary>
2023-01-16T09:01:53.8846943Z E AssertionError: [[ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8847414Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8848944Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 -1.000000e+00 9.007199e+15]] != [[ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8849590Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8850211Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 -0.000000e+00 9.007199e+15]]
2023-01-16T09:01:53.8850661Z E Falsifying example: test_tensorflow_FloorDiv(
2023-01-16T09:01:53.8851208Z E dtype_and_x=(['float32', 'float32'],
2023-01-16T09:01:53.8851788Z E [array([[-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8852271Z E -1.000000e+00],
2023-01-16T09:01:53.8852802Z E [-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8853266Z E -1.000000e+00],
2023-01-16T09:01:53.8854131Z E [-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8854645Z E -9.007199e+15]], dtype=float32),
2023-01-16T09:01:53.8855191Z E array([[-1.00000e+00, -1.00000e+00, -1.00000e+00, -1.00000e+00,
2023-01-16T09:01:53.8855891Z E -1.00000e+00],
2023-01-16T09:01:53.8856434Z E [-1.00000e+00, -1.00000e+00, -1.00000e+00, -1.00000e+00,
2023-01-16T09:01:53.8857163Z E -1.00000e+00],
2023-01-16T09:01:53.8857687Z E [-1.00000e+00, -1.00000e+00, -1.00000e+00, 8.50706e+37,
2023-01-16T09:01:53.8858195Z E -1.00000e+00]], dtype=float32)]),
2023-01-16T09:01:53.8858603Z E num_positional_args=0,
2023-01-16T09:01:53.8858983Z E native_array=[False],
2023-01-16T09:01:53.8859358Z E as_variable=[False],
2023-01-16T09:01:53.8859948Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.FloorDiv',
2023-01-16T09:01:53.8860496Z E frontend='tensorflow',
2023-01-16T09:01:53.8861071Z E on_device='cpu',
2023-01-16T09:01:53.8861426Z E )
2023-01-16T09:01:53.8861722Z E
2023-01-16T09:01:53.8862573Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2JkYGJkYGEAAkYGCGAc9Gz5/0DwL3nQuAcLG8Rof8DAIICphgEAYpEHOQ==') as a decorator on your test case
</details>
<details>
<summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_raw_ops.py::test_tensorflow_FloorDiv[cpu-ivy.functional.backends.numpy-False-False]</summary>
2023-01-16T09:01:53.8846943Z E AssertionError: [[ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8847414Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8848944Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 -1.000000e+00 9.007199e+15]] != [[ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8849590Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00 1.000000e+00]
2023-01-16T09:01:53.8850211Z E [ 1.000000e+00 1.000000e+00 1.000000e+00 -0.000000e+00 9.007199e+15]]
2023-01-16T09:01:53.8850661Z E Falsifying example: test_tensorflow_FloorDiv(
2023-01-16T09:01:53.8851208Z E dtype_and_x=(['float32', 'float32'],
2023-01-16T09:01:53.8851788Z E [array([[-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8852271Z E -1.000000e+00],
2023-01-16T09:01:53.8852802Z E [-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8853266Z E -1.000000e+00],
2023-01-16T09:01:53.8854131Z E [-1.000000e+00, -1.000000e+00, -1.000000e+00, -1.000000e+00,
2023-01-16T09:01:53.8854645Z E -9.007199e+15]], dtype=float32),
2023-01-16T09:01:53.8855191Z E array([[-1.00000e+00, -1.00000e+00, -1.00000e+00, -1.00000e+00,
2023-01-16T09:01:53.8855891Z E -1.00000e+00],
2023-01-16T09:01:53.8856434Z E [-1.00000e+00, -1.00000e+00, -1.00000e+00, -1.00000e+00,
2023-01-16T09:01:53.8857163Z E -1.00000e+00],
2023-01-16T09:01:53.8857687Z E [-1.00000e+00, -1.00000e+00, -1.00000e+00, 8.50706e+37,
2023-01-16T09:01:53.8858195Z E -1.00000e+00]], dtype=float32)]),
2023-01-16T09:01:53.8858603Z E num_positional_args=0,
2023-01-16T09:01:53.8858983Z E native_array=[False],
2023-01-16T09:01:53.8859358Z E as_variable=[False],
2023-01-16T09:01:53.8859948Z E fn_tree='ivy.functional.frontends.tensorflow.raw_ops.FloorDiv',
2023-01-16T09:01:53.8860496Z E frontend='tensorflow',
2023-01-16T09:01:53.8861071Z E on_device='cpu',
2023-01-16T09:01:53.8861426Z E )
2023-01-16T09:01:53.8861722Z E
2023-01-16T09:01:53.8862573Z E You can reproduce this example by temporarily adding @reproduce_failure('6.55.0', b'AXicY2JkYGJkYGEAAkYGCGAc9Gz5/0DwL3nQuAcLG8Rof8DAIICphgEAYpEHOQ==') as a decorator on your test case
</details>
| non_process | fix raw ops test tensorflow floordiv tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test tensorflow test raw ops py test tensorflow floordiv e assertionerror e e e e e falsifying example test tensorflow floordiv e dtype and x e array e e e e e dtype e array e e e e e dtype e num positional args e native array e as variable e fn tree ivy functional frontends tensorflow raw ops floordiv e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test raw ops py test tensorflow floordiv e assertionerror e e e e e falsifying example test tensorflow floordiv e dtype and x e array e e e e e dtype e array e e e e e dtype e num positional args e native array e as variable e fn tree ivy functional frontends tensorflow raw ops floordiv e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case | 0 |
20,611 | 27,279,543,779 | IssuesEvent | 2023-02-23 09:02:25 | googleapis/google-cloud-node | https://api.github.com/repos/googleapis/google-cloud-node | opened | Your .repo-metadata.json files have a problem 🤒 | type: process repo-metadata: lint | You have a problem with your .repo-metadata.json files:
Result of scan 📈:
* api_shortname '{{name}}' invalid in packages/gapic-node-templating/templates/bootstrap-templates/.repo-metadata.json
* api_shortname 'clouddms' invalid in packages/google-cloud-clouddms/.repo-metadata.json
* release_level must be equal to one of the allowed values in packages/google-cloud-contentwarehouse/.repo-metadata.json
* api_shortname 'filestore' invalid in packages/google-cloud-filestore/.repo-metadata.json
* api_shortname field missing from packages/google-cloud-run/.repo-metadata.json
* api_shortname field missing from packages/google-cloud-security-publicca/.repo-metadata.json
* api_shortname field missing from packages/google-cloud-video-stitcher/.repo-metadata.json
* api_shortname 'routing' invalid in packages/google-maps-routing/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions. | 1.0 | Your .repo-metadata.json files have a problem 🤒 - You have a problem with your .repo-metadata.json files:
Result of scan 📈:
* api_shortname '{{name}}' invalid in packages/gapic-node-templating/templates/bootstrap-templates/.repo-metadata.json
* api_shortname 'clouddms' invalid in packages/google-cloud-clouddms/.repo-metadata.json
* release_level must be equal to one of the allowed values in packages/google-cloud-contentwarehouse/.repo-metadata.json
* api_shortname 'filestore' invalid in packages/google-cloud-filestore/.repo-metadata.json
* api_shortname field missing from packages/google-cloud-run/.repo-metadata.json
* api_shortname field missing from packages/google-cloud-security-publicca/.repo-metadata.json
* api_shortname field missing from packages/google-cloud-video-stitcher/.repo-metadata.json
* api_shortname 'routing' invalid in packages/google-maps-routing/.repo-metadata.json
☝️ Once you address these problems, you can close this issue.
### Need help?
* [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field.
* [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**.
* Reach out to **go/github-automation** if you have any questions. | process | your repo metadata json files have a problem 🤒 you have a problem with your repo metadata json files result of scan 📈 api shortname name invalid in packages gapic node templating templates bootstrap templates repo metadata json api shortname clouddms invalid in packages google cloud clouddms repo metadata json release level must be equal to one of the allowed values in packages google cloud contentwarehouse repo metadata json api shortname filestore invalid in packages google cloud filestore repo metadata json api shortname field missing from packages google cloud run repo metadata json api shortname field missing from packages google cloud security publicca repo metadata json api shortname field missing from packages google cloud video stitcher repo metadata json api shortname routing invalid in packages google maps routing repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions | 1 |
22,036 | 4,769,515,982 | IssuesEvent | 2016-10-26 12:50:49 | SemanticMediaWiki/SemanticMediaWiki | https://api.github.com/repos/SemanticMediaWiki/SemanticMediaWiki | opened | Rework "INSTALL.md" | documentation | ### Issue
Since the upcoming 2.5 branch of SMW breaks compatibility with versions of MediaWiki 1.23.0 and lower the "INSTALL.md" should be reworked as it makes a lot of now unnecessary references to earlier versions and does not at all mention the worries introduced by MediaWiki 1.25+ ("composer.local.json"), etc.
| 1.0 | Rework "INSTALL.md" - ### Issue
Since the upcoming 2.5 branch of SMW breaks compatibility with versions of MediaWiki 1.23.0 and lower the "INSTALL.md" should be reworked as it makes a lot of now unnecessary references to earlier versions and does not at all mention the worries introduced by MediaWiki 1.25+ ("composer.local.json"), etc.
| non_process | rework install md issue since the upcoming branch of smw breaks compatibility with versions of mediawiki and lower the install md should be reworked as it makes a lot of now unnecessary references to earlier versions and does not at all mention the worries introduced by mediawiki composer local json etc | 0 |
9,635 | 12,599,661,592 | IssuesEvent | 2020-06-11 06:34:25 | allinurl/goaccess | https://api.github.com/repos/allinurl/goaccess | closed | How to seperate those statistics data by http domain(host)? | log-processing question | As the title mentioned, how can i classify the result by the http domain, for example, the top N requests data? | 1.0 | How to seperate those statistics data by http domain(host)? - As the title mentioned, how can i classify the result by the http domain, for example, the top N requests data? | process | how to seperate those statistics data by http domain host as the title mentioned how can i classify the result by the http domain for example the top n requests data | 1 |
16,593 | 21,642,161,485 | IssuesEvent | 2022-05-05 20:01:28 | googleapis/google-cloud-ruby | https://api.github.com/repos/googleapis/google-cloud-ruby | closed | Warning: a recent release failed | type: process | The following release PRs may have failed:
* #18200
* #18201
* #18202
* #18204
* #18205 | 1.0 | Warning: a recent release failed - The following release PRs may have failed:
* #18200
* #18201
* #18202
* #18204
* #18205 | process | warning a recent release failed the following release prs may have failed | 1 |
262,938 | 27,989,538,833 | IssuesEvent | 2023-03-27 01:40:18 | Killy85/game_ai_trainer | https://api.github.com/repos/Killy85/game_ai_trainer | opened | CVE-2023-25666 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl | Mend: dependency security vulnerability | ## CVE-2023-25666 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. Prior to versions 2.12.0 and 2.11.1, there is a floating point exception in AudioSpectrogram. A fix is included in TensorFlow version 2.12.0 and version 2.11.1.
<p>Publish Date: 2023-03-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25666>CVE-2023-25666</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-f637-vh3r-vfh2">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-f637-vh3r-vfh2</a></p>
<p>Release Date: 2023-03-24</p>
<p>Fix Resolution: tensorflow - 2.11.1,2.12.0, tensorflow-cpu - 2.11.1,2.12.0, tensorflow-gpu - 2.11.1,2.12.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-25666 (High) detected in tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2023-25666 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/d2/ea/ab2c8c0e81bd051cc1180b104c75a865ab0fc66c89be992c4b20bbf6d624/tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>
Dependency Hierarchy:
- :x: **tensorflow-1.13.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
TensorFlow is an open source platform for machine learning. Prior to versions 2.12.0 and 2.11.1, there is a floating point exception in AudioSpectrogram. A fix is included in TensorFlow version 2.12.0 and version 2.11.1.
<p>Publish Date: 2023-03-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25666>CVE-2023-25666</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-f637-vh3r-vfh2">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-f637-vh3r-vfh2</a></p>
<p>Release Date: 2023-03-24</p>
<p>Fix Resolution: tensorflow - 2.11.1,2.12.0, tensorflow-cpu - 2.11.1,2.12.0, tensorflow-gpu - 2.11.1,2.12.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in tensorflow whl cve high severity vulnerability vulnerable library tensorflow whl tensorflow is an open source machine learning framework for everyone library home page a href dependency hierarchy x tensorflow whl vulnerable library vulnerability details tensorflow is an open source platform for machine learning prior to versions and there is a floating point exception in audiospectrogram a fix is included in tensorflow version and version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with mend | 0 |
22,450 | 31,169,299,065 | IssuesEvent | 2023-08-16 22:56:41 | googleapis/gapic-generator-python | https://api.github.com/repos/googleapis/gapic-generator-python | closed | Add service YAML files to the golden integration tests | type: process | All of the golden files are missing the service YAML config files except [eventarc](https://github.com/googleapis/gapic-generator-python/blob/main/tests/integration/eventarc_v1.yaml). We should add/update these files to improve our integration testing. | 1.0 | Add service YAML files to the golden integration tests - All of the golden files are missing the service YAML config files except [eventarc](https://github.com/googleapis/gapic-generator-python/blob/main/tests/integration/eventarc_v1.yaml). We should add/update these files to improve our integration testing. | process | add service yaml files to the golden integration tests all of the golden files are missing the service yaml config files except we should add update these files to improve our integration testing | 1 |
67,880 | 14,891,996,988 | IssuesEvent | 2021-01-21 01:46:48 | phytomichael/KSA | https://api.github.com/repos/phytomichael/KSA | opened | CVE-2020-36180 (Medium) detected in jackson-databind-2.8.1.jar | security vulnerability | ## CVE-2020-36180 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /KSA/ksa/ksa/ksa-service-root/ksa-finance-service/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,KSA/ksa/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.cpdsadapter.DriverAdapterCPDS.
<p>Publish Date: 2021-01-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36180>CVE-2020-36180</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.1","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.1","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-36180","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.cpdsadapter.DriverAdapterCPDS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36180","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> --> | True | CVE-2020-36180 (Medium) detected in jackson-databind-2.8.1.jar - ## CVE-2020-36180 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.1.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /KSA/ksa/ksa/ksa-service-root/ksa-finance-service/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.1/jackson-databind-2.8.1.jar,KSA/ksa/ksa/ksa-web-root/ksa-web/target/ROOT/WEB-INF/lib/jackson-databind-2.8.1.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-1.4.0.RELEASE.jar (Root Library)
- :x: **jackson-databind-2.8.1.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.cpdsadapter.DriverAdapterCPDS.
<p>Publish Date: 2021-01-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36180>CVE-2020-36180</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>6.8</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.1","isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:1.4.0.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.8.1","isMinimumFixVersionAvailable":false}],"vulnerabilityIdentifier":"CVE-2020-36180","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.dbcp2.cpdsadapter.DriverAdapterCPDS.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36180","cvss2Severity":"medium","cvss2Score":"6.8","extraData":{}}</REMEDIATE> --> | non_process | cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file ksa ksa ksa ksa service root ksa finance service pom xml path to vulnerable library root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar ksa ksa ksa ksa web root ksa web target root web inf lib jackson databind jar dependency hierarchy spring boot starter web release jar root library x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons cpdsadapter driveradaptercpds publish date url a href cvss score details base score metrics not available isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons cpdsadapter driveradaptercpds vulnerabilityurl | 0 |
26,259 | 2,684,274,314 | IssuesEvent | 2015-03-28 20:35:34 | ConEmu/old-issues | https://api.github.com/repos/ConEmu/old-issues | closed | Crash whilst resizing Conemu window | 2–5 stars bug duplicate imported Priority-Medium | _From [col.brad...@gmail.com](https://code.google.com/u/103901031233724257395/) on January 24, 2013 00:00:36_
Required information! OS version: Win2k/WinXP/Vista/Win7/Win8 SP? x86/x64 ConEmu version: ? Far version (if you are using Far Manager): ? *Bug description* Conemu crashes whilst resizing window. *Steps to reproduction* 1. Open single instance of Conemu.
2. Set font to Consolas, 14pt, Clear-type.
3. Resize console window to maximum size on 1920x1080 monitor... CRASHY, CRASHY!
4. Workaround is to use Setting->Main->Size&Pos->Window size (cells).
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=904_ | 1.0 | Crash whilst resizing Conemu window - _From [col.brad...@gmail.com](https://code.google.com/u/103901031233724257395/) on January 24, 2013 00:00:36_
Required information! OS version: Win2k/WinXP/Vista/Win7/Win8 SP? x86/x64 ConEmu version: ? Far version (if you are using Far Manager): ? *Bug description* Conemu crashes whilst resizing window. *Steps to reproduction* 1. Open single instance of Conemu.
2. Set font to Consolas, 14pt, Clear-type.
3. Resize console window to maximum size on 1920x1080 monitor... CRASHY, CRASHY!
4. Workaround is to use Setting->Main->Size&Pos->Window size (cells).
_Original issue: http://code.google.com/p/conemu-maximus5/issues/detail?id=904_ | non_process | crash whilst resizing conemu window from on january required information os version winxp vista sp conemu version far version if you are using far manager bug description conemu crashes whilst resizing window steps to reproduction open single instance of conemu set font to consolas clear type resize console window to maximum size on monitor crashy crashy workaround is to use setting main size pos window size cells original issue | 0 |
6,783 | 9,917,114,974 | IssuesEvent | 2019-06-28 22:29:46 | HumanCellAtlas/dcp-community | https://api.github.com/repos/HumanCellAtlas/dcp-community | closed | Improving community notifications for charters | charter-process | - [ ] Use #dcp-announce[ments] instead of #dcp channel for charter _state changes_ | 1.0 | Improving community notifications for charters - - [ ] Use #dcp-announce[ments] instead of #dcp channel for charter _state changes_ | process | improving community notifications for charters use dcp announce instead of dcp channel for charter state changes | 1 |
14,027 | 16,826,829,384 | IssuesEvent | 2021-06-17 19:49:18 | hashgraph/hedera-mirror-node | https://api.github.com/repos/hashgraph/hedera-mirror-node | opened | Rename master branch to main | P1 enhancement process | **Problem**
In honor of [Juneteenth](https://en.wikipedia.org/wiki/Juneteenth), we want to rename our master branch to main as much of the industry already has.
**Solution**
- Change any references to master in the codebase to `main`
- Rename master branch to main in Github settings
- Adjust local git settings
```shell
git branch -m master main
git fetch origin
git branch -u origin/main main
```
**Alternatives**
**Additional Context**
| 1.0 | Rename master branch to main - **Problem**
In honor of [Juneteenth](https://en.wikipedia.org/wiki/Juneteenth), we want to rename our master branch to main as much of the industry already has.
**Solution**
- Change any references to master in the codebase to `main`
- Rename master branch to main in Github settings
- Adjust local git settings
```shell
git branch -m master main
git fetch origin
git branch -u origin/main main
```
**Alternatives**
**Additional Context**
| process | rename master branch to main problem in honor of we want to rename our master branch to main as much of the industry already has solution change any references to master in the codebase to main rename master branch to main in github settings adjust local git settings shell git branch m master main git fetch origin git branch u origin main main alternatives additional context | 1 |
77,556 | 27,049,323,444 | IssuesEvent | 2023-02-13 12:07:52 | notofonts/telugu | https://api.github.com/repos/notofonts/telugu | closed | The base line for noto sans Telugu is not in line with English and punctuation marks. | Type-Defect Script-Telugu Android FoundIn-1.x in-evaluation | ```
What steps will reproduce the problem?
1.Typing in Telugu make the baseline higher compared to English or hindi
2.The same happens for kannada too
3.
What is the expected output? What do you see instead?
The base line of Telugu and kannada are supposed to be in line with English and
hindi in order to use punctuations, but it appears to be higher.
What version of the product are you using? On what operating system?
KitKat 4.4.2 on nexus 4
Please provide any additional information below.
I happen to notice that there are regular/bold and regularUI/boldUI fonts and
it's supposed to use the UI version for contacts and Lockscreen etc and the
regular type for messaging and text display a cross the device, but what I
found was it is using the UI variant everywhere. When I replaced the UI version
with the normal version the Telugu text was being displayed on the same
baseline as English in messages etc but the bottom part is being chopped off in
Lockscreen or contacts etc. Please resolve this issue soon.
```
Original issue reported on code.google.com by `ashok.0...@gmail.com` on 29 Dec 2013 at 10:51
| 1.0 | The base line for noto sans Telugu is not in line with English and punctuation marks. - ```
What steps will reproduce the problem?
1.Typing in Telugu make the baseline higher compared to English or hindi
2.The same happens for kannada too
3.
What is the expected output? What do you see instead?
The base line of Telugu and kannada are supposed to be in line with English and
hindi in order to use punctuations, but it appears to be higher.
What version of the product are you using? On what operating system?
KitKat 4.4.2 on nexus 4
Please provide any additional information below.
I happen to notice that there are regular/bold and regularUI/boldUI fonts and
it's supposed to use the UI version for contacts and Lockscreen etc and the
regular type for messaging and text display a cross the device, but what I
found was it is using the UI variant everywhere. When I replaced the UI version
with the normal version the Telugu text was being displayed on the same
baseline as English in messages etc but the bottom part is being chopped off in
Lockscreen or contacts etc. Please resolve this issue soon.
```
Original issue reported on code.google.com by `ashok.0...@gmail.com` on 29 Dec 2013 at 10:51
| non_process | the base line for noto sans telugu is not in line with english and punctuation marks what steps will reproduce the problem typing in telugu make the baseline higher compared to english or hindi the same happens for kannada too what is the expected output what do you see instead the base line of telugu and kannada are supposed to be in line with english and hindi in order to use punctuations but it appears to be higher what version of the product are you using on what operating system kitkat on nexus please provide any additional information below i happen to notice that there are regular bold and regularui boldui fonts and it s supposed to use the ui version for contacts and lockscreen etc and the regular type for messaging and text display a cross the device but what i found was it is using the ui variant everywhere when i replaced the ui version with the normal version the telugu text was being displayed on the same baseline as english in messages etc but the bottom part is being chopped off in lockscreen or contacts etc please resolve this issue soon original issue reported on code google com by ashok gmail com on dec at | 0 |
6,072 | 8,909,442,947 | IssuesEvent | 2019-01-18 06:18:19 | mick-warehime/sixth_corp | https://api.github.com/repos/mick-warehime/sixth_corp | closed | remove union in factory constructors | development process | if isinstance(character, CharacterTypes):
data = character.data
else:
data = character | 1.0 | remove union in factory constructors - if isinstance(character, CharacterTypes):
data = character.data
else:
data = character | process | remove union in factory constructors if isinstance character charactertypes data character data else data character | 1 |
697,787 | 23,952,758,278 | IssuesEvent | 2022-09-12 12:53:06 | epicmaxco/vuestic-ui | https://api.github.com/repos/epicmaxco/vuestic-ui | opened | On dark theme sponsors are not very visible. | BUG good first issue LOW PRIORITY difficulty: I'm too young to die | 
Probably just needs an opaque white background for existing images, but you can investigate a bit.
Might be possible to do in context of markdown, github allowed some html before. | 1.0 | On dark theme sponsors are not very visible. - 
Probably just needs an opaque white background for existing images, but you can investigate a bit.
Might be possible to do in context of markdown, github allowed some html before. | non_process | on dark theme sponsors are not very visible probably just needs an opaque white background for existing images but you can investigate a bit might be possible to do in context of markdown github allowed some html before | 0 |
362,436 | 10,727,523,395 | IssuesEvent | 2019-10-28 11:55:16 | ISISScientificComputing/autoreduce | https://api.github.com/repos/ISISScientificComputing/autoreduce | closed | Refactor Hot Fixes from Python 3 Compatibility Upgrade | 🔑 Internal 🕐 Low priority | Issue raised by: [developer]
### What?
Many changes were made to the Autoreduction codebase to fix any incompatibilities when upgrading to Python 3 before end of life.
We should go back to create more permanent fixes for these changes where applicable to maintain a clean and readable codebase.
### Where?
1. All of Autoreduction codebase
2. QueueProcessors/AutoreductionProcessor/daemon.py
### How?
1. Change all usages of %s for string formatting to {} notation.
2. Put inside `with` and remove `pid_file.close()` on lines 113-116 inside QueueProcessors/AutoreductionProcessor/daemon.py if possible.
### How to test the issue is resolved
1. Check returned strings are formatted correctly
2. Check `pid_file` is read and closed correctly. | 1.0 | Refactor Hot Fixes from Python 3 Compatibility Upgrade - Issue raised by: [developer]
### What?
Many changes were made to the Autoreduction codebase to fix any incompatibilities when upgrading to Python 3 before end of life.
We should go back to create more permanent fixes for these changes where applicable to maintain a clean and readable codebase.
### Where?
1. All of Autoreduction codebase
2. QueueProcessors/AutoreductionProcessor/daemon.py
### How?
1. Change all usages of %s for string formatting to {} notation.
2. Put inside `with` and remove `pid_file.close()` on lines 113-116 inside QueueProcessors/AutoreductionProcessor/daemon.py if possible.
### How to test the issue is resolved
1. Check returned strings are formatted correctly
2. Check `pid_file` is read and closed correctly. | non_process | refactor hot fixes from python compatibility upgrade issue raised by what many changes were made to the autoreduction codebase to fix any incompatibilities when upgrading to python before end of life we should go back to create more permanent fixes for these changes where applicable to maintain a clean and readable codebase where all of autoreduction codebase queueprocessors autoreductionprocessor daemon py how change all usages of s for string formatting to notation put inside with and remove pid file close on lines inside queueprocessors autoreductionprocessor daemon py if possible how to test the issue is resolved check returned strings are formatted correctly check pid file is read and closed correctly | 0 |
16,502 | 2,906,427,067 | IssuesEvent | 2015-06-19 09:58:10 | cakephp/cakephp | https://api.github.com/repos/cakephp/cakephp | closed | 3.x PaginatorComponent defect | Defect |
If this is sortWhitelist: `sortWhitelist => []`,
and this is the url: `/controller/action?sort=field`.
The PaginatorComponent will sort by that field, ignoring that we don't want to allow sorting any field.
Instead of empty sortWhitelist, if the sortWhitelist isn't empty like this: `sortWhitelist => ['even_invalid_field']`,
then the sortWhitelist gets used by the PaginatorComponent.
```
public function validateSort(Table $object, array $options)
{
...
if (!empty($options['sortWhitelist'])) {
$field = key($options['order']);
$inWhitelist = in_array($field, $options['sortWhitelist'], true);
if (!$inWhitelist) {
$options['order'] = [];
return $options;
}
$validate = false;
}
...
}
```
It doesn't care about empty sortWhitelist.
| 1.0 | 3.x PaginatorComponent defect -
If this is sortWhitelist: `sortWhitelist => []`,
and this is the url: `/controller/action?sort=field`.
The PaginatorComponent will sort by that field, ignoring that we don't want to allow sorting any field.
Instead of empty sortWhitelist, if the sortWhitelist isn't empty like this: `sortWhitelist => ['even_invalid_field']`,
then the sortWhitelist gets used by the PaginatorComponent.
```
public function validateSort(Table $object, array $options)
{
...
if (!empty($options['sortWhitelist'])) {
$field = key($options['order']);
$inWhitelist = in_array($field, $options['sortWhitelist'], true);
if (!$inWhitelist) {
$options['order'] = [];
return $options;
}
$validate = false;
}
...
}
```
It doesn't care about empty sortWhitelist.
| non_process | x paginatorcomponent defect if this is sortwhitelist sortwhitelist and this is the url controller action sort field the paginatorcomponent will sort by that field ignoring that we don t want to allow sorting any field instead of empty sortwhitelist if the sortwhitelist isn t empty like this sortwhitelist then the sortwhitelist gets used by the paginatorcomponent public function validatesort table object array options if empty options field key options inwhitelist in array field options true if inwhitelist options return options validate false it doesn t care about empty sortwhitelist | 0 |
20,263 | 26,883,300,588 | IssuesEvent | 2023-02-05 22:07:57 | km4ack/pi-build | https://api.github.com/repos/km4ack/pi-build | closed | QSSTV Site Down | bug Slated in process | The [site](http://users.telenet.be/on4qz/qsstv/downloads/) where BAP downloads QSSTV is down. Thanks Bruce, W1BAW, for catching the issue. | 1.0 | QSSTV Site Down - The [site](http://users.telenet.be/on4qz/qsstv/downloads/) where BAP downloads QSSTV is down. Thanks Bruce, W1BAW, for catching the issue. | process | qsstv site down the where bap downloads qsstv is down thanks bruce for catching the issue | 1 |
8,458 | 11,631,219,684 | IssuesEvent | 2020-02-28 00:39:44 | geneontology/go-ontology | https://api.github.com/repos/geneontology/go-ontology | closed | regulation of mitotic DNA replication initiation from early origin | New term request PomBase cell cycle and DNA processes | by analogy with GO:0101017; for S. pombe rik1 and some phenotype LDs
id: GO:new
name: regulation of mitotic DNA replication initiation from early origin
namespace: biological_process
def: "Any process that modulates the frequency, rate or extent of firing from an early origin of replication involved in mitotic DNA replication." [PMID:22279046]
synonym: "regulation of early replication origin firing" EXACT []
is_a: GO:1903466 ! regulation of mitotic DNA replication initiation
early_origin_of_replication is SO:0002140 | 1.0 | regulation of mitotic DNA replication initiation from early origin - by analogy with GO:0101017; for S. pombe rik1 and some phenotype LDs
id: GO:new
name: regulation of mitotic DNA replication initiation from early origin
namespace: biological_process
def: "Any process that modulates the frequency, rate or extent of firing from an early origin of replication involved in mitotic DNA replication." [PMID:22279046]
synonym: "regulation of early replication origin firing" EXACT []
is_a: GO:1903466 ! regulation of mitotic DNA replication initiation
early_origin_of_replication is SO:0002140 | process | regulation of mitotic dna replication initiation from early origin by analogy with go for s pombe and some phenotype lds id go new name regulation of mitotic dna replication initiation from early origin namespace biological process def any process that modulates the frequency rate or extent of firing from an early origin of replication involved in mitotic dna replication synonym regulation of early replication origin firing exact is a go regulation of mitotic dna replication initiation early origin of replication is so | 1 |
19,395 | 25,539,287,645 | IssuesEvent | 2022-11-29 14:16:29 | ESMValGroup/ESMValCore | https://api.github.com/repos/ESMValGroup/ESMValCore | closed | Preprocessor `multimodel_statistics` fails when data have no time dimension | bug preprocessor | **Describe the bug**
Hi all, I'm developing some tests for the multimodel statistics using real data (#856), and I'm coming accross this bug:
Running multimodel statistics with a list of cubes with **no time dimension**, will result in
`ValueError: Cannot guess bounds for a coordinate of length 1.`
The bug can be reproduced by:
```python
cubes = [cube[0] for cube in cubes]
multi_model_statistics(cubes, span='full', statistics=['mean'])
```
<details>
<summary>See the full stack trace below.</summary>
```
timeseries_cubes_month = [<iris 'Cube' of air_temperature / (K) (time: 14; air_pressure: 2; latitude: 3; longitude: 2)>, <iris 'Cube' of air_te... 3; longitude: 2)>, <iris 'Cube' of air_temperature / (K) (time: 14; air_pressure: 2; latitude: 3; longitude: 2)>, ...]
@pytest.mark.functional
# @pytest.mark.xfail('ValueError')
def test_multimodel_no_time_dimension(timeseries_cubes_month):
"""Test statistic without time dimension using monthly data."""
span = 'full'
cubes = timeseries_cubes_month
cubes = [cube[0] for cube in cubes]
# ValueError: Cannot guess bounds for a coordinate of length 1.
> multimodel_test(cubes, span=span, statistic='mean')
tests/functional/test_multimodel.py:227:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
tests/functional/test_multimodel.py:125: in multimodel_test
output = multi_model_statistics(cubes, span=span, statistics=statistics)
esmvalcore/preprocessor/_multimodel.py:439: in multi_model_statistics
statistic_cube = _assemble_full_data(cubes, statistic)
esmvalcore/preprocessor/_multimodel.py:323: in _assemble_full_data
time_axis = [float(fl) for fl in _monthly_t(cubes)]
esmvalcore/preprocessor/_multimodel.py:277: in _monthly_t
days = {day for cube in cubes for day in _datetime_to_int_days(cube)}
esmvalcore/preprocessor/_multimodel.py:277: in <setcomp>
days = {day for cube in cubes for day in _datetime_to_int_days(cube)}
esmvalcore/preprocessor/_multimodel.py:206: in _datetime_to_int_days
cube = _align_yearly_axes(cube)
esmvalcore/preprocessor/_multimodel.py:231: in _align_yearly_axes
return regrid_time(cube, 'yr')
esmvalcore/preprocessor/_time.py:602: in regrid_time
cube.coord('time').guess_bounds()
../../miniconda3/envs/esmvaltool/lib/python3.8/site-packages/iris/coords.py:1564: in guess_bounds
self.bounds = self._guess_bounds(bound_position)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = DimCoord(array([49456.]), standard_name='time', units=Unit('days since 1850-1-1 00:00:00', calendar='365_day'), long_name='time', var_name='time', attributes={'_ChunkSizes': 1})
bound_position = 0.5
def _guess_bounds(self, bound_position=0.5):
"""
Return bounds for this coordinate based on its points.
Kwargs:
* bound_position:
The desired position of the bounds relative to the position
of the points.
Returns:
A numpy array of shape (len(self.points), 2).
.. note::
This method only works for coordinates with ``coord.ndim == 1``.
.. note::
If `iris.FUTURE.clip_latitudes` is True, then this method
will clip the coordinate bounds to the range [-90, 90] when:
- it is a `latitude` or `grid_latitude` coordinate,
- the units are degrees,
- all the points are in the range [-90, 90].
.. deprecated:: 2.0.0
The `iris.FUTURE.clip_latitudes` option is now deprecated
and is set to True by default. Please remove code which
relies on coordinate bounds being outside the range
[-90, 90].
"""
# XXX Consider moving into DimCoord
# ensure we have monotonic points
if not self.is_monotonic():
raise ValueError("Need monotonic points to generate bounds for %s"
% self.name())
if self.ndim != 1:
raise iris.exceptions.CoordinateMultiDimError(self)
if self.shape[0] < 2:
> raise ValueError('Cannot guess bounds for a coordinate of length '
'1.')
E ValueError: Cannot guess bounds for a coordinate of length 1.
```
</details>
| 1.0 | Preprocessor `multimodel_statistics` fails when data have no time dimension - **Describe the bug**
Hi all, I'm developing some tests for the multimodel statistics using real data (#856), and I'm coming accross this bug:
Running multimodel statistics with a list of cubes with **no time dimension**, will result in
`ValueError: Cannot guess bounds for a coordinate of length 1.`
The bug can be reproduced by:
```python
cubes = [cube[0] for cube in cubes]
multi_model_statistics(cubes, span='full', statistics=['mean'])
```
<details>
<summary>See the full stack trace below.</summary>
```
timeseries_cubes_month = [<iris 'Cube' of air_temperature / (K) (time: 14; air_pressure: 2; latitude: 3; longitude: 2)>, <iris 'Cube' of air_te... 3; longitude: 2)>, <iris 'Cube' of air_temperature / (K) (time: 14; air_pressure: 2; latitude: 3; longitude: 2)>, ...]
@pytest.mark.functional
# @pytest.mark.xfail('ValueError')
def test_multimodel_no_time_dimension(timeseries_cubes_month):
"""Test statistic without time dimension using monthly data."""
span = 'full'
cubes = timeseries_cubes_month
cubes = [cube[0] for cube in cubes]
# ValueError: Cannot guess bounds for a coordinate of length 1.
> multimodel_test(cubes, span=span, statistic='mean')
tests/functional/test_multimodel.py:227:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
tests/functional/test_multimodel.py:125: in multimodel_test
output = multi_model_statistics(cubes, span=span, statistics=statistics)
esmvalcore/preprocessor/_multimodel.py:439: in multi_model_statistics
statistic_cube = _assemble_full_data(cubes, statistic)
esmvalcore/preprocessor/_multimodel.py:323: in _assemble_full_data
time_axis = [float(fl) for fl in _monthly_t(cubes)]
esmvalcore/preprocessor/_multimodel.py:277: in _monthly_t
days = {day for cube in cubes for day in _datetime_to_int_days(cube)}
esmvalcore/preprocessor/_multimodel.py:277: in <setcomp>
days = {day for cube in cubes for day in _datetime_to_int_days(cube)}
esmvalcore/preprocessor/_multimodel.py:206: in _datetime_to_int_days
cube = _align_yearly_axes(cube)
esmvalcore/preprocessor/_multimodel.py:231: in _align_yearly_axes
return regrid_time(cube, 'yr')
esmvalcore/preprocessor/_time.py:602: in regrid_time
cube.coord('time').guess_bounds()
../../miniconda3/envs/esmvaltool/lib/python3.8/site-packages/iris/coords.py:1564: in guess_bounds
self.bounds = self._guess_bounds(bound_position)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = DimCoord(array([49456.]), standard_name='time', units=Unit('days since 1850-1-1 00:00:00', calendar='365_day'), long_name='time', var_name='time', attributes={'_ChunkSizes': 1})
bound_position = 0.5
def _guess_bounds(self, bound_position=0.5):
"""
Return bounds for this coordinate based on its points.
Kwargs:
* bound_position:
The desired position of the bounds relative to the position
of the points.
Returns:
A numpy array of shape (len(self.points), 2).
.. note::
This method only works for coordinates with ``coord.ndim == 1``.
.. note::
If `iris.FUTURE.clip_latitudes` is True, then this method
will clip the coordinate bounds to the range [-90, 90] when:
- it is a `latitude` or `grid_latitude` coordinate,
- the units are degrees,
- all the points are in the range [-90, 90].
.. deprecated:: 2.0.0
The `iris.FUTURE.clip_latitudes` option is now deprecated
and is set to True by default. Please remove code which
relies on coordinate bounds being outside the range
[-90, 90].
"""
# XXX Consider moving into DimCoord
# ensure we have monotonic points
if not self.is_monotonic():
raise ValueError("Need monotonic points to generate bounds for %s"
% self.name())
if self.ndim != 1:
raise iris.exceptions.CoordinateMultiDimError(self)
if self.shape[0] < 2:
> raise ValueError('Cannot guess bounds for a coordinate of length '
'1.')
E ValueError: Cannot guess bounds for a coordinate of length 1.
```
</details>
| process | preprocessor multimodel statistics fails when data have no time dimension describe the bug hi all i m developing some tests for the multimodel statistics using real data and i m coming accross this bug running multimodel statistics with a list of cubes with no time dimension will result in valueerror cannot guess bounds for a coordinate of length the bug can be reproduced by python cubes for cube in cubes multi model statistics cubes span full statistics see the full stack trace below timeseries cubes month pytest mark functional pytest mark xfail valueerror def test multimodel no time dimension timeseries cubes month test statistic without time dimension using monthly data span full cubes timeseries cubes month cubes for cube in cubes valueerror cannot guess bounds for a coordinate of length multimodel test cubes span span statistic mean tests functional test multimodel py tests functional test multimodel py in multimodel test output multi model statistics cubes span span statistics statistics esmvalcore preprocessor multimodel py in multi model statistics statistic cube assemble full data cubes statistic esmvalcore preprocessor multimodel py in assemble full data time axis esmvalcore preprocessor multimodel py in monthly t days day for cube in cubes for day in datetime to int days cube esmvalcore preprocessor multimodel py in days day for cube in cubes for day in datetime to int days cube esmvalcore preprocessor multimodel py in datetime to int days cube align yearly axes cube esmvalcore preprocessor multimodel py in align yearly axes return regrid time cube yr esmvalcore preprocessor time py in regrid time cube coord time guess bounds envs esmvaltool lib site packages iris coords py in guess bounds self bounds self guess bounds bound position self dimcoord array standard name time units unit days since calendar day long name time var name time attributes chunksizes bound position def guess bounds self bound position return bounds for this coordinate based on its points kwargs bound position the desired position of the bounds relative to the position of the points returns a numpy array of shape len self points note this method only works for coordinates with coord ndim note if iris future clip latitudes is true then this method will clip the coordinate bounds to the range when it is a latitude or grid latitude coordinate the units are degrees all the points are in the range deprecated the iris future clip latitudes option is now deprecated and is set to true by default please remove code which relies on coordinate bounds being outside the range xxx consider moving into dimcoord ensure we have monotonic points if not self is monotonic raise valueerror need monotonic points to generate bounds for s self name if self ndim raise iris exceptions coordinatemultidimerror self if self shape raise valueerror cannot guess bounds for a coordinate of length e valueerror cannot guess bounds for a coordinate of length | 1 |
9,717 | 12,716,589,366 | IssuesEvent | 2020-06-24 02:24:24 | OUDcollective/twenty20times | https://api.github.com/repos/OUDcollective/twenty20times | opened | Understanding the GitHub flow · GitHub Guides | workflow-process | 
## WHAT IS A PULL REQUEST
***GitHub Guides & Best Practices**
---
**Source URL**:
[https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/)
<table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
| 1.0 | Understanding the GitHub flow · GitHub Guides - 
## WHAT IS A PULL REQUEST
***GitHub Guides & Best Practices**
---
**Source URL**:
[https://guides.github.com/introduction/flow/](https://guides.github.com/introduction/flow/)
<table><tr><td><strong>Browser</strong></td><td>Chrome 84.0.4147.56</td></tr><tr><td><strong>OS</strong></td><td>Windows 10 64-bit</td></tr><tr><td><strong>Screen Size</strong></td><td>2560x1080</td></tr><tr><td><strong>Viewport Size</strong></td><td>2560x888</td></tr><tr><td><strong>Pixel Ratio</strong></td><td>@1x</td></tr><tr><td><strong>Zoom Level</strong></td><td>100%</td></tr></table>
| process | understanding the github flow · github guides what is a pull request github guides best practices source url browser chrome os windows bit screen size viewport size pixel ratio zoom level | 1 |
22,049 | 7,112,351,934 | IssuesEvent | 2018-01-17 16:46:19 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | Build is buggy for java 9. | :Core build | Try compiling with java 9:
First you will hit this. This can be fixed by using `assertEquals`, which also simplifies the test.
```
/home/rmuir/workspace/elasticsearch/core/src/test/java/org/elasticsearch/threadpool/ScalingThreadPoolTests.java:180: error: method invoked with incorrect number of arguments; expected 2, found 0
assertThat(stats(threadPool, threadPoolName).getThreads(), equalTo(128));
^
/home/rmuir/workspace/elasticsearch/core/src/test/java/org/elasticsearch/threadpool/ScalingThreadPoolTests.java:196: error: method invoked with incorrect number of arguments; expected 2, found 0
assertThat(stats(threadPool, threadPoolName).getCompleted(), equalTo(128L));
^
```
Next problem is that I am unsure `-source/-target` settings are being passed at all. Because I see class files referencing java 9 StringConcatFactory.
Next problem is that forbidden-apis screams, presumably because its being run in the wrong JVM (not the JAVA_HOME that is set, but the java 8 one running gradle itself). It sees StringConcatFactory and fails the build:
```
Caused by: org.gradle.internal.UncheckedException: de.thetaphi.forbiddenapis.ForbiddenApiException: Check for forbidden API calls failed: java.lang.ClassNotFoundException: java.lang.invoke.StringConcatFactory
at org.gradle.internal.UncheckedException.throwAsUncheckedException(UncheckedException.java:45)
at org.gradle.internal.reflect.JavaMethod.invoke(JavaMethod.java:78)
at org.gradle.api.internal.project.taskfactory.AnnotationProcessingTaskFactory$StandardTaskAction.doExecute(AnnotationProcessingTaskFactory.java:227)
at org.gradle.api.internal.project.taskfactory.AnnotationProcessingTaskFactory$StandardTaskAction.execute(AnnotationProcessingTaskFactory.java:220)
at org.gradle.api.internal.project.taskfactory.AnnotationProcessingTaskFactory$StandardTaskAction.execute(AnnotationProcessingTaskFactory.java:209)
at org.gradle.api.internal.AbstractTask$TaskActionWrapper.execute(AbstractTask.java:585)
at org.gradle.api.internal.AbstractTask$TaskActionWrapper.execute(AbstractTask.java:568)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeAction(ExecuteActionsTaskExecuter.java:80)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeActions(ExecuteActionsTaskExecuter.java:61)
... 68 more
Caused by: de.thetaphi.forbiddenapis.ForbiddenApiException: Check for forbidden API calls failed: java.lang.ClassNotFoundException: java.lang.invoke.StringConcatFactory
at de.thetaphi.forbiddenapis.Checker.run(Checker.java:550)
at de.thetaphi.forbiddenapis.gradle.CheckForbiddenApis.checkForbidden(CheckForbiddenApis.java:551)
at org.gradle.internal.reflect.JavaMethod.invoke(JavaMethod.java:75)
... 75 more
Caused by: java.lang.ClassNotFoundException: java.lang.invoke.StringConcatFactory
at de.thetaphi.forbiddenapis.Checker.getClassFromClassLoader(Checker.java:264)
at de.thetaphi.forbiddenapis.Checker.lookupRelatedClass(Checker.java:277)
at de.thetaphi.forbiddenapis.ClassScanner$2.checkMethodAccess(ClassScanner.java:349)
at de.thetaphi.forbiddenapis.ClassScanner$2.checkHandle(ClassScanner.java:411)
at de.thetaphi.forbiddenapis.ClassScanner$2.visitInvokeDynamicInsn(ClassScanner.java:497)
at de.thetaphi.forbiddenapis.asm.ClassReader.a(Unknown Source)
at de.thetaphi.forbiddenapis.asm.ClassReader.b(Unknown Source)
at de.thetaphi.forbiddenapis.asm.ClassReader.accept(Unknown Source)
at de.thetaphi.forbiddenapis.asm.ClassReader.accept(Unknown Source)
at de.thetaphi.forbiddenapis.Checker.checkClass(Checker.java:528)
at de.thetaphi.forbiddenapis.Checker.run(Checker.java:545)
... 77 more
BUILD FAILED
```
| 1.0 | Build is buggy for java 9. - Try compiling with java 9:
First you will hit this. This can be fixed by using `assertEquals`, which also simplifies the test.
```
/home/rmuir/workspace/elasticsearch/core/src/test/java/org/elasticsearch/threadpool/ScalingThreadPoolTests.java:180: error: method invoked with incorrect number of arguments; expected 2, found 0
assertThat(stats(threadPool, threadPoolName).getThreads(), equalTo(128));
^
/home/rmuir/workspace/elasticsearch/core/src/test/java/org/elasticsearch/threadpool/ScalingThreadPoolTests.java:196: error: method invoked with incorrect number of arguments; expected 2, found 0
assertThat(stats(threadPool, threadPoolName).getCompleted(), equalTo(128L));
^
```
Next problem is that I am unsure `-source/-target` settings are being passed at all. Because I see class files referencing java 9 StringConcatFactory.
Next problem is that forbidden-apis screams, presumably because its being run in the wrong JVM (not the JAVA_HOME that is set, but the java 8 one running gradle itself). It sees StringConcatFactory and fails the build:
```
Caused by: org.gradle.internal.UncheckedException: de.thetaphi.forbiddenapis.ForbiddenApiException: Check for forbidden API calls failed: java.lang.ClassNotFoundException: java.lang.invoke.StringConcatFactory
at org.gradle.internal.UncheckedException.throwAsUncheckedException(UncheckedException.java:45)
at org.gradle.internal.reflect.JavaMethod.invoke(JavaMethod.java:78)
at org.gradle.api.internal.project.taskfactory.AnnotationProcessingTaskFactory$StandardTaskAction.doExecute(AnnotationProcessingTaskFactory.java:227)
at org.gradle.api.internal.project.taskfactory.AnnotationProcessingTaskFactory$StandardTaskAction.execute(AnnotationProcessingTaskFactory.java:220)
at org.gradle.api.internal.project.taskfactory.AnnotationProcessingTaskFactory$StandardTaskAction.execute(AnnotationProcessingTaskFactory.java:209)
at org.gradle.api.internal.AbstractTask$TaskActionWrapper.execute(AbstractTask.java:585)
at org.gradle.api.internal.AbstractTask$TaskActionWrapper.execute(AbstractTask.java:568)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeAction(ExecuteActionsTaskExecuter.java:80)
at org.gradle.api.internal.tasks.execution.ExecuteActionsTaskExecuter.executeActions(ExecuteActionsTaskExecuter.java:61)
... 68 more
Caused by: de.thetaphi.forbiddenapis.ForbiddenApiException: Check for forbidden API calls failed: java.lang.ClassNotFoundException: java.lang.invoke.StringConcatFactory
at de.thetaphi.forbiddenapis.Checker.run(Checker.java:550)
at de.thetaphi.forbiddenapis.gradle.CheckForbiddenApis.checkForbidden(CheckForbiddenApis.java:551)
at org.gradle.internal.reflect.JavaMethod.invoke(JavaMethod.java:75)
... 75 more
Caused by: java.lang.ClassNotFoundException: java.lang.invoke.StringConcatFactory
at de.thetaphi.forbiddenapis.Checker.getClassFromClassLoader(Checker.java:264)
at de.thetaphi.forbiddenapis.Checker.lookupRelatedClass(Checker.java:277)
at de.thetaphi.forbiddenapis.ClassScanner$2.checkMethodAccess(ClassScanner.java:349)
at de.thetaphi.forbiddenapis.ClassScanner$2.checkHandle(ClassScanner.java:411)
at de.thetaphi.forbiddenapis.ClassScanner$2.visitInvokeDynamicInsn(ClassScanner.java:497)
at de.thetaphi.forbiddenapis.asm.ClassReader.a(Unknown Source)
at de.thetaphi.forbiddenapis.asm.ClassReader.b(Unknown Source)
at de.thetaphi.forbiddenapis.asm.ClassReader.accept(Unknown Source)
at de.thetaphi.forbiddenapis.asm.ClassReader.accept(Unknown Source)
at de.thetaphi.forbiddenapis.Checker.checkClass(Checker.java:528)
at de.thetaphi.forbiddenapis.Checker.run(Checker.java:545)
... 77 more
BUILD FAILED
```
| non_process | build is buggy for java try compiling with java first you will hit this this can be fixed by using assertequals which also simplifies the test home rmuir workspace elasticsearch core src test java org elasticsearch threadpool scalingthreadpooltests java error method invoked with incorrect number of arguments expected found assertthat stats threadpool threadpoolname getthreads equalto home rmuir workspace elasticsearch core src test java org elasticsearch threadpool scalingthreadpooltests java error method invoked with incorrect number of arguments expected found assertthat stats threadpool threadpoolname getcompleted equalto next problem is that i am unsure source target settings are being passed at all because i see class files referencing java stringconcatfactory next problem is that forbidden apis screams presumably because its being run in the wrong jvm not the java home that is set but the java one running gradle itself it sees stringconcatfactory and fails the build caused by org gradle internal uncheckedexception de thetaphi forbiddenapis forbiddenapiexception check for forbidden api calls failed java lang classnotfoundexception java lang invoke stringconcatfactory at org gradle internal uncheckedexception throwasuncheckedexception uncheckedexception java at org gradle internal reflect javamethod invoke javamethod java at org gradle api internal project taskfactory annotationprocessingtaskfactory standardtaskaction doexecute annotationprocessingtaskfactory java at org gradle api internal project taskfactory annotationprocessingtaskfactory standardtaskaction execute annotationprocessingtaskfactory java at org gradle api internal project taskfactory annotationprocessingtaskfactory standardtaskaction execute annotationprocessingtaskfactory java at org gradle api internal abstracttask taskactionwrapper execute abstracttask java at org gradle api internal abstracttask taskactionwrapper execute abstracttask java at org gradle api internal tasks execution executeactionstaskexecuter executeaction executeactionstaskexecuter java at org gradle api internal tasks execution executeactionstaskexecuter executeactions executeactionstaskexecuter java more caused by de thetaphi forbiddenapis forbiddenapiexception check for forbidden api calls failed java lang classnotfoundexception java lang invoke stringconcatfactory at de thetaphi forbiddenapis checker run checker java at de thetaphi forbiddenapis gradle checkforbiddenapis checkforbidden checkforbiddenapis java at org gradle internal reflect javamethod invoke javamethod java more caused by java lang classnotfoundexception java lang invoke stringconcatfactory at de thetaphi forbiddenapis checker getclassfromclassloader checker java at de thetaphi forbiddenapis checker lookuprelatedclass checker java at de thetaphi forbiddenapis classscanner checkmethodaccess classscanner java at de thetaphi forbiddenapis classscanner checkhandle classscanner java at de thetaphi forbiddenapis classscanner visitinvokedynamicinsn classscanner java at de thetaphi forbiddenapis asm classreader a unknown source at de thetaphi forbiddenapis asm classreader b unknown source at de thetaphi forbiddenapis asm classreader accept unknown source at de thetaphi forbiddenapis asm classreader accept unknown source at de thetaphi forbiddenapis checker checkclass checker java at de thetaphi forbiddenapis checker run checker java more build failed | 0 |
10,529 | 27,119,221,656 | IssuesEvent | 2023-02-15 21:14:33 | Azure/azure-sdk | https://api.github.com/repos/Azure/azure-sdk | closed | Board Review: Preview Features in Stable Packages | architecture board-review | The Cosmos team is requesting that we support shipping preview features in stable packages. Something like that already shipped in Cosmos Java packages, and the most recent request is for .NET Cosmos packages. We should talk at the arch board level if and how we want to support such requests.
The .NET platform supports some such scenarios. See https://github.com/dotnet/designs/blob/main/accepted/2021/preview-features/preview-features.md. The .NET platform solution does not work for the Azure SDK as it's not designed for Net Standard NuGet packages, but we met with the BCL team and we have a plan for how to make similar feature/checks working for Net Standard packages.
We also need to discuss how this would look for other languages. | 1.0 | Board Review: Preview Features in Stable Packages - The Cosmos team is requesting that we support shipping preview features in stable packages. Something like that already shipped in Cosmos Java packages, and the most recent request is for .NET Cosmos packages. We should talk at the arch board level if and how we want to support such requests.
The .NET platform supports some such scenarios. See https://github.com/dotnet/designs/blob/main/accepted/2021/preview-features/preview-features.md. The .NET platform solution does not work for the Azure SDK as it's not designed for Net Standard NuGet packages, but we met with the BCL team and we have a plan for how to make similar feature/checks working for Net Standard packages.
We also need to discuss how this would look for other languages. | non_process | board review preview features in stable packages the cosmos team is requesting that we support shipping preview features in stable packages something like that already shipped in cosmos java packages and the most recent request is for net cosmos packages we should talk at the arch board level if and how we want to support such requests the net platform supports some such scenarios see the net platform solution does not work for the azure sdk as it s not designed for net standard nuget packages but we met with the bcl team and we have a plan for how to make similar feature checks working for net standard packages we also need to discuss how this would look for other languages | 0 |
344,896 | 30,771,003,658 | IssuesEvent | 2023-07-30 22:37:45 | dancergraham/e57-python | https://api.github.com/repos/dancergraham/e57-python | closed | Tests: 🐇 bunny reading results | help wanted good first issue test | Test that the extents (max - min coordinates) of the bunny pointcloud are close to
Box dimensions
X: 0.155698
Y: 0.14731
Z: 0.120672
Global box center (median position)
X: -0.016840
Y: 0.113666
Z: -0.001537 | 1.0 | Tests: 🐇 bunny reading results - Test that the extents (max - min coordinates) of the bunny pointcloud are close to
Box dimensions
X: 0.155698
Y: 0.14731
Z: 0.120672
Global box center (median position)
X: -0.016840
Y: 0.113666
Z: -0.001537 | non_process | tests 🐇 bunny reading results test that the extents max min coordinates of the bunny pointcloud are close to box dimensions x y z global box center median position x y z | 0 |
1,096 | 3,563,661,759 | IssuesEvent | 2016-01-25 05:55:22 | kerubistan/kerub | https://api.github.com/repos/kerubistan/kerub | opened | Host power management | component:data processing enhancement | * Implement host power management methods (IPMI, WOL)
* Integrate into planner and executor
* allow entering host power management details on UI | 1.0 | Host power management - * Implement host power management methods (IPMI, WOL)
* Integrate into planner and executor
* allow entering host power management details on UI | process | host power management implement host power management methods ipmi wol integrate into planner and executor allow entering host power management details on ui | 1 |
19,558 | 25,882,147,989 | IssuesEvent | 2022-12-14 12:02:44 | gchq/stroom | https://api.github.com/repos/gchq/stroom | opened | Seeing `No effective streams found` errors when a ref data map contains XML data | bug f:processing | Works fine if the map contains simple string values. | 1.0 | Seeing `No effective streams found` errors when a ref data map contains XML data - Works fine if the map contains simple string values. | process | seeing no effective streams found errors when a ref data map contains xml data works fine if the map contains simple string values | 1 |
73,266 | 3,410,222,825 | IssuesEvent | 2015-12-04 19:10:46 | jamesmontemagno/Xamarin.Plugins | https://api.github.com/repos/jamesmontemagno/Xamarin.Plugins | opened | [Feature Request] Abstract Image Metadata & Return in MediaFile | enhancement Media priority-low | Originally from #76, what is requested is a way to abstract the metadata form the files into a common class that is returned with the file location in the Media File.
Things such as:
* Size
* Dimensions
* Date
* Exposure
* Camera Make
* Camera Model
* Etc.. | 1.0 | [Feature Request] Abstract Image Metadata & Return in MediaFile - Originally from #76, what is requested is a way to abstract the metadata form the files into a common class that is returned with the file location in the Media File.
Things such as:
* Size
* Dimensions
* Date
* Exposure
* Camera Make
* Camera Model
* Etc.. | non_process | abstract image metadata return in mediafile originally from what is requested is a way to abstract the metadata form the files into a common class that is returned with the file location in the media file things such as size dimensions date exposure camera make camera model etc | 0 |
7,393 | 10,520,436,847 | IssuesEvent | 2019-09-30 01:13:08 | qgis/QGIS | https://api.github.com/repos/qgis/QGIS | closed | Aggregate processing tool - Group by expression ($geometry) doesn't work | Bug Feedback Processing | Author Name: **Brent Edwards** (Brent Edwards)
Original Redmine Issue: [21740](https://issues.qgis.org/issues/21740)
Affected QGIS version: 3.6.0
Redmine category:processing/qgis
---
As the title states, when attempting to aggregate by geometry the results are inconclusive.
I tested the following expressions on a selection set of 2 identical polygons and they all produced the same result set of 2 polygons:
1) $geometry
2) geom_to_wkt($geometry)
3) centroid($geometry)
I was able to get the correct results by simply creating a virtual field with the same expression (ex. geom_to_wkt($geometry)) and the tool produced the correct result (single polygon with the agg fields).
| 1.0 | Aggregate processing tool - Group by expression ($geometry) doesn't work - Author Name: **Brent Edwards** (Brent Edwards)
Original Redmine Issue: [21740](https://issues.qgis.org/issues/21740)
Affected QGIS version: 3.6.0
Redmine category:processing/qgis
---
As the title states, when attempting to aggregate by geometry the results are inconclusive.
I tested the following expressions on a selection set of 2 identical polygons and they all produced the same result set of 2 polygons:
1) $geometry
2) geom_to_wkt($geometry)
3) centroid($geometry)
I was able to get the correct results by simply creating a virtual field with the same expression (ex. geom_to_wkt($geometry)) and the tool produced the correct result (single polygon with the agg fields).
| process | aggregate processing tool group by expression geometry doesn t work author name brent edwards brent edwards original redmine issue affected qgis version redmine category processing qgis as the title states when attempting to aggregate by geometry the results are inconclusive i tested the following expressions on a selection set of identical polygons and they all produced the same result set of polygons geometry geom to wkt geometry centroid geometry i was able to get the correct results by simply creating a virtual field with the same expression ex geom to wkt geometry and the tool produced the correct result single polygon with the agg fields | 1 |
8,317 | 11,485,924,023 | IssuesEvent | 2020-02-11 08:53:46 | geneontology/go-ontology | https://api.github.com/repos/geneontology/go-ontology | closed | ~obsolete~ merge GO:0010204 defense response signaling pathway, resistance gene-independent | multi-species process term merge | GO:0010204 defense response signaling pathway, resistance gene-independent
14 EXP annotations
This term should be obsoleted, becasue we shouldn't define things by what they aren't.
This would seem to apply to PAMP-mediated signalling, but could also apply to the non-immune system. There should always be a more precise term that could be used (or the parent)
| 1.0 | ~obsolete~ merge GO:0010204 defense response signaling pathway, resistance gene-independent - GO:0010204 defense response signaling pathway, resistance gene-independent
14 EXP annotations
This term should be obsoleted, becasue we shouldn't define things by what they aren't.
This would seem to apply to PAMP-mediated signalling, but could also apply to the non-immune system. There should always be a more precise term that could be used (or the parent)
| process | obsolete merge go defense response signaling pathway resistance gene independent go defense response signaling pathway resistance gene independent exp annotations this term should be obsoleted becasue we shouldn t define things by what they aren t this would seem to apply to pamp mediated signalling but could also apply to the non immune system there should always be a more precise term that could be used or the parent | 1 |
10,259 | 13,110,976,751 | IssuesEvent | 2020-08-04 21:49:20 | SCIInstitute/Seg3D | https://api.github.com/repos/SCIInstitute/Seg3D | opened | Better DICOM support | feature request file import software processes usability | - [ ] Ability to separate multiple volumes in the same folder
- [ ] Read in time series data such as: DWI/DTI and fMRI (can use for registration) | 1.0 | Better DICOM support - - [ ] Ability to separate multiple volumes in the same folder
- [ ] Read in time series data such as: DWI/DTI and fMRI (can use for registration) | process | better dicom support ability to separate multiple volumes in the same folder read in time series data such as dwi dti and fmri can use for registration | 1 |
138,953 | 18,797,799,672 | IssuesEvent | 2021-11-09 01:26:46 | rgordon95/conFusionAng | https://api.github.com/repos/rgordon95/conFusionAng | opened | CVE-2021-23400 (High) detected in nodemailer-2.7.2.tgz | security vulnerability | ## CVE-2021-23400 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nodemailer-2.7.2.tgz</b></p></summary>
<p>Easy as cake e-mail sending from your Node.js applications</p>
<p>Library home page: <a href="https://registry.npmjs.org/nodemailer/-/nodemailer-2.7.2.tgz">https://registry.npmjs.org/nodemailer/-/nodemailer-2.7.2.tgz</a></p>
<p>Path to dependency file: /conFusionAng/package.json</p>
<p>Path to vulnerable library: conFusionAng/node_modules/nodemailer/package.json</p>
<p>
Dependency Hierarchy:
- karma-2.0.5.tgz (Root Library)
- log4js-2.11.0.tgz
- :x: **nodemailer-2.7.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package nodemailer before 6.6.1 are vulnerable to HTTP Header Injection if unsanitized user input that may contain newlines and carriage returns is passed into an address object.
<p>Publish Date: 2021-06-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23400>CVE-2021-23400</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23400">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23400</a></p>
<p>Release Date: 2021-06-29</p>
<p>Fix Resolution: nodemailer - 6.6.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23400 (High) detected in nodemailer-2.7.2.tgz - ## CVE-2021-23400 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>nodemailer-2.7.2.tgz</b></p></summary>
<p>Easy as cake e-mail sending from your Node.js applications</p>
<p>Library home page: <a href="https://registry.npmjs.org/nodemailer/-/nodemailer-2.7.2.tgz">https://registry.npmjs.org/nodemailer/-/nodemailer-2.7.2.tgz</a></p>
<p>Path to dependency file: /conFusionAng/package.json</p>
<p>Path to vulnerable library: conFusionAng/node_modules/nodemailer/package.json</p>
<p>
Dependency Hierarchy:
- karma-2.0.5.tgz (Root Library)
- log4js-2.11.0.tgz
- :x: **nodemailer-2.7.2.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package nodemailer before 6.6.1 are vulnerable to HTTP Header Injection if unsanitized user input that may contain newlines and carriage returns is passed into an address object.
<p>Publish Date: 2021-06-29
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23400>CVE-2021-23400</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23400">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23400</a></p>
<p>Release Date: 2021-06-29</p>
<p>Fix Resolution: nodemailer - 6.6.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in nodemailer tgz cve high severity vulnerability vulnerable library nodemailer tgz easy as cake e mail sending from your node js applications library home page a href path to dependency file confusionang package json path to vulnerable library confusionang node modules nodemailer package json dependency hierarchy karma tgz root library tgz x nodemailer tgz vulnerable library vulnerability details the package nodemailer before are vulnerable to http header injection if unsanitized user input that may contain newlines and carriage returns is passed into an address object publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution nodemailer step up your open source security game with whitesource | 0 |
18,533 | 24,553,124,969 | IssuesEvent | 2022-10-12 14:01:02 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [PM] [VAPT] Red color disable icon should get displayed when admin mouse hovered on the disabled fields | Bug P2 Participant manager Process: Fixed Process: Tested QA Process: Tested dev | Red color disable icon should get displayed when the admin mouse hovered on the disabled fields
1. Participant details screen > Enrolled users > Enable invitation and Disable invitation
2. Forgot Password screen
3. My account screen | 3.0 | [PM] [VAPT] Red color disable icon should get displayed when admin mouse hovered on the disabled fields - Red color disable icon should get displayed when the admin mouse hovered on the disabled fields
1. Participant details screen > Enrolled users > Enable invitation and Disable invitation
2. Forgot Password screen
3. My account screen | process | red color disable icon should get displayed when admin mouse hovered on the disabled fields red color disable icon should get displayed when the admin mouse hovered on the disabled fields participant details screen enrolled users enable invitation and disable invitation forgot password screen my account screen | 1 |
20,615 | 27,288,630,083 | IssuesEvent | 2023-02-23 15:09:32 | googleapis/python-bigquery-sqlalchemy | https://api.github.com/repos/googleapis/python-bigquery-sqlalchemy | closed | compliance tests not running with pytest >= 7.0.0 | type: process priority: p3 api: bigquery | For some reason when the latest pytest is installed, I get
```
(compliance-3-8) (base) ➜ python-bigquery-sqlalchemy git:(tswast-prerelease) ✗ py.test -vv --junitxml=compliance_3.8_sponge_log.xml --reruns=3 --reruns-delay=60 '--only-rerun=403 Exceeded rate limits' '--only-rerun=409 Already Exists' '--only-rerun=404 Not found' '--only-rerun=400 Cannot execute DML over a non-existent table' tests/sqlalchemy_dialect_compliance/test_dialect_compliance.py
/Users/swast/src/github.com/googleapis/python-bigquery-sqlalchemy/.nox/compliance-3-8/lib/python3.8/site-packages/google/cloud/bigquery/client.py:535: UserWarning: Cannot create BigQuery Storage client, the dependency google-cloud-bigquery-storage is not installed.
warnings.warn(
====================================================================== test session starts =======================================================================
platform darwin -- Python 3.8.10, pytest-7.0.1, pluggy-1.0.0 -- /Users/swast/src/github.com/googleapis/python-bigquery-sqlalchemy/.nox/compliance-3-8/bin/python
cachedir: .pytest_cache
rootdir: /Users/swast/src/github.com/googleapis/python-bigquery-sqlalchemy, configfile: setup.cfg
plugins: rerunfailures-10.2
collected 0 items
---------------------- generated xml file: /Users/swast/src/github.com/googleapis/python-bigquery-sqlalchemy/compliance_3.8_sponge_log.xml -----------------------
===================================================================== no tests ran in 1.52s =====================================================================
```
And it fails with exit code 5. https://docs.pytest.org/en/latest/reference/exit-codes.html "No tests were collected" | 1.0 | compliance tests not running with pytest >= 7.0.0 - For some reason when the latest pytest is installed, I get
```
(compliance-3-8) (base) ➜ python-bigquery-sqlalchemy git:(tswast-prerelease) ✗ py.test -vv --junitxml=compliance_3.8_sponge_log.xml --reruns=3 --reruns-delay=60 '--only-rerun=403 Exceeded rate limits' '--only-rerun=409 Already Exists' '--only-rerun=404 Not found' '--only-rerun=400 Cannot execute DML over a non-existent table' tests/sqlalchemy_dialect_compliance/test_dialect_compliance.py
/Users/swast/src/github.com/googleapis/python-bigquery-sqlalchemy/.nox/compliance-3-8/lib/python3.8/site-packages/google/cloud/bigquery/client.py:535: UserWarning: Cannot create BigQuery Storage client, the dependency google-cloud-bigquery-storage is not installed.
warnings.warn(
====================================================================== test session starts =======================================================================
platform darwin -- Python 3.8.10, pytest-7.0.1, pluggy-1.0.0 -- /Users/swast/src/github.com/googleapis/python-bigquery-sqlalchemy/.nox/compliance-3-8/bin/python
cachedir: .pytest_cache
rootdir: /Users/swast/src/github.com/googleapis/python-bigquery-sqlalchemy, configfile: setup.cfg
plugins: rerunfailures-10.2
collected 0 items
---------------------- generated xml file: /Users/swast/src/github.com/googleapis/python-bigquery-sqlalchemy/compliance_3.8_sponge_log.xml -----------------------
===================================================================== no tests ran in 1.52s =====================================================================
```
And it fails with exit code 5. https://docs.pytest.org/en/latest/reference/exit-codes.html "No tests were collected" | process | compliance tests not running with pytest for some reason when the latest pytest is installed i get compliance base ➜ python bigquery sqlalchemy git tswast prerelease ✗ py test vv junitxml compliance sponge log xml reruns reruns delay only rerun exceeded rate limits only rerun already exists only rerun not found only rerun cannot execute dml over a non existent table tests sqlalchemy dialect compliance test dialect compliance py users swast src github com googleapis python bigquery sqlalchemy nox compliance lib site packages google cloud bigquery client py userwarning cannot create bigquery storage client the dependency google cloud bigquery storage is not installed warnings warn test session starts platform darwin python pytest pluggy users swast src github com googleapis python bigquery sqlalchemy nox compliance bin python cachedir pytest cache rootdir users swast src github com googleapis python bigquery sqlalchemy configfile setup cfg plugins rerunfailures collected items generated xml file users swast src github com googleapis python bigquery sqlalchemy compliance sponge log xml no tests ran in and it fails with exit code no tests were collected | 1 |
19,318 | 25,468,502,169 | IssuesEvent | 2022-11-25 07:57:15 | GoogleCloudPlatform/fda-mystudies | https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies | closed | [iOS] Not getting the study completion pop up in the study activities screen even though user submitted all the activities | Bug P1 iOS Process: Fixed Process: Tested QA | Not getting the study completed pop-up is not getting displayed on the study activities screen even though the user submitted all the activities present in the activities screen
Actual result:

Expected result:

| 2.0 | [iOS] Not getting the study completion pop up in the study activities screen even though user submitted all the activities - Not getting the study completed pop-up is not getting displayed on the study activities screen even though the user submitted all the activities present in the activities screen
Actual result:

Expected result:

| process | not getting the study completion pop up in the study activities screen even though user submitted all the activities not getting the study completed pop up is not getting displayed on the study activities screen even though the user submitted all the activities present in the activities screen actual result expected result | 1 |
217,709 | 16,870,089,474 | IssuesEvent | 2021-06-22 02:30:06 | thesofproject/sof | https://api.github.com/repos/thesofproject/sof | opened | [BUG][ADL]dai_config error when aplaying HDMI1 on ADL | Intel Daily tests bug | **Describe the bug**
When running aplay on HDMI1 , dai_config error occured
**To Reproduce**
on ADLP_RVP_SDW, run aplay -Dhw:0,5 -r 48000 -c 2 -f S16_LE -d 10 /dev/zero -v -q
on ADLP_RVP_HDA , run aplay -Dhw:0,3 -r 48000 -c 2 -f S16_LE -d 10 /dev/zero -v -q
**Reproduction Rate**
100%
**Environment**
Kernel Branch: topic/sof-dev
Kernel Commit: b1c762a1
SOF Branch: main
SOF Commit: 86c9a037e212
Topology: sof-hda-generic-4ch-kwd.tplg for ADLP_RVP_HDA ; sof-adl-rt711-4ch.tpl for ADLP_RVP_SDW
Platform: **ADLP_RVP_HDA ;ADLP_RVP_SDW**
**Screenshots or console output**
[dmesg]
```shell
ubuntu@sh-adlp-rvp-hda-01:~$ aplay -Dhw:0,5 48K_Let_It_Go.wav -vvv
Playing WAVE '48K_Let_It_Go.wav' : Signed 16 bit Little Endian, Rate 48000 Hz, Stereo
aplay: set_params:1407: Unable to install hw params:
```
[sof-logger]
```
[ 153826040.189586] ( 0.000000) c0 dma-trace src/trace/dma-trace.c:343 ERROR FW ABI 0x3013000 DBG ABI 0x5003000 tag v1.8-rc1-236-g86c9a037e212 src hash 0x6560584a (ldc hash 0x6560584a)
[ 238.124991] ( 238.124985) c0 ll-schedule ./schedule/ll_schedule.c:393 INFO task add 0x9e126740 dma-trace-task <2b972272-c5b1-4b7e-926f-0fc5cb4c4690>
[ 255.729157] ( 17.604166) c0 ll-schedule ./schedule/ll_schedule.c:397 INFO task params pri 4 flags 0 start 500000 period 500000
[ 281.458322] ( 25.729166) c0 ll-schedule ./schedule/ll_schedule.c:309 INFO new added task->start 3377618 at 3351437
[ 298.541655] ( 17.083332) c0 ll-schedule ./schedule/ll_schedule.c:312 INFO num_tasks 2 total_num_tasks 2
[ 951.666629] ( 653.125000) c0 ipc src/ipc/ipc3/handler.c:634 INFO ipc: pm -> restore
[ 4861.874807] ( 3910.208252) c0 component src/ipc/ipc3/helper.c:418 INFO comp new dai <c2b00d27-ffbc-4150-a51a-245c79c5e54b> type 2 id 9.57
[ 4909.114388] ( 47.239582) c0 hda-dai 3.6 /drivers/intel/hda/hda.c:69 INFO hda_probe()
[ 4932.499804] ( 23.385416) c0 dai src/lib/dai.c:164 INFO dai_get type 3 index 6 new sref 1
[ 4955.416470] ( 22.916666) c0 hda-dma ..../intel/hda/hda-dma.c:790 INFO hda-dmac :7 -> probe
[ 5041.093550] ( 85.677078) c0 dma src/lib/dma.c:121 INFO dma_get() ID 7 sref = 1 busy channels 0
[ 5690.156024] ( 649.062500) c0 ipc src/ipc/ipc3/handler.c:536 INFO ipc: dai 3.6 -> config
[ 5709.687273] ( 19.531250) c0 dai src/lib/dai.c:164 INFO dai_get type 3 index 6 new sref 2
[ 5727.603939] ( 17.916666) c0 hda-dai 3.6 /drivers/intel/hda/hda.c:36 INFO hda_set_config(): channels 2 rate 48000
[ 5745.781022] ( 18.177082) c0 dai src/lib/dai.c:188 INFO dai_put type 3 index 6 new sref 1
[ 5763.958104] ( 18.177082) c0 ipc src/ipc/ipc3/dai.c:159 INFO ipc_comp_dai_config() dai type = 3 index = 6
[ 5783.906020] ( 19.947916) c0 dai 9.57 src/ipc/ipc3/dai.c:223 INFO dai_config() dai type = 3 index = 6 dd 0x9e122000
[ 6482.812242] ( 698.906250) c0 ipc src/ipc/ipc3/handler.c:536 INFO ipc: dai 3.6 -> config
[ 6502.187242] ( 19.375000) c0 dai src/lib/dai.c:164 INFO dai_get type 3 index 6 new sref 2
[ 6519.739324] ( 17.552082) c0 hda-dai 3.6 /drivers/intel/hda/hda.c:36 INFO hda_set_config(): channels 2 rate 48000
[ 6537.499740] ( 17.760416) c0 dai src/lib/dai.c:188 INFO dai_put type 3 index 6 new sref 1
[ 6556.614323] ( 19.114582) c0 ipc src/ipc/ipc3/dai.c:159 INFO ipc_comp_dai_config() dai type = 3 index = 6
warn: failed to fread() 3 params from the log for src/ipc/ipc3/dai.c:223
```
| 1.0 | [BUG][ADL]dai_config error when aplaying HDMI1 on ADL - **Describe the bug**
When running aplay on HDMI1 , dai_config error occured
**To Reproduce**
on ADLP_RVP_SDW, run aplay -Dhw:0,5 -r 48000 -c 2 -f S16_LE -d 10 /dev/zero -v -q
on ADLP_RVP_HDA , run aplay -Dhw:0,3 -r 48000 -c 2 -f S16_LE -d 10 /dev/zero -v -q
**Reproduction Rate**
100%
**Environment**
Kernel Branch: topic/sof-dev
Kernel Commit: b1c762a1
SOF Branch: main
SOF Commit: 86c9a037e212
Topology: sof-hda-generic-4ch-kwd.tplg for ADLP_RVP_HDA ; sof-adl-rt711-4ch.tpl for ADLP_RVP_SDW
Platform: **ADLP_RVP_HDA ;ADLP_RVP_SDW**
**Screenshots or console output**
[dmesg]
```shell
ubuntu@sh-adlp-rvp-hda-01:~$ aplay -Dhw:0,5 48K_Let_It_Go.wav -vvv
Playing WAVE '48K_Let_It_Go.wav' : Signed 16 bit Little Endian, Rate 48000 Hz, Stereo
aplay: set_params:1407: Unable to install hw params:
```
[sof-logger]
```
[ 153826040.189586] ( 0.000000) c0 dma-trace src/trace/dma-trace.c:343 ERROR FW ABI 0x3013000 DBG ABI 0x5003000 tag v1.8-rc1-236-g86c9a037e212 src hash 0x6560584a (ldc hash 0x6560584a)
[ 238.124991] ( 238.124985) c0 ll-schedule ./schedule/ll_schedule.c:393 INFO task add 0x9e126740 dma-trace-task <2b972272-c5b1-4b7e-926f-0fc5cb4c4690>
[ 255.729157] ( 17.604166) c0 ll-schedule ./schedule/ll_schedule.c:397 INFO task params pri 4 flags 0 start 500000 period 500000
[ 281.458322] ( 25.729166) c0 ll-schedule ./schedule/ll_schedule.c:309 INFO new added task->start 3377618 at 3351437
[ 298.541655] ( 17.083332) c0 ll-schedule ./schedule/ll_schedule.c:312 INFO num_tasks 2 total_num_tasks 2
[ 951.666629] ( 653.125000) c0 ipc src/ipc/ipc3/handler.c:634 INFO ipc: pm -> restore
[ 4861.874807] ( 3910.208252) c0 component src/ipc/ipc3/helper.c:418 INFO comp new dai <c2b00d27-ffbc-4150-a51a-245c79c5e54b> type 2 id 9.57
[ 4909.114388] ( 47.239582) c0 hda-dai 3.6 /drivers/intel/hda/hda.c:69 INFO hda_probe()
[ 4932.499804] ( 23.385416) c0 dai src/lib/dai.c:164 INFO dai_get type 3 index 6 new sref 1
[ 4955.416470] ( 22.916666) c0 hda-dma ..../intel/hda/hda-dma.c:790 INFO hda-dmac :7 -> probe
[ 5041.093550] ( 85.677078) c0 dma src/lib/dma.c:121 INFO dma_get() ID 7 sref = 1 busy channels 0
[ 5690.156024] ( 649.062500) c0 ipc src/ipc/ipc3/handler.c:536 INFO ipc: dai 3.6 -> config
[ 5709.687273] ( 19.531250) c0 dai src/lib/dai.c:164 INFO dai_get type 3 index 6 new sref 2
[ 5727.603939] ( 17.916666) c0 hda-dai 3.6 /drivers/intel/hda/hda.c:36 INFO hda_set_config(): channels 2 rate 48000
[ 5745.781022] ( 18.177082) c0 dai src/lib/dai.c:188 INFO dai_put type 3 index 6 new sref 1
[ 5763.958104] ( 18.177082) c0 ipc src/ipc/ipc3/dai.c:159 INFO ipc_comp_dai_config() dai type = 3 index = 6
[ 5783.906020] ( 19.947916) c0 dai 9.57 src/ipc/ipc3/dai.c:223 INFO dai_config() dai type = 3 index = 6 dd 0x9e122000
[ 6482.812242] ( 698.906250) c0 ipc src/ipc/ipc3/handler.c:536 INFO ipc: dai 3.6 -> config
[ 6502.187242] ( 19.375000) c0 dai src/lib/dai.c:164 INFO dai_get type 3 index 6 new sref 2
[ 6519.739324] ( 17.552082) c0 hda-dai 3.6 /drivers/intel/hda/hda.c:36 INFO hda_set_config(): channels 2 rate 48000
[ 6537.499740] ( 17.760416) c0 dai src/lib/dai.c:188 INFO dai_put type 3 index 6 new sref 1
[ 6556.614323] ( 19.114582) c0 ipc src/ipc/ipc3/dai.c:159 INFO ipc_comp_dai_config() dai type = 3 index = 6
warn: failed to fread() 3 params from the log for src/ipc/ipc3/dai.c:223
```
| non_process | dai config error when aplaying on adl describe the bug when running aplay on dai config error occured to reproduce on adlp rvp sdw run aplay dhw r c f le d dev zero v q on adlp rvp hda run aplay dhw r c f le d dev zero v q reproduction rate environment kernel branch topic sof dev kernel commit sof branch main sof commit topology sof hda generic kwd tplg for adlp rvp hda sof adl tpl for adlp rvp sdw platform adlp rvp hda adlp rvp sdw screenshots or console output shell ubuntu sh adlp rvp hda aplay dhw let it go wav vvv playing wave let it go wav signed bit little endian rate hz stereo aplay set params unable to install hw params dma trace src trace dma trace c error fw abi dbg abi tag src hash ldc hash ll schedule schedule ll schedule c info task add dma trace task ll schedule schedule ll schedule c info task params pri flags start period ll schedule schedule ll schedule c info new added task start at ll schedule schedule ll schedule c info num tasks total num tasks ipc src ipc handler c info ipc pm restore component src ipc helper c info comp new dai type id hda dai drivers intel hda hda c info hda probe dai src lib dai c info dai get type index new sref hda dma intel hda hda dma c info hda dmac probe dma src lib dma c info dma get id sref busy channels ipc src ipc handler c info ipc dai config dai src lib dai c info dai get type index new sref hda dai drivers intel hda hda c info hda set config channels rate dai src lib dai c info dai put type index new sref ipc src ipc dai c info ipc comp dai config dai type index dai src ipc dai c info dai config dai type index dd ipc src ipc handler c info ipc dai config dai src lib dai c info dai get type index new sref hda dai drivers intel hda hda c info hda set config channels rate dai src lib dai c info dai put type index new sref ipc src ipc dai c info ipc comp dai config dai type index warn failed to fread params from the log for src ipc dai c | 0 |
240,052 | 19,989,489,444 | IssuesEvent | 2022-01-31 03:24:48 | datafuselabs/databend | https://api.github.com/repos/datafuselabs/databend | closed | Refine the stateful tests directory | C-testing C-good first issue | https://github.com/datafuselabs/databend/tree/3e1b530b9fa8ab16c140c9f4f16e582d9e768b8c/tests/suites/1_stateful
Move `12_0000_copy_s3.*` to `copy` folder, and rename to `00_0000_copy_s3`.
Move `12_0001_streaming_load.*` to `load` folder, and rename to `00_0000_streaming_load`. | 1.0 | Refine the stateful tests directory - https://github.com/datafuselabs/databend/tree/3e1b530b9fa8ab16c140c9f4f16e582d9e768b8c/tests/suites/1_stateful
Move `12_0000_copy_s3.*` to `copy` folder, and rename to `00_0000_copy_s3`.
Move `12_0001_streaming_load.*` to `load` folder, and rename to `00_0000_streaming_load`. | non_process | refine the stateful tests directory move copy to copy folder and rename to copy move streaming load to load folder and rename to streaming load | 0 |
560,511 | 16,598,521,722 | IssuesEvent | 2021-06-01 16:06:47 | appjusto/admin | https://api.github.com/repos/appjusto/admin | closed | Order situation "declined" in BO drawer | backoffice enhancement priority 0 | - [x] Show declined option;
- [x] Show declined issues (canceled). | 1.0 | Order situation "declined" in BO drawer - - [x] Show declined option;
- [x] Show declined issues (canceled). | non_process | order situation declined in bo drawer show declined option show declined issues canceled | 0 |
10,551 | 13,338,816,745 | IssuesEvent | 2020-08-28 11:46:33 | arcus-azure/arcus.messaging | https://api.github.com/repos/arcus-azure/arcus.messaging | closed | Provide catch-all message handler | area:message-processing enhancement feature | Provide a catch-all message handler that can be used in case no matching message handler was found. | 1.0 | Provide catch-all message handler - Provide a catch-all message handler that can be used in case no matching message handler was found. | process | provide catch all message handler provide a catch all message handler that can be used in case no matching message handler was found | 1 |
13,158 | 15,588,428,053 | IssuesEvent | 2021-03-18 06:22:50 | q191201771/lal | https://api.github.com/repos/q191201771/lal | closed | RTSP 1080视频花屏 | #Bug *In process *Waiting reply | ### 测试环境
- 测试版本:lal_v0.19.1_linux,默认配置
### 出现问题
- RTSP over udp 推流1080p视频:
`ffmpeg -re -stream_loop -1 -i "bbb_sunflower_1080p_30fps_normal.mp4" -an -vcodec copy -f rtsp rtsp://localhost:5544/live/test110`
- ffplay播放rtsp,看几秒钟后兔子出洞后花屏,每次重新播放花屏的时间和位置都一样。
- 转rtmp播放或者改为rtsp over tcp推流均没有问题
- RTSP over udp 推流小码流文件(source.200kbps.768x320.flv)没有这个问题
### 测试资源
- bbb_sunflower_1080p_30fps_normal.mp4 [国外](https://mirror.clarkson.edu/blender/demo/movies/BBB/)
- bbb_sunflower_1080p_30fps_normal.mp4 [国内](https://pan.baidu.com/s/1jI9BOTw)
| 1.0 | RTSP 1080视频花屏 - ### 测试环境
- 测试版本:lal_v0.19.1_linux,默认配置
### 出现问题
- RTSP over udp 推流1080p视频:
`ffmpeg -re -stream_loop -1 -i "bbb_sunflower_1080p_30fps_normal.mp4" -an -vcodec copy -f rtsp rtsp://localhost:5544/live/test110`
- ffplay播放rtsp,看几秒钟后兔子出洞后花屏,每次重新播放花屏的时间和位置都一样。
- 转rtmp播放或者改为rtsp over tcp推流均没有问题
- RTSP over udp 推流小码流文件(source.200kbps.768x320.flv)没有这个问题
### 测试资源
- bbb_sunflower_1080p_30fps_normal.mp4 [国外](https://mirror.clarkson.edu/blender/demo/movies/BBB/)
- bbb_sunflower_1080p_30fps_normal.mp4 [国内](https://pan.baidu.com/s/1jI9BOTw)
| process | rtsp 测试环境 测试版本:lal linux 默认配置 出现问题 rtsp over udp ffmpeg re stream loop i bbb sunflower normal an vcodec copy f rtsp rtsp localhost live ffplay播放rtsp,看几秒钟后兔子出洞后花屏,每次重新播放花屏的时间和位置都一样。 转rtmp播放或者改为rtsp over tcp推流均没有问题 rtsp over udp 推流小码流文件 source flv 没有这个问题 测试资源 bbb sunflower normal bbb sunflower normal | 1 |
192,999 | 22,216,033,451 | IssuesEvent | 2022-06-08 01:49:00 | AlexRogalskiy/github-action-json-fields | https://api.github.com/repos/AlexRogalskiy/github-action-json-fields | closed | CVE-2021-44907 (High) detected in qs-6.5.2.tgz - autoclosed | security vulnerability | ## CVE-2021-44907 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.5.2.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.5.2.tgz">https://registry.npmjs.org/qs/-/qs-6.5.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/qs/package.json,/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- coveralls-3.1.0.tgz (Root Library)
- request-2.88.2.tgz
- :x: **qs-6.5.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-json-fields/commit/b0a7808a1dde00217547fbb9b206a6ba2a8862f0">b0a7808a1dde00217547fbb9b206a6ba2a8862f0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Denial of Service vulnerability exists in qs up to 6.8.0 due to insufficient sanitization of property in the gs.parse function. The merge() function allows the assignment of properties on an array in the query. For any property being assigned, a value in the array is converted to an object containing these properties. Essentially, this means that the property whose expected type is Array always has to be checked with Array.isArray() by the user. This may not be obvious to the user and can cause unexpected behavior.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44907>CVE-2021-44907</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44907">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44907</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: qs - 6.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-44907 (High) detected in qs-6.5.2.tgz - autoclosed - ## CVE-2021-44907 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>qs-6.5.2.tgz</b></p></summary>
<p>A querystring parser that supports nesting and arrays, with a depth limit</p>
<p>Library home page: <a href="https://registry.npmjs.org/qs/-/qs-6.5.2.tgz">https://registry.npmjs.org/qs/-/qs-6.5.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/qs/package.json,/node_modules/qs/package.json</p>
<p>
Dependency Hierarchy:
- coveralls-3.1.0.tgz (Root Library)
- request-2.88.2.tgz
- :x: **qs-6.5.2.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/github-action-json-fields/commit/b0a7808a1dde00217547fbb9b206a6ba2a8862f0">b0a7808a1dde00217547fbb9b206a6ba2a8862f0</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Denial of Service vulnerability exists in qs up to 6.8.0 due to insufficient sanitization of property in the gs.parse function. The merge() function allows the assignment of properties on an array in the query. For any property being assigned, a value in the array is converted to an object containing these properties. Essentially, this means that the property whose expected type is Array always has to be checked with Array.isArray() by the user. This may not be obvious to the user and can cause unexpected behavior.
<p>Publish Date: 2022-03-17
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-44907>CVE-2021-44907</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44907">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-44907</a></p>
<p>Release Date: 2022-03-17</p>
<p>Fix Resolution: qs - 6.8.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in qs tgz autoclosed cve high severity vulnerability vulnerable library qs tgz a querystring parser that supports nesting and arrays with a depth limit library home page a href path to dependency file package json path to vulnerable library node modules npm node modules qs package json node modules qs package json dependency hierarchy coveralls tgz root library request tgz x qs tgz vulnerable library found in head commit a href vulnerability details a denial of service vulnerability exists in qs up to due to insufficient sanitization of property in the gs parse function the merge function allows the assignment of properties on an array in the query for any property being assigned a value in the array is converted to an object containing these properties essentially this means that the property whose expected type is array always has to be checked with array isarray by the user this may not be obvious to the user and can cause unexpected behavior publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution qs step up your open source security game with whitesource | 0 |
7,591 | 10,703,369,858 | IssuesEvent | 2019-10-24 09:25:39 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | ShardProcess: Trying to add a disposable to a DisposableStore that has already been disposed of | bug shared-process | Saw this shutting down quickly after starting up:
```
Error: Trying to add a disposable to a DisposableStore that has already been disposed of. The added object will be leaked!
at DisposableStore.add (/home/bpasero/development/monaco/out/vs/base/common/lifecycle.js:116:30)
at EventEmitter.<anonymous> (/home/bpasero/development/monaco/out/vs/code/electron-main/sharedProcess.js:91:33)
at Object.onceWrapper (events.js:288:20)
at EventEmitter.emit (events.js:200:13)
at WebContents.<anonymous> (/home/bpasero/development/monaco/.build/electron/resources/electron.asar/browser/api/web-contents.js:335:21)
at WebContents.emit (events.js:200:13)
``` | 1.0 | ShardProcess: Trying to add a disposable to a DisposableStore that has already been disposed of - Saw this shutting down quickly after starting up:
```
Error: Trying to add a disposable to a DisposableStore that has already been disposed of. The added object will be leaked!
at DisposableStore.add (/home/bpasero/development/monaco/out/vs/base/common/lifecycle.js:116:30)
at EventEmitter.<anonymous> (/home/bpasero/development/monaco/out/vs/code/electron-main/sharedProcess.js:91:33)
at Object.onceWrapper (events.js:288:20)
at EventEmitter.emit (events.js:200:13)
at WebContents.<anonymous> (/home/bpasero/development/monaco/.build/electron/resources/electron.asar/browser/api/web-contents.js:335:21)
at WebContents.emit (events.js:200:13)
``` | process | shardprocess trying to add a disposable to a disposablestore that has already been disposed of saw this shutting down quickly after starting up error trying to add a disposable to a disposablestore that has already been disposed of the added object will be leaked at disposablestore add home bpasero development monaco out vs base common lifecycle js at eventemitter home bpasero development monaco out vs code electron main sharedprocess js at object oncewrapper events js at eventemitter emit events js at webcontents home bpasero development monaco build electron resources electron asar browser api web contents js at webcontents emit events js | 1 |
12,537 | 14,972,532,978 | IssuesEvent | 2021-01-27 23:02:43 | panther-labs/panther | https://api.github.com/repos/panther-labs/panther | opened | Salesforce Log Puller | epic p0 team:data processing | ### Description
As an analyst, I want to be able to pull important Salesforce audit logs via a SaaS log experience.
### RFC
N/A
### Designs
TBD
### Acceptance Criteria
- step by step SaaS onboarding experience for Salesforce logs
- Agreed upon set of logs to support in MVP (after discussing with customers and log processing team) | 1.0 | Salesforce Log Puller - ### Description
As an analyst, I want to be able to pull important Salesforce audit logs via a SaaS log experience.
### RFC
N/A
### Designs
TBD
### Acceptance Criteria
- step by step SaaS onboarding experience for Salesforce logs
- Agreed upon set of logs to support in MVP (after discussing with customers and log processing team) | process | salesforce log puller description as an analyst i want to be able to pull important salesforce audit logs via a saas log experience rfc n a designs tbd acceptance criteria step by step saas onboarding experience for salesforce logs agreed upon set of logs to support in mvp after discussing with customers and log processing team | 1 |
709,057 | 24,366,467,626 | IssuesEvent | 2022-10-03 15:31:50 | TheGameCreators/AGK-Studio | https://api.github.com/repos/TheGameCreators/AGK-Studio | closed | Loss of Color in TextFontImages [Classic] | bug AGK-Core Low priority | FontImages have now lost color since the 2022.09.08 update (presumably related to the known issue re: the new [Text Shader](https://forum.thegamecreators.com/thread/228674#msg2671863) commands in Classic):
<img width="641" alt="wf" src="https://user-images.githubusercontent.com/52613918/193424637-da0c655e-18e4-4ba8-a80c-01424c7f4a43.png">
forum post [HERE](https://forum.thegamecreators.com/thread/228674#msg2671903) complete with project sample.
WIN 10
issue carries into HTML export, et al. | 1.0 | Loss of Color in TextFontImages [Classic] - FontImages have now lost color since the 2022.09.08 update (presumably related to the known issue re: the new [Text Shader](https://forum.thegamecreators.com/thread/228674#msg2671863) commands in Classic):
<img width="641" alt="wf" src="https://user-images.githubusercontent.com/52613918/193424637-da0c655e-18e4-4ba8-a80c-01424c7f4a43.png">
forum post [HERE](https://forum.thegamecreators.com/thread/228674#msg2671903) complete with project sample.
WIN 10
issue carries into HTML export, et al. | non_process | loss of color in textfontimages fontimages have now lost color since the update presumably related to the known issue re the new commands in classic img width alt wf src forum post complete with project sample win issue carries into html export et al | 0 |
60,384 | 3,125,996,295 | IssuesEvent | 2015-09-08 06:17:57 | NuGet/NuGetGallery | https://api.github.com/repos/NuGet/NuGetGallery | closed | Add Downloads by Client Report to Statistics Page | Priority - 2 | - This should show a graph.
- The data is from on a json blob created by the worker.
<!---
@huboard:{"order":4.25}
-->
| 1.0 | Add Downloads by Client Report to Statistics Page - - This should show a graph.
- The data is from on a json blob created by the worker.
<!---
@huboard:{"order":4.25}
-->
| non_process | add downloads by client report to statistics page this should show a graph the data is from on a json blob created by the worker huboard order | 0 |
23,049 | 3,755,593,109 | IssuesEvent | 2016-03-12 19:27:48 | RomanGolovanov/aMetro | https://api.github.com/repos/RomanGolovanov/aMetro | closed | Список карт для обновления при переходе через нотификацию не показывается | auto-migrated Component-Logic Priority-Medium Type-Defect | ```
What steps will reproduce the problem?
1. Переггрузить телефон;
2. Включить инет;
3. Дождаться проверки каталога, перейти по
нотификации, что есть обновления.
What is the expected output? What do you see instead?
Ожидается увидеть список массового
обновления карт. Показывается просто
список карт. Если же приложение уже было
запущено, ошибка повторяется через раз.
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `G.Glaur...@gmail.com` on 24 Aug 2010 at 9:22 | 1.0 | Список карт для обновления при переходе через нотификацию не показывается - ```
What steps will reproduce the problem?
1. Переггрузить телефон;
2. Включить инет;
3. Дождаться проверки каталога, перейти по
нотификации, что есть обновления.
What is the expected output? What do you see instead?
Ожидается увидеть список массового
обновления карт. Показывается просто
список карт. Если же приложение уже было
запущено, ошибка повторяется через раз.
Please use labels and text to provide additional information.
```
Original issue reported on code.google.com by `G.Glaur...@gmail.com` on 24 Aug 2010 at 9:22 | non_process | список карт для обновления при переходе через нотификацию не показывается what steps will reproduce the problem переггрузить телефон включить инет дождаться проверки каталога перейти по нотификации что есть обновления what is the expected output what do you see instead ожидается увидеть список массового обновления карт показывается просто список карт если же приложение уже было запущено ошибка повторяется через раз please use labels and text to provide additional information original issue reported on code google com by g glaur gmail com on aug at | 0 |
21,540 | 3,518,268,932 | IssuesEvent | 2016-01-12 12:01:10 | Virtual-Labs/problem-solving-iiith | https://api.github.com/repos/Virtual-Labs/problem-solving-iiith | reopened | QA_Permutation_UI | Category :UI Defect raised on: 26-11-2015 Developed by:IIIT Hyd Release Number Severity :S3 Status :Open Version Number :1.1 | Defect Description:
In the Landing page of "Permutation" experiment, the 'Home' &'Problem Solving Lab' links are present outside of the page width instead the links should be placed within the page limit inorder to maintain the page utility .
Actual Result:
In the Landing page of "Permutation" experiment,the 'Home' &'Problem Solving Lab' links are placed outside of the page width.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/problem-solving-iiith/blob/master/test-cases/integration_test-cases/Permutation/Permutation_01_Usability_smk.org

| 1.0 | QA_Permutation_UI - Defect Description:
In the Landing page of "Permutation" experiment, the 'Home' &'Problem Solving Lab' links are present outside of the page width instead the links should be placed within the page limit inorder to maintain the page utility .
Actual Result:
In the Landing page of "Permutation" experiment,the 'Home' &'Problem Solving Lab' links are placed outside of the page width.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/problem-solving-iiith/blob/master/test-cases/integration_test-cases/Permutation/Permutation_01_Usability_smk.org

| non_process | qa permutation ui defect description in the landing page of permutation experiment the home problem solving lab links are present outside of the page width instead the links should be placed within the page limit inorder to maintain the page utility actual result in the landing page of permutation experiment the home problem solving lab links are placed outside of the page width environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link | 0 |
9,040 | 6,735,020,458 | IssuesEvent | 2017-10-18 20:11:00 | Microsoft/BotBuilder | https://api.github.com/repos/Microsoft/BotBuilder | closed | [Question] Bot posts duplicate messages when bad connectivity | .NET SDK performance question Skype | <!--
Have a question?
Ask on http://stackoverflow.com/questions/tagged/botframework using the tag "botframework".
-->
## System Information
* SDK Language: .NET
* SDK Version: v.3.9.0
* Development Environment: <!-- localhost, Auzure Bot Service, App Service, other -->
## Bot Info <!-- As registered in the Bot Developer Portal at https://dev.botframework.com -->
* Bot name: Planner Bot
* Bot handle: taskplannerbot
* App ID: 971ebb48-c75a-471c-bee8-3f85033430b6
* Connected channels: Skype
## Issue Description
Bot posts (random number of) duplicated messages on Skype channel when bad connectivity.
## Code Example
<!-- Share your bot code that causes the problem. A complete working example is best. -->
## Steps to Reproduce the Issue
1. Make Skype running in a bad internet connection.
2. Randomly post message out from Bot.
## Expected Behavior
I am not sure if Bot Builder itself has some special ensure of message receive and attempt logic but it should not post duplicate messages even if it failed to ensure message received.
## Actual Results

After some repeat post messages it produces an exception that I could catch (in screenshot is when it post "Something went wrong..." message) which already attached here - [log.xlsx](https://github.com/Microsoft/BotBuilder/files/1199972/log.xlsx)
| True | [Question] Bot posts duplicate messages when bad connectivity - <!--
Have a question?
Ask on http://stackoverflow.com/questions/tagged/botframework using the tag "botframework".
-->
## System Information
* SDK Language: .NET
* SDK Version: v.3.9.0
* Development Environment: <!-- localhost, Auzure Bot Service, App Service, other -->
## Bot Info <!-- As registered in the Bot Developer Portal at https://dev.botframework.com -->
* Bot name: Planner Bot
* Bot handle: taskplannerbot
* App ID: 971ebb48-c75a-471c-bee8-3f85033430b6
* Connected channels: Skype
## Issue Description
Bot posts (random number of) duplicated messages on Skype channel when bad connectivity.
## Code Example
<!-- Share your bot code that causes the problem. A complete working example is best. -->
## Steps to Reproduce the Issue
1. Make Skype running in a bad internet connection.
2. Randomly post message out from Bot.
## Expected Behavior
I am not sure if Bot Builder itself has some special ensure of message receive and attempt logic but it should not post duplicate messages even if it failed to ensure message received.
## Actual Results

After some repeat post messages it produces an exception that I could catch (in screenshot is when it post "Something went wrong..." message) which already attached here - [log.xlsx](https://github.com/Microsoft/BotBuilder/files/1199972/log.xlsx)
| non_process | bot posts duplicate messages when bad connectivity have a question ask on using the tag botframework system information sdk language net sdk version v development environment bot info bot name planner bot bot handle taskplannerbot app id connected channels skype issue description bot posts random number of duplicated messages on skype channel when bad connectivity code example steps to reproduce the issue make skype running in a bad internet connection randomly post message out from bot expected behavior i am not sure if bot builder itself has some special ensure of message receive and attempt logic but it should not post duplicate messages even if it failed to ensure message received actual results after some repeat post messages it produces an exception that i could catch in screenshot is when it post something went wrong message which already attached here | 0 |
160,985 | 20,120,376,021 | IssuesEvent | 2022-02-08 01:12:55 | arohablue/BlockDockServer | https://api.github.com/repos/arohablue/BlockDockServer | opened | postgresql-9.4.1211.jre7.jar: 2 vulnerabilities (highest severity is: 9.8) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-9.4.1211.jre7.jar</b></p></summary>
<p>Java JDBC 4.1 (JRE 7+) driver for PostgreSQL database</p>
<p>Library home page: <a href="https://github.com/pgjdbc/pgjdbc-parent-poms">https://github.com/pgjdbc/pgjdbc-parent-poms</a></p>
<p>Path to dependency file: /BlockDockServer/build.gradle</p>
<p>Path to vulnerable library: /radle/caches/modules-2/files-2.1/org.postgresql/postgresql/9.4.1211.jre7/56b01e9e667f408818a6ef06a89598dbab80687d/postgresql-9.4.1211.jre7.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-21724](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | postgresql-9.4.1211.jre7.jar | Direct | org.postgresql:postgresql:42.2.25,42.3.2 | ❌ |
| [CVE-2020-13692](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.7 | postgresql-9.4.1211.jre7.jar | Direct | org.postgresql:postgresql:42.2.13 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-21724</summary>
### Vulnerable Library - <b>postgresql-9.4.1211.jre7.jar</b></p>
<p>Java JDBC 4.1 (JRE 7+) driver for PostgreSQL database</p>
<p>Library home page: <a href="https://github.com/pgjdbc/pgjdbc-parent-poms">https://github.com/pgjdbc/pgjdbc-parent-poms</a></p>
<p>Path to dependency file: /BlockDockServer/build.gradle</p>
<p>Path to vulnerable library: /radle/caches/modules-2/files-2.1/org.postgresql/postgresql/9.4.1211.jre7/56b01e9e667f408818a6ef06a89598dbab80687d/postgresql-9.4.1211.jre7.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-9.4.1211.jre7.jar** (Vulnerable Library)
</p>
<p></p>
### Vulnerability Details
<p>
pgjdbc is the offical PostgreSQL JDBC Driver. A security hole was found in the jdbc driver for postgresql database while doing security research. The system using the postgresql library will be attacked when attacker control the jdbc url or properties. pgjdbc instantiates plugin instances based on class names provided via `authenticationPluginClassName`, `sslhostnameverifier`, `socketFactory`, `sslfactory`, `sslpasswordcallback` connection properties. However, the driver did not verify if the class implements the expected interface before instantiating the class. This can lead to code execution loaded via arbitrary classes. Users using plugins are advised to upgrade. There are no known workarounds for this issue.
<p>Publish Date: 2022-02-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724>CVE-2022-21724</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-v7wg-cpwc-24m4">https://github.com/advisories/GHSA-v7wg-cpwc-24m4</a></p>
<p>Release Date: 2022-02-02</p>
<p>Fix Resolution: org.postgresql:postgresql:42.2.25,42.3.2</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-13692</summary>
### Vulnerable Library - <b>postgresql-9.4.1211.jre7.jar</b></p>
<p>Java JDBC 4.1 (JRE 7+) driver for PostgreSQL database</p>
<p>Library home page: <a href="https://github.com/pgjdbc/pgjdbc-parent-poms">https://github.com/pgjdbc/pgjdbc-parent-poms</a></p>
<p>Path to dependency file: /BlockDockServer/build.gradle</p>
<p>Path to vulnerable library: /radle/caches/modules-2/files-2.1/org.postgresql/postgresql/9.4.1211.jre7/56b01e9e667f408818a6ef06a89598dbab80687d/postgresql-9.4.1211.jre7.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-9.4.1211.jre7.jar** (Vulnerable Library)
</p>
<p></p>
### Vulnerability Details
<p>
PostgreSQL JDBC Driver (aka PgJDBC) before 42.2.13 allows XXE.
<p>Publish Date: 2020-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692>CVE-2020-13692</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.13">https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.13</a></p>
<p>Release Date: 2020-06-04</p>
<p>Fix Resolution: org.postgresql:postgresql:42.2.13</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.postgresql","packageName":"postgresql","packageVersion":"9.4.1211.jre7","packageFilePaths":["/BlockDockServer/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.postgresql:postgresql:9.4.1211.jre7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.postgresql:postgresql:42.2.25,42.3.2","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2022-21724","vulnerabilityDetails":"pgjdbc is the offical PostgreSQL JDBC Driver. A security hole was found in the jdbc driver for postgresql database while doing security research. The system using the postgresql library will be attacked when attacker control the jdbc url or properties. pgjdbc instantiates plugin instances based on class names provided via `authenticationPluginClassName`, `sslhostnameverifier`, `socketFactory`, `sslfactory`, `sslpasswordcallback` connection properties. However, the driver did not verify if the class implements the expected interface before instantiating the class. This can lead to code execution loaded via arbitrary classes. Users using plugins are advised to upgrade. There are no known workarounds for this issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.postgresql","packageName":"postgresql","packageVersion":"9.4.1211.jre7","packageFilePaths":["/BlockDockServer/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.postgresql:postgresql:9.4.1211.jre7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.postgresql:postgresql:42.2.13","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2020-13692","vulnerabilityDetails":"PostgreSQL JDBC Driver (aka PgJDBC) before 42.2.13 allows XXE.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692","cvss3Severity":"high","cvss3Score":"7.7","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> --> | True | postgresql-9.4.1211.jre7.jar: 2 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>postgresql-9.4.1211.jre7.jar</b></p></summary>
<p>Java JDBC 4.1 (JRE 7+) driver for PostgreSQL database</p>
<p>Library home page: <a href="https://github.com/pgjdbc/pgjdbc-parent-poms">https://github.com/pgjdbc/pgjdbc-parent-poms</a></p>
<p>Path to dependency file: /BlockDockServer/build.gradle</p>
<p>Path to vulnerable library: /radle/caches/modules-2/files-2.1/org.postgresql/postgresql/9.4.1211.jre7/56b01e9e667f408818a6ef06a89598dbab80687d/postgresql-9.4.1211.jre7.jar</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-21724](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | postgresql-9.4.1211.jre7.jar | Direct | org.postgresql:postgresql:42.2.25,42.3.2 | ❌ |
| [CVE-2020-13692](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.7 | postgresql-9.4.1211.jre7.jar | Direct | org.postgresql:postgresql:42.2.13 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-21724</summary>
### Vulnerable Library - <b>postgresql-9.4.1211.jre7.jar</b></p>
<p>Java JDBC 4.1 (JRE 7+) driver for PostgreSQL database</p>
<p>Library home page: <a href="https://github.com/pgjdbc/pgjdbc-parent-poms">https://github.com/pgjdbc/pgjdbc-parent-poms</a></p>
<p>Path to dependency file: /BlockDockServer/build.gradle</p>
<p>Path to vulnerable library: /radle/caches/modules-2/files-2.1/org.postgresql/postgresql/9.4.1211.jre7/56b01e9e667f408818a6ef06a89598dbab80687d/postgresql-9.4.1211.jre7.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-9.4.1211.jre7.jar** (Vulnerable Library)
</p>
<p></p>
### Vulnerability Details
<p>
pgjdbc is the offical PostgreSQL JDBC Driver. A security hole was found in the jdbc driver for postgresql database while doing security research. The system using the postgresql library will be attacked when attacker control the jdbc url or properties. pgjdbc instantiates plugin instances based on class names provided via `authenticationPluginClassName`, `sslhostnameverifier`, `socketFactory`, `sslfactory`, `sslpasswordcallback` connection properties. However, the driver did not verify if the class implements the expected interface before instantiating the class. This can lead to code execution loaded via arbitrary classes. Users using plugins are advised to upgrade. There are no known workarounds for this issue.
<p>Publish Date: 2022-02-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724>CVE-2022-21724</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>9.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-v7wg-cpwc-24m4">https://github.com/advisories/GHSA-v7wg-cpwc-24m4</a></p>
<p>Release Date: 2022-02-02</p>
<p>Fix Resolution: org.postgresql:postgresql:42.2.25,42.3.2</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2020-13692</summary>
### Vulnerable Library - <b>postgresql-9.4.1211.jre7.jar</b></p>
<p>Java JDBC 4.1 (JRE 7+) driver for PostgreSQL database</p>
<p>Library home page: <a href="https://github.com/pgjdbc/pgjdbc-parent-poms">https://github.com/pgjdbc/pgjdbc-parent-poms</a></p>
<p>Path to dependency file: /BlockDockServer/build.gradle</p>
<p>Path to vulnerable library: /radle/caches/modules-2/files-2.1/org.postgresql/postgresql/9.4.1211.jre7/56b01e9e667f408818a6ef06a89598dbab80687d/postgresql-9.4.1211.jre7.jar</p>
<p>
Dependency Hierarchy:
- :x: **postgresql-9.4.1211.jre7.jar** (Vulnerable Library)
</p>
<p></p>
### Vulnerability Details
<p>
PostgreSQL JDBC Driver (aka PgJDBC) before 42.2.13 allows XXE.
<p>Publish Date: 2020-06-04
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692>CVE-2020-13692</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.13">https://jdbc.postgresql.org/documentation/changelog.html#version_42.2.13</a></p>
<p>Release Date: 2020-06-04</p>
<p>Fix Resolution: org.postgresql:postgresql:42.2.13</p>
</p>
<p></p>
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details>
<!-- <REMEDIATE>[{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.postgresql","packageName":"postgresql","packageVersion":"9.4.1211.jre7","packageFilePaths":["/BlockDockServer/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.postgresql:postgresql:9.4.1211.jre7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.postgresql:postgresql:42.2.25,42.3.2","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2022-21724","vulnerabilityDetails":"pgjdbc is the offical PostgreSQL JDBC Driver. A security hole was found in the jdbc driver for postgresql database while doing security research. The system using the postgresql library will be attacked when attacker control the jdbc url or properties. pgjdbc instantiates plugin instances based on class names provided via `authenticationPluginClassName`, `sslhostnameverifier`, `socketFactory`, `sslfactory`, `sslpasswordcallback` connection properties. However, the driver did not verify if the class implements the expected interface before instantiating the class. This can lead to code execution loaded via arbitrary classes. Users using plugins are advised to upgrade. There are no known workarounds for this issue.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21724","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}},{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.postgresql","packageName":"postgresql","packageVersion":"9.4.1211.jre7","packageFilePaths":["/BlockDockServer/build.gradle"],"isTransitiveDependency":false,"dependencyTree":"org.postgresql:postgresql:9.4.1211.jre7","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.postgresql:postgresql:42.2.13","isBinary":false}],"baseBranches":[],"vulnerabilityIdentifier":"CVE-2020-13692","vulnerabilityDetails":"PostgreSQL JDBC Driver (aka PgJDBC) before 42.2.13 allows XXE.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-13692","cvss3Severity":"high","cvss3Score":"7.7","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"Low"},"extraData":{}}]</REMEDIATE> --> | non_process | postgresql jar vulnerabilities highest severity is vulnerable library postgresql jar java jdbc jre driver for postgresql database library home page a href path to dependency file blockdockserver build gradle path to vulnerable library radle caches modules files org postgresql postgresql postgresql jar vulnerabilities cve severity cvss dependency type fixed in remediation available high postgresql jar direct org postgresql postgresql ❌ high postgresql jar direct org postgresql postgresql ❌ details cve vulnerable library postgresql jar java jdbc jre driver for postgresql database library home page a href path to dependency file blockdockserver build gradle path to vulnerable library radle caches modules files org postgresql postgresql postgresql jar dependency hierarchy x postgresql jar vulnerable library vulnerability details pgjdbc is the offical postgresql jdbc driver a security hole was found in the jdbc driver for postgresql database while doing security research the system using the postgresql library will be attacked when attacker control the jdbc url or properties pgjdbc instantiates plugin instances based on class names provided via authenticationpluginclassname sslhostnameverifier socketfactory sslfactory sslpasswordcallback connection properties however the driver did not verify if the class implements the expected interface before instantiating the class this can lead to code execution loaded via arbitrary classes users using plugins are advised to upgrade there are no known workarounds for this issue publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org postgresql postgresql step up your open source security game with whitesource cve vulnerable library postgresql jar java jdbc jre driver for postgresql database library home page a href path to dependency file blockdockserver build gradle path to vulnerable library radle caches modules files org postgresql postgresql postgresql jar dependency hierarchy x postgresql jar vulnerable library vulnerability details postgresql jdbc driver aka pgjdbc before allows xxe publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org postgresql postgresql step up your open source security game with whitesource istransitivedependency false dependencytree org postgresql postgresql isminimumfixversionavailable true minimumfixversion org postgresql postgresql isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails pgjdbc is the offical postgresql jdbc driver a security hole was found in the jdbc driver for postgresql database while doing security research the system using the postgresql library will be attacked when attacker control the jdbc url or properties pgjdbc instantiates plugin instances based on class names provided via authenticationpluginclassname sslhostnameverifier socketfactory sslfactory sslpasswordcallback connection properties however the driver did not verify if the class implements the expected interface before instantiating the class this can lead to code execution loaded via arbitrary classes users using plugins are advised to upgrade there are no known workarounds for this issue vulnerabilityurl istransitivedependency false dependencytree org postgresql postgresql isminimumfixversionavailable true minimumfixversion org postgresql postgresql isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails postgresql jdbc driver aka pgjdbc before allows xxe vulnerabilityurl | 0 |
155,073 | 19,765,661,807 | IssuesEvent | 2022-01-17 01:40:55 | tuanducdesign/email-sending-verification | https://api.github.com/repos/tuanducdesign/email-sending-verification | opened | CVE-2021-23424 (High) detected in ansi-html-0.0.7.tgz | security vulnerability | ## CVE-2021-23424 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansi-html-0.0.7.tgz</b></p></summary>
<p>An elegant lib that converts the chalked (ANSI) text to HTML.</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-html/-/ansi-html-0.0.7.tgz">https://registry.npmjs.org/ansi-html/-/ansi-html-0.0.7.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/ansi-html/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.0.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- :x: **ansi-html-0.0.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tuanducdesign/email-sending-verification/commit/6c0c478a4770af860b9a85980a5919a9f37493b7">6c0c478a4770af860b9a85980a5919a9f37493b7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects all versions of package ansi-html. If an attacker provides a malicious string, it will get stuck processing the input for an extremely long time.
<p>Publish Date: 2021-08-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23424>CVE-2021-23424</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-23424 (High) detected in ansi-html-0.0.7.tgz - ## CVE-2021-23424 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>ansi-html-0.0.7.tgz</b></p></summary>
<p>An elegant lib that converts the chalked (ANSI) text to HTML.</p>
<p>Library home page: <a href="https://registry.npmjs.org/ansi-html/-/ansi-html-0.0.7.tgz">https://registry.npmjs.org/ansi-html/-/ansi-html-0.0.7.tgz</a></p>
<p>Path to dependency file: /client/package.json</p>
<p>Path to vulnerable library: /client/node_modules/ansi-html/package.json</p>
<p>
Dependency Hierarchy:
- react-scripts-4.0.0.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- :x: **ansi-html-0.0.7.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/tuanducdesign/email-sending-verification/commit/6c0c478a4770af860b9a85980a5919a9f37493b7">6c0c478a4770af860b9a85980a5919a9f37493b7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
This affects all versions of package ansi-html. If an attacker provides a malicious string, it will get stuck processing the input for an extremely long time.
<p>Publish Date: 2021-08-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23424>CVE-2021-23424</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve high detected in ansi html tgz cve high severity vulnerability vulnerable library ansi html tgz an elegant lib that converts the chalked ansi text to html library home page a href path to dependency file client package json path to vulnerable library client node modules ansi html package json dependency hierarchy react scripts tgz root library webpack dev server tgz x ansi html tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects all versions of package ansi html if an attacker provides a malicious string it will get stuck processing the input for an extremely long time publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with whitesource | 0 |
54,213 | 23,202,492,461 | IssuesEvent | 2022-08-01 23:30:38 | ArctosDB/arctos | https://api.github.com/repos/ArctosDB/arctos | closed | Advice Request: Tracking Extinct and Endangered Species in Arctos | Function-Taxonomy/Identification Enhancement Service-related | Has anyone come up with a way to track extinct / endangered / threatened species in Arctos? Whether it is a report or project or attribute, I'm curious to know how this has been addressed.
As a part of our permit process, we report specimens in our holdings to IDNR of endangered and threatened species held in our collections. We also just started a project photographing extinct bird species in our collections. It would be a lot easier to find specimens for these types of projects through an internal Arctos search. Maybe it could be a new feature that is tied to taxonomy? | 1.0 | Advice Request: Tracking Extinct and Endangered Species in Arctos - Has anyone come up with a way to track extinct / endangered / threatened species in Arctos? Whether it is a report or project or attribute, I'm curious to know how this has been addressed.
As a part of our permit process, we report specimens in our holdings to IDNR of endangered and threatened species held in our collections. We also just started a project photographing extinct bird species in our collections. It would be a lot easier to find specimens for these types of projects through an internal Arctos search. Maybe it could be a new feature that is tied to taxonomy? | non_process | advice request tracking extinct and endangered species in arctos has anyone come up with a way to track extinct endangered threatened species in arctos whether it is a report or project or attribute i m curious to know how this has been addressed as a part of our permit process we report specimens in our holdings to idnr of endangered and threatened species held in our collections we also just started a project photographing extinct bird species in our collections it would be a lot easier to find specimens for these types of projects through an internal arctos search maybe it could be a new feature that is tied to taxonomy | 0 |
19,089 | 25,140,813,109 | IssuesEvent | 2022-11-09 22:56:16 | open-telemetry/opentelemetry-collector-contrib | https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib | closed | [resourceprocessor] Specify list of resource attributes allowed to reach exporters | Stale processor/resource | **Is your feature request related to a problem? Please describe.**
As exporters can use [resource_to_telemetry_conversion](https://github.com/open-telemetry/opentelemetry-collector/blob/main/exporter/exporterhelper/README.md) to convert resource attributes into metric labels, ideally there should be a way to easily filter resource attributes.
**Describe the solution you'd like**
It is possible use the [delete action](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/resourceprocessor), but [include and exclude](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/attributesprocessor#includeexclude-filtering) that attributesprocessor has would be helpful.
**Describe alternatives you've considered**
I could look at all the resource attributes that exist and drop them, but, as I only want around 3 resource attributes, include would be easier.
**Additional context**
I'm using [awscontainerinsightreceiver](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/receiver/awscontainerinsightreceiver).
| 1.0 | [resourceprocessor] Specify list of resource attributes allowed to reach exporters - **Is your feature request related to a problem? Please describe.**
As exporters can use [resource_to_telemetry_conversion](https://github.com/open-telemetry/opentelemetry-collector/blob/main/exporter/exporterhelper/README.md) to convert resource attributes into metric labels, ideally there should be a way to easily filter resource attributes.
**Describe the solution you'd like**
It is possible use the [delete action](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/resourceprocessor), but [include and exclude](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/processor/attributesprocessor#includeexclude-filtering) that attributesprocessor has would be helpful.
**Describe alternatives you've considered**
I could look at all the resource attributes that exist and drop them, but, as I only want around 3 resource attributes, include would be easier.
**Additional context**
I'm using [awscontainerinsightreceiver](https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/receiver/awscontainerinsightreceiver).
| process | specify list of resource attributes allowed to reach exporters is your feature request related to a problem please describe as exporters can use to convert resource attributes into metric labels ideally there should be a way to easily filter resource attributes describe the solution you d like it is possible use the but that attributesprocessor has would be helpful describe alternatives you ve considered i could look at all the resource attributes that exist and drop them but as i only want around resource attributes include would be easier additional context i m using | 1 |
16,075 | 20,248,946,964 | IssuesEvent | 2022-02-14 16:07:28 | Bone008/orbiteye | https://api.github.com/repos/Bone008/orbiteye | opened | Process "Purpose" column for SATCAT data | data processing | Using the multiple separated sheets from [Celestrak](https://celestrak.com/NORAD/elements/) label the purpose of SATCAT rows where possible. | 1.0 | Process "Purpose" column for SATCAT data - Using the multiple separated sheets from [Celestrak](https://celestrak.com/NORAD/elements/) label the purpose of SATCAT rows where possible. | process | process purpose column for satcat data using the multiple separated sheets from label the purpose of satcat rows where possible | 1 |
74,934 | 9,167,723,230 | IssuesEvent | 2019-03-02 16:11:30 | boostorg/beast | https://api.github.com/repos/boostorg/beast | opened | string_base ideas | Design Feature | We can refactor `static_string`:
* Call it `char_string_base`, make it a normal class (i.e. not a template)
* Eschew `char_traits` and `CharT`, this is for `char` **only** and no locales
* Express all operations in terms of a virtual member `realloc`
* function by function compatibility with `std;:string`
* Full tests and coverage,
Now we can rewrite `static_string` using `char_string_base`, where `realloc` just throws if the static buffer is exceeded.
Then we can write a new class, a "flexible string" that has both an inline buffer, and an allocator. When the inline buffer overflows, the implementation switches to a dynamic allocation. This can replace some of the code that currently does this manually (e.g. in the parser and RFC range adaptors).
| 1.0 | string_base ideas - We can refactor `static_string`:
* Call it `char_string_base`, make it a normal class (i.e. not a template)
* Eschew `char_traits` and `CharT`, this is for `char` **only** and no locales
* Express all operations in terms of a virtual member `realloc`
* function by function compatibility with `std;:string`
* Full tests and coverage,
Now we can rewrite `static_string` using `char_string_base`, where `realloc` just throws if the static buffer is exceeded.
Then we can write a new class, a "flexible string" that has both an inline buffer, and an allocator. When the inline buffer overflows, the implementation switches to a dynamic allocation. This can replace some of the code that currently does this manually (e.g. in the parser and RFC range adaptors).
| non_process | string base ideas we can refactor static string call it char string base make it a normal class i e not a template eschew char traits and chart this is for char only and no locales express all operations in terms of a virtual member realloc function by function compatibility with std string full tests and coverage now we can rewrite static string using char string base where realloc just throws if the static buffer is exceeded then we can write a new class a flexible string that has both an inline buffer and an allocator when the inline buffer overflows the implementation switches to a dynamic allocation this can replace some of the code that currently does this manually e g in the parser and rfc range adaptors | 0 |
345,618 | 24,867,565,867 | IssuesEvent | 2022-10-27 13:06:29 | HyperloopUPV-H8/ST-LIB | https://api.github.com/repos/HyperloopUPV-H8/ST-LIB | closed | Problemas y soluciones a la hora de usar la FLASH | documentation | ## Los problemas que nos podemos encontrar usando la FLASH son los siguientes:
- Lo primero, hay que tener en cuenta que la FLASH es bastante delicada, un fallo en la alimentación o que esta sea inestable durante un proceso de borrado, escritura o incluso cuando esta no esta realizando ninguna operación podría causar problemas.
- Hay que tener en cuenta que programar con el debuguer, desde el STM32Cube o desde el STM32Programmer, mientras la FLASH estaba en uso (escibiendo o borrando) puede traer problemas. Es posible que el valor de la FLASH quede modificado y/o alguna flag quede mal puesta y en cuanto intentamos leer o progamar la FLASH el MCU entra en hard fault.
- Cuando se quiere programar desde el STM32Programmer es muy importante deshabilita la opción de _Run after programming_ porque esta opción por algún motivo que no he podido determinar provoca que el MCU entre en hard fault.

## Como solucionar problemas:
- Si lo anterior ocurre (el MCU entra en hard fault cada vez que queremos usar una funcion del módulo) la solución normal es ir al STM32Programmer, seleccionar la ventana _Erasing&Programming_ en el apartado de _Erase flash memory_ seleccionaremos los sectores del 4 al 7 y finalmente le daremos al botón de _Erase selected sectors_.

- Un indicador común de que la FLASH tiene algún problema es observarla desde el STM32Programmer y observar patrones de 0x00000000 y 0xFFFFFFFF alternados cada ciertos bytes.
- Si has realizado y comprobado lo anterior y la placa sigue entrando en hard fault seguramente no sea causado por la memoria FLASH.
## Consejos para evitar problemas:
- Estaría bastante bien que las placas de este año tengan alguna forma para indicar que alguna operación en la FLASH esta en marcha para evitar que se realicen operaciones que puedan corromper la FLASH.
- También es importante que cuando una placa reciba la orden de saltar al bootloader, antes compruebe que ninguna operación se esta realizando en la FLASH y esperar a que acabe si alguna esta en marcha antes de saltar. | 1.0 | Problemas y soluciones a la hora de usar la FLASH - ## Los problemas que nos podemos encontrar usando la FLASH son los siguientes:
- Lo primero, hay que tener en cuenta que la FLASH es bastante delicada, un fallo en la alimentación o que esta sea inestable durante un proceso de borrado, escritura o incluso cuando esta no esta realizando ninguna operación podría causar problemas.
- Hay que tener en cuenta que programar con el debuguer, desde el STM32Cube o desde el STM32Programmer, mientras la FLASH estaba en uso (escibiendo o borrando) puede traer problemas. Es posible que el valor de la FLASH quede modificado y/o alguna flag quede mal puesta y en cuanto intentamos leer o progamar la FLASH el MCU entra en hard fault.
- Cuando se quiere programar desde el STM32Programmer es muy importante deshabilita la opción de _Run after programming_ porque esta opción por algún motivo que no he podido determinar provoca que el MCU entre en hard fault.

## Como solucionar problemas:
- Si lo anterior ocurre (el MCU entra en hard fault cada vez que queremos usar una funcion del módulo) la solución normal es ir al STM32Programmer, seleccionar la ventana _Erasing&Programming_ en el apartado de _Erase flash memory_ seleccionaremos los sectores del 4 al 7 y finalmente le daremos al botón de _Erase selected sectors_.

- Un indicador común de que la FLASH tiene algún problema es observarla desde el STM32Programmer y observar patrones de 0x00000000 y 0xFFFFFFFF alternados cada ciertos bytes.
- Si has realizado y comprobado lo anterior y la placa sigue entrando en hard fault seguramente no sea causado por la memoria FLASH.
## Consejos para evitar problemas:
- Estaría bastante bien que las placas de este año tengan alguna forma para indicar que alguna operación en la FLASH esta en marcha para evitar que se realicen operaciones que puedan corromper la FLASH.
- También es importante que cuando una placa reciba la orden de saltar al bootloader, antes compruebe que ninguna operación se esta realizando en la FLASH y esperar a que acabe si alguna esta en marcha antes de saltar. | non_process | problemas y soluciones a la hora de usar la flash los problemas que nos podemos encontrar usando la flash son los siguientes lo primero hay que tener en cuenta que la flash es bastante delicada un fallo en la alimentación o que esta sea inestable durante un proceso de borrado escritura o incluso cuando esta no esta realizando ninguna operación podría causar problemas hay que tener en cuenta que programar con el debuguer desde el o desde el mientras la flash estaba en uso escibiendo o borrando puede traer problemas es posible que el valor de la flash quede modificado y o alguna flag quede mal puesta y en cuanto intentamos leer o progamar la flash el mcu entra en hard fault cuando se quiere programar desde el es muy importante deshabilita la opción de run after programming porque esta opción por algún motivo que no he podido determinar provoca que el mcu entre en hard fault como solucionar problemas si lo anterior ocurre el mcu entra en hard fault cada vez que queremos usar una funcion del módulo la solución normal es ir al seleccionar la ventana erasing programming en el apartado de erase flash memory seleccionaremos los sectores del al y finalmente le daremos al botón de erase selected sectors un indicador común de que la flash tiene algún problema es observarla desde el y observar patrones de y alternados cada ciertos bytes si has realizado y comprobado lo anterior y la placa sigue entrando en hard fault seguramente no sea causado por la memoria flash consejos para evitar problemas estaría bastante bien que las placas de este año tengan alguna forma para indicar que alguna operación en la flash esta en marcha para evitar que se realicen operaciones que puedan corromper la flash también es importante que cuando una placa reciba la orden de saltar al bootloader antes compruebe que ninguna operación se esta realizando en la flash y esperar a que acabe si alguna esta en marcha antes de saltar | 0 |
3,679 | 6,713,470,781 | IssuesEvent | 2017-10-13 13:35:53 | decidim/decidim | https://api.github.com/repos/decidim/decidim | closed | Bug when inviting moderators | component: processes in-progress type: bug | # This is a Bug Report
#### :tophat: Description
When inviting a moderator (email not associated with any account) on a participatory process the moderator gets stuck at the "Password / Confirm Password" step and is not able to validate its account.
This what the bug looks like.

**updates**
- Can replicate the same beug inviting an admin on a participatory process.
- Inviting an admin (with no account) on the whole org works.
We are using 0.6 stable.
I've been having more issues related to admin invitations and user roles. Will document further on this issue.
| 1.0 | Bug when inviting moderators - # This is a Bug Report
#### :tophat: Description
When inviting a moderator (email not associated with any account) on a participatory process the moderator gets stuck at the "Password / Confirm Password" step and is not able to validate its account.
This what the bug looks like.

**updates**
- Can replicate the same beug inviting an admin on a participatory process.
- Inviting an admin (with no account) on the whole org works.
We are using 0.6 stable.
I've been having more issues related to admin invitations and user roles. Will document further on this issue.
| process | bug when inviting moderators this is a bug report tophat description when inviting a moderator email not associated with any account on a participatory process the moderator gets stuck at the password confirm password step and is not able to validate its account this what the bug looks like updates can replicate the same beug inviting an admin on a participatory process inviting an admin with no account on the whole org works we are using stable i ve been having more issues related to admin invitations and user roles will document further on this issue | 1 |
9,724 | 12,719,803,417 | IssuesEvent | 2020-06-24 09:53:32 | prisma/prisma-client-js | https://api.github.com/repos/prisma/prisma-client-js | opened | Feature Request: Using Prisma Client as query builder only and get the SQL queries like `prisma.user.findOne().getRawQuery()` | kind/feature process/candidate team/product | ## Problem
Adding this feature opens the possibility to use Prisma Client as a query builder only.
Meaning getting the SQL queries without actually hitting the database.
What Lucas wanted is to be able to visualize the query that is going to execute in some sort of playground editor.
## Suggested solution
`prisma.user.findOne().getRawQuery()` would return the SQL query to be executed.
## Alternatives
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional context
See Public Slack Community https://prisma.slack.com/archives/CCWDULGUW/p1591855689031600 | 1.0 | Feature Request: Using Prisma Client as query builder only and get the SQL queries like `prisma.user.findOne().getRawQuery()` - ## Problem
Adding this feature opens the possibility to use Prisma Client as a query builder only.
Meaning getting the SQL queries without actually hitting the database.
What Lucas wanted is to be able to visualize the query that is going to execute in some sort of playground editor.
## Suggested solution
`prisma.user.findOne().getRawQuery()` would return the SQL query to be executed.
## Alternatives
<!-- A clear and concise description of any alternative solutions or features you've considered. -->
## Additional context
See Public Slack Community https://prisma.slack.com/archives/CCWDULGUW/p1591855689031600 | process | feature request using prisma client as query builder only and get the sql queries like prisma user findone getrawquery problem adding this feature opens the possibility to use prisma client as a query builder only meaning getting the sql queries without actually hitting the database what lucas wanted is to be able to visualize the query that is going to execute in some sort of playground editor suggested solution prisma user findone getrawquery would return the sql query to be executed alternatives additional context see public slack community | 1 |
15,918 | 20,120,969,956 | IssuesEvent | 2022-02-08 02:14:30 | streamnative/flink | https://api.github.com/repos/streamnative/flink | closed | [BUG][FLINK-24668] Pulsar consumption data repeated output | compute/data-processing type/bug | `flink-connector-pulsar` uses the `RecordsBySplits` of flink-connector-base to process data, But recordsInCurrentSplit and splitsIterator are not empty when nextSplit `splitSiterator.hasNext () = false`. | 1.0 | [BUG][FLINK-24668] Pulsar consumption data repeated output - `flink-connector-pulsar` uses the `RecordsBySplits` of flink-connector-base to process data, But recordsInCurrentSplit and splitsIterator are not empty when nextSplit `splitSiterator.hasNext () = false`. | process | pulsar consumption data repeated output flink connector pulsar uses the recordsbysplits of flink connector base to process data but recordsincurrentsplit and splitsiterator are not empty when nextsplit splitsiterator hasnext false | 1 |
115,855 | 4,683,113,168 | IssuesEvent | 2016-10-09 16:28:59 | chartjs/Chart.js | https://api.github.com/repos/chartjs/Chart.js | closed | Tooltips & x-axis mode v2.2.0 | Category: Bug Needs Investigation Priority: p1 | Hi,
1 - Tooltips arrow default position is "right" and we can see the arrow blink from the right to its calculated position during fade in and fade out animation.

2 - For bar chart, x-axis mode (thanks for this option @Mesonyx) depends of "stacked" param
if stacked is true, mode = label

if stacked is false, mode = single

So i'm wondering if "x-axis" shouldn't be an option instead of a "mode". It will give the liberty to have x-axis + single or x-axis + label.
3 - For line chart, when pointRadius = 0 (for a clean design or when there is a lot of data points) tooltip is visible only if your cursor is on the 1px corresponding to the X.

It is possible to have something like padding left/right = space between two points / 2 for the "hover zone" of the tooltip.

| 1.0 | Tooltips & x-axis mode v2.2.0 - Hi,
1 - Tooltips arrow default position is "right" and we can see the arrow blink from the right to its calculated position during fade in and fade out animation.

2 - For bar chart, x-axis mode (thanks for this option @Mesonyx) depends of "stacked" param
if stacked is true, mode = label

if stacked is false, mode = single

So i'm wondering if "x-axis" shouldn't be an option instead of a "mode". It will give the liberty to have x-axis + single or x-axis + label.
3 - For line chart, when pointRadius = 0 (for a clean design or when there is a lot of data points) tooltip is visible only if your cursor is on the 1px corresponding to the X.

It is possible to have something like padding left/right = space between two points / 2 for the "hover zone" of the tooltip.

| non_process | tooltips x axis mode hi tooltips arrow default position is right and we can see the arrow blink from the right to its calculated position during fade in and fade out animation for bar chart x axis mode thanks for this option mesonyx depends of stacked param if stacked is true mode label if stacked is false mode single so i m wondering if x axis shouldn t be an option instead of a mode it will give the liberty to have x axis single or x axis label for line chart when pointradius for a clean design or when there is a lot of data points tooltip is visible only if your cursor is on the corresponding to the x it is possible to have something like padding left right space between two points for the hover zone of the tooltip | 0 |
5,030 | 7,851,532,610 | IssuesEvent | 2018-06-20 12:04:23 | allinurl/goaccess | https://api.github.com/repos/allinurl/goaccess | closed | Error in report when import multple logs | log-processing | 
When I import multiple files, the report is wrong.
It look like all of them will be count into the day when the logs was imported.
In the photo, June 18 is the day which I ran command to import | 1.0 | Error in report when import multple logs - 
When I import multiple files, the report is wrong.
It look like all of them will be count into the day when the logs was imported.
In the photo, June 18 is the day which I ran command to import | process | error in report when import multple logs when i import multiple files the report is wrong it look like all of them will be count into the day when the logs was imported in the photo june is the day which i ran command to import | 1 |
553,871 | 16,384,348,280 | IssuesEvent | 2021-05-17 08:31:22 | MathiasReker/Delfinen | https://api.github.com/repos/MathiasReker/Delfinen | opened | DisciplineModel fix | bug medium priority | Remove some of the code, maby it should be used another place
.. make sens when we need the class. | 1.0 | DisciplineModel fix - Remove some of the code, maby it should be used another place
.. make sens when we need the class. | non_process | disciplinemodel fix remove some of the code maby it should be used another place make sens when we need the class | 0 |
9,013 | 12,123,868,663 | IssuesEvent | 2020-04-22 13:23:41 | Torbjornsson/DATX05-Master_Thesis | https://api.github.com/repos/Torbjornsson/DATX05-Master_Thesis | opened | Process/Design | Section: Process | - [ ] Puzzles
- [ ] Tutorials
- [ ] Objects
- [ ] Environment
- [ ] Interactions
- [ ] Sound
- [ ] Test | 1.0 | Process/Design - - [ ] Puzzles
- [ ] Tutorials
- [ ] Objects
- [ ] Environment
- [ ] Interactions
- [ ] Sound
- [ ] Test | process | process design puzzles tutorials objects environment interactions sound test | 1 |
11,456 | 14,274,873,863 | IssuesEvent | 2020-11-22 06:54:26 | SerialLain3170/nagato | https://api.github.com/repos/SerialLain3170/nagato | opened | Flowtron: an Autoregressive Flow-based Generative Network for Text-to-Speech Synthesis | speech processing | # Overview

- Flowtron is autoregressive text-to-speech model that generates a sequence of mel-spectrogram frames.
- In affine coupling layer, the model incorporates speaker embeddings and text embeddings. | 1.0 | Flowtron: an Autoregressive Flow-based Generative Network for Text-to-Speech Synthesis - # Overview

- Flowtron is autoregressive text-to-speech model that generates a sequence of mel-spectrogram frames.
- In affine coupling layer, the model incorporates speaker embeddings and text embeddings. | process | flowtron an autoregressive flow based generative network for text to speech synthesis overview flowtron is autoregressive text to speech model that generates a sequence of mel spectrogram frames in affine coupling layer the model incorporates speaker embeddings and text embeddings | 1 |
22,155 | 30,695,551,392 | IssuesEvent | 2023-07-26 18:17:57 | metabase/metabase | https://api.github.com/repos/metabase/metabase | reopened | Performance regression in nested queries | Type:Bug Priority:P1 Reporting/Dashboards Querying/Nested Queries .Team/QueryProcessor :hammer_and_wrench: .Escalation | ### Describe the bug
Let's compare these 3 scenarios:
1) scenario 1
- create a question: count of people where source = 'Google', save it as a number
- add it 30 times to a dashboard, save it
- refresh

2) scenario 2
- create a question: select all from people, save it as people_1
- create a question out of people_1: select all from people_1 where source = 'Google', save it as people_2
- create a question out of people_2: count, save it as a number as people_3
- add it 30 times to a dashboard, save it
- refresh

3) scenario 3
- let's grab people_3, turn it to a SQL question and save it as people_3_with_sql
- create a new dashboard and add it 30 times, save it, refresh

### To Reproduce
Above
### Expected behavior
If queries are the same, you should get the responses with the same speed
### Logs
_No response_
### Information about your Metabase installation
```JSON
v47, but this is hitting a few 46 customers
```
### Severity
P1
### Additional context
Why does the frontend render them 1 by one? | 1.0 | Performance regression in nested queries - ### Describe the bug
Let's compare these 3 scenarios:
1) scenario 1
- create a question: count of people where source = 'Google', save it as a number
- add it 30 times to a dashboard, save it
- refresh

2) scenario 2
- create a question: select all from people, save it as people_1
- create a question out of people_1: select all from people_1 where source = 'Google', save it as people_2
- create a question out of people_2: count, save it as a number as people_3
- add it 30 times to a dashboard, save it
- refresh

3) scenario 3
- let's grab people_3, turn it to a SQL question and save it as people_3_with_sql
- create a new dashboard and add it 30 times, save it, refresh

### To Reproduce
Above
### Expected behavior
If queries are the same, you should get the responses with the same speed
### Logs
_No response_
### Information about your Metabase installation
```JSON
v47, but this is hitting a few 46 customers
```
### Severity
P1
### Additional context
Why does the frontend render them 1 by one? | process | performance regression in nested queries describe the bug let s compare these scenarios scenario create a question count of people where source google save it as a number add it times to a dashboard save it refresh scenario create a question select all from people save it as people create a question out of people select all from people where source google save it as people create a question out of people count save it as a number as people add it times to a dashboard save it refresh scenario let s grab people turn it to a sql question and save it as people with sql create a new dashboard and add it times save it refresh to reproduce above expected behavior if queries are the same you should get the responses with the same speed logs no response information about your metabase installation json but this is hitting a few customers severity additional context why does the frontend render them by one | 1 |
418,215 | 28,114,021,610 | IssuesEvent | 2023-03-31 09:22:06 | ltzehan/ped | https://api.github.com/repos/ltzehan/ped | opened | Formatting error in description of `add` syntax | type.DocumentationBug severity.VeryLow | From how the other commands are formatted, this looks like it should be formatted as bullet points

<!--session: 1680252418031-b536a5a7-94fd-4e54-924e-bf7b9d014f9f-->
<!--Version: Web v3.4.7--> | 1.0 | Formatting error in description of `add` syntax - From how the other commands are formatted, this looks like it should be formatted as bullet points

<!--session: 1680252418031-b536a5a7-94fd-4e54-924e-bf7b9d014f9f-->
<!--Version: Web v3.4.7--> | non_process | formatting error in description of add syntax from how the other commands are formatted this looks like it should be formatted as bullet points | 0 |
15,658 | 19,846,969,633 | IssuesEvent | 2022-01-21 07:53:45 | ooi-data/CE06ISSM-MFD37-03-DOSTAD000-recovered_host-dosta_abcdjm_ctdbp_dcl_instrument_recovered | https://api.github.com/repos/ooi-data/CE06ISSM-MFD37-03-DOSTAD000-recovered_host-dosta_abcdjm_ctdbp_dcl_instrument_recovered | opened | 🛑 Processing failed: ValueError | process | ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:53:44.982293.
## Details
Flow name: `CE06ISSM-MFD37-03-DOSTAD000-recovered_host-dosta_abcdjm_ctdbp_dcl_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
| 1.0 | 🛑 Processing failed: ValueError - ## Overview
`ValueError` found in `processing_task` task during run ended on 2022-01-21T07:53:44.982293.
## Details
Flow name: `CE06ISSM-MFD37-03-DOSTAD000-recovered_host-dosta_abcdjm_ctdbp_dcl_instrument_recovered`
Task name: `processing_task`
Error type: `ValueError`
Error message: not enough values to unpack (expected 3, got 0)
<details>
<summary>Traceback</summary>
```
Traceback (most recent call last):
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/pipeline.py", line 165, in processing
final_path = finalize_data_stream(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 84, in finalize_data_stream
append_to_zarr(mod_ds, final_store, enc, logger=logger)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/__init__.py", line 357, in append_to_zarr
_append_zarr(store, mod_ds)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/ooi_harvester/processor/utils.py", line 187, in _append_zarr
existing_arr.append(var_data.values)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 519, in values
return _as_array_or_item(self._data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/variable.py", line 259, in _as_array_or_item
data = np.asarray(data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 1541, in __array__
x = self.compute()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 288, in compute
(result,) = compute(self, traverse=False, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/base.py", line 571, in compute
results = schedule(dsk, keys, **kwargs)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/threaded.py", line 79, in get
results = get_async(
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 507, in get_async
raise_exception(exc, tb)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 315, in reraise
raise exc
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/local.py", line 220, in execute_task
result = _execute_task(task, data)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/core.py", line 119, in _execute_task
return func(*(_execute_task(a, cache) for a in args))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/dask/array/core.py", line 116, in getter
c = np.asarray(c)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 357, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 551, in __array__
self._ensure_cached()
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 548, in _ensure_cached
self.array = NumpyIndexingAdapter(np.asarray(self.array))
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 521, in __array__
return np.asarray(self.array, dtype=dtype)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/core/indexing.py", line 422, in __array__
return np.asarray(array[self.key], dtype=None)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/xarray/backends/zarr.py", line 73, in __getitem__
return array[key.tuple]
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 673, in __getitem__
return self.get_basic_selection(selection, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 798, in get_basic_selection
return self._get_basic_selection_nd(selection=selection, out=out,
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 841, in _get_basic_selection_nd
return self._get_selection(indexer=indexer, out=out, fields=fields)
File "/srv/conda/envs/notebook/lib/python3.9/site-packages/zarr/core.py", line 1135, in _get_selection
lchunk_coords, lchunk_selection, lout_selection = zip(*indexer)
ValueError: not enough values to unpack (expected 3, got 0)
```
</details>
| process | 🛑 processing failed valueerror overview valueerror found in processing task task during run ended on details flow name recovered host dosta abcdjm ctdbp dcl instrument recovered task name processing task error type valueerror error message not enough values to unpack expected got traceback traceback most recent call last file srv conda envs notebook lib site packages ooi harvester processor pipeline py line in processing final path finalize data stream file srv conda envs notebook lib site packages ooi harvester processor init py line in finalize data stream append to zarr mod ds final store enc logger logger file srv conda envs notebook lib site packages ooi harvester processor init py line in append to zarr append zarr store mod ds file srv conda envs notebook lib site packages ooi harvester processor utils py line in append zarr existing arr append var data values file srv conda envs notebook lib site packages xarray core variable py line in values return as array or item self data file srv conda envs notebook lib site packages xarray core variable py line in as array or item data np asarray data file srv conda envs notebook lib site packages dask array core py line in array x self compute file srv conda envs notebook lib site packages dask base py line in compute result compute self traverse false kwargs file srv conda envs notebook lib site packages dask base py line in compute results schedule dsk keys kwargs file srv conda envs notebook lib site packages dask threaded py line in get results get async file srv conda envs notebook lib site packages dask local py line in get async raise exception exc tb file srv conda envs notebook lib site packages dask local py line in reraise raise exc file srv conda envs notebook lib site packages dask local py line in execute task result execute task task data file srv conda envs notebook lib site packages dask core py line in execute task return func execute task a cache for a in args file srv conda envs notebook lib site packages dask array core py line in getter c np asarray c file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array self ensure cached file srv conda envs notebook lib site packages xarray core indexing py line in ensure cached self array numpyindexingadapter np asarray self array file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray self array dtype dtype file srv conda envs notebook lib site packages xarray core indexing py line in array return np asarray array dtype none file srv conda envs notebook lib site packages xarray backends zarr py line in getitem return array file srv conda envs notebook lib site packages zarr core py line in getitem return self get basic selection selection fields fields file srv conda envs notebook lib site packages zarr core py line in get basic selection return self get basic selection nd selection selection out out file srv conda envs notebook lib site packages zarr core py line in get basic selection nd return self get selection indexer indexer out out fields fields file srv conda envs notebook lib site packages zarr core py line in get selection lchunk coords lchunk selection lout selection zip indexer valueerror not enough values to unpack expected got | 1 |
21,399 | 29,202,233,769 | IssuesEvent | 2023-05-21 00:38:03 | devssa/onde-codar-em-salvador | https://api.github.com/repos/devssa/onde-codar-em-salvador | closed | [Santa Lúcia, Belo Horizonte] Flutter Developer na Coodesh | SALVADOR MOBILE FIREBASE REQUISITOS IOS ANDROID PROCESSOS GITHUB FLUTTER AMP UMA DART MANUTENÇÃO ACESSIBILIDADE ALOCADO Stale | ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/desenvolvedor-flutter-165642989?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>PS GESTAO DE CARREIRA</strong> está em busca de <strong><ins>Flutter Developer</ins></strong> para compor seu time!</p>
<p>Se você busca de ingressar em uma empresa que possui uma proposta inovadora no mercado do agronegócio, sendo a maior referência da tecnologia que envolve esse mercado , essa vaga é a que você está procurando! Estamos contratando uma pessoa desenvolvedora Flutter para atuar em um time que está em constante crescimento e aprimoramento de tecnologia e conhecimento para contribui nesse cenário.</p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Desenvolver aplicações mobile para Android e iOS usando as tecnologias Flutter e Dart. Os aplicativos irão consumir API online para troca de dados e receber notificações Push;</li>
<li>Realizar manutenção em aplicativos já existentes desenvolvendo novas funcionalidades;</li>
<li>Desenvolver novos aplicativos e publicar nas lojas AppStore e PlayStore.</li>
</ul>
<p>Segunda à sexta, 08:00 às 18:00 - 44 Horas semanais.</p>
<p></p>
## PS GESTAO DE CARREIRA:
<p>Oferecemos serviços de recrutamento & seleção/hunting, assesment organizacional e desenvolvimento de carreira por meio de serviços customizados, soluções eficientes, práticas e objetivas. Conectamos candidatos à empresas de diversos ramos, buscando atender a todos da melhor maneira! Tudo isso foi desenvolvido para que o empresário ou gestor de RH possa dedicar tempo aos processos internos enquanto nós, uma empresa especializada, cuidamos para encontrar as melhores pessoas para sua empresa.</p>
<p></p><a href='https://coodesh.com/empresas/ps-gestao-de-carreira'>Veja mais no site</a>
## Habilidades:
- Flutter
- Dart
- Acessibilidade
## Local:
Santa Lúcia, Belo Horizonte
## Requisitos:
- Experiência com desenvolvimento de aplicativos Mobile em Flutter e Dart;
- Experiência no desenvolvimento de apps nestas tecnologias bem como experiência em publicas nas lojas;
- Experiência com Firebase e OneSignal para notificações Push será considerado um diferencial.
## Benefícios:
- Vale Transporte/Auxílio Combustível;
- Vale Alimentação;
- Plano de Saúde UNIMED Estadual;
- Plano Odontológico UNIMED;
- Seguro de Vida.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Flutter Developer na PS GESTAO DE CARREIRA](https://coodesh.com/vagas/desenvolvedor-flutter-165642989?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Regime
CLT
#### Categoria
Mobile | 1.0 | [Santa Lúcia, Belo Horizonte] Flutter Developer na Coodesh - ## Descrição da vaga:
Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios.
Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/desenvolvedor-flutter-165642989?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋
<p>A <strong>PS GESTAO DE CARREIRA</strong> está em busca de <strong><ins>Flutter Developer</ins></strong> para compor seu time!</p>
<p>Se você busca de ingressar em uma empresa que possui uma proposta inovadora no mercado do agronegócio, sendo a maior referência da tecnologia que envolve esse mercado , essa vaga é a que você está procurando! Estamos contratando uma pessoa desenvolvedora Flutter para atuar em um time que está em constante crescimento e aprimoramento de tecnologia e conhecimento para contribui nesse cenário.</p>
<p><strong>Responsabilidades:</strong></p>
<ul>
<li>Desenvolver aplicações mobile para Android e iOS usando as tecnologias Flutter e Dart. Os aplicativos irão consumir API online para troca de dados e receber notificações Push;</li>
<li>Realizar manutenção em aplicativos já existentes desenvolvendo novas funcionalidades;</li>
<li>Desenvolver novos aplicativos e publicar nas lojas AppStore e PlayStore.</li>
</ul>
<p>Segunda à sexta, 08:00 às 18:00 - 44 Horas semanais.</p>
<p></p>
## PS GESTAO DE CARREIRA:
<p>Oferecemos serviços de recrutamento & seleção/hunting, assesment organizacional e desenvolvimento de carreira por meio de serviços customizados, soluções eficientes, práticas e objetivas. Conectamos candidatos à empresas de diversos ramos, buscando atender a todos da melhor maneira! Tudo isso foi desenvolvido para que o empresário ou gestor de RH possa dedicar tempo aos processos internos enquanto nós, uma empresa especializada, cuidamos para encontrar as melhores pessoas para sua empresa.</p>
<p></p><a href='https://coodesh.com/empresas/ps-gestao-de-carreira'>Veja mais no site</a>
## Habilidades:
- Flutter
- Dart
- Acessibilidade
## Local:
Santa Lúcia, Belo Horizonte
## Requisitos:
- Experiência com desenvolvimento de aplicativos Mobile em Flutter e Dart;
- Experiência no desenvolvimento de apps nestas tecnologias bem como experiência em publicas nas lojas;
- Experiência com Firebase e OneSignal para notificações Push será considerado um diferencial.
## Benefícios:
- Vale Transporte/Auxílio Combustível;
- Vale Alimentação;
- Plano de Saúde UNIMED Estadual;
- Plano Odontológico UNIMED;
- Seguro de Vida.
## Como se candidatar:
Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Flutter Developer na PS GESTAO DE CARREIRA](https://coodesh.com/vagas/desenvolvedor-flutter-165642989?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open)
Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação.
## Labels
#### Alocação
Alocado
#### Regime
CLT
#### Categoria
Mobile | process | flutter developer na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a ps gestao de carreira está em busca de flutter developer para compor seu time se você busca de ingressar em uma empresa que possui uma proposta inovadora no mercado do agronegócio sendo a maior referência da tecnologia que envolve esse mercado essa vaga é a que você está procurando estamos contratando uma pessoa desenvolvedora flutter para atuar em um time que está em constante crescimento e aprimoramento de tecnologia e conhecimento para contribui nesse cenário responsabilidades desenvolver aplicações mobile para android e ios usando as tecnologias flutter e dart os aplicativos irão consumir api online para troca de dados e receber notificações push realizar manutenção em aplicativos já existentes desenvolvendo novas funcionalidades desenvolver novos aplicativos e publicar nas lojas appstore e playstore segunda à sexta às horas semanais ps gestao de carreira oferecemos serviços de recrutamento amp seleção hunting assesment organizacional e desenvolvimento de carreira por meio de serviços customizados soluções eficientes práticas e objetivas conectamos candidatos à empresas de diversos ramos buscando atender a todos da melhor maneira tudo isso foi desenvolvido para que o empresário ou gestor de rh possa dedicar tempo aos processos internos enquanto nós uma empresa especializada cuidamos para encontrar as melhores pessoas para sua empresa habilidades flutter dart acessibilidade local santa lúcia belo horizonte requisitos experiência com desenvolvimento de aplicativos mobile em flutter e dart experiência no desenvolvimento de apps nestas tecnologias bem como experiência em publicas nas lojas experiência com firebase e onesignal para notificações push será considerado um diferencial benefícios vale transporte auxílio combustível vale alimentação plano de saúde unimed estadual plano odontológico unimed seguro de vida como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado regime clt categoria mobile | 1 |
5,638 | 8,494,548,252 | IssuesEvent | 2018-10-28 22:31:35 | P0cL4bs/WiFi-Pumpkin | https://api.github.com/repos/P0cL4bs/WiFi-Pumpkin | closed | Device not showing in Pumpkin device list | enhancement in process priority solved | #### Please tell us details about your environment.
* Card wireless adapters name (please check if support AP/mode):
* Version used tool:
* Virtual Machine (yes or no and which):
* Operating System and version:
## What's the problem (or question)?
I've got Kali 2017.3 on a Raspberry Pi 3
I've ran the following:
apt-get install libssl1.0
apt-get install libjpeg-dev
pip2 install mitmproxy==0.18.2
pip install --upgrade service_identity
I have the following issue:
The pumpkin application runs, but when I connect my device to the pumpkin's SSID, the device does not show up in the connected devices list (except the very first time the device connects) even though the device says it's connected and the pumpkin activity section shows activity when you try to go to a site on a browser. Also, the device will not connect to the Internet even though the PI can connect via its Ethernet connection to my router. Here's the output from two attempts of starting, stopping, and connecting to the pumpkinAP.
root@kali:~# wifi-pumpkin
Loading GUI...
WiFi-Pumpkin Running!
[] Loading debugging mode
[] Current Session::ID [ODc2OTI=]
[] Configuring hostapd...
[] Enable forwarding in iptables...
[] Configuring dhcpd...
[] Sharing Internet Connections with NAT...
[New Thread 950 (hostapd)]
[New Thread 952 (Dns2Proxy)]
SSLstrip v0.9 + POC by Leonardo Nve
[*] TCPProxy running on port 80/8080:
TCPProxy::ftp status:On
TCPProxy::hexdump status:On
TCPProxy::kerberos status:On
TCPProxy::summary status:On
TCPProxy::imageCap status:On
TCPProxy::httpCap status:On
TCPProxy::emails status:On
AP::[pumpkinAP] Running...
AP::BSSID::[ac:e6:45:03:26:4d] CH 11
WARNING:root:POST Data (update.googleapis.com):
Thread::[hostapd] successfully stopped.
Thread::[Dns2Proxy] successfully stopped.
Thread::[sslstrip2] successfully stopped.
Thread::[TCPProxy] successfully stopped.
[] Loading debugging mode
[] Current Session::ID [Njg0OTM=]
[] Configuring hostapd...
[] Enable forwarding in iptables...
[] Configuring dhcpd...
[] Sharing Internet Connections with NAT...
[New Thread 1033 (hostapd)]
[New Thread 1035 (Dns2Proxy)]
SSLstrip v0.9 + POC by Leonardo Nve
[*] TCPProxy running on port 80/8080:
TCPProxy::ftp status:On
TCPProxy::hexdump status:On
TCPProxy::kerberos status:On
TCPProxy::summary status:On
TCPProxy::imageCap status:On
TCPProxy::httpCap status:On
TCPProxy::emails status:On
AP::[pumpkinAP] Running...
AP::BSSID::[ac:e6:45:03:26:4d] CH 11
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 54, in addNextWidget
for client in agent[key].keys():
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Thread::[hostapd] successfully stopped.
Thread::[Dns2Proxy] successfully stopped.
Thread::[sslstrip2] successfully stopped.
Thread::[TCPProxy] successfully stopped.
| 1.0 | Device not showing in Pumpkin device list - #### Please tell us details about your environment.
* Card wireless adapters name (please check if support AP/mode):
* Version used tool:
* Virtual Machine (yes or no and which):
* Operating System and version:
## What's the problem (or question)?
I've got Kali 2017.3 on a Raspberry Pi 3
I've ran the following:
apt-get install libssl1.0
apt-get install libjpeg-dev
pip2 install mitmproxy==0.18.2
pip install --upgrade service_identity
I have the following issue:
The pumpkin application runs, but when I connect my device to the pumpkin's SSID, the device does not show up in the connected devices list (except the very first time the device connects) even though the device says it's connected and the pumpkin activity section shows activity when you try to go to a site on a browser. Also, the device will not connect to the Internet even though the PI can connect via its Ethernet connection to my router. Here's the output from two attempts of starting, stopping, and connecting to the pumpkinAP.
root@kali:~# wifi-pumpkin
Loading GUI...
WiFi-Pumpkin Running!
[] Loading debugging mode
[] Current Session::ID [ODc2OTI=]
[] Configuring hostapd...
[] Enable forwarding in iptables...
[] Configuring dhcpd...
[] Sharing Internet Connections with NAT...
[New Thread 950 (hostapd)]
[New Thread 952 (Dns2Proxy)]
SSLstrip v0.9 + POC by Leonardo Nve
[*] TCPProxy running on port 80/8080:
TCPProxy::ftp status:On
TCPProxy::hexdump status:On
TCPProxy::kerberos status:On
TCPProxy::summary status:On
TCPProxy::imageCap status:On
TCPProxy::httpCap status:On
TCPProxy::emails status:On
AP::[pumpkinAP] Running...
AP::BSSID::[ac:e6:45:03:26:4d] CH 11
WARNING:root:POST Data (update.googleapis.com):
Thread::[hostapd] successfully stopped.
Thread::[Dns2Proxy] successfully stopped.
Thread::[sslstrip2] successfully stopped.
Thread::[TCPProxy] successfully stopped.
[] Loading debugging mode
[] Current Session::ID [Njg0OTM=]
[] Configuring hostapd...
[] Enable forwarding in iptables...
[] Configuring dhcpd...
[] Sharing Internet Connections with NAT...
[New Thread 1033 (hostapd)]
[New Thread 1035 (Dns2Proxy)]
SSLstrip v0.9 + POC by Leonardo Nve
[*] TCPProxy running on port 80/8080:
TCPProxy::ftp status:On
TCPProxy::hexdump status:On
TCPProxy::kerberos status:On
TCPProxy::summary status:On
TCPProxy::imageCap status:On
TCPProxy::httpCap status:On
TCPProxy::emails status:On
AP::[pumpkinAP] Running...
AP::BSSID::[ac:e6:45:03:26:4d] CH 11
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 54, in addNextWidget
for client in agent[key].keys():
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Traceback (most recent call last):
File "/usr/share/WiFi-Pumpkin/core/main.py", line 1064, in get_Hostapd_Response
self.TabInfoAP.delete_item(data)
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 71, in delete_item
self.addNextWidget(self.APclients[agent])
File "/usr/share/WiFi-Pumpkin/core/widgets/customiseds.py", line 51, in addNextWidget
self.items_widgets[agent.keys()[0]] = {}
AttributeError: 'str' object has no attribute 'keys'
Thread::[hostapd] successfully stopped.
Thread::[Dns2Proxy] successfully stopped.
Thread::[sslstrip2] successfully stopped.
Thread::[TCPProxy] successfully stopped.
| process | device not showing in pumpkin device list please tell us details about your environment card wireless adapters name please check if support ap mode version used tool virtual machine yes or no and which operating system and version what s the problem or question i ve got kali on a raspberry pi i ve ran the following apt get install apt get install libjpeg dev install mitmproxy pip install upgrade service identity i have the following issue the pumpkin application runs but when i connect my device to the pumpkin s ssid the device does not show up in the connected devices list except the very first time the device connects even though the device says it s connected and the pumpkin activity section shows activity when you try to go to a site on a browser also the device will not connect to the internet even though the pi can connect via its ethernet connection to my router here s the output from two attempts of starting stopping and connecting to the pumpkinap root kali wifi pumpkin loading gui wifi pumpkin running loading debugging mode current session id configuring hostapd enable forwarding in iptables configuring dhcpd sharing internet connections with nat sslstrip poc by leonardo nve tcpproxy running on port tcpproxy ftp status on tcpproxy hexdump status on tcpproxy kerberos status on tcpproxy summary status on tcpproxy imagecap status on tcpproxy httpcap status on tcpproxy emails status on ap running ap bssid ch warning root post data update googleapis com thread successfully stopped thread successfully stopped thread successfully stopped thread successfully stopped loading debugging mode current session id configuring hostapd enable forwarding in iptables configuring dhcpd sharing internet connections with nat sslstrip poc by leonardo nve tcpproxy running on port tcpproxy ftp status on tcpproxy hexdump status on tcpproxy kerberos status on tcpproxy summary status on tcpproxy imagecap status on tcpproxy httpcap status on tcpproxy emails status on ap running ap bssid ch traceback most recent call last file usr share wifi pumpkin core main py line in get hostapd response self tabinfoap delete item data file usr share wifi pumpkin core widgets customiseds py line in delete item self addnextwidget self apclients file usr share wifi pumpkin core widgets customiseds py line in addnextwidget for client in agent keys attributeerror str object has no attribute keys traceback most recent call last file usr share wifi pumpkin core main py line in get hostapd response self tabinfoap delete item data file usr share wifi pumpkin core widgets customiseds py line in delete item self addnextwidget self apclients file usr share wifi pumpkin core widgets customiseds py line in addnextwidget self items widgets attributeerror str object has no attribute keys traceback most recent call last file usr share wifi pumpkin core main py line in get hostapd response self tabinfoap delete item data file usr share wifi pumpkin core widgets customiseds py line in delete item self addnextwidget self apclients file usr share wifi pumpkin core widgets customiseds py line in addnextwidget self items widgets attributeerror str object has no attribute keys traceback most recent call last file usr share wifi pumpkin core main py line in get hostapd response self tabinfoap delete item data file usr share wifi pumpkin core widgets customiseds py line in delete item self addnextwidget self apclients file usr share wifi pumpkin core widgets customiseds py line in addnextwidget self items widgets attributeerror str object has no attribute keys traceback most recent call last file usr share wifi pumpkin core main py line in get hostapd response self tabinfoap delete item data file usr share wifi pumpkin core widgets customiseds py line in delete item self addnextwidget self apclients file usr share wifi pumpkin core widgets customiseds py line in addnextwidget self items widgets attributeerror str object has no attribute keys traceback most recent call last file usr share wifi pumpkin core main py line in get hostapd response self tabinfoap delete item data file usr share wifi pumpkin core widgets customiseds py line in delete item self addnextwidget self apclients file usr share wifi pumpkin core widgets customiseds py line in addnextwidget self items widgets attributeerror str object has no attribute keys thread successfully stopped thread successfully stopped thread successfully stopped thread successfully stopped | 1 |
727,753 | 25,045,830,807 | IssuesEvent | 2022-11-05 08:15:51 | googleapis/nodejs-ai-platform | https://api.github.com/repos/googleapis/nodejs-ai-platform | opened | AI platform feature values apis: "before all" hook: should create the perm_users entity type for "should import feature values" failed | type: bug priority: p1 flakybot: issue | This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 9345b15fcc22f60f00cd50442403fc1af4fabe71
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/fe76c9bc-cc8e-4a63-a10c-5ff3da0449f2), [Sponge](http://sponge2/fe76c9bc-cc8e-4a63-a10c-5ff3da0449f2)
status: failed
<details><summary>Test output</summary><br><pre>Timeout of 1200000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/workspace/samples/test/feature-values-samples.test.js)
Error: Timeout of 1200000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/workspace/samples/test/feature-values-samples.test.js)
at listOnTimeout (internal/timers.js:554:17)
at processTimers (internal/timers.js:497:7)</pre></details> | 1.0 | AI platform feature values apis: "before all" hook: should create the perm_users entity type for "should import feature values" failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/main/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: 9345b15fcc22f60f00cd50442403fc1af4fabe71
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/fe76c9bc-cc8e-4a63-a10c-5ff3da0449f2), [Sponge](http://sponge2/fe76c9bc-cc8e-4a63-a10c-5ff3da0449f2)
status: failed
<details><summary>Test output</summary><br><pre>Timeout of 1200000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/workspace/samples/test/feature-values-samples.test.js)
Error: Timeout of 1200000ms exceeded. For async tests and hooks, ensure "done()" is called; if returning a Promise, ensure it resolves. (/workspace/samples/test/feature-values-samples.test.js)
at listOnTimeout (internal/timers.js:554:17)
at processTimers (internal/timers.js:497:7)</pre></details> | non_process | ai platform feature values apis before all hook should create the perm users entity type for should import feature values failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves workspace samples test feature values samples test js error timeout of exceeded for async tests and hooks ensure done is called if returning a promise ensure it resolves workspace samples test feature values samples test js at listontimeout internal timers js at processtimers internal timers js | 0 |
15,128 | 18,872,366,761 | IssuesEvent | 2021-11-13 12:14:13 | RobertCraigie/prisma-client-py | https://api.github.com/repos/RobertCraigie/prisma-client-py | closed | Type validator is broken on python3.9.0-1 | bug/2-confirmed kind/bug process/candidate | <!--
Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output.
See https://prisma-client-py.readthedocs.io/en/latest/logging/ for how to enable additional logging output.
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
`tests/test_validator.py::test_valid` and `tests/test_validator.py::test_recursive` fail on `python3.9.1`
`test_valid` output:
```
―――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――― test_valid ――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――
def test_valid() -> None:
"""Basic usage with correct data"""
> validated = validate(types.IntFilter, {'equals': 1})
tests/test_validator.py:14:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
/private/tmp/tox/prisma-client-py/py39/lib/python3.9/site-packages/prisma/validator.py:81: in validate
instance = model.parse_obj(data)
pydantic/main.py:578: in pydantic.main.BaseModel.parse_obj
???
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E pydantic.error_wrappers.ValidationError: 7 validation errors for IntFilter
E not_in
E field required (type=value_error.missing)
E lt
E field required (type=value_error.missing)
E lte
E field required (type=value_error.missing)
E gt
E field required (type=value_error.missing)
E gte
E field required (type=value_error.missing)
E in
E field required (type=value_error.missing)
E not
E field required (type=value_error.missing)
pydantic/main.py:406: ValidationError
```
## How to reproduce
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
`tox -e py39 -- tests/test_validator.py`
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Tests should pass
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]--> Mac OS
- Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]--> SQLite
- Python version: <!--[Run `python -V` to see your Python version]--> Python 3.9.1
| 1.0 | Type validator is broken on python3.9.0-1 - <!--
Thanks for helping us improve Prisma Client Python! 🙏 Please follow the sections in the template and provide as much information as possible about your problem, e.g. by enabling additional logging output.
See https://prisma-client-py.readthedocs.io/en/latest/logging/ for how to enable additional logging output.
-->
## Bug description
<!-- A clear and concise description of what the bug is. -->
`tests/test_validator.py::test_valid` and `tests/test_validator.py::test_recursive` fail on `python3.9.1`
`test_valid` output:
```
―――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――― test_valid ――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――
def test_valid() -> None:
"""Basic usage with correct data"""
> validated = validate(types.IntFilter, {'equals': 1})
tests/test_validator.py:14:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
/private/tmp/tox/prisma-client-py/py39/lib/python3.9/site-packages/prisma/validator.py:81: in validate
instance = model.parse_obj(data)
pydantic/main.py:578: in pydantic.main.BaseModel.parse_obj
???
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
> ???
E pydantic.error_wrappers.ValidationError: 7 validation errors for IntFilter
E not_in
E field required (type=value_error.missing)
E lt
E field required (type=value_error.missing)
E lte
E field required (type=value_error.missing)
E gt
E field required (type=value_error.missing)
E gte
E field required (type=value_error.missing)
E in
E field required (type=value_error.missing)
E not
E field required (type=value_error.missing)
pydantic/main.py:406: ValidationError
```
## How to reproduce
<!--
Steps to reproduce the behavior:
1. Go to '...'
2. Change '....'
3. Run '....'
4. See error
-->
`tox -e py39 -- tests/test_validator.py`
## Expected behavior
<!-- A clear and concise description of what you expected to happen. -->
Tests should pass
## Environment & setup
<!-- In which environment does the problem occur -->
- OS: <!--[e.g. Mac OS, Windows, Debian, CentOS, ...]--> Mac OS
- Database: <!--[PostgreSQL, MySQL, MariaDB or SQLite]--> SQLite
- Python version: <!--[Run `python -V` to see your Python version]--> Python 3.9.1
| process | type validator is broken on thanks for helping us improve prisma client python 🙏 please follow the sections in the template and provide as much information as possible about your problem e g by enabling additional logging output see for how to enable additional logging output bug description tests test validator py test valid and tests test validator py test recursive fail on test valid output ―――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――― test valid ―――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――――― def test valid none basic usage with correct data validated validate types intfilter equals tests test validator py private tmp tox prisma client py lib site packages prisma validator py in validate instance model parse obj data pydantic main py in pydantic main basemodel parse obj e pydantic error wrappers validationerror validation errors for intfilter e not in e field required type value error missing e lt e field required type value error missing e lte e field required type value error missing e gt e field required type value error missing e gte e field required type value error missing e in e field required type value error missing e not e field required type value error missing pydantic main py validationerror how to reproduce steps to reproduce the behavior go to change run see error tox e tests test validator py expected behavior tests should pass environment setup os mac os database sqlite python version python | 1 |
345,423 | 30,810,449,434 | IssuesEvent | 2023-08-01 10:01:23 | wazuh/wazuh | https://api.github.com/repos/wazuh/wazuh | closed | Release 4.5.0 - Alpha 1 - Integration tests | type/test level/task | # Description
| Wazuh QA: Branch | Wazuh QA: Commit | Wazuh: Tag | Wazuh: Commit |
|:--:|:--:|:--:|:--:|
| `v4.5.0-alpha1`| https://github.com/wazuh/wazuh-qa/commit/ede149c9ab767d23d3d4f9ca461186dc660e45a6 | `v4.5.0-alpha1` |https://github.com/wazuh/wazuh/commit/39df1e74bfd035fea8948191ae195fc17eb7e2f3 |
We are going to check that the integration tests of the `4.5` branch of `wazuh-qa` work correctly using the [v4.5.0-alpha1](https://github.com/wazuh/wazuh/tree/v4.5.0-alpha1) version of `wazuh`.
The tests will be performed in Jenkins using `centOS` as the manager OS. As for the agents, `Linux`, `Windows` , `Solaris`, `macOS` will be used as required.
## Tests Integration - Status
#### Main RC issue
- https://github.com/wazuh/wazuh/issues/18058
#### References
|Color|Status |
|:--:|:--|
|🟢|All tests passed successfully|
|🟡|All tests passed but there are some warnings|
|🔴|Some tests have failures or errors|
|:large_blue_circle:|Test execution in progress|
|:black_circle:|To Do|
|🟠|Jenkins provision fails|
|:purple_circle:| All skipped |
## Test Integration - Results
<table>
<thead>
<tr>
<th style="width: 175px;">Name</th>
<th style="width: 499px;" colspan="6">Jenkins</th>
</tr>
</thead>
<tbody>
<tr>
<td style="width: 175px;">OS</td>
<td style="width: 208px;" colspan="2">Linux</td>
<td style="width: 79px;">Windows</td>
<td style="width: 97px;">Solaris</td>
<td style="width: 97px;" colspan="2">macOS</td>
</tr>
<tr>
<td style="width: 175px;">Target</td>
<td style="width: 103px;">Manager</td>
<td style="width: 99px;">Agent</td>
<td style="width: 79px;">Agent</td>
<td style="width: 97px;">Agent</td>
<td style="width: 97px;" colspan="2">Agent</td>
</tr>
<tr>
<td style="width: 175px;"><strong>active_response</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>agentd</strong></td>
<td style="width: 103px;">NA</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>analysisd</strong></td>
<td style="width: 103px;">🟢 </td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>api</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>authd</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;"> NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>enrollment<br/></strong></td>
<td style="width: 103px;">NA</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>fim</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="width: 97px;">🟢</td>
<td style="width: 97px;" colspan="2">🟢</td>
</tr>
<tr>
<td style="width: 175px;"><strong>gcloud<br/></strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>github<br /></strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>logcollector</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="width: 97px;">🟢</td>
<td style="width: 97px;" colspan="2">🔴</td>
</tr>
<tr>
<td style="width: 175px;"><strong>logtest<br/></strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>office365<br /></strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>remoted</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>rids</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>rootcheck</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>vulnerability_detector</strong></td>
<td style="width: 103px;">🔴</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>wazuh_db</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>syscollector</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🔴</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">🟢</td>
</tr>
</tbody>
</table>
## Evidence
**IT Launcher build**: https://ci.wazuh.info/job/Test_integration_launcher/180/
## Conclusion :red_circle:
Some tests failed, and a known issue was reported
- https://github.com/wazuh/wazuh-qa/issues/4230
leading to the creation of some new issues.
- https://github.com/wazuh/wazuh-jenkins/issues/5468
- https://github.com/wazuh/wazuh-qa/issues/4370
## Auditors validation
The definition of done for this one is the validation of the conclusions and the test results from all auditors.
All checks from below must be accepted in order to close this issue.
- [ ] @davidjiglesias
- [ ] @jnasselle | 1.0 | Release 4.5.0 - Alpha 1 - Integration tests - # Description
| Wazuh QA: Branch | Wazuh QA: Commit | Wazuh: Tag | Wazuh: Commit |
|:--:|:--:|:--:|:--:|
| `v4.5.0-alpha1`| https://github.com/wazuh/wazuh-qa/commit/ede149c9ab767d23d3d4f9ca461186dc660e45a6 | `v4.5.0-alpha1` |https://github.com/wazuh/wazuh/commit/39df1e74bfd035fea8948191ae195fc17eb7e2f3 |
We are going to check that the integration tests of the `4.5` branch of `wazuh-qa` work correctly using the [v4.5.0-alpha1](https://github.com/wazuh/wazuh/tree/v4.5.0-alpha1) version of `wazuh`.
The tests will be performed in Jenkins using `centOS` as the manager OS. As for the agents, `Linux`, `Windows` , `Solaris`, `macOS` will be used as required.
## Tests Integration - Status
#### Main RC issue
- https://github.com/wazuh/wazuh/issues/18058
#### References
|Color|Status |
|:--:|:--|
|🟢|All tests passed successfully|
|🟡|All tests passed but there are some warnings|
|🔴|Some tests have failures or errors|
|:large_blue_circle:|Test execution in progress|
|:black_circle:|To Do|
|🟠|Jenkins provision fails|
|:purple_circle:| All skipped |
## Test Integration - Results
<table>
<thead>
<tr>
<th style="width: 175px;">Name</th>
<th style="width: 499px;" colspan="6">Jenkins</th>
</tr>
</thead>
<tbody>
<tr>
<td style="width: 175px;">OS</td>
<td style="width: 208px;" colspan="2">Linux</td>
<td style="width: 79px;">Windows</td>
<td style="width: 97px;">Solaris</td>
<td style="width: 97px;" colspan="2">macOS</td>
</tr>
<tr>
<td style="width: 175px;">Target</td>
<td style="width: 103px;">Manager</td>
<td style="width: 99px;">Agent</td>
<td style="width: 79px;">Agent</td>
<td style="width: 97px;">Agent</td>
<td style="width: 97px;" colspan="2">Agent</td>
</tr>
<tr>
<td style="width: 175px;"><strong>active_response</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>agentd</strong></td>
<td style="width: 103px;">NA</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>analysisd</strong></td>
<td style="width: 103px;">🟢 </td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>api</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>authd</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;"> NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>enrollment<br/></strong></td>
<td style="width: 103px;">NA</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>fim</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="width: 97px;">🟢</td>
<td style="width: 97px;" colspan="2">🟢</td>
</tr>
<tr>
<td style="width: 175px;"><strong>gcloud<br/></strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>github<br /></strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>logcollector</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🟢</td>
<td style="width: 97px;">🟢</td>
<td style="width: 97px;" colspan="2">🔴</td>
</tr>
<tr>
<td style="width: 175px;"><strong>logtest<br/></strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>office365<br /></strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>remoted</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>rids</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>rootcheck</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>vulnerability_detector</strong></td>
<td style="width: 103px;">🔴</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>wazuh_db</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">NA</td>
<td style="width: 79px;">NA</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">NA</td>
</tr>
<tr>
<td style="width: 175px;"><strong>syscollector</strong></td>
<td style="width: 103px;">🟢</td>
<td style="width: 99px;">🟢</td>
<td style="width: 79px;">🔴</td>
<td style="text-align: center; width: 97px;">NA</td>
<td style="text-align: center; width: 97px;" colspan="2">🟢</td>
</tr>
</tbody>
</table>
## Evidence
**IT Launcher build**: https://ci.wazuh.info/job/Test_integration_launcher/180/
## Conclusion :red_circle:
Some tests failed, and a known issue was reported
- https://github.com/wazuh/wazuh-qa/issues/4230
leading to the creation of some new issues.
- https://github.com/wazuh/wazuh-jenkins/issues/5468
- https://github.com/wazuh/wazuh-qa/issues/4370
## Auditors validation
The definition of done for this one is the validation of the conclusions and the test results from all auditors.
All checks from below must be accepted in order to close this issue.
- [ ] @davidjiglesias
- [ ] @jnasselle | non_process | release alpha integration tests description wazuh qa branch wazuh qa commit wazuh tag wazuh commit we are going to check that the integration tests of the branch of wazuh qa work correctly using the version of wazuh the tests will be performed in jenkins using centos as the manager os as for the agents linux windows solaris macos will be used as required tests integration status main rc issue references color status 🟢 all tests passed successfully 🟡 all tests passed but there are some warnings 🔴 some tests have failures or errors large blue circle test execution in progress black circle to do 🟠 jenkins provision fails purple circle all skipped test integration results name jenkins os linux windows solaris macos target manager agent agent agent agent active response 🟢 🟢 🟢 na na agentd na 🟢 🟢 na na analysisd 🟢 na na na na api 🟢 na na na na authd 🟢 nbsp na na na na enrollment na 🟢 🟢 na na fim 🟢 🟢 🟢 🟢 🟢 gcloud 🟢 na na na na github 🟢 🟢 na na na logcollector 🟢 🟢 🟢 🟢 🔴 logtest 🟢 na na na na 🟢 🟢 na na na remoted 🟢 na na na na rids 🟢 na na na na rootcheck 🟢 na na na na vulnerability detector 🔴 na na na na wazuh db 🟢 na na na na syscollector 🟢 🟢 🔴 na 🟢 evidence it launcher build conclusion red circle some tests failed and a known issue was reported leading to the creation of some new issues auditors validation the definition of done for this one is the validation of the conclusions and the test results from all auditors all checks from below must be accepted in order to close this issue davidjiglesias jnasselle | 0 |
8,224 | 11,410,593,309 | IssuesEvent | 2020-02-01 00:03:23 | parcel-bundler/parcel | https://api.github.com/repos/parcel-bundler/parcel | closed | CSS glob import breaking on new file | :bug: Bug CSS Preprocessing Stale | # 🐛 bug report
Parcel throws an error when adding a new file whilst using glob imports in css files
## 🎛 Configuration (.babelrc, package.json, cli command)
Out of the box
## 🤔 Expected Behavior
Reloads watcher when new file is added in the globbed css import
## 😯 Current Behavior
Parcel throws an error saying file not found
`Error: ENOENT: no such file or directory`
## 💻 Code Sample
Folder Structure:
styles/one.css
styles/two.css
styles.three.css
main.css
Import declaration in main.css:
`@import './styles/*.css';`
## 🌍 Your Environment
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 1.12.3
| Node | 12.0.0
| npm/Yarn | 6.9.0/1.15.2
| Operating System | osx 10.14.4
| 1.0 | CSS glob import breaking on new file - # 🐛 bug report
Parcel throws an error when adding a new file whilst using glob imports in css files
## 🎛 Configuration (.babelrc, package.json, cli command)
Out of the box
## 🤔 Expected Behavior
Reloads watcher when new file is added in the globbed css import
## 😯 Current Behavior
Parcel throws an error saying file not found
`Error: ENOENT: no such file or directory`
## 💻 Code Sample
Folder Structure:
styles/one.css
styles/two.css
styles.three.css
main.css
Import declaration in main.css:
`@import './styles/*.css';`
## 🌍 Your Environment
| Software | Version(s) |
| ---------------- | ---------- |
| Parcel | 1.12.3
| Node | 12.0.0
| npm/Yarn | 6.9.0/1.15.2
| Operating System | osx 10.14.4
| process | css glob import breaking on new file 🐛 bug report parcel throws an error when adding a new file whilst using glob imports in css files 🎛 configuration babelrc package json cli command out of the box 🤔 expected behavior reloads watcher when new file is added in the globbed css import 😯 current behavior parcel throws an error saying file not found error enoent no such file or directory 💻 code sample folder structure styles one css styles two css styles three css main css import declaration in main css import styles css 🌍 your environment software version s parcel node npm yarn operating system osx | 1 |
17,351 | 23,173,027,310 | IssuesEvent | 2022-07-31 01:57:07 | lynnandtonic/nestflix.fun | https://api.github.com/repos/lynnandtonic/nestflix.fun | closed | Add Olympus: The Gods' House from “Reality Z” | suggested title in process | Please add as much of the following info as you can:
Title: Olympus: The Gods' House
Type (film/tv show): TV show - reality (foreign - Brazilian)
Film or show in which it appears: Reality Z
Is the parent film/show streaming anywhere? Yes - Netflix
About when in the parent film/show does it appear? Peppered throughout the first episode starting within the first five minutes up until about 18:00.
Actual footage of the film/show can be seen (yes/no)? Yes
Synopsis (written by me): Similar to the hit reality show, _Big Brother_, seven mortals must live together as Greek gods competing to see who gets to gain Zeus' (and the audience's) approval to ultimately stay on Mount Olympus.
Director: Brandão
Cast: Divina McCall (host), Jessica (Aphrodite), Marcos (Apollo), Augusto (Ares), Veronica (Athena), Cleide (Demeter), Madonna (Dionysus), TK (Hermes), and Brandão as the Voice of Zeus
| 1.0 | Add Olympus: The Gods' House from “Reality Z” - Please add as much of the following info as you can:
Title: Olympus: The Gods' House
Type (film/tv show): TV show - reality (foreign - Brazilian)
Film or show in which it appears: Reality Z
Is the parent film/show streaming anywhere? Yes - Netflix
About when in the parent film/show does it appear? Peppered throughout the first episode starting within the first five minutes up until about 18:00.
Actual footage of the film/show can be seen (yes/no)? Yes
Synopsis (written by me): Similar to the hit reality show, _Big Brother_, seven mortals must live together as Greek gods competing to see who gets to gain Zeus' (and the audience's) approval to ultimately stay on Mount Olympus.
Director: Brandão
Cast: Divina McCall (host), Jessica (Aphrodite), Marcos (Apollo), Augusto (Ares), Veronica (Athena), Cleide (Demeter), Madonna (Dionysus), TK (Hermes), and Brandão as the Voice of Zeus
| process | add olympus the gods house from “reality z” please add as much of the following info as you can title olympus the gods house type film tv show tv show reality foreign brazilian film or show in which it appears reality z is the parent film show streaming anywhere yes netflix about when in the parent film show does it appear peppered throughout the first episode starting within the first five minutes up until about actual footage of the film show can be seen yes no yes synopsis written by me similar to the hit reality show big brother seven mortals must live together as greek gods competing to see who gets to gain zeus and the audience s approval to ultimately stay on mount olympus director brandão cast divina mccall host jessica aphrodite marcos apollo augusto ares veronica athena cleide demeter madonna dionysus tk hermes and brandão as the voice of zeus | 1 |
6,085 | 8,941,487,170 | IssuesEvent | 2019-01-24 04:45:20 | googleapis/nodejs-compute | https://api.github.com/repos/googleapis/nodejs-compute | closed | License does not specify copyright correctly | type: process | Your [license](https://github.com/googleapis/nodejs-compute/blob/ee962dd9d1b5b068d5db6a7745cc0d722ac28ba4/LICENSE#L190) still says `Copyright [yyyy] [name of copyright owner]`. I think this needs to be updated. | 1.0 | License does not specify copyright correctly - Your [license](https://github.com/googleapis/nodejs-compute/blob/ee962dd9d1b5b068d5db6a7745cc0d722ac28ba4/LICENSE#L190) still says `Copyright [yyyy] [name of copyright owner]`. I think this needs to be updated. | process | license does not specify copyright correctly your still says copyright i think this needs to be updated | 1 |
112,413 | 17,089,599,716 | IssuesEvent | 2021-07-08 15:43:26 | turkdevops/atom | https://api.github.com/repos/turkdevops/atom | closed | CVE-2020-15168 (Medium) detected in node-fetch-1.7.3.tgz - autoclosed | security vulnerability | ## CVE-2020-15168 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-fetch-1.7.3.tgz</b></p></summary>
<p>A light-weight module that brings window.fetch to node.js and io.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz</a></p>
<p>Path to dependency file: atom/package.json</p>
<p>Path to vulnerable library: atom/node_modules/node-fetch/package.json</p>
<p>
Dependency Hierarchy:
- github-0.34.2.tgz (Root Library)
- react-relay-5.0.0.tgz
- fbjs-1.0.0.tgz
- isomorphic-fetch-2.2.1.tgz
- :x: **node-fetch-1.7.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/atom/commit/0dd3ca131788e5d5aeaf0e095ce6ca46d6df0b9f">0dd3ca131788e5d5aeaf0e095ce6ca46d6df0b9f</a></p>
<p>Found in base branch: <b>electron-upgrade</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
node-fetch before versions 2.6.1 and 3.0.0-beta.9 did not honor the size option after following a redirect, which means that when a content size was over the limit, a FetchError would never get thrown and the process would end without failure. For most people, this fix will have a little or no impact. However, if you are relying on node-fetch to gate files above a size, the impact could be significant, for example: If you don't double-check the size of the data after fetch() has completed, your JS thread could get tied up doing work on a large file (DoS) and/or cost you money in computing.
<p>Publish Date: 2020-09-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15168>CVE-2020-15168</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r">https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.6.1,3.0.0-beta.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-15168 (Medium) detected in node-fetch-1.7.3.tgz - autoclosed - ## CVE-2020-15168 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-fetch-1.7.3.tgz</b></p></summary>
<p>A light-weight module that brings window.fetch to node.js and io.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz">https://registry.npmjs.org/node-fetch/-/node-fetch-1.7.3.tgz</a></p>
<p>Path to dependency file: atom/package.json</p>
<p>Path to vulnerable library: atom/node_modules/node-fetch/package.json</p>
<p>
Dependency Hierarchy:
- github-0.34.2.tgz (Root Library)
- react-relay-5.0.0.tgz
- fbjs-1.0.0.tgz
- isomorphic-fetch-2.2.1.tgz
- :x: **node-fetch-1.7.3.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/turkdevops/atom/commit/0dd3ca131788e5d5aeaf0e095ce6ca46d6df0b9f">0dd3ca131788e5d5aeaf0e095ce6ca46d6df0b9f</a></p>
<p>Found in base branch: <b>electron-upgrade</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
node-fetch before versions 2.6.1 and 3.0.0-beta.9 did not honor the size option after following a redirect, which means that when a content size was over the limit, a FetchError would never get thrown and the process would end without failure. For most people, this fix will have a little or no impact. However, if you are relying on node-fetch to gate files above a size, the impact could be significant, for example: If you don't double-check the size of the data after fetch() has completed, your JS thread could get tied up doing work on a large file (DoS) and/or cost you money in computing.
<p>Publish Date: 2020-09-10
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15168>CVE-2020-15168</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r">https://github.com/node-fetch/node-fetch/security/advisories/GHSA-w7rc-rwvf-8q5r</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: 2.6.1,3.0.0-beta.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_process | cve medium detected in node fetch tgz autoclosed cve medium severity vulnerability vulnerable library node fetch tgz a light weight module that brings window fetch to node js and io js library home page a href path to dependency file atom package json path to vulnerable library atom node modules node fetch package json dependency hierarchy github tgz root library react relay tgz fbjs tgz isomorphic fetch tgz x node fetch tgz vulnerable library found in head commit a href found in base branch electron upgrade vulnerability details node fetch before versions and beta did not honor the size option after following a redirect which means that when a content size was over the limit a fetcherror would never get thrown and the process would end without failure for most people this fix will have a little or no impact however if you are relying on node fetch to gate files above a size the impact could be significant for example if you don t double check the size of the data after fetch has completed your js thread could get tied up doing work on a large file dos and or cost you money in computing publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution beta step up your open source security game with whitesource | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.