Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 5 112 | repo_url stringlengths 34 141 | action stringclasses 3 values | title stringlengths 1 844 | labels stringlengths 4 721 | body stringlengths 1 261k | index stringclasses 12 values | text_combine stringlengths 96 261k | label stringclasses 2 values | text stringlengths 96 248k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
163,770 | 20,364,107,661 | IssuesEvent | 2022-02-21 02:08:39 | violasarah2000/satx2 | https://api.github.com/repos/violasarah2000/satx2 | opened | CVE-2021-3664 (Medium) detected in url-parse-1.4.7.tgz | security vulnerability | ## CVE-2021-3664 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /integration/angular_cli/package.json</p>
<p>Path to vulnerable library: /integration/angular_cli/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.901.12.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- sockjs-client-1.4.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse is vulnerable to URL Redirection to Untrusted Site
<p>Publish Date: 2021-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3664>CVE-2021-3664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664</a></p>
<p>Release Date: 2021-07-26</p>
<p>Fix Resolution (url-parse): 1.5.2</p>
<p>Direct dependency fix Resolution (@angular-devkit/build-angular): 0.901.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-3664 (Medium) detected in url-parse-1.4.7.tgz - ## CVE-2021-3664 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /integration/angular_cli/package.json</p>
<p>Path to vulnerable library: /integration/angular_cli/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- build-angular-0.901.12.tgz (Root Library)
- webpack-dev-server-3.11.0.tgz
- sockjs-client-1.4.0.tgz
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
url-parse is vulnerable to URL Redirection to Untrusted Site
<p>Publish Date: 2021-07-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3664>CVE-2021-3664</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-3664</a></p>
<p>Release Date: 2021-07-26</p>
<p>Fix Resolution (url-parse): 1.5.2</p>
<p>Direct dependency fix Resolution (@angular-devkit/build-angular): 0.901.13</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file integration angular cli package json path to vulnerable library integration angular cli node modules url parse package json dependency hierarchy build angular tgz root library webpack dev server tgz sockjs client tgz x url parse tgz vulnerable library vulnerability details url parse is vulnerable to url redirection to untrusted site publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse direct dependency fix resolution angular devkit build angular step up your open source security game with whitesource | 0 |
125,332 | 10,340,071,362 | IssuesEvent | 2019-09-03 20:55:58 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | [UI] Cluster Templates - Azure cloud provider cluster creation from a cluster template in UI does not display the selections as per the cluster template | [zube]: To Test kind/bug-qa team/ui | Version: Master-head
Steps
1. Create a cluster template/revision T1/R1 in UI with cloud provider set and provide all the credentials
<img width="1243" alt="Screen Shot 2019-07-28 at 6 50 38 AM" src="https://user-images.githubusercontent.com/18536626/62001175-f8e2a080-b107-11e9-92ee-aaba0e57b366.png">
<img width="800" alt="Screen Shot 2019-07-28 at 6 52 42 AM" src="https://user-images.githubusercontent.com/18536626/62001984-1c621700-b119-11e9-9e1a-63b3fd831192.png">
2. Create Cluster and choose the template revision R1. Once the cluster template is selected, the screen shows the cloud provider disabled. The azure cloud provider selection in the template is not honored in UI.
<img width="1178" alt="Screen Shot 2019-07-28 at 8 16 03 AM" src="https://user-images.githubusercontent.com/18536626/62001518-fbe18f00-b10f-11e9-969d-2ef60ae79d50.png">
Cluster template API below:
```
"rancherKubernetesEngineConfig": {
"addonJobTimeout": 30,
"authentication": {
"strategy": "x509",
"type": "/v3/schemas/authnConfig"
},
"cloudProvider": {
"azureCloudProvider": {
"aadClientCertPath": null,
"aadClientId": "REDACTED",
"cloud": null,
"cloudProviderBackoff": false,
"cloudProviderBackoffDuration": 0,
"cloudProviderBackoffExponent": 0,
"cloudProviderBackoffJitter": 0,
"cloudProviderBackoffRetries": 0,
"cloudProviderRateLimit": false,
"cloudProviderRateLimitBucket": 0,
"cloudProviderRateLimitQPS": 0,
"loadBalancerSku": null,
"location": null,
"maximumLoadBalancerRuleCount": 0,
"primaryAvailabilitySetName": null,
"primaryScaleSetName": null,
"resourceGroup": null,
"routeTableName": null,
"securityGroupName": null,
"subnetName": null,
"subscriptionId": "REDACTED",
"tenantId": "REDACTED",
"type": "/v3/schemas/azureCloudProvider",
"useInstanceMetadata": false,
"useManagedIdentityExtension": false,
"vmType": null,
"vnetName": null,
"vnetResourceGroup": null
},
"name": "azure",
"type": "/v3/schemas/cloudProvider"
},
"ignoreDockerVersion": true,
"ingress": {
"provider": "nginx",
"type": "/v3/schemas/ingressConfig"
},
"kubernetesVersion": "1.13.x",
"monitoring": {
"provider": "metrics-server",
"type": "/v3/schemas/monitoringConfig"
},
```
| 1.0 | [UI] Cluster Templates - Azure cloud provider cluster creation from a cluster template in UI does not display the selections as per the cluster template - Version: Master-head
Steps
1. Create a cluster template/revision T1/R1 in UI with cloud provider set and provide all the credentials
<img width="1243" alt="Screen Shot 2019-07-28 at 6 50 38 AM" src="https://user-images.githubusercontent.com/18536626/62001175-f8e2a080-b107-11e9-92ee-aaba0e57b366.png">
<img width="800" alt="Screen Shot 2019-07-28 at 6 52 42 AM" src="https://user-images.githubusercontent.com/18536626/62001984-1c621700-b119-11e9-9e1a-63b3fd831192.png">
2. Create Cluster and choose the template revision R1. Once the cluster template is selected, the screen shows the cloud provider disabled. The azure cloud provider selection in the template is not honored in UI.
<img width="1178" alt="Screen Shot 2019-07-28 at 8 16 03 AM" src="https://user-images.githubusercontent.com/18536626/62001518-fbe18f00-b10f-11e9-969d-2ef60ae79d50.png">
Cluster template API below:
```
"rancherKubernetesEngineConfig": {
"addonJobTimeout": 30,
"authentication": {
"strategy": "x509",
"type": "/v3/schemas/authnConfig"
},
"cloudProvider": {
"azureCloudProvider": {
"aadClientCertPath": null,
"aadClientId": "REDACTED",
"cloud": null,
"cloudProviderBackoff": false,
"cloudProviderBackoffDuration": 0,
"cloudProviderBackoffExponent": 0,
"cloudProviderBackoffJitter": 0,
"cloudProviderBackoffRetries": 0,
"cloudProviderRateLimit": false,
"cloudProviderRateLimitBucket": 0,
"cloudProviderRateLimitQPS": 0,
"loadBalancerSku": null,
"location": null,
"maximumLoadBalancerRuleCount": 0,
"primaryAvailabilitySetName": null,
"primaryScaleSetName": null,
"resourceGroup": null,
"routeTableName": null,
"securityGroupName": null,
"subnetName": null,
"subscriptionId": "REDACTED",
"tenantId": "REDACTED",
"type": "/v3/schemas/azureCloudProvider",
"useInstanceMetadata": false,
"useManagedIdentityExtension": false,
"vmType": null,
"vnetName": null,
"vnetResourceGroup": null
},
"name": "azure",
"type": "/v3/schemas/cloudProvider"
},
"ignoreDockerVersion": true,
"ingress": {
"provider": "nginx",
"type": "/v3/schemas/ingressConfig"
},
"kubernetesVersion": "1.13.x",
"monitoring": {
"provider": "metrics-server",
"type": "/v3/schemas/monitoringConfig"
},
```
| non_priority | cluster templates azure cloud provider cluster creation from a cluster template in ui does not display the selections as per the cluster template version master head steps create a cluster template revision in ui with cloud provider set and provide all the credentials img width alt screen shot at am src img width alt screen shot at am src create cluster and choose the template revision once the cluster template is selected the screen shows the cloud provider disabled the azure cloud provider selection in the template is not honored in ui img width alt screen shot at am src cluster template api below rancherkubernetesengineconfig addonjobtimeout authentication strategy type schemas authnconfig cloudprovider azurecloudprovider aadclientcertpath null aadclientid redacted cloud null cloudproviderbackoff false cloudproviderbackoffduration cloudproviderbackoffexponent cloudproviderbackoffjitter cloudproviderbackoffretries cloudproviderratelimit false cloudproviderratelimitbucket cloudproviderratelimitqps loadbalancersku null location null maximumloadbalancerrulecount primaryavailabilitysetname null primaryscalesetname null resourcegroup null routetablename null securitygroupname null subnetname null subscriptionid redacted tenantid redacted type schemas azurecloudprovider useinstancemetadata false usemanagedidentityextension false vmtype null vnetname null vnetresourcegroup null name azure type schemas cloudprovider ignoredockerversion true ingress provider nginx type schemas ingressconfig kubernetesversion x monitoring provider metrics server type schemas monitoringconfig | 0 |
51,023 | 13,612,773,005 | IssuesEvent | 2020-09-23 10:49:23 | jgeraigery/E-Forms | https://api.github.com/repos/jgeraigery/E-Forms | opened | CVE-2018-16487 (Medium) detected in lodash-3.10.1.tgz, lodash-1.0.2.tgz | security vulnerability | ## CVE-2018-16487 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-3.10.1.tgz</b>, <b>lodash-1.0.2.tgz</b></p></summary>
<p>
<details><summary><b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: E-Forms/package.json</p>
<p>Path to vulnerable library: E-Forms/node_modules/karma/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.5.0.tgz (Root Library)
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>lodash-1.0.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, and extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p>
<p>Path to dependency file: E-Forms/package.json</p>
<p>Path to vulnerable library: E-Forms/node_modules/glob-watcher/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- gulp-3.9.1.tgz (Root Library)
- vinyl-fs-0.3.14.tgz
- glob-watcher-0.0.6.tgz
- gaze-0.5.2.tgz
- globule-0.1.0.tgz
- :x: **lodash-1.0.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/E-Forms/commit/d2563d81800b6075ab02a5ae4913e8aec8f4cc32">d2563d81800b6075ab02a5ae4913e8aec8f4cc32</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.
<p>Publish Date: 2019-02-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p>
<p>Release Date: 2019-02-01</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"karma:1.5.0;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.0.2","isTransitiveDependency":true,"dependencyTree":"gulp:3.9.1;vinyl-fs:0.3.14;glob-watcher:0.0.6;gaze:0.5.2;globule:0.1.0;lodash:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2018-16487","vulnerabilityDetails":"A prototype pollution vulnerability was found in lodash \u003c4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487","cvss3Severity":"medium","cvss3Score":"5.6","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2018-16487 (Medium) detected in lodash-3.10.1.tgz, lodash-1.0.2.tgz - ## CVE-2018-16487 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-3.10.1.tgz</b>, <b>lodash-1.0.2.tgz</b></p></summary>
<p>
<details><summary><b>lodash-3.10.1.tgz</b></p></summary>
<p>The modern build of lodash modular utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p>
<p>Path to dependency file: E-Forms/package.json</p>
<p>Path to vulnerable library: E-Forms/node_modules/karma/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- karma-1.5.0.tgz (Root Library)
- :x: **lodash-3.10.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>lodash-1.0.2.tgz</b></p></summary>
<p>A utility library delivering consistency, customization, performance, and extras.</p>
<p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p>
<p>Path to dependency file: E-Forms/package.json</p>
<p>Path to vulnerable library: E-Forms/node_modules/glob-watcher/node_modules/lodash/package.json</p>
<p>
Dependency Hierarchy:
- gulp-3.9.1.tgz (Root Library)
- vinyl-fs-0.3.14.tgz
- glob-watcher-0.0.6.tgz
- gaze-0.5.2.tgz
- globule-0.1.0.tgz
- :x: **lodash-1.0.2.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/E-Forms/commit/d2563d81800b6075ab02a5ae4913e8aec8f4cc32">d2563d81800b6075ab02a5ae4913e8aec8f4cc32</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A prototype pollution vulnerability was found in lodash <4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.
<p>Publish Date: 2019-02-01
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487>CVE-2018-16487</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487">https://bugzilla.redhat.com/show_bug.cgi?id=CVE-2018-16487</a></p>
<p>Release Date: 2019-02-01</p>
<p>Fix Resolution: 4.17.11</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"3.10.1","isTransitiveDependency":true,"dependencyTree":"karma:1.5.0;lodash:3.10.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"},{"packageType":"javascript/Node.js","packageName":"lodash","packageVersion":"1.0.2","isTransitiveDependency":true,"dependencyTree":"gulp:3.9.1;vinyl-fs:0.3.14;glob-watcher:0.0.6;gaze:0.5.2;globule:0.1.0;lodash:1.0.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"4.17.11"}],"vulnerabilityIdentifier":"CVE-2018-16487","vulnerabilityDetails":"A prototype pollution vulnerability was found in lodash \u003c4.17.11 where the functions merge, mergeWith, and defaultsDeep can be tricked into adding or modifying properties of Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16487","cvss3Severity":"medium","cvss3Score":"5.6","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"Low","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_priority | cve medium detected in lodash tgz lodash tgz cve medium severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file e forms package json path to vulnerable library e forms node modules karma node modules lodash package json dependency hierarchy karma tgz root library x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file e forms package json path to vulnerable library e forms node modules glob watcher node modules lodash package json dependency hierarchy gulp tgz root library vinyl fs tgz glob watcher tgz gaze tgz globule tgz x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails a prototype pollution vulnerability was found in lodash where the functions merge mergewith and defaultsdeep can be tricked into adding or modifying properties of object prototype vulnerabilityurl | 0 |
431,540 | 30,238,816,550 | IssuesEvent | 2023-07-06 12:12:56 | mango-gremlin/Teamprojekt-Diagnosis-System | https://api.github.com/repos/mango-gremlin/Teamprojekt-Diagnosis-System | opened | Complete REAMDE for lumpy skin | documentation low importance | Complete the REAMDE file for the lumpy skin model.
Currently, there is a READMEpre.md file in Project/ML/Lumpy Skin.
Find the important information in there | 1.0 | Complete REAMDE for lumpy skin - Complete the REAMDE file for the lumpy skin model.
Currently, there is a READMEpre.md file in Project/ML/Lumpy Skin.
Find the important information in there | non_priority | complete reamde for lumpy skin complete the reamde file for the lumpy skin model currently there is a readmepre md file in project ml lumpy skin find the important information in there | 0 |
343,674 | 30,683,651,168 | IssuesEvent | 2023-07-26 10:49:24 | opendatahub-io/distributed-workloads | https://api.github.com/repos/opendatahub-io/distributed-workloads | opened | Understand the test Implementation and coverage for Ray Cluster test | testing | - Go through the code and understand the test Implementation for Ray cluster tests
- Run the tests manually https://github.com/opendatahub-io/distributed-workloads/tree/main/tests
| 1.0 | Understand the test Implementation and coverage for Ray Cluster test - - Go through the code and understand the test Implementation for Ray cluster tests
- Run the tests manually https://github.com/opendatahub-io/distributed-workloads/tree/main/tests
| non_priority | understand the test implementation and coverage for ray cluster test go through the code and understand the test implementation for ray cluster tests run the tests manually | 0 |
392,217 | 26,930,512,410 | IssuesEvent | 2023-02-07 16:30:31 | Pachka/mwss-App | https://api.github.com/repos/Pachka/mwss-App | opened | Minor typos in "How to use" tab | documentation simulation | Various small typos in "how to use" text should be corrected.
NB: Lulla started drafting a suggested replacement text in the google doc. | 1.0 | Minor typos in "How to use" tab - Various small typos in "how to use" text should be corrected.
NB: Lulla started drafting a suggested replacement text in the google doc. | non_priority | minor typos in how to use tab various small typos in how to use text should be corrected nb lulla started drafting a suggested replacement text in the google doc | 0 |
301,217 | 26,027,568,954 | IssuesEvent | 2022-12-21 17:42:16 | microsoft/vscode-python | https://api.github.com/repos/microsoft/vscode-python | closed | Enable runtime invocation parameters for test runs | feature-request area-testing needs proposal | It does not currently appear possible to pass arguments to the test runner when executing or debugging unit tests. This is especially problematic when tests are tagged and one typically wants to execute different test groups or group combinations.
| 1.0 | Enable runtime invocation parameters for test runs - It does not currently appear possible to pass arguments to the test runner when executing or debugging unit tests. This is especially problematic when tests are tagged and one typically wants to execute different test groups or group combinations.
| non_priority | enable runtime invocation parameters for test runs it does not currently appear possible to pass arguments to the test runner when executing or debugging unit tests this is especially problematic when tests are tagged and one typically wants to execute different test groups or group combinations | 0 |
269,101 | 28,960,003,409 | IssuesEvent | 2023-05-10 01:07:25 | ChoeMinji/mongo-r4.4.6 | https://api.github.com/repos/ChoeMinji/mongo-r4.4.6 | reopened | CVE-2019-6292 (Medium) detected in mongor5.1.0-rc0 | Mend: dependency security vulnerability | ## CVE-2019-6292 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mongor5.1.0-rc0</b></p></summary>
<p>
<p>The MongoDB Database</p>
<p>Library home page: <a href=https://github.com/mongodb/mongo.git>https://github.com/mongodb/mongo.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/mongo-r4.4.6/commit/9c4537f1af3987a4f237e73712977c87c207c818">9c4537f1af3987a4f237e73712977c87c207c818</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/third_party/yaml-cpp-0.6.2/yaml-cpp/src/singledocparser.cpp</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/third_party/yaml-cpp-0.6.2/yaml-cpp/src/singledocparser.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in singledocparser.cpp in yaml-cpp (aka LibYaml-C++) 0.6.2. Stack Exhaustion occurs in YAML::SingleDocParser, and there is a stack consumption problem caused by recursive stack frames: HandleCompactMap, HandleMap, HandleFlowSequence, HandleSequence, HandleNode. Remote attackers could leverage this vulnerability to cause a denial-of-service via a cpp file.
<p>Publish Date: 2019-01-15
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-6292>CVE-2019-6292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-6292">https://nvd.nist.gov/vuln/detail/CVE-2019-6292</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: libyaml-cpp0.3 - 0.6.3-1,0.6.3-1;libyaml-cpp0.5 - 0.6.3-1,0.6.3-1;libyaml-cpp0.6 - 0.6.3-1,0.6.3-1;libyaml-cpp-dev - 0.6.3-1,0.6.3-1;libyaml-cpp0.5v5 - 0.6.3-1,0.6.3-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-6292 (Medium) detected in mongor5.1.0-rc0 - ## CVE-2019-6292 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mongor5.1.0-rc0</b></p></summary>
<p>
<p>The MongoDB Database</p>
<p>Library home page: <a href=https://github.com/mongodb/mongo.git>https://github.com/mongodb/mongo.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/ChoeMinji/mongo-r4.4.6/commit/9c4537f1af3987a4f237e73712977c87c207c818">9c4537f1af3987a4f237e73712977c87c207c818</a></p>
<p>Found in base branch: <b>main</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/third_party/yaml-cpp-0.6.2/yaml-cpp/src/singledocparser.cpp</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/src/third_party/yaml-cpp-0.6.2/yaml-cpp/src/singledocparser.cpp</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in singledocparser.cpp in yaml-cpp (aka LibYaml-C++) 0.6.2. Stack Exhaustion occurs in YAML::SingleDocParser, and there is a stack consumption problem caused by recursive stack frames: HandleCompactMap, HandleMap, HandleFlowSequence, HandleSequence, HandleNode. Remote attackers could leverage this vulnerability to cause a denial-of-service via a cpp file.
<p>Publish Date: 2019-01-15
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-6292>CVE-2019-6292</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2019-6292">https://nvd.nist.gov/vuln/detail/CVE-2019-6292</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: libyaml-cpp0.3 - 0.6.3-1,0.6.3-1;libyaml-cpp0.5 - 0.6.3-1,0.6.3-1;libyaml-cpp0.6 - 0.6.3-1,0.6.3-1;libyaml-cpp-dev - 0.6.3-1,0.6.3-1;libyaml-cpp0.5v5 - 0.6.3-1,0.6.3-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in cve medium severity vulnerability vulnerable library the mongodb database library home page a href found in head commit a href found in base branch main vulnerable source files src third party yaml cpp yaml cpp src singledocparser cpp src third party yaml cpp yaml cpp src singledocparser cpp vulnerability details an issue was discovered in singledocparser cpp in yaml cpp aka libyaml c stack exhaustion occurs in yaml singledocparser and there is a stack consumption problem caused by recursive stack frames handlecompactmap handlemap handleflowsequence handlesequence handlenode remote attackers could leverage this vulnerability to cause a denial of service via a cpp file publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libyaml libyaml libyaml libyaml cpp dev libyaml step up your open source security game with mend | 0 |
62,414 | 8,607,467,817 | IssuesEvent | 2018-11-17 22:46:34 | fga-eps-mds/2018.2-Integra-Vendas | https://api.github.com/repos/fga-eps-mds/2018.2-Integra-Vendas | closed | Documentação da Sprint 10 | 0-Scrum Master 2-Documentation | <!--- Descreva a atividade que deve ser feita para atender a issue --->
Fazer a documentação da Sprint 10.
**Tarefas**
- [ ] Planejamento da sprint;
- [ ] Retrospectiva da sprint;
- [ ] Revisão da sprint;
- [ ] Métricas.
**Observações**
* A *issue* deve ser pontuada;
* A *issue* deve ser delegada a alguém;
* A *issue* deve ter *labels*;
* A *issue* deve pertencer a uma *milestone*. | 1.0 | Documentação da Sprint 10 - <!--- Descreva a atividade que deve ser feita para atender a issue --->
Fazer a documentação da Sprint 10.
**Tarefas**
- [ ] Planejamento da sprint;
- [ ] Retrospectiva da sprint;
- [ ] Revisão da sprint;
- [ ] Métricas.
**Observações**
* A *issue* deve ser pontuada;
* A *issue* deve ser delegada a alguém;
* A *issue* deve ter *labels*;
* A *issue* deve pertencer a uma *milestone*. | non_priority | documentação da sprint fazer a documentação da sprint tarefas planejamento da sprint retrospectiva da sprint revisão da sprint métricas observações a issue deve ser pontuada a issue deve ser delegada a alguém a issue deve ter labels a issue deve pertencer a uma milestone | 0 |
115,575 | 9,805,302,607 | IssuesEvent | 2019-06-12 08:41:02 | Students-of-the-city-of-Kostroma/Student-timetable | https://api.github.com/repos/Students-of-the-city-of-Kostroma/Student-timetable | opened | Исправить сбои в автотестах расположенных в файле UT_Delete_CTrainingProfile | Auto test Script Unit test | Выявить причину сбоев в автотестах.
При необходимости исправить автотесты и сценарии
#422 Script | 2.0 | Исправить сбои в автотестах расположенных в файле UT_Delete_CTrainingProfile - Выявить причину сбоев в автотестах.
При необходимости исправить автотесты и сценарии
#422 Script | non_priority | исправить сбои в автотестах расположенных в файле ut delete ctrainingprofile выявить причину сбоев в автотестах при необходимости исправить автотесты и сценарии script | 0 |
176,645 | 14,593,057,882 | IssuesEvent | 2020-12-19 20:42:27 | MIDIBlocks/handsfree | https://api.github.com/repos/MIDIBlocks/handsfree | closed | pinchScroll plugin for hands | documentation enhancement | - [x] Create the plugin
- [x] Document
- [x] Create an example using it
## Misc
- [x] Make sidebar wider on desktop, so that the debugger is more visible | 1.0 | pinchScroll plugin for hands - - [x] Create the plugin
- [x] Document
- [x] Create an example using it
## Misc
- [x] Make sidebar wider on desktop, so that the debugger is more visible | non_priority | pinchscroll plugin for hands create the plugin document create an example using it misc make sidebar wider on desktop so that the debugger is more visible | 0 |
19,539 | 6,734,341,419 | IssuesEvent | 2017-10-18 17:42:27 | hashicorp/packer | https://api.github.com/repos/hashicorp/packer | closed | lxd will not produce any artifacts when name parameter is set. | bug builder/lxd | **Affected versions:** Packer 1.1.0
**Host platform:** Ubuntu 16.04.3 LTS
**Builder:** lxd
**Provisioner:** ansible-local with Ansible v2.3.2.0
Below packer config fails to produce any artifacts
```
{
"type": "lxd",
"name": "u1",
"image": "ubuntu-daily:xenial"
}
```
Here is the output
```
Building lxd xenial
==> Builds finished but no artifacts were created.
```
If I remove `name` leaving just
```
{
"type": "lxd",
"image": "ubuntu-daily:xenial"
}
```
Then all works as expected
```
Building lxd xenial
lxd output will be in this color.
==> lxd: Creating container...
==> lxd: Provisioning with shell script: /tmp/packer-shell220599120
==> lxd: Stopping container...
==> lxd: Publishing container...
==> lxd: Created image: 6d56b1d0dd667f92e7d7431202d332b33d06530552fdc9615029475e2159c833
==> lxd: Unregistering and deleting deleting container...
```
```
2017/09/22 16:27:47 [INFO] Packer version: 1.1.0
2017/09/22 16:27:47 Packer Target OS/Arch: linux amd64
2017/09/22 16:27:47 Built with Go Version: go1.9
2017/09/22 16:27:47 Detected home directory from env var: /home/lucaszb
2017/09/22 16:27:47 Using internal plugin for qemu
2017/09/22 16:27:47 Using internal plugin for amazon-ebssurrogate
2017/09/22 16:27:47 Using internal plugin for cloudstack
2017/09/22 16:27:47 Using internal plugin for docker
2017/09/22 16:27:47 Using internal plugin for hyperv-iso
2017/09/22 16:27:47 Using internal plugin for lxc
2017/09/22 16:27:47 Using internal plugin for parallels-iso
2017/09/22 16:27:47 Using internal plugin for amazon-ebs
2017/09/22 16:27:47 Using internal plugin for amazon-instance
2017/09/22 16:27:47 Using internal plugin for azure-arm
2017/09/22 16:27:47 Using internal plugin for googlecompute
2017/09/22 16:27:47 Using internal plugin for lxd
2017/09/22 16:27:47 Using internal plugin for virtualbox-ovf
2017/09/22 16:27:47 Using internal plugin for amazon-chroot
2017/09/22 16:27:47 Using internal plugin for file
2017/09/22 16:27:47 Using internal plugin for oracle-oci
2017/09/22 16:27:47 Using internal plugin for parallels-pvm
2017/09/22 16:27:47 Using internal plugin for digitalocean
2017/09/22 16:27:47 Using internal plugin for oneandone
2017/09/22 16:27:47 Using internal plugin for profitbricks
2017/09/22 16:27:47 Using internal plugin for triton
2017/09/22 16:27:47 Using internal plugin for vmware-vmx
2017/09/22 16:27:47 Using internal plugin for alicloud-ecs
2017/09/22 16:27:47 Using internal plugin for amazon-ebsvolume
2017/09/22 16:27:47 Using internal plugin for null
2017/09/22 16:27:47 Using internal plugin for openstack
2017/09/22 16:27:47 Using internal plugin for virtualbox-iso
2017/09/22 16:27:47 Using internal plugin for vmware-iso
2017/09/22 16:27:47 Using internal plugin for shell-local
2017/09/22 16:27:47 Using internal plugin for chef-solo
2017/09/22 16:27:47 Using internal plugin for converge
2017/09/22 16:27:47 Using internal plugin for chef-client
2017/09/22 16:27:47 Using internal plugin for file
2017/09/22 16:27:47 Using internal plugin for salt-masterless
2017/09/22 16:27:47 Using internal plugin for ansible
2017/09/22 16:27:47 Using internal plugin for ansible-local
2017/09/22 16:27:47 Using internal plugin for shell
2017/09/22 16:27:47 Using internal plugin for windows-restart
2017/09/22 16:27:47 Using internal plugin for puppet-masterless
2017/09/22 16:27:47 Using internal plugin for puppet-server
2017/09/22 16:27:47 Using internal plugin for powershell
2017/09/22 16:27:47 Using internal plugin for windows-shell
2017/09/22 16:27:47 Using internal plugin for amazon-import
2017/09/22 16:27:47 Using internal plugin for compress
2017/09/22 16:27:47 Using internal plugin for docker-push
2017/09/22 16:27:47 Using internal plugin for manifest
2017/09/22 16:27:47 Using internal plugin for vagrant-cloud
2017/09/22 16:27:47 Using internal plugin for alicloud-import
2017/09/22 16:27:47 Using internal plugin for artifice
2017/09/22 16:27:47 Using internal plugin for atlas
2017/09/22 16:27:47 Using internal plugin for docker-import
2017/09/22 16:27:47 Using internal plugin for checksum
2017/09/22 16:27:47 Using internal plugin for docker-save
2017/09/22 16:27:47 Using internal plugin for googlecompute-export
2017/09/22 16:27:47 Using internal plugin for vsphere
2017/09/22 16:27:47 Using internal plugin for docker-tag
2017/09/22 16:27:47 Using internal plugin for shell-local
2017/09/22 16:27:47 Using internal plugin for vagrant
2017/09/22 16:27:47 Using internal plugin for vsphere-template
2017/09/22 16:27:47 Detected home directory from env var: /home/lucaszb
2017/09/22 16:27:47 Attempting to open config file: /home/lucaszb/.packerconfig
2017/09/22 16:27:47 [WARN] Config file doesn't exist: /home/lucaszb/.packerconfig
2017/09/22 16:27:47 Packer config: &{DisableCheckpoint:false DisableCheckpointSignature:false PluginMinPort:10000 PluginMaxPort:25000 Builders:map[digitalocean:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-digitalocean amazon-ebsvolume:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-ebsvolume hyperv-iso:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-hyperv-iso amazon-ebs:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-ebs vmware-vmx:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-vmware-vmx vmware-iso:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-vmware-iso file:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-file oracle-oci:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-oracle-oci oneandone:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-oneandone null:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-null qemu:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-qemu virtualbox-ovf:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-virtualbox-ovf amazon-ebssurrogate:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-ebssurrogate triton:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-triton alicloud-ecs:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-alicloud-ecs amazon-instance:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-instance lxd:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-lxd amazon-chroot:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-chroot profitbricks:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-profitbricks openstack:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-openstack docker:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-docker parallels-iso:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-parallels-iso azure-arm:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-azure-arm cloudstack:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-cloudstack lxc:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-lxc virtualbox-iso:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-virtualbox-iso googlecompute:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-googlecompute parallels-pvm:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-parallels-pvm] PostProcessors:map[amazon-import:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-amazon-import docker-push:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-docker-push manifest:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-manifest checksum:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-checksum googlecompute-export:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-googlecompute-export vagrant:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-vagrant docker-import:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-docker-import docker-save:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-docker-save vsphere-template:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-vsphere-template compress:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-compress alicloud-import:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-alicloud-import vsphere:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-vsphere vagrant-cloud:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-vagrant-cloud artifice:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-artifice atlas:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-atlas docker-tag:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-docker-tag shell-local:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-shell-local] Provisioners:map[shell:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-shell windows-restart:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-windows-restart powershell:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-powershell windows-shell:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-windows-shell file:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-file ansible:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-ansible ansible-local:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-ansible-local puppet-masterless:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-puppet-masterless salt-masterless:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-salt-masterless puppet-server:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-puppet-server chef-solo:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-chef-solo converge:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-converge chef-client:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-chef-client shell-local:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-shell-local]}
2017/09/22 16:27:47 Detected home directory from env var: /home/lucaszb
2017/09/22 16:27:47 Setting cache directory: /home/lucaszb/...packer_cache
2017/09/22 16:27:47 ui: Debug mode enabled. Builds will not be parallelized.
2017/09/22 16:27:47 Build debug mode: true
2017/09/22 16:27:47 Force build: false
2017/09/22 16:27:47 On error:
2017/09/22 16:27:47 Waiting on builds to complete...
2017/09/22 16:27:47 Builds completed. Waiting on interrupt barrier...
2017/09/22 16:27:47 ui:
==> Builds finished but no artifacts were created.
2017/09/22 16:27:47 Detected home directory from env var: /home/lucaszb
2017/09/22 16:27:47 waiting for all plugin processes to complete...
```
| 1.0 | lxd will not produce any artifacts when name parameter is set. - **Affected versions:** Packer 1.1.0
**Host platform:** Ubuntu 16.04.3 LTS
**Builder:** lxd
**Provisioner:** ansible-local with Ansible v2.3.2.0
Below packer config fails to produce any artifacts
```
{
"type": "lxd",
"name": "u1",
"image": "ubuntu-daily:xenial"
}
```
Here is the output
```
Building lxd xenial
==> Builds finished but no artifacts were created.
```
If I remove `name` leaving just
```
{
"type": "lxd",
"image": "ubuntu-daily:xenial"
}
```
Then all works as expected
```
Building lxd xenial
lxd output will be in this color.
==> lxd: Creating container...
==> lxd: Provisioning with shell script: /tmp/packer-shell220599120
==> lxd: Stopping container...
==> lxd: Publishing container...
==> lxd: Created image: 6d56b1d0dd667f92e7d7431202d332b33d06530552fdc9615029475e2159c833
==> lxd: Unregistering and deleting deleting container...
```
```
2017/09/22 16:27:47 [INFO] Packer version: 1.1.0
2017/09/22 16:27:47 Packer Target OS/Arch: linux amd64
2017/09/22 16:27:47 Built with Go Version: go1.9
2017/09/22 16:27:47 Detected home directory from env var: /home/lucaszb
2017/09/22 16:27:47 Using internal plugin for qemu
2017/09/22 16:27:47 Using internal plugin for amazon-ebssurrogate
2017/09/22 16:27:47 Using internal plugin for cloudstack
2017/09/22 16:27:47 Using internal plugin for docker
2017/09/22 16:27:47 Using internal plugin for hyperv-iso
2017/09/22 16:27:47 Using internal plugin for lxc
2017/09/22 16:27:47 Using internal plugin for parallels-iso
2017/09/22 16:27:47 Using internal plugin for amazon-ebs
2017/09/22 16:27:47 Using internal plugin for amazon-instance
2017/09/22 16:27:47 Using internal plugin for azure-arm
2017/09/22 16:27:47 Using internal plugin for googlecompute
2017/09/22 16:27:47 Using internal plugin for lxd
2017/09/22 16:27:47 Using internal plugin for virtualbox-ovf
2017/09/22 16:27:47 Using internal plugin for amazon-chroot
2017/09/22 16:27:47 Using internal plugin for file
2017/09/22 16:27:47 Using internal plugin for oracle-oci
2017/09/22 16:27:47 Using internal plugin for parallels-pvm
2017/09/22 16:27:47 Using internal plugin for digitalocean
2017/09/22 16:27:47 Using internal plugin for oneandone
2017/09/22 16:27:47 Using internal plugin for profitbricks
2017/09/22 16:27:47 Using internal plugin for triton
2017/09/22 16:27:47 Using internal plugin for vmware-vmx
2017/09/22 16:27:47 Using internal plugin for alicloud-ecs
2017/09/22 16:27:47 Using internal plugin for amazon-ebsvolume
2017/09/22 16:27:47 Using internal plugin for null
2017/09/22 16:27:47 Using internal plugin for openstack
2017/09/22 16:27:47 Using internal plugin for virtualbox-iso
2017/09/22 16:27:47 Using internal plugin for vmware-iso
2017/09/22 16:27:47 Using internal plugin for shell-local
2017/09/22 16:27:47 Using internal plugin for chef-solo
2017/09/22 16:27:47 Using internal plugin for converge
2017/09/22 16:27:47 Using internal plugin for chef-client
2017/09/22 16:27:47 Using internal plugin for file
2017/09/22 16:27:47 Using internal plugin for salt-masterless
2017/09/22 16:27:47 Using internal plugin for ansible
2017/09/22 16:27:47 Using internal plugin for ansible-local
2017/09/22 16:27:47 Using internal plugin for shell
2017/09/22 16:27:47 Using internal plugin for windows-restart
2017/09/22 16:27:47 Using internal plugin for puppet-masterless
2017/09/22 16:27:47 Using internal plugin for puppet-server
2017/09/22 16:27:47 Using internal plugin for powershell
2017/09/22 16:27:47 Using internal plugin for windows-shell
2017/09/22 16:27:47 Using internal plugin for amazon-import
2017/09/22 16:27:47 Using internal plugin for compress
2017/09/22 16:27:47 Using internal plugin for docker-push
2017/09/22 16:27:47 Using internal plugin for manifest
2017/09/22 16:27:47 Using internal plugin for vagrant-cloud
2017/09/22 16:27:47 Using internal plugin for alicloud-import
2017/09/22 16:27:47 Using internal plugin for artifice
2017/09/22 16:27:47 Using internal plugin for atlas
2017/09/22 16:27:47 Using internal plugin for docker-import
2017/09/22 16:27:47 Using internal plugin for checksum
2017/09/22 16:27:47 Using internal plugin for docker-save
2017/09/22 16:27:47 Using internal plugin for googlecompute-export
2017/09/22 16:27:47 Using internal plugin for vsphere
2017/09/22 16:27:47 Using internal plugin for docker-tag
2017/09/22 16:27:47 Using internal plugin for shell-local
2017/09/22 16:27:47 Using internal plugin for vagrant
2017/09/22 16:27:47 Using internal plugin for vsphere-template
2017/09/22 16:27:47 Detected home directory from env var: /home/lucaszb
2017/09/22 16:27:47 Attempting to open config file: /home/lucaszb/.packerconfig
2017/09/22 16:27:47 [WARN] Config file doesn't exist: /home/lucaszb/.packerconfig
2017/09/22 16:27:47 Packer config: &{DisableCheckpoint:false DisableCheckpointSignature:false PluginMinPort:10000 PluginMaxPort:25000 Builders:map[digitalocean:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-digitalocean amazon-ebsvolume:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-ebsvolume hyperv-iso:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-hyperv-iso amazon-ebs:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-ebs vmware-vmx:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-vmware-vmx vmware-iso:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-vmware-iso file:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-file oracle-oci:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-oracle-oci oneandone:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-oneandone null:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-null qemu:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-qemu virtualbox-ovf:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-virtualbox-ovf amazon-ebssurrogate:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-ebssurrogate triton:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-triton alicloud-ecs:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-alicloud-ecs amazon-instance:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-instance lxd:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-lxd amazon-chroot:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-amazon-chroot profitbricks:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-profitbricks openstack:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-openstack docker:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-docker parallels-iso:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-parallels-iso azure-arm:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-azure-arm cloudstack:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-cloudstack lxc:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-lxc virtualbox-iso:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-virtualbox-iso googlecompute:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-googlecompute parallels-pvm:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-builder-parallels-pvm] PostProcessors:map[amazon-import:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-amazon-import docker-push:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-docker-push manifest:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-manifest checksum:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-checksum googlecompute-export:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-googlecompute-export vagrant:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-vagrant docker-import:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-docker-import docker-save:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-docker-save vsphere-template:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-vsphere-template compress:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-compress alicloud-import:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-alicloud-import vsphere:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-vsphere vagrant-cloud:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-vagrant-cloud artifice:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-artifice atlas:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-atlas docker-tag:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-docker-tag shell-local:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-post-processor-shell-local] Provisioners:map[shell:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-shell windows-restart:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-windows-restart powershell:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-powershell windows-shell:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-windows-shell file:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-file ansible:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-ansible ansible-local:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-ansible-local puppet-masterless:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-puppet-masterless salt-masterless:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-salt-masterless puppet-server:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-puppet-server chef-solo:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-chef-solo converge:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-converge chef-client:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-chef-client shell-local:/usr/local/bin/packer-PACKERSPACE-plugin-PACKERSPACE-packer-provisioner-shell-local]}
2017/09/22 16:27:47 Detected home directory from env var: /home/lucaszb
2017/09/22 16:27:47 Setting cache directory: /home/lucaszb/...packer_cache
2017/09/22 16:27:47 ui: Debug mode enabled. Builds will not be parallelized.
2017/09/22 16:27:47 Build debug mode: true
2017/09/22 16:27:47 Force build: false
2017/09/22 16:27:47 On error:
2017/09/22 16:27:47 Waiting on builds to complete...
2017/09/22 16:27:47 Builds completed. Waiting on interrupt barrier...
2017/09/22 16:27:47 ui:
==> Builds finished but no artifacts were created.
2017/09/22 16:27:47 Detected home directory from env var: /home/lucaszb
2017/09/22 16:27:47 waiting for all plugin processes to complete...
```
| non_priority | lxd will not produce any artifacts when name parameter is set affected versions packer host platform ubuntu lts builder lxd provisioner ansible local with ansible below packer config fails to produce any artifacts type lxd name image ubuntu daily xenial here is the output building lxd xenial builds finished but no artifacts were created if i remove name leaving just type lxd image ubuntu daily xenial then all works as expected building lxd xenial lxd output will be in this color lxd creating container lxd provisioning with shell script tmp packer lxd stopping container lxd publishing container lxd created image lxd unregistering and deleting deleting container packer version packer target os arch linux built with go version detected home directory from env var home lucaszb using internal plugin for qemu using internal plugin for amazon ebssurrogate using internal plugin for cloudstack using internal plugin for docker using internal plugin for hyperv iso using internal plugin for lxc using internal plugin for parallels iso using internal plugin for amazon ebs using internal plugin for amazon instance using internal plugin for azure arm using internal plugin for googlecompute using internal plugin for lxd using internal plugin for virtualbox ovf using internal plugin for amazon chroot using internal plugin for file using internal plugin for oracle oci using internal plugin for parallels pvm using internal plugin for digitalocean using internal plugin for oneandone using internal plugin for profitbricks using internal plugin for triton using internal plugin for vmware vmx using internal plugin for alicloud ecs using internal plugin for amazon ebsvolume using internal plugin for null using internal plugin for openstack using internal plugin for virtualbox iso using internal plugin for vmware iso using internal plugin for shell local using internal plugin for chef solo using internal plugin for converge using internal plugin for chef client using internal plugin for file using internal plugin for salt masterless using internal plugin for ansible using internal plugin for ansible local using internal plugin for shell using internal plugin for windows restart using internal plugin for puppet masterless using internal plugin for puppet server using internal plugin for powershell using internal plugin for windows shell using internal plugin for amazon import using internal plugin for compress using internal plugin for docker push using internal plugin for manifest using internal plugin for vagrant cloud using internal plugin for alicloud import using internal plugin for artifice using internal plugin for atlas using internal plugin for docker import using internal plugin for checksum using internal plugin for docker save using internal plugin for googlecompute export using internal plugin for vsphere using internal plugin for docker tag using internal plugin for shell local using internal plugin for vagrant using internal plugin for vsphere template detected home directory from env var home lucaszb attempting to open config file home lucaszb packerconfig config file doesn t exist home lucaszb packerconfig packer config disablecheckpoint false disablecheckpointsignature false pluginminport pluginmaxport builders map postprocessors map provisioners map detected home directory from env var home lucaszb setting cache directory home lucaszb packer cache ui debug mode enabled builds will not be parallelized build debug mode true force build false on error waiting on builds to complete builds completed waiting on interrupt barrier ui builds finished but no artifacts were created detected home directory from env var home lucaszb waiting for all plugin processes to complete | 0 |
149,843 | 23,537,625,452 | IssuesEvent | 2022-08-19 23:47:54 | dgkf/parttime | https://api.github.com/repos/dgkf/parttime | closed | Reduce parttime fields (remove `secfrac` and `tzmin`) | help wanted design (impl.) | I've been considering whether some of these fields should be collapsed.
- `secfrac` is a bit redundant with `sec` (we could easily just include the fractional component)
- `tzhour` and `tzmin` could be collapsed into a single `tz` field (I haven't decided yet what the units of this field would be - leaning toward hours since it's most often how tz offsets are communicated).
Redundancy is inevitable (since you could just as well store everything as seconds + origin with an uncertainty). What makes these specific fields questionable is that they are rarely used in isolation and could easily be recovered from the reduced forms.
Collapsing these would simplify some of the formatting, reduce the object size and simplify the data structure slightly. | 1.0 | Reduce parttime fields (remove `secfrac` and `tzmin`) - I've been considering whether some of these fields should be collapsed.
- `secfrac` is a bit redundant with `sec` (we could easily just include the fractional component)
- `tzhour` and `tzmin` could be collapsed into a single `tz` field (I haven't decided yet what the units of this field would be - leaning toward hours since it's most often how tz offsets are communicated).
Redundancy is inevitable (since you could just as well store everything as seconds + origin with an uncertainty). What makes these specific fields questionable is that they are rarely used in isolation and could easily be recovered from the reduced forms.
Collapsing these would simplify some of the formatting, reduce the object size and simplify the data structure slightly. | non_priority | reduce parttime fields remove secfrac and tzmin i ve been considering whether some of these fields should be collapsed secfrac is a bit redundant with sec we could easily just include the fractional component tzhour and tzmin could be collapsed into a single tz field i haven t decided yet what the units of this field would be leaning toward hours since it s most often how tz offsets are communicated redundancy is inevitable since you could just as well store everything as seconds origin with an uncertainty what makes these specific fields questionable is that they are rarely used in isolation and could easily be recovered from the reduced forms collapsing these would simplify some of the formatting reduce the object size and simplify the data structure slightly | 0 |
233,487 | 18,990,772,696 | IssuesEvent | 2021-11-22 06:55:35 | stylelint/vscode-stylelint | https://api.github.com/repos/stylelint/vscode-stylelint | closed | Non-custom LSP event name | Suggestion Refactor Tests | The custom event defined here:
https://github.com/stylelint/vscode-stylelint/blob/b07fb506452b720a28e0f3fb01f48fc5c27becbf/src/server/types.ts#L34
is named `textDocument/didRegisterDocumentFormattingEditProvider`. I wonder why not `stylelint/didRegisterDocumentFormattingEditProvider` to make it obvious that it's a custom one?
Not really causing any issues. Would just expect the events in the `textDocument` namespace to be valid per spec. | 1.0 | Non-custom LSP event name - The custom event defined here:
https://github.com/stylelint/vscode-stylelint/blob/b07fb506452b720a28e0f3fb01f48fc5c27becbf/src/server/types.ts#L34
is named `textDocument/didRegisterDocumentFormattingEditProvider`. I wonder why not `stylelint/didRegisterDocumentFormattingEditProvider` to make it obvious that it's a custom one?
Not really causing any issues. Would just expect the events in the `textDocument` namespace to be valid per spec. | non_priority | non custom lsp event name the custom event defined here is named textdocument didregisterdocumentformattingeditprovider i wonder why not stylelint didregisterdocumentformattingeditprovider to make it obvious that it s a custom one not really causing any issues would just expect the events in the textdocument namespace to be valid per spec | 0 |
30,166 | 14,441,679,194 | IssuesEvent | 2020-12-07 17:06:05 | frontendbr/forum | https://api.github.com/repos/frontendbr/forum | closed | Ferramentas para medir e melhorar performance de apps híbridos (cordova e ionic) | Performance [Dúvida] | Fala galera beleza?! Vocês conhecem alguma ferramenta ou forma de medir e melhorar performance de apps híbridos? | True | Ferramentas para medir e melhorar performance de apps híbridos (cordova e ionic) - Fala galera beleza?! Vocês conhecem alguma ferramenta ou forma de medir e melhorar performance de apps híbridos? | non_priority | ferramentas para medir e melhorar performance de apps híbridos cordova e ionic fala galera beleza vocês conhecem alguma ferramenta ou forma de medir e melhorar performance de apps híbridos | 0 |
40,379 | 16,472,771,410 | IssuesEvent | 2021-05-23 18:50:42 | ErickCR12/GymTEC | https://api.github.com/repos/ErickCR12/GymTEC | opened | Tarea 3 | API Service Funcionalidad 21: Gestión de tipos | Crear una solicitud POST que recibe toda la información necesaria para la agregación de un tipo de planilla a la base de datos SQL. De manera default se tienen los tipos de planilla Pago mensual, pago por hora y pago por clase. | 1.0 | Tarea 3 - Crear una solicitud POST que recibe toda la información necesaria para la agregación de un tipo de planilla a la base de datos SQL. De manera default se tienen los tipos de planilla Pago mensual, pago por hora y pago por clase. | non_priority | tarea crear una solicitud post que recibe toda la información necesaria para la agregación de un tipo de planilla a la base de datos sql de manera default se tienen los tipos de planilla pago mensual pago por hora y pago por clase | 0 |
9,178 | 12,227,034,803 | IssuesEvent | 2020-05-03 13:38:53 | jyn514/rcc | https://api.github.com/repos/jyn514/rcc | opened | [ICE] preprocessor needs a recursion guard | ICE fuzz preprocessor | ### Code
<!-- The code that caused the panic goes here.
This should also include the error message you got. -->
Modified from the gcc torture suite.
```c
/* Add a few "extern int Xxxxxx ();" declarations. */
#define DEF(x) extern int x;
#define LIM1(x) DEF(x##0); DEF(x##1); DEF(x##2); DEF(x##3); DEF(x##4); \
DEF(x##5); DEF(x##6); DEF(x##7); DEF(x##8); DEF(x##9);
#define LIM2(x) LIM1(x##0) LIM1(x##1) LIM1(x##2) LIM1(x##3) LIM1(x##4) \
LIM1(x##5) LIM1(x##6) LIM1(x##7) LIM1(x##8) LIM1(x##9)
#define LIM3(x) LIM2(x##0) LIM2(x##1) LIM2(x##2) LIM2(x##3) LIM2(x##4) \
LIM2(x##5) LIM2(x##6) LIM2(x##7) LIM2(x##8) LIM2(x##9)
#define LIM4(x) LIM3(x##0) LIM3(x##1) LIM3(x##2) LIM3(x##3) LIM3(x##4) \
LIM3(x##5) LIM3(x##6) LIM3(x##7) LIM3(x##8) LIM3(x##9)
LIM4 (X);
thread 'main' has overflowed its stack
fatal runtime error: stack overflow
Aborted (core dumped)
```
### Expected behavior
<!-- A description of what you expected to happen.
If you're not sure (e.g. this is invalid code),
paste the output of another compiler
(I like `clang -x c - -Wall -Wextra -pedantic`) -->
It should give a fatal error and exit normally, like for nested expressions.
<details><summary>Backtrace</summary>
<!-- The output of `RUST_BACKTRACE=1 cargo run` goes here. -->
```
Program received signal SIGSEGV, Segmentation fault.
0x0000555555a16c41 in rcc::lex::cpp::PreProcessor::next_replacement_token (
self=0x7fffffff9760) at src/lex/cpp.rs:281
281 if let Some(replacement) = self.pending.pop_front() {
(gdb) where
#0 0x0000555555a16c41 in rcc::lex::cpp::PreProcessor::next_replacement_token (
self=0x7fffffff9760) at src/lex/cpp.rs:281
#1 0x0000555555a1707b in rcc::lex::cpp::PreProcessor::match_next (
self=0x7fffffff9760, token=...) at src/lex/cpp.rs:292
#2 0x0000555555a1aece in rcc::lex::cpp::PreProcessor::replace_function (
self=0x7fffffff9760, name=..., start=613) at src/lex/cpp.rs:708
#3 0x0000555555a1aba2 in rcc::lex::cpp::PreProcessor::replace_id (
self=0x7fffffff9760, name=..., location=...) at src/lex/cpp.rs:696
#4 0x0000555555a17a1c in rcc::lex::cpp::PreProcessor::handle_token (
self=0x7fffffff9760, token=..., location=...) at src/lex/cpp.rs:342
#5 0x0000555555a16533 in <rcc::lex::cpp::PreProcessor as core::iter::traits::iterator::Iterator>::next (self=0x7fffffff9760) at src/lex/cpp.rs:240
#6 0x0000555555a1b85d in rcc::lex::cpp::PreProcessor::replace_function (
self=0x7fffffff9760, name=..., start=613) at src/lex/cpp.rs:775
#7 0x0000555555a1aba2 in rcc::lex::cpp::PreProcessor::replace_id (
self=0x7fffffff9760, name=..., location=...) at src/lex/cpp.rs:696
#8 0x0000555555a17a1c in rcc::lex::cpp::PreProcessor::handle_token (
self=0x7fffffff9760, token=..., location=...) at src/lex/cpp.rs:342
... several thousand more lines ...
```
</details>
| 1.0 | [ICE] preprocessor needs a recursion guard - ### Code
<!-- The code that caused the panic goes here.
This should also include the error message you got. -->
Modified from the gcc torture suite.
```c
/* Add a few "extern int Xxxxxx ();" declarations. */
#define DEF(x) extern int x;
#define LIM1(x) DEF(x##0); DEF(x##1); DEF(x##2); DEF(x##3); DEF(x##4); \
DEF(x##5); DEF(x##6); DEF(x##7); DEF(x##8); DEF(x##9);
#define LIM2(x) LIM1(x##0) LIM1(x##1) LIM1(x##2) LIM1(x##3) LIM1(x##4) \
LIM1(x##5) LIM1(x##6) LIM1(x##7) LIM1(x##8) LIM1(x##9)
#define LIM3(x) LIM2(x##0) LIM2(x##1) LIM2(x##2) LIM2(x##3) LIM2(x##4) \
LIM2(x##5) LIM2(x##6) LIM2(x##7) LIM2(x##8) LIM2(x##9)
#define LIM4(x) LIM3(x##0) LIM3(x##1) LIM3(x##2) LIM3(x##3) LIM3(x##4) \
LIM3(x##5) LIM3(x##6) LIM3(x##7) LIM3(x##8) LIM3(x##9)
LIM4 (X);
thread 'main' has overflowed its stack
fatal runtime error: stack overflow
Aborted (core dumped)
```
### Expected behavior
<!-- A description of what you expected to happen.
If you're not sure (e.g. this is invalid code),
paste the output of another compiler
(I like `clang -x c - -Wall -Wextra -pedantic`) -->
It should give a fatal error and exit normally, like for nested expressions.
<details><summary>Backtrace</summary>
<!-- The output of `RUST_BACKTRACE=1 cargo run` goes here. -->
```
Program received signal SIGSEGV, Segmentation fault.
0x0000555555a16c41 in rcc::lex::cpp::PreProcessor::next_replacement_token (
self=0x7fffffff9760) at src/lex/cpp.rs:281
281 if let Some(replacement) = self.pending.pop_front() {
(gdb) where
#0 0x0000555555a16c41 in rcc::lex::cpp::PreProcessor::next_replacement_token (
self=0x7fffffff9760) at src/lex/cpp.rs:281
#1 0x0000555555a1707b in rcc::lex::cpp::PreProcessor::match_next (
self=0x7fffffff9760, token=...) at src/lex/cpp.rs:292
#2 0x0000555555a1aece in rcc::lex::cpp::PreProcessor::replace_function (
self=0x7fffffff9760, name=..., start=613) at src/lex/cpp.rs:708
#3 0x0000555555a1aba2 in rcc::lex::cpp::PreProcessor::replace_id (
self=0x7fffffff9760, name=..., location=...) at src/lex/cpp.rs:696
#4 0x0000555555a17a1c in rcc::lex::cpp::PreProcessor::handle_token (
self=0x7fffffff9760, token=..., location=...) at src/lex/cpp.rs:342
#5 0x0000555555a16533 in <rcc::lex::cpp::PreProcessor as core::iter::traits::iterator::Iterator>::next (self=0x7fffffff9760) at src/lex/cpp.rs:240
#6 0x0000555555a1b85d in rcc::lex::cpp::PreProcessor::replace_function (
self=0x7fffffff9760, name=..., start=613) at src/lex/cpp.rs:775
#7 0x0000555555a1aba2 in rcc::lex::cpp::PreProcessor::replace_id (
self=0x7fffffff9760, name=..., location=...) at src/lex/cpp.rs:696
#8 0x0000555555a17a1c in rcc::lex::cpp::PreProcessor::handle_token (
self=0x7fffffff9760, token=..., location=...) at src/lex/cpp.rs:342
... several thousand more lines ...
```
</details>
| non_priority | preprocessor needs a recursion guard code the code that caused the panic goes here this should also include the error message you got modified from the gcc torture suite c add a few extern int xxxxxx declarations define def x extern int x define x def x def x def x def x def x def x def x def x def x def x define x x x x x x x x x x x define x x x x x x x x x x x define x x x x x x x x x x x x thread main has overflowed its stack fatal runtime error stack overflow aborted core dumped expected behavior a description of what you expected to happen if you re not sure e g this is invalid code paste the output of another compiler i like clang x c wall wextra pedantic it should give a fatal error and exit normally like for nested expressions backtrace program received signal sigsegv segmentation fault in rcc lex cpp preprocessor next replacement token self at src lex cpp rs if let some replacement self pending pop front gdb where in rcc lex cpp preprocessor next replacement token self at src lex cpp rs in rcc lex cpp preprocessor match next self token at src lex cpp rs in rcc lex cpp preprocessor replace function self name start at src lex cpp rs in rcc lex cpp preprocessor replace id self name location at src lex cpp rs in rcc lex cpp preprocessor handle token self token location at src lex cpp rs in next self at src lex cpp rs in rcc lex cpp preprocessor replace function self name start at src lex cpp rs in rcc lex cpp preprocessor replace id self name location at src lex cpp rs in rcc lex cpp preprocessor handle token self token location at src lex cpp rs several thousand more lines | 0 |
13,823 | 16,587,868,162 | IssuesEvent | 2021-06-01 01:24:09 | lsmacedo/spotifyt-back-end | https://api.github.com/repos/lsmacedo/spotifyt-back-end | opened | Rodar scripts de pre-commit e pre-push | process | Pre-commit: rodar lint
Pre-push: rodar testes unitários e E2E | 1.0 | Rodar scripts de pre-commit e pre-push - Pre-commit: rodar lint
Pre-push: rodar testes unitários e E2E | non_priority | rodar scripts de pre commit e pre push pre commit rodar lint pre push rodar testes unitários e | 0 |
145,153 | 13,144,820,935 | IssuesEvent | 2020-08-08 00:18:34 | hackforla/civic-opportunity | https://api.github.com/repos/hackforla/civic-opportunity | opened | GA Modeling Analysis | documentation special projects | ### Overview
Establish setup for modeling analysis
### Action Items
### Resources/Instructions
| 1.0 | GA Modeling Analysis - ### Overview
Establish setup for modeling analysis
### Action Items
### Resources/Instructions
| non_priority | ga modeling analysis overview establish setup for modeling analysis action items resources instructions | 0 |
303,316 | 26,198,913,878 | IssuesEvent | 2023-01-03 15:45:47 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: backup/mixed-version-basic failed | C-test-failure O-robot O-roachtest release-blocker T-disaster-recovery branch-release-22.2 | roachtest.backup/mixed-version-basic [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/backup/mixed-version-basic) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1):
```
test artifacts and logs in: /artifacts/backup/mixed-version-basic/run_1
(test_impl.go:292).Fatalf: job failed: Use of BACKUP with incremental requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out.
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #92445 roachtest: backup/mixed-version-basic failed [C-test-failure O-roachtest O-robot T-disaster-recovery branch-master release-blocker]
</p>
</details>
/cc @cockroachdb/disaster-recovery
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*backup/mixed-version-basic.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-22907 | 2.0 | roachtest: backup/mixed-version-basic failed - roachtest.backup/mixed-version-basic [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/8147282?buildTab=artifacts#/backup/mixed-version-basic) on release-22.2 @ [07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1](https://github.com/cockroachdb/cockroach/commits/07a53a36601e9ca5fcffcff55f69b43c6dfbf1c1):
```
test artifacts and logs in: /artifacts/backup/mixed-version-basic/run_1
(test_impl.go:292).Fatalf: job failed: Use of BACKUP with incremental requires an enterprise license. Your evaluation license expired on December 30, 2022. If you're interested in getting a new license, please contact subscriptions@cockroachlabs.com and we can help you out.
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=true</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #92445 roachtest: backup/mixed-version-basic failed [C-test-failure O-roachtest O-robot T-disaster-recovery branch-master release-blocker]
</p>
</details>
/cc @cockroachdb/disaster-recovery
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*backup/mixed-version-basic.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-22907 | non_priority | roachtest backup mixed version basic failed roachtest backup mixed version basic with on release test artifacts and logs in artifacts backup mixed version basic run test impl go fatalf job failed use of backup with incremental requires an enterprise license your evaluation license expired on december if you re interested in getting a new license please contact subscriptions cockroachlabs com and we can help you out parameters roachtest cloud gce roachtest cpu roachtest encrypted true roachtest ssd help see see same failure on other branches roachtest backup mixed version basic failed cc cockroachdb disaster recovery jira issue crdb | 0 |
122,324 | 16,105,590,485 | IssuesEvent | 2021-04-27 14:35:57 | MozillaFoundation/Design | https://api.github.com/repos/MozillaFoundation/Design | closed | [Email templates] Prep Figma File for Dev | MarketingOps design | Prepping Figma file to include necessary measurements and annotations and export any images/icons. Might have to clarify further with Will on what he needs.
| 1.0 | [Email templates] Prep Figma File for Dev - Prepping Figma file to include necessary measurements and annotations and export any images/icons. Might have to clarify further with Will on what he needs.
| non_priority | prep figma file for dev prepping figma file to include necessary measurements and annotations and export any images icons might have to clarify further with will on what he needs | 0 |
298,814 | 25,859,202,644 | IssuesEvent | 2022-12-13 15:46:59 | Open-Attestation/statuspage | https://api.github.com/repos/Open-Attestation/statuspage | closed | 🛑 OpenAttestation is down | status open-attestation | In [`3c76dad`](https://github.com/Open-Attestation/statuspage/commit/3c76dadc1fa349dd898b18ac7232e3b8f026e113
), OpenAttestation (https://www.openattestation.com) was **down**:
- HTTP code: 503
- Response time: 5112 ms
| 1.0 | 🛑 OpenAttestation is down - In [`3c76dad`](https://github.com/Open-Attestation/statuspage/commit/3c76dadc1fa349dd898b18ac7232e3b8f026e113
), OpenAttestation (https://www.openattestation.com) was **down**:
- HTTP code: 503
- Response time: 5112 ms
| non_priority | 🛑 openattestation is down in openattestation was down http code response time ms | 0 |
250,536 | 27,099,278,908 | IssuesEvent | 2023-02-15 07:09:10 | ilan-WS/cloudify-system-tests | https://api.github.com/repos/ilan-WS/cloudify-system-tests | opened | CVE-2023-0286 (High) detected in cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl | security vulnerability | ## CVE-2023-0286 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</b></p></summary>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl">https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt,/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There is a type confusion vulnerability relating to X.400 address processing inside an X.509 GeneralName. X.400 addresses were parsed as an ASN1_STRING but the public structure definition for GENERAL_NAME incorrectly specified the type of the x400Address field as ASN1_TYPE. This field is subsequently interpreted by the OpenSSL function GENERAL_NAME_cmp as an ASN1_TYPE rather than an ASN1_STRING. When CRL checking is enabled (i.e. the application sets the X509_V_FLAG_CRL_CHECK flag), this vulnerability may allow an attacker to pass arbitrary pointers to a memcmp call, enabling them to read memory contents or enact a denial of service. In most cases, the attack requires the attacker to provide both the certificate chain and CRL, neither of which need to have a valid signature. If the attacker only controls one of these inputs, the other input must already contain an X.400 address as a CRL distribution point, which is uncommon. As such, this vulnerability is most likely to only affect applications which have implemented their own functionality for retrieving CRLs over a network.
<p>Publish Date: 2023-02-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0286>CVE-2023-0286</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openssl.org/news/vulnerabilities.html">https://www.openssl.org/news/vulnerabilities.html</a></p>
<p>Release Date: 2023-02-08</p>
<p>Fix Resolution: openssl-3.0.8</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| True | CVE-2023-0286 (High) detected in cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl - ## CVE-2023-0286 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</b></p></summary>
<p>cryptography is a package which provides cryptographic recipes and primitives to Python developers.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl">https://files.pythonhosted.org/packages/b2/26/7af637e6a7e87258b963f1731c5982fb31cd507f0d90d91836e446955d02/cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt,/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **cryptography-3.4.7-cp36-abi3-manylinux2014_x86_64.whl** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
There is a type confusion vulnerability relating to X.400 address processing inside an X.509 GeneralName. X.400 addresses were parsed as an ASN1_STRING but the public structure definition for GENERAL_NAME incorrectly specified the type of the x400Address field as ASN1_TYPE. This field is subsequently interpreted by the OpenSSL function GENERAL_NAME_cmp as an ASN1_TYPE rather than an ASN1_STRING. When CRL checking is enabled (i.e. the application sets the X509_V_FLAG_CRL_CHECK flag), this vulnerability may allow an attacker to pass arbitrary pointers to a memcmp call, enabling them to read memory contents or enact a denial of service. In most cases, the attack requires the attacker to provide both the certificate chain and CRL, neither of which need to have a valid signature. If the attacker only controls one of these inputs, the other input must already contain an X.400 address as a CRL distribution point, which is uncommon. As such, this vulnerability is most likely to only affect applications which have implemented their own functionality for retrieving CRLs over a network.
<p>Publish Date: 2023-02-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-0286>CVE-2023-0286</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.openssl.org/news/vulnerabilities.html">https://www.openssl.org/news/vulnerabilities.html</a></p>
<p>Release Date: 2023-02-08</p>
<p>Fix Resolution: openssl-3.0.8</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
| non_priority | cve high detected in cryptography whl cve high severity vulnerability vulnerable library cryptography whl cryptography is a package which provides cryptographic recipes and primitives to python developers library home page a href path to dependency file requirements txt path to vulnerable library requirements txt requirements txt dependency hierarchy x cryptography whl vulnerable library found in base branch master vulnerability details there is a type confusion vulnerability relating to x address processing inside an x generalname x addresses were parsed as an string but the public structure definition for general name incorrectly specified the type of the field as type this field is subsequently interpreted by the openssl function general name cmp as an type rather than an string when crl checking is enabled i e the application sets the v flag crl check flag this vulnerability may allow an attacker to pass arbitrary pointers to a memcmp call enabling them to read memory contents or enact a denial of service in most cases the attack requires the attacker to provide both the certificate chain and crl neither of which need to have a valid signature if the attacker only controls one of these inputs the other input must already contain an x address as a crl distribution point which is uncommon as such this vulnerability is most likely to only affect applications which have implemented their own functionality for retrieving crls over a network publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution openssl check this box to open an automated fix pr | 0 |
5,888 | 3,690,331,745 | IssuesEvent | 2016-02-25 19:36:45 | egel/dotfiles | https://api.github.com/repos/egel/dotfiles | closed | add mac support | building process enhancement | add installation software for macs
- check is brew installed (throw error)
- editorconfig | 1.0 | add mac support - add installation software for macs
- check is brew installed (throw error)
- editorconfig | non_priority | add mac support add installation software for macs check is brew installed throw error editorconfig | 0 |
83,299 | 10,324,880,860 | IssuesEvent | 2019-09-01 13:04:32 | SAP/fundamental-ngx | https://api.github.com/repos/SAP/fundamental-ngx | closed | documentation - enhance the Reactive Form examples | documentation | The idea is to enhance the Reactive Forms examples (one enabled and one disabled) and list the states underneath.
- [ ] - components to review
- [ ] - calendar
- [ ] - checkbox
- [ ] - combobox
- [ ] - datepicker
- [ ] - datetimepicker
- [ ] - input
- [ ] - inputGroup
- [ ] - localizationEditor
- [ ] - multiInput
- [ ] - radioButton
- [ ] - select
- [ ] - native Select
- [ ] - textArea
- [ ] - time
- [ ] - timepicker
- [ ] - toggle
- [ ] - Unify the documentation for all the components
- [ ] - add enabled and disabled example | 1.0 | documentation - enhance the Reactive Form examples - The idea is to enhance the Reactive Forms examples (one enabled and one disabled) and list the states underneath.
- [ ] - components to review
- [ ] - calendar
- [ ] - checkbox
- [ ] - combobox
- [ ] - datepicker
- [ ] - datetimepicker
- [ ] - input
- [ ] - inputGroup
- [ ] - localizationEditor
- [ ] - multiInput
- [ ] - radioButton
- [ ] - select
- [ ] - native Select
- [ ] - textArea
- [ ] - time
- [ ] - timepicker
- [ ] - toggle
- [ ] - Unify the documentation for all the components
- [ ] - add enabled and disabled example | non_priority | documentation enhance the reactive form examples the idea is to enhance the reactive forms examples one enabled and one disabled and list the states underneath components to review calendar checkbox combobox datepicker datetimepicker input inputgroup localizationeditor multiinput radiobutton select native select textarea time timepicker toggle unify the documentation for all the components add enabled and disabled example | 0 |
363,900 | 25,473,133,316 | IssuesEvent | 2022-11-25 12:01:49 | fetchai/ledger-subquery | https://api.github.com/repos/fetchai/ledger-subquery | closed | Add issue templates | C-documentation | ## Acceptance criteria
Following the example set in the [cosmpy](https://github.com/fetchai/cosmpy/tree/master/.github/ISSUE_TEMPLATE) repo, this repo should have distinct issue templates for bug reports and feature requests. | 1.0 | Add issue templates - ## Acceptance criteria
Following the example set in the [cosmpy](https://github.com/fetchai/cosmpy/tree/master/.github/ISSUE_TEMPLATE) repo, this repo should have distinct issue templates for bug reports and feature requests. | non_priority | add issue templates acceptance criteria following the example set in the repo this repo should have distinct issue templates for bug reports and feature requests | 0 |
104,463 | 8,972,723,675 | IssuesEvent | 2019-01-29 19:03:25 | MicrosoftDocs/visualstudio-docs | https://api.github.com/repos/MicrosoftDocs/visualstudio-docs | closed | Videos on Coded UI test walk through page cannot be played | area - test doc feedback easy | The issue is on the following page.
https://docs.microsoft.com/en-us/visualstudio/test/walkthrough-creating-editing-and-maintaining-a-coded-ui-test
The video links do not work. Clicking the link results in "HTTP Error 503. The service is unavailable."
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 449c8c80-5302-0d8b-c5a0-04b56ccb4f97
* Version Independent ID: 16c81ae7-bda9-6599-e64c-a8f73602513e
* Content: [Create a Coded UI Test in Visual Studio - Visual Studio](https://docs.microsoft.com/en-us/visualstudio/test/walkthrough-creating-editing-and-maintaining-a-coded-ui-test)
* Content Source: [docs/test/walkthrough-creating-editing-and-maintaining-a-coded-ui-test.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/master/docs/test/walkthrough-creating-editing-and-maintaining-a-coded-ui-test.md)
* Product: **visual-studio-dev15**
* GitHub Login: @gewarren
* Microsoft Alias: **gewarren** | 1.0 | Videos on Coded UI test walk through page cannot be played - The issue is on the following page.
https://docs.microsoft.com/en-us/visualstudio/test/walkthrough-creating-editing-and-maintaining-a-coded-ui-test
The video links do not work. Clicking the link results in "HTTP Error 503. The service is unavailable."
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: 449c8c80-5302-0d8b-c5a0-04b56ccb4f97
* Version Independent ID: 16c81ae7-bda9-6599-e64c-a8f73602513e
* Content: [Create a Coded UI Test in Visual Studio - Visual Studio](https://docs.microsoft.com/en-us/visualstudio/test/walkthrough-creating-editing-and-maintaining-a-coded-ui-test)
* Content Source: [docs/test/walkthrough-creating-editing-and-maintaining-a-coded-ui-test.md](https://github.com/MicrosoftDocs/visualstudio-docs/blob/master/docs/test/walkthrough-creating-editing-and-maintaining-a-coded-ui-test.md)
* Product: **visual-studio-dev15**
* GitHub Login: @gewarren
* Microsoft Alias: **gewarren** | non_priority | videos on coded ui test walk through page cannot be played the issue is on the following page the video links do not work clicking the link results in http error the service is unavailable document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product visual studio github login gewarren microsoft alias gewarren | 0 |
79,935 | 29,661,542,958 | IssuesEvent | 2023-06-10 08:04:46 | vector-im/element-web | https://api.github.com/repos/vector-im/element-web | closed | Element creates room on wrong server | T-Defect | ### Steps to reproduce
# First prepare the following setting:
* On homeserver HS create user U.
* On homeserver SS create space S and invite user U to it.
* User U accepts the invitation to space S.
* On homeserver SS grant administration rights for space S to user U.
# Then reproduce the bug like this:
* User U logs in to homeserver HS using Element.
* User U navigates Element to the overview page of space S.
* User U clicks the "Add" button on the overview page of space S and invokes the "New Room" menu item to create the new room R.
### Outcome
# Expected Result
* The room R is created on homeserver SS, because it was created using the "New Room" button on the overview page of space S, and space S is hosted on homeserver SS. The benefit is that the room R will survive the eventual dismission of homeserver HS.
# Actual Result
* The room R is created on homeserver HS, because it was created by User U, and user U's homeserver is HS. The drawback is that the room R will become unusable at the eventual dismission of homeserver HS.
### Operating system
Linux
### Browser information
Chrome
### URL for webapp
Latest
### Application version
Version Element: 1.11.32, Version Olm: 3.2.14
### Homeserver
Synapse (Latest)
### Will you send logs?
No | 1.0 | Element creates room on wrong server - ### Steps to reproduce
# First prepare the following setting:
* On homeserver HS create user U.
* On homeserver SS create space S and invite user U to it.
* User U accepts the invitation to space S.
* On homeserver SS grant administration rights for space S to user U.
# Then reproduce the bug like this:
* User U logs in to homeserver HS using Element.
* User U navigates Element to the overview page of space S.
* User U clicks the "Add" button on the overview page of space S and invokes the "New Room" menu item to create the new room R.
### Outcome
# Expected Result
* The room R is created on homeserver SS, because it was created using the "New Room" button on the overview page of space S, and space S is hosted on homeserver SS. The benefit is that the room R will survive the eventual dismission of homeserver HS.
# Actual Result
* The room R is created on homeserver HS, because it was created by User U, and user U's homeserver is HS. The drawback is that the room R will become unusable at the eventual dismission of homeserver HS.
### Operating system
Linux
### Browser information
Chrome
### URL for webapp
Latest
### Application version
Version Element: 1.11.32, Version Olm: 3.2.14
### Homeserver
Synapse (Latest)
### Will you send logs?
No | non_priority | element creates room on wrong server steps to reproduce first prepare the following setting on homeserver hs create user u on homeserver ss create space s and invite user u to it user u accepts the invitation to space s on homeserver ss grant administration rights for space s to user u then reproduce the bug like this user u logs in to homeserver hs using element user u navigates element to the overview page of space s user u clicks the add button on the overview page of space s and invokes the new room menu item to create the new room r outcome expected result the room r is created on homeserver ss because it was created using the new room button on the overview page of space s and space s is hosted on homeserver ss the benefit is that the room r will survive the eventual dismission of homeserver hs actual result the room r is created on homeserver hs because it was created by user u and user u s homeserver is hs the drawback is that the room r will become unusable at the eventual dismission of homeserver hs operating system linux browser information chrome url for webapp latest application version version element version olm homeserver synapse latest will you send logs no | 0 |
170,507 | 14,262,329,767 | IssuesEvent | 2020-11-20 12:48:07 | SchlossLab/mikropml | https://api.github.com/repos/SchlossLab/mikropml | closed | Preprocessing Data Vignette text correction | documentation | In the binary data example, the text says the preprocessing output is list of length two: `dat_transformed` and `grp_feats` but the example output shows a third item `removed_feats` that is not mentioned in the text just below the example. The text may need updating to reflect the third item in the output. | 1.0 | Preprocessing Data Vignette text correction - In the binary data example, the text says the preprocessing output is list of length two: `dat_transformed` and `grp_feats` but the example output shows a third item `removed_feats` that is not mentioned in the text just below the example. The text may need updating to reflect the third item in the output. | non_priority | preprocessing data vignette text correction in the binary data example the text says the preprocessing output is list of length two dat transformed and grp feats but the example output shows a third item removed feats that is not mentioned in the text just below the example the text may need updating to reflect the third item in the output | 0 |
45,521 | 24,089,973,892 | IssuesEvent | 2022-09-19 14:03:35 | irods/irods | https://api.github.com/repos/irods/irods | closed | Is these Iget and Iput speed correct? | resolved/invalid question performance | I got 2 SSD with read and write speed around 2500 MB/s. They are mounted as /mnt/client and /mnt/server.
I create a vault in /mnt/server and create resource threre. iRODS is run in local host.
I got the this speed with N = 4, file size = 25 GB :
`iput speed = 974.424 MB/s`
`iget speed = 887.365 MB/s
`
Do you think this speed is normal or I can do something to optimize it further more? | True | Is these Iget and Iput speed correct? - I got 2 SSD with read and write speed around 2500 MB/s. They are mounted as /mnt/client and /mnt/server.
I create a vault in /mnt/server and create resource threre. iRODS is run in local host.
I got the this speed with N = 4, file size = 25 GB :
`iput speed = 974.424 MB/s`
`iget speed = 887.365 MB/s
`
Do you think this speed is normal or I can do something to optimize it further more? | non_priority | is these iget and iput speed correct i got ssd with read and write speed around mb s they are mounted as mnt client and mnt server i create a vault in mnt server and create resource threre irods is run in local host i got the this speed with n file size gb iput speed mb s iget speed mb s do you think this speed is normal or i can do something to optimize it further more | 0 |
9,421 | 6,888,853,182 | IssuesEvent | 2017-11-22 08:12:27 | zcash/zcash | https://api.github.com/repos/zcash/zcash | opened | Investigate z_shieldcoinbase usage of AvailableCoins | bug performance wallet | There may be a degradation in coin selection performance of ~10% due to two calls to AvailableCoins. | True | Investigate z_shieldcoinbase usage of AvailableCoins - There may be a degradation in coin selection performance of ~10% due to two calls to AvailableCoins. | non_priority | investigate z shieldcoinbase usage of availablecoins there may be a degradation in coin selection performance of due to two calls to availablecoins | 0 |
137,553 | 11,140,848,181 | IssuesEvent | 2019-12-21 17:34:08 | fastlane/fastlane | https://api.github.com/repos/fastlane/fastlane | closed | Dry run mode for Actions.sh | status: waiting-for-reply topic: tests | Trying to improve the testing framework, but also make it less intrusive for the source code. Inspired by the discussions in #14471.
One idea is to add a new parameter to `Actions.sh` which will be used in tests and when running `fastlane --dry-run`:
### A. Existing code
```ruby
# do not need to update existing code to keep previous behavior
Actions.sh('git rev-list HEAD')
# or explicitly
Actions.sh('git rev-list HEAD', dry_run: :return_commandline)
```
This will run the command in production, and return 'git rev-parse HEAD' in dry-run / tests, as it is now.
### B. Dry-run supporting code with side effects
```ruby
Actions.sh('rm -v .checkpoint', dry_run: "removed '.checkpoint'\n")
```
This command will not run in testing / dry-run mode, but `Actions.sh` will return `removed '.checkpoint'` instead of the command itself.
### C. Dry-run supporting code without side effects / hermetic
```ruby
Actions.sh('git rev-parse HEAD', dry_run: :no_side_effects)
```
**This is most useful for tests**. We run the command in production / tests / dry-run, but expect no side effects from it.
---
I'm interested in feedback from fastlane team, and I am happy to start implementing a PoC if you like the idea. | 1.0 | Dry run mode for Actions.sh - Trying to improve the testing framework, but also make it less intrusive for the source code. Inspired by the discussions in #14471.
One idea is to add a new parameter to `Actions.sh` which will be used in tests and when running `fastlane --dry-run`:
### A. Existing code
```ruby
# do not need to update existing code to keep previous behavior
Actions.sh('git rev-list HEAD')
# or explicitly
Actions.sh('git rev-list HEAD', dry_run: :return_commandline)
```
This will run the command in production, and return 'git rev-parse HEAD' in dry-run / tests, as it is now.
### B. Dry-run supporting code with side effects
```ruby
Actions.sh('rm -v .checkpoint', dry_run: "removed '.checkpoint'\n")
```
This command will not run in testing / dry-run mode, but `Actions.sh` will return `removed '.checkpoint'` instead of the command itself.
### C. Dry-run supporting code without side effects / hermetic
```ruby
Actions.sh('git rev-parse HEAD', dry_run: :no_side_effects)
```
**This is most useful for tests**. We run the command in production / tests / dry-run, but expect no side effects from it.
---
I'm interested in feedback from fastlane team, and I am happy to start implementing a PoC if you like the idea. | non_priority | dry run mode for actions sh trying to improve the testing framework but also make it less intrusive for the source code inspired by the discussions in one idea is to add a new parameter to actions sh which will be used in tests and when running fastlane dry run a existing code ruby do not need to update existing code to keep previous behavior actions sh git rev list head or explicitly actions sh git rev list head dry run return commandline this will run the command in production and return git rev parse head in dry run tests as it is now b dry run supporting code with side effects ruby actions sh rm v checkpoint dry run removed checkpoint n this command will not run in testing dry run mode but actions sh will return removed checkpoint instead of the command itself c dry run supporting code without side effects hermetic ruby actions sh git rev parse head dry run no side effects this is most useful for tests we run the command in production tests dry run but expect no side effects from it i m interested in feedback from fastlane team and i am happy to start implementing a poc if you like the idea | 0 |
202,662 | 15,294,544,666 | IssuesEvent | 2021-02-24 02:43:42 | ampproject/amphtml | https://api.github.com/repos/ampproject/amphtml | closed | Video docking visual tests are flaky due to scrollbars | P2: Soon Related to: Flaky Tests WG: components | For example, https://percy.io/ampproject/amphtml/builds/8998973/changed/510237008
For more flaky builds, go to https://percy.io/ampproject/amphtml, search for the text "auto-approved", and click through to the diffs.
Right now, this is the main cause of failing CI builds. | 1.0 | Video docking visual tests are flaky due to scrollbars - For example, https://percy.io/ampproject/amphtml/builds/8998973/changed/510237008
For more flaky builds, go to https://percy.io/ampproject/amphtml, search for the text "auto-approved", and click through to the diffs.
Right now, this is the main cause of failing CI builds. | non_priority | video docking visual tests are flaky due to scrollbars for example for more flaky builds go to search for the text auto approved and click through to the diffs right now this is the main cause of failing ci builds | 0 |
139,921 | 18,869,682,209 | IssuesEvent | 2021-11-13 01:06:48 | samq-ghdemo/Forrester-Demo | https://api.github.com/repos/samq-ghdemo/Forrester-Demo | closed | WS-2017-3772 (High) detected in underscore.string-3.3.5.tgz - autoclosed | security vulnerability | ## WS-2017-3772 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore.string-3.3.5.tgz</b></p></summary>
<p>String manipulation extensions for Underscore.js javascript library.</p>
<p>Library home page: <a href="https://registry.npmjs.org/underscore.string/-/underscore.string-3.3.5.tgz">https://registry.npmjs.org/underscore.string/-/underscore.string-3.3.5.tgz</a></p>
<p>Path to dependency file: Forrester-Demo/package.json</p>
<p>Path to vulnerable library: Forrester-Demo/node_modules/underscore.string/package.json</p>
<p>
Dependency Hierarchy:
- grunt-jsbeautifier-0.2.13.tgz (Root Library)
- :x: **underscore.string-3.3.5.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Regular Expression Denial of Service (ReDoS) vulnerability was found in underscore.string 2.4.0 through 3.3.5.
<p>Publish Date: 2017-09-08
<p>URL: <a href=https://github.com/esamattis/underscore.string/issues/510>WS-2017-3772</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"underscore.string","packageVersion":"3.3.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-jsbeautifier:0.2.13;underscore.string:3.3.5","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-3772","vulnerabilityDetails":"Regular Expression Denial of Service (ReDoS) vulnerability was found in underscore.string 2.4.0 through 3.3.5.","vulnerabilityUrl":"https://github.com/esamattis/underscore.string/issues/510","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | WS-2017-3772 (High) detected in underscore.string-3.3.5.tgz - autoclosed - ## WS-2017-3772 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>underscore.string-3.3.5.tgz</b></p></summary>
<p>String manipulation extensions for Underscore.js javascript library.</p>
<p>Library home page: <a href="https://registry.npmjs.org/underscore.string/-/underscore.string-3.3.5.tgz">https://registry.npmjs.org/underscore.string/-/underscore.string-3.3.5.tgz</a></p>
<p>Path to dependency file: Forrester-Demo/package.json</p>
<p>Path to vulnerable library: Forrester-Demo/node_modules/underscore.string/package.json</p>
<p>
Dependency Hierarchy:
- grunt-jsbeautifier-0.2.13.tgz (Root Library)
- :x: **underscore.string-3.3.5.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Regular Expression Denial of Service (ReDoS) vulnerability was found in underscore.string 2.4.0 through 3.3.5.
<p>Publish Date: 2017-09-08
<p>URL: <a href=https://github.com/esamattis/underscore.string/issues/510>WS-2017-3772</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"underscore.string","packageVersion":"3.3.5","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"grunt-jsbeautifier:0.2.13;underscore.string:3.3.5","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2017-3772","vulnerabilityDetails":"Regular Expression Denial of Service (ReDoS) vulnerability was found in underscore.string 2.4.0 through 3.3.5.","vulnerabilityUrl":"https://github.com/esamattis/underscore.string/issues/510","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_priority | ws high detected in underscore string tgz autoclosed ws high severity vulnerability vulnerable library underscore string tgz string manipulation extensions for underscore js javascript library library home page a href path to dependency file forrester demo package json path to vulnerable library forrester demo node modules underscore string package json dependency hierarchy grunt jsbeautifier tgz root library x underscore string tgz vulnerable library found in base branch master vulnerability details regular expression denial of service redos vulnerability was found in underscore string through publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree grunt jsbeautifier underscore string isminimumfixversionavailable false basebranches vulnerabilityidentifier ws vulnerabilitydetails regular expression denial of service redos vulnerability was found in underscore string through vulnerabilityurl | 0 |
201,831 | 23,041,611,513 | IssuesEvent | 2022-07-23 08:15:15 | wbpascal/Azuria | https://api.github.com/repos/wbpascal/Azuria | opened | CVE-2017-0247 (High) detected in system.net.http.4.3.0.nupkg, system.text.encodings.web.4.3.0.nupkg | security vulnerability | ## CVE-2017-0247 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>system.net.http.4.3.0.nupkg</b>, <b>system.text.encodings.web.4.3.0.nupkg</b></p></summary>
<p>
<details><summary><b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /Azuria.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- newtonsoft.json.10.0.2.nupkg (Root Library)
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
</details>
<details><summary><b>system.text.encodings.web.4.3.0.nupkg</b></p></summary>
<p>Provides types for encoding and escaping strings for use in JavaScript, HyperText Markup Language (H...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.text.encodings.web.4.3.0.nupkg">https://api.nuget.org/packages/system.text.encodings.web.4.3.0.nupkg</a></p>
<p>Path to dependency file: /Azuria.Test/Azuria.Test.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.text.encodings.web/4.3.0/system.text.encodings.web.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- Azuria.Test.Core-1.0.0 (Root Library)
- microsoft.aspnetcore.webutilities.1.1.1.nupkg
- :x: **system.text.encodings.web.4.3.0.nupkg** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/wbpascal/Azuria/commit/d81a0c428805d5623a9ca9b13049fab702250d18">d81a0c428805d5623a9ca9b13049fab702250d18</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A denial of service vulnerability exists when the ASP.NET Core fails to properly validate web requests. NOTE: Microsoft has not commented on third-party claims that the issue is that the TextEncoder.EncodeCore function in the System.Text.Encodings.Web package in ASP.NET Core Mvc before 1.0.4 and 1.1.x before 1.1.3 allows remote attackers to cause a denial of service by leveraging failure to properly calculate the length of 4-byte characters in the Unicode Non-Character range.
<p>Publish Date: 2017-05-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-0247>CVE-2017-0247</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2017-05-12</p>
<p>Fix Resolution: System.Text.Encodings.Web - 4.0.1,4.3.1;System.Net.Http - 4.1.2,4.3.2;System.Net.Http.WinHttpHandler - 4.0.2,4.5.4;System.Net.Security - 4.0.1,4.3.1;System.Net.WebSockets.Client - 4.0.1,4.3.1;Microsoft.AspNetCore.Mvc - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Core - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Abstractions - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ApiExplorer - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Cors - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.DataAnnotations - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Json - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Xml - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Localization - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor.Host - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.TagHelpers - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ViewFeatures - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.WebApiCompatShim - 1.0.4,1.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-0247 (High) detected in system.net.http.4.3.0.nupkg, system.text.encodings.web.4.3.0.nupkg - ## CVE-2017-0247 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>system.net.http.4.3.0.nupkg</b>, <b>system.text.encodings.web.4.3.0.nupkg</b></p></summary>
<p>
<details><summary><b>system.net.http.4.3.0.nupkg</b></p></summary>
<p>Provides a programming interface for modern HTTP applications, including HTTP client components that...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.net.http.4.3.0.nupkg">https://api.nuget.org/packages/system.net.http.4.3.0.nupkg</a></p>
<p>Path to dependency file: /Azuria.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.net.http/4.3.0/system.net.http.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- newtonsoft.json.10.0.2.nupkg (Root Library)
- netstandard.library.1.6.1.nupkg
- :x: **system.net.http.4.3.0.nupkg** (Vulnerable Library)
</details>
<details><summary><b>system.text.encodings.web.4.3.0.nupkg</b></p></summary>
<p>Provides types for encoding and escaping strings for use in JavaScript, HyperText Markup Language (H...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.text.encodings.web.4.3.0.nupkg">https://api.nuget.org/packages/system.text.encodings.web.4.3.0.nupkg</a></p>
<p>Path to dependency file: /Azuria.Test/Azuria.Test.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.text.encodings.web/4.3.0/system.text.encodings.web.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- Azuria.Test.Core-1.0.0 (Root Library)
- microsoft.aspnetcore.webutilities.1.1.1.nupkg
- :x: **system.text.encodings.web.4.3.0.nupkg** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/wbpascal/Azuria/commit/d81a0c428805d5623a9ca9b13049fab702250d18">d81a0c428805d5623a9ca9b13049fab702250d18</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A denial of service vulnerability exists when the ASP.NET Core fails to properly validate web requests. NOTE: Microsoft has not commented on third-party claims that the issue is that the TextEncoder.EncodeCore function in the System.Text.Encodings.Web package in ASP.NET Core Mvc before 1.0.4 and 1.1.x before 1.1.3 allows remote attackers to cause a denial of service by leveraging failure to properly calculate the length of 4-byte characters in the Unicode Non-Character range.
<p>Publish Date: 2017-05-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-0247>CVE-2017-0247</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2017-05-12</p>
<p>Fix Resolution: System.Text.Encodings.Web - 4.0.1,4.3.1;System.Net.Http - 4.1.2,4.3.2;System.Net.Http.WinHttpHandler - 4.0.2,4.5.4;System.Net.Security - 4.0.1,4.3.1;System.Net.WebSockets.Client - 4.0.1,4.3.1;Microsoft.AspNetCore.Mvc - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Core - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Abstractions - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ApiExplorer - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Cors - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.DataAnnotations - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Json - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Formatters.Xml - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Localization - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor.Host - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.Razor - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.TagHelpers - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.ViewFeatures - 1.0.4,1.1.3;Microsoft.AspNetCore.Mvc.WebApiCompatShim - 1.0.4,1.1.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in system net http nupkg system text encodings web nupkg cve high severity vulnerability vulnerable libraries system net http nupkg system text encodings web nupkg system net http nupkg provides a programming interface for modern http applications including http client components that library home page a href path to dependency file azuria csproj path to vulnerable library home wss scanner nuget packages system net http system net http nupkg dependency hierarchy newtonsoft json nupkg root library netstandard library nupkg x system net http nupkg vulnerable library system text encodings web nupkg provides types for encoding and escaping strings for use in javascript hypertext markup language h library home page a href path to dependency file azuria test azuria test csproj path to vulnerable library home wss scanner nuget packages system text encodings web system text encodings web nupkg dependency hierarchy azuria test core root library microsoft aspnetcore webutilities nupkg x system text encodings web nupkg vulnerable library found in head commit a href found in base branch master vulnerability details a denial of service vulnerability exists when the asp net core fails to properly validate web requests note microsoft has not commented on third party claims that the issue is that the textencoder encodecore function in the system text encodings web package in asp net core mvc before and x before allows remote attackers to cause a denial of service by leveraging failure to properly calculate the length of byte characters in the unicode non character range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution system text encodings web system net http system net http winhttphandler system net security system net websockets client microsoft aspnetcore mvc microsoft aspnetcore mvc core microsoft aspnetcore mvc abstractions microsoft aspnetcore mvc apiexplorer microsoft aspnetcore mvc cors microsoft aspnetcore mvc dataannotations microsoft aspnetcore mvc formatters json microsoft aspnetcore mvc formatters xml microsoft aspnetcore mvc localization microsoft aspnetcore mvc razor host microsoft aspnetcore mvc razor microsoft aspnetcore mvc taghelpers microsoft aspnetcore mvc viewfeatures microsoft aspnetcore mvc webapicompatshim step up your open source security game with mend | 0 |
53,345 | 13,261,447,464 | IssuesEvent | 2020-08-20 19:55:05 | icecube-trac/tix4 | https://api.github.com/repos/icecube-trac/tix4 | closed | [clast] no sphinx documentation (Trac #1210) | Migrated from Trac combo reconstruction defect | Add something to `resources/docs/index.rst`. Even a link to the doxygen would be useful.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1210">https://code.icecube.wisc.edu/projects/icecube/ticket/1210</a>, reported by david.schultzand owned by markw04</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:55",
"_ts": "1550067295757382",
"description": "Add something to `resources/docs/index.rst`. Even a link to the doxygen would be useful.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"time": "2015-08-19T19:28:52",
"component": "combo reconstruction",
"summary": "[clast] no sphinx documentation",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "markw04",
"type": "defect"
}
```
</p>
</details>
| 1.0 | [clast] no sphinx documentation (Trac #1210) - Add something to `resources/docs/index.rst`. Even a link to the doxygen would be useful.
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/1210">https://code.icecube.wisc.edu/projects/icecube/ticket/1210</a>, reported by david.schultzand owned by markw04</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:14:55",
"_ts": "1550067295757382",
"description": "Add something to `resources/docs/index.rst`. Even a link to the doxygen would be useful.",
"reporter": "david.schultz",
"cc": "",
"resolution": "fixed",
"time": "2015-08-19T19:28:52",
"component": "combo reconstruction",
"summary": "[clast] no sphinx documentation",
"priority": "blocker",
"keywords": "",
"milestone": "",
"owner": "markw04",
"type": "defect"
}
```
</p>
</details>
| non_priority | no sphinx documentation trac add something to resources docs index rst even a link to the doxygen would be useful migrated from json status closed changetime ts description add something to resources docs index rst even a link to the doxygen would be useful reporter david schultz cc resolution fixed time component combo reconstruction summary no sphinx documentation priority blocker keywords milestone owner type defect | 0 |
88,397 | 25,392,883,776 | IssuesEvent | 2022-11-22 06:01:11 | SynoCommunity/spksrc | https://api.github.com/repos/SynoCommunity/spksrc | closed | ImageMagick not converting SVG to Bitmap | enhancement build/meson build/rust | _For new Package Requests, see the guidelines_
### Setup
_Package Name:_ ImageMagick
_Package Version:_ 7.0.8-4
_NAS Model:_ DS413
_NAS Architecture:_ FREESCALE QorIQ P1022
_DSM version:_ DSM 6.2.2-24922 Update 4
### Expected behavior
Possibility to convert SVG files to bitmap
### Actual behavior
Not possible to convert SVG files to Bitmap one. Trying to do so generate errors.
### Steps to reproduce
_1._ loggin with ssh client
_2._ have a svg file to convert : "my_drawing.svg" for example
_3._ try command : convert my_drawing.svg my_drawing.jpg
_4._ you'll get the error :
`convert: UnableToOpenConfigureFile 'delegates.xml' @ warning/configure.c/GetConfigureOptions/677.
convert: NotAuthorized 'MVG' @ error/constitute.c/IsCoderAuthorized/408.
convert: NoImagesDefined 'dessin.jpg' @ error/convert.c/ConvertImageCommand/3258.`
| 2.0 | ImageMagick not converting SVG to Bitmap - _For new Package Requests, see the guidelines_
### Setup
_Package Name:_ ImageMagick
_Package Version:_ 7.0.8-4
_NAS Model:_ DS413
_NAS Architecture:_ FREESCALE QorIQ P1022
_DSM version:_ DSM 6.2.2-24922 Update 4
### Expected behavior
Possibility to convert SVG files to bitmap
### Actual behavior
Not possible to convert SVG files to Bitmap one. Trying to do so generate errors.
### Steps to reproduce
_1._ loggin with ssh client
_2._ have a svg file to convert : "my_drawing.svg" for example
_3._ try command : convert my_drawing.svg my_drawing.jpg
_4._ you'll get the error :
`convert: UnableToOpenConfigureFile 'delegates.xml' @ warning/configure.c/GetConfigureOptions/677.
convert: NotAuthorized 'MVG' @ error/constitute.c/IsCoderAuthorized/408.
convert: NoImagesDefined 'dessin.jpg' @ error/convert.c/ConvertImageCommand/3258.`
| non_priority | imagemagick not converting svg to bitmap for new package requests see the guidelines setup package name imagemagick package version nas model nas architecture freescale qoriq dsm version dsm update expected behavior possibility to convert svg files to bitmap actual behavior not possible to convert svg files to bitmap one trying to do so generate errors steps to reproduce loggin with ssh client have a svg file to convert my drawing svg for example try command convert my drawing svg my drawing jpg you ll get the error convert unabletoopenconfigurefile delegates xml warning configure c getconfigureoptions convert notauthorized mvg error constitute c iscoderauthorized convert noimagesdefined dessin jpg error convert c convertimagecommand | 0 |
34,890 | 12,304,367,276 | IssuesEvent | 2020-05-11 20:21:50 | aws-amplify/amplify-js | https://api.github.com/repos/aws-amplify/amplify-js | closed | How to detect revoked access token | Auth Cognito Security question | ** Which Category is your question related to? **
Authentication
** What AWS Services are you utilizing? **
Cognitio
** Provide additional details e.g. code snippets **
I am using globalSignOut to revoke the access token for a User. My Java code checks the validity of the access token but does not detect a revoked access token since this information is not included in the token. How can I check if a token has been revoked. | True | How to detect revoked access token - ** Which Category is your question related to? **
Authentication
** What AWS Services are you utilizing? **
Cognitio
** Provide additional details e.g. code snippets **
I am using globalSignOut to revoke the access token for a User. My Java code checks the validity of the access token but does not detect a revoked access token since this information is not included in the token. How can I check if a token has been revoked. | non_priority | how to detect revoked access token which category is your question related to authentication what aws services are you utilizing cognitio provide additional details e g code snippets i am using globalsignout to revoke the access token for a user my java code checks the validity of the access token but does not detect a revoked access token since this information is not included in the token how can i check if a token has been revoked | 0 |
54,970 | 11,355,932,199 | IssuesEvent | 2020-01-24 21:17:37 | GSA/code-gov-front-end | https://api.github.com/repos/GSA/code-gov-front-end | closed | Fixes to Open Task `Types` | [issue-type] bug [issue-type] good first issue [skill-level] beginner code.gov help wanted | <!-- Issues should follow our Issue Guidelines, which are at https://github.com/GSA/code-gov-front-end/blob/master/CONTRIBUTING.md#issue-guidelines -->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
**To Reproduce**
<!-- Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
-->
1. View Open Tasks for GSA https://code.gov/open-tasks?&agencies=GSA&page=1&size=10
2. Take a look a the Type meta data for `Reduce Image Sizes` and several other tasks. The type is listed as `good`.
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
Type should be `good first issue`.
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->
<img width="971" alt="Screen Shot 2019-05-07 at 5 34 16 PM" src="https://user-images.githubusercontent.com/2197515/57335070-8b6a3980-70ef-11e9-8ea6-73d09d0de356.png">
<img width="286" alt="Screen Shot 2019-05-07 at 5 34 24 PM" src="https://user-images.githubusercontent.com/2197515/57335071-8b6a3980-70ef-11e9-86e0-189d7dc29e1f.png">
<img width="334" alt="Screen Shot 2019-05-07 at 5 38 14 PM" src="https://user-images.githubusercontent.com/2197515/57335072-8b6a3980-70ef-11e9-8b7a-2b044db5127a.png">
**Additional context**
<!-- Add any other context about the problem here. -->
| 1.0 | Fixes to Open Task `Types` - <!-- Issues should follow our Issue Guidelines, which are at https://github.com/GSA/code-gov-front-end/blob/master/CONTRIBUTING.md#issue-guidelines -->
**Describe the bug**
<!-- A clear and concise description of what the bug is. -->
**To Reproduce**
<!-- Steps to reproduce the behavior:
1. Go to '...'
2. Click on '....'
3. Scroll down to '....'
4. See error
-->
1. View Open Tasks for GSA https://code.gov/open-tasks?&agencies=GSA&page=1&size=10
2. Take a look a the Type meta data for `Reduce Image Sizes` and several other tasks. The type is listed as `good`.
**Expected behavior**
<!-- A clear and concise description of what you expected to happen. -->
Type should be `good first issue`.
**Screenshots**
<!-- If applicable, add screenshots to help explain your problem. -->
<img width="971" alt="Screen Shot 2019-05-07 at 5 34 16 PM" src="https://user-images.githubusercontent.com/2197515/57335070-8b6a3980-70ef-11e9-8ea6-73d09d0de356.png">
<img width="286" alt="Screen Shot 2019-05-07 at 5 34 24 PM" src="https://user-images.githubusercontent.com/2197515/57335071-8b6a3980-70ef-11e9-86e0-189d7dc29e1f.png">
<img width="334" alt="Screen Shot 2019-05-07 at 5 38 14 PM" src="https://user-images.githubusercontent.com/2197515/57335072-8b6a3980-70ef-11e9-8b7a-2b044db5127a.png">
**Additional context**
<!-- Add any other context about the problem here. -->
| non_priority | fixes to open task types describe the bug to reproduce steps to reproduce the behavior go to click on scroll down to see error view open tasks for gsa take a look a the type meta data for reduce image sizes and several other tasks the type is listed as good expected behavior type should be good first issue screenshots img width alt screen shot at pm src img width alt screen shot at pm src img width alt screen shot at pm src additional context | 0 |
79,836 | 15,586,239,257 | IssuesEvent | 2021-03-18 01:29:29 | jyothsna/votingapplication | https://api.github.com/repos/jyothsna/votingapplication | opened | CVE-2020-36181 (High) detected in jackson-databind-2.9.6.jar, jackson-databind-2.9.0.jar | security vulnerability | ## CVE-2020-36181 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.0.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: votingapplication/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.3.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.0.3.RELEASE.jar
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: votingapplication/pom.xml</p>
<p>Path to vulnerable library: epository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.cpdsadapter.DriverAdapterCPDS.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36181>CVE-2020-36181</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/3004">https://github.com/FasterXML/jackson-databind/issues/3004</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-36181 (High) detected in jackson-databind-2.9.6.jar, jackson-databind-2.9.0.jar - ## CVE-2020-36181 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.0.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: votingapplication/build.gradle</p>
<p>Path to vulnerable library: /root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar,/root/.gradle/caches/modules-2/files-2.1/com.fasterxml.jackson.core/jackson-databind/2.9.6/cfa4f316351a91bfd95cb0644c6a2c95f52db1fc/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.0.3.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.0.3.RELEASE.jar
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: votingapplication/pom.xml</p>
<p>Path to vulnerable library: epository/com/fasterxml/jackson/core/jackson-databind/2.9.0/jackson-databind-2.9.0.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.0.jar** (Vulnerable Library)
</details>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.cpdsadapter.DriverAdapterCPDS.
<p>Publish Date: 2021-01-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36181>CVE-2020-36181</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/3004">https://github.com/FasterXML/jackson-databind/issues/3004</a></p>
<p>Release Date: 2021-01-06</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in jackson databind jar jackson databind jar cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file votingapplication build gradle path to vulnerable library root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar root gradle caches modules files com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file votingapplication pom xml path to vulnerable library epository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp cpdsadapter driveradaptercpds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource | 0 |
344,681 | 30,752,302,367 | IssuesEvent | 2023-07-28 20:32:45 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | backupccl: TestBackupRestoreTenant timed out on `WaitForTenantReadiness` | C-bug skipped-test T-multitenant db-cy-23 | Run: https://teamcity.cockroachdb.com/viewLog.html?buildId=11074294&buildTypeId=Cockroach_BazelEssentialCi
The call over here - https://github.com/cockroachdb/cockroach/blob/9c510f9abdcd0d52e04f620ce5fa283c54d6ef46/pkg/ccl/backupccl/backup_test.go#L7358 seems to be stuck in this select clause - https://github.com/cockroachdb/cockroach/blob/ecc50b7a1936328f1e0add890bff05ef820c2344/pkg/server/testserver.go#L1048. I haven't dug beyond that yet. I'll post more updates as I do.
Jira issue: CRDB-30152 | 1.0 | backupccl: TestBackupRestoreTenant timed out on `WaitForTenantReadiness` - Run: https://teamcity.cockroachdb.com/viewLog.html?buildId=11074294&buildTypeId=Cockroach_BazelEssentialCi
The call over here - https://github.com/cockroachdb/cockroach/blob/9c510f9abdcd0d52e04f620ce5fa283c54d6ef46/pkg/ccl/backupccl/backup_test.go#L7358 seems to be stuck in this select clause - https://github.com/cockroachdb/cockroach/blob/ecc50b7a1936328f1e0add890bff05ef820c2344/pkg/server/testserver.go#L1048. I haven't dug beyond that yet. I'll post more updates as I do.
Jira issue: CRDB-30152 | non_priority | backupccl testbackuprestoretenant timed out on waitfortenantreadiness run the call over here seems to be stuck in this select clause i haven t dug beyond that yet i ll post more updates as i do jira issue crdb | 0 |
155,935 | 19,803,139,374 | IssuesEvent | 2022-01-19 01:33:05 | artkamote/examples | https://api.github.com/repos/artkamote/examples | opened | CVE-2020-9488 (Low) detected in log4j-1.2.17.jar | security vulnerability | ## CVE-2020-9488 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to dependency file: /aws-java-simple-http-endpoint/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- aws-lambda-java-log4j-1.0.0.jar (Root Library)
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-9488 (Low) detected in log4j-1.2.17.jar - ## CVE-2020-9488 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>log4j-1.2.17.jar</b></p></summary>
<p>Apache Log4j 1.2</p>
<p>Path to dependency file: /aws-java-simple-http-endpoint/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/log4j/log4j/1.2.17/log4j-1.2.17.jar</p>
<p>
Dependency Hierarchy:
- aws-lambda-java-log4j-1.0.0.jar (Root Library)
- :x: **log4j-1.2.17.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Improper validation of certificate with host mismatch in Apache Log4j SMTP appender. This could allow an SMTPS connection to be intercepted by a man-in-the-middle attack which could leak any log messages sent through that appender.
<p>Publish Date: 2020-04-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9488>CVE-2020-9488</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>3.7</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://issues.apache.org/jira/browse/LOG4J2-2819">https://issues.apache.org/jira/browse/LOG4J2-2819</a></p>
<p>Release Date: 2020-04-27</p>
<p>Fix Resolution: org.apache.logging.log4j:log4j-core:2.13.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve low detected in jar cve low severity vulnerability vulnerable library jar apache path to dependency file aws java simple http endpoint pom xml path to vulnerable library home wss scanner repository jar dependency hierarchy aws lambda java jar root library x jar vulnerable library found in base branch master vulnerability details improper validation of certificate with host mismatch in apache smtp appender this could allow an smtps connection to be intercepted by a man in the middle attack which could leak any log messages sent through that appender publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache logging core step up your open source security game with whitesource | 0 |
160,293 | 12,506,696,217 | IssuesEvent | 2020-06-02 13:01:00 | pingcap/tidb | https://api.github.com/repos/pingcap/tidb | opened | global_vars_cache_test.go:testGVCSuite.TestSimple failed | component/test | global_vars_cache_test.go:testGVCSuite.TestSimple
```
[2020-06-02T09:58:31.704Z] ----------------------------------------------------------------------
[2020-06-02T09:58:31.704Z] FAIL: global_vars_cache_test.go:38: testGVCSuite.TestSimple
[2020-06-02T09:58:31.704Z]
[2020-06-02T09:58:31.704Z] global_vars_cache_test.go:85:
[2020-06-02T09:58:31.704Z] ...
[2020-06-02T09:58:31.704Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testleak/leaktest.go:135:
[2020-06-02T09:58:31.704Z] c.Errorf("Test %s check-count %d appears to have leaked: %v", c.TestName(), cnt, g)
[2020-06-02T09:58:31.704Z] ... Error: Test testGVCSuite.TestSimple check-count 50 appears to have leaked: github.com/pingcap/tidb/domain.(*Domain).loadSchemaInLoop(0xc007494ea0, 0x2faf080)
[2020-06-02T09:58:31.704Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/domain/domain.go:505 +0x254
[2020-06-02T09:58:31.704Z] created by github.com/pingcap/tidb/domain.(*Domain).Init
[2020-06-02T09:58:31.704Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/domain/domain.go:740 +0x5d1
[2020-06-02T09:58:31.704Z]
```
Latest failed builds:
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/38202/display/redirect
| 1.0 | global_vars_cache_test.go:testGVCSuite.TestSimple failed - global_vars_cache_test.go:testGVCSuite.TestSimple
```
[2020-06-02T09:58:31.704Z] ----------------------------------------------------------------------
[2020-06-02T09:58:31.704Z] FAIL: global_vars_cache_test.go:38: testGVCSuite.TestSimple
[2020-06-02T09:58:31.704Z]
[2020-06-02T09:58:31.704Z] global_vars_cache_test.go:85:
[2020-06-02T09:58:31.704Z] ...
[2020-06-02T09:58:31.704Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/util/testleak/leaktest.go:135:
[2020-06-02T09:58:31.704Z] c.Errorf("Test %s check-count %d appears to have leaked: %v", c.TestName(), cnt, g)
[2020-06-02T09:58:31.704Z] ... Error: Test testGVCSuite.TestSimple check-count 50 appears to have leaked: github.com/pingcap/tidb/domain.(*Domain).loadSchemaInLoop(0xc007494ea0, 0x2faf080)
[2020-06-02T09:58:31.704Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/domain/domain.go:505 +0x254
[2020-06-02T09:58:31.704Z] created by github.com/pingcap/tidb/domain.(*Domain).Init
[2020-06-02T09:58:31.704Z] /home/jenkins/agent/workspace/tidb_ghpr_unit_test/go/src/github.com/pingcap/tidb/domain/domain.go:740 +0x5d1
[2020-06-02T09:58:31.704Z]
```
Latest failed builds:
https://internal.pingcap.net/idc-jenkins/job/tidb_ghpr_unit_test/38202/display/redirect
| non_priority | global vars cache test go testgvcsuite testsimple failed global vars cache test go testgvcsuite testsimple fail global vars cache test go testgvcsuite testsimple global vars cache test go home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb util testleak leaktest go c errorf test s check count d appears to have leaked v c testname cnt g error test testgvcsuite testsimple check count appears to have leaked github com pingcap tidb domain domain loadschemainloop home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb domain domain go created by github com pingcap tidb domain domain init home jenkins agent workspace tidb ghpr unit test go src github com pingcap tidb domain domain go latest failed builds | 0 |
79,650 | 28,495,924,935 | IssuesEvent | 2023-04-18 14:12:41 | vector-im/element-desktop | https://api.github.com/repos/vector-im/element-desktop | opened | Desktop notifications now show an app ID on windows | T-Defect | 
This might be a recent windows update - notifications have only recently gotten the app name put at the top of the notification. The screenshot is from Element Nightly, but I believe Windows has a length limit so it might be trimming 'nightly' off the apparent ID | 1.0 | Desktop notifications now show an app ID on windows - 
This might be a recent windows update - notifications have only recently gotten the app name put at the top of the notification. The screenshot is from Element Nightly, but I believe Windows has a length limit so it might be trimming 'nightly' off the apparent ID | non_priority | desktop notifications now show an app id on windows this might be a recent windows update notifications have only recently gotten the app name put at the top of the notification the screenshot is from element nightly but i believe windows has a length limit so it might be trimming nightly off the apparent id | 0 |
6,839 | 5,683,074,549 | IssuesEvent | 2017-04-13 11:37:45 | learningequality/ka-lite | https://api.github.com/repos/learningequality/ka-lite | opened | Clean up old code in kalite.contentload | refactoring and performance | ## Summary
Most of `kalite.contentload` seems to be outdated, both on distributed and central server.
Here's a snippet from `kalite.contentload.api_views`:
```
"""
This file implements both the central and distributed server sides of
a handshake to download KA data.
Why does the central server have to be involved?
- because we have exactly one API Key for KA, and we don't want to share it
with distributed server accounts.
- because we don't trust KA to keep their API static; by putting the central
server in the middle, we can easily update, and distributed servers don't break.
Here's how it works:
* On the distributed server, there is a button on the facility user's "account" page with a button"Download data from KA".
* That button has a link to a distributed server url. The user clicks it.
* That distributed server view sets up a proper URL/request to the central server, then redirects that central server URL.
* The central server tries to authenticate to KA (forwarding users to KA), with a call-back URL when that succeeds.
* The user authenticates with KA, and KA oauth is returned to the central server.
* The central server then uses the KA API to get the user data, interpret it, massage it, and compute (our) relevant quantities.
* The central server then uses a distributed server call-back URL to POST the downloaded user data.
* The distributed server gets that data, loads it, saves it, and then redirects the user--to their account page.
* The account page shows again, this time including the imported KA data
"""
```
## System information
- Operating system: all
- Version: develop | True | Clean up old code in kalite.contentload - ## Summary
Most of `kalite.contentload` seems to be outdated, both on distributed and central server.
Here's a snippet from `kalite.contentload.api_views`:
```
"""
This file implements both the central and distributed server sides of
a handshake to download KA data.
Why does the central server have to be involved?
- because we have exactly one API Key for KA, and we don't want to share it
with distributed server accounts.
- because we don't trust KA to keep their API static; by putting the central
server in the middle, we can easily update, and distributed servers don't break.
Here's how it works:
* On the distributed server, there is a button on the facility user's "account" page with a button"Download data from KA".
* That button has a link to a distributed server url. The user clicks it.
* That distributed server view sets up a proper URL/request to the central server, then redirects that central server URL.
* The central server tries to authenticate to KA (forwarding users to KA), with a call-back URL when that succeeds.
* The user authenticates with KA, and KA oauth is returned to the central server.
* The central server then uses the KA API to get the user data, interpret it, massage it, and compute (our) relevant quantities.
* The central server then uses a distributed server call-back URL to POST the downloaded user data.
* The distributed server gets that data, loads it, saves it, and then redirects the user--to their account page.
* The account page shows again, this time including the imported KA data
"""
```
## System information
- Operating system: all
- Version: develop | non_priority | clean up old code in kalite contentload summary most of kalite contentload seems to be outdated both on distributed and central server here s a snippet from kalite contentload api views this file implements both the central and distributed server sides of a handshake to download ka data why does the central server have to be involved because we have exactly one api key for ka and we don t want to share it with distributed server accounts because we don t trust ka to keep their api static by putting the central server in the middle we can easily update and distributed servers don t break here s how it works on the distributed server there is a button on the facility user s account page with a button download data from ka that button has a link to a distributed server url the user clicks it that distributed server view sets up a proper url request to the central server then redirects that central server url the central server tries to authenticate to ka forwarding users to ka with a call back url when that succeeds the user authenticates with ka and ka oauth is returned to the central server the central server then uses the ka api to get the user data interpret it massage it and compute our relevant quantities the central server then uses a distributed server call back url to post the downloaded user data the distributed server gets that data loads it saves it and then redirects the user to their account page the account page shows again this time including the imported ka data system information operating system all version develop | 0 |
308,195 | 26,585,261,748 | IssuesEvent | 2023-01-22 23:17:38 | xbmc/xbmc | https://api.github.com/repos/xbmc/xbmc | closed | Kodi is unable to play an m3u8 file, but can play m3u8 inside it | Triage: Tested and not reproduced | <!--- Please fill out this template to the best of your ability. You can always edit this issue once you have created it. -->
<!--- Read the following link before you create a new problem report: https://kodi.wiki/view/HOW-TO:Submit_a_bug_report -->
## Bug report
### Describe the bug
Here is a clear and concise description of what the problem is:
The file [test file](https://cdn-ue1-prod.tsv2.amagi.tv/avod/simplestream-lds-ldtimln/ODYPPCS001EP001/ODYPPCS001EP001.m3u8)
Kodi can not play it. Inside this m3u8 file, there are links for different versions of the stream. Kodi can play each of these stream, when directly called.
A note: MPV can play this file without any problem.
## Expected Behavior
Here is a clear and concise description of what was expected to happen:
Kodi can play the stream throughout the initial m3u8 file.
## Actual Behavior
Kodi fails badly to play the stream.
## Possible Fix
### To Reproduce
Steps to reproduce the behavior:
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
<!--- Put your text below this line -->
1. [ Link to the problematic file](https://cdn-ue1-prod.tsv2.amagi.tv/avod/simplestream-lds-ldtimln/ODYPPCS001EP001/ODYPPCS001EP001.m3u8)
2. Launch it through Kodi
3.
### Debuglog
<!--- Put your text below this line -->
<!--- A debuglog is always mandatory when creating an issue. Provide one! -->
The debuglog can be found here:
[log.txt](https://github.com/xbmc/xbmc/files/10471307/log.txt)
### Screenshots
Here are some links or screenshots to help explain the problem:
<!--- Put your text below this line -->
## Additional context or screenshots (if appropriate)
Here is some additional context or explanation that might help:
<!--- How has this bug affected you? What were you trying to accomplish? -->
<!--- Put your text below this line -->
### Your Environment
Used Operating system:
<!--- Include as many relevant details about the environment you experienced the bug in -->
<!--- Put your text below this line. Checkboxes can easily be ticked once issue is created -->
- [] Android
- [ ] iOS
- [ ] tvOS
- [x] Linux
- [ ] macOS
- [] Windows
- [ ] Windows UWP
- Operating system version/name:
- Kodi version: Kdo Matrix 19.5, Kodi Nexus 20.0
<!--- End of this issue -->
*note: Once the issue is made we require you to update it with new information or Kodi versions should that be required.
Team Kodi will consider your problem report however, we will not make any promises the problem will be solved.* | 1.0 | Kodi is unable to play an m3u8 file, but can play m3u8 inside it - <!--- Please fill out this template to the best of your ability. You can always edit this issue once you have created it. -->
<!--- Read the following link before you create a new problem report: https://kodi.wiki/view/HOW-TO:Submit_a_bug_report -->
## Bug report
### Describe the bug
Here is a clear and concise description of what the problem is:
The file [test file](https://cdn-ue1-prod.tsv2.amagi.tv/avod/simplestream-lds-ldtimln/ODYPPCS001EP001/ODYPPCS001EP001.m3u8)
Kodi can not play it. Inside this m3u8 file, there are links for different versions of the stream. Kodi can play each of these stream, when directly called.
A note: MPV can play this file without any problem.
## Expected Behavior
Here is a clear and concise description of what was expected to happen:
Kodi can play the stream throughout the initial m3u8 file.
## Actual Behavior
Kodi fails badly to play the stream.
## Possible Fix
### To Reproduce
Steps to reproduce the behavior:
<!--- Provide a link to a live example, or an unambiguous set of steps to -->
<!--- reproduce this bug. Include code to reproduce, if relevant -->
<!--- Put your text below this line -->
1. [ Link to the problematic file](https://cdn-ue1-prod.tsv2.amagi.tv/avod/simplestream-lds-ldtimln/ODYPPCS001EP001/ODYPPCS001EP001.m3u8)
2. Launch it through Kodi
3.
### Debuglog
<!--- Put your text below this line -->
<!--- A debuglog is always mandatory when creating an issue. Provide one! -->
The debuglog can be found here:
[log.txt](https://github.com/xbmc/xbmc/files/10471307/log.txt)
### Screenshots
Here are some links or screenshots to help explain the problem:
<!--- Put your text below this line -->
## Additional context or screenshots (if appropriate)
Here is some additional context or explanation that might help:
<!--- How has this bug affected you? What were you trying to accomplish? -->
<!--- Put your text below this line -->
### Your Environment
Used Operating system:
<!--- Include as many relevant details about the environment you experienced the bug in -->
<!--- Put your text below this line. Checkboxes can easily be ticked once issue is created -->
- [] Android
- [ ] iOS
- [ ] tvOS
- [x] Linux
- [ ] macOS
- [] Windows
- [ ] Windows UWP
- Operating system version/name:
- Kodi version: Kdo Matrix 19.5, Kodi Nexus 20.0
<!--- End of this issue -->
*note: Once the issue is made we require you to update it with new information or Kodi versions should that be required.
Team Kodi will consider your problem report however, we will not make any promises the problem will be solved.* | non_priority | kodi is unable to play an file but can play inside it bug report describe the bug here is a clear and concise description of what the problem is the file kodi can not play it inside this file there are links for different versions of the stream kodi can play each of these stream when directly called a note mpv can play this file without any problem expected behavior here is a clear and concise description of what was expected to happen kodi can play the stream throughout the initial file actual behavior kodi fails badly to play the stream possible fix to reproduce steps to reproduce the behavior launch it through kodi debuglog the debuglog can be found here screenshots here are some links or screenshots to help explain the problem additional context or screenshots if appropriate here is some additional context or explanation that might help your environment used operating system android ios tvos linux macos windows windows uwp operating system version name kodi version kdo matrix kodi nexus note once the issue is made we require you to update it with new information or kodi versions should that be required team kodi will consider your problem report however we will not make any promises the problem will be solved | 0 |
339,411 | 24,619,166,133 | IssuesEvent | 2022-10-15 17:54:31 | Megabit/Blazorise | https://api.github.com/repos/Megabit/Blazorise | closed | Carousel behaviour | Area: Documentation Type: Possible Bug Status: Repro Missing | I have the carousel working but with one big issue, every time an image slides out of view it is followed by a blank screen before the next image slides in. I am using just 3 images. Also, is there a way of slowing down the slide frequency?
| 1.0 | Carousel behaviour - I have the carousel working but with one big issue, every time an image slides out of view it is followed by a blank screen before the next image slides in. I am using just 3 images. Also, is there a way of slowing down the slide frequency?
| non_priority | carousel behaviour i have the carousel working but with one big issue every time an image slides out of view it is followed by a blank screen before the next image slides in i am using just images also is there a way of slowing down the slide frequency | 0 |
53,481 | 6,730,462,761 | IssuesEvent | 2017-10-18 01:05:53 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | Documentation and behavior flaws with PackageReference | Area:ToolRestore ClosedAs:ByDesign NeedsTriageDiscussion Style:PackageReference Type:Docs | NuGet.exe 4.1.0.2450
VS2017 Enterprise 15.1(26403.7) - NuGet 4.1.0
1. The documentation lists the Version under PackageReference as an attribute, but when a reference is added to a project by the VS Addin, it is stored as an element. The NuGet executable does not seem to have any problem with this, but it is rather confusing.
2. The VS Addin will always force the complete version number into this Version Element, which is inconvenient if it was configured to use a floating version, you will loose the floating part after an update in VS. It also seems the VS Addin does not recognize the floating part and will interprete 1.1.1.* as 1.1.1, which is something completely different in behavior. (the exe does not seem to have this problem and offers the latest version available in range). A solution could be to use the attribute for the (manually probably) permissive range definition and the element for the current used version in VS. And the two tools (addin and exe) should behave the same on those definitions.
3. The PackageReferences are not automatically transferred into a nupkg file if the exe is used (Nuget.exe pack UC.vbproj). (Tested on a CI server with only build tools; a nuspec file is present in the project with the same name as the project file to get a little more control over description, icon etc.)
Doc available at:
https://docs.microsoft.com/en-us/nuget/consume-packages/package-references-in-project-files
| 1.0 | Documentation and behavior flaws with PackageReference - NuGet.exe 4.1.0.2450
VS2017 Enterprise 15.1(26403.7) - NuGet 4.1.0
1. The documentation lists the Version under PackageReference as an attribute, but when a reference is added to a project by the VS Addin, it is stored as an element. The NuGet executable does not seem to have any problem with this, but it is rather confusing.
2. The VS Addin will always force the complete version number into this Version Element, which is inconvenient if it was configured to use a floating version, you will loose the floating part after an update in VS. It also seems the VS Addin does not recognize the floating part and will interprete 1.1.1.* as 1.1.1, which is something completely different in behavior. (the exe does not seem to have this problem and offers the latest version available in range). A solution could be to use the attribute for the (manually probably) permissive range definition and the element for the current used version in VS. And the two tools (addin and exe) should behave the same on those definitions.
3. The PackageReferences are not automatically transferred into a nupkg file if the exe is used (Nuget.exe pack UC.vbproj). (Tested on a CI server with only build tools; a nuspec file is present in the project with the same name as the project file to get a little more control over description, icon etc.)
Doc available at:
https://docs.microsoft.com/en-us/nuget/consume-packages/package-references-in-project-files
| non_priority | documentation and behavior flaws with packagereference nuget exe enterprise nuget the documentation lists the version under packagereference as an attribute but when a reference is added to a project by the vs addin it is stored as an element the nuget executable does not seem to have any problem with this but it is rather confusing the vs addin will always force the complete version number into this version element which is inconvenient if it was configured to use a floating version you will loose the floating part after an update in vs it also seems the vs addin does not recognize the floating part and will interprete as which is something completely different in behavior the exe does not seem to have this problem and offers the latest version available in range a solution could be to use the attribute for the manually probably permissive range definition and the element for the current used version in vs and the two tools addin and exe should behave the same on those definitions the packagereferences are not automatically transferred into a nupkg file if the exe is used nuget exe pack uc vbproj tested on a ci server with only build tools a nuspec file is present in the project with the same name as the project file to get a little more control over description icon etc doc available at | 0 |
16,014 | 5,197,585,612 | IssuesEvent | 2017-01-23 15:59:47 | project-lighthouse/camera-app.ios | https://api.github.com/repos/project-lighthouse/camera-app.ios | closed | Extend `Feedback` with a way to run platform TTS | Code - Cocoa Feature - Audio | This will let users access the voice that they're used to. | 1.0 | Extend `Feedback` with a way to run platform TTS - This will let users access the voice that they're used to. | non_priority | extend feedback with a way to run platform tts this will let users access the voice that they re used to | 0 |
58,730 | 14,345,517,981 | IssuesEvent | 2020-11-28 19:25:33 | AlexRogalskiy/object-mappers-playground | https://api.github.com/repos/AlexRogalskiy/object-mappers-playground | opened | CVE-2009-2625 (Medium) detected in xercesImpl-2.9.1.jar | security vulnerability | ## CVE-2009-2625 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.9.1.jar</b></p></summary>
<p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the
Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI),
a complete framework for building parser components and configurations that is extremely
modular and easy to program.</p>
<p>Path to dependency file: object-mappers-playground/modules/objectmappers-all/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar</p>
<p>
Dependency Hierarchy:
- milyn-smooks-all-1.7.0.jar (Root Library)
- milyn-commons-1.7.0.jar
- :x: **xercesImpl-2.9.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/object-mappers-playground/commit/51396b5bdc926331c2a288f1f8de822ff8828666">51396b5bdc926331c2a288f1f8de822ff8828666</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XMLScanner.java in Apache Xerces2 Java, as used in Sun Java Runtime Environment (JRE) in JDK and JRE 6 before Update 15 and JDK and JRE 5.0 before Update 20, and in other products, allows remote attackers to cause a denial of service (infinite loop and application hang) via malformed XML input, as demonstrated by the Codenomicon XML fuzzing framework.
<p>Publish Date: 2009-08-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2009-2625>CVE-2009-2625</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://www.securitytracker.com/id?1022680">http://www.securitytracker.com/id?1022680</a></p>
<p>Release Date: 2017-12-31</p>
<p>Fix Resolution: The vendor has issued a fix for Windows, Solaris, and Linux:
* JDK and JRE 6 Update 15 or later
* JDK and JRE 5.0 Update 20 or later
Java SE releases are available at:
JDK and JRE 6 Update 15:
http://java.sun.com/javase/downloads/index.jsp
JRE 6 Update 15:
http://java.com/
through the Java Update tool for Microsoft Windows users.
JDK 6 Update 15 for Solaris is available in the following patches:
* Java SE 6 Update 15 (as delivered in patch 125136-16)
* Java SE 6 Update 15 (as delivered in patch 125137-16 (64bit))
* Java SE 6_x86 Update 15 (as delivered in patch 125138-16)
* Java SE 6_x86 Update 15 (as delivered in patch 125139-16 (64bit))
JDK and JRE 5.0 Update 20:
http://java.sun.com/javase/downloads/index_jdk5.jsp
JDK 5.0 Update 20 for Solaris is available in the following patches:
* J2SE 5.0 Update 18 (as delivered in patch 118666-21)
* J2SE 5.0 Update 18 (as delivered in patch 118667-21 (64bit))
* J2SE 5.0_x86 Update 18 (as delivered in patch 118668-21)
* J2SE 5.0_x86 Update 18 (as delivered in patch 118669-21 (64bit))
Java SE for Business releases are available at:
http://www.sun.com/software/javaseforbusiness/getit_download.jsp
Note: When installing a new version of the product from a source other than a Solaris patch, it is recommended that the old affected versions be removed from your system. To remove old affected versions on the Windows platform, please see:
http://www.java.com/en/download/help/5000010800.xml
The vendor's advisory is available at:
http://sunsolve.sun.com/search/document.do?assetkey=1-66-263489-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2009-2625 (Medium) detected in xercesImpl-2.9.1.jar - ## CVE-2009-2625 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.9.1.jar</b></p></summary>
<p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the
Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI),
a complete framework for building parser components and configurations that is extremely
modular and easy to program.</p>
<p>Path to dependency file: object-mappers-playground/modules/objectmappers-all/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar,/home/wss-scanner/.m2/repository/xerces/xercesImpl/2.9.1/xercesImpl-2.9.1.jar</p>
<p>
Dependency Hierarchy:
- milyn-smooks-all-1.7.0.jar (Root Library)
- milyn-commons-1.7.0.jar
- :x: **xercesImpl-2.9.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/object-mappers-playground/commit/51396b5bdc926331c2a288f1f8de822ff8828666">51396b5bdc926331c2a288f1f8de822ff8828666</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XMLScanner.java in Apache Xerces2 Java, as used in Sun Java Runtime Environment (JRE) in JDK and JRE 6 before Update 15 and JDK and JRE 5.0 before Update 20, and in other products, allows remote attackers to cause a denial of service (infinite loop and application hang) via malformed XML input, as demonstrated by the Codenomicon XML fuzzing framework.
<p>Publish Date: 2009-08-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2009-2625>CVE-2009-2625</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://www.securitytracker.com/id?1022680">http://www.securitytracker.com/id?1022680</a></p>
<p>Release Date: 2017-12-31</p>
<p>Fix Resolution: The vendor has issued a fix for Windows, Solaris, and Linux:
* JDK and JRE 6 Update 15 or later
* JDK and JRE 5.0 Update 20 or later
Java SE releases are available at:
JDK and JRE 6 Update 15:
http://java.sun.com/javase/downloads/index.jsp
JRE 6 Update 15:
http://java.com/
through the Java Update tool for Microsoft Windows users.
JDK 6 Update 15 for Solaris is available in the following patches:
* Java SE 6 Update 15 (as delivered in patch 125136-16)
* Java SE 6 Update 15 (as delivered in patch 125137-16 (64bit))
* Java SE 6_x86 Update 15 (as delivered in patch 125138-16)
* Java SE 6_x86 Update 15 (as delivered in patch 125139-16 (64bit))
JDK and JRE 5.0 Update 20:
http://java.sun.com/javase/downloads/index_jdk5.jsp
JDK 5.0 Update 20 for Solaris is available in the following patches:
* J2SE 5.0 Update 18 (as delivered in patch 118666-21)
* J2SE 5.0 Update 18 (as delivered in patch 118667-21 (64bit))
* J2SE 5.0_x86 Update 18 (as delivered in patch 118668-21)
* J2SE 5.0_x86 Update 18 (as delivered in patch 118669-21 (64bit))
Java SE for Business releases are available at:
http://www.sun.com/software/javaseforbusiness/getit_download.jsp
Note: When installing a new version of the product from a source other than a Solaris patch, it is recommended that the old affected versions be removed from your system. To remove old affected versions on the Windows platform, please see:
http://www.java.com/en/download/help/5000010800.xml
The vendor's advisory is available at:
http://sunsolve.sun.com/search/document.do?assetkey=1-66-263489-1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in xercesimpl jar cve medium severity vulnerability vulnerable library xercesimpl jar is the next generation of high performance fully compliant xml parsers in the apache xerces family this new version of xerces introduces the xerces native interface xni a complete framework for building parser components and configurations that is extremely modular and easy to program path to dependency file object mappers playground modules objectmappers all pom xml path to vulnerable library home wss scanner repository xerces xercesimpl xercesimpl jar home wss scanner repository xerces xercesimpl xercesimpl jar dependency hierarchy milyn smooks all jar root library milyn commons jar x xercesimpl jar vulnerable library found in head commit a href found in base branch master vulnerability details xmlscanner java in apache java as used in sun java runtime environment jre in jdk and jre before update and jdk and jre before update and in other products allows remote attackers to cause a denial of service infinite loop and application hang via malformed xml input as demonstrated by the codenomicon xml fuzzing framework publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution the vendor has issued a fix for windows solaris and linux jdk and jre update or later jdk and jre update or later java se releases are available at jdk and jre update jre update through the java update tool for microsoft windows users jdk update for solaris is available in the following patches java se update as delivered in patch java se update as delivered in patch java se update as delivered in patch java se update as delivered in patch jdk and jre update jdk update for solaris is available in the following patches update as delivered in patch update as delivered in patch update as delivered in patch update as delivered in patch java se for business releases are available at note when installing a new version of the product from a source other than a solaris patch it is recommended that the old affected versions be removed from your system to remove old affected versions on the windows platform please see the vendor s advisory is available at step up your open source security game with whitesource | 0 |
355,529 | 25,175,959,342 | IssuesEvent | 2022-11-11 09:17:07 | Thing1Thing2/pe | https://api.github.com/repos/Thing1Thing2/pe | opened | "private contact detail" term not used anywhere | severity.VeryLow type.DocumentationBug | The term `private contact detail` is mentioned in the glossary but not used anywhere in UG or DG.

<!--session: 1668150936310-328337a8-0287-4e2b-8277-cac763619f8e-->
<!--Version: Web v3.4.4--> | 1.0 | "private contact detail" term not used anywhere - The term `private contact detail` is mentioned in the glossary but not used anywhere in UG or DG.

<!--session: 1668150936310-328337a8-0287-4e2b-8277-cac763619f8e-->
<!--Version: Web v3.4.4--> | non_priority | private contact detail term not used anywhere the term private contact detail is mentioned in the glossary but not used anywhere in ug or dg | 0 |
168,387 | 14,147,464,567 | IssuesEvent | 2020-11-10 20:51:18 | aws/aws-sdk-java-v2 | https://api.github.com/repos/aws/aws-sdk-java-v2 | opened | No examples of Query partition and sort key using Dyanmodb Enhanced client | documentation needs-triage | <!--- Provide a general summary of the issue in the Title above -->
The example code and documentation does not provide example of how to query table or index with sort key. I am looking for example where key=X and sort key between A and B. This seems like basic use case but i have spent hours trying to figure out.
## Describe the issue
If you look at example below and assume the sort key is regDate. how will you query Id= X and regDate between 2020-01-01 and 2020-11-01?
https://github.com/awsdocs/aws-doc-sdk-examples/blob/master/javav2/example_code/dynamodb/src/main/java/com/example/dynamodb/EnhancedQueryRecords.java
| 1.0 | No examples of Query partition and sort key using Dyanmodb Enhanced client - <!--- Provide a general summary of the issue in the Title above -->
The example code and documentation does not provide example of how to query table or index with sort key. I am looking for example where key=X and sort key between A and B. This seems like basic use case but i have spent hours trying to figure out.
## Describe the issue
If you look at example below and assume the sort key is regDate. how will you query Id= X and regDate between 2020-01-01 and 2020-11-01?
https://github.com/awsdocs/aws-doc-sdk-examples/blob/master/javav2/example_code/dynamodb/src/main/java/com/example/dynamodb/EnhancedQueryRecords.java
| non_priority | no examples of query partition and sort key using dyanmodb enhanced client the example code and documentation does not provide example of how to query table or index with sort key i am looking for example where key x and sort key between a and b this seems like basic use case but i have spent hours trying to figure out describe the issue if you look at example below and assume the sort key is regdate how will you query id x and regdate between and | 0 |
87,352 | 17,212,381,804 | IssuesEvent | 2021-07-19 07:06:44 | rust-lang/rust | https://api.github.com/repos/rust-lang/rust | closed | Inneficient code generation, unless splitting part of the code into #[inline(always)] sub-function | A-codegen I-slow T-compiler | (from #43702)
Both examples are completely equivalent after inlining, but the second one gets compiled into a more efficient assembly:
```Rust
pub fn square(mut buf: &mut [u8], self_buf: &[u8; 16*8]) {
while buf.len() >= self_buf.len() {
let (l, r) = {buf}.split_at_mut(self_buf.len());
buf = r;
for i in 0..16*8 { l[i] ^= self_buf[i]; }
}
}
```
https://rust.godbolt.org/z/mXrwhv
```Rust
#[inline(always)]
fn xor(l: &mut [u8], self_buf: &[u8; 16*8]) {
for i in 0..16*8 { l[i] ^= self_buf[i]; }
}
pub fn square(mut buf: &mut [u8], self_buf: &[u8; 16*8]) {
while buf.len() >= self_buf.len() {
let (l, r) = {buf}.split_at_mut(self_buf.len());
buf = r;
xor(l, self_buf);
}
}
```
https://rust.godbolt.org/z/Uk3X3Z | 1.0 | Inneficient code generation, unless splitting part of the code into #[inline(always)] sub-function - (from #43702)
Both examples are completely equivalent after inlining, but the second one gets compiled into a more efficient assembly:
```Rust
pub fn square(mut buf: &mut [u8], self_buf: &[u8; 16*8]) {
while buf.len() >= self_buf.len() {
let (l, r) = {buf}.split_at_mut(self_buf.len());
buf = r;
for i in 0..16*8 { l[i] ^= self_buf[i]; }
}
}
```
https://rust.godbolt.org/z/mXrwhv
```Rust
#[inline(always)]
fn xor(l: &mut [u8], self_buf: &[u8; 16*8]) {
for i in 0..16*8 { l[i] ^= self_buf[i]; }
}
pub fn square(mut buf: &mut [u8], self_buf: &[u8; 16*8]) {
while buf.len() >= self_buf.len() {
let (l, r) = {buf}.split_at_mut(self_buf.len());
buf = r;
xor(l, self_buf);
}
}
```
https://rust.godbolt.org/z/Uk3X3Z | non_priority | inneficient code generation unless splitting part of the code into sub function from both examples are completely equivalent after inlining but the second one gets compiled into a more efficient assembly rust pub fn square mut buf mut self buf while buf len self buf len let l r buf split at mut self buf len buf r for i in l self buf rust fn xor l mut self buf for i in l self buf pub fn square mut buf mut self buf while buf len self buf len let l r buf split at mut self buf len buf r xor l self buf | 0 |
131,247 | 10,686,190,406 | IssuesEvent | 2019-10-22 14:01:26 | commercetools/merchant-center-application-kit | https://api.github.com/repos/commercetools/merchant-center-application-kit | closed | Error in rtl test when using ProjectExtensionProviderForImageRegex | 🐛 Type: Bug 👨⚕️ Type: Tests | **Describe the bug**
When I `renderApp`, in a rtl test, a component that renders app-kit `ProjectExtensionProviderForImageRegex` I get an error.
**To Reproduce**
Steps to reproduce the behavior:
1. Create a rtl test
2. use `renderApp` with a component that renders `ProjectExtensionProviderForImageRegex`
3. See error
```
Error: Network error: No more mocked responses for the query: query FetchProjectExtensionImageRegex {
projectExtension {
id
imageRegex {
thumb {
...ImageRegex
}
small {
...ImageRegex
}
}
}
}
fragment ImageRegex on ImageRegexOptions {
flag
search
replace
}
```
**Expected behavior**
No error should occur. | 1.0 | Error in rtl test when using ProjectExtensionProviderForImageRegex - **Describe the bug**
When I `renderApp`, in a rtl test, a component that renders app-kit `ProjectExtensionProviderForImageRegex` I get an error.
**To Reproduce**
Steps to reproduce the behavior:
1. Create a rtl test
2. use `renderApp` with a component that renders `ProjectExtensionProviderForImageRegex`
3. See error
```
Error: Network error: No more mocked responses for the query: query FetchProjectExtensionImageRegex {
projectExtension {
id
imageRegex {
thumb {
...ImageRegex
}
small {
...ImageRegex
}
}
}
}
fragment ImageRegex on ImageRegexOptions {
flag
search
replace
}
```
**Expected behavior**
No error should occur. | non_priority | error in rtl test when using projectextensionproviderforimageregex describe the bug when i renderapp in a rtl test a component that renders app kit projectextensionproviderforimageregex i get an error to reproduce steps to reproduce the behavior create a rtl test use renderapp with a component that renders projectextensionproviderforimageregex see error error network error no more mocked responses for the query query fetchprojectextensionimageregex projectextension id imageregex thumb imageregex small imageregex fragment imageregex on imageregexoptions flag search replace expected behavior no error should occur | 0 |
95,377 | 8,556,909,644 | IssuesEvent | 2018-11-08 14:28:57 | elifesciences/elife-xpub | https://api.github.com/repos/elifesciences/elife-xpub | closed | Complete end2end journal login integration with profile menu check | Testing | @giorgiosironi says (see #648)
> Only extension of the tests to perform more checks on the xpub page (Continue vs other pages) is missing.
| 1.0 | Complete end2end journal login integration with profile menu check - @giorgiosironi says (see #648)
> Only extension of the tests to perform more checks on the xpub page (Continue vs other pages) is missing.
| non_priority | complete journal login integration with profile menu check giorgiosironi says see only extension of the tests to perform more checks on the xpub page continue vs other pages is missing | 0 |
104,356 | 11,404,699,138 | IssuesEvent | 2020-01-31 10:22:02 | OpenMined/PySyft | https://api.github.com/repos/OpenMined/PySyft | opened | Docs on adding a new message type | documentation | **Is your feature request related to a problem? Please describe.**
While implementing a new message type for syft workers, I didn't found a clear documentation on how to do so, I used some help from slack and grep utility to find what need to be updated.
**Describe the solution you'd like**
I think it would be great if we can put a doc describing how to do so, @karlhigley suggested a README file under syft/messaging, another possibility is to add a quick tutorial on this.
| 1.0 | Docs on adding a new message type - **Is your feature request related to a problem? Please describe.**
While implementing a new message type for syft workers, I didn't found a clear documentation on how to do so, I used some help from slack and grep utility to find what need to be updated.
**Describe the solution you'd like**
I think it would be great if we can put a doc describing how to do so, @karlhigley suggested a README file under syft/messaging, another possibility is to add a quick tutorial on this.
| non_priority | docs on adding a new message type is your feature request related to a problem please describe while implementing a new message type for syft workers i didn t found a clear documentation on how to do so i used some help from slack and grep utility to find what need to be updated describe the solution you d like i think it would be great if we can put a doc describing how to do so karlhigley suggested a readme file under syft messaging another possibility is to add a quick tutorial on this | 0 |
163,520 | 20,363,842,903 | IssuesEvent | 2022-02-21 01:35:08 | szb512/cypress | https://api.github.com/repos/szb512/cypress | opened | CVE-2022-0686 (Medium) detected in url-parse-1.4.7.tgz | security vulnerability | ## CVE-2022-0686 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /cypress/packages/server/package.json</p>
<p>Path to vulnerable library: /packages/server/node_modules/url-parse/package.json,/packages/server/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.8.
<p>Publish Date: 2022-02-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0686>CVE-2022-0686</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686</a></p>
<p>Release Date: 2022-02-19</p>
<p>Fix Resolution: url-parse - 1.5.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-0686 (Medium) detected in url-parse-1.4.7.tgz - ## CVE-2022-0686 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>url-parse-1.4.7.tgz</b></p></summary>
<p>Small footprint URL parser that works seamlessly across Node.js and browser environments</p>
<p>Library home page: <a href="https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz">https://registry.npmjs.org/url-parse/-/url-parse-1.4.7.tgz</a></p>
<p>Path to dependency file: /cypress/packages/server/package.json</p>
<p>Path to vulnerable library: /packages/server/node_modules/url-parse/package.json,/packages/server/node_modules/url-parse/package.json</p>
<p>
Dependency Hierarchy:
- :x: **url-parse-1.4.7.tgz** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Authorization Bypass Through User-Controlled Key in NPM url-parse prior to 1.5.8.
<p>Publish Date: 2022-02-19
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0686>CVE-2022-0686</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-0686</a></p>
<p>Release Date: 2022-02-19</p>
<p>Fix Resolution: url-parse - 1.5.8</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in url parse tgz cve medium severity vulnerability vulnerable library url parse tgz small footprint url parser that works seamlessly across node js and browser environments library home page a href path to dependency file cypress packages server package json path to vulnerable library packages server node modules url parse package json packages server node modules url parse package json dependency hierarchy x url parse tgz vulnerable library vulnerability details authorization bypass through user controlled key in npm url parse prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution url parse step up your open source security game with whitesource | 0 |
45,373 | 13,110,050,896 | IssuesEvent | 2020-08-04 19:53:27 | ronakjain2012/eShop-server | https://api.github.com/repos/ronakjain2012/eShop-server | opened | WS-2019-0019 (Medium) detected in braces-1.8.5.tgz | security vulnerability | ## WS-2019-0019 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>braces-1.8.5.tgz</b></p></summary>
<p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p>
<p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/eShop-server/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/eShop-server/node_modules/micromatch/node_modules/braces/package.json</p>
<p>
Dependency Hierarchy:
- babel-cli-6.26.0.tgz (Root Library)
- chokidar-1.7.0.tgz
- anymatch-1.3.2.tgz
- micromatch-2.3.11.tgz
- :x: **braces-1.8.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ronakjain2012/eShop-server/commit/58a30d0751202fabb10d47dc9b8578f17acf0c69">58a30d0751202fabb10d47dc9b8578f17acf0c69</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.
<p>Publish Date: 2018-02-16
<p>URL: <a href=https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451>WS-2019-0019</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p>
<p>Release Date: 2019-02-21</p>
<p>Fix Resolution: 2.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0019 (Medium) detected in braces-1.8.5.tgz - ## WS-2019-0019 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>braces-1.8.5.tgz</b></p></summary>
<p>Fastest brace expansion for node.js, with the most complete support for the Bash 4.3 braces specification.</p>
<p>Library home page: <a href="https://registry.npmjs.org/braces/-/braces-1.8.5.tgz">https://registry.npmjs.org/braces/-/braces-1.8.5.tgz</a></p>
<p>Path to dependency file: /tmp/ws-scm/eShop-server/package.json</p>
<p>Path to vulnerable library: /tmp/ws-scm/eShop-server/node_modules/micromatch/node_modules/braces/package.json</p>
<p>
Dependency Hierarchy:
- babel-cli-6.26.0.tgz (Root Library)
- chokidar-1.7.0.tgz
- anymatch-1.3.2.tgz
- micromatch-2.3.11.tgz
- :x: **braces-1.8.5.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ronakjain2012/eShop-server/commit/58a30d0751202fabb10d47dc9b8578f17acf0c69">58a30d0751202fabb10d47dc9b8578f17acf0c69</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Version of braces prior to 2.3.1 are vulnerable to Regular Expression Denial of Service (ReDoS). Untrusted input may cause catastrophic backtracking while matching regular expressions. This can cause the application to be unresponsive leading to Denial of Service.
<p>Publish Date: 2018-02-16
<p>URL: <a href=https://github.com/micromatch/braces/commit/abdafb0cae1e0c00f184abbadc692f4eaa98f451>WS-2019-0019</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/786">https://www.npmjs.com/advisories/786</a></p>
<p>Release Date: 2019-02-21</p>
<p>Fix Resolution: 2.3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in braces tgz ws medium severity vulnerability vulnerable library braces tgz fastest brace expansion for node js with the most complete support for the bash braces specification library home page a href path to dependency file tmp ws scm eshop server package json path to vulnerable library tmp ws scm eshop server node modules micromatch node modules braces package json dependency hierarchy babel cli tgz root library chokidar tgz anymatch tgz micromatch tgz x braces tgz vulnerable library found in head commit a href vulnerability details version of braces prior to are vulnerable to regular expression denial of service redos untrusted input may cause catastrophic backtracking while matching regular expressions this can cause the application to be unresponsive leading to denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
306,118 | 26,436,506,706 | IssuesEvent | 2023-01-15 13:06:46 | kubernetes/autoscaler | https://api.github.com/repos/kubernetes/autoscaler | closed | SIGSEGV in aws-sdk-go tests | test cluster-autoscaler area/provider/aws kind/bug lifecycle/rotten | <!--
Please answer these questions before submitting your bug report. Thanks!
-->
**Which component are you using?**:
Cluster Autoscaler
<!--
Which autoscaling component hosted in this repository (cluster-autoscaler, vertical-pod-autoscaler, addon-resizer, helm charts) is the bug in?
-->
**What version of the component are you using?**:
Head (master)
<!--
What version of the relevant component are you using? Either the image tag or helm chart version.
-->
**What did you expect to happen?**:
Tests pass OK
**What happened instead?**:
```
--- FAIL: TestRequestCancelRetry (0.00s)
panic: runtime error: invalid memory address or nil pointer dereference [recovered]
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x18 pc=0xa9bd39]
goroutine 435 [running]:
testing.tRunner.func1.2({0xb2c880, 0x1b00b30})
/opt/hostedtoolcache/go/1.17.5/x64/src/testing/testing.go:1209 +0x36c
testing.tRunner.func1()
/opt/hostedtoolcache/go/1.17.5/x64/src/testing/testing.go:1212 +0x3b6
panic({0xb2c880, 0x1b00b30})
/opt/hostedtoolcache/go/1.17.5/x64/src/runtime/panic.go:1047 +0x266
k8s.io/autoscaler/cluster-autoscaler/cloudprovider/aws/aws-sdk-go/aws/request_test.TestRequestCancelRetry(0xc00036[48](https://github.com/kubernetes/autoscaler/runs/7869582469?check_suite_focus=true#step:7:49)20)
/home/runner/work/autoscaler/autoscaler/go/src/k8s.io/autoscaler/cluster-autoscaler/cloudprovider/aws/aws-sdk-go/aws/request/http_request_retry_test.go:31 +0xa19
testing.tRunner(0xc000364820, 0xbd6e28)
/opt/hostedtoolcache/go/1.17.5/x64/src/testing/testing.go:1259 +0x230
created by testing.(*T).Run
/opt/hostedtoolcache/go/1.17.5/x64/src/testing/testing.go:1306 +0x727
FAIL k8s.io/autoscaler/cluster-autoscaler/cloudprovider/aws/aws-sdk-go/aws/request 16.5[50](https://github.com/kubernetes/autoscaler/runs/7869582469?check_suite_focus=true#step:7:51)s
``` | 1.0 | SIGSEGV in aws-sdk-go tests - <!--
Please answer these questions before submitting your bug report. Thanks!
-->
**Which component are you using?**:
Cluster Autoscaler
<!--
Which autoscaling component hosted in this repository (cluster-autoscaler, vertical-pod-autoscaler, addon-resizer, helm charts) is the bug in?
-->
**What version of the component are you using?**:
Head (master)
<!--
What version of the relevant component are you using? Either the image tag or helm chart version.
-->
**What did you expect to happen?**:
Tests pass OK
**What happened instead?**:
```
--- FAIL: TestRequestCancelRetry (0.00s)
panic: runtime error: invalid memory address or nil pointer dereference [recovered]
panic: runtime error: invalid memory address or nil pointer dereference
[signal SIGSEGV: segmentation violation code=0x1 addr=0x18 pc=0xa9bd39]
goroutine 435 [running]:
testing.tRunner.func1.2({0xb2c880, 0x1b00b30})
/opt/hostedtoolcache/go/1.17.5/x64/src/testing/testing.go:1209 +0x36c
testing.tRunner.func1()
/opt/hostedtoolcache/go/1.17.5/x64/src/testing/testing.go:1212 +0x3b6
panic({0xb2c880, 0x1b00b30})
/opt/hostedtoolcache/go/1.17.5/x64/src/runtime/panic.go:1047 +0x266
k8s.io/autoscaler/cluster-autoscaler/cloudprovider/aws/aws-sdk-go/aws/request_test.TestRequestCancelRetry(0xc00036[48](https://github.com/kubernetes/autoscaler/runs/7869582469?check_suite_focus=true#step:7:49)20)
/home/runner/work/autoscaler/autoscaler/go/src/k8s.io/autoscaler/cluster-autoscaler/cloudprovider/aws/aws-sdk-go/aws/request/http_request_retry_test.go:31 +0xa19
testing.tRunner(0xc000364820, 0xbd6e28)
/opt/hostedtoolcache/go/1.17.5/x64/src/testing/testing.go:1259 +0x230
created by testing.(*T).Run
/opt/hostedtoolcache/go/1.17.5/x64/src/testing/testing.go:1306 +0x727
FAIL k8s.io/autoscaler/cluster-autoscaler/cloudprovider/aws/aws-sdk-go/aws/request 16.5[50](https://github.com/kubernetes/autoscaler/runs/7869582469?check_suite_focus=true#step:7:51)s
``` | non_priority | sigsegv in aws sdk go tests please answer these questions before submitting your bug report thanks which component are you using cluster autoscaler which autoscaling component hosted in this repository cluster autoscaler vertical pod autoscaler addon resizer helm charts is the bug in what version of the component are you using head master what version of the relevant component are you using either the image tag or helm chart version what did you expect to happen tests pass ok what happened instead fail testrequestcancelretry panic runtime error invalid memory address or nil pointer dereference panic runtime error invalid memory address or nil pointer dereference goroutine testing trunner opt hostedtoolcache go src testing testing go testing trunner opt hostedtoolcache go src testing testing go panic opt hostedtoolcache go src runtime panic go io autoscaler cluster autoscaler cloudprovider aws aws sdk go aws request test testrequestcancelretry home runner work autoscaler autoscaler go src io autoscaler cluster autoscaler cloudprovider aws aws sdk go aws request http request retry test go testing trunner opt hostedtoolcache go src testing testing go created by testing t run opt hostedtoolcache go src testing testing go fail io autoscaler cluster autoscaler cloudprovider aws aws sdk go aws request | 0 |
225,010 | 24,805,683,977 | IssuesEvent | 2022-10-25 04:08:56 | samq-democorp/Umbraco-CMS | https://api.github.com/repos/samq-democorp/Umbraco-CMS | closed | Umbraco.Cms.Tests.Integration-1.0.0: 2 vulnerabilities (highest severity is: 7.5) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Umbraco.Cms.Tests.Integration-1.0.0</b></p></summary>
<p></p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/5.0.0/system.security.cryptography.xml.5.0.0.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/samq-democorp/Umbraco-CMS/commit/07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080">07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2019-0820](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | system.text.regularexpressions.4.3.0.nupkg | Transitive | N/A | ❌ |
| [CVE-2022-34716](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34716) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | system.security.cryptography.xml.5.0.0.nupkg | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0820</summary>
### Vulnerable Library - <b>system.text.regularexpressions.4.3.0.nupkg</b></p>
<p>Provides the System.Text.RegularExpressions.Regex class, an implementation of a regular expression e...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.text.regularexpressions.4.3.0.nupkg">https://api.nuget.org/packages/system.text.regularexpressions.4.3.0.nupkg</a></p>
<p>Path to dependency file: /tests/Umbraco.Tests.Benchmarks/Umbraco.Tests.Benchmarks.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.text.regularexpressions/4.3.0/system.text.regularexpressions.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- Umbraco.Cms.Tests.Integration-1.0.0 (Root Library)
- umbraco.cms.infrastructure.9.5.0.nupkg
- markdown.2.2.1.nupkg
- :x: **system.text.regularexpressions.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-democorp/Umbraco-CMS/commit/07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080">07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080</a></p>
<p>Found in base branch: <b>v9/contrib</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka '.NET Framework and .NET Core Denial of Service Vulnerability'. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981.
Mend Note: After conducting further research, Mend has determined that CVE-2019-0820 only affects environments with versions 4.3.0 and 4.3.1 only on netcore50 environment of system.text.regularexpressions.nupkg.
<p>Publish Date: 2019-05-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820>CVE-2019-0820</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cmhx-cq75-c4mj">https://github.com/advisories/GHSA-cmhx-cq75-c4mj</a></p>
<p>Release Date: 2019-05-16</p>
<p>Fix Resolution: System.Text.RegularExpressions - 4.3.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-34716</summary>
### Vulnerable Library - <b>system.security.cryptography.xml.5.0.0.nupkg</b></p>
<p>Provides classes to support the creation and validation of XML digital signatures. The classes in th...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.security.cryptography.xml.5.0.0.nupkg">https://api.nuget.org/packages/system.security.cryptography.xml.5.0.0.nupkg</a></p>
<p>Path to dependency file: /tests/Umbraco.Tests.Common/Umbraco.Tests.Common.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/5.0.0/system.security.cryptography.xml.5.0.0.nupkg</p>
<p>
Dependency Hierarchy:
- Umbraco.Cms.Tests.Integration-1.0.0 (Root Library)
- umbraco.cms.web.backoffice.9.5.0.nupkg
- umbraco.cms.examine.lucene.9.5.0.nupkg
- examine.2.0.1.nupkg
- microsoft.aspnetcore.dataprotection.5.0.5.nupkg
- :x: **system.security.cryptography.xml.5.0.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-democorp/Umbraco-CMS/commit/07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080">07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080</a></p>
<p>Found in base branch: <b>v9/contrib</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Microsoft is releasing this security advisory to provide information about a vulnerability in .NET Core 3.1 and .NET 6.0. An information disclosure vulnerability exists in .NET Core 3.1 and .NET 6.0 that could lead to unauthorized access of privileged information.
## Affected software
* Any .NET 6.0 application running on .NET 6.0.7 or earlier.
* Any .NET Core 3.1 applicaiton running on .NET Core 3.1.27 or earlier.
## Patches
* If you're using .NET 6.0, you should download and install Runtime 6.0.8 or SDK 6.0.108 (for Visual Studio 2022 v17.1) from https://dotnet.microsoft.com/download/dotnet-core/6.0.
* If you're using .NET Core 3.1, you should download and install Runtime 3.1.28 (for Visual Studio 2019 v16.9) from https://dotnet.microsoft.com/download/dotnet-core/3.1.
<p>Publish Date: 2022-08-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34716>CVE-2022-34716</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-2m65-m22p-9wjw">https://github.com/advisories/GHSA-2m65-m22p-9wjw</a></p>
<p>Release Date: 2022-08-09</p>
<p>Fix Resolution: Microsoft.AspNetCore.App.Runtime.linux-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.osx-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x86 - 3.1.28,6.0.8;System.Security.Cryptography.Xml - 4.7.1,6.0.1</p>
</p>
<p></p>
</details> | True | Umbraco.Cms.Tests.Integration-1.0.0: 2 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Umbraco.Cms.Tests.Integration-1.0.0</b></p></summary>
<p></p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/5.0.0/system.security.cryptography.xml.5.0.0.nupkg</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/samq-democorp/Umbraco-CMS/commit/07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080">07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2019-0820](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | system.text.regularexpressions.4.3.0.nupkg | Transitive | N/A | ❌ |
| [CVE-2022-34716](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34716) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.9 | system.security.cryptography.xml.5.0.0.nupkg | Transitive | N/A | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-0820</summary>
### Vulnerable Library - <b>system.text.regularexpressions.4.3.0.nupkg</b></p>
<p>Provides the System.Text.RegularExpressions.Regex class, an implementation of a regular expression e...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.text.regularexpressions.4.3.0.nupkg">https://api.nuget.org/packages/system.text.regularexpressions.4.3.0.nupkg</a></p>
<p>Path to dependency file: /tests/Umbraco.Tests.Benchmarks/Umbraco.Tests.Benchmarks.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.text.regularexpressions/4.3.0/system.text.regularexpressions.4.3.0.nupkg</p>
<p>
Dependency Hierarchy:
- Umbraco.Cms.Tests.Integration-1.0.0 (Root Library)
- umbraco.cms.infrastructure.9.5.0.nupkg
- markdown.2.2.1.nupkg
- :x: **system.text.regularexpressions.4.3.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-democorp/Umbraco-CMS/commit/07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080">07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080</a></p>
<p>Found in base branch: <b>v9/contrib</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A denial of service vulnerability exists when .NET Framework and .NET Core improperly process RegEx strings, aka '.NET Framework and .NET Core Denial of Service Vulnerability'. This CVE ID is unique from CVE-2019-0980, CVE-2019-0981.
Mend Note: After conducting further research, Mend has determined that CVE-2019-0820 only affects environments with versions 4.3.0 and 4.3.1 only on netcore50 environment of system.text.regularexpressions.nupkg.
<p>Publish Date: 2019-05-16
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-0820>CVE-2019-0820</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cmhx-cq75-c4mj">https://github.com/advisories/GHSA-cmhx-cq75-c4mj</a></p>
<p>Release Date: 2019-05-16</p>
<p>Fix Resolution: System.Text.RegularExpressions - 4.3.1</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-34716</summary>
### Vulnerable Library - <b>system.security.cryptography.xml.5.0.0.nupkg</b></p>
<p>Provides classes to support the creation and validation of XML digital signatures. The classes in th...</p>
<p>Library home page: <a href="https://api.nuget.org/packages/system.security.cryptography.xml.5.0.0.nupkg">https://api.nuget.org/packages/system.security.cryptography.xml.5.0.0.nupkg</a></p>
<p>Path to dependency file: /tests/Umbraco.Tests.Common/Umbraco.Tests.Common.csproj</p>
<p>Path to vulnerable library: /home/wss-scanner/.nuget/packages/system.security.cryptography.xml/5.0.0/system.security.cryptography.xml.5.0.0.nupkg</p>
<p>
Dependency Hierarchy:
- Umbraco.Cms.Tests.Integration-1.0.0 (Root Library)
- umbraco.cms.web.backoffice.9.5.0.nupkg
- umbraco.cms.examine.lucene.9.5.0.nupkg
- examine.2.0.1.nupkg
- microsoft.aspnetcore.dataprotection.5.0.5.nupkg
- :x: **system.security.cryptography.xml.5.0.0.nupkg** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/samq-democorp/Umbraco-CMS/commit/07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080">07d00f9f09c53bd7fd2cc157f7b57dbcbbc93080</a></p>
<p>Found in base branch: <b>v9/contrib</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
Microsoft is releasing this security advisory to provide information about a vulnerability in .NET Core 3.1 and .NET 6.0. An information disclosure vulnerability exists in .NET Core 3.1 and .NET 6.0 that could lead to unauthorized access of privileged information.
## Affected software
* Any .NET 6.0 application running on .NET 6.0.7 or earlier.
* Any .NET Core 3.1 applicaiton running on .NET Core 3.1.27 or earlier.
## Patches
* If you're using .NET 6.0, you should download and install Runtime 6.0.8 or SDK 6.0.108 (for Visual Studio 2022 v17.1) from https://dotnet.microsoft.com/download/dotnet-core/6.0.
* If you're using .NET Core 3.1, you should download and install Runtime 3.1.28 (for Visual Studio 2019 v16.9) from https://dotnet.microsoft.com/download/dotnet-core/3.1.
<p>Publish Date: 2022-08-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-34716>CVE-2022-34716</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.9</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-2m65-m22p-9wjw">https://github.com/advisories/GHSA-2m65-m22p-9wjw</a></p>
<p>Release Date: 2022-08-09</p>
<p>Fix Resolution: Microsoft.AspNetCore.App.Runtime.linux-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-musl-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.linux-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.osx-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-arm64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x64 - 3.1.28,6.0.8;Microsoft.AspNetCore.App.Runtime.win-x86 - 3.1.28,6.0.8;System.Security.Cryptography.Xml - 4.7.1,6.0.1</p>
</p>
<p></p>
</details> | non_priority | umbraco cms tests integration vulnerabilities highest severity is autoclosed vulnerable library umbraco cms tests integration path to vulnerable library home wss scanner nuget packages system security cryptography xml system security cryptography xml nupkg found in head commit a href vulnerabilities cve severity cvss dependency type fixed in remediation available high system text regularexpressions nupkg transitive n a medium system security cryptography xml nupkg transitive n a details cve vulnerable library system text regularexpressions nupkg provides the system text regularexpressions regex class an implementation of a regular expression e library home page a href path to dependency file tests umbraco tests benchmarks umbraco tests benchmarks csproj path to vulnerable library home wss scanner nuget packages system text regularexpressions system text regularexpressions nupkg dependency hierarchy umbraco cms tests integration root library umbraco cms infrastructure nupkg markdown nupkg x system text regularexpressions nupkg vulnerable library found in head commit a href found in base branch contrib vulnerability details a denial of service vulnerability exists when net framework and net core improperly process regex strings aka net framework and net core denial of service vulnerability this cve id is unique from cve cve mend note after conducting further research mend has determined that cve only affects environments with versions and only on environment of system text regularexpressions nupkg publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution system text regularexpressions cve vulnerable library system security cryptography xml nupkg provides classes to support the creation and validation of xml digital signatures the classes in th library home page a href path to dependency file tests umbraco tests common umbraco tests common csproj path to vulnerable library home wss scanner nuget packages system security cryptography xml system security cryptography xml nupkg dependency hierarchy umbraco cms tests integration root library umbraco cms web backoffice nupkg umbraco cms examine lucene nupkg examine nupkg microsoft aspnetcore dataprotection nupkg x system security cryptography xml nupkg vulnerable library found in head commit a href found in base branch contrib vulnerability details microsoft is releasing this security advisory to provide information about a vulnerability in net core and net an information disclosure vulnerability exists in net core and net that could lead to unauthorized access of privileged information affected software any net application running on net or earlier any net core applicaiton running on net core or earlier patches if you re using net you should download and install runtime or sdk for visual studio from if you re using net core you should download and install runtime for visual studio from publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft aspnetcore app runtime linux arm microsoft aspnetcore app runtime linux microsoft aspnetcore app runtime linux musl arm microsoft aspnetcore app runtime linux musl microsoft aspnetcore app runtime linux musl microsoft aspnetcore app runtime linux microsoft aspnetcore app runtime osx microsoft aspnetcore app runtime win arm microsoft aspnetcore app runtime win microsoft aspnetcore app runtime win microsoft aspnetcore app runtime win system security cryptography xml | 0 |
172,107 | 14,350,332,429 | IssuesEvent | 2020-11-29 20:25:06 | andrewdimmer/csi-3370-software-project | https://api.github.com/repos/andrewdimmer/csi-3370-software-project | closed | [Display_01] Define Testing Plan for Display_01 Production Classes | Display_01 documentation | **What is the use case? Please describe.**
For test driven development, it is important to determine a test plan before just jumping in and writing tests and code. For this sprint task, you'll be developing a testing plan for all of the classes in the `edu.oakland.production.display01` package. Note: This issue is similar to #51, but is tracked differently as helper and production classes have different deliverable dates.
**Prerequisites**
Please makes sure that #38 is closed and merged to `main` before starting this, to ensure that all of the documentation you need for this is in the correct spot.
**Sprint Task Setup**
To be done by @EC01EC , @stevecatherman , or @casher-ou before the rest of the team starts working on it:
1. Use the documentation at https://github.com/andrewdimmer/csi-3370-software-project/tree/main/docs/uml-class-diagrams/display01/production to identify the classes that require a test plan.
2. Edit this issue message to add a checkbox for each class, and assign it to the correct team member(s). Be sure to balance your team between this issue at #51 . Note that the teams assigned here should also be the same teams in charge of building the appropriate classes in #67 .
**Software Setup**
For this task, you'll be updating PlantUML diagrams. To preview them on your local machine, set up:
- [Visual Studio Code](https://code.visualstudio.com/)
- [PlantUML Visual Studio Code Extension](https://marketplace.visualstudio.com/items?itemName=jebbs.plantuml)
- [Graphviz](https://graphviz.org/)
After that, you can open a .puml document, right click on it, and "Preview Current Diagram".
If you don't want to set that up though, you don't have to. When you commit, GitHub has been configured to automatically update all the diagrams, and you can push to GitHub then view the diagrams from there and/or pull them back to your current machine.
**Steps**
To be done by each team member for each class they are developing a test plan for:
1. Each team member should watch the first ~9 minutes of @andrewdimmer 's [Test Driven Development Crash Course](https://youtu.be/eSc0aY4GhPs)
2. One person should create a branch off of main called `define-Name_Of_Test_Class` (ex. the `Demo` test plan gets defined on a branch called `define-DemoTests`). Everyone else in the programming pair or group should then checkout that branch.
3. Open the Test Class UML Diagram at `/docs/uml-class-diagrams/display01/test/Name_Of_Test_Class/Name_Of_Test_Class.puml`.
4. Using the information in the video, record your tests on the Test Class UML Diagram you opened in step 3. Need help with the PlantUML Diagram itself? Check out https://plantuml.com/class-diagram. Feel free to use the existing class diagrams already provided at https://github.com/andrewdimmer/csi-3370-software-project/tree/main/docs/uml-class-diagrams/display01/production to help guide you in creating the test classes!
5. Commit your changes with the message `[Display_01] Define Name_Of_Test_Class`
6. Push your changes
7. Create a pull request with the name `[Display_01] Define Name_Of_Test_Class` and the rest of the information like description, tags, etc. For reference, the settings on the right hand side should match this issue, with the exception that the assignees on the PR are just those who are pushing code.
8. Once your test plan is complete, add @andrewdimmer as a reviewer.
**Tasks to Complete**
- [x] Allocate Team Members to Classes - @EC01EC , @stevecatherman , or @casher-ou
- [x] Define DisplayGpsInterfaceImplementationTests - @reedsikorskiOU or @DLidz
- [x] Define DisplayGpsManagerImplementationTests - @atpoynter or @elilley15
- [x] Define DisplayGpsReceiverImplementationTests - @EC01EC or @casher-ou
**Additional context**
Note: if for some reason you need to change a public interface or shared data type, reach out to @andrewdimmer who can help you manage making that BREAKING CHANGE.
Note: if you run into merge conflicts, let @andrewdimmer know and he can help you resolve them. | 1.0 | [Display_01] Define Testing Plan for Display_01 Production Classes - **What is the use case? Please describe.**
For test driven development, it is important to determine a test plan before just jumping in and writing tests and code. For this sprint task, you'll be developing a testing plan for all of the classes in the `edu.oakland.production.display01` package. Note: This issue is similar to #51, but is tracked differently as helper and production classes have different deliverable dates.
**Prerequisites**
Please makes sure that #38 is closed and merged to `main` before starting this, to ensure that all of the documentation you need for this is in the correct spot.
**Sprint Task Setup**
To be done by @EC01EC , @stevecatherman , or @casher-ou before the rest of the team starts working on it:
1. Use the documentation at https://github.com/andrewdimmer/csi-3370-software-project/tree/main/docs/uml-class-diagrams/display01/production to identify the classes that require a test plan.
2. Edit this issue message to add a checkbox for each class, and assign it to the correct team member(s). Be sure to balance your team between this issue at #51 . Note that the teams assigned here should also be the same teams in charge of building the appropriate classes in #67 .
**Software Setup**
For this task, you'll be updating PlantUML diagrams. To preview them on your local machine, set up:
- [Visual Studio Code](https://code.visualstudio.com/)
- [PlantUML Visual Studio Code Extension](https://marketplace.visualstudio.com/items?itemName=jebbs.plantuml)
- [Graphviz](https://graphviz.org/)
After that, you can open a .puml document, right click on it, and "Preview Current Diagram".
If you don't want to set that up though, you don't have to. When you commit, GitHub has been configured to automatically update all the diagrams, and you can push to GitHub then view the diagrams from there and/or pull them back to your current machine.
**Steps**
To be done by each team member for each class they are developing a test plan for:
1. Each team member should watch the first ~9 minutes of @andrewdimmer 's [Test Driven Development Crash Course](https://youtu.be/eSc0aY4GhPs)
2. One person should create a branch off of main called `define-Name_Of_Test_Class` (ex. the `Demo` test plan gets defined on a branch called `define-DemoTests`). Everyone else in the programming pair or group should then checkout that branch.
3. Open the Test Class UML Diagram at `/docs/uml-class-diagrams/display01/test/Name_Of_Test_Class/Name_Of_Test_Class.puml`.
4. Using the information in the video, record your tests on the Test Class UML Diagram you opened in step 3. Need help with the PlantUML Diagram itself? Check out https://plantuml.com/class-diagram. Feel free to use the existing class diagrams already provided at https://github.com/andrewdimmer/csi-3370-software-project/tree/main/docs/uml-class-diagrams/display01/production to help guide you in creating the test classes!
5. Commit your changes with the message `[Display_01] Define Name_Of_Test_Class`
6. Push your changes
7. Create a pull request with the name `[Display_01] Define Name_Of_Test_Class` and the rest of the information like description, tags, etc. For reference, the settings on the right hand side should match this issue, with the exception that the assignees on the PR are just those who are pushing code.
8. Once your test plan is complete, add @andrewdimmer as a reviewer.
**Tasks to Complete**
- [x] Allocate Team Members to Classes - @EC01EC , @stevecatherman , or @casher-ou
- [x] Define DisplayGpsInterfaceImplementationTests - @reedsikorskiOU or @DLidz
- [x] Define DisplayGpsManagerImplementationTests - @atpoynter or @elilley15
- [x] Define DisplayGpsReceiverImplementationTests - @EC01EC or @casher-ou
**Additional context**
Note: if for some reason you need to change a public interface or shared data type, reach out to @andrewdimmer who can help you manage making that BREAKING CHANGE.
Note: if you run into merge conflicts, let @andrewdimmer know and he can help you resolve them. | non_priority | define testing plan for display production classes what is the use case please describe for test driven development it is important to determine a test plan before just jumping in and writing tests and code for this sprint task you ll be developing a testing plan for all of the classes in the edu oakland production package note this issue is similar to but is tracked differently as helper and production classes have different deliverable dates prerequisites please makes sure that is closed and merged to main before starting this to ensure that all of the documentation you need for this is in the correct spot sprint task setup to be done by stevecatherman or casher ou before the rest of the team starts working on it use the documentation at to identify the classes that require a test plan edit this issue message to add a checkbox for each class and assign it to the correct team member s be sure to balance your team between this issue at note that the teams assigned here should also be the same teams in charge of building the appropriate classes in software setup for this task you ll be updating plantuml diagrams to preview them on your local machine set up after that you can open a puml document right click on it and preview current diagram if you don t want to set that up though you don t have to when you commit github has been configured to automatically update all the diagrams and you can push to github then view the diagrams from there and or pull them back to your current machine steps to be done by each team member for each class they are developing a test plan for each team member should watch the first minutes of andrewdimmer s one person should create a branch off of main called define name of test class ex the demo test plan gets defined on a branch called define demotests everyone else in the programming pair or group should then checkout that branch open the test class uml diagram at docs uml class diagrams test name of test class name of test class puml using the information in the video record your tests on the test class uml diagram you opened in step need help with the plantuml diagram itself check out feel free to use the existing class diagrams already provided at to help guide you in creating the test classes commit your changes with the message define name of test class push your changes create a pull request with the name define name of test class and the rest of the information like description tags etc for reference the settings on the right hand side should match this issue with the exception that the assignees on the pr are just those who are pushing code once your test plan is complete add andrewdimmer as a reviewer tasks to complete allocate team members to classes stevecatherman or casher ou define displaygpsinterfaceimplementationtests reedsikorskiou or dlidz define displaygpsmanagerimplementationtests atpoynter or define displaygpsreceiverimplementationtests or casher ou additional context note if for some reason you need to change a public interface or shared data type reach out to andrewdimmer who can help you manage making that breaking change note if you run into merge conflicts let andrewdimmer know and he can help you resolve them | 0 |
39,131 | 5,219,459,405 | IssuesEvent | 2017-01-26 19:09:03 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [Tests] TopHitsAggregatorTests#testInsideTerms fails reproducibly | :Aggregations test v6.0.0 | This one fails reproducibly on master (2abe948cd7b492f583cae066f59904c0f4253d20):
https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+5.x+multijob-unix-compatibility/os=ubuntu/475
'''
gradle :core:test -Dtests.seed=E1AB0588265D530F -Dtests.class=org.elasticsearch.search.aggregations.metrics.tophits.TopHitsAggregatorTests -Dtests.method="testInsideTerms" -Dtests.security.manager=true -Dtests.locale=en-IE -Dtests.timezone=Israel
'''
Error looks like this:
```
ERROR 0.04s J1 | TopHitsAggregatorTests.testInsideTerms <<< FAILURES!
> Throwable #1: java.lang.ClassCastException: org.apache.lucene.index.ParallelCompositeReader cannot be cast to org.apache.lucene.index.DirectoryReader
> at __randomizedtesting.SeedInfo.seed([E1AB0588265D530F:11E1E897EE56B196]:0)
> at org.elasticsearch.search.aggregations.support.ValuesSource$Bytes$WithOrdinals$FieldData.globalOrdinalsValues(ValuesSource.java:150)
> at org.elasticsearch.search.aggregations.support.ValuesSource$Bytes$WithOrdinals.globalMaxOrd(ValuesSource.java:123)
> at org.elasticsearch.search.aggregations.bucket.terms.TermsAggregatorFactory.doCreateInternal(TermsAggregatorFactory.java:118)
> at org.elasticsearch.search.aggregations.support.ValuesSourceAggregatorFactory.createInternal(ValuesSourceAggregatorFactory.java:55)
> at org.elasticsearch.search.aggregations.AggregatorFactory.create(AggregatorFactory.java:221)
> at org.elasticsearch.search.aggregations.AggregatorTestCase.createAggregator(AggregatorTestCase.java:121)
> at org.elasticsearch.search.aggregations.AggregatorTestCase.searchAndReduce(AggregatorTestCase.java:167)
> at org.elasticsearch.search.aggregations.metrics.tophits.TopHitsAggregatorTests.testCase(TopHitsAggregatorTests.java:135)
> at org.elasticsearch.search.aggregations.metrics.tophits.TopHitsAggregatorTests.testInsideTerms(TopHitsAggregatorTests.java:83)
> at java.lang.Thread.run(Thread.java:745)
```
| 1.0 | [Tests] TopHitsAggregatorTests#testInsideTerms fails reproducibly - This one fails reproducibly on master (2abe948cd7b492f583cae066f59904c0f4253d20):
https://elasticsearch-ci.elastic.co/job/elastic+elasticsearch+5.x+multijob-unix-compatibility/os=ubuntu/475
'''
gradle :core:test -Dtests.seed=E1AB0588265D530F -Dtests.class=org.elasticsearch.search.aggregations.metrics.tophits.TopHitsAggregatorTests -Dtests.method="testInsideTerms" -Dtests.security.manager=true -Dtests.locale=en-IE -Dtests.timezone=Israel
'''
Error looks like this:
```
ERROR 0.04s J1 | TopHitsAggregatorTests.testInsideTerms <<< FAILURES!
> Throwable #1: java.lang.ClassCastException: org.apache.lucene.index.ParallelCompositeReader cannot be cast to org.apache.lucene.index.DirectoryReader
> at __randomizedtesting.SeedInfo.seed([E1AB0588265D530F:11E1E897EE56B196]:0)
> at org.elasticsearch.search.aggregations.support.ValuesSource$Bytes$WithOrdinals$FieldData.globalOrdinalsValues(ValuesSource.java:150)
> at org.elasticsearch.search.aggregations.support.ValuesSource$Bytes$WithOrdinals.globalMaxOrd(ValuesSource.java:123)
> at org.elasticsearch.search.aggregations.bucket.terms.TermsAggregatorFactory.doCreateInternal(TermsAggregatorFactory.java:118)
> at org.elasticsearch.search.aggregations.support.ValuesSourceAggregatorFactory.createInternal(ValuesSourceAggregatorFactory.java:55)
> at org.elasticsearch.search.aggregations.AggregatorFactory.create(AggregatorFactory.java:221)
> at org.elasticsearch.search.aggregations.AggregatorTestCase.createAggregator(AggregatorTestCase.java:121)
> at org.elasticsearch.search.aggregations.AggregatorTestCase.searchAndReduce(AggregatorTestCase.java:167)
> at org.elasticsearch.search.aggregations.metrics.tophits.TopHitsAggregatorTests.testCase(TopHitsAggregatorTests.java:135)
> at org.elasticsearch.search.aggregations.metrics.tophits.TopHitsAggregatorTests.testInsideTerms(TopHitsAggregatorTests.java:83)
> at java.lang.Thread.run(Thread.java:745)
```
| non_priority | tophitsaggregatortests testinsideterms fails reproducibly this one fails reproducibly on master gradle core test dtests seed dtests class org elasticsearch search aggregations metrics tophits tophitsaggregatortests dtests method testinsideterms dtests security manager true dtests locale en ie dtests timezone israel error looks like this error tophitsaggregatortests testinsideterms failures throwable java lang classcastexception org apache lucene index parallelcompositereader cannot be cast to org apache lucene index directoryreader at randomizedtesting seedinfo seed at org elasticsearch search aggregations support valuessource bytes withordinals fielddata globalordinalsvalues valuessource java at org elasticsearch search aggregations support valuessource bytes withordinals globalmaxord valuessource java at org elasticsearch search aggregations bucket terms termsaggregatorfactory docreateinternal termsaggregatorfactory java at org elasticsearch search aggregations support valuessourceaggregatorfactory createinternal valuessourceaggregatorfactory java at org elasticsearch search aggregations aggregatorfactory create aggregatorfactory java at org elasticsearch search aggregations aggregatortestcase createaggregator aggregatortestcase java at org elasticsearch search aggregations aggregatortestcase searchandreduce aggregatortestcase java at org elasticsearch search aggregations metrics tophits tophitsaggregatortests testcase tophitsaggregatortests java at org elasticsearch search aggregations metrics tophits tophitsaggregatortests testinsideterms tophitsaggregatortests java at java lang thread run thread java | 0 |
314,747 | 23,535,530,094 | IssuesEvent | 2022-08-19 20:13:56 | scipy/scipy | https://api.github.com/repos/scipy/scipy | closed | Improve documentation for smoothing in interpolate.UnivariateSpline and interpolate.splrep | scipy.interpolate Documentation | **Problem:**
For [`scipy.interpolate.UnivariateSpline`](https://docs.scipy.org/doc/scipy/reference/generated/scipy.interpolate.UnivariateSpline.html#scipy.interpolate.UnivariateSpline) and [`scipy.interpolate.splrep`](https://docs.scipy.org/doc/scipy/reference/generated/scipy.interpolate.splrep.html#scipy.interpolate.splrep), the user is asked to supply a smoothing factor (`s`) that is used to smooth the spline fit through noisy data. At first glance, this is easy, because `s=0` provides no smoothing, only interpolation. A larger value of `s` provides more smoothing. However, in practice, I have been finding `s` through a tedious process of trial and error, and I imagine that many users are doing the same. As discussed in https://github.com/scipy/scipy/issues/4975#issuecomment-112870150 and (more recently) in
https://github.com/PyAbel/PyAbel/issues/276#issuecomment-616966199, the documentation for how to select an appropriate `s` is somewhat confusing.
In the case where users correctly input the weights array (`w`) as the inverse of the standard deviation, then `s` is correctly set to `s = len(w)`, which is a good starting point and covered in the documentation. However, for users that use the default weights array (`w = ones(len(x))`), then the default `s` is much too high and they aren't offered any guidance on how to select an appropriate `s`.
**Suggested solution:**
As discussed in the docs for [splrep ](https://docs.scipy.org/doc/scipy-0.14.0/reference/generated/scipy.interpolate.splrep.html#scipy.interpolate.splrep), the amount of smoothness is given by `sum((w * (y - g))**2,axis=0) <= s`, where w is the weights, y is the data, and g is the spline. So, if the weights are equal to 1 everywhere, then we are left with `sum((y-g)**2) <= s`, which is `(number of points) * (standard dev of residuals)**2`. If we assume that we want the standard deviation of the fit residuals to be equal to the noise (standard deviation) of the data, then we would want `s = (number of points) * (standard deviation)**2`.
So, I think that the documentation for `s` should include something like the following:
"If the weights are all set to 1 (default), then the appropriate value of `s` will be somewhere near len(x)*sd**2, where `sd` is is an estimate of the standard deviation of the data points." Another sentence should be included to emphasize that the user *must* set either the `w` or the `s` to achieve sensible results.
In addition, the documentation for `UnivariateSpline` should mention that the default value of the `w` array is `ones(len(x))`. `splrep` already does this, but `UnivariateSpline` just says that "If None (default), weights are all equal.". Since the absolute value of the weights matters for the selection of `s`, it's important to say what the actual value of `w[i]` is.
**Alternatives:**
I don't think that these are great ideas, but since the form asks for alternatives, here is what comes to mind:
1) More invasive: One could argue that UnivariateSpline should work reasonably, out-of-the-box, without the user needing to input the weights or smoothing variables. This would amount to setting the default `s = len(x) * 1e-4`, or assuming a 1% error in the datapoints.
2) More invasive: the functions could show a warning if the user hasn't set either `w` or `s`.
**Side-note:**
`splrep` sets the default `s = len(x) - sqrt(2*(len(x))` while `UnivariateSpline` sets the default `s = len(x)`. These functions are both fitting the same splines, right? Shouldn't they use the same default? Their documentation is also quite different, and it might make sense to take the best of both docs and combine them. Pragmatically, I'm guessing that most users will be using UnivariateSpline (it's the most user-friendly), but splrep has more comprehensive documentation.
EDIT: Here is a previous issue from many years ago when `s = None` by default: https://github.com/scipy/scipy/issues/726
EDIT2: Here is the link to the section of the FITPACK code where there is a lengthy (but fairly confusing) discussion of the smoothing factor: https://github.com/scipy/scipy/blob/master/scipy/interpolate/fitpack/curfit.f#L141 | 1.0 | Improve documentation for smoothing in interpolate.UnivariateSpline and interpolate.splrep - **Problem:**
For [`scipy.interpolate.UnivariateSpline`](https://docs.scipy.org/doc/scipy/reference/generated/scipy.interpolate.UnivariateSpline.html#scipy.interpolate.UnivariateSpline) and [`scipy.interpolate.splrep`](https://docs.scipy.org/doc/scipy/reference/generated/scipy.interpolate.splrep.html#scipy.interpolate.splrep), the user is asked to supply a smoothing factor (`s`) that is used to smooth the spline fit through noisy data. At first glance, this is easy, because `s=0` provides no smoothing, only interpolation. A larger value of `s` provides more smoothing. However, in practice, I have been finding `s` through a tedious process of trial and error, and I imagine that many users are doing the same. As discussed in https://github.com/scipy/scipy/issues/4975#issuecomment-112870150 and (more recently) in
https://github.com/PyAbel/PyAbel/issues/276#issuecomment-616966199, the documentation for how to select an appropriate `s` is somewhat confusing.
In the case where users correctly input the weights array (`w`) as the inverse of the standard deviation, then `s` is correctly set to `s = len(w)`, which is a good starting point and covered in the documentation. However, for users that use the default weights array (`w = ones(len(x))`), then the default `s` is much too high and they aren't offered any guidance on how to select an appropriate `s`.
**Suggested solution:**
As discussed in the docs for [splrep ](https://docs.scipy.org/doc/scipy-0.14.0/reference/generated/scipy.interpolate.splrep.html#scipy.interpolate.splrep), the amount of smoothness is given by `sum((w * (y - g))**2,axis=0) <= s`, where w is the weights, y is the data, and g is the spline. So, if the weights are equal to 1 everywhere, then we are left with `sum((y-g)**2) <= s`, which is `(number of points) * (standard dev of residuals)**2`. If we assume that we want the standard deviation of the fit residuals to be equal to the noise (standard deviation) of the data, then we would want `s = (number of points) * (standard deviation)**2`.
So, I think that the documentation for `s` should include something like the following:
"If the weights are all set to 1 (default), then the appropriate value of `s` will be somewhere near len(x)*sd**2, where `sd` is is an estimate of the standard deviation of the data points." Another sentence should be included to emphasize that the user *must* set either the `w` or the `s` to achieve sensible results.
In addition, the documentation for `UnivariateSpline` should mention that the default value of the `w` array is `ones(len(x))`. `splrep` already does this, but `UnivariateSpline` just says that "If None (default), weights are all equal.". Since the absolute value of the weights matters for the selection of `s`, it's important to say what the actual value of `w[i]` is.
**Alternatives:**
I don't think that these are great ideas, but since the form asks for alternatives, here is what comes to mind:
1) More invasive: One could argue that UnivariateSpline should work reasonably, out-of-the-box, without the user needing to input the weights or smoothing variables. This would amount to setting the default `s = len(x) * 1e-4`, or assuming a 1% error in the datapoints.
2) More invasive: the functions could show a warning if the user hasn't set either `w` or `s`.
**Side-note:**
`splrep` sets the default `s = len(x) - sqrt(2*(len(x))` while `UnivariateSpline` sets the default `s = len(x)`. These functions are both fitting the same splines, right? Shouldn't they use the same default? Their documentation is also quite different, and it might make sense to take the best of both docs and combine them. Pragmatically, I'm guessing that most users will be using UnivariateSpline (it's the most user-friendly), but splrep has more comprehensive documentation.
EDIT: Here is a previous issue from many years ago when `s = None` by default: https://github.com/scipy/scipy/issues/726
EDIT2: Here is the link to the section of the FITPACK code where there is a lengthy (but fairly confusing) discussion of the smoothing factor: https://github.com/scipy/scipy/blob/master/scipy/interpolate/fitpack/curfit.f#L141 | non_priority | improve documentation for smoothing in interpolate univariatespline and interpolate splrep problem for and the user is asked to supply a smoothing factor s that is used to smooth the spline fit through noisy data at first glance this is easy because s provides no smoothing only interpolation a larger value of s provides more smoothing however in practice i have been finding s through a tedious process of trial and error and i imagine that many users are doing the same as discussed in and more recently in the documentation for how to select an appropriate s is somewhat confusing in the case where users correctly input the weights array w as the inverse of the standard deviation then s is correctly set to s len w which is a good starting point and covered in the documentation however for users that use the default weights array w ones len x then the default s is much too high and they aren t offered any guidance on how to select an appropriate s suggested solution as discussed in the docs for the amount of smoothness is given by sum w y g axis s where w is the weights y is the data and g is the spline so if the weights are equal to everywhere then we are left with sum y g s which is number of points standard dev of residuals if we assume that we want the standard deviation of the fit residuals to be equal to the noise standard deviation of the data then we would want s number of points standard deviation so i think that the documentation for s should include something like the following if the weights are all set to default then the appropriate value of s will be somewhere near len x sd where sd is is an estimate of the standard deviation of the data points another sentence should be included to emphasize that the user must set either the w or the s to achieve sensible results in addition the documentation for univariatespline should mention that the default value of the w array is ones len x splrep already does this but univariatespline just says that if none default weights are all equal since the absolute value of the weights matters for the selection of s it s important to say what the actual value of w is alternatives i don t think that these are great ideas but since the form asks for alternatives here is what comes to mind more invasive one could argue that univariatespline should work reasonably out of the box without the user needing to input the weights or smoothing variables this would amount to setting the default s len x or assuming a error in the datapoints more invasive the functions could show a warning if the user hasn t set either w or s side note splrep sets the default s len x sqrt len x while univariatespline sets the default s len x these functions are both fitting the same splines right shouldn t they use the same default their documentation is also quite different and it might make sense to take the best of both docs and combine them pragmatically i m guessing that most users will be using univariatespline it s the most user friendly but splrep has more comprehensive documentation edit here is a previous issue from many years ago when s none by default here is the link to the section of the fitpack code where there is a lengthy but fairly confusing discussion of the smoothing factor | 0 |
436,549 | 30,558,219,365 | IssuesEvent | 2023-07-20 13:05:22 | nearform/bioconductor.org | https://api.github.com/repos/nearform/bioconductor.org | closed | Review links to prevent orphaned pages | documentation | As we start to restructure the site it's important to ensure we're not losing links to pages. Producing a sitemap will enable us to understand what's linked from where, and ensure that pages aren't orphaned.
## Acceptance Criteria
* Produce a list of all the pages with number of inlinks
* Produce a list of all the pages with their outlinks | 1.0 | Review links to prevent orphaned pages - As we start to restructure the site it's important to ensure we're not losing links to pages. Producing a sitemap will enable us to understand what's linked from where, and ensure that pages aren't orphaned.
## Acceptance Criteria
* Produce a list of all the pages with number of inlinks
* Produce a list of all the pages with their outlinks | non_priority | review links to prevent orphaned pages as we start to restructure the site it s important to ensure we re not losing links to pages producing a sitemap will enable us to understand what s linked from where and ensure that pages aren t orphaned acceptance criteria produce a list of all the pages with number of inlinks produce a list of all the pages with their outlinks | 0 |
143,206 | 21,958,836,370 | IssuesEvent | 2022-05-24 14:16:05 | protocol/retriev | https://api.github.com/repos/protocol/retriev | opened | Naming [product design step 0] | dif/medium kind/discussion effort/hours need/author-input topic/product-design | We’re having a serie of issues related to the design of the visual identity and UX of the alpha.
**_Step 0_** is a brainstorming session hold by @irenegia and @nicola to come up with narrative/metaphor to communicate the product and identify each part/moment of the protocol!
This will help us set the limits and directions for the next stage of design.
@claudio (that I’m adding to the repository) will join this process to better learn about the protocol itself.
| 1.0 | Naming [product design step 0] - We’re having a serie of issues related to the design of the visual identity and UX of the alpha.
**_Step 0_** is a brainstorming session hold by @irenegia and @nicola to come up with narrative/metaphor to communicate the product and identify each part/moment of the protocol!
This will help us set the limits and directions for the next stage of design.
@claudio (that I’m adding to the repository) will join this process to better learn about the protocol itself.
| non_priority | naming we’re having a serie of issues related to the design of the visual identity and ux of the alpha step is a brainstorming session hold by irenegia and nicola to come up with narrative metaphor to communicate the product and identify each part moment of the protocol this will help us set the limits and directions for the next stage of design claudio that i’m adding to the repository will join this process to better learn about the protocol itself | 0 |
447,680 | 31,719,137,770 | IssuesEvent | 2023-09-10 07:41:28 | manojadams/mui-forms | https://api.github.com/repos/manojadams/mui-forms | closed | Checkbox control: Add support for boolean value | documentation enhancement | ### Checkbox should support boolean values from list of options:
For example:
```
[{
"label": "Yes",
"value": true
}, {
"label": "No",
"value": false
}]
``` | 1.0 | Checkbox control: Add support for boolean value - ### Checkbox should support boolean values from list of options:
For example:
```
[{
"label": "Yes",
"value": true
}, {
"label": "No",
"value": false
}]
``` | non_priority | checkbox control add support for boolean value checkbox should support boolean values from list of options for example label yes value true label no value false | 0 |
5,024 | 3,491,156,324 | IssuesEvent | 2016-01-04 14:26:36 | Dragovorn/MCCW-Reborn | https://api.github.com/repos/Dragovorn/MCCW-Reborn | closed | Add Town Hall | Build Team In Development To-Do | The build team is working on this, I need to make them a github account so people know they are working on it. | 1.0 | Add Town Hall - The build team is working on this, I need to make them a github account so people know they are working on it. | non_priority | add town hall the build team is working on this i need to make them a github account so people know they are working on it | 0 |
47,973 | 13,298,461,911 | IssuesEvent | 2020-08-25 08:18:07 | YauheniPo/Atlas_Popo_GUI_Framework | https://api.github.com/repos/YauheniPo/Atlas_Popo_GUI_Framework | opened | WS-2019-0379 (Medium) detected in commons-codec-1.10.jar | security vulnerability | ## WS-2019-0379 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: /tmp/ws-scm/Atlas_Popo_GUI_Framework/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p>
<p>
Dependency Hierarchy:
- webdrivermanager-3.4.0.jar (Root Library)
- httpclient-4.5.6.jar
- :x: **commons-codec-1.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/YauheniPo/Atlas_Popo_GUI_Framework/commit/865d1eab5956d1b38c4acaaa788e8fb4996089b5">865d1eab5956d1b38c4acaaa788e8fb4996089b5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p>
<p>Release Date: 2019-05-12</p>
<p>Fix Resolution: 1.13-RC1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0379 (Medium) detected in commons-codec-1.10.jar - ## WS-2019-0379 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.10.jar</b></p></summary>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: /tmp/ws-scm/Atlas_Popo_GUI_Framework/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/commons-codec/commons-codec/1.10/commons-codec-1.10.jar</p>
<p>
Dependency Hierarchy:
- webdrivermanager-3.4.0.jar (Root Library)
- httpclient-4.5.6.jar
- :x: **commons-codec-1.10.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/YauheniPo/Atlas_Popo_GUI_Framework/commit/865d1eab5956d1b38c4acaaa788e8fb4996089b5">865d1eab5956d1b38c4acaaa788e8fb4996089b5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p>
<p>Release Date: 2019-05-12</p>
<p>Fix Resolution: 1.13-RC1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file tmp ws scm atlas popo gui framework pom xml path to vulnerable library home wss scanner repository commons codec commons codec commons codec jar dependency hierarchy webdrivermanager jar root library httpclient jar x commons codec jar vulnerable library found in head commit a href vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
57,049 | 8,138,598,601 | IssuesEvent | 2018-08-20 15:10:02 | pyca/cryptography | https://api.github.com/repos/pyca/cryptography | opened | Add an FAQ about abi3 wheels | documentation good first issue | There seems to be a lot of bugs about these lately.
Let's have an FAQ we can point people at. | 1.0 | Add an FAQ about abi3 wheels - There seems to be a lot of bugs about these lately.
Let's have an FAQ we can point people at. | non_priority | add an faq about wheels there seems to be a lot of bugs about these lately let s have an faq we can point people at | 0 |
159,943 | 20,085,940,171 | IssuesEvent | 2022-02-05 01:15:24 | logbie/HyperGAN | https://api.github.com/repos/logbie/HyperGAN | opened | CVE-2022-21727 (High) detected in tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl | security vulnerability | ## CVE-2022-21727 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /HyperGAN/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Tensorflow is an Open Source Machine Learning Framework. The implementation of shape inference for `Dequantize` is vulnerable to an integer overflow weakness. The `axis` argument can be `-1` (the default value for the optional argument) or any other positive value at most the number of dimensions of the input. Unfortunately, the upper bound is not checked, and, since the code computes `axis + 1`, an attacker can trigger an integer overflow. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range.
<p>Publish Date: 2022-02-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21727>CVE-2022-21727</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-c6fh-56w7-fvjw">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-c6fh-56w7-fvjw</a></p>
<p>Release Date: 2022-02-03</p>
<p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1;tensorflow-cpu - 2.5.3,2.6.3,2.7.1;tensorflow-gpu - 2.5.3,2.6.3,2.7.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-21727 (High) detected in tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl - ## CVE-2022-21727 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</b></p></summary>
<p>TensorFlow is an open source machine learning framework for everyone.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl">https://files.pythonhosted.org/packages/0a/93/c7bca39b23aae45cd2e85ad3871c81eccc63b9c5276e926511e2e5b0879d/tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /HyperGAN/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **tensorflow_gpu-2.1.0-cp36-cp36m-manylinux2010_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Tensorflow is an Open Source Machine Learning Framework. The implementation of shape inference for `Dequantize` is vulnerable to an integer overflow weakness. The `axis` argument can be `-1` (the default value for the optional argument) or any other positive value at most the number of dimensions of the input. Unfortunately, the upper bound is not checked, and, since the code computes `axis + 1`, an attacker can trigger an integer overflow. The fix will be included in TensorFlow 2.8.0. We will also cherrypick this commit on TensorFlow 2.7.1, TensorFlow 2.6.3, and TensorFlow 2.5.3, as these are also affected and still in supported range.
<p>Publish Date: 2022-02-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-21727>CVE-2022-21727</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.6</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/tensorflow/tensorflow/security/advisories/GHSA-c6fh-56w7-fvjw">https://github.com/tensorflow/tensorflow/security/advisories/GHSA-c6fh-56w7-fvjw</a></p>
<p>Release Date: 2022-02-03</p>
<p>Fix Resolution: tensorflow - 2.5.3,2.6.3,2.7.1;tensorflow-cpu - 2.5.3,2.6.3,2.7.1;tensorflow-gpu - 2.5.3,2.6.3,2.7.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve high detected in tensorflow gpu whl cve high severity vulnerability vulnerable library tensorflow gpu whl tensorflow is an open source machine learning framework for everyone library home page a href path to dependency file requirements txt path to vulnerable library hypergan requirements txt dependency hierarchy x tensorflow gpu whl vulnerable library vulnerability details tensorflow is an open source machine learning framework the implementation of shape inference for dequantize is vulnerable to an integer overflow weakness the axis argument can be the default value for the optional argument or any other positive value at most the number of dimensions of the input unfortunately the upper bound is not checked and since the code computes axis an attacker can trigger an integer overflow the fix will be included in tensorflow we will also cherrypick this commit on tensorflow tensorflow and tensorflow as these are also affected and still in supported range publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution tensorflow tensorflow cpu tensorflow gpu step up your open source security game with whitesource | 0 |
100,378 | 11,188,778,074 | IssuesEvent | 2020-01-02 06:58:58 | VikasPandey121/Digital-Clock | https://api.github.com/repos/VikasPandey121/Digital-Clock | opened | Update README.md | documentation good first issue | Your task is to :
- Go through the README.md File and lookup for the grammatical error or any missing information
- Attach screenshots of all screens in a proper format in README.md file | 1.0 | Update README.md - Your task is to :
- Go through the README.md File and lookup for the grammatical error or any missing information
- Attach screenshots of all screens in a proper format in README.md file | non_priority | update readme md your task is to go through the readme md file and lookup for the grammatical error or any missing information attach screenshots of all screens in a proper format in readme md file | 0 |
299 | 2,580,686,612 | IssuesEvent | 2015-02-13 19:20:33 | reapp/reapp-ui | https://api.github.com/repos/reapp/reapp-ui | opened | Optimize ComponentProps | performance | Notice `...this.componentProps()` used all over the place?
This is a helper function we made to add `ref`, `styles`, `className`, `id`, `ref` AND animation styles on all our nodes. It's awesome, but it's called *all* the time when animations are running.
And so, this is a great place to look. I think there is some big caching optimizations available here. I started looking at them, but backed out because it's not entirely simple. You'll probably introduce some state. Something like
```
componentWillMount() {
this._cachedProps = {};
}
```
And from there figure out how to cache them without breaking stuff. But this would definitely improve animation performance big time. | True | Optimize ComponentProps - Notice `...this.componentProps()` used all over the place?
This is a helper function we made to add `ref`, `styles`, `className`, `id`, `ref` AND animation styles on all our nodes. It's awesome, but it's called *all* the time when animations are running.
And so, this is a great place to look. I think there is some big caching optimizations available here. I started looking at them, but backed out because it's not entirely simple. You'll probably introduce some state. Something like
```
componentWillMount() {
this._cachedProps = {};
}
```
And from there figure out how to cache them without breaking stuff. But this would definitely improve animation performance big time. | non_priority | optimize componentprops notice this componentprops used all over the place this is a helper function we made to add ref styles classname id ref and animation styles on all our nodes it s awesome but it s called all the time when animations are running and so this is a great place to look i think there is some big caching optimizations available here i started looking at them but backed out because it s not entirely simple you ll probably introduce some state something like componentwillmount this cachedprops and from there figure out how to cache them without breaking stuff but this would definitely improve animation performance big time | 0 |
24,606 | 7,529,643,055 | IssuesEvent | 2018-04-14 07:40:35 | ElektraInitiative/libelektra | https://api.github.com/repos/ElektraInitiative/libelektra | opened | have maintainable list of build server admins | build | Feature request: it would be great to have an easy-to-maintain list of admins shared between all build jobs. Admins then can say "add to whitelist" so that collaborators are allowed to trigger jobs.
@ingwinlu wrote:
> I tried to add `allowMembersOfWhitelistedOrgsAsAdmin` to all build jobs today but seems like I can still not trigger a build all (see #1863) properly and only some jobs get executed
I do not have any idea why you cannot trigger jobs. Seems like misconfiguration in most of the jobs? Maybe the "add to whitelist" did not work properly in your case?
> @markus2330 https://github.com/janinko/ghprb/issues/416#issuecomment-266254688
As I understood it, building jobs only needs the "add to whitelist" by an admin. And this always worked, you are the first case where it did not. | 1.0 | have maintainable list of build server admins - Feature request: it would be great to have an easy-to-maintain list of admins shared between all build jobs. Admins then can say "add to whitelist" so that collaborators are allowed to trigger jobs.
@ingwinlu wrote:
> I tried to add `allowMembersOfWhitelistedOrgsAsAdmin` to all build jobs today but seems like I can still not trigger a build all (see #1863) properly and only some jobs get executed
I do not have any idea why you cannot trigger jobs. Seems like misconfiguration in most of the jobs? Maybe the "add to whitelist" did not work properly in your case?
> @markus2330 https://github.com/janinko/ghprb/issues/416#issuecomment-266254688
As I understood it, building jobs only needs the "add to whitelist" by an admin. And this always worked, you are the first case where it did not. | non_priority | have maintainable list of build server admins feature request it would be great to have an easy to maintain list of admins shared between all build jobs admins then can say add to whitelist so that collaborators are allowed to trigger jobs ingwinlu wrote i tried to add allowmembersofwhitelistedorgsasadmin to all build jobs today but seems like i can still not trigger a build all see properly and only some jobs get executed i do not have any idea why you cannot trigger jobs seems like misconfiguration in most of the jobs maybe the add to whitelist did not work properly in your case as i understood it building jobs only needs the add to whitelist by an admin and this always worked you are the first case where it did not | 0 |
30,998 | 14,737,291,950 | IssuesEvent | 2021-01-07 01:25:25 | mozilla-mobile/fenix | https://api.github.com/repos/mozilla-mobile/fenix | closed | [Bug] Fix strict mode violation on Nightly caused by org.mozilla.fenix.components.Analytics.getExperiments | eng:performance 🐞 bug | Fix the following strict mode violation:
https://gist.github.com/travis79/88779741ced2bddf03fe68797c3faa8f | True | [Bug] Fix strict mode violation on Nightly caused by org.mozilla.fenix.components.Analytics.getExperiments - Fix the following strict mode violation:
https://gist.github.com/travis79/88779741ced2bddf03fe68797c3faa8f | non_priority | fix strict mode violation on nightly caused by org mozilla fenix components analytics getexperiments fix the following strict mode violation | 0 |
182,919 | 21,678,053,617 | IssuesEvent | 2022-05-09 01:13:01 | metao1/springboot-redis-lettuce | https://api.github.com/repos/metao1/springboot-redis-lettuce | opened | CVE-2022-24823 (Medium) detected in netty-common-4.1.75.Final.jar | security vulnerability | ## CVE-2022-24823 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-common-4.1.75.Final.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/io.netty/netty-common/4.1.75.Final/84c32aa0e36b450d739c25576f5185e3ab9e7f66/netty-common-4.1.75.Final.jar</p>
<p>
Dependency Hierarchy:
- lettuce-core-6.1.8.RELEASE.jar (Root Library)
- :x: **netty-common-4.1.75.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/metao1/springboot-redis-lettuce/commit/e5a817563306daff5aae55f885899f97c1ab555f">e5a817563306daff5aae55f885899f97c1ab555f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user.
<p>Publish Date: 2022-05-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823>CVE-2022-24823</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p>
<p>Release Date: 2022-05-06</p>
<p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-24823 (Medium) detected in netty-common-4.1.75.Final.jar - ## CVE-2022-24823 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-common-4.1.75.Final.jar</b></p></summary>
<p></p>
<p>Library home page: <a href="https://netty.io/">https://netty.io/</a></p>
<p>Path to dependency file: /build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/io.netty/netty-common/4.1.75.Final/84c32aa0e36b450d739c25576f5185e3ab9e7f66/netty-common-4.1.75.Final.jar</p>
<p>
Dependency Hierarchy:
- lettuce-core-6.1.8.RELEASE.jar (Root Library)
- :x: **netty-common-4.1.75.Final.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/metao1/springboot-redis-lettuce/commit/e5a817563306daff5aae55f885899f97c1ab555f">e5a817563306daff5aae55f885899f97c1ab555f</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user.
<p>Publish Date: 2022-05-06
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823>CVE-2022-24823</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p>
<p>Release Date: 2022-05-06</p>
<p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in netty common final jar cve medium severity vulnerability vulnerable library netty common final jar library home page a href path to dependency file build gradle path to vulnerable library home wss scanner gradle caches modules files io netty netty common final netty common final jar dependency hierarchy lettuce core release jar root library x netty common final jar vulnerable library found in head commit a href found in base branch master vulnerability details netty is an open source asynchronous event driven network application framework the package io netty netty codec http prior to version final contains an insufficient fix for cve when netty s multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled this only impacts applications running on java version and lower additionally this vulnerability impacts code running on unix like systems and very old versions of mac osx and windows as they all share the system temporary directory between all users version final contains a patch for this vulnerability as a workaround specify one s own java io tmpdir when starting the jvm or use defaulthttpdatafactory setbasedir to set the directory to something that is only readable by the current user publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all io netty netty common final step up your open source security game with whitesource | 0 |
367,615 | 25,753,907,317 | IssuesEvent | 2022-12-08 15:05:41 | smw0807/pubg_your.stat | https://api.github.com/repos/smw0807/pubg_your.stat | closed | 수치 변환 | documentation | 몇몇 누적 데이터 필드들 변환 필요...
누적된 데이터들이다 보니 지나치게 높은 항목들이 있음
시간, 거리 관련 데이터들은 기본 값이 분, 미터로 제공해주기 때문에
몇만 분, 몇십만 미터 등등의 데이터로 표기됨
보기 편한 단위로 변환해주는 작업 필요
| 1.0 | 수치 변환 - 몇몇 누적 데이터 필드들 변환 필요...
누적된 데이터들이다 보니 지나치게 높은 항목들이 있음
시간, 거리 관련 데이터들은 기본 값이 분, 미터로 제공해주기 때문에
몇만 분, 몇십만 미터 등등의 데이터로 표기됨
보기 편한 단위로 변환해주는 작업 필요
| non_priority | 수치 변환 몇몇 누적 데이터 필드들 변환 필요 누적된 데이터들이다 보니 지나치게 높은 항목들이 있음 시간 거리 관련 데이터들은 기본 값이 분 미터로 제공해주기 때문에 몇만 분 몇십만 미터 등등의 데이터로 표기됨 보기 편한 단위로 변환해주는 작업 필요 | 0 |
68,053 | 17,138,361,984 | IssuesEvent | 2021-07-13 06:40:50 | google/mediapipe | https://api.github.com/repos/google/mediapipe | closed | build_bazel_rules_apple download failed | platform:desktop stalled stat:awaiting response type:build/install | run the below command in ubuntu18.04 x86_64
```GLOG_logtostderr=1 bazel run --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop/hello_world:hello_world```
but it throws out the problem:
ERROR: An error occurred during the fetch of repository 'build_bazel_rules_apple':
java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.19.0/rules_apple.0.19.0.tar.gz] to /root/.cache/bazel/_bazel_root/4884d566396e9b67b62185751879ad14/external/build_bazel_rules_apple/rules_apple.0.19.0.tar.gz: Read timed out
ERROR: no such package '@build_bazel_rules_apple//apple': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.19.0/rules_apple.0.19.0.tar.gz] to /root/.cache/bazel/_bazel_root/4884d566396e9b67b62185751879ad14/external/build_bazel_rules_apple/rules_apple.0.19.0.tar.gz: Read timed out
ERROR: no such package '@build_bazel_rules_apple//apple': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.19.0/rules_apple.0.19.0.tar.gz] to /root/.cache/bazel/_bazel_root/4884d566396e9b67b62185751879ad14/external/build_bazel_rules_apple/rules_apple.0.19.0.tar.gz: Read timed out
ERROR: no such package '@org_tensorflow//tensorflow': java.io.IOException: Error downloading [https://mirror.bazel.build/github.com/tensorflow/tensorflow/archive/7c09d15f9fcc14343343c247ebf5b8e0afe3e4aa.tar.gz, https://github.com/tensorflow/tensorflow/archive/7c09d15f9fcc14343343c247ebf5b8e0afe3e4aa.tar.gz] to /root/.cache/bazel/_bazel_root/4884d566396e9b67b62185751879ad14/external/org_tensorflow/7c09d15f9fcc14343343c247ebf5b8e0afe3e4aa.tar.gz: Tried to reconnect at offset 10,759,001 but server didn't support it
and i can't download tensorflow. is there any alternetive?
| 1.0 | build_bazel_rules_apple download failed - run the below command in ubuntu18.04 x86_64
```GLOG_logtostderr=1 bazel run --define MEDIAPIPE_DISABLE_GPU=1 mediapipe/examples/desktop/hello_world:hello_world```
but it throws out the problem:
ERROR: An error occurred during the fetch of repository 'build_bazel_rules_apple':
java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.19.0/rules_apple.0.19.0.tar.gz] to /root/.cache/bazel/_bazel_root/4884d566396e9b67b62185751879ad14/external/build_bazel_rules_apple/rules_apple.0.19.0.tar.gz: Read timed out
ERROR: no such package '@build_bazel_rules_apple//apple': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.19.0/rules_apple.0.19.0.tar.gz] to /root/.cache/bazel/_bazel_root/4884d566396e9b67b62185751879ad14/external/build_bazel_rules_apple/rules_apple.0.19.0.tar.gz: Read timed out
ERROR: no such package '@build_bazel_rules_apple//apple': java.io.IOException: Error downloading [https://github.com/bazelbuild/rules_apple/releases/download/0.19.0/rules_apple.0.19.0.tar.gz] to /root/.cache/bazel/_bazel_root/4884d566396e9b67b62185751879ad14/external/build_bazel_rules_apple/rules_apple.0.19.0.tar.gz: Read timed out
ERROR: no such package '@org_tensorflow//tensorflow': java.io.IOException: Error downloading [https://mirror.bazel.build/github.com/tensorflow/tensorflow/archive/7c09d15f9fcc14343343c247ebf5b8e0afe3e4aa.tar.gz, https://github.com/tensorflow/tensorflow/archive/7c09d15f9fcc14343343c247ebf5b8e0afe3e4aa.tar.gz] to /root/.cache/bazel/_bazel_root/4884d566396e9b67b62185751879ad14/external/org_tensorflow/7c09d15f9fcc14343343c247ebf5b8e0afe3e4aa.tar.gz: Tried to reconnect at offset 10,759,001 but server didn't support it
and i can't download tensorflow. is there any alternetive?
| non_priority | build bazel rules apple download failed run the below command in glog logtostderr bazel run define mediapipe disable gpu mediapipe examples desktop hello world hello world but it throws out the problem error an error occurred during the fetch of repository build bazel rules apple java io ioexception error downloading to root cache bazel bazel root external build bazel rules apple rules apple tar gz read timed out error no such package build bazel rules apple apple java io ioexception error downloading to root cache bazel bazel root external build bazel rules apple rules apple tar gz read timed out error no such package build bazel rules apple apple java io ioexception error downloading to root cache bazel bazel root external build bazel rules apple rules apple tar gz read timed out error no such package org tensorflow tensorflow java io ioexception error downloading to root cache bazel bazel root external org tensorflow tar gz tried to reconnect at offset but server didn t support it and i can t download tensorflow is there any alternetive | 0 |
256,612 | 22,067,168,080 | IssuesEvent | 2022-05-31 05:27:38 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: sqlsmith/setup=seed/setting=no-ddl failed | C-test-failure O-robot O-roachtest branch-master release-blocker T-sql-queries | roachtest.sqlsmith/setup=seed/setting=no-ddl [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5299852?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5299852?buildTab=artifacts#/sqlsmith/setup=seed/setting=no-ddl) on master @ [1e2cc61b58dc14386bb68dca59814874648931c2](https://github.com/cockroachdb/cockroach/commits/1e2cc61b58dc14386bb68dca59814874648931c2):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /artifacts/sqlsmith/setup=seed/setting=no-ddl/run_1
sqlsmith.go:265,sqlsmith.go:305,test_runner.go:884: error: pq: internal error: expected *DString, found tree.dNull
stmt:
SELECT
'67 years 6 mons 834 days 19:14:45.200516':::INTERVAL AS col_50,
10043:::INT8 AS col_51,
(-1.9870857726031228):::FLOAT8 AS col_52,
COALESCE(tab_23.crdb_internal_mvcc_timestamp, (-90413478223369345.65):::DECIMAL) AS col_53,
'{"*9DC}])9": [[1.6711536670477731, true], {}, {"c": [{}]}], "D5^LuGd1B": [null], "H~{{XZL8G)sw": [], "a": [], "foobar": [[]]}':::JSONB
AS col_54,
(now():::DATE::DATE::DATE - crdb_internal.num_inverted_index_entries(tab_23._string::STRING, tab_23._int2::INT8)::INT8::INT8)::DATE
AS col_55,
tab_23._float8 AS col_56,
tab_23._inet AS col_57,
e'\x13b]5``\x0bP+':::STRING AS col_58
FROM
defaultdb.public.seed@[0] AS tab_23
WHERE
(NOT tab_23._bool)
ORDER BY
tab_23._int2 ASC, tab_23._int2 DESC
LIMIT
78:::INT8;
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=seed/setting=no-ddl.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-16165 | 2.0 | roachtest: sqlsmith/setup=seed/setting=no-ddl failed - roachtest.sqlsmith/setup=seed/setting=no-ddl [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5299852?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5299852?buildTab=artifacts#/sqlsmith/setup=seed/setting=no-ddl) on master @ [1e2cc61b58dc14386bb68dca59814874648931c2](https://github.com/cockroachdb/cockroach/commits/1e2cc61b58dc14386bb68dca59814874648931c2):
```
The test failed on branch=master, cloud=gce:
test artifacts and logs in: /artifacts/sqlsmith/setup=seed/setting=no-ddl/run_1
sqlsmith.go:265,sqlsmith.go:305,test_runner.go:884: error: pq: internal error: expected *DString, found tree.dNull
stmt:
SELECT
'67 years 6 mons 834 days 19:14:45.200516':::INTERVAL AS col_50,
10043:::INT8 AS col_51,
(-1.9870857726031228):::FLOAT8 AS col_52,
COALESCE(tab_23.crdb_internal_mvcc_timestamp, (-90413478223369345.65):::DECIMAL) AS col_53,
'{"*9DC}])9": [[1.6711536670477731, true], {}, {"c": [{}]}], "D5^LuGd1B": [null], "H~{{XZL8G)sw": [], "a": [], "foobar": [[]]}':::JSONB
AS col_54,
(now():::DATE::DATE::DATE - crdb_internal.num_inverted_index_entries(tab_23._string::STRING, tab_23._int2::INT8)::INT8::INT8)::DATE
AS col_55,
tab_23._float8 AS col_56,
tab_23._inet AS col_57,
e'\x13b]5``\x0bP+':::STRING AS col_58
FROM
defaultdb.public.seed@[0] AS tab_23
WHERE
(NOT tab_23._bool)
ORDER BY
tab_23._int2 ASC, tab_23._int2 DESC
LIMIT
78:::INT8;
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*sqlsmith/setup=seed/setting=no-ddl.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
Jira issue: CRDB-16165 | non_priority | roachtest sqlsmith setup seed setting no ddl failed roachtest sqlsmith setup seed setting no ddl with on master the test failed on branch master cloud gce test artifacts and logs in artifacts sqlsmith setup seed setting no ddl run sqlsmith go sqlsmith go test runner go error pq internal error expected dstring found tree dnull stmt select years mons days interval as col as col as col coalesce tab crdb internal mvcc timestamp decimal as col c h sw a foobar jsonb as col now date date date crdb internal num inverted index entries tab string string tab date as col tab as col tab inet as col e string as col from defaultdb public seed as tab where not tab bool order by tab asc tab desc limit help see see cc cockroachdb sql queries jira issue crdb | 0 |
36,110 | 5,031,580,541 | IssuesEvent | 2016-12-16 07:49:09 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | github.com/cockroachdb/cockroach/pkg/sql/distsql: TestHashRouter failed under stress | Robot test-failure | SHA: https://github.com/cockroachdb/cockroach/commits/8d9859c68d863a7d1381bb1b56d7d4bd6dadf031
Parameters:
```
COCKROACH_PROPOSER_EVALUATED_KV=true
TAGS=
GOFLAGS=
```
Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=92381&tab=buildLog
```
routers_test.go:103: rows [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] and [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] in different buckets
routers_test.go:103: rows [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] and ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] in different buckets
routers_test.go:103: rows [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] and [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] in different buckets
routers_test.go:103: rows [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] and [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] in different buckets
routers_test.go:103: rows [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] and ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] in different buckets
routers_test.go:103: rows [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] and [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] in different buckets
routers_test.go:103: rows [NULL 1.3014694194430971 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 1086849928.669756868 e'\x13rZ\x11m'] and [e'\x1a +0+\\m\\' 0.0595383958595917 '' b'\bQ\x13\x1d\xfd6\x04H' 85488810207331899180000000000000 e'\x13rZ\x11m'] in different buckets
routers_test.go:103: rows [NULL 0.14466573028119212 e'\u059F' b'?C' 12678999017128137610 e'\x13rZ\x11m'] and [e'\x1a +0+\\m\\' 0.0595383958595917 '' b'\bQ\x13\x1d\xfd6\x04H' 85488810207331899180000000000000 e'\x13rZ\x11m'] in different buckets
routers_test.go:103: rows [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] and [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] in different buckets
routers_test.go:103: rows [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] and ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] in different buckets
routers_test.go:103: rows [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] and [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] in different buckets
routers_test.go:103: rows [e'\x1a +0+\\m\\' 0.0595383958595917 '' b'\bQ\x13\x1d\xfd6\x04H' 85488810207331899180000000000000 e'\x13rZ\x11m'] and [NULL 1.3014694194430971 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 1086849928.669756868 e'\x13rZ\x11m'] in different buckets
routers_test.go:103: rows [e'\x1a +0+\\m\\' 0.0595383958595917 '' b'\bQ\x13\x1d\xfd6\x04H' 85488810207331899180000000000000 e'\x13rZ\x11m'] and [NULL 0.14466573028119212 e'\u059F' b'?C' 12678999017128137610 e'\x13rZ\x11m'] in different buckets
routers_test.go:103: rows [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] and [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] in different buckets
routers_test.go:103: rows [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] and [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] in different buckets
routers_test.go:103: rows [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] and [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] in different buckets
routers_test.go:103: rows ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] and [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] in different buckets
routers_test.go:103: rows ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] and [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] in different buckets
routers_test.go:103: rows ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] and [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] in different buckets
routers_test.go:103: rows [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] and [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] in different buckets
routers_test.go:103: rows [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] and [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] in different buckets
routers_test.go:103: rows [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] and [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] in different buckets
``` | 1.0 | github.com/cockroachdb/cockroach/pkg/sql/distsql: TestHashRouter failed under stress - SHA: https://github.com/cockroachdb/cockroach/commits/8d9859c68d863a7d1381bb1b56d7d4bd6dadf031
Parameters:
```
COCKROACH_PROPOSER_EVALUATED_KV=true
TAGS=
GOFLAGS=
```
Stress build found a failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=92381&tab=buildLog
```
routers_test.go:103: rows [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] and [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] in different buckets
routers_test.go:103: rows [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] and ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] in different buckets
routers_test.go:103: rows [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] and [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] in different buckets
routers_test.go:103: rows [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] and [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] in different buckets
routers_test.go:103: rows [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] and ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] in different buckets
routers_test.go:103: rows [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] and [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] in different buckets
routers_test.go:103: rows [NULL 1.3014694194430971 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 1086849928.669756868 e'\x13rZ\x11m'] and [e'\x1a +0+\\m\\' 0.0595383958595917 '' b'\bQ\x13\x1d\xfd6\x04H' 85488810207331899180000000000000 e'\x13rZ\x11m'] in different buckets
routers_test.go:103: rows [NULL 0.14466573028119212 e'\u059F' b'?C' 12678999017128137610 e'\x13rZ\x11m'] and [e'\x1a +0+\\m\\' 0.0595383958595917 '' b'\bQ\x13\x1d\xfd6\x04H' 85488810207331899180000000000000 e'\x13rZ\x11m'] in different buckets
routers_test.go:103: rows [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] and [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] in different buckets
routers_test.go:103: rows [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] and ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] in different buckets
routers_test.go:103: rows [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] and [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] in different buckets
routers_test.go:103: rows [e'\x1a +0+\\m\\' 0.0595383958595917 '' b'\bQ\x13\x1d\xfd6\x04H' 85488810207331899180000000000000 e'\x13rZ\x11m'] and [NULL 1.3014694194430971 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 1086849928.669756868 e'\x13rZ\x11m'] in different buckets
routers_test.go:103: rows [e'\x1a +0+\\m\\' 0.0595383958595917 '' b'\bQ\x13\x1d\xfd6\x04H' 85488810207331899180000000000000 e'\x13rZ\x11m'] and [NULL 0.14466573028119212 e'\u059F' b'?C' 12678999017128137610 e'\x13rZ\x11m'] in different buckets
routers_test.go:103: rows [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] and [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] in different buckets
routers_test.go:103: rows [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] and [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] in different buckets
routers_test.go:103: rows [e'\x0b:snpgAR0' 1.0169933991924955 e'\u059F' b'\xa6\xfe`\xddgx\xe3\b' 50098762.46547676937 ''] and [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] in different buckets
routers_test.go:103: rows ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] and [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] in different buckets
routers_test.go:103: rows ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] and [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] in different buckets
routers_test.go:103: rows ['>gA' -0.9195696262880189 e'\u059F' b'\xc2' 46577444482967552030000000000000 ''] and [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] in different buckets
routers_test.go:103: rows [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] and [e'vc7H\fS(2' -1.7275869978035117 '' b'c /r[\xde' 69.00683718486054427 ''] in different buckets
routers_test.go:103: rows [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] and [e'\nu' -0.9849645988690314 '' b'\bQ\x13\x1d\xfd6\x04H' 7457758506596949.421 ''] in different buckets
routers_test.go:103: rows [NULL -0.304072764526957 e'\u059F' b'?C' 42576448126598073750000000000 ''] and [e'!\r]HF\x1e\x1c\n' 2.268375350793319 '' b'>K' 1086849928.669756868 ''] in different buckets
``` | non_priority | github com cockroachdb cockroach pkg sql distsql testhashrouter failed under stress sha parameters cockroach proposer evaluated kv true tags goflags stress build found a failed test routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows hf n b k and in different buckets routers test go rows hf n b k and in different buckets routers test go rows hf n b k and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and hf n b k in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and hf n b k in different buckets routers test go rows and in different buckets routers test go rows and in different buckets routers test go rows and hf n b k in different buckets | 0 |
44,433 | 5,825,028,730 | IssuesEvent | 2017-05-07 18:00:42 | Shopify/shipit-engine | https://api.github.com/repos/Shopify/shipit-engine | closed | new stack UI needs a little love | bug Design | The repo field on the new stack page has line breaks that are not pretty:
<img alt="shipit 2015-07-03 11-10-42" src="https://cloud.githubusercontent.com/assets/118850/8501521/4a4fefb8-2174-11e5-8448-f79e28d0a99d.png">
| 1.0 | new stack UI needs a little love - The repo field on the new stack page has line breaks that are not pretty:
<img alt="shipit 2015-07-03 11-10-42" src="https://cloud.githubusercontent.com/assets/118850/8501521/4a4fefb8-2174-11e5-8448-f79e28d0a99d.png">
| non_priority | new stack ui needs a little love the repo field on the new stack page has line breaks that are not pretty img alt shipit src | 0 |
276,419 | 30,455,737,979 | IssuesEvent | 2023-07-16 21:33:58 | OpenLiberty/open-liberty | https://api.github.com/repos/OpenLiberty/open-liberty | closed | Upgrade jose4j library version used by acme | team:Core Security target:23006 | Update acme to use the latest version of the jose4j library (0.9.3).
For: https://github.ibm.com/websphere/wlp-scan/issues/281
PR: https://github.com/OpenLiberty/open-liberty/pull/25208 | True | Upgrade jose4j library version used by acme - Update acme to use the latest version of the jose4j library (0.9.3).
For: https://github.ibm.com/websphere/wlp-scan/issues/281
PR: https://github.com/OpenLiberty/open-liberty/pull/25208 | non_priority | upgrade library version used by acme update acme to use the latest version of the library for pr | 0 |
19,412 | 26,937,068,030 | IssuesEvent | 2023-02-07 21:40:02 | xCollateral/VulkanMod | https://api.github.com/repos/xCollateral/VulkanMod | closed | [1.19.3] Journeymap mixin issue. | compatibility | Looks like you are fully rewriting this method.
https://github.com/xCollateral/VulkanMod/blob/edf26c7888800b2d795767d12cb383409361e10b/src/main/java/net/vulkanmod/mixin/debug/DebugHudM.java#L80
You do not really need to rewrite the whole method.
I had a user report a crash when JM and Vulcan are together, details and crash report are here.
https://github.com/TeamJM/journeymap/issues/547
This is my mixin inject point in that method.
```
@Inject(method = "drawGameInformation(Lcom/mojang/blaze3d/vertex/PoseStack;)V",
at = @At(value = "INVOKE", target = "Ljava/util/List;add(Ljava/lang/Object;)Z",
shift = At.Shift.AFTER,
ordinal = 2),
locals = LocalCapture.CAPTURE_FAILSOFT)
``` | True | [1.19.3] Journeymap mixin issue. - Looks like you are fully rewriting this method.
https://github.com/xCollateral/VulkanMod/blob/edf26c7888800b2d795767d12cb383409361e10b/src/main/java/net/vulkanmod/mixin/debug/DebugHudM.java#L80
You do not really need to rewrite the whole method.
I had a user report a crash when JM and Vulcan are together, details and crash report are here.
https://github.com/TeamJM/journeymap/issues/547
This is my mixin inject point in that method.
```
@Inject(method = "drawGameInformation(Lcom/mojang/blaze3d/vertex/PoseStack;)V",
at = @At(value = "INVOKE", target = "Ljava/util/List;add(Ljava/lang/Object;)Z",
shift = At.Shift.AFTER,
ordinal = 2),
locals = LocalCapture.CAPTURE_FAILSOFT)
``` | non_priority | journeymap mixin issue looks like you are fully rewriting this method you do not really need to rewrite the whole method i had a user report a crash when jm and vulcan are together details and crash report are here this is my mixin inject point in that method inject method drawgameinformation lcom mojang vertex posestack v at at value invoke target ljava util list add ljava lang object z shift at shift after ordinal locals localcapture capture failsoft | 0 |
308,024 | 26,574,622,456 | IssuesEvent | 2023-01-21 16:49:43 | junit-team/junit5 | https://api.github.com/repos/junit-team/junit5 | opened | Parameters are not validated in local `@MethodSource` factory method | type: bug component: Jupiter theme: parameterized tests theme: diagnostics | ## Overview
The current implementation of support for local factory method names that accept arguments in `@MethodSource` does not validate the specified parameters. Rather, the lookup is performed based solely on the names of the parameters. Furthermore, the lookup ignores the parameters if there is only one factory method with the specified name.
This can lead to incorrect configuration being silently ignored or confusing error messages when the specified factory method is overloaded.
## Examples
Given the following extensions:
```java
class IntegerResolver implements ParameterResolver {
@Override
public boolean supportsParameter(ParameterContext pc, ExtensionContext ec) {
return pc.getParameter().getType() == int.class;
}
@Override
public Object resolveParameter(ParameterContext pc, ExtensionContext ec) {
return 2;
}
}
class IntegerArrayResolver implements ParameterResolver {
@Override
public boolean supportsParameter(ParameterContext pc, ExtensionContext ec) {
return pc.getParameter().getType() == int[].class;
}
@Override
public Object resolveParameter(ParameterContext pc, ExtensionContext ec) {
return new int[] { 2, 3 };
}
}
```
And given the following test class:
```java
@ExtendWith({IntegerResolver.class, IntegerArrayResolver.class})
class MethodSourceTests {
@ParameterizedTest
// @MethodSource("example.MethodSourceTests#factory(dog)")
@MethodSource("factory(dog)")
void test(String argument) {
assertTrue(argument.startsWith("2"));
}
static Stream<Arguments> factory(int quantities) {
return Stream.of(arguments(quantities + " apples"), arguments(quantities + " lemons"));
}
// static Stream<Arguments> factory(int[] quantities) {
// return Stream.of(arguments(quantities[0] + " apples"), arguments(quantities[0] + " lemons"));
// }
}
```
If we run the test "as is", it will pass because the `dog` parameter is not validated. It is in fact completely ignored since there are no "competing" overloaded `factory` methods with the same name.
Whereas, if we attempt to do the same using the fully qualified method name (FQMN) syntax (`@MethodSource("example.MethodSourceTests#factory(dog)")`), we get the following exception.
> org.junit.platform.commons.JUnitException: Failed to load parameter type [dog] for method [factory] in class [example.MethodSourceTests].
Similarly, if we supply a valid parameter type for a factory method that does not exist using the FQMN syntax (`@MethodSource("example.MethodSourceTests#factory(java.lang.Integer)")`), we get the following exception.
> org.junit.platform.commons.JUnitException: Could not find factory method [factory(java.lang.Integer)] in class [example.MethodSourceTests]
In contrast, using the local factory method syntax (`@MethodSource("factory(java.lang.Integer)")`), the test will pass since the parameters are ignored (as with the `dog` example).
## Deliverables
- [ ] Validate factory method parameter types supplied to `@MethodSource` using the local factory method name syntax.
- [ ] Ensure that error messages for configuration errors encountered when using the local factory method name syntax align with the error messages produced for similar configuration errors when using the fully qualified method name syntax.
| 1.0 | Parameters are not validated in local `@MethodSource` factory method - ## Overview
The current implementation of support for local factory method names that accept arguments in `@MethodSource` does not validate the specified parameters. Rather, the lookup is performed based solely on the names of the parameters. Furthermore, the lookup ignores the parameters if there is only one factory method with the specified name.
This can lead to incorrect configuration being silently ignored or confusing error messages when the specified factory method is overloaded.
## Examples
Given the following extensions:
```java
class IntegerResolver implements ParameterResolver {
@Override
public boolean supportsParameter(ParameterContext pc, ExtensionContext ec) {
return pc.getParameter().getType() == int.class;
}
@Override
public Object resolveParameter(ParameterContext pc, ExtensionContext ec) {
return 2;
}
}
class IntegerArrayResolver implements ParameterResolver {
@Override
public boolean supportsParameter(ParameterContext pc, ExtensionContext ec) {
return pc.getParameter().getType() == int[].class;
}
@Override
public Object resolveParameter(ParameterContext pc, ExtensionContext ec) {
return new int[] { 2, 3 };
}
}
```
And given the following test class:
```java
@ExtendWith({IntegerResolver.class, IntegerArrayResolver.class})
class MethodSourceTests {
@ParameterizedTest
// @MethodSource("example.MethodSourceTests#factory(dog)")
@MethodSource("factory(dog)")
void test(String argument) {
assertTrue(argument.startsWith("2"));
}
static Stream<Arguments> factory(int quantities) {
return Stream.of(arguments(quantities + " apples"), arguments(quantities + " lemons"));
}
// static Stream<Arguments> factory(int[] quantities) {
// return Stream.of(arguments(quantities[0] + " apples"), arguments(quantities[0] + " lemons"));
// }
}
```
If we run the test "as is", it will pass because the `dog` parameter is not validated. It is in fact completely ignored since there are no "competing" overloaded `factory` methods with the same name.
Whereas, if we attempt to do the same using the fully qualified method name (FQMN) syntax (`@MethodSource("example.MethodSourceTests#factory(dog)")`), we get the following exception.
> org.junit.platform.commons.JUnitException: Failed to load parameter type [dog] for method [factory] in class [example.MethodSourceTests].
Similarly, if we supply a valid parameter type for a factory method that does not exist using the FQMN syntax (`@MethodSource("example.MethodSourceTests#factory(java.lang.Integer)")`), we get the following exception.
> org.junit.platform.commons.JUnitException: Could not find factory method [factory(java.lang.Integer)] in class [example.MethodSourceTests]
In contrast, using the local factory method syntax (`@MethodSource("factory(java.lang.Integer)")`), the test will pass since the parameters are ignored (as with the `dog` example).
## Deliverables
- [ ] Validate factory method parameter types supplied to `@MethodSource` using the local factory method name syntax.
- [ ] Ensure that error messages for configuration errors encountered when using the local factory method name syntax align with the error messages produced for similar configuration errors when using the fully qualified method name syntax.
| non_priority | parameters are not validated in local methodsource factory method overview the current implementation of support for local factory method names that accept arguments in methodsource does not validate the specified parameters rather the lookup is performed based solely on the names of the parameters furthermore the lookup ignores the parameters if there is only one factory method with the specified name this can lead to incorrect configuration being silently ignored or confusing error messages when the specified factory method is overloaded examples given the following extensions java class integerresolver implements parameterresolver override public boolean supportsparameter parametercontext pc extensioncontext ec return pc getparameter gettype int class override public object resolveparameter parametercontext pc extensioncontext ec return class integerarrayresolver implements parameterresolver override public boolean supportsparameter parametercontext pc extensioncontext ec return pc getparameter gettype int class override public object resolveparameter parametercontext pc extensioncontext ec return new int and given the following test class java extendwith integerresolver class integerarrayresolver class class methodsourcetests parameterizedtest methodsource example methodsourcetests factory dog methodsource factory dog void test string argument asserttrue argument startswith static stream factory int quantities return stream of arguments quantities apples arguments quantities lemons static stream factory int quantities return stream of arguments quantities apples arguments quantities lemons if we run the test as is it will pass because the dog parameter is not validated it is in fact completely ignored since there are no competing overloaded factory methods with the same name whereas if we attempt to do the same using the fully qualified method name fqmn syntax methodsource example methodsourcetests factory dog we get the following exception org junit platform commons junitexception failed to load parameter type for method in class similarly if we supply a valid parameter type for a factory method that does not exist using the fqmn syntax methodsource example methodsourcetests factory java lang integer we get the following exception org junit platform commons junitexception could not find factory method in class in contrast using the local factory method syntax methodsource factory java lang integer the test will pass since the parameters are ignored as with the dog example deliverables validate factory method parameter types supplied to methodsource using the local factory method name syntax ensure that error messages for configuration errors encountered when using the local factory method name syntax align with the error messages produced for similar configuration errors when using the fully qualified method name syntax | 0 |
418,563 | 28,119,722,715 | IssuesEvent | 2023-03-31 13:25:44 | NHMDenmark/Herbarium-Sheets-workstation | https://api.github.com/repos/NHMDenmark/Herbarium-Sheets-workstation | closed | Make additions to training document to account for exceptional cases | documentation workstation NHMD feedback needed workflow | We need protocols for all of these:
- [ ] Multi specimen sheets in layers
- [ ] Multi sheet specimens
- [ ] Specimens on irregular sized paper
- [ ] Mixed specimens
- [ ] Damaged specimens
- [ ] Loose specimens
- [ ] Specimens with no text
- [ ] Specimens with text on back
- [ ] Specimens with no room for the barcode label
- [ ] Specimens that already has a barcode
- [ ] Obvious erroneous specimens
- [ ] Plus of course all the tech issues we will run into
- [ ] When to use notes field
| 1.0 | Make additions to training document to account for exceptional cases - We need protocols for all of these:
- [ ] Multi specimen sheets in layers
- [ ] Multi sheet specimens
- [ ] Specimens on irregular sized paper
- [ ] Mixed specimens
- [ ] Damaged specimens
- [ ] Loose specimens
- [ ] Specimens with no text
- [ ] Specimens with text on back
- [ ] Specimens with no room for the barcode label
- [ ] Specimens that already has a barcode
- [ ] Obvious erroneous specimens
- [ ] Plus of course all the tech issues we will run into
- [ ] When to use notes field
| non_priority | make additions to training document to account for exceptional cases we need protocols for all of these multi specimen sheets in layers multi sheet specimens specimens on irregular sized paper mixed specimens damaged specimens loose specimens specimens with no text specimens with text on back specimens with no room for the barcode label specimens that already has a barcode obvious erroneous specimens plus of course all the tech issues we will run into when to use notes field | 0 |
47,981 | 10,192,956,339 | IssuesEvent | 2019-08-12 12:34:25 | swagger-api/swagger-editor | https://api.github.com/repos/swagger-api/swagger-editor | closed | Feature request - Namespace UI on swagger editor | cat: code generation type: feature | I like the new version swagger editor, but there is one feature which would top it all off in my opinion.
I think it would be really cool if you could give the namespace name, or a form to fill out before the client/server libraries download.
Let me know what you guys think 👍 | 1.0 | Feature request - Namespace UI on swagger editor - I like the new version swagger editor, but there is one feature which would top it all off in my opinion.
I think it would be really cool if you could give the namespace name, or a form to fill out before the client/server libraries download.
Let me know what you guys think 👍 | non_priority | feature request namespace ui on swagger editor i like the new version swagger editor but there is one feature which would top it all off in my opinion i think it would be really cool if you could give the namespace name or a form to fill out before the client server libraries download let me know what you guys think 👍 | 0 |
210,160 | 23,739,022,914 | IssuesEvent | 2022-08-31 10:41:25 | SacleuxBenoit/test-nodejs | https://api.github.com/repos/SacleuxBenoit/test-nodejs | closed | CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js - autoclosed | security vulnerability | ## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: /front/node_modules/sockjs/examples/echo/index.html</p>
<p>Path to vulnerable library: /front/node_modules/sockjs/examples/echo/index.html,/front/node_modules/sockjs/examples/express-3.x/index.html,/front/node_modules/sockjs/examples/hapi/html/index.html,/front/node_modules/sockjs/examples/express/index.html,/front/node_modules/sockjs/examples/multiplex/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SacleuxBenoit/test-nodejs/commit/33514995c8ae2f63bbd14d446b77b1f82e188ba3">33514995c8ae2f63bbd14d446b77b1f82e188ba3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js - autoclosed - ## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: /front/node_modules/sockjs/examples/echo/index.html</p>
<p>Path to vulnerable library: /front/node_modules/sockjs/examples/echo/index.html,/front/node_modules/sockjs/examples/express-3.x/index.html,/front/node_modules/sockjs/examples/hapi/html/index.html,/front/node_modules/sockjs/examples/express/index.html,/front/node_modules/sockjs/examples/multiplex/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/SacleuxBenoit/test-nodejs/commit/33514995c8ae2f63bbd14d446b77b1f82e188ba3">33514995c8ae2f63bbd14d446b77b1f82e188ba3</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in jquery min js autoclosed cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file front node modules sockjs examples echo index html path to vulnerable library front node modules sockjs examples echo index html front node modules sockjs examples express x index html front node modules sockjs examples hapi html index html front node modules sockjs examples express index html front node modules sockjs examples multiplex index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
423,678 | 28,808,190,446 | IssuesEvent | 2023-05-03 00:48:47 | khcho0125/DSM-Delivery-Server | https://api.github.com/repos/khcho0125/DSM-Delivery-Server | closed | Change UUID Legary to Integer Value | documentation | ## AS-IS
- UUID's unconvenience in the application, so I decide to remove this legary
## TO-BE
- UUID → Int = More better readability and performance | 1.0 | Change UUID Legary to Integer Value - ## AS-IS
- UUID's unconvenience in the application, so I decide to remove this legary
## TO-BE
- UUID → Int = More better readability and performance | non_priority | change uuid legary to integer value as is uuid s unconvenience in the application so i decide to remove this legary to be uuid → int more better readability and performance | 0 |
132,833 | 10,765,993,705 | IssuesEvent | 2019-11-01 12:39:24 | flutter/flutter | https://api.github.com/repos/flutter/flutter | closed | Standardize testing environments of flutter_tools dependencies | a: tests team tool | Instead of using our own machines and deleting stuff to reproduce states of missing dependencies to test things like #9580, create an official process to test on unified initial environments.
Something lightweight like Docker is ideal. Then we can have different images for each development environment combinations. Definitely seems possible for Android SDK https://hub.docker.com/r/runmymind/docker-android-sdk/~/dockerfile/. There's even a flutter image https://hub.docker.com/r/brianegan/flutter/. Not sure how easy it is to get xcode running in docker though.
Otherwise we can create images of OS X in virtualbox | 1.0 | Standardize testing environments of flutter_tools dependencies - Instead of using our own machines and deleting stuff to reproduce states of missing dependencies to test things like #9580, create an official process to test on unified initial environments.
Something lightweight like Docker is ideal. Then we can have different images for each development environment combinations. Definitely seems possible for Android SDK https://hub.docker.com/r/runmymind/docker-android-sdk/~/dockerfile/. There's even a flutter image https://hub.docker.com/r/brianegan/flutter/. Not sure how easy it is to get xcode running in docker though.
Otherwise we can create images of OS X in virtualbox | non_priority | standardize testing environments of flutter tools dependencies instead of using our own machines and deleting stuff to reproduce states of missing dependencies to test things like create an official process to test on unified initial environments something lightweight like docker is ideal then we can have different images for each development environment combinations definitely seems possible for android sdk there s even a flutter image not sure how easy it is to get xcode running in docker though otherwise we can create images of os x in virtualbox | 0 |
159,677 | 13,769,525,716 | IssuesEvent | 2020-10-07 18:45:25 | torch2424/as-bind | https://api.github.com/repos/torch2424/as-bind | opened | How to add more types / plans going forward | documentation good first issue | Was chatting in the AS discord about AS bind, and figured this would be good to document in an issue as well :smile:
---
1. Identify the data id in: https://github.com/torch2424/as-bind/blob/master/lib/assembly/as-bind.ts#L15
2. Pull the data out of linear memory by it's layout: https://www.assemblyscript.org/interoperability.html#class-layout or https://www.assemblyscript.org/memory.html#internals
I do think that library would be helpful for general Data management in a JS app, and could allow for a "Everything is a data view", making things similar on the implementation side. But I wouldn't want to force users into using a specific library, or inflate the code size (It's not super big at 11kb, but currently we sit around 4kb + tree-shakable): https://bundlephobia.com/result?p=structurae@3.3.0
What we probably want to do, per the steps I mentioned above, is expose the Standard Library automatically in the as-bind entry file.
And for custom objects, is offer a way in both JS and AS land to pass around custom objects using like a function call.
That also being said, I know per the working meeting, there's plans to build an AS bindgen soon-ish. Which would then be the preffered way to handle this. :slight_smile: But I don't know how soon is soon haha! | 1.0 | How to add more types / plans going forward - Was chatting in the AS discord about AS bind, and figured this would be good to document in an issue as well :smile:
---
1. Identify the data id in: https://github.com/torch2424/as-bind/blob/master/lib/assembly/as-bind.ts#L15
2. Pull the data out of linear memory by it's layout: https://www.assemblyscript.org/interoperability.html#class-layout or https://www.assemblyscript.org/memory.html#internals
I do think that library would be helpful for general Data management in a JS app, and could allow for a "Everything is a data view", making things similar on the implementation side. But I wouldn't want to force users into using a specific library, or inflate the code size (It's not super big at 11kb, but currently we sit around 4kb + tree-shakable): https://bundlephobia.com/result?p=structurae@3.3.0
What we probably want to do, per the steps I mentioned above, is expose the Standard Library automatically in the as-bind entry file.
And for custom objects, is offer a way in both JS and AS land to pass around custom objects using like a function call.
That also being said, I know per the working meeting, there's plans to build an AS bindgen soon-ish. Which would then be the preffered way to handle this. :slight_smile: But I don't know how soon is soon haha! | non_priority | how to add more types plans going forward was chatting in the as discord about as bind and figured this would be good to document in an issue as well smile identify the data id in pull the data out of linear memory by it s layout or i do think that library would be helpful for general data management in a js app and could allow for a everything is a data view making things similar on the implementation side but i wouldn t want to force users into using a specific library or inflate the code size it s not super big at but currently we sit around tree shakable what we probably want to do per the steps i mentioned above is expose the standard library automatically in the as bind entry file and for custom objects is offer a way in both js and as land to pass around custom objects using like a function call that also being said i know per the working meeting there s plans to build an as bindgen soon ish which would then be the preffered way to handle this slight smile but i don t know how soon is soon haha | 0 |
259,655 | 22,504,503,467 | IssuesEvent | 2022-06-23 14:27:58 | MPMG-DCC-UFMG/F01 | https://api.github.com/repos/MPMG-DCC-UFMG/F01 | opened | Teste de generalizacao para a tag Terceiro Setor - Dados de Parcerias - Piau | generalization test development | DoD: Realizar o teste de Generalização do validador da tag Terceiro Setor - Dados de Parcerias para o Município de Piau. | 1.0 | Teste de generalizacao para a tag Terceiro Setor - Dados de Parcerias - Piau - DoD: Realizar o teste de Generalização do validador da tag Terceiro Setor - Dados de Parcerias para o Município de Piau. | non_priority | teste de generalizacao para a tag terceiro setor dados de parcerias piau dod realizar o teste de generalização do validador da tag terceiro setor dados de parcerias para o município de piau | 0 |
78,990 | 22,581,815,347 | IssuesEvent | 2022-06-28 12:19:54 | elastic/beats | https://api.github.com/repos/elastic/beats | closed | Build 46 for 8.3 with status FAILURE | automation ci-reported Team:Elastic-Agent-Data-Plane build-failures |
## :broken_heart: Tests Failed
<!-- BUILD BADGES-->
> _the below badges are clickable and redirect to their specific view in the CI or DOCS_
[](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//pipeline) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//tests) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//changes) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//artifacts) [](http://beats_null.docs-preview.app.elstc.co/diff) [](https://ci-stats.elastic.co/app/apm/services/beats-ci/transactions/view?rangeFrom=2022-06-28T07:24:39.656Z&rangeTo=2022-06-28T07:44:39.656Z&transactionName=Beats/beats/8.3&transactionType=job&latencyAggregationType=avg&traceId=974160cb047e3df95a0df6fba5995df1&transactionId=56f08c13620a77e0)
<!-- BUILD SUMMARY-->
<details><summary>Expand to view the summary</summary>
<p>
#### Build stats
* Start Time: 2022-06-28T07:34:39.656+0000
* Duration: 105 min 55 sec
#### Test stats :test_tube:
| Test | Results |
| ------------ | :-----------------------------: |
| Failed | 1 |
| Passed | 24019 |
| Skipped | 2249 |
| Total | 26269 |
</p>
</details>
<!-- TEST RESULTS IF ANY-->
### Test errors [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//tests)
<details><summary>Expand to view the tests failures</summary><p>
##### `Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/mongodb/replstatus`
<ul>
<details><summary>Expand to view the error details</summary><p>
```
Failed
```
</p></details>
<details><summary>Expand to view the stacktrace</summary><p>
```
=== RUN TestFetch
replstatus_integration_test.go:53: Expected 0 error, had 1. [error getting replication status: error running command "replSetGetStatus" on admin db: (InvalidReplicaSetConfig) Our replica set config is invalid or we are not a member of it]
--- FAIL: TestFetch (12.92s)
```
</p></details>
</ul>
</p></details>
<!-- STEPS ERRORS IF ANY -->
### Steps errors [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//pipeline)
<details><summary>Expand to view the steps failures</summary>
<p>
> Show only the first 10 steps failures
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 29 min 26 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/16977/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 20 min 8 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/23885/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 20 min 52 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/24087/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `winlogbeat-windows-2022-windows-2022 - Install Python`
<ul>
<li>Took 2 min 23 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/13471/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `winlogbeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 53 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/14076/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `x-pack/filebeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 55 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/14236/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `x-pack/functionbeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 25 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/15375/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `x-pack/osquerybeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 53 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/14047/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `x-pack/packetbeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 54 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/15537/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `Error signal`
<ul>
<li>Took 0 min 0 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/24103/log/?start=0">here</a></li>
<li>Description: <code>Error "hudson.AbortException: script returned exit code 1"</code></l1>
</ul>
</p>
</details>
| 1.0 | Build 46 for 8.3 with status FAILURE -
## :broken_heart: Tests Failed
<!-- BUILD BADGES-->
> _the below badges are clickable and redirect to their specific view in the CI or DOCS_
[](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//pipeline) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//tests) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//changes) [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//artifacts) [](http://beats_null.docs-preview.app.elstc.co/diff) [](https://ci-stats.elastic.co/app/apm/services/beats-ci/transactions/view?rangeFrom=2022-06-28T07:24:39.656Z&rangeTo=2022-06-28T07:44:39.656Z&transactionName=Beats/beats/8.3&transactionType=job&latencyAggregationType=avg&traceId=974160cb047e3df95a0df6fba5995df1&transactionId=56f08c13620a77e0)
<!-- BUILD SUMMARY-->
<details><summary>Expand to view the summary</summary>
<p>
#### Build stats
* Start Time: 2022-06-28T07:34:39.656+0000
* Duration: 105 min 55 sec
#### Test stats :test_tube:
| Test | Results |
| ------------ | :-----------------------------: |
| Failed | 1 |
| Passed | 24019 |
| Skipped | 2249 |
| Total | 26269 |
</p>
</details>
<!-- TEST RESULTS IF ANY-->
### Test errors [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//tests)
<details><summary>Expand to view the tests failures</summary><p>
##### `Build&Test / metricbeat-goIntegTest / TestFetch – github.com/elastic/beats/v7/metricbeat/module/mongodb/replstatus`
<ul>
<details><summary>Expand to view the error details</summary><p>
```
Failed
```
</p></details>
<details><summary>Expand to view the stacktrace</summary><p>
```
=== RUN TestFetch
replstatus_integration_test.go:53: Expected 0 error, had 1. [error getting replication status: error running command "replSetGetStatus" on admin db: (InvalidReplicaSetConfig) Our replica set config is invalid or we are not a member of it]
--- FAIL: TestFetch (12.92s)
```
</p></details>
</ul>
</p></details>
<!-- STEPS ERRORS IF ANY -->
### Steps errors [](https://beats-ci.elastic.co/blue/organizations/jenkins/Beats%2Fbeats%2F8.3/detail/8.3/46//pipeline)
<details><summary>Expand to view the steps failures</summary>
<p>
> Show only the first 10 steps failures
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 29 min 26 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/16977/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 20 min 8 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/23885/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `metricbeat-goIntegTest - mage goIntegTest`
<ul>
<li>Took 20 min 52 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/24087/log/?start=0">here</a></li>
<li>Description: <code>mage goIntegTest</code></l1>
</ul>
##### `winlogbeat-windows-2022-windows-2022 - Install Python`
<ul>
<li>Took 2 min 23 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/13471/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `winlogbeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 53 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/14076/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `x-pack/filebeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 55 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/14236/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `x-pack/functionbeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 25 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/15375/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `x-pack/osquerybeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 53 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/14047/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `x-pack/packetbeat-windows-2016-windows-2016 - Install Python`
<ul>
<li>Took 2 min 54 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/15537/log/?start=0">here</a></li>
<li>Description: <code>.ci/scripts/install-tools.bat</code></l1>
</ul>
##### `Error signal`
<ul>
<li>Took 0 min 0 sec . View more details <a href="https://beats-ci.elastic.co//blue/rest/organizations/jenkins/pipelines/Beats/pipelines/beats/pipelines/8.3/runs/46/steps/24103/log/?start=0">here</a></li>
<li>Description: <code>Error "hudson.AbortException: script returned exit code 1"</code></l1>
</ul>
</p>
</details>
| non_priority | build for with status failure broken heart tests failed the below badges are clickable and redirect to their specific view in the ci or docs expand to view the summary build stats start time duration min sec test stats test tube test results failed passed skipped total test errors expand to view the tests failures build test metricbeat gointegtest testfetch – github com elastic beats metricbeat module mongodb replstatus expand to view the error details failed expand to view the stacktrace run testfetch replstatus integration test go expected error had fail testfetch steps errors expand to view the steps failures show only the first steps failures metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest metricbeat gointegtest mage gointegtest took min sec view more details a href description mage gointegtest winlogbeat windows windows install python took min sec view more details a href description ci scripts install tools bat winlogbeat windows windows install python took min sec view more details a href description ci scripts install tools bat x pack filebeat windows windows install python took min sec view more details a href description ci scripts install tools bat x pack functionbeat windows windows install python took min sec view more details a href description ci scripts install tools bat x pack osquerybeat windows windows install python took min sec view more details a href description ci scripts install tools bat x pack packetbeat windows windows install python took min sec view more details a href description ci scripts install tools bat error signal took min sec view more details a href description error hudson abortexception script returned exit code | 0 |
69,577 | 30,335,310,175 | IssuesEvent | 2023-07-11 09:11:58 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Powershell examples appear to be incorrect | app-service/svc triaged assigned-to-author doc-enhancement Pri1 escalated-product-team | Powershell examples do not appear to work any longer.
In particular the command to list hotswap logs doesn't list any logs
> Get-AzLog -ResourceGroup staging -StartTime 2023-06-01 -Caller SlotSwapJobProcessor
WARNING: 12:37:23 PM - *** The namespace for all the model classes will change from Microsoft.Azure.Management.Monitor.Management.Models to Microsoft.Azure.Management.Monitor.Models in future releases.
WARNING: 12:37:23 PM - *** The namespace for output classes will be uniform for all classes in future releases to make it independent of modifications in the model classes.
WARNING: [Get-AzActivityLog] Parameter deprecation: The DetailedOutput parameter will be deprecated in a future breaking change release.
>
and the command to cancel a deployment just says "Cannot modify this site because another operation is in progress."
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: f6e09089-1ae2-8943-5ce2-9d48f458c81f
* Version Independent ID: ba780cba-f604-b0a4-a81a-23c7d2384762
* Content: [Set up staging environments - Azure App Service](https://learn.microsoft.com/en-us/azure/app-service/deploy-staging-slots#cancel-a-pending-swap-swap-with-review-and-restore-the-source-slot-configuration)
* Content Source: [articles/app-service/deploy-staging-slots.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/app-service/deploy-staging-slots.md)
* Service: **app-service**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin** | 1.0 | Powershell examples appear to be incorrect - Powershell examples do not appear to work any longer.
In particular the command to list hotswap logs doesn't list any logs
> Get-AzLog -ResourceGroup staging -StartTime 2023-06-01 -Caller SlotSwapJobProcessor
WARNING: 12:37:23 PM - *** The namespace for all the model classes will change from Microsoft.Azure.Management.Monitor.Management.Models to Microsoft.Azure.Management.Monitor.Models in future releases.
WARNING: 12:37:23 PM - *** The namespace for output classes will be uniform for all classes in future releases to make it independent of modifications in the model classes.
WARNING: [Get-AzActivityLog] Parameter deprecation: The DetailedOutput parameter will be deprecated in a future breaking change release.
>
and the command to cancel a deployment just says "Cannot modify this site because another operation is in progress."
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: f6e09089-1ae2-8943-5ce2-9d48f458c81f
* Version Independent ID: ba780cba-f604-b0a4-a81a-23c7d2384762
* Content: [Set up staging environments - Azure App Service](https://learn.microsoft.com/en-us/azure/app-service/deploy-staging-slots#cancel-a-pending-swap-swap-with-review-and-restore-the-source-slot-configuration)
* Content Source: [articles/app-service/deploy-staging-slots.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/app-service/deploy-staging-slots.md)
* Service: **app-service**
* GitHub Login: @cephalin
* Microsoft Alias: **cephalin** | non_priority | powershell examples appear to be incorrect powershell examples do not appear to work any longer in particular the command to list hotswap logs doesn t list any logs get azlog resourcegroup staging starttime caller slotswapjobprocessor warning pm the namespace for all the model classes will change from microsoft azure management monitor management models to microsoft azure management monitor models in future releases warning pm the namespace for output classes will be uniform for all classes in future releases to make it independent of modifications in the model classes warning parameter deprecation the detailedoutput parameter will be deprecated in a future breaking change release and the command to cancel a deployment just says cannot modify this site because another operation is in progress document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service app service github login cephalin microsoft alias cephalin | 0 |
52,472 | 7,765,313,764 | IssuesEvent | 2018-06-02 01:49:13 | eslint/eslint | https://api.github.com/repos/eslint/eslint | closed | Docs: Move custom parser docs into their own page? | accepted documentation | Just raising this as a minor information architecture question.
Custom parsers can potentially be very complicated (if they support `parseForESLint` and expose parser services, a custom scope manager, and a custom set of visitor keys). Right now, the documentation for custom parsers is all on the "Working with Plugins" page. I think it would make sense to extract custom parsers to their own page and to have "Working with Plugins" link to that page (maybe only retaining the information needed for exposing or consuming the custom parser in the plugin itself).
On the other hand, maybe this could confuse new users (who might, e.g., think they need to write custom parsers when they really don't in most cases).
Curious to hear people's thoughts. | 1.0 | Docs: Move custom parser docs into their own page? - Just raising this as a minor information architecture question.
Custom parsers can potentially be very complicated (if they support `parseForESLint` and expose parser services, a custom scope manager, and a custom set of visitor keys). Right now, the documentation for custom parsers is all on the "Working with Plugins" page. I think it would make sense to extract custom parsers to their own page and to have "Working with Plugins" link to that page (maybe only retaining the information needed for exposing or consuming the custom parser in the plugin itself).
On the other hand, maybe this could confuse new users (who might, e.g., think they need to write custom parsers when they really don't in most cases).
Curious to hear people's thoughts. | non_priority | docs move custom parser docs into their own page just raising this as a minor information architecture question custom parsers can potentially be very complicated if they support parseforeslint and expose parser services a custom scope manager and a custom set of visitor keys right now the documentation for custom parsers is all on the working with plugins page i think it would make sense to extract custom parsers to their own page and to have working with plugins link to that page maybe only retaining the information needed for exposing or consuming the custom parser in the plugin itself on the other hand maybe this could confuse new users who might e g think they need to write custom parsers when they really don t in most cases curious to hear people s thoughts | 0 |
2,166 | 2,887,419,064 | IssuesEvent | 2015-06-12 14:52:03 | mitchellh/packer | https://api.github.com/repos/mitchellh/packer | closed | x509 errors | bug builder/openstack | Getting error when attempting to use packer provisioner against RDO - Icehouse Openstack.
Build 'openstack' errored: Post https://ost.<domain>.com:5000/v2.0/tokens: x509: certificate signed by unknown authority
I'm running Packer on Mac OSX 10.9.5. Seemingly no capability to accept/ignore self-signed certs exists.
This issue is Mac specific - the problem seems to be from golang itself. In src/pkg/crypto/x509/root_cgo_darwin.go they attempt to get the system default trusted roots using the keychain. But it appears it only grabs the ‘System Roots’ keychain, and does not make use of any certificates marked as trusted in the ‘System’ or ‘login’ keychain
debug --
https://gist.github.com/civik/d1d4be73b72287e05a64 | 1.0 | x509 errors - Getting error when attempting to use packer provisioner against RDO - Icehouse Openstack.
Build 'openstack' errored: Post https://ost.<domain>.com:5000/v2.0/tokens: x509: certificate signed by unknown authority
I'm running Packer on Mac OSX 10.9.5. Seemingly no capability to accept/ignore self-signed certs exists.
This issue is Mac specific - the problem seems to be from golang itself. In src/pkg/crypto/x509/root_cgo_darwin.go they attempt to get the system default trusted roots using the keychain. But it appears it only grabs the ‘System Roots’ keychain, and does not make use of any certificates marked as trusted in the ‘System’ or ‘login’ keychain
debug --
https://gist.github.com/civik/d1d4be73b72287e05a64 | non_priority | errors getting error when attempting to use packer provisioner against rdo icehouse openstack build openstack errored post certificate signed by unknown authority i m running packer on mac osx seemingly no capability to accept ignore self signed certs exists this issue is mac specific the problem seems to be from golang itself in src pkg crypto root cgo darwin go they attempt to get the system default trusted roots using the keychain but it appears it only grabs the ‘system roots’ keychain and does not make use of any certificates marked as trusted in the ‘system’ or ‘login’ keychain debug | 0 |
49,084 | 20,574,149,467 | IssuesEvent | 2022-03-04 01:25:31 | Azure/azure-sdk-for-net | https://api.github.com/repos/Azure/azure-sdk-for-net | closed | [BUG] ObjectReplicationSourceProperties is null | Storage Service Attention Client customer-reported bug needs-team-attention | ### Library name and version
Azure.Storage.Blobs v12.10.0
### Describe the bug
We are testing the Object Replication feature for blobs. Based on this [link ](https://docs.microsoft.com/en-us/azure/storage/blobs/object-replication-configure?tabs=powershell#check-the-replication-status-of-a-blob) we should be able to check the replication status of the blobs through the ObjectReplicationSourceProperties. However, the property is coming as null for all blobs. As a result, blob.ObjectReplicationSourceProperties[0].Rules[0].ReplicationStatus does not work and throws null reference exception. We are testing this on an account where the replication has already taken place and the replicated blobs can be seen in the destination account. However, the ObjectReplicationSourceProperties is coming as null for blobs in both source and destination accounts. In the Azure portal, the replication properties are visible.
### Expected behavior
ObjectReplicationSourceProperties[0].Rules[0].ReplicationStatus should return complete or failure for blobs in storage accounts where replication policy has been applied.
### Actual behavior
ObjectReplicationSourceProperties is coming as null.
### Reproduction Steps
1) Followed the steps in this page for the portal to create the replication policies: https://docs.microsoft.com/en-us/azure/storage/blobs/object-replication-configure?tabs=portal
2) Used the code below to list and iterate through the blobs. Replace the variables in <> for the account, container, access key
```
using System;
using System.Collections.Generic;
using System.Linq;
using System.Text;
using System.Threading.Tasks;
using Azure;
using Azure.Storage;
using Azure.Storage.Blobs;
using Azure.Storage.Blobs.Models;
namespace TestBlobReplication
{
class Program
{
//Azure storage account information
private static readonly string storageAccountName = <**STORAGE ACCOUNT NAME**>;
private static readonly string storageContainerName = <**CONTAINER NAME**>;
private static readonly string AzureStoragePrimaryKey = <**PRIMARY ACCESS KEY**>;
private static readonly string AzureStorageConnectionString = $"DefaultEndpointsProtocol=https;AccountName={storageAccountName};AccountKey={AzureStoragePrimaryKey};EndpointSuffix=core.windows.net";
private static BlobContainerClient blobContainerClient;
public async Task<BlobContainerClient> GetBlobContainerReferenceAsync(string containerName)
{
var blobContainerClient = new BlobContainerClient(AzureStorageConnectionString, containerName);
bool exists = await blobContainerClient.ExistsAsync();
if (exists)
{
Console.WriteLine($"{DateTime.Now} \n2. Found Container \n{containerName}\n");
}
else
{
await blobContainerClient.CreateAsync();
Console.WriteLine($"{DateTime.Now} Blob container: {containerName} inside storage account: {blobContainerClient.AccountName} does not exist, container created");
//throw new InvalidOperationException("Storage Account/Container not found");
}
return blobContainerClient;
}
public async Task<IList<BlobItem>> ListBlobMetadataAsync(
string prefix,
string continuationToken,
int maxResultCount = 5000)
{
// TODO: make this.container auto initialize
IAsyncEnumerable<Page<BlobItem>> blobPages = blobContainerClient.GetBlobsAsync(
BlobTraits.Metadata,
BlobStates.None,
prefix)
.AsPages(continuationToken, maxResultCount);
IAsyncEnumerator<Page<BlobItem>> enumerator = blobPages.GetAsyncEnumerator();
IList<BlobItem> resultSegment = new List<BlobItem>();
string newContinuationToken = null;
try
{
while (await enumerator.MoveNextAsync())
{
Page<BlobItem> blobPage = enumerator.Current;
foreach (BlobItem blobItem in blobPage.Values)
{
resultSegment.Add(blobItem);
}
newContinuationToken = blobPage.ContinuationToken;
}
}
finally
{
await enumerator.DisposeAsync();
}
return resultSegment;
}
static async Task Main(string[] args)
{
Console.WriteLine($"{DateTime.Now} Beginning operations...");
Program p = new Program();
blobContainerClient = await p.GetBlobContainerReferenceAsync(storageContainerName);
//BlobClient blobClient = blobContainerClient.GetBlobClient(blobId);
IList<BlobItem> blobs = await p.ListBlobMetadataAsync(null, null);
foreach (var blob in blobs)
{
var metadata = blob.Metadata;
var createdTime = blob.Properties.CreatedOn;
var replicated = blob.ObjectReplicationSourceProperties[0].Rules[0].ReplicationStatus;
}
}
}
}
```
### Environment
.NET CORE 3.1
Visual Studio 2019 | 1.0 | [BUG] ObjectReplicationSourceProperties is null - ### Library name and version
Azure.Storage.Blobs v12.10.0
### Describe the bug
We are testing the Object Replication feature for blobs. Based on this [link ](https://docs.microsoft.com/en-us/azure/storage/blobs/object-replication-configure?tabs=powershell#check-the-replication-status-of-a-blob) we should be able to check the replication status of the blobs through the ObjectReplicationSourceProperties. However, the property is coming as null for all blobs. As a result, blob.ObjectReplicationSourceProperties[0].Rules[0].ReplicationStatus does not work and throws null reference exception. We are testing this on an account where the replication has already taken place and the replicated blobs can be seen in the destination account. However, the ObjectReplicationSourceProperties is coming as null for blobs in both source and destination accounts. In the Azure portal, the replication properties are visible.
### Expected behavior
ObjectReplicationSourceProperties[0].Rules[0].ReplicationStatus should return complete or failure for blobs in storage accounts where replication policy has been applied.
### Actual behavior
ObjectReplicationSourceProperties is coming as null.
### Reproduction Steps
1) Followed the steps in this page for the portal to create the replication policies: https://docs.microsoft.com/en-us/azure/storage/blobs/object-replication-configure?tabs=portal
2) Used the code below to list and iterate through the blobs. Replace the variables in <> for the account, container, access key
```
using System;
using System.Collections.Generic;
using System.Linq;
using System.Text;
using System.Threading.Tasks;
using Azure;
using Azure.Storage;
using Azure.Storage.Blobs;
using Azure.Storage.Blobs.Models;
namespace TestBlobReplication
{
class Program
{
//Azure storage account information
private static readonly string storageAccountName = <**STORAGE ACCOUNT NAME**>;
private static readonly string storageContainerName = <**CONTAINER NAME**>;
private static readonly string AzureStoragePrimaryKey = <**PRIMARY ACCESS KEY**>;
private static readonly string AzureStorageConnectionString = $"DefaultEndpointsProtocol=https;AccountName={storageAccountName};AccountKey={AzureStoragePrimaryKey};EndpointSuffix=core.windows.net";
private static BlobContainerClient blobContainerClient;
public async Task<BlobContainerClient> GetBlobContainerReferenceAsync(string containerName)
{
var blobContainerClient = new BlobContainerClient(AzureStorageConnectionString, containerName);
bool exists = await blobContainerClient.ExistsAsync();
if (exists)
{
Console.WriteLine($"{DateTime.Now} \n2. Found Container \n{containerName}\n");
}
else
{
await blobContainerClient.CreateAsync();
Console.WriteLine($"{DateTime.Now} Blob container: {containerName} inside storage account: {blobContainerClient.AccountName} does not exist, container created");
//throw new InvalidOperationException("Storage Account/Container not found");
}
return blobContainerClient;
}
public async Task<IList<BlobItem>> ListBlobMetadataAsync(
string prefix,
string continuationToken,
int maxResultCount = 5000)
{
// TODO: make this.container auto initialize
IAsyncEnumerable<Page<BlobItem>> blobPages = blobContainerClient.GetBlobsAsync(
BlobTraits.Metadata,
BlobStates.None,
prefix)
.AsPages(continuationToken, maxResultCount);
IAsyncEnumerator<Page<BlobItem>> enumerator = blobPages.GetAsyncEnumerator();
IList<BlobItem> resultSegment = new List<BlobItem>();
string newContinuationToken = null;
try
{
while (await enumerator.MoveNextAsync())
{
Page<BlobItem> blobPage = enumerator.Current;
foreach (BlobItem blobItem in blobPage.Values)
{
resultSegment.Add(blobItem);
}
newContinuationToken = blobPage.ContinuationToken;
}
}
finally
{
await enumerator.DisposeAsync();
}
return resultSegment;
}
static async Task Main(string[] args)
{
Console.WriteLine($"{DateTime.Now} Beginning operations...");
Program p = new Program();
blobContainerClient = await p.GetBlobContainerReferenceAsync(storageContainerName);
//BlobClient blobClient = blobContainerClient.GetBlobClient(blobId);
IList<BlobItem> blobs = await p.ListBlobMetadataAsync(null, null);
foreach (var blob in blobs)
{
var metadata = blob.Metadata;
var createdTime = blob.Properties.CreatedOn;
var replicated = blob.ObjectReplicationSourceProperties[0].Rules[0].ReplicationStatus;
}
}
}
}
```
### Environment
.NET CORE 3.1
Visual Studio 2019 | non_priority | objectreplicationsourceproperties is null library name and version azure storage blobs describe the bug we are testing the object replication feature for blobs based on this we should be able to check the replication status of the blobs through the objectreplicationsourceproperties however the property is coming as null for all blobs as a result blob objectreplicationsourceproperties rules replicationstatus does not work and throws null reference exception we are testing this on an account where the replication has already taken place and the replicated blobs can be seen in the destination account however the objectreplicationsourceproperties is coming as null for blobs in both source and destination accounts in the azure portal the replication properties are visible expected behavior objectreplicationsourceproperties rules replicationstatus should return complete or failure for blobs in storage accounts where replication policy has been applied actual behavior objectreplicationsourceproperties is coming as null reproduction steps followed the steps in this page for the portal to create the replication policies used the code below to list and iterate through the blobs replace the variables in for the account container access key using system using system collections generic using system linq using system text using system threading tasks using azure using azure storage using azure storage blobs using azure storage blobs models namespace testblobreplication class program azure storage account information private static readonly string storageaccountname private static readonly string storagecontainername private static readonly string azurestorageprimarykey private static readonly string azurestorageconnectionstring defaultendpointsprotocol https accountname storageaccountname accountkey azurestorageprimarykey endpointsuffix core windows net private static blobcontainerclient blobcontainerclient public async task getblobcontainerreferenceasync string containername var blobcontainerclient new blobcontainerclient azurestorageconnectionstring containername bool exists await blobcontainerclient existsasync if exists console writeline datetime now found container n containername n else await blobcontainerclient createasync console writeline datetime now blob container containername inside storage account blobcontainerclient accountname does not exist container created throw new invalidoperationexception storage account container not found return blobcontainerclient public async task listblobmetadataasync string prefix string continuationtoken int maxresultcount todo make this container auto initialize iasyncenumerable blobpages blobcontainerclient getblobsasync blobtraits metadata blobstates none prefix aspages continuationtoken maxresultcount iasyncenumerator enumerator blobpages getasyncenumerator ilist resultsegment new list string newcontinuationtoken null try while await enumerator movenextasync page blobpage enumerator current foreach blobitem blobitem in blobpage values resultsegment add blobitem newcontinuationtoken blobpage continuationtoken finally await enumerator disposeasync return resultsegment static async task main string args console writeline datetime now beginning operations program p new program blobcontainerclient await p getblobcontainerreferenceasync storagecontainername blobclient blobclient blobcontainerclient getblobclient blobid ilist blobs await p listblobmetadataasync null null foreach var blob in blobs var metadata blob metadata var createdtime blob properties createdon var replicated blob objectreplicationsourceproperties rules replicationstatus environment net core visual studio | 0 |
9,472 | 6,311,569,689 | IssuesEvent | 2017-07-23 20:41:45 | clojurewerkz/elastisch | https://api.github.com/repos/clojurewerkz/elastisch | closed | `c.e.r.response/ok?` erroneously reports false when updating a document | low-hanging fruit usability | Elastisch’s [Getting Started](https://github.com/clojurewerkz/elastisch.docs/blob/2e28162c779ffdbe3e620970682a763389bcadda/articles/getting_started.md#checking-responses) and [Indexing](https://github.com/clojurewerkz/elastisch.docs/blob/2e28162c779ffdbe3e620970682a763389bcadda/articles/indexing.md#with-http-client) pages make it sound like `clojurewerkz.elastisch.rest.response/ok?` should _always_ be used to determine if an Elasticsearch command succeeded, but it can give false negatives. Specifically, as the following REPL interactions demonstrate, it won't recognize a successful document update. I believe this is because `c.e.r.response/ok?` merely delegates to `c.e.r.response/created?` but [successful _updates_ return `"created": false`](https://www.elastic.co/guide/en/elasticsearch/guide/1.x/update-doc.html).
``` clojure
(require '(clojurewerkz.elastisch.rest [document :as es.document]
[response :as es.response]))
;= nil
(def es-conn (clojurewerkz.elastisch.rest/connect))
;= #'user/es-conn
(es.document/put es-conn "example-index" "example-mapping" "example-id"
{"example_attribute" "initial value"})
;= {:_index "example-index", :_type "example-mapping", :_id "example-id", :_version 1, :created true}
(es.response/ok? *1)
;= true
(es.document/put es-conn "example-index" "example-mapping" "example-id"
{"example_attribute" "*updated* value"})
;= {:_index "example-index", :_type "example-mapping", :_id "example-id", :_version 2, :created false}
(es.response/ok? *1)
;= false
(es.document/get es-conn "example-index" "example-mapping" "example-id")
;= {:_index "example-index", :_type "example-mapping", :_id "example-id", :_version 2, :found true, :_source {:example_attribute "*updated* value"}
```
Note that the returned document contains the updated value, indicating that the PUT did actually succeed.
| True | `c.e.r.response/ok?` erroneously reports false when updating a document - Elastisch’s [Getting Started](https://github.com/clojurewerkz/elastisch.docs/blob/2e28162c779ffdbe3e620970682a763389bcadda/articles/getting_started.md#checking-responses) and [Indexing](https://github.com/clojurewerkz/elastisch.docs/blob/2e28162c779ffdbe3e620970682a763389bcadda/articles/indexing.md#with-http-client) pages make it sound like `clojurewerkz.elastisch.rest.response/ok?` should _always_ be used to determine if an Elasticsearch command succeeded, but it can give false negatives. Specifically, as the following REPL interactions demonstrate, it won't recognize a successful document update. I believe this is because `c.e.r.response/ok?` merely delegates to `c.e.r.response/created?` but [successful _updates_ return `"created": false`](https://www.elastic.co/guide/en/elasticsearch/guide/1.x/update-doc.html).
``` clojure
(require '(clojurewerkz.elastisch.rest [document :as es.document]
[response :as es.response]))
;= nil
(def es-conn (clojurewerkz.elastisch.rest/connect))
;= #'user/es-conn
(es.document/put es-conn "example-index" "example-mapping" "example-id"
{"example_attribute" "initial value"})
;= {:_index "example-index", :_type "example-mapping", :_id "example-id", :_version 1, :created true}
(es.response/ok? *1)
;= true
(es.document/put es-conn "example-index" "example-mapping" "example-id"
{"example_attribute" "*updated* value"})
;= {:_index "example-index", :_type "example-mapping", :_id "example-id", :_version 2, :created false}
(es.response/ok? *1)
;= false
(es.document/get es-conn "example-index" "example-mapping" "example-id")
;= {:_index "example-index", :_type "example-mapping", :_id "example-id", :_version 2, :found true, :_source {:example_attribute "*updated* value"}
```
Note that the returned document contains the updated value, indicating that the PUT did actually succeed.
| non_priority | c e r response ok erroneously reports false when updating a document elastisch’s and pages make it sound like clojurewerkz elastisch rest response ok should always be used to determine if an elasticsearch command succeeded but it can give false negatives specifically as the following repl interactions demonstrate it won t recognize a successful document update i believe this is because c e r response ok merely delegates to c e r response created but clojure require clojurewerkz elastisch rest nil def es conn clojurewerkz elastisch rest connect user es conn es document put es conn example index example mapping example id example attribute initial value index example index type example mapping id example id version created true es response ok true es document put es conn example index example mapping example id example attribute updated value index example index type example mapping id example id version created false es response ok false es document get es conn example index example mapping example id index example index type example mapping id example id version found true source example attribute updated value note that the returned document contains the updated value indicating that the put did actually succeed | 0 |
169,441 | 20,841,700,489 | IssuesEvent | 2022-03-21 01:20:40 | dmartinez777/color-picker-test | https://api.github.com/repos/dmartinez777/color-picker-test | opened | CVE-2022-24773 (Medium) detected in node-forge-0.10.0.tgz | security vulnerability | ## CVE-2022-24773 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not properly check `DigestInfo` for a proper ASN.1 structure. This can lead to successful verification with signatures that contain invalid structures but a valid digest. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24773>CVE-2022-24773</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24773">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24773</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution: node-forge - 1.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-24773 (Medium) detected in node-forge-0.10.0.tgz - ## CVE-2022-24773 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-forge-0.10.0.tgz</b></p></summary>
<p>JavaScript implementations of network transports, cryptography, ciphers, PKI, message digests, and various utilities.</p>
<p>Library home page: <a href="https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz">https://registry.npmjs.org/node-forge/-/node-forge-0.10.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/node-forge/package.json</p>
<p>
Dependency Hierarchy:
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Forge (also called `node-forge`) is a native implementation of Transport Layer Security in JavaScript. Prior to version 1.3.0, RSA PKCS#1 v1.5 signature verification code does not properly check `DigestInfo` for a proper ASN.1 structure. This can lead to successful verification with signatures that contain invalid structures but a valid digest. The issue has been addressed in `node-forge` version 1.3.0. There are currently no known workarounds.
<p>Publish Date: 2022-03-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24773>CVE-2022-24773</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24773">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24773</a></p>
<p>Release Date: 2022-03-18</p>
<p>Fix Resolution: node-forge - 1.3.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_priority | cve medium detected in node forge tgz cve medium severity vulnerability vulnerable library node forge tgz javascript implementations of network transports cryptography ciphers pki message digests and various utilities library home page a href path to dependency file package json path to vulnerable library node modules node forge package json dependency hierarchy found in base branch master vulnerability details forge also called node forge is a native implementation of transport layer security in javascript prior to version rsa pkcs signature verification code does not properly check digestinfo for a proper asn structure this can lead to successful verification with signatures that contain invalid structures but a valid digest the issue has been addressed in node forge version there are currently no known workarounds publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution node forge step up your open source security game with whitesource | 0 |
235,424 | 19,346,011,230 | IssuesEvent | 2021-12-15 10:52:44 | mozilla-mobile/firefox-ios | https://api.github.com/repos/mozilla-mobile/firefox-ios | opened | [XCUITest] Create automated test to check Customize Homescreen feature | eng:automation eng:ui-test | There is a [manual test](Customize Homepage) part of the smoketest that would be nice to have automated | 1.0 | [XCUITest] Create automated test to check Customize Homescreen feature - There is a [manual test](Customize Homepage) part of the smoketest that would be nice to have automated | non_priority | create automated test to check customize homescreen feature there is a customize homepage part of the smoketest that would be nice to have automated | 0 |
180,612 | 30,536,084,831 | IssuesEvent | 2023-07-19 17:27:31 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | opened | Interview VFS teams to learn about DI experiences | service-design | ## Purpose
Let's learn from VFS teams about what made DI clear, unclear, difficult, time-consuming, easier, etc
## Tasks
- [ ] Get list of teams from Governance who have gone through DI recently
- [ ] Schedule sessions
- [ ] Interview VFS teams
- [ ] Synthesize takeaways
## Acceptance Criteria
- [ ] Documented learnings
| 1.0 | Interview VFS teams to learn about DI experiences - ## Purpose
Let's learn from VFS teams about what made DI clear, unclear, difficult, time-consuming, easier, etc
## Tasks
- [ ] Get list of teams from Governance who have gone through DI recently
- [ ] Schedule sessions
- [ ] Interview VFS teams
- [ ] Synthesize takeaways
## Acceptance Criteria
- [ ] Documented learnings
| non_priority | interview vfs teams to learn about di experiences purpose let s learn from vfs teams about what made di clear unclear difficult time consuming easier etc tasks get list of teams from governance who have gone through di recently schedule sessions interview vfs teams synthesize takeaways acceptance criteria documented learnings | 0 |
311,692 | 23,400,549,234 | IssuesEvent | 2022-08-12 07:29:02 | lorenzo-rovigatti/oxDNA | https://api.github.com/repos/lorenzo-rovigatti/oxDNA | closed | `restart_step_counter` is required when it shouldn't be | documentation | The documentation (https://lorenzo-rovigatti.github.io/oxDNA/input.html) says that the default value of `restart_step_counter` is `false`. However, when I don't include it in my input file, I get the following error:
``ERROR: Mandatory key `restart_step_counter' not found``
Again, happy to provide an explicit input file if that's helpful
| 1.0 | `restart_step_counter` is required when it shouldn't be - The documentation (https://lorenzo-rovigatti.github.io/oxDNA/input.html) says that the default value of `restart_step_counter` is `false`. However, when I don't include it in my input file, I get the following error:
``ERROR: Mandatory key `restart_step_counter' not found``
Again, happy to provide an explicit input file if that's helpful
| non_priority | restart step counter is required when it shouldn t be the documentation says that the default value of restart step counter is false however when i don t include it in my input file i get the following error error mandatory key restart step counter not found again happy to provide an explicit input file if that s helpful | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.