Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
425,484
12,341,042,572
IssuesEvent
2020-05-14 21:05:36
huridocs/uwazi
https://api.github.com/repos/huridocs/uwazi
closed
Fix intermitent tests for CSV Export
Bug Priority: High Status: Sprint
- [x] Use supertest to test res.download and not route return - [x] Check validation (specially `search` which is currently an array) - [x] Test properly the file unlink, this probably means mocking the file generation naming. Please ensure that the name can not be duplicated if tests are run in paralell - [x] Test for correct passing of the user to the search function (perhaps @daneryl can help set this up quicker) - [x] test that the exporter got called correctly with the desired params, even if its reply is mocked. A good approach is to tailor the reply according to the arguments passed, in that way you are testing, with the response, the arguments passed. - [x] Look if there is merit to create a separate file for the `api/export/` route?
1.0
Fix intermitent tests for CSV Export - - [x] Use supertest to test res.download and not route return - [x] Check validation (specially `search` which is currently an array) - [x] Test properly the file unlink, this probably means mocking the file generation naming. Please ensure that the name can not be duplicated if tests are run in paralell - [x] Test for correct passing of the user to the search function (perhaps @daneryl can help set this up quicker) - [x] test that the exporter got called correctly with the desired params, even if its reply is mocked. A good approach is to tailor the reply according to the arguments passed, in that way you are testing, with the response, the arguments passed. - [x] Look if there is merit to create a separate file for the `api/export/` route?
non_main
fix intermitent tests for csv export use supertest to test res download and not route return check validation specially search which is currently an array test properly the file unlink this probably means mocking the file generation naming please ensure that the name can not be duplicated if tests are run in paralell test for correct passing of the user to the search function perhaps daneryl can help set this up quicker test that the exporter got called correctly with the desired params even if its reply is mocked a good approach is to tailor the reply according to the arguments passed in that way you are testing with the response the arguments passed look if there is merit to create a separate file for the api export route
0
224,606
24,782,438,722
IssuesEvent
2022-10-24 06:54:13
sast-automation-dev/vulnado-43
https://api.github.com/repos/sast-automation-dev/vulnado-43
opened
handlebars-4.1.0.min.js: 4 vulnerabilities (highest severity is: 9.8)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (handlebars version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-19919](https://www.mend.io/vulnerability-database/CVE-2019-19919) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | handlebars-4.1.0.min.js | Direct | 4.3.0 | &#10060; | | [WS-2019-0064](https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | handlebars-4.1.0.min.js | Direct | 1.0.6-2,4.0.14,4.1.2 | &#10060; | | [WS-2019-0332](https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | handlebars-4.1.0.min.js | Direct | handlebars - 4.5.3 | &#10060; | | [WS-2019-0331](https://github.com/wycats/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | handlebars-4.1.0.min.js | Direct | handlebars - 4.5.2 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-19919</summary> ### Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.1.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of handlebars prior to 4.3.0 are vulnerable to Prototype Pollution leading to Remote Code Execution. Templates may alter an Object's __proto__ and __defineGetter__ properties, which may allow an attacker to execute arbitrary code through crafted payloads. <p>Publish Date: Dec 20, 2019 11:15:00 PM <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19919>CVE-2019-19919</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1164">https://www.npmjs.com/advisories/1164</a></p> <p>Release Date: Dec 20, 2019 11:15:00 PM</p> <p>Fix Resolution: 4.3.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2019-0064</summary> ### Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.1.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server. <p>Publish Date: Jan 30, 2019 9:21:44 PM <p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p> <p>Release Date: Jan 30, 2019 9:21:44 PM</p> <p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> WS-2019-0332</summary> ### Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.1.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Arbitrary Code Execution vulnerability found in handlebars before 4.5.3. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.It is due to an incomplete fix for a WS-2019-0331. <p>Publish Date: Nov 17, 2019 8:29:47 PM <p>URL: <a href=https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7>WS-2019-0332</a></p> </p> <p></p> ### CVSS 2 Score Details (<b>5.0</b>) <p> Base Score Metrics not available</p> </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p> <p>Release Date: Nov 17, 2019 8:29:47 PM</p> <p>Fix Resolution: handlebars - 4.5.3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> WS-2019-0331</summary> ### Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.1.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Arbitrary Code Execution vulnerability found in handlebars before 4.5.2. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system. <p>Publish Date: Nov 13, 2019 8:41:36 PM <p>URL: <a href=https://github.com/wycats/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e>WS-2019-0331</a></p> </p> <p></p> ### CVSS 2 Score Details (<b>5.0</b>) <p> Base Score Metrics not available</p> </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p> <p>Release Date: Nov 13, 2019 8:41:36 PM</p> <p>Fix Resolution: handlebars - 4.5.2</p> </p> <p></p> </details>
True
handlebars-4.1.0.min.js: 4 vulnerabilities (highest severity is: 9.8) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (handlebars version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2019-19919](https://www.mend.io/vulnerability-database/CVE-2019-19919) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 9.8 | handlebars-4.1.0.min.js | Direct | 4.3.0 | &#10060; | | [WS-2019-0064](https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.3 | handlebars-4.1.0.min.js | Direct | 1.0.6-2,4.0.14,4.1.2 | &#10060; | | [WS-2019-0332](https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | handlebars-4.1.0.min.js | Direct | handlebars - 4.5.3 | &#10060; | | [WS-2019-0331](https://github.com/wycats/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.0 | handlebars-4.1.0.min.js | Direct | handlebars - 4.5.2 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2019-19919</summary> ### Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.1.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of handlebars prior to 4.3.0 are vulnerable to Prototype Pollution leading to Remote Code Execution. Templates may alter an Object's __proto__ and __defineGetter__ properties, which may allow an attacker to execute arbitrary code through crafted payloads. <p>Publish Date: Dec 20, 2019 11:15:00 PM <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19919>CVE-2019-19919</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>9.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1164">https://www.npmjs.com/advisories/1164</a></p> <p>Release Date: Dec 20, 2019 11:15:00 PM</p> <p>Fix Resolution: 4.3.0</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> WS-2019-0064</summary> ### Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.1.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Versions of handlebars prior to 4.0.14 are vulnerable to Prototype Pollution. Templates may alter an Objects' prototype, thus allowing an attacker to execute arbitrary code on the server. <p>Publish Date: Jan 30, 2019 9:21:44 PM <p>URL: <a href=https://github.com/wycats/handlebars.js/compare/v4.1.1...v4.1.2>WS-2019-0064</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/755/versions">https://www.npmjs.com/advisories/755/versions</a></p> <p>Release Date: Jan 30, 2019 9:21:44 PM</p> <p>Fix Resolution: 1.0.6-2,4.0.14,4.1.2</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> WS-2019-0332</summary> ### Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.1.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Arbitrary Code Execution vulnerability found in handlebars before 4.5.3. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system.It is due to an incomplete fix for a WS-2019-0331. <p>Publish Date: Nov 17, 2019 8:29:47 PM <p>URL: <a href=https://github.com/wycats/handlebars.js/commit/198887808780bbef9dba67a8af68ece091d5baa7>WS-2019-0332</a></p> </p> <p></p> ### CVSS 2 Score Details (<b>5.0</b>) <p> Base Score Metrics not available</p> </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1324">https://www.npmjs.com/advisories/1324</a></p> <p>Release Date: Nov 17, 2019 8:29:47 PM</p> <p>Fix Resolution: handlebars - 4.5.3</p> </p> <p></p> </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> WS-2019-0331</summary> ### Vulnerable Library - <b>handlebars-4.1.0.min.js</b></p> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.1.0/handlebars.min.js</a></p> <p>Path to dependency file: /client/login.html</p> <p>Path to vulnerable library: /client/login.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.1.0.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/sast-automation-dev/vulnado-43/commit/72ba4cd71f4291bcca0fcc9e1020d14d9115260e">72ba4cd71f4291bcca0fcc9e1020d14d9115260e</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> Arbitrary Code Execution vulnerability found in handlebars before 4.5.2. Lookup helper fails to validate templates. Attack may submit templates that execute arbitrary JavaScript in the system. <p>Publish Date: Nov 13, 2019 8:41:36 PM <p>URL: <a href=https://github.com/wycats/handlebars.js/commit/d54137810a49939fd2ad01a91a34e182ece4528e>WS-2019-0331</a></p> </p> <p></p> ### CVSS 2 Score Details (<b>5.0</b>) <p> Base Score Metrics not available</p> </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1316">https://www.npmjs.com/advisories/1316</a></p> <p>Release Date: Nov 13, 2019 8:41:36 PM</p> <p>Fix Resolution: handlebars - 4.5.2</p> </p> <p></p> </details>
non_main
handlebars min js vulnerabilities highest severity is vulnerable library handlebars min js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file client login html path to vulnerable library client login html found in head commit a href vulnerabilities cve severity cvss dependency type fixed in handlebars version remediation available high handlebars min js direct high handlebars min js direct medium handlebars min js direct handlebars medium handlebars min js direct handlebars details cve vulnerable library handlebars min js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file client login html path to vulnerable library client login html dependency hierarchy x handlebars min js vulnerable library found in head commit a href found in base branch master vulnerability details versions of handlebars prior to are vulnerable to prototype pollution leading to remote code execution templates may alter an object s proto and definegetter properties which may allow an attacker to execute arbitrary code through crafted payloads publish date dec pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date dec pm fix resolution ws vulnerable library handlebars min js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file client login html path to vulnerable library client login html dependency hierarchy x handlebars min js vulnerable library found in head commit a href found in base branch master vulnerability details versions of handlebars prior to are vulnerable to prototype pollution templates may alter an objects prototype thus allowing an attacker to execute arbitrary code on the server publish date jan pm url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date jan pm fix resolution ws vulnerable library handlebars min js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file client login html path to vulnerable library client login html dependency hierarchy x handlebars min js vulnerable library found in head commit a href found in base branch master vulnerability details arbitrary code execution vulnerability found in handlebars before lookup helper fails to validate templates attack may submit templates that execute arbitrary javascript in the system it is due to an incomplete fix for a ws publish date nov pm url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date nov pm fix resolution handlebars ws vulnerable library handlebars min js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file client login html path to vulnerable library client login html dependency hierarchy x handlebars min js vulnerable library found in head commit a href found in base branch master vulnerability details arbitrary code execution vulnerability found in handlebars before lookup helper fails to validate templates attack may submit templates that execute arbitrary javascript in the system publish date nov pm url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date nov pm fix resolution handlebars
0
312,324
9,545,809,291
IssuesEvent
2019-05-01 18:07:59
medic/medic
https://api.github.com/repos/medic/medic
opened
JavaScript errors and crash stacks not logged to the console
Priority: 1 - High Type: Bug
**Describe the bug** Not all JavaScript errors and crash stacks are reported to the console. **To Reproduce** Steps to reproduce the behavior: 1. Open `webapp/webapp/src/js/controllers/inbox.js` in an editor and delete `Auth` from the ngInjection list. 1. Rebuild app 1. Login to the webapp **Expected behavior** The app should fail to load. And there should be an error in the console explaining why. **Observed behavior** The console has no errors. I think this bug is important and labelled it Priority 1 because these console errors are often our only way to investigate/understand live-site issues on the client.
1.0
JavaScript errors and crash stacks not logged to the console - **Describe the bug** Not all JavaScript errors and crash stacks are reported to the console. **To Reproduce** Steps to reproduce the behavior: 1. Open `webapp/webapp/src/js/controllers/inbox.js` in an editor and delete `Auth` from the ngInjection list. 1. Rebuild app 1. Login to the webapp **Expected behavior** The app should fail to load. And there should be an error in the console explaining why. **Observed behavior** The console has no errors. I think this bug is important and labelled it Priority 1 because these console errors are often our only way to investigate/understand live-site issues on the client.
non_main
javascript errors and crash stacks not logged to the console describe the bug not all javascript errors and crash stacks are reported to the console to reproduce steps to reproduce the behavior open webapp webapp src js controllers inbox js in an editor and delete auth from the nginjection list rebuild app login to the webapp expected behavior the app should fail to load and there should be an error in the console explaining why observed behavior the console has no errors i think this bug is important and labelled it priority because these console errors are often our only way to investigate understand live site issues on the client
0
590,038
17,769,346,703
IssuesEvent
2021-08-30 11:48:58
o3de/o3de
https://api.github.com/repos/o3de/o3de
opened
Unable to add the AutomatedTesting project to the Project Manager
kind/bug needs-sig needs-triage priority/major
**Describe the bug** It is not possible to add the AutomatedTesting project to the Project Manager when it is launched from the o3de-install folder (after following the [Pre-built SDK engine guide](https://o3deorg.netlify.app/docs/welcome-guide/setup/setup-from-github/)). The workaround for the issue is to manually add the path to the project into the o3de_manifest.json file. Please refer to the attached video for more details. **To Reproduce** Steps to reproduce the behavior: 1. Follow the [Pre-built SDK engine guide](https://o3deorg.netlify.app/docs/welcome-guide/setup/setup-from-github/). 2. Launch Editor.exe from the o3de-install/bin/Windows/profile folder. 3. Try to add the AutomatedTesting project to the Project Manager. **Expected behavior** AutomatedTesting project is added to the Project Manager successfully. **Video** https://user-images.githubusercontent.com/86953108/131334244-bd32d226-227b-4409-b2e8-f8cdce42046b.mp4 **Desktop/Device:** - Device: PC - OS: Windows - Version 10 - CPU AMD Ryzen 5 3600 - GPU Nvidia RTX 2060 SUPER - Memory 16GB
1.0
Unable to add the AutomatedTesting project to the Project Manager - **Describe the bug** It is not possible to add the AutomatedTesting project to the Project Manager when it is launched from the o3de-install folder (after following the [Pre-built SDK engine guide](https://o3deorg.netlify.app/docs/welcome-guide/setup/setup-from-github/)). The workaround for the issue is to manually add the path to the project into the o3de_manifest.json file. Please refer to the attached video for more details. **To Reproduce** Steps to reproduce the behavior: 1. Follow the [Pre-built SDK engine guide](https://o3deorg.netlify.app/docs/welcome-guide/setup/setup-from-github/). 2. Launch Editor.exe from the o3de-install/bin/Windows/profile folder. 3. Try to add the AutomatedTesting project to the Project Manager. **Expected behavior** AutomatedTesting project is added to the Project Manager successfully. **Video** https://user-images.githubusercontent.com/86953108/131334244-bd32d226-227b-4409-b2e8-f8cdce42046b.mp4 **Desktop/Device:** - Device: PC - OS: Windows - Version 10 - CPU AMD Ryzen 5 3600 - GPU Nvidia RTX 2060 SUPER - Memory 16GB
non_main
unable to add the automatedtesting project to the project manager describe the bug it is not possible to add the automatedtesting project to the project manager when it is launched from the install folder after following the the workaround for the issue is to manually add the path to the project into the manifest json file please refer to the attached video for more details to reproduce steps to reproduce the behavior follow the launch editor exe from the install bin windows profile folder try to add the automatedtesting project to the project manager expected behavior automatedtesting project is added to the project manager successfully video desktop device device pc os windows version cpu amd ryzen gpu nvidia rtx super memory
0
1,075
4,892,751,448
IssuesEvent
2016-11-18 20:44:16
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
opened
Deprecate `brew cask update`
awaiting maintainer feedback
`brew cask update` is simply an alias to `brew update`. Even so, it’s somewhat common for me to see people recommend `brew cask upgrade`, or even worse, `brew update && brew cask update`. Especially now that two projects are integrated, I think we should just scrape `brew cask update` altogether, having a transitional period where it will encourage users to simply `brew update` instead. Pinging @caskroom/maintainers @MikeMcquaid.
True
Deprecate `brew cask update` - `brew cask update` is simply an alias to `brew update`. Even so, it’s somewhat common for me to see people recommend `brew cask upgrade`, or even worse, `brew update && brew cask update`. Especially now that two projects are integrated, I think we should just scrape `brew cask update` altogether, having a transitional period where it will encourage users to simply `brew update` instead. Pinging @caskroom/maintainers @MikeMcquaid.
main
deprecate brew cask update brew cask update is simply an alias to brew update even so it’s somewhat common for me to see people recommend brew cask upgrade or even worse brew update brew cask update especially now that two projects are integrated i think we should just scrape brew cask update altogether having a transitional period where it will encourage users to simply brew update instead pinging caskroom maintainers mikemcquaid
1
23,415
6,419,940,139
IssuesEvent
2017-08-08 22:30:23
xhqiao89/HydroDesktop_test
https://api.github.com/repos/xhqiao89/HydroDesktop_test
closed
Change units on graphs
CodePlex
<b>nmhydro[CodePlex]</b> <br />The prerequisite for this feature is implementing feature 8377 (Add support for unit conversion). This feature should be implemented as a quotChange Unitsquot ribbon button in the Graph tab. When the user clicks quotChange Unitsquot a new form is displayed. This form shows the list of all current variables. For each variable, it shows the current unit and a quotnew unitquot drop down. The quotnew unitquot drop down is populated with all units to which the current unit can be converted. For example if the current unit is meter then the list contains values such as millimeter, centimeter, kilometer, inch, foot, yard, mile.. When the user selects one of the options, the graphs for the selected variable are re-drawn using the new units and any future graphs of series with the selected variable will also be displayed using the selected units. [Reported originally by Stephen]
1.0
Change units on graphs - <b>nmhydro[CodePlex]</b> <br />The prerequisite for this feature is implementing feature 8377 (Add support for unit conversion). This feature should be implemented as a quotChange Unitsquot ribbon button in the Graph tab. When the user clicks quotChange Unitsquot a new form is displayed. This form shows the list of all current variables. For each variable, it shows the current unit and a quotnew unitquot drop down. The quotnew unitquot drop down is populated with all units to which the current unit can be converted. For example if the current unit is meter then the list contains values such as millimeter, centimeter, kilometer, inch, foot, yard, mile.. When the user selects one of the options, the graphs for the selected variable are re-drawn using the new units and any future graphs of series with the selected variable will also be displayed using the selected units. [Reported originally by Stephen]
non_main
change units on graphs nmhydro the prerequisite for this feature is implementing feature add support for unit conversion this feature should be implemented as a quotchange unitsquot ribbon button in the graph tab when the user clicks quotchange unitsquot a new form is displayed this form shows the list of all current variables for each variable it shows the current unit and a quotnew unitquot drop down the quotnew unitquot drop down is populated with all units to which the current unit can be converted for example if the current unit is meter then the list contains values such as millimeter centimeter kilometer inch foot yard mile when the user selects one of the options the graphs for the selected variable are re drawn using the new units and any future graphs of series with the selected variable will also be displayed using the selected units
0
443,885
30,961,248,370
IssuesEvent
2023-08-08 04:18:46
open-feature/flagd
https://api.github.com/repos/open-feature/flagd
opened
[DOC] Point from /docs/other_resources/systemd_service.md to website
documentation Needs Triage
### Change in the documentation Reword this page https://github.com/open-feature/flagd/blob/main/docs/other_resources/systemd_service.md to instead point to https://flagd.dev/nonk8s/systemservice/
1.0
[DOC] Point from /docs/other_resources/systemd_service.md to website - ### Change in the documentation Reword this page https://github.com/open-feature/flagd/blob/main/docs/other_resources/systemd_service.md to instead point to https://flagd.dev/nonk8s/systemservice/
non_main
point from docs other resources systemd service md to website change in the documentation reword this page to instead point to
0
5,159
26,271,460,759
IssuesEvent
2023-01-06 17:23:42
cosmos/ibc-rs
https://api.github.com/repos/cosmos/ibc-rs
closed
Refactor packet handler proofs
A: good first issue O: maintainability
Inline proofs similar to #217. Specifically, the [channel `verify` module](https://github.com/cosmos/ibc-rs/blob/07d4949482f0f7184e3b815f8b54421fe00fed23/crates/ibc/src/core/ics04_channel/handler.rs#L32) would be removed completely.
True
Refactor packet handler proofs - Inline proofs similar to #217. Specifically, the [channel `verify` module](https://github.com/cosmos/ibc-rs/blob/07d4949482f0f7184e3b815f8b54421fe00fed23/crates/ibc/src/core/ics04_channel/handler.rs#L32) would be removed completely.
main
refactor packet handler proofs inline proofs similar to specifically the would be removed completely
1
2,429
8,620,893,575
IssuesEvent
2018-11-20 16:03:54
simplesamlphp/simplesamlphp
https://api.github.com/repos/simplesamlphp/simplesamlphp
closed
Pre migrate templates: move logic to calling page
enhancement maintainability started
Sometimes, elaborate logic or a function in the template is there to select something from data in the context. Move this to the calling page and dump the result in the context. These cannot be found with a quick grep through the code, but an `if` or a loop that only uses what's in the context to set something only used in the template is the giveaway.
True
Pre migrate templates: move logic to calling page - Sometimes, elaborate logic or a function in the template is there to select something from data in the context. Move this to the calling page and dump the result in the context. These cannot be found with a quick grep through the code, but an `if` or a loop that only uses what's in the context to set something only used in the template is the giveaway.
main
pre migrate templates move logic to calling page sometimes elaborate logic or a function in the template is there to select something from data in the context move this to the calling page and dump the result in the context these cannot be found with a quick grep through the code but an if or a loop that only uses what s in the context to set something only used in the template is the giveaway
1
5,474
5,003,684,944
IssuesEvent
2016-12-12 00:32:25
pandas-dev/pandas
https://api.github.com/repos/pandas-dev/pandas
opened
PERF: use StringHashTable for value_counts / duplicated with strings
Difficulty Intermediate Effort Low Performance Strings
xref https://github.com/pandas-dev/pandas/pull/14859 - [ ] ``.value_counts()`` - [ ] ``.duplicated()``
True
PERF: use StringHashTable for value_counts / duplicated with strings - xref https://github.com/pandas-dev/pandas/pull/14859 - [ ] ``.value_counts()`` - [ ] ``.duplicated()``
non_main
perf use stringhashtable for value counts duplicated with strings xref value counts duplicated
0
110,169
23,883,647,373
IssuesEvent
2022-09-08 05:17:30
usdigitalresponse/univaf
https://api.github.com/repos/usdigitalresponse/univaf
opened
Consider using DataDog HTTP API for metrics
ops code quality api
We currently send metrics to data using DogStatsD, which requires having the DataDog agent running. This has worked pretty well for us on AWS/ECS, but is less than ideal on Render — you can’t have sidecar containers in Render (so the agent can’t run alongside the app). Instead, the recommended solution is to run the agent as its own service. Unfortunately, running it as its own service means: - The `host` tag is always wrong (it refers to the service the agent is running in, not whoever is sending metrics). - The agent sends normal machine stats (CPU usage, memory, etc.) for itself, which is entirely pointless in this situation. - The agent costs $18 on DataDog to be a host. One alternative is to use Docker on Render and embed the agent as a background process in the app’s Docker container. That’s generally considered to be a poor pattern for containers and adds a lot of complexity (and we are switching to Render to *reduce* complexity!). The better answer here is probably to use DataDog’s HTTP API to send metrics. A quick look around NPM makes [datadog-metrics](https://www.npmjs.com/package/datadog-metrics) look like the best option here — it gives us a very similar interface to the hot-shots package we are currently using for StatsD, and buffers & batches metric submission so we don’t have a lot of overhead from API requests. There’s also [dogapi](https://www.npmjs.com/package/dogapi), but the API is not as nice and it doesn’t do any buffering. One issue that has to be solved is dev/test support. If you create an instance of the datadog-metrics client with no valid connection info, it throws errors when you send. (This works fine with hot-shots because StatsD is UDP-based, so it just fires messages off into the void and doesn’t care if nobody receives them.) We’d need to make a mock version or some kind of wrapper to use when a DataDog API key is not configured.
1.0
Consider using DataDog HTTP API for metrics - We currently send metrics to data using DogStatsD, which requires having the DataDog agent running. This has worked pretty well for us on AWS/ECS, but is less than ideal on Render — you can’t have sidecar containers in Render (so the agent can’t run alongside the app). Instead, the recommended solution is to run the agent as its own service. Unfortunately, running it as its own service means: - The `host` tag is always wrong (it refers to the service the agent is running in, not whoever is sending metrics). - The agent sends normal machine stats (CPU usage, memory, etc.) for itself, which is entirely pointless in this situation. - The agent costs $18 on DataDog to be a host. One alternative is to use Docker on Render and embed the agent as a background process in the app’s Docker container. That’s generally considered to be a poor pattern for containers and adds a lot of complexity (and we are switching to Render to *reduce* complexity!). The better answer here is probably to use DataDog’s HTTP API to send metrics. A quick look around NPM makes [datadog-metrics](https://www.npmjs.com/package/datadog-metrics) look like the best option here — it gives us a very similar interface to the hot-shots package we are currently using for StatsD, and buffers & batches metric submission so we don’t have a lot of overhead from API requests. There’s also [dogapi](https://www.npmjs.com/package/dogapi), but the API is not as nice and it doesn’t do any buffering. One issue that has to be solved is dev/test support. If you create an instance of the datadog-metrics client with no valid connection info, it throws errors when you send. (This works fine with hot-shots because StatsD is UDP-based, so it just fires messages off into the void and doesn’t care if nobody receives them.) We’d need to make a mock version or some kind of wrapper to use when a DataDog API key is not configured.
non_main
consider using datadog http api for metrics we currently send metrics to data using dogstatsd which requires having the datadog agent running this has worked pretty well for us on aws ecs but is less than ideal on render — you can’t have sidecar containers in render so the agent can’t run alongside the app instead the recommended solution is to run the agent as its own service unfortunately running it as its own service means the host tag is always wrong it refers to the service the agent is running in not whoever is sending metrics the agent sends normal machine stats cpu usage memory etc for itself which is entirely pointless in this situation the agent costs on datadog to be a host one alternative is to use docker on render and embed the agent as a background process in the app’s docker container that’s generally considered to be a poor pattern for containers and adds a lot of complexity and we are switching to render to reduce complexity the better answer here is probably to use datadog’s http api to send metrics a quick look around npm makes look like the best option here — it gives us a very similar interface to the hot shots package we are currently using for statsd and buffers batches metric submission so we don’t have a lot of overhead from api requests there’s also but the api is not as nice and it doesn’t do any buffering one issue that has to be solved is dev test support if you create an instance of the datadog metrics client with no valid connection info it throws errors when you send this works fine with hot shots because statsd is udp based so it just fires messages off into the void and doesn’t care if nobody receives them we’d need to make a mock version or some kind of wrapper to use when a datadog api key is not configured
0
133,713
29,505,755,484
IssuesEvent
2023-06-03 09:35:35
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
opened
Simplify error messages from `pulumi convert`
kind/enhancement area/codegen
## Hello! <!-- Please leave this section as-is, it's designed to help others in the community know how to interact with our GitHub issues. --> - Vote on this issue by adding a 👍 reaction - If you want to implement this feature, comment to let us know (we'll work with you on design, scheduling, etc.) ## Issue details Currently, when programgen fails during a `pulumi convert`, we output PCL-related errors. But we don't want PCL to show up as a concept to users. The preference is to show a simpler error message rather than showing any PCL-related errors. The PCL errors should be emitted in advanced diagnostic logs that can be accessed with a higher verbosity (for us to use to help users through a conversion), but otherwise, a more generic "couldn't convert" error should be shown. How good can we make the generic error? Can we tell the user what resource has the error? What property of that resource?
1.0
Simplify error messages from `pulumi convert` - ## Hello! <!-- Please leave this section as-is, it's designed to help others in the community know how to interact with our GitHub issues. --> - Vote on this issue by adding a 👍 reaction - If you want to implement this feature, comment to let us know (we'll work with you on design, scheduling, etc.) ## Issue details Currently, when programgen fails during a `pulumi convert`, we output PCL-related errors. But we don't want PCL to show up as a concept to users. The preference is to show a simpler error message rather than showing any PCL-related errors. The PCL errors should be emitted in advanced diagnostic logs that can be accessed with a higher verbosity (for us to use to help users through a conversion), but otherwise, a more generic "couldn't convert" error should be shown. How good can we make the generic error? Can we tell the user what resource has the error? What property of that resource?
non_main
simplify error messages from pulumi convert hello vote on this issue by adding a 👍 reaction if you want to implement this feature comment to let us know we ll work with you on design scheduling etc issue details currently when programgen fails during a pulumi convert we output pcl related errors but we don t want pcl to show up as a concept to users the preference is to show a simpler error message rather than showing any pcl related errors the pcl errors should be emitted in advanced diagnostic logs that can be accessed with a higher verbosity for us to use to help users through a conversion but otherwise a more generic couldn t convert error should be shown how good can we make the generic error can we tell the user what resource has the error what property of that resource
0
242,898
26,277,866,627
IssuesEvent
2023-01-07 01:21:59
AlexRogalskiy/scala-patterns
https://api.github.com/repos/AlexRogalskiy/scala-patterns
opened
CVE-2022-21670 (Medium) detected in markdown-it-12.0.4.tgz
security vulnerability
## CVE-2022-21670 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>markdown-it-12.0.4.tgz</b></p></summary> <p>Markdown-it - modern pluggable markdown parser.</p> <p>Library home page: <a href="https://registry.npmjs.org/markdown-it/-/markdown-it-12.0.4.tgz">https://registry.npmjs.org/markdown-it/-/markdown-it-12.0.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/markdown-it/package.json</p> <p> Dependency Hierarchy: - markdownlint-cli-0.27.1.tgz (Root Library) - markdownlint-0.23.1.tgz - :x: **markdown-it-12.0.4.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> markdown-it is a Markdown parser. Prior to version 1.3.2, special patterns with length greater than 50 thousand characterss could slow down the parser significantly. Users should upgrade to version 12.3.2 to receive a patch. There are no known workarounds aside from upgrading. <p>Publish Date: 2022-01-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-21670>CVE-2022-21670</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c">https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution (markdown-it): 12.3.2</p> <p>Direct dependency fix Resolution (markdownlint-cli): 0.31.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-21670 (Medium) detected in markdown-it-12.0.4.tgz - ## CVE-2022-21670 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>markdown-it-12.0.4.tgz</b></p></summary> <p>Markdown-it - modern pluggable markdown parser.</p> <p>Library home page: <a href="https://registry.npmjs.org/markdown-it/-/markdown-it-12.0.4.tgz">https://registry.npmjs.org/markdown-it/-/markdown-it-12.0.4.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/markdown-it/package.json</p> <p> Dependency Hierarchy: - markdownlint-cli-0.27.1.tgz (Root Library) - markdownlint-0.23.1.tgz - :x: **markdown-it-12.0.4.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> markdown-it is a Markdown parser. Prior to version 1.3.2, special patterns with length greater than 50 thousand characterss could slow down the parser significantly. Users should upgrade to version 12.3.2 to receive a patch. There are no known workarounds aside from upgrading. <p>Publish Date: 2022-01-10 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-21670>CVE-2022-21670</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c">https://github.com/markdown-it/markdown-it/security/advisories/GHSA-6vfc-qv3f-vr6c</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution (markdown-it): 12.3.2</p> <p>Direct dependency fix Resolution (markdownlint-cli): 0.31.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in markdown it tgz cve medium severity vulnerability vulnerable library markdown it tgz markdown it modern pluggable markdown parser library home page a href path to dependency file package json path to vulnerable library node modules markdown it package json dependency hierarchy markdownlint cli tgz root library markdownlint tgz x markdown it tgz vulnerable library found in base branch master vulnerability details markdown it is a markdown parser prior to version special patterns with length greater than thousand characterss could slow down the parser significantly users should upgrade to version to receive a patch there are no known workarounds aside from upgrading publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution markdown it direct dependency fix resolution markdownlint cli step up your open source security game with mend
0
27,386
21,685,385,697
IssuesEvent
2022-05-09 10:44:58
effektio/effektio
https://api.github.com/repos/effektio/effektio
opened
Deploy Chatterbox
infrastructure s-chat
https://github.com/vector-im/chatterbox - an in-website chat-box linked to the support channel of the specific project room.
1.0
Deploy Chatterbox - https://github.com/vector-im/chatterbox - an in-website chat-box linked to the support channel of the specific project room.
non_main
deploy chatterbox an in website chat box linked to the support channel of the specific project room
0
2,851
10,219,697,498
IssuesEvent
2019-08-15 19:16:35
lrozenblyum/chess
https://api.github.com/repos/lrozenblyum/chess
opened
Winboard player construction - simplify
maintainability winboard
During implementation of #278 it has become obvious that winboard player creation via static 'factory' method looks out-of-common style. It is also against OOP principles. We should allow WinboardPlayer creation via poor constructor. Risks: a) 'singleton' enforcement won't be as easy (although we don't have it right now) b) some tests which were depending on our 2-phase construction of the player (with separate initCommander call) may behave differently and not test well what's needed. Requires good investigation.
True
Winboard player construction - simplify - During implementation of #278 it has become obvious that winboard player creation via static 'factory' method looks out-of-common style. It is also against OOP principles. We should allow WinboardPlayer creation via poor constructor. Risks: a) 'singleton' enforcement won't be as easy (although we don't have it right now) b) some tests which were depending on our 2-phase construction of the player (with separate initCommander call) may behave differently and not test well what's needed. Requires good investigation.
main
winboard player construction simplify during implementation of it has become obvious that winboard player creation via static factory method looks out of common style it is also against oop principles we should allow winboardplayer creation via poor constructor risks a singleton enforcement won t be as easy although we don t have it right now b some tests which were depending on our phase construction of the player with separate initcommander call may behave differently and not test well what s needed requires good investigation
1
110,859
9,480,324,817
IssuesEvent
2019-04-20 16:49:50
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed test: TestLint
C-test-failure O-robot
The following tests appear to have failed on master (lint): TestLint/TestVet, TestLint/TestVet: TestLint/TestVet/shadow, TestLint You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestLint). [#1251247](https://teamcity.cockroachdb.com/viewLog.html?buildId=1251247): ``` TestLint --- FAIL: lint/TestLint (237.270s) TestLint/TestVet: TestLint/TestVet/shadow ...0 +0x42 fp=0x7ffc59860a60 sp=0x7ffc59860a08 pc=0x423092 lint_test.go:1308: runtime.(*mheap).allocSpanLocked(0xa03820, 0x1, 0xa1bfe8, 0x0) lint_test.go:1308: /usr/local/go/src/runtime/mheap.go:848 +0x337 fp=0x7ffc59860aa0 sp=0x7ffc59860a60 pc=0x422f17 lint_test.go:1308: runtime.(*mheap).alloc_m(0xa03820, 0x1, 0x2a, 0x0) lint_test.go:1308: /usr/local/go/src/runtime/mheap.go:692 +0x119 fp=0x7ffc59860ae0 sp=0x7ffc59860aa0 pc=0x422729 lint_test.go:1308: runtime.(*mheap).alloc.func1() lint_test.go:1308: /usr/local/go/src/runtime/mheap.go:759 +0x4c fp=0x7ffc59860b18 sp=0x7ffc59860ae0 pc=0x455f7c lint_test.go:1308: runtime.(*mheap).alloc(0xa03820, 0x1, 0x1002a, 0x0) lint_test.go:1308: /usr/local/go/src/runtime/mheap.go:758 +0x8a fp=0x7ffc59860b68 sp=0x7ffc59860b18 pc=0x4229ca lint_test.go:1308: runtime.(*mcentral).grow(0xa055d8, 0x0) lint_test.go:1308: /usr/local/go/src/runtime/mcentral.go:232 +0x94 fp=0x7ffc59860bb0 sp=0x7ffc59860b68 pc=0x416744 lint_test.go:1308: runtime.(*mcentral).cacheSpan(0xa055d8, 0x7feb8f2c8000) lint_test.go:1308: /usr/local/go/src/runtime/mcentral.go:106 +0x2f8 fp=0x7ffc59860bf8 sp=0x7ffc59860bb0 pc=0x416298 lint_test.go:1308: runtime.(*mcache).refill(0x7feb8f2c8000, 0x2a) lint_test.go:1308: /usr/local/go/src/runtime/mcache.go:122 +0x95 fp=0x7ffc59860c28 sp=0x7ffc59860bf8 pc=0x415e55 lint_test.go:1308: runtime.(*mcache).nextFree.func1() lint_test.go:1308: /usr/local/go/src/runtime/malloc.go:749 +0x32 fp=0x7ffc59860c48 sp=0x7ffc59860c28 pc=0x455392 lint_test.go:1308: runtime.(*mcache).nextFree(0x7feb8f2c8000, 0xa1c02a, 0x4000, 0x7feb8f2c8000, 0x7ffc59860d08) lint_test.go:1308: /usr/local/go/src/runtime/malloc.go:748 +0xb6 fp=0x7ffc59860ca0 sp=0x7ffc59860c48 pc=0x40b2f6 lint_test.go:1308: runtime.mallocgc(0x180, 0x77c180, 0x7ffc59860d01, 0x7feb8f2cc000) lint_test.go:1308: /usr/local/go/src/runtime/malloc.go:903 +0x793 fp=0x7ffc59860d40 sp=0x7ffc59860ca0 pc=0x40bc43 lint_test.go:1308: runtime.newobject(0x77c180, 0xa1c040) lint_test.go:1308: /usr/local/go/src/runtime/malloc.go:1032 +0x38 fp=0x7ffc59860d70 sp=0x7ffc59860d40 pc=0x40c028 lint_test.go:1308: runtime.malg(0x7feb00008000, 0x7feb8f2c8000) lint_test.go:1308: /usr/local/go/src/runtime/proc.go:3288 +0x31 fp=0x7ffc59860db0 sp=0x7ffc59860d70 pc=0x434621 lint_test.go:1308: runtime.mpreinit(0x9fe300) lint_test.go:1308: /usr/local/go/src/runtime/os_linux.go:311 +0x29 fp=0x7ffc59860dd0 sp=0x7ffc59860db0 pc=0x429529 lint_test.go:1308: runtime.mcommoninit(0x9fe300) lint_test.go:1308: /usr/local/go/src/runtime/proc.go:624 +0xc1 fp=0x7ffc59860e08 sp=0x7ffc59860dd0 pc=0x42df41 lint_test.go:1308: runtime.schedinit() lint_test.go:1308: /usr/local/go/src/runtime/proc.go:546 +0x89 fp=0x7ffc59860e70 sp=0x7ffc59860e08 pc=0x42dc09 lint_test.go:1308: runtime.rt0_go(0x7ffc59860ea8, 0x2, 0x7ffc59860ea8, 0x0, 0x0, 0x2, 0x7ffc59862ae9, 0x7ffc59862b1d, 0x0, 0x7ffc59862b41, ...) lint_test.go:1308: /usr/local/go/src/runtime/asm_amd64.s:195 +0x11a fp=0x7ffc59860e78 sp=0x7ffc59860e70 pc=0x457aca TestLint/TestVet --- FAIL: lint/TestLint: TestLint/TestVet (904.290s) ------- Stdout: ------- === PAUSE TestLint/TestVet ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: TestLint - The following tests appear to have failed on master (lint): TestLint/TestVet, TestLint/TestVet: TestLint/TestVet/shadow, TestLint You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestLint). [#1251247](https://teamcity.cockroachdb.com/viewLog.html?buildId=1251247): ``` TestLint --- FAIL: lint/TestLint (237.270s) TestLint/TestVet: TestLint/TestVet/shadow ...0 +0x42 fp=0x7ffc59860a60 sp=0x7ffc59860a08 pc=0x423092 lint_test.go:1308: runtime.(*mheap).allocSpanLocked(0xa03820, 0x1, 0xa1bfe8, 0x0) lint_test.go:1308: /usr/local/go/src/runtime/mheap.go:848 +0x337 fp=0x7ffc59860aa0 sp=0x7ffc59860a60 pc=0x422f17 lint_test.go:1308: runtime.(*mheap).alloc_m(0xa03820, 0x1, 0x2a, 0x0) lint_test.go:1308: /usr/local/go/src/runtime/mheap.go:692 +0x119 fp=0x7ffc59860ae0 sp=0x7ffc59860aa0 pc=0x422729 lint_test.go:1308: runtime.(*mheap).alloc.func1() lint_test.go:1308: /usr/local/go/src/runtime/mheap.go:759 +0x4c fp=0x7ffc59860b18 sp=0x7ffc59860ae0 pc=0x455f7c lint_test.go:1308: runtime.(*mheap).alloc(0xa03820, 0x1, 0x1002a, 0x0) lint_test.go:1308: /usr/local/go/src/runtime/mheap.go:758 +0x8a fp=0x7ffc59860b68 sp=0x7ffc59860b18 pc=0x4229ca lint_test.go:1308: runtime.(*mcentral).grow(0xa055d8, 0x0) lint_test.go:1308: /usr/local/go/src/runtime/mcentral.go:232 +0x94 fp=0x7ffc59860bb0 sp=0x7ffc59860b68 pc=0x416744 lint_test.go:1308: runtime.(*mcentral).cacheSpan(0xa055d8, 0x7feb8f2c8000) lint_test.go:1308: /usr/local/go/src/runtime/mcentral.go:106 +0x2f8 fp=0x7ffc59860bf8 sp=0x7ffc59860bb0 pc=0x416298 lint_test.go:1308: runtime.(*mcache).refill(0x7feb8f2c8000, 0x2a) lint_test.go:1308: /usr/local/go/src/runtime/mcache.go:122 +0x95 fp=0x7ffc59860c28 sp=0x7ffc59860bf8 pc=0x415e55 lint_test.go:1308: runtime.(*mcache).nextFree.func1() lint_test.go:1308: /usr/local/go/src/runtime/malloc.go:749 +0x32 fp=0x7ffc59860c48 sp=0x7ffc59860c28 pc=0x455392 lint_test.go:1308: runtime.(*mcache).nextFree(0x7feb8f2c8000, 0xa1c02a, 0x4000, 0x7feb8f2c8000, 0x7ffc59860d08) lint_test.go:1308: /usr/local/go/src/runtime/malloc.go:748 +0xb6 fp=0x7ffc59860ca0 sp=0x7ffc59860c48 pc=0x40b2f6 lint_test.go:1308: runtime.mallocgc(0x180, 0x77c180, 0x7ffc59860d01, 0x7feb8f2cc000) lint_test.go:1308: /usr/local/go/src/runtime/malloc.go:903 +0x793 fp=0x7ffc59860d40 sp=0x7ffc59860ca0 pc=0x40bc43 lint_test.go:1308: runtime.newobject(0x77c180, 0xa1c040) lint_test.go:1308: /usr/local/go/src/runtime/malloc.go:1032 +0x38 fp=0x7ffc59860d70 sp=0x7ffc59860d40 pc=0x40c028 lint_test.go:1308: runtime.malg(0x7feb00008000, 0x7feb8f2c8000) lint_test.go:1308: /usr/local/go/src/runtime/proc.go:3288 +0x31 fp=0x7ffc59860db0 sp=0x7ffc59860d70 pc=0x434621 lint_test.go:1308: runtime.mpreinit(0x9fe300) lint_test.go:1308: /usr/local/go/src/runtime/os_linux.go:311 +0x29 fp=0x7ffc59860dd0 sp=0x7ffc59860db0 pc=0x429529 lint_test.go:1308: runtime.mcommoninit(0x9fe300) lint_test.go:1308: /usr/local/go/src/runtime/proc.go:624 +0xc1 fp=0x7ffc59860e08 sp=0x7ffc59860dd0 pc=0x42df41 lint_test.go:1308: runtime.schedinit() lint_test.go:1308: /usr/local/go/src/runtime/proc.go:546 +0x89 fp=0x7ffc59860e70 sp=0x7ffc59860e08 pc=0x42dc09 lint_test.go:1308: runtime.rt0_go(0x7ffc59860ea8, 0x2, 0x7ffc59860ea8, 0x0, 0x0, 0x2, 0x7ffc59862ae9, 0x7ffc59862b1d, 0x0, 0x7ffc59862b41, ...) lint_test.go:1308: /usr/local/go/src/runtime/asm_amd64.s:195 +0x11a fp=0x7ffc59860e78 sp=0x7ffc59860e70 pc=0x457aca TestLint/TestVet --- FAIL: lint/TestLint: TestLint/TestVet (904.290s) ------- Stdout: ------- === PAUSE TestLint/TestVet ``` Please assign, take a look and update the issue accordingly.
non_main
teamcity failed test testlint the following tests appear to have failed on master lint testlint testvet testlint testvet testlint testvet shadow testlint you may want to check testlint fail lint testlint testlint testvet testlint testvet shadow fp sp pc lint test go runtime mheap allocspanlocked lint test go usr local go src runtime mheap go fp sp pc lint test go runtime mheap alloc m lint test go usr local go src runtime mheap go fp sp pc lint test go runtime mheap alloc lint test go usr local go src runtime mheap go fp sp pc lint test go runtime mheap alloc lint test go usr local go src runtime mheap go fp sp pc lint test go runtime mcentral grow lint test go usr local go src runtime mcentral go fp sp pc lint test go runtime mcentral cachespan lint test go usr local go src runtime mcentral go fp sp pc lint test go runtime mcache refill lint test go usr local go src runtime mcache go fp sp pc lint test go runtime mcache nextfree lint test go usr local go src runtime malloc go fp sp pc lint test go runtime mcache nextfree lint test go usr local go src runtime malloc go fp sp pc lint test go runtime mallocgc lint test go usr local go src runtime malloc go fp sp pc lint test go runtime newobject lint test go usr local go src runtime malloc go fp sp pc lint test go runtime malg lint test go usr local go src runtime proc go fp sp pc lint test go runtime mpreinit lint test go usr local go src runtime os linux go fp sp pc lint test go runtime mcommoninit lint test go usr local go src runtime proc go fp sp pc lint test go runtime schedinit lint test go usr local go src runtime proc go fp sp pc lint test go runtime go lint test go usr local go src runtime asm s fp sp pc testlint testvet fail lint testlint testlint testvet stdout pause testlint testvet please assign take a look and update the issue accordingly
0
2,852
10,237,432,593
IssuesEvent
2019-08-19 13:53:52
zaproxy/zaproxy
https://api.github.com/repos/zaproxy/zaproxy
opened
forced browse: Use logger instead of standard out
IdealFirstBug Maintainability Usability add-on good first issue
the `bruteforce` extension could be further migrated from use of `System.out.println` to a standard ZAP logger. Then it would benefit from timestamps, etc which would potentially make troubleshooting easier.
True
forced browse: Use logger instead of standard out - the `bruteforce` extension could be further migrated from use of `System.out.println` to a standard ZAP logger. Then it would benefit from timestamps, etc which would potentially make troubleshooting easier.
main
forced browse use logger instead of standard out the bruteforce extension could be further migrated from use of system out println to a standard zap logger then it would benefit from timestamps etc which would potentially make troubleshooting easier
1
5,643
28,369,876,415
IssuesEvent
2023-04-12 16:10:41
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
opened
Allow listening for updates in BrokerTopologyListener
kind/toil area/reliability area/maintainability component/gateway
**Description** In order for the job push' `ClientStreamService` to detect brokers being added/removed, we need a way to get this information via the topology. We could use a plain membership listener, but that one is unaware of whether a node is a broker or a gateway, whereas the `BrokerTopologyManager` can already provide this information. We should add the following capabilities: - [ ] Add a new listener which gets notified when a broker is added or removed - [ ] When the listener is initially added, it is also initialized the current list of known brokers, to avoid race conditions - [ ] The listener can be removed via its identity
True
Allow listening for updates in BrokerTopologyListener - **Description** In order for the job push' `ClientStreamService` to detect brokers being added/removed, we need a way to get this information via the topology. We could use a plain membership listener, but that one is unaware of whether a node is a broker or a gateway, whereas the `BrokerTopologyManager` can already provide this information. We should add the following capabilities: - [ ] Add a new listener which gets notified when a broker is added or removed - [ ] When the listener is initially added, it is also initialized the current list of known brokers, to avoid race conditions - [ ] The listener can be removed via its identity
main
allow listening for updates in brokertopologylistener description in order for the job push clientstreamservice to detect brokers being added removed we need a way to get this information via the topology we could use a plain membership listener but that one is unaware of whether a node is a broker or a gateway whereas the brokertopologymanager can already provide this information we should add the following capabilities add a new listener which gets notified when a broker is added or removed when the listener is initially added it is also initialized the current list of known brokers to avoid race conditions the listener can be removed via its identity
1
25,471
11,173,315,838
IssuesEvent
2019-12-29 13:23:36
openSUSE/libpathrs
https://api.github.com/repos/openSUSE/libpathrs
opened
readlinkat2(AT_EMPTY_PATH)
security upstream/linux
Right now it's not possible to `readlink` an `O_PATH | O_NOFOLLOW` handle to a symlink. While this isn't strictly required for security (after all, `readlinkat` will give you an error if the target is not a symlink and we don't need special handling for magic-links) this would be a huge nice-to-have.
True
readlinkat2(AT_EMPTY_PATH) - Right now it's not possible to `readlink` an `O_PATH | O_NOFOLLOW` handle to a symlink. While this isn't strictly required for security (after all, `readlinkat` will give you an error if the target is not a symlink and we don't need special handling for magic-links) this would be a huge nice-to-have.
non_main
at empty path right now it s not possible to readlink an o path o nofollow handle to a symlink while this isn t strictly required for security after all readlinkat will give you an error if the target is not a symlink and we don t need special handling for magic links this would be a huge nice to have
0
2,040
6,887,616,738
IssuesEvent
2017-11-22 00:28:10
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
ec2_vpc_route_table not updating?
affects_2.3 aws bug_report cloud needs_info needs_template support:core waiting_on_maintainer
_From @mattgrahamdevops on February 15, 2016 23:21_ Not sure if Im doing something wrong, but I can add the routes using ec2_vpc_route_table module but if I terminate the NAT instances that are in the routes (instance_id), you see the "black hole" in AWS GUI but for some reason when I go to run my playbook again, it creates new NAT's, gets the instance id's and then attempts to apply them to the route table but fails. If I manually go in and delete the "black hole" routes and run the playbook, its fine. Version: ``` ansible 2.0.0.2 ``` playbook code: ``` name: App Private Route Table ec2_vpc_route_table: vpc_id: "{{ vpc_id }}" region: "{{ aws_region }}" subnets: "{{ item.subnet }}" tags: Name: "{{ env | default('test') }}app_private{{ item.az }}" routes: - dest: 0.0.0.0/0 instance_id: "{{ item.instance }}" with_items: - { Name: app_a, subnet: "{{ cidr }}.5.0/24", instance: "{{ nat_servers.results[0].tagged_instances[0].id }}", az: a } - { Name: app_b, subnet: "{{ cidr }}.6.0/24", instance: "{{ nat_servers.results[1].tagged_instances[0].id }}", az: b } ``` Error: ``` An exception occurred during task execution. To see the full traceback, use -vvv. The error was: TypeError: argument of type 'NoneType' is not iterable failed: [localhost] => (item={u'subnet': u'10.40.5.0/24', u'az': u'a', u'Name': u'app_a', u'instance': u'i-fddcfa22'}) => {"failed": true, "item": {"Name": "app_a", "az": "a", "instance": "i-fddcfa22", "subnet": "10.40.5.0/24"}, "parsed": false} An exception occurred during task execution. To see the full traceback, use -vvv. The error was: TypeError: argument of type 'NoneType' is not iterable failed: [localhost] => (item={u'subnet': u'10.40.6.0/24', u'az': u'b', u'Name': u'app_b', u'instance': u'i-d6b1b908'}) => {"failed": true, "item": {"Name": "app_b", "az": "b", "instance": "i-d6b1b908", "subnet": "10.40.6.0/24"}, "parsed": false} ``` Current route table looks like: ``` 0.0.0.0/0 eni-af3902c8 / i-3a0b2de5 Black Hole No ``` _Copied from original issue: ansible/ansible-modules-extras#1674_
True
ec2_vpc_route_table not updating? - _From @mattgrahamdevops on February 15, 2016 23:21_ Not sure if Im doing something wrong, but I can add the routes using ec2_vpc_route_table module but if I terminate the NAT instances that are in the routes (instance_id), you see the "black hole" in AWS GUI but for some reason when I go to run my playbook again, it creates new NAT's, gets the instance id's and then attempts to apply them to the route table but fails. If I manually go in and delete the "black hole" routes and run the playbook, its fine. Version: ``` ansible 2.0.0.2 ``` playbook code: ``` name: App Private Route Table ec2_vpc_route_table: vpc_id: "{{ vpc_id }}" region: "{{ aws_region }}" subnets: "{{ item.subnet }}" tags: Name: "{{ env | default('test') }}app_private{{ item.az }}" routes: - dest: 0.0.0.0/0 instance_id: "{{ item.instance }}" with_items: - { Name: app_a, subnet: "{{ cidr }}.5.0/24", instance: "{{ nat_servers.results[0].tagged_instances[0].id }}", az: a } - { Name: app_b, subnet: "{{ cidr }}.6.0/24", instance: "{{ nat_servers.results[1].tagged_instances[0].id }}", az: b } ``` Error: ``` An exception occurred during task execution. To see the full traceback, use -vvv. The error was: TypeError: argument of type 'NoneType' is not iterable failed: [localhost] => (item={u'subnet': u'10.40.5.0/24', u'az': u'a', u'Name': u'app_a', u'instance': u'i-fddcfa22'}) => {"failed": true, "item": {"Name": "app_a", "az": "a", "instance": "i-fddcfa22", "subnet": "10.40.5.0/24"}, "parsed": false} An exception occurred during task execution. To see the full traceback, use -vvv. The error was: TypeError: argument of type 'NoneType' is not iterable failed: [localhost] => (item={u'subnet': u'10.40.6.0/24', u'az': u'b', u'Name': u'app_b', u'instance': u'i-d6b1b908'}) => {"failed": true, "item": {"Name": "app_b", "az": "b", "instance": "i-d6b1b908", "subnet": "10.40.6.0/24"}, "parsed": false} ``` Current route table looks like: ``` 0.0.0.0/0 eni-af3902c8 / i-3a0b2de5 Black Hole No ``` _Copied from original issue: ansible/ansible-modules-extras#1674_
main
vpc route table not updating from mattgrahamdevops on february not sure if im doing something wrong but i can add the routes using vpc route table module but if i terminate the nat instances that are in the routes instance id you see the black hole in aws gui but for some reason when i go to run my playbook again it creates new nat s gets the instance id s and then attempts to apply them to the route table but fails if i manually go in and delete the black hole routes and run the playbook its fine version ansible playbook code name app private route table vpc route table vpc id vpc id region aws region subnets item subnet tags name env default test app private item az routes dest instance id item instance with items name app a subnet cidr instance nat servers results tagged instances id az a name app b subnet cidr instance nat servers results tagged instances id az b error an exception occurred during task execution to see the full traceback use vvv the error was typeerror argument of type nonetype is not iterable failed item u subnet u u az u a u name u app a u instance u i failed true item name app a az a instance i subnet parsed false an exception occurred during task execution to see the full traceback use vvv the error was typeerror argument of type nonetype is not iterable failed item u subnet u u az u b u name u app b u instance u i failed true item name app b az b instance i subnet parsed false current route table looks like eni i black hole no copied from original issue ansible ansible modules extras
1
127,935
10,500,676,576
IssuesEvent
2019-09-26 11:03:52
Omeljusik/teenjob_development
https://api.github.com/repos/Omeljusik/teenjob_development
closed
[Admin_Org] При незаполнении любого из пунктов кроме описания, при попытке "опубликовать" сайт ломается. Одинаково для Волонтерство, Стажировки, Мероприятия.
critical ready to test
Запись экрана: https://drive.google.com/file/d/1D5lv_25HBIEiyR_GZ5hx7Kk95W2uj2SI/view?usp=sharing ## Как должно быть - [ ] при незаполнении необходимого для заполнения поля высвечивается предупреждение "Заполните это поле" - [ ] из стажировок/волонтерства убрать * с "Телефон*" ![image](https://user-images.githubusercontent.com/32821205/64859548-386f3880-d633-11e9-9d37-ce03bed90e21.png) - [ ] дбавить * к названию, городу, возрасту, профобласти, описанию - [ ] добавить обозначение "Описание" перед редакционной панелью ![image](https://user-images.githubusercontent.com/32821205/64861115-ccdb9a00-d637-11e9-9e48-2662119db698.png) ### Сделать редакционную панель в 2 раза длиннее ![image](https://user-images.githubusercontent.com/32821205/64861044-a4ec3680-d637-11e9-9ab7-5e717fb90c55.png) ### Вместо заголовка "Новое объявление" при создании стажировки "Стажировка", при создании волонтерства "Волонтерство", при создании мероприятия "Мероприятие" ![image](https://user-images.githubusercontent.com/32821205/64861018-900fa300-d637-11e9-978d-580d2f20b0df.png) ### Все макеты (исправленные): Волонтерство/Стажировки https://www.figma.com/file/OD9PcqDHdfLJHDLJi34l3wAH/teenjob?node-id=1201%3A3 Мероприятия https://www.figma.com/file/OD9PcqDHdfLJHDLJi34l3wAH/teenjob?node-id=1205%3A5
1.0
[Admin_Org] При незаполнении любого из пунктов кроме описания, при попытке "опубликовать" сайт ломается. Одинаково для Волонтерство, Стажировки, Мероприятия. - Запись экрана: https://drive.google.com/file/d/1D5lv_25HBIEiyR_GZ5hx7Kk95W2uj2SI/view?usp=sharing ## Как должно быть - [ ] при незаполнении необходимого для заполнения поля высвечивается предупреждение "Заполните это поле" - [ ] из стажировок/волонтерства убрать * с "Телефон*" ![image](https://user-images.githubusercontent.com/32821205/64859548-386f3880-d633-11e9-9d37-ce03bed90e21.png) - [ ] дбавить * к названию, городу, возрасту, профобласти, описанию - [ ] добавить обозначение "Описание" перед редакционной панелью ![image](https://user-images.githubusercontent.com/32821205/64861115-ccdb9a00-d637-11e9-9e48-2662119db698.png) ### Сделать редакционную панель в 2 раза длиннее ![image](https://user-images.githubusercontent.com/32821205/64861044-a4ec3680-d637-11e9-9ab7-5e717fb90c55.png) ### Вместо заголовка "Новое объявление" при создании стажировки "Стажировка", при создании волонтерства "Волонтерство", при создании мероприятия "Мероприятие" ![image](https://user-images.githubusercontent.com/32821205/64861018-900fa300-d637-11e9-978d-580d2f20b0df.png) ### Все макеты (исправленные): Волонтерство/Стажировки https://www.figma.com/file/OD9PcqDHdfLJHDLJi34l3wAH/teenjob?node-id=1201%3A3 Мероприятия https://www.figma.com/file/OD9PcqDHdfLJHDLJi34l3wAH/teenjob?node-id=1205%3A5
non_main
при незаполнении любого из пунктов кроме описания при попытке опубликовать сайт ломается одинаково для волонтерство стажировки мероприятия запись экрана как должно быть при незаполнении необходимого для заполнения поля высвечивается предупреждение заполните это поле из стажировок волонтерства убрать с телефон дбавить к названию городу возрасту профобласти описанию добавить обозначение описание перед редакционной панелью сделать редакционную панель в раза длиннее вместо заголовка новое объявление при создании стажировки стажировка при создании волонтерства волонтерство при создании мероприятия мероприятие все макеты исправленные волонтерство стажировки мероприятия
0
4,529
23,542,526,150
IssuesEvent
2022-08-20 16:18:16
spack/spack
https://api.github.com/repos/spack/spack
opened
Unit tests in `bindist.py` depend on the order of execution
bug maintainers
Due to a fixture that is shared among tests: https://github.com/spack/spack/blob/fff929d5ab08fc21a3fbe795c10e255f993aa39e/lib/spack/spack/test/bindist.py#L49-L54 the unit tests are building a build cache, and later tests depend on the results of previous tests. If the scope of the fixture is changed to `function` tests start failing
True
Unit tests in `bindist.py` depend on the order of execution - Due to a fixture that is shared among tests: https://github.com/spack/spack/blob/fff929d5ab08fc21a3fbe795c10e255f993aa39e/lib/spack/spack/test/bindist.py#L49-L54 the unit tests are building a build cache, and later tests depend on the results of previous tests. If the scope of the fixture is changed to `function` tests start failing
main
unit tests in bindist py depend on the order of execution due to a fixture that is shared among tests the unit tests are building a build cache and later tests depend on the results of previous tests if the scope of the fixture is changed to function tests start failing
1
43,793
11,305,844,941
IssuesEvent
2020-01-18 09:19:28
pdulth/tmp2
https://api.github.com/repos/pdulth/tmp2
opened
[32] The Kitalpha zip name must contain the word "incubation"
Build Kitalpha enhancement verified
Because Kitalpha is in incubation, the word "incubation" in the zip file names for release. Ex: kitalpha_[eclipse target]_[release number]-[date]-incubation => kitalpha_juno_0.5.0-20140822_085151-incubation Benoit `Eclipse-555993` `Polarsys-32` `@benoit.langlois` `2014-08-28`
1.0
[32] The Kitalpha zip name must contain the word "incubation" - Because Kitalpha is in incubation, the word "incubation" in the zip file names for release. Ex: kitalpha_[eclipse target]_[release number]-[date]-incubation => kitalpha_juno_0.5.0-20140822_085151-incubation Benoit `Eclipse-555993` `Polarsys-32` `@benoit.langlois` `2014-08-28`
non_main
the kitalpha zip name must contain the word incubation because kitalpha is in incubation the word incubation in the zip file names for release ex kitalpha incubation kitalpha juno incubation benoit eclipse polarsys benoit langlois
0
719
4,309,544,991
IssuesEvent
2016-07-21 16:20:04
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
Minecraft: should support alternate names for items.
Maintainer Approved Suggestion Triggering
Item names are too hard to guess. They should have alternate names or should be stored with the most common term. 'shovel' for example, isn't in the item list, but rather stored in the plural form as 'shovels'. Some people may say "how to craft shovels" whereas others might prefer "how to craft a shovel". I doubt anyone will guess "Jack-O-Lantern". @engvik ------ IA Page: http://duck.co/ia/view/minecraft
True
Minecraft: should support alternate names for items. - Item names are too hard to guess. They should have alternate names or should be stored with the most common term. 'shovel' for example, isn't in the item list, but rather stored in the plural form as 'shovels'. Some people may say "how to craft shovels" whereas others might prefer "how to craft a shovel". I doubt anyone will guess "Jack-O-Lantern". @engvik ------ IA Page: http://duck.co/ia/view/minecraft
main
minecraft should support alternate names for items item names are too hard to guess they should have alternate names or should be stored with the most common term shovel for example isn t in the item list but rather stored in the plural form as shovels some people may say how to craft shovels whereas others might prefer how to craft a shovel i doubt anyone will guess jack o lantern engvik ia page
1
1,929
6,607,591,449
IssuesEvent
2017-09-19 07:45:17
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
remote script directory is not removed from `sys.path` when `remote_tmp` is a link
affects_2.3 bug_report module needs_maintainer needs_triage support:community
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME * tempfile * remote files handling in general ##### ANSIBLE VERSION ``` ansible 2.3.0.0 config file = /tmp/original/ansible.cfg configured module search path = Default w/o overrides python version = 2.7.12 (default, Nov 19 2016, 06:48:10) [GCC 5.4.0 20160609] ``` ##### CONFIGURATION `ansible.cfg`: ``` [defaults] remote_tmp: /tmp/link ``` ##### OS / ENVIRONMENT ``` Linux ic5-16 4.4.0-93-generic #116-Ubuntu SMP Fri Aug 11 21:17:51 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux ``` ``` Linux Mint 18.1 Serena \n \l ``` ##### SUMMARY When `remote_tmp` is a link it is not removed from `sys.path` in uploaded module. In case of `tempfile`, it clashes with original python module which it uses. ##### STEPS TO REPRODUCE ``` mkdir -p /tmp/original /tmp/ansible ln -s /tmp/original /tmp/link echo "[defaults]\nremote_tmp: /tmp/link" > /tmp/ansible/ansible.cfg cd /tmp/ansible && ansible -m tempfile localhost -vvv ``` ##### EXPECTED RESULTS Module `tempfile` should succeed. ##### ACTUAL RESULTS ``` [WARNING]: provided hosts list is empty, only localhost is available Loading callback plugin minimal of type stdout, v2.0 from /usr/lib/python2.7/dist-packages/ansible/plugins/callback/__init__.pyc META: ran handlers Using module file /usr/lib/python2.7/dist-packages/ansible/modules/files/tempfile.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: kostrzewa <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /tmp/link/ansible-tmp-1505799554.81-197426653923554 `" && echo ansible-tmp-1505799554.81-197426653923554="` echo /tmp/link/ansible-tmp-1505799554.81-197426653923554 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpXX2me4 TO /tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /tmp/link/ansible-tmp-1505799554.81-197426653923554/ /tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py; rm -rf "/tmp/link/ansible-tmp-1505799554.81-197426653923554/" > /dev/null 2>&1 && sleep 0' The full traceback is: Traceback (most recent call last): File "/tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py", line 136, in <module> shutil.rmtree(temp_path) NameError: name 'temp_path' is not defined localhost | FAILED! => { "changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py\", line 136, in <module>\n shutil.rmtree(temp_path)\nNameError: name 'temp_path' is not defined\n", "module_stdout": "", "msg": "MODULE FAILURE", "rc": 0 } ``` ##### EXTRA INFO This happens because: * remote module is invoked through path with a link * in order to remove script's directory from `sys.path` module handling code resolves the path to script directory with `os.path.abspath()` * in `sys.path` the script directory is registered with real path thus script directory is not removed from `sys.path` and because ansible module has the same name as python module when it calls `import tempfile` it ends up importing itself. When running ansible with `ANSIBLE_KEEP_REMOTE_FILES=1` and invoking the module directly, with some debug prints, I get: ``` /tmp/link$ /tmp/link/ansible-tmp-1505798980.71-111952868906477/tempfile.py scriptdir: /tmp/link/ansible-tmp-1505798980.71-111952868906477 sys.path: ['/tmp/original/ansible-tmp-1505798980.71-111952868906477', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/usr/lib/python2.7/lib-old', '/usr/lib/python2.7/lib-dynload', '/usr/local/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages/PILcompat', '/usr/lib/python2.7/dist-packages/gst-0.10', '/usr/lib/python2.7/dist-packages/gtk-2.0', '/usr/lib/pymodules/python2.7'] scriptdir: /tmp/link/ansible-tmp-1505798980.71-111952868906477 sys.path: ['/tmp/original/ansible-tmp-1505798980.71-111952868906477', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/usr/lib/python2.7/lib-old', '/usr/lib/python2.7/lib-dynload', '/usr/local/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages/PILcompat', '/usr/lib/python2.7/dist-packages/gst-0.10', '/usr/lib/python2.7/dist-packages/gtk-2.0', '/usr/lib/pymodules/python2.7'] Traceback (most recent call last): File "/tmp/link/ansible-tmp-1505798980.71-111952868906477/tempfile.py", line 281, in <module> shutil.rmtree(temp_path) NameError: name 'temp_path' is not defined ``` After changing: ``` 56 scriptdir = os.path.dirname(os.path.abspath(__main__.__file__)) ``` to: ``` 56 scriptdir = os.path.dirname(os.path.realpath(__main__.__file__)) ``` the module starts to work.
True
remote script directory is not removed from `sys.path` when `remote_tmp` is a link - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME * tempfile * remote files handling in general ##### ANSIBLE VERSION ``` ansible 2.3.0.0 config file = /tmp/original/ansible.cfg configured module search path = Default w/o overrides python version = 2.7.12 (default, Nov 19 2016, 06:48:10) [GCC 5.4.0 20160609] ``` ##### CONFIGURATION `ansible.cfg`: ``` [defaults] remote_tmp: /tmp/link ``` ##### OS / ENVIRONMENT ``` Linux ic5-16 4.4.0-93-generic #116-Ubuntu SMP Fri Aug 11 21:17:51 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux ``` ``` Linux Mint 18.1 Serena \n \l ``` ##### SUMMARY When `remote_tmp` is a link it is not removed from `sys.path` in uploaded module. In case of `tempfile`, it clashes with original python module which it uses. ##### STEPS TO REPRODUCE ``` mkdir -p /tmp/original /tmp/ansible ln -s /tmp/original /tmp/link echo "[defaults]\nremote_tmp: /tmp/link" > /tmp/ansible/ansible.cfg cd /tmp/ansible && ansible -m tempfile localhost -vvv ``` ##### EXPECTED RESULTS Module `tempfile` should succeed. ##### ACTUAL RESULTS ``` [WARNING]: provided hosts list is empty, only localhost is available Loading callback plugin minimal of type stdout, v2.0 from /usr/lib/python2.7/dist-packages/ansible/plugins/callback/__init__.pyc META: ran handlers Using module file /usr/lib/python2.7/dist-packages/ansible/modules/files/tempfile.py <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: kostrzewa <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo /tmp/link/ansible-tmp-1505799554.81-197426653923554 `" && echo ansible-tmp-1505799554.81-197426653923554="` echo /tmp/link/ansible-tmp-1505799554.81-197426653923554 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpXX2me4 TO /tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py <127.0.0.1> EXEC /bin/sh -c 'chmod u+x /tmp/link/ansible-tmp-1505799554.81-197426653923554/ /tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py && sleep 0' <127.0.0.1> EXEC /bin/sh -c '/usr/bin/python /tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py; rm -rf "/tmp/link/ansible-tmp-1505799554.81-197426653923554/" > /dev/null 2>&1 && sleep 0' The full traceback is: Traceback (most recent call last): File "/tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py", line 136, in <module> shutil.rmtree(temp_path) NameError: name 'temp_path' is not defined localhost | FAILED! => { "changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/link/ansible-tmp-1505799554.81-197426653923554/tempfile.py\", line 136, in <module>\n shutil.rmtree(temp_path)\nNameError: name 'temp_path' is not defined\n", "module_stdout": "", "msg": "MODULE FAILURE", "rc": 0 } ``` ##### EXTRA INFO This happens because: * remote module is invoked through path with a link * in order to remove script's directory from `sys.path` module handling code resolves the path to script directory with `os.path.abspath()` * in `sys.path` the script directory is registered with real path thus script directory is not removed from `sys.path` and because ansible module has the same name as python module when it calls `import tempfile` it ends up importing itself. When running ansible with `ANSIBLE_KEEP_REMOTE_FILES=1` and invoking the module directly, with some debug prints, I get: ``` /tmp/link$ /tmp/link/ansible-tmp-1505798980.71-111952868906477/tempfile.py scriptdir: /tmp/link/ansible-tmp-1505798980.71-111952868906477 sys.path: ['/tmp/original/ansible-tmp-1505798980.71-111952868906477', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/usr/lib/python2.7/lib-old', '/usr/lib/python2.7/lib-dynload', '/usr/local/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages/PILcompat', '/usr/lib/python2.7/dist-packages/gst-0.10', '/usr/lib/python2.7/dist-packages/gtk-2.0', '/usr/lib/pymodules/python2.7'] scriptdir: /tmp/link/ansible-tmp-1505798980.71-111952868906477 sys.path: ['/tmp/original/ansible-tmp-1505798980.71-111952868906477', '/usr/lib/python2.7', '/usr/lib/python2.7/plat-x86_64-linux-gnu', '/usr/lib/python2.7/lib-tk', '/usr/lib/python2.7/lib-old', '/usr/lib/python2.7/lib-dynload', '/usr/local/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages', '/usr/lib/python2.7/dist-packages/PILcompat', '/usr/lib/python2.7/dist-packages/gst-0.10', '/usr/lib/python2.7/dist-packages/gtk-2.0', '/usr/lib/pymodules/python2.7'] Traceback (most recent call last): File "/tmp/link/ansible-tmp-1505798980.71-111952868906477/tempfile.py", line 281, in <module> shutil.rmtree(temp_path) NameError: name 'temp_path' is not defined ``` After changing: ``` 56 scriptdir = os.path.dirname(os.path.abspath(__main__.__file__)) ``` to: ``` 56 scriptdir = os.path.dirname(os.path.realpath(__main__.__file__)) ``` the module starts to work.
main
remote script directory is not removed from sys path when remote tmp is a link issue type bug report component name tempfile remote files handling in general ansible version ansible config file tmp original ansible cfg configured module search path default w o overrides python version default nov configuration ansible cfg remote tmp tmp link os environment linux generic ubuntu smp fri aug utc gnu linux linux mint serena n l summary when remote tmp is a link it is not removed from sys path in uploaded module in case of tempfile it clashes with original python module which it uses steps to reproduce mkdir p tmp original tmp ansible ln s tmp original tmp link echo nremote tmp tmp link tmp ansible ansible cfg cd tmp ansible ansible m tempfile localhost vvv expected results module tempfile should succeed actual results provided hosts list is empty only localhost is available loading callback plugin minimal of type stdout from usr lib dist packages ansible plugins callback init pyc meta ran handlers using module file usr lib dist packages ansible modules files tempfile py establish local connection for user kostrzewa exec bin sh c umask mkdir p echo tmp link ansible tmp echo ansible tmp echo tmp link ansible tmp sleep put tmp to tmp link ansible tmp tempfile py exec bin sh c chmod u x tmp link ansible tmp tmp link ansible tmp tempfile py sleep exec bin sh c usr bin python tmp link ansible tmp tempfile py rm rf tmp link ansible tmp dev null sleep the full traceback is traceback most recent call last file tmp link ansible tmp tempfile py line in shutil rmtree temp path nameerror name temp path is not defined localhost failed changed false failed true module stderr traceback most recent call last n file tmp link ansible tmp tempfile py line in n shutil rmtree temp path nnameerror name temp path is not defined n module stdout msg module failure rc extra info this happens because remote module is invoked through path with a link in order to remove script s directory from sys path module handling code resolves the path to script directory with os path abspath in sys path the script directory is registered with real path thus script directory is not removed from sys path and because ansible module has the same name as python module when it calls import tempfile it ends up importing itself when running ansible with ansible keep remote files and invoking the module directly with some debug prints i get tmp link tmp link ansible tmp tempfile py scriptdir tmp link ansible tmp sys path scriptdir tmp link ansible tmp sys path traceback most recent call last file tmp link ansible tmp tempfile py line in shutil rmtree temp path nameerror name temp path is not defined after changing scriptdir os path dirname os path abspath main file to scriptdir os path dirname os path realpath main file the module starts to work
1
221,290
17,010,902,794
IssuesEvent
2021-07-02 04:18:55
devxas/airta-home
https://api.github.com/repos/devxas/airta-home
opened
design for seperate sitemap engine as microservice
documentation enhancement
extract current sitemap logic from airta-engine, to handle a refactored sitemap logic and function.
1.0
design for seperate sitemap engine as microservice - extract current sitemap logic from airta-engine, to handle a refactored sitemap logic and function.
non_main
design for seperate sitemap engine as microservice extract current sitemap logic from airta engine to handle a refactored sitemap logic and function
0
536,327
15,707,591,943
IssuesEvent
2021-03-26 19:07:33
itslupus/gamersnet
https://api.github.com/repos/itslupus/gamersnet
closed
Game based search
high priority user story
**Description**: As a user, I want to be able to search for posts to find players that are playing my game **Acceptance Criteria**: Search posts by name of games **Dev Tasks**: [Backend endpoint to fetch posts](https://github.com/itslupus/gamersnet/issues/15) [Frontend UI to select games](https://github.com/itslupus/gamersnet/issues/16) [Frontend UI to display posts by games](https://github.com/itslupus/gamersnet/issues/17) **Story Points (1 - 5**): 3
1.0
Game based search - **Description**: As a user, I want to be able to search for posts to find players that are playing my game **Acceptance Criteria**: Search posts by name of games **Dev Tasks**: [Backend endpoint to fetch posts](https://github.com/itslupus/gamersnet/issues/15) [Frontend UI to select games](https://github.com/itslupus/gamersnet/issues/16) [Frontend UI to display posts by games](https://github.com/itslupus/gamersnet/issues/17) **Story Points (1 - 5**): 3
non_main
game based search description as a user i want to be able to search for posts to find players that are playing my game acceptance criteria search posts by name of games dev tasks story points
0
4,388
22,339,500,172
IssuesEvent
2022-06-14 22:21:46
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Is there a suggested practice for accessing SecretsManager from a "sam local start-api" context?
maintainer/need-response stage/needs-triage
In my lambda I'm invoking the SecretsManager very straightforwardly: ```javascript const smClient = new AWS.SecretsManager(); // further down . . . const secretResponse = await smClient .getSecretValue({ SecretId: process.env.MY_SECRET_ARN, }) .promise() ``` This works when deployed. In my template I've made sure the lambda has the correct permissions (several things (like "Events") left out for readability): ```yaml MyFunction: Type: AWS::Serverless::Function Properties: Environment: Variables: MY_SECRET_ARN: !Ref MySecretArn PackageType: Zip Policies: - AWSLambdaBasicExecutionRole - AWSSecretsManagerGetSecretValuePolicy: SecretArn: !Ref MySecretArn ``` It makes sense why I should be getting an error on the call to SecretsManager ("The security token included in the request is expired"), but is there a way that I _ought_ to be setting this up so that I can enjoy running the lambdas locally when I'm developing?
True
Is there a suggested practice for accessing SecretsManager from a "sam local start-api" context? - In my lambda I'm invoking the SecretsManager very straightforwardly: ```javascript const smClient = new AWS.SecretsManager(); // further down . . . const secretResponse = await smClient .getSecretValue({ SecretId: process.env.MY_SECRET_ARN, }) .promise() ``` This works when deployed. In my template I've made sure the lambda has the correct permissions (several things (like "Events") left out for readability): ```yaml MyFunction: Type: AWS::Serverless::Function Properties: Environment: Variables: MY_SECRET_ARN: !Ref MySecretArn PackageType: Zip Policies: - AWSLambdaBasicExecutionRole - AWSSecretsManagerGetSecretValuePolicy: SecretArn: !Ref MySecretArn ``` It makes sense why I should be getting an error on the call to SecretsManager ("The security token included in the request is expired"), but is there a way that I _ought_ to be setting this up so that I can enjoy running the lambdas locally when I'm developing?
main
is there a suggested practice for accessing secretsmanager from a sam local start api context in my lambda i m invoking the secretsmanager very straightforwardly javascript const smclient new aws secretsmanager further down const secretresponse await smclient getsecretvalue secretid process env my secret arn promise this works when deployed in my template i ve made sure the lambda has the correct permissions several things like events left out for readability yaml myfunction type aws serverless function properties environment variables my secret arn ref mysecretarn packagetype zip policies awslambdabasicexecutionrole awssecretsmanagergetsecretvaluepolicy secretarn ref mysecretarn it makes sense why i should be getting an error on the call to secretsmanager the security token included in the request is expired but is there a way that i ought to be setting this up so that i can enjoy running the lambdas locally when i m developing
1
45,885
18,885,059,121
IssuesEvent
2021-11-15 06:31:51
openshift/odo
https://api.github.com/repos/openshift/odo
closed
`odo catalog describe service postgresoperator.v5.0.3/PostgresCluster` is not helpfull
priority/High area/service-operators
For example the output mentions `instances` that should be array of the object, but it doesn't say how the objects should look like ``` instances ([]object) -required- ``` This is the same for every embedded object mentioned in the spec, there are no detail about their structure. /priority high /area service-operators
1.0
`odo catalog describe service postgresoperator.v5.0.3/PostgresCluster` is not helpfull - For example the output mentions `instances` that should be array of the object, but it doesn't say how the objects should look like ``` instances ([]object) -required- ``` This is the same for every embedded object mentioned in the spec, there are no detail about their structure. /priority high /area service-operators
non_main
odo catalog describe service postgresoperator postgrescluster is not helpfull for example the output mentions instances that should be array of the object but it doesn t say how the objects should look like instances object required this is the same for every embedded object mentioned in the spec there are no detail about their structure priority high area service operators
0
226,401
24,947,086,052
IssuesEvent
2022-11-01 01:50:13
rvvergara/advanced-web-developer-bootcamp-notes-examples-and-exercises
https://api.github.com/repos/rvvergara/advanced-web-developer-bootcamp-notes-examples-and-exercises
opened
CVE-2022-37620 (Medium) detected in html-minifier-3.5.19.tgz, html-minifier-3.5.18.tgz
security vulnerability
## CVE-2022-37620 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>html-minifier-3.5.19.tgz</b>, <b>html-minifier-3.5.18.tgz</b></p></summary> <p> <details><summary><b>html-minifier-3.5.19.tgz</b></p></summary> <p>Highly configurable, well-tested, JavaScript-based HTML minifier.</p> <p>Library home page: <a href="https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.19.tgz">https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.19.tgz</a></p> <p>Path to dependency file: /Section30-Building-A-FullStack-App-With-React/fullstack-todo-app/todos-react-frontent/package.json</p> <p>Path to vulnerable library: /Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.4.tgz (Root Library) - html-webpack-plugin-2.29.0.tgz - :x: **html-minifier-3.5.19.tgz** (Vulnerable Library) </details> <details><summary><b>html-minifier-3.5.18.tgz</b></p></summary> <p>Highly configurable, well-tested, JavaScript-based HTML minifier.</p> <p>Library home page: <a href="https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.18.tgz">https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.18.tgz</a></p> <p>Path to dependency file: /Section26-Create-React-App-and-Props/recipe-props-starter/package.json</p> <p>Path to vulnerable library: /Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.4.tgz (Root Library) - html-webpack-plugin-2.29.0.tgz - :x: **html-minifier-3.5.18.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/rvvergara/advanced-web-developer-bootcamp-notes-examples-and-exercises/commit/d80f181a2f4dd1169fbef38813309ffdd4ad82d4">d80f181a2f4dd1169fbef38813309ffdd4ad82d4</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular Expression Denial of Service (ReDoS) flaw was found in kangax html-minifier 4.0.0 via the candidate variable in htmlminifier.js. <p>Publish Date: 2022-10-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37620>CVE-2022-37620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-37620 (Medium) detected in html-minifier-3.5.19.tgz, html-minifier-3.5.18.tgz - ## CVE-2022-37620 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>html-minifier-3.5.19.tgz</b>, <b>html-minifier-3.5.18.tgz</b></p></summary> <p> <details><summary><b>html-minifier-3.5.19.tgz</b></p></summary> <p>Highly configurable, well-tested, JavaScript-based HTML minifier.</p> <p>Library home page: <a href="https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.19.tgz">https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.19.tgz</a></p> <p>Path to dependency file: /Section30-Building-A-FullStack-App-With-React/fullstack-todo-app/todos-react-frontent/package.json</p> <p>Path to vulnerable library: /Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/memory-game-exercise/node_modules/html-minifier/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.4.tgz (Root Library) - html-webpack-plugin-2.29.0.tgz - :x: **html-minifier-3.5.19.tgz** (Vulnerable Library) </details> <details><summary><b>html-minifier-3.5.18.tgz</b></p></summary> <p>Highly configurable, well-tested, JavaScript-based HTML minifier.</p> <p>Library home page: <a href="https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.18.tgz">https://registry.npmjs.org/html-minifier/-/html-minifier-3.5.18.tgz</a></p> <p>Path to dependency file: /Section26-Create-React-App-and-Props/recipe-props-starter/package.json</p> <p>Path to vulnerable library: /Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json,/Section28-The-Virtual-DOM-Events-and-Forms/recipe-app-with-state/node_modules/html-minifier/package.json</p> <p> Dependency Hierarchy: - react-scripts-1.1.4.tgz (Root Library) - html-webpack-plugin-2.29.0.tgz - :x: **html-minifier-3.5.18.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/rvvergara/advanced-web-developer-bootcamp-notes-examples-and-exercises/commit/d80f181a2f4dd1169fbef38813309ffdd4ad82d4">d80f181a2f4dd1169fbef38813309ffdd4ad82d4</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular Expression Denial of Service (ReDoS) flaw was found in kangax html-minifier 4.0.0 via the candidate variable in htmlminifier.js. <p>Publish Date: 2022-10-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-37620>CVE-2022-37620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in html minifier tgz html minifier tgz cve medium severity vulnerability vulnerable libraries html minifier tgz html minifier tgz html minifier tgz highly configurable well tested javascript based html minifier library home page a href path to dependency file building a fullstack app with react fullstack todo app todos react frontent package json path to vulnerable library the virtual dom events and forms memory game exercise node modules html minifier package json the virtual dom events and forms memory game exercise node modules html minifier package json the virtual dom events and forms memory game exercise node modules html minifier package json the virtual dom events and forms memory game exercise node modules html minifier package json the virtual dom events and forms memory game exercise node modules html minifier package json the virtual dom events and forms memory game exercise node modules html minifier package json dependency hierarchy react scripts tgz root library html webpack plugin tgz x html minifier tgz vulnerable library html minifier tgz highly configurable well tested javascript based html minifier library home page a href path to dependency file create react app and props recipe props starter package json path to vulnerable library the virtual dom events and forms recipe app with state node modules html minifier package json the virtual dom events and forms recipe app with state node modules html minifier package json the virtual dom events and forms recipe app with state node modules html minifier package json the virtual dom events and forms recipe app with state node modules html minifier package json the virtual dom events and forms recipe app with state node modules html minifier package json the virtual dom events and forms recipe app with state node modules html minifier package json dependency hierarchy react scripts tgz root library html webpack plugin tgz x html minifier tgz vulnerable library found in head commit a href found in base branch master vulnerability details a regular expression denial of service redos flaw was found in kangax html minifier via the candidate variable in htmlminifier js publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href step up your open source security game with mend
0
2,335
8,360,412,090
IssuesEvent
2018-10-03 11:27:35
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
Flush cache with Redis fact cache gives Unexpected Exception
affects_2.6 bug module needs_maintainer needs_triage support:community support:core traceback
<!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and devel branch are affected too --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY With Redis as a fact cache, ansible-playbook run with the --flush-cache option results in: ERROR! Unexpected Exception, this is probably a bug: u'my.hostname' Probably a regression of: https://github.com/ansible/ansible/issues/31148 ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ansible-playbook, fact cache, Redis ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible 2.6.4 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /bin/ansible python version = 2.7.5 (default, Jul 13 2018, 13:06:57) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below CACHE_PLUGIN(/etc/ansible/ansible.cfg) = redis CACHE_PLUGIN_CONNECTION(/etc/ansible/ansible.cfg) = localhost:6379:0 CACHE_PLUGIN_TIMEOUT(/etc/ansible/ansible.cfg) = 86400 DEFAULT_GATHERING(/etc/ansible/ansible.cfg) = smart HOST_KEY_CHECKING(/etc/ansible/ansible.cfg) = False INVENTORY_ENABLED(/etc/ansible/ansible.cfg) = ['host_list', 'foreman', 'yaml', 'constructed', 'script', 'ini'] RETRY_FILES_ENABLED(/etc/ansible/ansible.cfg) = False ``` ##### OS / ENVIRONMENT CentOS Linux release 7.5.1804 (Core) ansible.noarch 2.6.4-1.el7 @epel ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> yum install redis systemctl start redis make required changes to /etc/ansible/ansible.cfg to enable fact caching to your local Redis instance run your playbook without --flush-cache to generate cache entries run the playbook again with the --flush-cache option and note the error repeat the test again after doing a: redis-cli FLUSHALL run ansible-playbook again, even with flush-cache option at this point it works since there's no cache data run once more with --flush-cache and note the failure <!--- Paste example playbooks or commands between quotes below --> ```yaml --- - hosts: - all tasks: - name: ping ping: ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS ansible-playbook run with the --flush-cache option succeeds without error ##### ACTUAL RESULTS ansible-playbook run with the --flush-cache option fails with an error ```paste below ansible-playbook -i /etc/ansible/puppet.foreman.yml ./refresh-ansible-cache.yml -f 25 --flush-cache -u rundeck --private-key /home/rundeck/.ssh/id_rsa --limit kafka-2* -vvv ansible-playbook 2.6.4 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /bin/ansible-playbook python version = 2.7.5 (default, Jul 13 2018, 13:06:57) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] Using /etc/ansible/ansible.cfg as config file Parsed /etc/ansible/puppet.foreman.yml inventory source with foreman plugin ERROR! Unexpected Exception, this is probably a bug: u'my.host.name' the full traceback was: Traceback (most recent call last): File "/bin/ansible-playbook", line 118, in <module> exit_code = cli.run() File "/usr/lib/python2.7/site-packages/ansible/cli/playbook.py", line 116, in run self._flush_cache(inventory, variable_manager) File "/usr/lib/python2.7/site-packages/ansible/cli/playbook.py", line 196, in _flush_cache variable_manager.clear_facts(hostname) File "/usr/lib/python2.7/site-packages/ansible/vars/manager.py", line 578, in clear_facts del self._fact_cache[hostname] File "/usr/lib/python2.7/site-packages/ansible/plugins/cache/__init__.py", line 273, in __delitem__ self._plugin.delete(key) File "/usr/lib/python2.7/site-packages/ansible/plugins/cache/redis.py", line 119, in delete del self._cache[key] KeyError: u'my.host.name' ```
True
Flush cache with Redis fact cache gives Unexpected Exception - <!--- Verify first that your issue is not already reported on GitHub --> <!--- Also test if the latest release and devel branch are affected too --> <!--- Complete *all* sections as described, this form is processed automatically --> ##### SUMMARY With Redis as a fact cache, ansible-playbook run with the --flush-cache option results in: ERROR! Unexpected Exception, this is probably a bug: u'my.hostname' Probably a regression of: https://github.com/ansible/ansible/issues/31148 ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ansible-playbook, fact cache, Redis ##### ANSIBLE VERSION <!--- Paste verbatim output from "ansible --version" between quotes --> ```paste below ansible 2.6.4 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /bin/ansible python version = 2.7.5 (default, Jul 13 2018, 13:06:57) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] ``` ##### CONFIGURATION <!--- Paste verbatim output from "ansible-config dump --only-changed" between quotes --> ```paste below CACHE_PLUGIN(/etc/ansible/ansible.cfg) = redis CACHE_PLUGIN_CONNECTION(/etc/ansible/ansible.cfg) = localhost:6379:0 CACHE_PLUGIN_TIMEOUT(/etc/ansible/ansible.cfg) = 86400 DEFAULT_GATHERING(/etc/ansible/ansible.cfg) = smart HOST_KEY_CHECKING(/etc/ansible/ansible.cfg) = False INVENTORY_ENABLED(/etc/ansible/ansible.cfg) = ['host_list', 'foreman', 'yaml', 'constructed', 'script', 'ini'] RETRY_FILES_ENABLED(/etc/ansible/ansible.cfg) = False ``` ##### OS / ENVIRONMENT CentOS Linux release 7.5.1804 (Core) ansible.noarch 2.6.4-1.el7 @epel ##### STEPS TO REPRODUCE <!--- Describe exactly how to reproduce the problem, using a minimal test-case --> yum install redis systemctl start redis make required changes to /etc/ansible/ansible.cfg to enable fact caching to your local Redis instance run your playbook without --flush-cache to generate cache entries run the playbook again with the --flush-cache option and note the error repeat the test again after doing a: redis-cli FLUSHALL run ansible-playbook again, even with flush-cache option at this point it works since there's no cache data run once more with --flush-cache and note the failure <!--- Paste example playbooks or commands between quotes below --> ```yaml --- - hosts: - all tasks: - name: ping ping: ``` <!--- HINT: You can paste gist.github.com links for larger files --> ##### EXPECTED RESULTS ansible-playbook run with the --flush-cache option succeeds without error ##### ACTUAL RESULTS ansible-playbook run with the --flush-cache option fails with an error ```paste below ansible-playbook -i /etc/ansible/puppet.foreman.yml ./refresh-ansible-cache.yml -f 25 --flush-cache -u rundeck --private-key /home/rundeck/.ssh/id_rsa --limit kafka-2* -vvv ansible-playbook 2.6.4 config file = /etc/ansible/ansible.cfg configured module search path = [u'/root/.ansible/plugins/modules', u'/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python2.7/site-packages/ansible executable location = /bin/ansible-playbook python version = 2.7.5 (default, Jul 13 2018, 13:06:57) [GCC 4.8.5 20150623 (Red Hat 4.8.5-28)] Using /etc/ansible/ansible.cfg as config file Parsed /etc/ansible/puppet.foreman.yml inventory source with foreman plugin ERROR! Unexpected Exception, this is probably a bug: u'my.host.name' the full traceback was: Traceback (most recent call last): File "/bin/ansible-playbook", line 118, in <module> exit_code = cli.run() File "/usr/lib/python2.7/site-packages/ansible/cli/playbook.py", line 116, in run self._flush_cache(inventory, variable_manager) File "/usr/lib/python2.7/site-packages/ansible/cli/playbook.py", line 196, in _flush_cache variable_manager.clear_facts(hostname) File "/usr/lib/python2.7/site-packages/ansible/vars/manager.py", line 578, in clear_facts del self._fact_cache[hostname] File "/usr/lib/python2.7/site-packages/ansible/plugins/cache/__init__.py", line 273, in __delitem__ self._plugin.delete(key) File "/usr/lib/python2.7/site-packages/ansible/plugins/cache/redis.py", line 119, in delete del self._cache[key] KeyError: u'my.host.name' ```
main
flush cache with redis fact cache gives unexpected exception summary with redis as a fact cache ansible playbook run with the flush cache option results in error unexpected exception this is probably a bug u my hostname probably a regression of issue type bug report component name ansible playbook fact cache redis ansible version paste below ansible config file etc ansible ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location bin ansible python version default jul configuration paste below cache plugin etc ansible ansible cfg redis cache plugin connection etc ansible ansible cfg localhost cache plugin timeout etc ansible ansible cfg default gathering etc ansible ansible cfg smart host key checking etc ansible ansible cfg false inventory enabled etc ansible ansible cfg retry files enabled etc ansible ansible cfg false os environment centos linux release core ansible noarch epel steps to reproduce yum install redis systemctl start redis make required changes to etc ansible ansible cfg to enable fact caching to your local redis instance run your playbook without flush cache to generate cache entries run the playbook again with the flush cache option and note the error repeat the test again after doing a redis cli flushall run ansible playbook again even with flush cache option at this point it works since there s no cache data run once more with flush cache and note the failure yaml hosts all tasks name ping ping expected results ansible playbook run with the flush cache option succeeds without error actual results ansible playbook run with the flush cache option fails with an error paste below ansible playbook i etc ansible puppet foreman yml refresh ansible cache yml f flush cache u rundeck private key home rundeck ssh id rsa limit kafka vvv ansible playbook config file etc ansible ansible cfg configured module search path ansible python module location usr lib site packages ansible executable location bin ansible playbook python version default jul using etc ansible ansible cfg as config file parsed etc ansible puppet foreman yml inventory source with foreman plugin error unexpected exception this is probably a bug u my host name the full traceback was traceback most recent call last file bin ansible playbook line in exit code cli run file usr lib site packages ansible cli playbook py line in run self flush cache inventory variable manager file usr lib site packages ansible cli playbook py line in flush cache variable manager clear facts hostname file usr lib site packages ansible vars manager py line in clear facts del self fact cache file usr lib site packages ansible plugins cache init py line in delitem self plugin delete key file usr lib site packages ansible plugins cache redis py line in delete del self cache keyerror u my host name
1
333,533
24,378,604,619
IssuesEvent
2022-10-04 05:09:15
Divested-Mobile/Mull-Fenix
https://api.github.com/repos/Divested-Mobile/Mull-Fenix
closed
Flutter web pages: html versions blurry & no scrolling, canvaskit only blank screen
documentation
[Flutter](https://flutter.dev/) apps compiled to html web versions display much blurrier than in other browsers, and scrolling doesn't work. Also tested Fennec & original Firefox Android which have no problem with scrolling and have much less blurry items. Examples with description where scrolling should be: - https://gallery.flutter.dev/#/ -> open the "MATERIAL" category: On other browsers you get a scroll-able list, on Mull only the first 4 items are visible. - https://bdlukaa.github.io/fluent_ui/ -> for example if you open the "Icons" page, you get a very long scroll-able list on other browsers, on Mull: no scrolling. When forcing the web renderer to canvaskit (default = auto which serves html version to mobile browsers) Mull displays only a blank screen. Other browsers (inculuding Firefox based ones such as Fennec) display normal, though load times are long since canvaskit require a big initial download)
1.0
Flutter web pages: html versions blurry & no scrolling, canvaskit only blank screen - [Flutter](https://flutter.dev/) apps compiled to html web versions display much blurrier than in other browsers, and scrolling doesn't work. Also tested Fennec & original Firefox Android which have no problem with scrolling and have much less blurry items. Examples with description where scrolling should be: - https://gallery.flutter.dev/#/ -> open the "MATERIAL" category: On other browsers you get a scroll-able list, on Mull only the first 4 items are visible. - https://bdlukaa.github.io/fluent_ui/ -> for example if you open the "Icons" page, you get a very long scroll-able list on other browsers, on Mull: no scrolling. When forcing the web renderer to canvaskit (default = auto which serves html version to mobile browsers) Mull displays only a blank screen. Other browsers (inculuding Firefox based ones such as Fennec) display normal, though load times are long since canvaskit require a big initial download)
non_main
flutter web pages html versions blurry no scrolling canvaskit only blank screen apps compiled to html web versions display much blurrier than in other browsers and scrolling doesn t work also tested fennec original firefox android which have no problem with scrolling and have much less blurry items examples with description where scrolling should be open the material category on other browsers you get a scroll able list on mull only the first items are visible for example if you open the icons page you get a very long scroll able list on other browsers on mull no scrolling when forcing the web renderer to canvaskit default auto which serves html version to mobile browsers mull displays only a blank screen other browsers inculuding firefox based ones such as fennec display normal though load times are long since canvaskit require a big initial download
0
558,069
16,525,457,825
IssuesEvent
2021-05-26 19:28:21
GoogleCloudPlatform/cloud-code-intellij
https://api.github.com/repos/GoogleCloudPlatform/cloud-code-intellij
closed
CompletionException from lsp4j.jsonrpc.json.StreamMessageConsumer
area/lsp kind/bug priority/p3
``` java.util.concurrent.CompletionException: org.eclipse.lsp4j.jsonrpc.JsonRpcException: java.io.IOException: _ at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319) at java.base/java.util.concurrent.CompletableFuture$UniAccept.tryFire(CompletableFuture.java:718) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture$AsyncRun.run(CompletableFuture.java:1742) at java.base/java.util.concurrent.CompletableFuture$AsyncRun.exec(CompletableFuture.java:1728) at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:177) Caused by: org.eclipse.lsp4j.jsonrpc.JsonRpcException: java.io.IOException: _ at org.eclipse.lsp4j.jsonrpc.json.StreamMessageConsumer.consume(StreamMessageConsumer.java:72) at org.wso2.lsp4intellij.client.languageserver.wrapper.MessageHandler.lambda$apply$0(MessageHandler.java:45) at org.eclipse.lsp4j.jsonrpc.RemoteEndpoint.lambda$handleRequest$1(RemoteEndpoint.java:281) at java.base/java.util.concurrent.CompletableFuture$UniAccept.tryFire(CompletableFuture.java:714) ... 8 more Caused by: java.io.IOException: _ at java.base/java.io.FileOutputStream.writeBytes(Native Method) at java.base/java.io.FileOutputStream.write(FileOutputStream.java:354) at java.base/java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) at java.base/java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) at org.eclipse.lsp4j.jsonrpc.json.StreamMessageConsumer.consume(StreamMessageConsumer.java:69) ``` https://listnrcrash.corp.google.com/product/717949/report/1000000801159646158 CC 20.6.1 IDEA 2019.3.5
1.0
CompletionException from lsp4j.jsonrpc.json.StreamMessageConsumer - ``` java.util.concurrent.CompletionException: org.eclipse.lsp4j.jsonrpc.JsonRpcException: java.io.IOException: _ at java.base/java.util.concurrent.CompletableFuture.encodeThrowable(CompletableFuture.java:314) at java.base/java.util.concurrent.CompletableFuture.completeThrowable(CompletableFuture.java:319) at java.base/java.util.concurrent.CompletableFuture$UniAccept.tryFire(CompletableFuture.java:718) at java.base/java.util.concurrent.CompletableFuture.postComplete(CompletableFuture.java:506) at java.base/java.util.concurrent.CompletableFuture$AsyncRun.run(CompletableFuture.java:1742) at java.base/java.util.concurrent.CompletableFuture$AsyncRun.exec(CompletableFuture.java:1728) at java.base/java.util.concurrent.ForkJoinTask.doExec(ForkJoinTask.java:290) at java.base/java.util.concurrent.ForkJoinPool$WorkQueue.topLevelExec(ForkJoinPool.java:1020) at java.base/java.util.concurrent.ForkJoinPool.scan(ForkJoinPool.java:1656) at java.base/java.util.concurrent.ForkJoinPool.runWorker(ForkJoinPool.java:1594) at java.base/java.util.concurrent.ForkJoinWorkerThread.run(ForkJoinWorkerThread.java:177) Caused by: org.eclipse.lsp4j.jsonrpc.JsonRpcException: java.io.IOException: _ at org.eclipse.lsp4j.jsonrpc.json.StreamMessageConsumer.consume(StreamMessageConsumer.java:72) at org.wso2.lsp4intellij.client.languageserver.wrapper.MessageHandler.lambda$apply$0(MessageHandler.java:45) at org.eclipse.lsp4j.jsonrpc.RemoteEndpoint.lambda$handleRequest$1(RemoteEndpoint.java:281) at java.base/java.util.concurrent.CompletableFuture$UniAccept.tryFire(CompletableFuture.java:714) ... 8 more Caused by: java.io.IOException: _ at java.base/java.io.FileOutputStream.writeBytes(Native Method) at java.base/java.io.FileOutputStream.write(FileOutputStream.java:354) at java.base/java.io.BufferedOutputStream.flushBuffer(BufferedOutputStream.java:81) at java.base/java.io.BufferedOutputStream.flush(BufferedOutputStream.java:142) at org.eclipse.lsp4j.jsonrpc.json.StreamMessageConsumer.consume(StreamMessageConsumer.java:69) ``` https://listnrcrash.corp.google.com/product/717949/report/1000000801159646158 CC 20.6.1 IDEA 2019.3.5
non_main
completionexception from jsonrpc json streammessageconsumer java util concurrent completionexception org eclipse jsonrpc jsonrpcexception java io ioexception at java base java util concurrent completablefuture encodethrowable completablefuture java at java base java util concurrent completablefuture completethrowable completablefuture java at java base java util concurrent completablefuture uniaccept tryfire completablefuture java at java base java util concurrent completablefuture postcomplete completablefuture java at java base java util concurrent completablefuture asyncrun run completablefuture java at java base java util concurrent completablefuture asyncrun exec completablefuture java at java base java util concurrent forkjointask doexec forkjointask java at java base java util concurrent forkjoinpool workqueue toplevelexec forkjoinpool java at java base java util concurrent forkjoinpool scan forkjoinpool java at java base java util concurrent forkjoinpool runworker forkjoinpool java at java base java util concurrent forkjoinworkerthread run forkjoinworkerthread java caused by org eclipse jsonrpc jsonrpcexception java io ioexception at org eclipse jsonrpc json streammessageconsumer consume streammessageconsumer java at org client languageserver wrapper messagehandler lambda apply messagehandler java at org eclipse jsonrpc remoteendpoint lambda handlerequest remoteendpoint java at java base java util concurrent completablefuture uniaccept tryfire completablefuture java more caused by java io ioexception at java base java io fileoutputstream writebytes native method at java base java io fileoutputstream write fileoutputstream java at java base java io bufferedoutputstream flushbuffer bufferedoutputstream java at java base java io bufferedoutputstream flush bufferedoutputstream java at org eclipse jsonrpc json streammessageconsumer consume streammessageconsumer java cc idea
0
2,360
8,415,670,096
IssuesEvent
2018-10-13 17:02:30
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
Bug in error handling Librato Annotation module
affects_2.0 bug module needs_maintainer support:community traceback
From @casperbakker on 2016-05-19T10:25:48Z ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME monitoring/librato_annotation ##### ANSIBLE VERSION ``` ansible 2.0.2.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT Ubuntu 14.04 ##### SUMMARY Apparently my Librato credentials are incorrect. But the var 'e' at https://github.com/ansible/ansible-modules-extras/blob/2a0c5e2a8fd7ed3ce6d6eedd08e85e01e1617113/monitoring/librato_annotation.py#L136 seems to be not placed right. I have no knowledge about Python, but seems like a bug to me. Result: ``` fatal: [IP]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/home/ubuntu/.ansible/tmp/ansible-tmp-1463652618.03-36649702854390/librato_annotation\", line 3003, in <module>\r\n main()\r\n File \"/home/ubuntu/.ansible/tmp/ansible-tmp-1463652618.03-36649702854390/librato_annotation\", line 157, in main\r\n post_annotation(module)\r\n File \"/home/ubuntu/.ansible/tmp/ansible-tmp-1463652618.03-36649702854390/librato_annotation\", line 137, in post_annotation\r\n module.fail_json(msg=\"Request Failed\", reason=e.reason)\r\nNameError: global name 'e' is not defined\r\n", "msg": "MODULE FAILURE", "parsed": false} ``` ##### STEPS TO REPRODUCE ``` - name: Annotate Librato librato_annotation: user: "{{ secret_librato_username }}" api_key: "{{ secret_librato_api_key }}" title: New deploy name: app-deploys source: "{{ application_env }}" when: '"production" in group_names' ``` ##### EXPECTED RESULTS I expect a 'normal' error message with "Request Failed" ##### ACTUAL RESULTS Got a stacktrace about "global name 'e' is not defined" Copied from original issue: ansible/ansible-modules-extras#2256
True
Bug in error handling Librato Annotation module - From @casperbakker on 2016-05-19T10:25:48Z ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME monitoring/librato_annotation ##### ANSIBLE VERSION ``` ansible 2.0.2.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### OS / ENVIRONMENT Ubuntu 14.04 ##### SUMMARY Apparently my Librato credentials are incorrect. But the var 'e' at https://github.com/ansible/ansible-modules-extras/blob/2a0c5e2a8fd7ed3ce6d6eedd08e85e01e1617113/monitoring/librato_annotation.py#L136 seems to be not placed right. I have no knowledge about Python, but seems like a bug to me. Result: ``` fatal: [IP]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/home/ubuntu/.ansible/tmp/ansible-tmp-1463652618.03-36649702854390/librato_annotation\", line 3003, in <module>\r\n main()\r\n File \"/home/ubuntu/.ansible/tmp/ansible-tmp-1463652618.03-36649702854390/librato_annotation\", line 157, in main\r\n post_annotation(module)\r\n File \"/home/ubuntu/.ansible/tmp/ansible-tmp-1463652618.03-36649702854390/librato_annotation\", line 137, in post_annotation\r\n module.fail_json(msg=\"Request Failed\", reason=e.reason)\r\nNameError: global name 'e' is not defined\r\n", "msg": "MODULE FAILURE", "parsed": false} ``` ##### STEPS TO REPRODUCE ``` - name: Annotate Librato librato_annotation: user: "{{ secret_librato_username }}" api_key: "{{ secret_librato_api_key }}" title: New deploy name: app-deploys source: "{{ application_env }}" when: '"production" in group_names' ``` ##### EXPECTED RESULTS I expect a 'normal' error message with "Request Failed" ##### ACTUAL RESULTS Got a stacktrace about "global name 'e' is not defined" Copied from original issue: ansible/ansible-modules-extras#2256
main
bug in error handling librato annotation module from casperbakker on issue type bug report component name monitoring librato annotation ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides os environment ubuntu summary apparently my librato credentials are incorrect but the var e at seems to be not placed right i have no knowledge about python but seems like a bug to me result fatal failed changed false failed true module stderr module stdout traceback most recent call last r n file home ubuntu ansible tmp ansible tmp librato annotation line in r n main r n file home ubuntu ansible tmp ansible tmp librato annotation line in main r n post annotation module r n file home ubuntu ansible tmp ansible tmp librato annotation line in post annotation r n module fail json msg request failed reason e reason r nnameerror global name e is not defined r n msg module failure parsed false steps to reproduce name annotate librato librato annotation user secret librato username api key secret librato api key title new deploy name app deploys source application env when production in group names expected results i expect a normal error message with request failed actual results got a stacktrace about global name e is not defined copied from original issue ansible ansible modules extras
1
1,663
6,574,059,391
IssuesEvent
2017-09-11 11:17:55
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
synchronize become_user not honored when become_method: su is used
affects_2.3 bug_report waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> synchronize ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Red Hat Enterprise Linux Server release 6.7 (Santiago) ##### SUMMARY <!--- Explain the problem briefly --> become_method: su should allow the remote execution to occur as the become_user specified. However, that is not occurring. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Specify become and synchronize parameters as below. Execute as follows and provide pwd for become_user ansible-playbook dir1_sync.yml -i ansible/hosts --ask-become --diff <!--- Paste example playbooks or commands between quotes below --> ``` become: true become_method: su become_user: ruser tasks: - name: "Verify dir1 sync" synchronize: src: "/home/luser/dir1" dest: "/apps/dir1" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I expect the remote commands to run as the become_user specified. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> The remote command did not run as the become_user. If I change the dest: path to be /tmp, the command executes but the resulting directory on the remote machine is owned by the ssh user. This issue is similar to https://github.com/ansible/ansible-modules-core/issues/4508 but the workaround doesn't work since using become_method: su. <!--- Paste verbatim command output between quotes below --> ``` fatal: [10.193.239.53]: FAILED! => {"changed": false, "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --rsh 'ssh -S none -o StrictHostKeyChecking=no' --out-format='<<CHANGED>>%i %n%L' \"/home/luser/dir1" \"10.193.239.53:/apps\"", "failed": true, "msg": "rsync: mkdir \"/apps\" failed: Permission denied (13)\nrsync error: error in file IO (code 11) at main.c(576) [receiver=3.0.6]\nrsync: connection unexpectedly closed (9 bytes received so far) [sender]\nrsync error: error in rsync protocol data stream (code 12) at io.c(600) [sender=3.0.6]\n", "rc": 12} ```
True
synchronize become_user not honored when become_method: su is used - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> synchronize ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Red Hat Enterprise Linux Server release 6.7 (Santiago) ##### SUMMARY <!--- Explain the problem briefly --> become_method: su should allow the remote execution to occur as the become_user specified. However, that is not occurring. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> Specify become and synchronize parameters as below. Execute as follows and provide pwd for become_user ansible-playbook dir1_sync.yml -i ansible/hosts --ask-become --diff <!--- Paste example playbooks or commands between quotes below --> ``` become: true become_method: su become_user: ruser tasks: - name: "Verify dir1 sync" synchronize: src: "/home/luser/dir1" dest: "/apps/dir1" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> I expect the remote commands to run as the become_user specified. ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> The remote command did not run as the become_user. If I change the dest: path to be /tmp, the command executes but the resulting directory on the remote machine is owned by the ssh user. This issue is similar to https://github.com/ansible/ansible-modules-core/issues/4508 but the workaround doesn't work since using become_method: su. <!--- Paste verbatim command output between quotes below --> ``` fatal: [10.193.239.53]: FAILED! => {"changed": false, "cmd": "/usr/bin/rsync --delay-updates -F --compress --archive --rsh 'ssh -S none -o StrictHostKeyChecking=no' --out-format='<<CHANGED>>%i %n%L' \"/home/luser/dir1" \"10.193.239.53:/apps\"", "failed": true, "msg": "rsync: mkdir \"/apps\" failed: Permission denied (13)\nrsync error: error in file IO (code 11) at main.c(576) [receiver=3.0.6]\nrsync: connection unexpectedly closed (9 bytes received so far) [sender]\nrsync error: error in rsync protocol data stream (code 12) at io.c(600) [sender=3.0.6]\n", "rc": 12} ```
main
synchronize become user not honored when become method su is used issue type bug report component name synchronize ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific red hat enterprise linux server release santiago summary become method su should allow the remote execution to occur as the become user specified however that is not occurring steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used specify become and synchronize parameters as below execute as follows and provide pwd for become user ansible playbook sync yml i ansible hosts ask become diff become true become method su become user ruser tasks name verify sync synchronize src home luser dest apps expected results i expect the remote commands to run as the become user specified actual results the remote command did not run as the become user if i change the dest path to be tmp the command executes but the resulting directory on the remote machine is owned by the ssh user this issue is similar to but the workaround doesn t work since using become method su fatal failed changed false cmd usr bin rsync delay updates f compress archive rsh ssh s none o stricthostkeychecking no out format i n l home luser apps failed true msg rsync mkdir apps failed permission denied nrsync error error in file io code at main c nrsync connection unexpectedly closed bytes received so far nrsync error error in rsync protocol data stream code at io c n rc
1
2,505
8,655,459,652
IssuesEvent
2018-11-27 16:00:30
codestation/qcma
https://api.github.com/repos/codestation/qcma
closed
When transferring homebrew files using psp savedata, *.lnk files will be omitted.
unmaintained
If the data contains a file with *.lnk extension, qcma will ignore it, and omit it in the upload. To fix it, QListDB::recursiveScanRootDirectory should be changed a little. `dir.entryInfoList(QDir::AllEntries | QDir::NoDotAndDotDot, QDir::Time);` To `dir.entryInfoList(QDir::AllEntries | QDir::NoDotAndDotDot | QDir::System, QDir::Time);`
True
When transferring homebrew files using psp savedata, *.lnk files will be omitted. - If the data contains a file with *.lnk extension, qcma will ignore it, and omit it in the upload. To fix it, QListDB::recursiveScanRootDirectory should be changed a little. `dir.entryInfoList(QDir::AllEntries | QDir::NoDotAndDotDot, QDir::Time);` To `dir.entryInfoList(QDir::AllEntries | QDir::NoDotAndDotDot | QDir::System, QDir::Time);`
main
when transferring homebrew files using psp savedata lnk files will be omitted if the data contains a file with lnk extension qcma will ignore it and omit it in the upload to fix it qlistdb recursivescanrootdirectory should be changed a little dir entryinfolist qdir allentries qdir nodotanddotdot qdir time to dir entryinfolist qdir allentries qdir nodotanddotdot qdir system qdir time
1
3,613
14,614,903,225
IssuesEvent
2020-12-22 10:38:36
melisMirza/SWE573_project
https://api.github.com/repos/melisMirza/SWE573_project
opened
Dockerize the Django Application
backend maintainance
Dockerize a trial django application. Main purpose is to understand the docker-compose.yaml, requirements.txt and containerization of the application. Run the application using Docker.
True
Dockerize the Django Application - Dockerize a trial django application. Main purpose is to understand the docker-compose.yaml, requirements.txt and containerization of the application. Run the application using Docker.
main
dockerize the django application dockerize a trial django application main purpose is to understand the docker compose yaml requirements txt and containerization of the application run the application using docker
1
4,323
21,731,062,430
IssuesEvent
2022-05-11 12:03:58
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
closed
Fix type_options validation for Money type
type: bug work: backend restricted: maintainers status: started
## Reproduce 1. Set up a Money column. 1. Submit a `PATCH` request to the columns API, e.g. `/api/db/v0/tables/15/columns/52/` 1. Send: ```json5 { "type": "MATHESAR_TYPES.MATHESAR_MONEY", "type_options": { "precision": 100, "scale": 2 } } ``` Expect success. Receive: ```json [ { "code": 4210, "message": "Unknown type_option passed", "field": null, "detail": null } ] ``` 1. Send: ```json5 { "type": "MATHESAR_TYPES.MATHESAR_MONEY", "type_options": { "precision": null, "scale": null } } ``` Expect success. Receive: ```json [ { "code": 2023, "field": "precision", "message": "This field may not be null.", "detail": {} }, { "code": 2023, "field": "scale", "message": "This field may not be null.", "detail": {} } ] ```
True
Fix type_options validation for Money type - ## Reproduce 1. Set up a Money column. 1. Submit a `PATCH` request to the columns API, e.g. `/api/db/v0/tables/15/columns/52/` 1. Send: ```json5 { "type": "MATHESAR_TYPES.MATHESAR_MONEY", "type_options": { "precision": 100, "scale": 2 } } ``` Expect success. Receive: ```json [ { "code": 4210, "message": "Unknown type_option passed", "field": null, "detail": null } ] ``` 1. Send: ```json5 { "type": "MATHESAR_TYPES.MATHESAR_MONEY", "type_options": { "precision": null, "scale": null } } ``` Expect success. Receive: ```json [ { "code": 2023, "field": "precision", "message": "This field may not be null.", "detail": {} }, { "code": 2023, "field": "scale", "message": "This field may not be null.", "detail": {} } ] ```
main
fix type options validation for money type reproduce set up a money column submit a patch request to the columns api e g api db tables columns send type mathesar types mathesar money type options precision scale expect success receive json code message unknown type option passed field null detail null send type mathesar types mathesar money type options precision null scale null expect success receive json code field precision message this field may not be null detail code field scale message this field may not be null detail
1
2,716
9,545,498,119
IssuesEvent
2019-05-01 17:14:04
Homebrew/homebrew-cask
https://api.github.com/repos/Homebrew/homebrew-cask
closed
vulkan-sdk cask
awaiting maintainer feedback
Last August @vitorgalvao and I discussed in [this thread](https://github.com/Homebrew/homebrew-cask/issues/51330#event-1958163894) a cask for the vulkan-sdk. In your message, you said: > @apenngrace I’m sympathetic to your case, but I doubt this is used by enough people to warrant the trouble (would have to check the analytics). Regarding user interest, I have had pretty consistent daily traffic for my cask (even though it is a private tap). Would you re-consider allowing this cask to be part of the homebrew-cask repo again? As we discussed months ago, there is currently no build from source option for this project... even though it is theoretically possible. In the meantime, there is duplicate formulae appearing such as `vulkan-headers` that only installs headers whereas my cask installs all assets in a functional manner that can be tested on the command line with the `vulkaninfo` command after installation.
True
vulkan-sdk cask - Last August @vitorgalvao and I discussed in [this thread](https://github.com/Homebrew/homebrew-cask/issues/51330#event-1958163894) a cask for the vulkan-sdk. In your message, you said: > @apenngrace I’m sympathetic to your case, but I doubt this is used by enough people to warrant the trouble (would have to check the analytics). Regarding user interest, I have had pretty consistent daily traffic for my cask (even though it is a private tap). Would you re-consider allowing this cask to be part of the homebrew-cask repo again? As we discussed months ago, there is currently no build from source option for this project... even though it is theoretically possible. In the meantime, there is duplicate formulae appearing such as `vulkan-headers` that only installs headers whereas my cask installs all assets in a functional manner that can be tested on the command line with the `vulkaninfo` command after installation.
main
vulkan sdk cask last august vitorgalvao and i discussed in a cask for the vulkan sdk in your message you said apenngrace i’m sympathetic to your case but i doubt this is used by enough people to warrant the trouble would have to check the analytics regarding user interest i have had pretty consistent daily traffic for my cask even though it is a private tap would you re consider allowing this cask to be part of the homebrew cask repo again as we discussed months ago there is currently no build from source option for this project even though it is theoretically possible in the meantime there is duplicate formulae appearing such as vulkan headers that only installs headers whereas my cask installs all assets in a functional manner that can be tested on the command line with the vulkaninfo command after installation
1
2,702
9,488,317,602
IssuesEvent
2019-04-22 19:14:15
hydroshare/hydroshare
https://api.github.com/repos/hydroshare/hydroshare
closed
3 Listing Pages dont behave the same over Organization as First Author
Maintainability Needs Discussion RESOURCE UI COMPONENT
I just created a res on QAT with account czo_boulder/123. It has only one Author which is an Organization. http://qat.hydroshare.org:8000/resource/9cdc19783a574c289e87a4f416ef3785/ I shared this resource with CZO Boulder Group with View Permission. Its Resource Landing Page shows Author correctly. ![image](https://user-images.githubusercontent.com/10070599/55963290-c469d800-5c2f-11e9-97b8-df9b4dc0e19d.png) This resource shows up in 3 listings: Listing 1: My Resource Listing shows Organisation as First Author, which is correct I think ![image](https://user-images.githubusercontent.com/10070599/55963332-db102f00-5c2f-11e9-93fe-8638b4c35471.png) Listing 2: CZO Boulder's Group Resource Listing --- First Author is empty ![image](https://user-images.githubusercontent.com/10070599/55963363-e82d1e00-5c2f-11e9-94bb-0bbf4b8d88e7.png) Listing 3: CZO Community Resource Listing -- First Author is empty ![image](https://user-images.githubusercontent.com/10070599/55963388-f7ac6700-5c2f-11e9-8777-2dbf947749d9.png)
True
3 Listing Pages dont behave the same over Organization as First Author - I just created a res on QAT with account czo_boulder/123. It has only one Author which is an Organization. http://qat.hydroshare.org:8000/resource/9cdc19783a574c289e87a4f416ef3785/ I shared this resource with CZO Boulder Group with View Permission. Its Resource Landing Page shows Author correctly. ![image](https://user-images.githubusercontent.com/10070599/55963290-c469d800-5c2f-11e9-97b8-df9b4dc0e19d.png) This resource shows up in 3 listings: Listing 1: My Resource Listing shows Organisation as First Author, which is correct I think ![image](https://user-images.githubusercontent.com/10070599/55963332-db102f00-5c2f-11e9-93fe-8638b4c35471.png) Listing 2: CZO Boulder's Group Resource Listing --- First Author is empty ![image](https://user-images.githubusercontent.com/10070599/55963363-e82d1e00-5c2f-11e9-94bb-0bbf4b8d88e7.png) Listing 3: CZO Community Resource Listing -- First Author is empty ![image](https://user-images.githubusercontent.com/10070599/55963388-f7ac6700-5c2f-11e9-8777-2dbf947749d9.png)
main
listing pages dont behave the same over organization as first author i just created a res on qat with account czo boulder it has only one author which is an organization i shared this resource with czo boulder group with view permission its resource landing page shows author correctly this resource shows up in listings listing my resource listing shows organisation as first author which is correct i think listing czo boulder s group resource listing first author is empty listing czo community resource listing first author is empty
1
5,226
26,507,019,456
IssuesEvent
2023-01-18 14:32:27
precice/precice
https://api.github.com/repos/precice/precice
closed
Define primitives in terms of vertices
enhancement usability maintainability
**Please describe the problem you are trying to solve.** Currently, triangles are defined in terms of edges an edges are defined in terms of vertices. This makes defining triangles in adapter codes and integration tests tedious: ``` meshID = getMeshID(); setMeshVertices(...); for e in edges: setMeshEdge(...); for t in trianges setMeshTriangle(...); ``` This will become extremely tedious once we support tetrahedral elements. For consistency, they would need to be defined using triangles. As the `mesh::Triangle` is implemented this way as well, we need to still create `mesh::Edge`s for each triangle. Meaning `setMeshTrianglesWithEdges`, needs to check existing edges in order to avoid duplications, which is the reason why the API call is so slow. ``` meshID = getMeshID(); setMeshVertices(...); for t in trianges setMeshTriangleWithEdges(...); ``` **Describe the solution you propose.** Implement the following approach for connectivity: * Elements are defined by vertices only. * Sub-elements such as the edges of a triangle are implicitly defined via the triangle. * Update the API implementation to still allow constructing triangles from edgeIDs. Example Triangle Strip of 6 vertices and 4 triangles: Scenario | #Triangles | #Edges --- | --- | --- All edges | 4 | 12 Unique edges | 4 | 9 Implicit edges | 4 | 0 ## Pros: * Easier to use API with faster calls for setting connectivity information. * No handling of edge IDs in adapters. * Faster internals as we don't need to handle edges * Lower memory consumption as meshes won't contain implicit primitives. This will be very noticeable once we implement voluminous primitives such as tetrahedra. * Lower memory consumption as we don't need special maps to access unique vertices of triangles via the edges any more. * Lower initialization times as the meshes will be smaller, so there will be less to communicate. * This would make all tests that require meshes with connectivity as well as integration tests way cleaner. ## Cons: * The repartitioning will filter out complete primitives including all implicit ones. We could implement an algorithm that generates leftover edges, but this will need some thought. * This will make primitive ownership #963 more difficult to implement as primitives may not exist in every participant (1Triangle vs 1 Edge due to filtered vertex). This will impact scaled-consistent mappings as well as the watch-integral for parallel participants. * The nearest-projection needs to be redesigned. Problem is that an _implied_ edge of a triangle would be the better choice, but the current implementation would choose the best _explicitly_ defined edge instead. All implied edges need to be indexed for a correct implementation. **Example:** Defined primitives are triangle ABC and edge AD. Point E will project outside of the triangle ABC and find a suitable projection to AD. So, E will be projected to AD, even though the implicit edge AB would be the more suitable choice. ![bitmap](https://user-images.githubusercontent.com/13552216/146365428-e9cda696-c0fb-48f1-81aa-a7cf2f249129.png) **Describe alternatives you've considered** Leave it as it is. **Additional context** * Volume coupling requiring tetrahedra #468 * Primitive ownership #963 * Parallel support for scaled-consistent mappings #962 * Overweight mesh primitives #483
True
Define primitives in terms of vertices - **Please describe the problem you are trying to solve.** Currently, triangles are defined in terms of edges an edges are defined in terms of vertices. This makes defining triangles in adapter codes and integration tests tedious: ``` meshID = getMeshID(); setMeshVertices(...); for e in edges: setMeshEdge(...); for t in trianges setMeshTriangle(...); ``` This will become extremely tedious once we support tetrahedral elements. For consistency, they would need to be defined using triangles. As the `mesh::Triangle` is implemented this way as well, we need to still create `mesh::Edge`s for each triangle. Meaning `setMeshTrianglesWithEdges`, needs to check existing edges in order to avoid duplications, which is the reason why the API call is so slow. ``` meshID = getMeshID(); setMeshVertices(...); for t in trianges setMeshTriangleWithEdges(...); ``` **Describe the solution you propose.** Implement the following approach for connectivity: * Elements are defined by vertices only. * Sub-elements such as the edges of a triangle are implicitly defined via the triangle. * Update the API implementation to still allow constructing triangles from edgeIDs. Example Triangle Strip of 6 vertices and 4 triangles: Scenario | #Triangles | #Edges --- | --- | --- All edges | 4 | 12 Unique edges | 4 | 9 Implicit edges | 4 | 0 ## Pros: * Easier to use API with faster calls for setting connectivity information. * No handling of edge IDs in adapters. * Faster internals as we don't need to handle edges * Lower memory consumption as meshes won't contain implicit primitives. This will be very noticeable once we implement voluminous primitives such as tetrahedra. * Lower memory consumption as we don't need special maps to access unique vertices of triangles via the edges any more. * Lower initialization times as the meshes will be smaller, so there will be less to communicate. * This would make all tests that require meshes with connectivity as well as integration tests way cleaner. ## Cons: * The repartitioning will filter out complete primitives including all implicit ones. We could implement an algorithm that generates leftover edges, but this will need some thought. * This will make primitive ownership #963 more difficult to implement as primitives may not exist in every participant (1Triangle vs 1 Edge due to filtered vertex). This will impact scaled-consistent mappings as well as the watch-integral for parallel participants. * The nearest-projection needs to be redesigned. Problem is that an _implied_ edge of a triangle would be the better choice, but the current implementation would choose the best _explicitly_ defined edge instead. All implied edges need to be indexed for a correct implementation. **Example:** Defined primitives are triangle ABC and edge AD. Point E will project outside of the triangle ABC and find a suitable projection to AD. So, E will be projected to AD, even though the implicit edge AB would be the more suitable choice. ![bitmap](https://user-images.githubusercontent.com/13552216/146365428-e9cda696-c0fb-48f1-81aa-a7cf2f249129.png) **Describe alternatives you've considered** Leave it as it is. **Additional context** * Volume coupling requiring tetrahedra #468 * Primitive ownership #963 * Parallel support for scaled-consistent mappings #962 * Overweight mesh primitives #483
main
define primitives in terms of vertices please describe the problem you are trying to solve currently triangles are defined in terms of edges an edges are defined in terms of vertices this makes defining triangles in adapter codes and integration tests tedious meshid getmeshid setmeshvertices for e in edges setmeshedge for t in trianges setmeshtriangle this will become extremely tedious once we support tetrahedral elements for consistency they would need to be defined using triangles as the mesh triangle is implemented this way as well we need to still create mesh edge s for each triangle meaning setmeshtriangleswithedges needs to check existing edges in order to avoid duplications which is the reason why the api call is so slow meshid getmeshid setmeshvertices for t in trianges setmeshtrianglewithedges describe the solution you propose implement the following approach for connectivity elements are defined by vertices only sub elements such as the edges of a triangle are implicitly defined via the triangle update the api implementation to still allow constructing triangles from edgeids example triangle strip of vertices and triangles scenario triangles edges all edges unique edges implicit edges pros easier to use api with faster calls for setting connectivity information no handling of edge ids in adapters faster internals as we don t need to handle edges lower memory consumption as meshes won t contain implicit primitives this will be very noticeable once we implement voluminous primitives such as tetrahedra lower memory consumption as we don t need special maps to access unique vertices of triangles via the edges any more lower initialization times as the meshes will be smaller so there will be less to communicate this would make all tests that require meshes with connectivity as well as integration tests way cleaner cons the repartitioning will filter out complete primitives including all implicit ones we could implement an algorithm that generates leftover edges but this will need some thought this will make primitive ownership more difficult to implement as primitives may not exist in every participant vs edge due to filtered vertex this will impact scaled consistent mappings as well as the watch integral for parallel participants the nearest projection needs to be redesigned problem is that an implied edge of a triangle would be the better choice but the current implementation would choose the best explicitly defined edge instead all implied edges need to be indexed for a correct implementation example defined primitives are triangle abc and edge ad point e will project outside of the triangle abc and find a suitable projection to ad so e will be projected to ad even though the implicit edge ab would be the more suitable choice describe alternatives you ve considered leave it as it is additional context volume coupling requiring tetrahedra primitive ownership parallel support for scaled consistent mappings overweight mesh primitives
1
4,417
22,740,492,328
IssuesEvent
2022-07-07 02:58:34
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
closed
[MAINTAIN] RESTfulSummarizedExperiment error load
maintain
<!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** <details> ``` * installing *source* package ‘tenXplore’ ... ** using staged installation ** R ** data ** inst ** byte-compile and prepare package for lazy loading Error: class "RESTfulSummarizedExperiment" is not exported by 'namespace:restfulSE' Execution halted ERROR: lazy loading failed for package ‘tenXplore’ * removing ‘/build/r-tenxplore/src/tenXplore’ * restoring previous ‘/build/r-tenxplore/src/tenXplore’ ``` </details> **Packages (please complete the following information):** - Package Name: [e.g. iqtree] **Description** Add any other context about the problem here.
True
[MAINTAIN] RESTfulSummarizedExperiment error load - <!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** <details> ``` * installing *source* package ‘tenXplore’ ... ** using staged installation ** R ** data ** inst ** byte-compile and prepare package for lazy loading Error: class "RESTfulSummarizedExperiment" is not exported by 'namespace:restfulSE' Execution halted ERROR: lazy loading failed for package ‘tenXplore’ * removing ‘/build/r-tenxplore/src/tenXplore’ * restoring previous ‘/build/r-tenxplore/src/tenXplore’ ``` </details> **Packages (please complete the following information):** - Package Name: [e.g. iqtree] **Description** Add any other context about the problem here.
main
restfulsummarizedexperiment error load please report the error of one package in one issue use multi issues to report multi bugs thanks log of the bug installing source package ‘tenxplore’ using staged installation r data inst byte compile and prepare package for lazy loading error class restfulsummarizedexperiment is not exported by namespace restfulse execution halted error lazy loading failed for package ‘tenxplore’ removing ‘ build r tenxplore src tenxplore’ restoring previous ‘ build r tenxplore src tenxplore’ packages please complete the following information package name description add any other context about the problem here
1
167,521
20,726,190,614
IssuesEvent
2022-03-14 02:22:00
tqa236/scala_exercism
https://api.github.com/repos/tqa236/scala_exercism
opened
CVE-2020-36188 (High) detected in jackson-databind-2.9.0.jar
security vulnerability
## CVE-2020-36188 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.9.0.jar</p> <p> Dependency Hierarchy: - sbt-coveralls-1.2.7.jar (Root Library) - jackson-module-scala_2.12-2.9.0.jar - :x: **jackson-databind-2.9.0.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188>CVE-2020-36188</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36188 (High) detected in jackson-databind-2.9.0.jar - ## CVE-2020-36188 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: /home/wss-scanner/.ivy2/cache/com.fasterxml.jackson.core/jackson-databind/bundles/jackson-databind-2.9.0.jar</p> <p> Dependency Hierarchy: - sbt-coveralls-1.2.7.jar (Root Library) - jackson-module-scala_2.12-2.9.0.jar - :x: **jackson-databind-2.9.0.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to com.newrelic.agent.deps.ch.qos.logback.core.db.JNDIConnectionSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36188>CVE-2020-36188</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2996">https://github.com/FasterXML/jackson-databind/issues/2996</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library home wss scanner cache com fasterxml jackson core jackson databind bundles jackson databind jar dependency hierarchy sbt coveralls jar root library jackson module scala jar x jackson databind jar vulnerable library found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com newrelic agent deps ch qos logback core db jndiconnectionsource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource
0
206,814
16,057,666,512
IssuesEvent
2021-04-23 08:03:36
dewarian/real-time-web-2021
https://api.github.com/repos/dewarian/real-time-web-2021
closed
Describe socket events
documentation
Feed back from numerous people is that the socket events aren't described, what they do or when they are called. # Solution - [x] Describe what a socket event is. - [x] Write markdown table for the socket events.
1.0
Describe socket events - Feed back from numerous people is that the socket events aren't described, what they do or when they are called. # Solution - [x] Describe what a socket event is. - [x] Write markdown table for the socket events.
non_main
describe socket events feed back from numerous people is that the socket events aren t described what they do or when they are called solution describe what a socket event is write markdown table for the socket events
0
2,815
10,082,420,222
IssuesEvent
2019-07-25 11:11:46
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
closed
Do use a proper Log category
Area: analyzer Area: maintainability feature
Often, loggers are initialized as following: ```c# ILog Log = LogManager.GetLogger(MethodBase.GetCurrentMethod().DeclaringType); ``` Unfortunately, that results in the situation that the logs contain no useful categories. Instead, developers try to control the log output via the different severities (Debug, Info, ...). As a result, the logs contain a lot of garbage. Hence, when analyzing log files, it is unclear which log statement belongs into which category. Therefore, we should file an issue if the loggers are initialized as above (via types).
True
Do use a proper Log category - Often, loggers are initialized as following: ```c# ILog Log = LogManager.GetLogger(MethodBase.GetCurrentMethod().DeclaringType); ``` Unfortunately, that results in the situation that the logs contain no useful categories. Instead, developers try to control the log output via the different severities (Debug, Info, ...). As a result, the logs contain a lot of garbage. Hence, when analyzing log files, it is unclear which log statement belongs into which category. Therefore, we should file an issue if the loggers are initialized as above (via types).
main
do use a proper log category often loggers are initialized as following c ilog log logmanager getlogger methodbase getcurrentmethod declaringtype unfortunately that results in the situation that the logs contain no useful categories instead developers try to control the log output via the different severities debug info as a result the logs contain a lot of garbage hence when analyzing log files it is unclear which log statement belongs into which category therefore we should file an issue if the loggers are initialized as above via types
1
764
4,364,165,033
IssuesEvent
2016-08-03 05:07:14
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_service: wrong logic or processing services/scale
bug_report cloud docker waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_service ##### ANSIBLE VERSION ``` ansible 2.1.0.0 ``` ##### CONFIGURATION nothing special ##### OS / ENVIRONMENT Centos 7, docker-1.11.2, docker-compose 1.7.1 ##### SUMMARY docker_service launches services, not mentioned in service parameter ##### STEPS TO REPRODUCE ``` [root@ba0 test]# cat docker-compose.yml version: "2" services: centos1: image: centos:centos7 command: sleep 1d centos2: image: centos:centos7 command: sleep 1d centos3: image: centos:centos7 command: sleep 1d ansible -m docker_service -c local localhost -a "project_src=/tmp/test/ services=centos2 scale={'centos1':2} state=present" ``` ##### EXPECTED RESULTS I expect that only 'centos2' service launched ##### ACTUAL RESULTS Both centos2 and scaled centos1 launched. ``` localhost | SUCCESS => { "ansible_facts": { "centos1": { "test_centos1_1": { "cmd": [ "sleep", "1d" ], "image": "centos:centos7", "labels": { "build-date": "2016-06-02", "com.docker.compose.config-hash": "0d0b17f36bed81ed3668740e739677660e8f65e393cb9c3f57c3d0887c16ecc0", "com.docker.compose.container-number": "1", "com.docker.compose.oneoff": "False", "com.docker.compose.project": "test", "com.docker.compose.service": "centos1", "com.docker.compose.version": "1.7.1", "license": "GPLv2", "name": "CentOS Base Image", "vendor": "CentOS" }, "networks": { "test_default": { "IPAddress": "172.20.0.3", "IPPrefixLen": 16, "aliases": [ "centos1", "cba992c086f6" ], "globalIPv6": "", "globalIPv6PrefixLen": 0, "links": null, "macAddress": "02:42:ac:14:00:03" } }, "state": { "running": true, "status": "running" } }, "test_centos1_2": { "cmd": [ "sleep", "1d" ], "image": "centos:centos7", "labels": { "build-date": "2016-06-02", "com.docker.compose.config-hash": "0d0b17f36bed81ed3668740e739677660e8f65e393cb9c3f57c3d0887c16ecc0", "com.docker.compose.container-number": "2", "com.docker.compose.oneoff": "False", "com.docker.compose.project": "test", "com.docker.compose.service": "centos1", "com.docker.compose.version": "1.7.1", "license": "GPLv2", "name": "CentOS Base Image", "vendor": "CentOS" }, "networks": { "test_default": { "IPAddress": "172.20.0.4", "IPPrefixLen": 16, "aliases": [ "30249f7e074e", "centos1" ], "globalIPv6": "", "globalIPv6PrefixLen": 0, "links": null, "macAddress": "02:42:ac:14:00:04" } }, "state": { "running": true, "status": "running" } } }, "centos2": { "test_centos2_1": { "cmd": [ "sleep", "1d" ], "image": "centos:centos7", "labels": { "build-date": "2016-06-02", "com.docker.compose.config-hash": "0d0b17f36bed81ed3668740e739677660e8f65e393cb9c3f57c3d0887c16ecc0", "com.docker.compose.container-number": "1", "com.docker.compose.oneoff": "False", "com.docker.compose.project": "test", "com.docker.compose.service": "centos2", "com.docker.compose.version": "1.7.1", "license": "GPLv2", "name": "CentOS Base Image", "vendor": "CentOS" }, "networks": { "test_default": { "IPAddress": "172.20.0.2", "IPPrefixLen": 16, "aliases": [ "81c91cb6007b", "centos2" ], "globalIPv6": "", "globalIPv6PrefixLen": 0, "links": null, "macAddress": "02:42:ac:14:00:02" } }, "state": { "running": true, "status": "running" } } }, "centos3": {} }, "changed": true } ```
True
docker_service: wrong logic or processing services/scale - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_service ##### ANSIBLE VERSION ``` ansible 2.1.0.0 ``` ##### CONFIGURATION nothing special ##### OS / ENVIRONMENT Centos 7, docker-1.11.2, docker-compose 1.7.1 ##### SUMMARY docker_service launches services, not mentioned in service parameter ##### STEPS TO REPRODUCE ``` [root@ba0 test]# cat docker-compose.yml version: "2" services: centos1: image: centos:centos7 command: sleep 1d centos2: image: centos:centos7 command: sleep 1d centos3: image: centos:centos7 command: sleep 1d ansible -m docker_service -c local localhost -a "project_src=/tmp/test/ services=centos2 scale={'centos1':2} state=present" ``` ##### EXPECTED RESULTS I expect that only 'centos2' service launched ##### ACTUAL RESULTS Both centos2 and scaled centos1 launched. ``` localhost | SUCCESS => { "ansible_facts": { "centos1": { "test_centos1_1": { "cmd": [ "sleep", "1d" ], "image": "centos:centos7", "labels": { "build-date": "2016-06-02", "com.docker.compose.config-hash": "0d0b17f36bed81ed3668740e739677660e8f65e393cb9c3f57c3d0887c16ecc0", "com.docker.compose.container-number": "1", "com.docker.compose.oneoff": "False", "com.docker.compose.project": "test", "com.docker.compose.service": "centos1", "com.docker.compose.version": "1.7.1", "license": "GPLv2", "name": "CentOS Base Image", "vendor": "CentOS" }, "networks": { "test_default": { "IPAddress": "172.20.0.3", "IPPrefixLen": 16, "aliases": [ "centos1", "cba992c086f6" ], "globalIPv6": "", "globalIPv6PrefixLen": 0, "links": null, "macAddress": "02:42:ac:14:00:03" } }, "state": { "running": true, "status": "running" } }, "test_centos1_2": { "cmd": [ "sleep", "1d" ], "image": "centos:centos7", "labels": { "build-date": "2016-06-02", "com.docker.compose.config-hash": "0d0b17f36bed81ed3668740e739677660e8f65e393cb9c3f57c3d0887c16ecc0", "com.docker.compose.container-number": "2", "com.docker.compose.oneoff": "False", "com.docker.compose.project": "test", "com.docker.compose.service": "centos1", "com.docker.compose.version": "1.7.1", "license": "GPLv2", "name": "CentOS Base Image", "vendor": "CentOS" }, "networks": { "test_default": { "IPAddress": "172.20.0.4", "IPPrefixLen": 16, "aliases": [ "30249f7e074e", "centos1" ], "globalIPv6": "", "globalIPv6PrefixLen": 0, "links": null, "macAddress": "02:42:ac:14:00:04" } }, "state": { "running": true, "status": "running" } } }, "centos2": { "test_centos2_1": { "cmd": [ "sleep", "1d" ], "image": "centos:centos7", "labels": { "build-date": "2016-06-02", "com.docker.compose.config-hash": "0d0b17f36bed81ed3668740e739677660e8f65e393cb9c3f57c3d0887c16ecc0", "com.docker.compose.container-number": "1", "com.docker.compose.oneoff": "False", "com.docker.compose.project": "test", "com.docker.compose.service": "centos2", "com.docker.compose.version": "1.7.1", "license": "GPLv2", "name": "CentOS Base Image", "vendor": "CentOS" }, "networks": { "test_default": { "IPAddress": "172.20.0.2", "IPPrefixLen": 16, "aliases": [ "81c91cb6007b", "centos2" ], "globalIPv6": "", "globalIPv6PrefixLen": 0, "links": null, "macAddress": "02:42:ac:14:00:02" } }, "state": { "running": true, "status": "running" } } }, "centos3": {} }, "changed": true } ```
main
docker service wrong logic or processing services scale issue type bug report component name docker service ansible version ansible configuration nothing special os environment centos docker docker compose summary docker service launches services not mentioned in service parameter steps to reproduce cat docker compose yml version services image centos command sleep image centos command sleep image centos command sleep ansible m docker service c local localhost a project src tmp test services scale state present expected results i expect that only service launched actual results both and scaled launched localhost success ansible facts test cmd sleep image centos labels build date com docker compose config hash com docker compose container number com docker compose oneoff false com docker compose project test com docker compose service com docker compose version license name centos base image vendor centos networks test default ipaddress ipprefixlen aliases links null macaddress ac state running true status running test cmd sleep image centos labels build date com docker compose config hash com docker compose container number com docker compose oneoff false com docker compose project test com docker compose service com docker compose version license name centos base image vendor centos networks test default ipaddress ipprefixlen aliases links null macaddress ac state running true status running test cmd sleep image centos labels build date com docker compose config hash com docker compose container number com docker compose oneoff false com docker compose project test com docker compose service com docker compose version license name centos base image vendor centos networks test default ipaddress ipprefixlen aliases links null macaddress ac state running true status running changed true
1
3,755
15,788,513,092
IssuesEvent
2021-04-01 20:55:13
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
`DataTable[size="short"]` breaks alignment of `Select`s arrow icon:
component: data-table component: select status: waiting for maintainer response 💬 type: bug 🐛
## What package(s) are you using? - [ ] `carbon-components` - [x] `carbon-components-react` ## Detailed description > Is this issue related to a specific component? - `DataTable` - `Select` > Please create a reduced test case in CodeSandbox https://codesandbox.io/s/eager-taussig-lfolm?file=/src/index.js ## Additional information Offending style (works fine without it): ```css .bx--data-table--short .bx--select__arrow { top: 0.875rem; } ```
True
`DataTable[size="short"]` breaks alignment of `Select`s arrow icon: - ## What package(s) are you using? - [ ] `carbon-components` - [x] `carbon-components-react` ## Detailed description > Is this issue related to a specific component? - `DataTable` - `Select` > Please create a reduced test case in CodeSandbox https://codesandbox.io/s/eager-taussig-lfolm?file=/src/index.js ## Additional information Offending style (works fine without it): ```css .bx--data-table--short .bx--select__arrow { top: 0.875rem; } ```
main
datatable breaks alignment of select s arrow icon what package s are you using carbon components carbon components react detailed description is this issue related to a specific component datatable select please create a reduced test case in codesandbox additional information offending style works fine without it css bx data table short bx select arrow top
1
50,544
7,610,588,468
IssuesEvent
2018-05-01 09:14:18
ipfs/ipfs-cluster
https://api.github.com/repos/ipfs/ipfs-cluster
closed
docs: clarify how to start a multi-node cluster
documentation in progress
Just going to list out a few of the issues I hit and where I felt the docs fell down in assisting with troubleshooting them. The first and main issue that I encountered was getting the ipfs-cluster-service nodes to be able to communicate with each other. This was partly due to me having the incorrect expectations about how the cluster would go about forming, basically, if they had the same `CLUSTER_SECRET` they would auto-discover each other similar to ipfs nodes, when within the same network, which they were in this case. Things that I found confusing about the cluster creation process: - `ipfs-cluster-service init` cmd creates config directory and service.json file. There appears to be no way of providing a service.json file to `init`, which when using a docker container means you have to run the container so it runs `init` and then stopping the container and editing `service.json` and then starting the container again. (Note: I was able to do this by mounting named volumes). - The chicken-and-egg aspect of starting a cluster isn't very clear in the documentation. I found this really confusing due to the aforementioned false expectation about member auto-discovery based on the shared `CLUSTER_SECRET`. To clarify what I mean by 'chicken-and-egg', the requirement to start one node then take that nodes multiaddress and add it to service.json of the other nodes as the bootstrap address (if doing dynamic cluster membership) The other thing is how the documentation is split over several documents and it doesn't always make sense why some are split and some are not.
1.0
docs: clarify how to start a multi-node cluster - Just going to list out a few of the issues I hit and where I felt the docs fell down in assisting with troubleshooting them. The first and main issue that I encountered was getting the ipfs-cluster-service nodes to be able to communicate with each other. This was partly due to me having the incorrect expectations about how the cluster would go about forming, basically, if they had the same `CLUSTER_SECRET` they would auto-discover each other similar to ipfs nodes, when within the same network, which they were in this case. Things that I found confusing about the cluster creation process: - `ipfs-cluster-service init` cmd creates config directory and service.json file. There appears to be no way of providing a service.json file to `init`, which when using a docker container means you have to run the container so it runs `init` and then stopping the container and editing `service.json` and then starting the container again. (Note: I was able to do this by mounting named volumes). - The chicken-and-egg aspect of starting a cluster isn't very clear in the documentation. I found this really confusing due to the aforementioned false expectation about member auto-discovery based on the shared `CLUSTER_SECRET`. To clarify what I mean by 'chicken-and-egg', the requirement to start one node then take that nodes multiaddress and add it to service.json of the other nodes as the bootstrap address (if doing dynamic cluster membership) The other thing is how the documentation is split over several documents and it doesn't always make sense why some are split and some are not.
non_main
docs clarify how to start a multi node cluster just going to list out a few of the issues i hit and where i felt the docs fell down in assisting with troubleshooting them the first and main issue that i encountered was getting the ipfs cluster service nodes to be able to communicate with each other this was partly due to me having the incorrect expectations about how the cluster would go about forming basically if they had the same cluster secret they would auto discover each other similar to ipfs nodes when within the same network which they were in this case things that i found confusing about the cluster creation process ipfs cluster service init cmd creates config directory and service json file there appears to be no way of providing a service json file to init which when using a docker container means you have to run the container so it runs init and then stopping the container and editing service json and then starting the container again note i was able to do this by mounting named volumes the chicken and egg aspect of starting a cluster isn t very clear in the documentation i found this really confusing due to the aforementioned false expectation about member auto discovery based on the shared cluster secret to clarify what i mean by chicken and egg the requirement to start one node then take that nodes multiaddress and add it to service json of the other nodes as the bootstrap address if doing dynamic cluster membership the other thing is how the documentation is split over several documents and it doesn t always make sense why some are split and some are not
0
5,448
27,267,322,663
IssuesEvent
2023-02-22 19:08:07
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
The overflowMenu doesn't pop up if using states and renderIcon
type: bug 🐛 status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬 status: needs reproduction
## What package(s) are you using? - [ ] `carbon-components` - [x] `carbon-components-react` ## Detailed description When using overflowMenu with state change, the popup menu will be closed automatically when changing state (perhaps because of re-rendering?) so I could not see the popup menu. We also found that `onClose()` will be called twice when the popup menu closes. The other finding is if we don’t use `renderIcon` , the behavior looks normal. > Is this issue related to a specific component? OverflowMenu > What did you expect to happen? What happened instead? What would you like to > see changed? The overflowMenu with renderIcon can display the menu items when changing states. > What browser are you working in? Chrome > What version of the Carbon Design System are you using? > What offering/product do you work on? Any pressing ship or release dates we > should be aware of? ## Steps to reproduce the issue 1. From the sample code below in CodeSanbox, click the menu, and no menu item displayed. > Please create a reduced test case in CodeSandbox https://codesandbox.io/s/festive-cohen-shp8y ## Additional information - Screenshots or code - Notes
True
The overflowMenu doesn't pop up if using states and renderIcon - ## What package(s) are you using? - [ ] `carbon-components` - [x] `carbon-components-react` ## Detailed description When using overflowMenu with state change, the popup menu will be closed automatically when changing state (perhaps because of re-rendering?) so I could not see the popup menu. We also found that `onClose()` will be called twice when the popup menu closes. The other finding is if we don’t use `renderIcon` , the behavior looks normal. > Is this issue related to a specific component? OverflowMenu > What did you expect to happen? What happened instead? What would you like to > see changed? The overflowMenu with renderIcon can display the menu items when changing states. > What browser are you working in? Chrome > What version of the Carbon Design System are you using? > What offering/product do you work on? Any pressing ship or release dates we > should be aware of? ## Steps to reproduce the issue 1. From the sample code below in CodeSanbox, click the menu, and no menu item displayed. > Please create a reduced test case in CodeSandbox https://codesandbox.io/s/festive-cohen-shp8y ## Additional information - Screenshots or code - Notes
main
the overflowmenu doesn t pop up if using states and rendericon what package s are you using carbon components carbon components react detailed description when using overflowmenu with state change the popup menu will be closed automatically when changing state perhaps because of re rendering so i could not see the popup menu we also found that onclose will be called twice when the popup menu closes the other finding is if we don’t use rendericon the behavior looks normal is this issue related to a specific component overflowmenu what did you expect to happen what happened instead what would you like to see changed the overflowmenu with rendericon can display the menu items when changing states what browser are you working in chrome what version of the carbon design system are you using what offering product do you work on any pressing ship or release dates we should be aware of steps to reproduce the issue from the sample code below in codesanbox click the menu and no menu item displayed please create a reduced test case in codesandbox additional information screenshots or code notes
1
247,171
7,904,593,644
IssuesEvent
2018-07-02 05:43:08
fossasia/susper.com
https://api.github.com/repos/fossasia/susper.com
closed
Speech to text not working on Chrome browser on mobile
Priority: High bug
Speech to text does not work on the Chrome browser on mobile. A message appears "Please check your microphone".
1.0
Speech to text not working on Chrome browser on mobile - Speech to text does not work on the Chrome browser on mobile. A message appears "Please check your microphone".
non_main
speech to text not working on chrome browser on mobile speech to text does not work on the chrome browser on mobile a message appears please check your microphone
0
8,954
3,812,241,654
IssuesEvent
2016-03-27 12:36:35
vasl-developers/vasl
https://api.github.com/repos/vasl-developers/vasl
opened
Range missing from LOS near overlays
2 - Code bug
Previously, whether the LOS checking was disabled or not, drawing a string would always give you the range (handy if the LOS is long and you want a quick confirmation of the range, even if you already know the LOS is clear). Now, if the LOS crosses any part of an overlay, the range is either reported as "0" or just left blank.
1.0
Range missing from LOS near overlays - Previously, whether the LOS checking was disabled or not, drawing a string would always give you the range (handy if the LOS is long and you want a quick confirmation of the range, even if you already know the LOS is clear). Now, if the LOS crosses any part of an overlay, the range is either reported as "0" or just left blank.
non_main
range missing from los near overlays previously whether the los checking was disabled or not drawing a string would always give you the range handy if the los is long and you want a quick confirmation of the range even if you already know the los is clear now if the los crosses any part of an overlay the range is either reported as or just left blank
0
2,004
6,718,163,347
IssuesEvent
2017-10-15 09:03:00
Kristinita/Erics-Green-Room
https://api.github.com/repos/Kristinita/Erics-Green-Room
closed
[Feature request] Неправильные похожие ответы
enhancement need-maintainer
### 1. Запрос Если будет реализован запрос **#2**,          неплохо было бы, если составителю пакета можно было бы указывать неправильные ответы. Если игрок введёт какой-либо из них → приём ответов от пользователя прекращается → очко за данный ответ пользователю в любом случае не будет присвоено → начинается следующий вопрос. ### 2. Примеры желаемого поведения Предложу реализацию при помощи нового типа метаданных — `*-error`. #### 1. Стереобат ```markdown https://i.imgur.com/lS3NCHb.png Верхняя ступень, на которой размещались колонны*Стилобат*-info-Не путать со стереобатом*-error-Стереобат*-proof-200 ``` Если игрок введёт «стереобат», 1. Не должно высветиться, что это похожий вариант; 1. Каким-то образом будет показано, что ответ неправильный, и сразу (после времени на комментарий) начнётся следующий вопрос. #### 2. Швейцария Упрощённый пример, если не понятно. ```markdown Берн — столица ЭТОЙ страны*Швейцария*-error-Швеция ``` Вариант «Швеция» не должен быть показан, как близкий, его следует засчитывать как неправильный. ### 3. Аргументация Цель «близких вариантов» в комнате — дать пользователю возможность исправиться в случае опечаток и не более того. При этом: 1. Нежелательны случайные совпадения. В случае со Швецией и Швейцарией игрок, думавший о совсем другой стране, случайно угадывает ответ. 1. Многие вопросы как раз и рассчитаны на то, чтобы игрок различал понятия, к примеру: `стилобат` и `стереобат`, `портик` и `портал`, `Нордкин` и `Нордкап`. А сейчас игрок вбивает одно из понятий → видит, что ответ близкий → вбивает второе, правильное. Это снижает эффективность прокачки понятий в комнатах. Спасибо.
True
[Feature request] Неправильные похожие ответы - ### 1. Запрос Если будет реализован запрос **#2**,          неплохо было бы, если составителю пакета можно было бы указывать неправильные ответы. Если игрок введёт какой-либо из них → приём ответов от пользователя прекращается → очко за данный ответ пользователю в любом случае не будет присвоено → начинается следующий вопрос. ### 2. Примеры желаемого поведения Предложу реализацию при помощи нового типа метаданных — `*-error`. #### 1. Стереобат ```markdown https://i.imgur.com/lS3NCHb.png Верхняя ступень, на которой размещались колонны*Стилобат*-info-Не путать со стереобатом*-error-Стереобат*-proof-200 ``` Если игрок введёт «стереобат», 1. Не должно высветиться, что это похожий вариант; 1. Каким-то образом будет показано, что ответ неправильный, и сразу (после времени на комментарий) начнётся следующий вопрос. #### 2. Швейцария Упрощённый пример, если не понятно. ```markdown Берн — столица ЭТОЙ страны*Швейцария*-error-Швеция ``` Вариант «Швеция» не должен быть показан, как близкий, его следует засчитывать как неправильный. ### 3. Аргументация Цель «близких вариантов» в комнате — дать пользователю возможность исправиться в случае опечаток и не более того. При этом: 1. Нежелательны случайные совпадения. В случае со Швецией и Швейцарией игрок, думавший о совсем другой стране, случайно угадывает ответ. 1. Многие вопросы как раз и рассчитаны на то, чтобы игрок различал понятия, к примеру: `стилобат` и `стереобат`, `портик` и `портал`, `Нордкин` и `Нордкап`. А сейчас игрок вбивает одно из понятий → видит, что ответ близкий → вбивает второе, правильное. Это снижает эффективность прокачки понятий в комнатах. Спасибо.
main
неправильные похожие ответы запрос если будет реализован запрос          неплохо было бы если составителю пакета можно было бы указывать неправильные ответы если игрок введёт какой либо из них → приём ответов от пользователя прекращается → очко за данный ответ пользователю в любом случае не будет присвоено → начинается следующий вопрос примеры желаемого поведения предложу реализацию при помощи нового типа метаданных — error стереобат markdown верхняя ступень на которой размещались колонны стилобат info не путать со стереобатом error стереобат proof если игрок введёт «стереобат» не должно высветиться что это похожий вариант каким то образом будет показано что ответ неправильный и сразу после времени на комментарий начнётся следующий вопрос швейцария упрощённый пример если не понятно markdown берн — столица этой страны швейцария error швеция вариант «швеция» не должен быть показан как близкий его следует засчитывать как неправильный аргументация цель «близких вариантов» в комнате — дать пользователю возможность исправиться в случае опечаток и не более того при этом нежелательны случайные совпадения в случае со швецией и швейцарией игрок думавший о совсем другой стране случайно угадывает ответ многие вопросы как раз и рассчитаны на то чтобы игрок различал понятия к примеру стилобат и стереобат портик и портал нордкин и нордкап а сейчас игрок вбивает одно из понятий → видит что ответ близкий → вбивает второе правильное это снижает эффективность прокачки понятий в комнатах спасибо
1
321,556
23,860,972,725
IssuesEvent
2022-09-07 07:00:13
cerus/maps
https://api.github.com/repos/cerus/maps
closed
Add documentation
documentation good first issue hacktoberfest
The code is lacking some documentation. This is a nice starter issue if you're kind of new to Open Source or Spigot or just want to work on something simple. (You don't have to comment everything if you don't want to, but it would be nice if you document at least one module.) Comment if you want to work on this. If you have any questions or need some help feel free to contact me.
1.0
Add documentation - The code is lacking some documentation. This is a nice starter issue if you're kind of new to Open Source or Spigot or just want to work on something simple. (You don't have to comment everything if you don't want to, but it would be nice if you document at least one module.) Comment if you want to work on this. If you have any questions or need some help feel free to contact me.
non_main
add documentation the code is lacking some documentation this is a nice starter issue if you re kind of new to open source or spigot or just want to work on something simple you don t have to comment everything if you don t want to but it would be nice if you document at least one module comment if you want to work on this if you have any questions or need some help feel free to contact me
0
1,845
6,577,385,115
IssuesEvent
2017-09-12 00:32:29
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Yum module should support disableexcludes option
affects_2.0 feature_idea waiting_on_maintainer
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME yum module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 ``` ##### SUMMARY Core module yum, should support more yum options. Disabling excluded packages in repository (--disableexcludes) is a handy one.
True
Yum module should support disableexcludes option - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME yum module ##### ANSIBLE VERSION ``` ansible 2.0.1.0 ``` ##### SUMMARY Core module yum, should support more yum options. Disabling excluded packages in repository (--disableexcludes) is a handy one.
main
yum module should support disableexcludes option issue type feature idea component name yum module ansible version ansible summary core module yum should support more yum options disabling excluded packages in repository disableexcludes is a handy one
1
1,483
6,416,006,777
IssuesEvent
2017-08-08 14:00:21
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
vca_vapp errors in setting the computer_name to vm_name for API 5.5, 5.1, 1.5
affects_2.1 bug_report cloud vmware waiting_on_maintainer
I have two cloud providers I use, one that uses API version 5.6, and one that uses 5.5 This module works great on version API version 5.6 but when I attempt to create a vapp on the vcd with API version 5.5 (of 5.1, or 1.5 which are the supported APIs on that cloud provider) -- I receive an error. ........ "failed": true, "msg": "Error in setting the computer_name to vm_name"} I've run into this before with pyvcloud and basically just implemented my own 'renaming' method https://gist.github.com/lasko/9ce419800d115e33a8c2 I'm not sure how this could be incorporated into this module, but it sure would save me a lot of headache.
True
vca_vapp errors in setting the computer_name to vm_name for API 5.5, 5.1, 1.5 - I have two cloud providers I use, one that uses API version 5.6, and one that uses 5.5 This module works great on version API version 5.6 but when I attempt to create a vapp on the vcd with API version 5.5 (of 5.1, or 1.5 which are the supported APIs on that cloud provider) -- I receive an error. ........ "failed": true, "msg": "Error in setting the computer_name to vm_name"} I've run into this before with pyvcloud and basically just implemented my own 'renaming' method https://gist.github.com/lasko/9ce419800d115e33a8c2 I'm not sure how this could be incorporated into this module, but it sure would save me a lot of headache.
main
vca vapp errors in setting the computer name to vm name for api i have two cloud providers i use one that uses api version and one that uses this module works great on version api version but when i attempt to create a vapp on the vcd with api version of or which are the supported apis on that cloud provider i receive an error failed true msg error in setting the computer name to vm name i ve run into this before with pyvcloud and basically just implemented my own renaming method i m not sure how this could be incorporated into this module but it sure would save me a lot of headache
1
3,498
13,651,412,525
IssuesEvent
2020-09-27 01:01:56
amyjko/faculty
https://api.github.com/repos/amyjko/faculty
closed
Commitments: maintainability
maintainability
I quickly hacked the promises page together without much mind for maintainability. Go through, comment, refactor for reuse.
True
Commitments: maintainability - I quickly hacked the promises page together without much mind for maintainability. Go through, comment, refactor for reuse.
main
commitments maintainability i quickly hacked the promises page together without much mind for maintainability go through comment refactor for reuse
1
190,850
15,257,079,905
IssuesEvent
2021-02-20 23:15:49
Donovanphenry/Hide_And_Seq
https://api.github.com/repos/Donovanphenry/Hide_And_Seq
closed
Download dataset in notebook instance
documentation
Instead of manually downloading the dataset and uploading it for every notebook instance, can we instead download the dataset directly to the notebook?
1.0
Download dataset in notebook instance - Instead of manually downloading the dataset and uploading it for every notebook instance, can we instead download the dataset directly to the notebook?
non_main
download dataset in notebook instance instead of manually downloading the dataset and uploading it for every notebook instance can we instead download the dataset directly to the notebook
0
3,861
17,030,269,069
IssuesEvent
2021-07-04 12:18:33
skku-npc/skku-coding-platform
https://api.github.com/repos/skku-npc/skku-coding-platform
closed
Replace tar-simditor-markdown
frontend maintain security
## Issue 더 이상 지원되지 않는 패키지인 `tar-simditor-markdown` 대체하기 (`pages/admin/components/Simditor.vue`) ## Why? ### Cross-Site Scripting (XSS) Issue 구버전의 jquery를 사용해 XSS Issue 발생 [XSS(Cross-Site Scripting) 이란?](https://noirstar.tistory.com/266) [NPM Advisory Info](https://www.npmjs.com/advisories/328) ### No Longer Maintained 2017년 이후로 업데이트 X https://github.com/itargaryen/tar-simditor ![image](https://user-images.githubusercontent.com/19747913/112109192-98fe0f00-8bf4-11eb-9daf-4ab3b29d2182.png)
True
Replace tar-simditor-markdown - ## Issue 더 이상 지원되지 않는 패키지인 `tar-simditor-markdown` 대체하기 (`pages/admin/components/Simditor.vue`) ## Why? ### Cross-Site Scripting (XSS) Issue 구버전의 jquery를 사용해 XSS Issue 발생 [XSS(Cross-Site Scripting) 이란?](https://noirstar.tistory.com/266) [NPM Advisory Info](https://www.npmjs.com/advisories/328) ### No Longer Maintained 2017년 이후로 업데이트 X https://github.com/itargaryen/tar-simditor ![image](https://user-images.githubusercontent.com/19747913/112109192-98fe0f00-8bf4-11eb-9daf-4ab3b29d2182.png)
main
replace tar simditor markdown issue 더 이상 지원되지 않는 패키지인 tar simditor markdown 대체하기 pages admin components simditor vue why cross site scripting xss issue 구버전의 jquery를 사용해 xss issue 발생 no longer maintained 이후로 업데이트 x
1
44,292
12,101,439,576
IssuesEvent
2020-04-20 15:13:00
codesmithtools/Templates
https://api.github.com/repos/codesmithtools/Templates
closed
Invalid Cast (check your mapping for property type mismatches); setter of System.Object
Framework-NHibernate Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. Attempt to load entity from the database 2. 3. What is the expected output? What do you see instead? I'm expecting a loaded entity returned, instead I'm getting an error. What version of the product are you using? PLINQO.NH-v1.0.0, CodeSmith Generator Professional 5.3.4 Please provide any additional information below. See the error generated here: http://screencast.com/t/OGrS9grh I have checked the mapping files and I cannot find a problem. I'm able to save entities to the database without a problem, it's just loading them back. I've pasted the mapping file below, against this SQL server table: http://screencast.com/t/KJYOcViOsc <?xml version="1.0" encoding="utf-8" ?> <hibernate-mapping xmlns="urn:nhibernate-mapping-2.2" namespace="Domain.Data" assembly="Domain"> <class name="Company" table="[dbo].[Company]" > <id name="CompanyId" column="[CompanyId]" type="Int32" > <generator class="native" /> </id> <property name="Name" column="[Name]" type="String" not-null="true" length="100" /> <property name="CreatedDate" column="[CreatedDate]" type="DateTime" not-null="true" /> <property name="CreatedBy" column="[CreatedBy]" type="Int32" not-null="true" /> <property name="LastUpdatedDate" column="[LastUpdatedDate]" type="DateTime" not-null="true" /> <property name="LastUpdatedBy" column="[LastUpdatedBy]" type="Int32" not-null="true" /> <bag name="UserList" inverse="true" cascade="all-delete-orphan" lazy="true" > <key> <column name="[CompanyId]" /> </key> <one-to-many class="User" /> </bag> </class> </hibernate-mapping> Here's the full stacktrace of the error: at NHibernate.Tuple.Entity.PocoEntityTuplizer.SetPropertyValuesWithOptimizer(Object entity, Object[] values) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Tuple\Entity\PocoEntityTuplizer.cs:line 308 at NHibernate.Tuple.Entity.PocoEntityTuplizer.SetPropertyValues(Object entity, Object[] values) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Tuple\Entity\PocoEntityTuplizer.cs:line 289 at NHibernate.Persister.Entity.AbstractEntityPersister.SetPropertyValues(Object obj, Object[] values, EntityMode entityMode) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Persister\Entity\AbstractEntityPersister.cs:line 3858 at NHibernate.Engine.TwoPhaseLoad.InitializeEntity(Object entity, Boolean readOnly, ISessionImplementor session, PreLoadEvent preLoadEvent, PostLoadEvent postLoadEvent) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Engine\TwoPhaseLoad.cs:line 97 at NHibernate.Loader.Loader.InitializeEntitiesAndCollections(IList hydratedObjects, Object resultSetId, ISessionImplementor session, Boolean readOnly) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 603 at NHibernate.Loader.Loader.DoQuery(ISessionImplementor session, QueryParameters queryParameters, Boolean returnProxies) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 472 at NHibernate.Loader.Loader.DoQueryAndInitializeNonLazyCollections(ISessionImplementor session, QueryParameters queryParameters, Boolean returnProxies) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 243 at NHibernate.Loader.Loader.DoList(ISessionImplementor session, QueryParameters queryParameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 1712 at NHibernate.Loader.Loader.ListIgnoreQueryCache(ISessionImplementor session, QueryParameters queryParameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 1601 at NHibernate.Loader.Loader.List(ISessionImplementor session, QueryParameters queryParameters, ISet`1 querySpaces, IType[] resultTypes) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 1595 at NHibernate.Hql.Ast.ANTLR.Loader.QueryLoader.List(ISessionImplementor session, QueryParameters queryParameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Hql\Ast\ANTLR\Loader\QueryLoader.cs:line 300 at NHibernate.Hql.Ast.ANTLR.QueryTranslatorImpl.List(ISessionImplementor session, QueryParameters queryParameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Hql\Ast\ANTLR\QueryTranslatorImpl.cs:line 111 at NHibernate.Engine.Query.HQLQueryPlan.PerformList(QueryParameters queryParameters, ISessionImplementor session, IList results) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Engine\Query\HQLQueryPlan.cs:line 105 at NHibernate.Impl.SessionImpl.List(IQueryExpression queryExpression, QueryParameters queryParameters, IList results) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Impl\SessionImpl.cs:line 661 at NHibernate.Impl.AbstractSessionImpl.List(IQueryExpression queryExpression, QueryParameters parameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Impl\AbstractSessionImpl.cs:line 92 at NHibernate.Impl.ExpressionQueryImpl.List() in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Impl\ExpressionQueryImpl.cs:line 63 at NHibernate.Linq.DefaultQueryProvider.ExecuteQuery(NhLinqExpression nhLinqExpression, IQuery query, NhLinqExpression nhQuery) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Linq\DefaultQueryProvider.cs:line 103 at NHibernate.Linq.DefaultQueryProvider.Execute(Expression expression) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Linq\DefaultQueryProvider.cs:line 35 at NHibernate.Linq.DefaultQueryProvider.Execute[TResult](Expression expression) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Linq\DefaultQueryProvider.cs:line 40 at Remotion.Linq.QueryableBase`1.GetEnumerator() in c:\build\Remotion\working\Relinq\Core\QueryableBase.cs:line 132 at CodeSmith.Data.NHibernate.Table`1.GetEnumerator() in \PLINQO-NH\Source\CodeSmith.Data.NHibernate\Entities\Tables.cs:line 30 at System.Collections.Generic.List`1..ctor(IEnumerable`1 collection) at System.Linq.Enumerable.ToList[TSource](IEnumerable`1 source) at .Domain.Data.CompanyExtensions.GetAll(IQueryable`1 query) in \code\.Domain\Data\Queries\Company.generated.cs:line 486 at .Domain.Data.Company.GetAll(DataContext dbContext) in \code\.Domain\Data\Entities\Company.generated.cs:line 327 at .Domain.Data.Company.GetAll() in \code\.Domain\Data\Entities\Company.generated.cs:line 321 at .Domain.Test.Base_Tests.CompanyFixture.LoadAll() in \code\.Domain.Test\Base Tests\CompanyFixture.cs:line 186 ``` Original issue reported on code.google.com by `ben.emp...@gmail.com` on 1 Jul 2011 at 4:46
1.0
Invalid Cast (check your mapping for property type mismatches); setter of System.Object - ``` What steps will reproduce the problem? 1. Attempt to load entity from the database 2. 3. What is the expected output? What do you see instead? I'm expecting a loaded entity returned, instead I'm getting an error. What version of the product are you using? PLINQO.NH-v1.0.0, CodeSmith Generator Professional 5.3.4 Please provide any additional information below. See the error generated here: http://screencast.com/t/OGrS9grh I have checked the mapping files and I cannot find a problem. I'm able to save entities to the database without a problem, it's just loading them back. I've pasted the mapping file below, against this SQL server table: http://screencast.com/t/KJYOcViOsc <?xml version="1.0" encoding="utf-8" ?> <hibernate-mapping xmlns="urn:nhibernate-mapping-2.2" namespace="Domain.Data" assembly="Domain"> <class name="Company" table="[dbo].[Company]" > <id name="CompanyId" column="[CompanyId]" type="Int32" > <generator class="native" /> </id> <property name="Name" column="[Name]" type="String" not-null="true" length="100" /> <property name="CreatedDate" column="[CreatedDate]" type="DateTime" not-null="true" /> <property name="CreatedBy" column="[CreatedBy]" type="Int32" not-null="true" /> <property name="LastUpdatedDate" column="[LastUpdatedDate]" type="DateTime" not-null="true" /> <property name="LastUpdatedBy" column="[LastUpdatedBy]" type="Int32" not-null="true" /> <bag name="UserList" inverse="true" cascade="all-delete-orphan" lazy="true" > <key> <column name="[CompanyId]" /> </key> <one-to-many class="User" /> </bag> </class> </hibernate-mapping> Here's the full stacktrace of the error: at NHibernate.Tuple.Entity.PocoEntityTuplizer.SetPropertyValuesWithOptimizer(Object entity, Object[] values) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Tuple\Entity\PocoEntityTuplizer.cs:line 308 at NHibernate.Tuple.Entity.PocoEntityTuplizer.SetPropertyValues(Object entity, Object[] values) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Tuple\Entity\PocoEntityTuplizer.cs:line 289 at NHibernate.Persister.Entity.AbstractEntityPersister.SetPropertyValues(Object obj, Object[] values, EntityMode entityMode) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Persister\Entity\AbstractEntityPersister.cs:line 3858 at NHibernate.Engine.TwoPhaseLoad.InitializeEntity(Object entity, Boolean readOnly, ISessionImplementor session, PreLoadEvent preLoadEvent, PostLoadEvent postLoadEvent) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Engine\TwoPhaseLoad.cs:line 97 at NHibernate.Loader.Loader.InitializeEntitiesAndCollections(IList hydratedObjects, Object resultSetId, ISessionImplementor session, Boolean readOnly) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 603 at NHibernate.Loader.Loader.DoQuery(ISessionImplementor session, QueryParameters queryParameters, Boolean returnProxies) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 472 at NHibernate.Loader.Loader.DoQueryAndInitializeNonLazyCollections(ISessionImplementor session, QueryParameters queryParameters, Boolean returnProxies) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 243 at NHibernate.Loader.Loader.DoList(ISessionImplementor session, QueryParameters queryParameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 1712 at NHibernate.Loader.Loader.ListIgnoreQueryCache(ISessionImplementor session, QueryParameters queryParameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 1601 at NHibernate.Loader.Loader.List(ISessionImplementor session, QueryParameters queryParameters, ISet`1 querySpaces, IType[] resultTypes) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Loader\Loader.cs:line 1595 at NHibernate.Hql.Ast.ANTLR.Loader.QueryLoader.List(ISessionImplementor session, QueryParameters queryParameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Hql\Ast\ANTLR\Loader\QueryLoader.cs:line 300 at NHibernate.Hql.Ast.ANTLR.QueryTranslatorImpl.List(ISessionImplementor session, QueryParameters queryParameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Hql\Ast\ANTLR\QueryTranslatorImpl.cs:line 111 at NHibernate.Engine.Query.HQLQueryPlan.PerformList(QueryParameters queryParameters, ISessionImplementor session, IList results) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Engine\Query\HQLQueryPlan.cs:line 105 at NHibernate.Impl.SessionImpl.List(IQueryExpression queryExpression, QueryParameters queryParameters, IList results) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Impl\SessionImpl.cs:line 661 at NHibernate.Impl.AbstractSessionImpl.List(IQueryExpression queryExpression, QueryParameters parameters) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Impl\AbstractSessionImpl.cs:line 92 at NHibernate.Impl.ExpressionQueryImpl.List() in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Impl\ExpressionQueryImpl.cs:line 63 at NHibernate.Linq.DefaultQueryProvider.ExecuteQuery(NhLinqExpression nhLinqExpression, IQuery query, NhLinqExpression nhQuery) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Linq\DefaultQueryProvider.cs:line 103 at NHibernate.Linq.DefaultQueryProvider.Execute(Expression expression) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Linq\DefaultQueryProvider.cs:line 35 at NHibernate.Linq.DefaultQueryProvider.Execute[TResult](Expression expression) in E:\Downloads\NHibernate-3.2.0.Beta1-src\src\NHibernate\Linq\DefaultQueryProvider.cs:line 40 at Remotion.Linq.QueryableBase`1.GetEnumerator() in c:\build\Remotion\working\Relinq\Core\QueryableBase.cs:line 132 at CodeSmith.Data.NHibernate.Table`1.GetEnumerator() in \PLINQO-NH\Source\CodeSmith.Data.NHibernate\Entities\Tables.cs:line 30 at System.Collections.Generic.List`1..ctor(IEnumerable`1 collection) at System.Linq.Enumerable.ToList[TSource](IEnumerable`1 source) at .Domain.Data.CompanyExtensions.GetAll(IQueryable`1 query) in \code\.Domain\Data\Queries\Company.generated.cs:line 486 at .Domain.Data.Company.GetAll(DataContext dbContext) in \code\.Domain\Data\Entities\Company.generated.cs:line 327 at .Domain.Data.Company.GetAll() in \code\.Domain\Data\Entities\Company.generated.cs:line 321 at .Domain.Test.Base_Tests.CompanyFixture.LoadAll() in \code\.Domain.Test\Base Tests\CompanyFixture.cs:line 186 ``` Original issue reported on code.google.com by `ben.emp...@gmail.com` on 1 Jul 2011 at 4:46
non_main
invalid cast check your mapping for property type mismatches setter of system object what steps will reproduce the problem attempt to load entity from the database what is the expected output what do you see instead i m expecting a loaded entity returned instead i m getting an error what version of the product are you using plinqo nh codesmith generator professional please provide any additional information below see the error generated here i have checked the mapping files and i cannot find a problem i m able to save entities to the database without a problem it s just loading them back i ve pasted the mapping file below against this sql server table hibernate mapping xmlns urn nhibernate mapping namespace domain data assembly domain here s the full stacktrace of the error at nhibernate tuple entity pocoentitytuplizer setpropertyvalueswithoptimizer object entity object values in e downloads nhibernate src src nhibernate tuple entity pocoentitytuplizer cs line at nhibernate tuple entity pocoentitytuplizer setpropertyvalues object entity object values in e downloads nhibernate src src nhibernate tuple entity pocoentitytuplizer cs line at nhibernate persister entity abstractentitypersister setpropertyvalues object obj object values entitymode entitymode in e downloads nhibernate src src nhibernate persister entity abstractentitypersister cs line at nhibernate engine twophaseload initializeentity object entity boolean readonly isessionimplementor session preloadevent preloadevent postloadevent postloadevent in e downloads nhibernate src src nhibernate engine twophaseload cs line at nhibernate loader loader initializeentitiesandcollections ilist hydratedobjects object resultsetid isessionimplementor session boolean readonly in e downloads nhibernate src src nhibernate loader loader cs line at nhibernate loader loader doquery isessionimplementor session queryparameters queryparameters boolean returnproxies in e downloads nhibernate src src nhibernate loader loader cs line at nhibernate loader loader doqueryandinitializenonlazycollections isessionimplementor session queryparameters queryparameters boolean returnproxies in e downloads nhibernate src src nhibernate loader loader cs line at nhibernate loader loader dolist isessionimplementor session queryparameters queryparameters in e downloads nhibernate src src nhibernate loader loader cs line at nhibernate loader loader listignorequerycache isessionimplementor session queryparameters queryparameters in e downloads nhibernate src src nhibernate loader loader cs line at nhibernate loader loader list isessionimplementor session queryparameters queryparameters iset queryspaces itype resulttypes in e downloads nhibernate src src nhibernate loader loader cs line at nhibernate hql ast antlr loader queryloader list isessionimplementor session queryparameters queryparameters in e downloads nhibernate src src nhibernate hql ast antlr loader queryloader cs line at nhibernate hql ast antlr querytranslatorimpl list isessionimplementor session queryparameters queryparameters in e downloads nhibernate src src nhibernate hql ast antlr querytranslatorimpl cs line at nhibernate engine query hqlqueryplan performlist queryparameters queryparameters isessionimplementor session ilist results in e downloads nhibernate src src nhibernate engine query hqlqueryplan cs line at nhibernate impl sessionimpl list iqueryexpression queryexpression queryparameters queryparameters ilist results in e downloads nhibernate src src nhibernate impl sessionimpl cs line at nhibernate impl abstractsessionimpl list iqueryexpression queryexpression queryparameters parameters in e downloads nhibernate src src nhibernate impl abstractsessionimpl cs line at nhibernate impl expressionqueryimpl list in e downloads nhibernate src src nhibernate impl expressionqueryimpl cs line at nhibernate linq defaultqueryprovider executequery nhlinqexpression nhlinqexpression iquery query nhlinqexpression nhquery in e downloads nhibernate src src nhibernate linq defaultqueryprovider cs line at nhibernate linq defaultqueryprovider execute expression expression in e downloads nhibernate src src nhibernate linq defaultqueryprovider cs line at nhibernate linq defaultqueryprovider execute expression expression in e downloads nhibernate src src nhibernate linq defaultqueryprovider cs line at remotion linq queryablebase getenumerator in c build remotion working relinq core queryablebase cs line at codesmith data nhibernate table getenumerator in plinqo nh source codesmith data nhibernate entities tables cs line at system collections generic list ctor ienumerable collection at system linq enumerable tolist ienumerable source at domain data companyextensions getall iqueryable query in code domain data queries company generated cs line at domain data company getall datacontext dbcontext in code domain data entities company generated cs line at domain data company getall in code domain data entities company generated cs line at domain test base tests companyfixture loadall in code domain test base tests companyfixture cs line original issue reported on code google com by ben emp gmail com on jul at
0
809,859
30,215,013,531
IssuesEvent
2023-07-05 15:02:05
AxonFramework/.github
https://api.github.com/repos/AxonFramework/.github
closed
Explain all projects under the Axon Framework organization
Priority 1: Must Status: In Progress Type: Feature
The main `README.md` should reference all the repositories present in the [Axon Framework](https://github.com/AxonFramework/) GitHub organization. Furthermore, a reference to the documentation and code samples should also be included. Thus, this includes: - Axon Framework - Documentation - Samples - Extensions - BOM - Intellij Plugin
1.0
Explain all projects under the Axon Framework organization - The main `README.md` should reference all the repositories present in the [Axon Framework](https://github.com/AxonFramework/) GitHub organization. Furthermore, a reference to the documentation and code samples should also be included. Thus, this includes: - Axon Framework - Documentation - Samples - Extensions - BOM - Intellij Plugin
non_main
explain all projects under the axon framework organization the main readme md should reference all the repositories present in the github organization furthermore a reference to the documentation and code samples should also be included thus this includes axon framework documentation samples extensions bom intellij plugin
0
2,158
7,503,896,061
IssuesEvent
2018-04-10 00:27:15
Microsoft/DXUT
https://api.github.com/repos/Microsoft/DXUT
opened
Retire VS 2013 projects
maintainence
Planning to drop support for the VS 2013 toolset, which involves removing the following projects: ``DXUT_2013.vcxproj`` ``DXUT_2013_Win10.vcxproj`` > Note that I will also remove ``Windows10SDKVS13_x64.props`` and ``Windows10SDKVS13_x86.props``
True
Retire VS 2013 projects - Planning to drop support for the VS 2013 toolset, which involves removing the following projects: ``DXUT_2013.vcxproj`` ``DXUT_2013_Win10.vcxproj`` > Note that I will also remove ``Windows10SDKVS13_x64.props`` and ``Windows10SDKVS13_x86.props``
main
retire vs projects planning to drop support for the vs toolset which involves removing the following projects dxut vcxproj dxut vcxproj note that i will also remove props and props
1
326,782
28,018,532,834
IssuesEvent
2023-03-28 02:07:56
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
closed
There is no loading icon in the 'Search' box of the 'Manage Access' dialog when running a search
:heavy_check_mark: merged 🧪 testing :beetle: regression :computer: windows :gear: adls gen2
**Storage Explorer Version**: 1.29.0-dev **Build Number**: 20230327.1 **Branch**: main **Platform/OS**: Windows 10 **Architecture**: ia32 **How Found**: From running test cases **Regression From**: Previous release (1.28.1) ## Steps to Reproduce ## 1. Expand one ADLS Gen2 storage account -> Blob Containers. 2. Right click one blob container -> Click 'Manage Access Control Lists...'. 3. Click 'Add' button on the dialog -> Click 'Search'. 4. Check whether there is a loading icon in the 'Search' box. ## Expected Experience ## There is a loading icon in the 'Search' box. ![image](https://user-images.githubusercontent.com/41351993/227874497-39ca296c-5c91-44a4-ab63-d66dd7ba2c9d.png) ## Actual Experience ## There is no loading icon in the 'Search' box. ![image](https://user-images.githubusercontent.com/41351993/227874484-dc8eb931-029b-41a5-b541-2cceee159a48.png) ## Additional Context ## This issue does not reproduce on Linux Ubuntu 22.04/MacOS Ventura 13.2.1 (Apple M1 Pro).
1.0
There is no loading icon in the 'Search' box of the 'Manage Access' dialog when running a search - **Storage Explorer Version**: 1.29.0-dev **Build Number**: 20230327.1 **Branch**: main **Platform/OS**: Windows 10 **Architecture**: ia32 **How Found**: From running test cases **Regression From**: Previous release (1.28.1) ## Steps to Reproduce ## 1. Expand one ADLS Gen2 storage account -> Blob Containers. 2. Right click one blob container -> Click 'Manage Access Control Lists...'. 3. Click 'Add' button on the dialog -> Click 'Search'. 4. Check whether there is a loading icon in the 'Search' box. ## Expected Experience ## There is a loading icon in the 'Search' box. ![image](https://user-images.githubusercontent.com/41351993/227874497-39ca296c-5c91-44a4-ab63-d66dd7ba2c9d.png) ## Actual Experience ## There is no loading icon in the 'Search' box. ![image](https://user-images.githubusercontent.com/41351993/227874484-dc8eb931-029b-41a5-b541-2cceee159a48.png) ## Additional Context ## This issue does not reproduce on Linux Ubuntu 22.04/MacOS Ventura 13.2.1 (Apple M1 Pro).
non_main
there is no loading icon in the search box of the manage access dialog when running a search storage explorer version dev build number branch main platform os windows architecture how found from running test cases regression from previous release steps to reproduce expand one adls storage account blob containers right click one blob container click manage access control lists click add button on the dialog click search check whether there is a loading icon in the search box expected experience there is a loading icon in the search box actual experience there is no loading icon in the search box additional context this issue does not reproduce on linux ubuntu macos ventura apple pro
0
793,165
27,984,998,233
IssuesEvent
2023-03-26 15:44:38
literalpie/storybook-framework-qwik
https://api.github.com/repos/literalpie/storybook-framework-qwik
closed
[BUG] Get "The Qwik Component was not invocated correctly" after upgrading @builder.io/qwik to v0.24.0
high-priority
**Reproduce** Download the content in `storybook-framework-qwik/packages/[qwik-lib](https://github.com/literalpie/storybook-framework-qwik/tree/main/packages/qwik-lib)`. Run `yarn install` and then `yarn storybook` and Storybook will work. Go to `package.json`, change `@builder.io/qwik` to `^0.24.0`, run `yarn install` and then `yarn storybook` and Storybook will open with the following error: <img width="1652" alt="Skärmavbild 2023-03-25 kl 13 21 58" src="https://user-images.githubusercontent.com/9481000/227717656-3890ad50-d904-4818-a0dd-ad715d9fe232.png"> Issue was caused by this [release](https://github.com/BuilderIO/qwik/releases/tag/v0.24.0).
1.0
[BUG] Get "The Qwik Component was not invocated correctly" after upgrading @builder.io/qwik to v0.24.0 - **Reproduce** Download the content in `storybook-framework-qwik/packages/[qwik-lib](https://github.com/literalpie/storybook-framework-qwik/tree/main/packages/qwik-lib)`. Run `yarn install` and then `yarn storybook` and Storybook will work. Go to `package.json`, change `@builder.io/qwik` to `^0.24.0`, run `yarn install` and then `yarn storybook` and Storybook will open with the following error: <img width="1652" alt="Skärmavbild 2023-03-25 kl 13 21 58" src="https://user-images.githubusercontent.com/9481000/227717656-3890ad50-d904-4818-a0dd-ad715d9fe232.png"> Issue was caused by this [release](https://github.com/BuilderIO/qwik/releases/tag/v0.24.0).
non_main
get the qwik component was not invocated correctly after upgrading builder io qwik to reproduce download the content in storybook framework qwik packages run yarn install and then yarn storybook and storybook will work go to package json change builder io qwik to run yarn install and then yarn storybook and storybook will open with the following error img width alt skärmavbild kl src issue was caused by this
0
4,767
24,547,280,635
IssuesEvent
2022-10-12 09:45:30
gemarkode/Codebase
https://api.github.com/repos/gemarkode/Codebase
opened
Attention Hacktoberfest Users
question hacktoberfest hacktoberfest2022 maintainers
This is a repo that we created for Open Source purposes, not just for Hacktoberfest only. <strong>Please explain the meaning of the code you sent in the PR message. Like This [Pull #36](https://github.com/gemarkode/Codebase/pull/36) </strong> Please Read [CONTRIBUTING.md](https://github.com/gemarkode/Codebase/blob/main/CONTRIBUTING.md) (coming soon) Before Contributing.
True
Attention Hacktoberfest Users - This is a repo that we created for Open Source purposes, not just for Hacktoberfest only. <strong>Please explain the meaning of the code you sent in the PR message. Like This [Pull #36](https://github.com/gemarkode/Codebase/pull/36) </strong> Please Read [CONTRIBUTING.md](https://github.com/gemarkode/Codebase/blob/main/CONTRIBUTING.md) (coming soon) Before Contributing.
main
attention hacktoberfest users this is a repo that we created for open source purposes not just for hacktoberfest only please explain the meaning of the code you sent in the pr message like this please read coming soon before contributing
1
45,483
18,732,498,171
IssuesEvent
2021-11-04 00:19:23
Azure/azure-sdk-for-go
https://api.github.com/repos/Azure/azure-sdk-for-go
closed
[service-bus] ServiceBusAdministrationClient
Service Bus Client feature-request
Track 2 SDKs should have a top level client that manages communication with the Service Bus ATOM endpoint (basically - a similar endpoint to what you can accomplish with ARM, but using a Service Bus connection string). Work: - [ ] Adding in Create/Delete/Get support for queues: https://github.com/Azure/azure-sdk-for-go/pull/15790 - [ ] GetQueueRuntimeProperties, ListQueues, UpdateQueues - [ ] Subscriptions - [ ] #16045 - [ ] Topics - [ ] GetNamespaceProperties Issues that I've seen that we should address as part of this work: - [#229: Blank error returned when using a key without Manage permission](https://github.com/Azure/azure-service-bus-go/issues/229) Future considerations: - Generate using autorest and the swagger file for the endpoint: [link](https://raw.githubusercontent.com/Azure/azure-rest-api-specs/sb_dataplane_namespace/specification/servicebus/data-plane/servicebus-swagger.json). @chlowell gave this a shot but it looks like our autorest impl for Go isn't quite up to the task yet. For now we'll ship using the existing EntityManager, with a unified surface in `AdminClient`.
1.0
[service-bus] ServiceBusAdministrationClient - Track 2 SDKs should have a top level client that manages communication with the Service Bus ATOM endpoint (basically - a similar endpoint to what you can accomplish with ARM, but using a Service Bus connection string). Work: - [ ] Adding in Create/Delete/Get support for queues: https://github.com/Azure/azure-sdk-for-go/pull/15790 - [ ] GetQueueRuntimeProperties, ListQueues, UpdateQueues - [ ] Subscriptions - [ ] #16045 - [ ] Topics - [ ] GetNamespaceProperties Issues that I've seen that we should address as part of this work: - [#229: Blank error returned when using a key without Manage permission](https://github.com/Azure/azure-service-bus-go/issues/229) Future considerations: - Generate using autorest and the swagger file for the endpoint: [link](https://raw.githubusercontent.com/Azure/azure-rest-api-specs/sb_dataplane_namespace/specification/servicebus/data-plane/servicebus-swagger.json). @chlowell gave this a shot but it looks like our autorest impl for Go isn't quite up to the task yet. For now we'll ship using the existing EntityManager, with a unified surface in `AdminClient`.
non_main
servicebusadministrationclient track sdks should have a top level client that manages communication with the service bus atom endpoint basically a similar endpoint to what you can accomplish with arm but using a service bus connection string work adding in create delete get support for queues getqueueruntimeproperties listqueues updatequeues subscriptions topics getnamespaceproperties issues that i ve seen that we should address as part of this work future considerations generate using autorest and the swagger file for the endpoint chlowell gave this a shot but it looks like our autorest impl for go isn t quite up to the task yet for now we ll ship using the existing entitymanager with a unified surface in adminclient
0
4,876
25,021,274,273
IssuesEvent
2022-11-04 01:04:35
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Sort list of schemas
type: enhancement work: backend work: frontend status: ready restricted: maintainers
On the Database Page, the schemas appear in random order. ![image](https://user-images.githubusercontent.com/42411/199863005-2cc007c5-3224-489b-86de-80d924010833.png) I think they should be listed alphabetically. I'm not sure whether it would be better to handle this on the back end or the front end. Any thoughts @mathemancer @rajatvijay?
True
Sort list of schemas - On the Database Page, the schemas appear in random order. ![image](https://user-images.githubusercontent.com/42411/199863005-2cc007c5-3224-489b-86de-80d924010833.png) I think they should be listed alphabetically. I'm not sure whether it would be better to handle this on the back end or the front end. Any thoughts @mathemancer @rajatvijay?
main
sort list of schemas on the database page the schemas appear in random order i think they should be listed alphabetically i m not sure whether it would be better to handle this on the back end or the front end any thoughts mathemancer rajatvijay
1
42,778
11,062,102,271
IssuesEvent
2019-12-11 08:51:54
eclipse/sumo
https://api.github.com/repos/eclipse/sumo
closed
outdated nightly build on windows machines
c:build
``` 2019-11-29 01:00:00.908000: Running msvc12 build using python 2.7.11 (v2.7.11:6d1b6a68f775, Dec 5 2015, 20:40:30) [MSC v.1500 64 bit (AMD64)]. fatal: unable to access 'https://github.com/eclipse/sumo/': Failed to connect to github.com port 443: Connection refused ``` ``` 2019-11-29 02:00:51.616000: Running msvc12 build using python 2.7.11 (v2.7.11:6d1b6a68f775, Dec 5 2015, 20:40:30) [MSC v.1500 64 bit (AMD64)]. fatal: unable to access 'https://github.com/eclipse/sumo/': Failed to connect to github.com port 443: Connection refused 2019-11-29 02:00:58.819000: Cleaning directory of Visual Studio 12 2013 Win64. ``` linux build is up-to-date at least
1.0
outdated nightly build on windows machines - ``` 2019-11-29 01:00:00.908000: Running msvc12 build using python 2.7.11 (v2.7.11:6d1b6a68f775, Dec 5 2015, 20:40:30) [MSC v.1500 64 bit (AMD64)]. fatal: unable to access 'https://github.com/eclipse/sumo/': Failed to connect to github.com port 443: Connection refused ``` ``` 2019-11-29 02:00:51.616000: Running msvc12 build using python 2.7.11 (v2.7.11:6d1b6a68f775, Dec 5 2015, 20:40:30) [MSC v.1500 64 bit (AMD64)]. fatal: unable to access 'https://github.com/eclipse/sumo/': Failed to connect to github.com port 443: Connection refused 2019-11-29 02:00:58.819000: Cleaning directory of Visual Studio 12 2013 Win64. ``` linux build is up-to-date at least
non_main
outdated nightly build on windows machines running build using python dec fatal unable to access failed to connect to github com port connection refused running build using python dec fatal unable to access failed to connect to github com port connection refused cleaning directory of visual studio linux build is up to date at least
0
558
4,007,192,723
IssuesEvent
2016-05-12 17:16:58
duckduckgo/zeroclickinfo-spice
https://api.github.com/repos/duckduckgo/zeroclickinfo-spice
closed
Advanced Calculator: Compute integral gives wrong result
External Maintainer Input Requested
I've tried the sample search for "compute integral" but the result is a little bit off: ![duckduckgo_compute_intergral](https://cloud.githubusercontent.com/assets/8273519/15090925/0accde62-1441-11e6-88f1-7f48a0a48a3f.png) Same with other powers apparently. The result for x^a always seems to return x^(a+2)/(a+2). ------ IA Page: http://duck.co/ia/view/symbolab [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @levaly
True
Advanced Calculator: Compute integral gives wrong result - I've tried the sample search for "compute integral" but the result is a little bit off: ![duckduckgo_compute_intergral](https://cloud.githubusercontent.com/assets/8273519/15090925/0accde62-1441-11e6-88f1-7f48a0a48a3f.png) Same with other powers apparently. The result for x^a always seems to return x^(a+2)/(a+2). ------ IA Page: http://duck.co/ia/view/symbolab [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @levaly
main
advanced calculator compute integral gives wrong result i ve tried the sample search for compute integral but the result is a little bit off same with other powers apparently the result for x a always seems to return x a a ia page levaly
1
556,903
16,494,524,298
IssuesEvent
2021-05-25 08:51:37
wso2/product-apim
https://api.github.com/repos/wso2/product-apim
opened
APIM-3.2 - Unable to update the endpoint security configs of the migrated APIs
Priority/Normal Type/Bug
### Description: Users are unable to update the endpoint configurations of the migrated APIs. When a user tries to update endpoint configs from the publisher UI, the changes will not persist and does not reflect in the synapse config file or registry properly. This issue does not occurs for the newly created APIs. For migrated APIs, it happens under the following conditions. - Migrated version needs to be APIM-3.2 - Endpoint security has to be configured in the API prior to the migration. Endpoint configuration section in the registry of a migrated API is malformed as follows. - Migrated API ` { "endpoint_type":"http", "failOver":"False", "endpoint_security":{ "password":"", "type":"BASIC", "production":{ "password":"123456789", "customParameters":"null", "type":"BASIC", "enabled":true, "username":"apigateway" }, "username":"apigateway" }, "production_endpoints":{ "url":"https://e2e2-api-consent.coms-dev.aws.de.pri.o2.com" } } ` - Newly created APIs `{ "endpoint_type":"http", "endpoint_security":{ "production":{ "password":"abc123", "tokenUrl":null, "clientId":null, "clientSecret":null, "customParameters":"{}", "additionalProperties":{ }, "type":"BASIC", "grantType":null, "enabled":true, "uniqueIdentifier":null, "username":"test" } }, "production_endpoints":{ "template_not_supported":false, "url":"https:\/\/run.mocky.io\/v3\/80579333-5f98-4104-a97c-c9f024a2e0a9" } } ` ### Steps to reproduce: 1. Migrate APIM to APIM-3.2. 2. Choose an API which already configured endpoint configurations in the previous APIM version. 3. Try to update the endpoint configurations from the publisher UI. 4. The changes will not be reflect in the synapse config file or registry. ### Affected Product Version: APIM-3.2
1.0
APIM-3.2 - Unable to update the endpoint security configs of the migrated APIs - ### Description: Users are unable to update the endpoint configurations of the migrated APIs. When a user tries to update endpoint configs from the publisher UI, the changes will not persist and does not reflect in the synapse config file or registry properly. This issue does not occurs for the newly created APIs. For migrated APIs, it happens under the following conditions. - Migrated version needs to be APIM-3.2 - Endpoint security has to be configured in the API prior to the migration. Endpoint configuration section in the registry of a migrated API is malformed as follows. - Migrated API ` { "endpoint_type":"http", "failOver":"False", "endpoint_security":{ "password":"", "type":"BASIC", "production":{ "password":"123456789", "customParameters":"null", "type":"BASIC", "enabled":true, "username":"apigateway" }, "username":"apigateway" }, "production_endpoints":{ "url":"https://e2e2-api-consent.coms-dev.aws.de.pri.o2.com" } } ` - Newly created APIs `{ "endpoint_type":"http", "endpoint_security":{ "production":{ "password":"abc123", "tokenUrl":null, "clientId":null, "clientSecret":null, "customParameters":"{}", "additionalProperties":{ }, "type":"BASIC", "grantType":null, "enabled":true, "uniqueIdentifier":null, "username":"test" } }, "production_endpoints":{ "template_not_supported":false, "url":"https:\/\/run.mocky.io\/v3\/80579333-5f98-4104-a97c-c9f024a2e0a9" } } ` ### Steps to reproduce: 1. Migrate APIM to APIM-3.2. 2. Choose an API which already configured endpoint configurations in the previous APIM version. 3. Try to update the endpoint configurations from the publisher UI. 4. The changes will not be reflect in the synapse config file or registry. ### Affected Product Version: APIM-3.2
non_main
apim unable to update the endpoint security configs of the migrated apis description users are unable to update the endpoint configurations of the migrated apis when a user tries to update endpoint configs from the publisher ui the changes will not persist and does not reflect in the synapse config file or registry properly this issue does not occurs for the newly created apis for migrated apis it happens under the following conditions migrated version needs to be apim endpoint security has to be configured in the api prior to the migration endpoint configuration section in the registry of a migrated api is malformed as follows migrated api endpoint type http failover false endpoint security password type basic production password customparameters null type basic enabled true username apigateway username apigateway production endpoints url newly created apis endpoint type http endpoint security production password tokenurl null clientid null clientsecret null customparameters additionalproperties type basic granttype null enabled true uniqueidentifier null username test production endpoints template not supported false url https run mocky io steps to reproduce migrate apim to apim choose an api which already configured endpoint configurations in the previous apim version try to update the endpoint configurations from the publisher ui the changes will not be reflect in the synapse config file or registry affected product version apim
0
5,110
26,032,000,991
IssuesEvent
2022-12-21 22:27:41
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
sam init workflow for layers
type/feature stage/pm-review maintainer/need-response
<!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). --> ### Describe your idea/feature/enhancement The sam cli command ' sam init ' is a great workflow, and I think having an additional selection (besides just runtimes) would be for layers. For example: sam init Which template source would you like to use? 1 - AWS Quick Start Templates 2 - Custom Template Location Choice: 1 Which runtime would you like to use? 1 - nodejs12.x 2 - python3.8 3 - ruby2.7 4 - go1.x 5 - java11 6 - dotnetcore3.1 7 - nodejs10.x 8 - python3.7 9 - python3.6 10 - python2.7 11 - ruby2.5 12 - java8 13 - dotnetcore2.1 14 - dotnetcore2.0 15 - dotnetcore1.0 16 - layer <-- Suggested addition ### Proposal If we can get an idea of a starting point to hit the ground running on making this addition, I think following the similar processes for the runtime workflows could make this quick work. The normal recommendations for creating layers is to spin up an EC2 instance, or of course just making a new template specifically for layers; however I think having it as a dedicated workflow would greatly simplify and speed up layer creations. Things to consider: 1. Will this require any updates to the [SAM Spec](https://github.com/awslabs/serverless-application-model) I don't believe this is going to warrant an update to the SAM Spec as it's already in the spec; this is more along the lines of making it faster and more user friendly for users to create layers solely. In addition to layers, having a documented workflow for the community at large to create new/modify "Quick Start Templates" could be a beneficial side effect. One such idea that comes to mind is enabling X-Ray in a workflow, which can be a bit more hands on when you need to define the API Gateway resource. A workflow could make it bit quicker. ### Additional Details
True
sam init workflow for layers - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed). --> ### Describe your idea/feature/enhancement The sam cli command ' sam init ' is a great workflow, and I think having an additional selection (besides just runtimes) would be for layers. For example: sam init Which template source would you like to use? 1 - AWS Quick Start Templates 2 - Custom Template Location Choice: 1 Which runtime would you like to use? 1 - nodejs12.x 2 - python3.8 3 - ruby2.7 4 - go1.x 5 - java11 6 - dotnetcore3.1 7 - nodejs10.x 8 - python3.7 9 - python3.6 10 - python2.7 11 - ruby2.5 12 - java8 13 - dotnetcore2.1 14 - dotnetcore2.0 15 - dotnetcore1.0 16 - layer <-- Suggested addition ### Proposal If we can get an idea of a starting point to hit the ground running on making this addition, I think following the similar processes for the runtime workflows could make this quick work. The normal recommendations for creating layers is to spin up an EC2 instance, or of course just making a new template specifically for layers; however I think having it as a dedicated workflow would greatly simplify and speed up layer creations. Things to consider: 1. Will this require any updates to the [SAM Spec](https://github.com/awslabs/serverless-application-model) I don't believe this is going to warrant an update to the SAM Spec as it's already in the spec; this is more along the lines of making it faster and more user friendly for users to create layers solely. In addition to layers, having a documented workflow for the community at large to create new/modify "Quick Start Templates" could be a beneficial side effect. One such idea that comes to mind is enabling X-Ray in a workflow, which can be a bit more hands on when you need to define the API Gateway resource. A workflow could make it bit quicker. ### Additional Details
main
sam init workflow for layers describe your idea feature enhancement the sam cli command sam init is a great workflow and i think having an additional selection besides just runtimes would be for layers for example sam init which template source would you like to use aws quick start templates custom template location choice which runtime would you like to use x x x layer suggested addition proposal if we can get an idea of a starting point to hit the ground running on making this addition i think following the similar processes for the runtime workflows could make this quick work the normal recommendations for creating layers is to spin up an instance or of course just making a new template specifically for layers however i think having it as a dedicated workflow would greatly simplify and speed up layer creations things to consider will this require any updates to the i don t believe this is going to warrant an update to the sam spec as it s already in the spec this is more along the lines of making it faster and more user friendly for users to create layers solely in addition to layers having a documented workflow for the community at large to create new modify quick start templates could be a beneficial side effect one such idea that comes to mind is enabling x ray in a workflow which can be a bit more hands on when you need to define the api gateway resource a workflow could make it bit quicker additional details
1
5,523
27,615,419,503
IssuesEvent
2023-03-09 18:57:37
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
opened
[MAINTAIN] easycodeml and phylosuite License notice
maintain
# [**easycodeml**](https://github.com/BioEasy/EasyCodeML/issues/20) - Based on [this commit](https://github.com/BioEasy/EasyCodeML/commit/a655455558b75763cbf39ce4666fd7e6897824ad), they just change the LGPL license and don't release source code with LGPL - Based on [PAML](https://github.com/abacus-gene/paml) license, GPL, all binary and source code should be distributed as GPL, until [latest commit](https://github.com/BioEasy/EasyCodeML/commit/e55d61d795b2ae4f6cee494e7041d7a69a32f690), easycodeml doesn't distribute the all binary compiled from source code and source code of paml with GPL license # [**phylosuite**](https://github.com/dongzhang0725/PhyloSuite/issues/59) - Based on the GPL license phylosuite followed, they should distribute the source, however, no source code is distributed between 1.2.2 and 1.2.3, see [comparison](https://github.com/dongzhang0725/PhyloSuite/compare/1.2.2...1.2.3), the author just uploaded the zipped binary file.
True
[MAINTAIN] easycodeml and phylosuite License notice - # [**easycodeml**](https://github.com/BioEasy/EasyCodeML/issues/20) - Based on [this commit](https://github.com/BioEasy/EasyCodeML/commit/a655455558b75763cbf39ce4666fd7e6897824ad), they just change the LGPL license and don't release source code with LGPL - Based on [PAML](https://github.com/abacus-gene/paml) license, GPL, all binary and source code should be distributed as GPL, until [latest commit](https://github.com/BioEasy/EasyCodeML/commit/e55d61d795b2ae4f6cee494e7041d7a69a32f690), easycodeml doesn't distribute the all binary compiled from source code and source code of paml with GPL license # [**phylosuite**](https://github.com/dongzhang0725/PhyloSuite/issues/59) - Based on the GPL license phylosuite followed, they should distribute the source, however, no source code is distributed between 1.2.2 and 1.2.3, see [comparison](https://github.com/dongzhang0725/PhyloSuite/compare/1.2.2...1.2.3), the author just uploaded the zipped binary file.
main
easycodeml and phylosuite license notice based on they just change the lgpl license and don t release source code with lgpl based on license gpl all binary and source code should be distributed as gpl until easycodeml doesn t distribute the all binary compiled from source code and source code of paml with gpl license based on the gpl license phylosuite followed they should distribute the source however no source code is distributed between and see the author just uploaded the zipped binary file
1
5,114
26,038,420,114
IssuesEvent
2022-12-22 08:11:05
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
Replace TopologyListeners with Atomix Pub-Sub mechanics
kind/toil scope/broker area/maintainability
Currently in the DeploymentDistributor and SubscriptionCommandSender we use the Topology, more specific a TopologyPartitionListener to determine the leader for a partition. If we could remove these listener usage we would be able to remove the complete Topology stuff from the broker, since the topology request is now directly answered on the gateway with the help of atomix. To remove the partition listener we could use atomix Pub-Sub mechanics. We could create for each partition a subscription topic. On the subscription command sending we will then send direct to the specific partition subscription and for deployment distribution we need only one subscription, where the first node distributes this deployment over the specific topic.
True
Replace TopologyListeners with Atomix Pub-Sub mechanics - Currently in the DeploymentDistributor and SubscriptionCommandSender we use the Topology, more specific a TopologyPartitionListener to determine the leader for a partition. If we could remove these listener usage we would be able to remove the complete Topology stuff from the broker, since the topology request is now directly answered on the gateway with the help of atomix. To remove the partition listener we could use atomix Pub-Sub mechanics. We could create for each partition a subscription topic. On the subscription command sending we will then send direct to the specific partition subscription and for deployment distribution we need only one subscription, where the first node distributes this deployment over the specific topic.
main
replace topologylisteners with atomix pub sub mechanics currently in the deploymentdistributor and subscriptioncommandsender we use the topology more specific a topologypartitionlistener to determine the leader for a partition if we could remove these listener usage we would be able to remove the complete topology stuff from the broker since the topology request is now directly answered on the gateway with the help of atomix to remove the partition listener we could use atomix pub sub mechanics we could create for each partition a subscription topic on the subscription command sending we will then send direct to the specific partition subscription and for deployment distribution we need only one subscription where the first node distributes this deployment over the specific topic
1
305,716
26,406,363,503
IssuesEvent
2023-01-13 08:23:40
kyma-project/kyma
https://api.github.com/repos/kyma-project/kyma
closed
Commerce Mock Tests timeout on multiple pipelines
kind/bug kind/failing-test
<!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** <!-- Provide a clear and concise description of the problem. Describe where it appears, when it occurred, and what it affects. --> The commerce mock tests, which are part of the fast integration tests time out multiple times on multiple pipelines, which seem to be one of the reasons for flakiness. Some of the test runs can be found [here](https://storage.googleapis.com/kyma-prow-logs/pr-logs/pull/kyma-project_control-plane/1519/pre-main-control-plane-reconciler-e2e/1511086525536800768/build-log.txt) and [here](https://storage.googleapis.com/kyma-prow-logs/pr-logs/pull/kyma-project_control-plane/1519/pre-main-control-plane-reconciler-e2e/1511328231528075264/build-log.txt). <!-- Provide relevant technical details such as the Kubernetes version, the cluster name and provider, the Kyma version, the browser name and version, or the operating system. --> **Expected result** <!-- Describe what you expect to happen. --> The commerce mock is initialised properly and the tests succeed (do not timeout). **Actual result** <!-- Describe what happens instead. --> The tests time out with the following error: ``` 1) Executing Standard Testsuite: CommerceMock Tests: CommerceMock test fixture should be ready: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 2) Executing Standard Testsuite: CommerceMock Tests: in-cluster event should be delivered (structured and binary mode): Error: Fetching published event responded with error: AssertionError: The same event id expected in the result: expected { podName: 'lastorder-vsf92-6dbfb4b797-jphdd' } to have nested property 'event.id' at convertAxiosError (utils/index.js:97:12) at /home/prow/go/src/github.com/kyma-project/kyma/tests/fast-integration/test/fixtures/commerce-mock/index.js:873:15 at processTicksAndRejections (internal/process/task_queues.js:95:5) at async checkInClusterEventDeliveryHelper (test/fixtures/commerce-mock/index.js:860:10) at async checkInClusterEventDelivery (test/fixtures/commerce-mock/index.js:825:3) at async Context.<anonymous> (test/1-commerce-mock.js:58:7) 3) Executing Standard Testsuite: CommerceMock Tests: function should be reachable through secured API Rule: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 4) Executing Standard Testsuite: CommerceMock Tests: order.created.v1 event should trigger the lastorder function: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 5) Executing Standard Testsuite: CommerceMock Tests: order.created.v1 cloud event in structured mode should trigger the lastorder function: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 6) Executing Standard Testsuite: CommerceMock Tests: order.created.v1 cloud event in binary mode should trigger the lastorder function: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 7) Executing Standard Testsuite: CommerceMock Tests: should add, update and delete a service: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) ``` **Steps to reproduce** <!-- List the steps to follow to reproduce the bug. Attach any files, links, code samples, or screenshots that could help in investigating the problem. --> **Troubleshooting** <!-- Describe the steps you have already taken to solve the issue. -->
1.0
Commerce Mock Tests timeout on multiple pipelines - <!-- Thank you for your contribution. Before you submit the issue: 1. Search open and closed issues for duplicates. 2. Read the contributing guidelines. --> **Description** <!-- Provide a clear and concise description of the problem. Describe where it appears, when it occurred, and what it affects. --> The commerce mock tests, which are part of the fast integration tests time out multiple times on multiple pipelines, which seem to be one of the reasons for flakiness. Some of the test runs can be found [here](https://storage.googleapis.com/kyma-prow-logs/pr-logs/pull/kyma-project_control-plane/1519/pre-main-control-plane-reconciler-e2e/1511086525536800768/build-log.txt) and [here](https://storage.googleapis.com/kyma-prow-logs/pr-logs/pull/kyma-project_control-plane/1519/pre-main-control-plane-reconciler-e2e/1511328231528075264/build-log.txt). <!-- Provide relevant technical details such as the Kubernetes version, the cluster name and provider, the Kyma version, the browser name and version, or the operating system. --> **Expected result** <!-- Describe what you expect to happen. --> The commerce mock is initialised properly and the tests succeed (do not timeout). **Actual result** <!-- Describe what happens instead. --> The tests time out with the following error: ``` 1) Executing Standard Testsuite: CommerceMock Tests: CommerceMock test fixture should be ready: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 2) Executing Standard Testsuite: CommerceMock Tests: in-cluster event should be delivered (structured and binary mode): Error: Fetching published event responded with error: AssertionError: The same event id expected in the result: expected { podName: 'lastorder-vsf92-6dbfb4b797-jphdd' } to have nested property 'event.id' at convertAxiosError (utils/index.js:97:12) at /home/prow/go/src/github.com/kyma-project/kyma/tests/fast-integration/test/fixtures/commerce-mock/index.js:873:15 at processTicksAndRejections (internal/process/task_queues.js:95:5) at async checkInClusterEventDeliveryHelper (test/fixtures/commerce-mock/index.js:860:10) at async checkInClusterEventDelivery (test/fixtures/commerce-mock/index.js:825:3) at async Context.<anonymous> (test/1-commerce-mock.js:58:7) 3) Executing Standard Testsuite: CommerceMock Tests: function should be reachable through secured API Rule: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 4) Executing Standard Testsuite: CommerceMock Tests: order.created.v1 event should trigger the lastorder function: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 5) Executing Standard Testsuite: CommerceMock Tests: order.created.v1 cloud event in structured mode should trigger the lastorder function: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 6) Executing Standard Testsuite: CommerceMock Tests: order.created.v1 cloud event in binary mode should trigger the lastorder function: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) 7) Executing Standard Testsuite: CommerceMock Tests: should add, update and delete a service: Error: Wait for VirtualService commerce-mock timeout (20000 ms) at Timeout._onTimeout (utils/index.js:404:20) at listOnTimeout (internal/timers.js:557:17) at processTimers (internal/timers.js:500:7) ``` **Steps to reproduce** <!-- List the steps to follow to reproduce the bug. Attach any files, links, code samples, or screenshots that could help in investigating the problem. --> **Troubleshooting** <!-- Describe the steps you have already taken to solve the issue. -->
non_main
commerce mock tests timeout on multiple pipelines thank you for your contribution before you submit the issue search open and closed issues for duplicates read the contributing guidelines description provide a clear and concise description of the problem describe where it appears when it occurred and what it affects the commerce mock tests which are part of the fast integration tests time out multiple times on multiple pipelines which seem to be one of the reasons for flakiness some of the test runs can be found and expected result the commerce mock is initialised properly and the tests succeed do not timeout actual result the tests time out with the following error executing standard testsuite commercemock tests commercemock test fixture should be ready error wait for virtualservice commerce mock timeout ms at timeout ontimeout utils index js at listontimeout internal timers js at processtimers internal timers js executing standard testsuite commercemock tests in cluster event should be delivered structured and binary mode error fetching published event responded with error assertionerror the same event id expected in the result expected podname lastorder jphdd to have nested property event id at convertaxioserror utils index js at home prow go src github com kyma project kyma tests fast integration test fixtures commerce mock index js at processticksandrejections internal process task queues js at async checkinclustereventdeliveryhelper test fixtures commerce mock index js at async checkinclustereventdelivery test fixtures commerce mock index js at async context test commerce mock js executing standard testsuite commercemock tests function should be reachable through secured api rule error wait for virtualservice commerce mock timeout ms at timeout ontimeout utils index js at listontimeout internal timers js at processtimers internal timers js executing standard testsuite commercemock tests order created event should trigger the lastorder function error wait for virtualservice commerce mock timeout ms at timeout ontimeout utils index js at listontimeout internal timers js at processtimers internal timers js executing standard testsuite commercemock tests order created cloud event in structured mode should trigger the lastorder function error wait for virtualservice commerce mock timeout ms at timeout ontimeout utils index js at listontimeout internal timers js at processtimers internal timers js executing standard testsuite commercemock tests order created cloud event in binary mode should trigger the lastorder function error wait for virtualservice commerce mock timeout ms at timeout ontimeout utils index js at listontimeout internal timers js at processtimers internal timers js executing standard testsuite commercemock tests should add update and delete a service error wait for virtualservice commerce mock timeout ms at timeout ontimeout utils index js at listontimeout internal timers js at processtimers internal timers js steps to reproduce troubleshooting
0
53,211
6,304,996,333
IssuesEvent
2017-07-21 17:16:03
conde2/DC-UFSCar-ES2-201701-BoxTesters
https://api.github.com/repos/conde2/DC-UFSCar-ES2-201701-BoxTesters
closed
Importação de itens bibliográficos na base corrente - ReTeste
Teste
Refazer o teste para as manutenções realizadas
1.0
Importação de itens bibliográficos na base corrente - ReTeste - Refazer o teste para as manutenções realizadas
non_main
importação de itens bibliográficos na base corrente reteste refazer o teste para as manutenções realizadas
0
2,605
8,838,983,685
IssuesEvent
2019-01-05 23:37:22
OpenLightingProject/ola
https://api.github.com/repos/OpenLightingProject/ola
opened
Check our Run Length Encoding can't cause buffer overflows
Component-Plugin Difficulty-Medium Language-C++ Maintainability Type-Task
Such as here: https://bugs.wireshark.org/bugzilla/show_bug.cgi?id=5539 Confirm, fix if necessary, and add tests for the future for the following usages - [ ] ESP Net - [ ] Shownet - [ ] ola::dmx::RunLengthEncoder - [ ] USB Pro (change packets?)
True
Check our Run Length Encoding can't cause buffer overflows - Such as here: https://bugs.wireshark.org/bugzilla/show_bug.cgi?id=5539 Confirm, fix if necessary, and add tests for the future for the following usages - [ ] ESP Net - [ ] Shownet - [ ] ola::dmx::RunLengthEncoder - [ ] USB Pro (change packets?)
main
check our run length encoding can t cause buffer overflows such as here confirm fix if necessary and add tests for the future for the following usages esp net shownet ola dmx runlengthencoder usb pro change packets
1
849
4,506,717,858
IssuesEvent
2016-09-02 05:54:24
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
DigitalOcean key generation
bug_report cloud digital_ocean feature_idea P3 waiting_on_maintainer
##### Issue Type: Bug Report ##### Component Name: digital_ocean_sshkey ##### Ansible Version: ansible 1.8.2 configured module search path = None ##### Environment: Mac OS X ##### Summary: If an SSH key exists in digitalocean, trying to "state=present" the same SSH public key with another name fails ##### Steps To Reproduce: - Create an SSH key on the WebUI, and name it key1 - run this ansible task: ``` local_action: digital_ocean_sshkey state=present name=key2 ssh_pub_key={{digitalocean_ssh_key_file}} client_id={{digitalocean_client_id}} api_key={{digitalocean_api_key}} ``` ##### Expected Results: One of: - An explanatory error message - the key being renamed - the id of the already existing key being returned (possibly with the name key1) ##### Actual Results: ``` failed: [127.0.0.1 -> 127.0.0.1] => {"failed": true} msg: SSH Key failed to be created ```
True
DigitalOcean key generation - ##### Issue Type: Bug Report ##### Component Name: digital_ocean_sshkey ##### Ansible Version: ansible 1.8.2 configured module search path = None ##### Environment: Mac OS X ##### Summary: If an SSH key exists in digitalocean, trying to "state=present" the same SSH public key with another name fails ##### Steps To Reproduce: - Create an SSH key on the WebUI, and name it key1 - run this ansible task: ``` local_action: digital_ocean_sshkey state=present name=key2 ssh_pub_key={{digitalocean_ssh_key_file}} client_id={{digitalocean_client_id}} api_key={{digitalocean_api_key}} ``` ##### Expected Results: One of: - An explanatory error message - the key being renamed - the id of the already existing key being returned (possibly with the name key1) ##### Actual Results: ``` failed: [127.0.0.1 -> 127.0.0.1] => {"failed": true} msg: SSH Key failed to be created ```
main
digitalocean key generation issue type bug report component name digital ocean sshkey ansible version ansible configured module search path none environment mac os x summary if an ssh key exists in digitalocean trying to state present the same ssh public key with another name fails steps to reproduce create an ssh key on the webui and name it run this ansible task local action digital ocean sshkey state present name ssh pub key digitalocean ssh key file client id digitalocean client id api key digitalocean api key expected results one of an explanatory error message the key being renamed the id of the already existing key being returned possibly with the name actual results failed failed true msg ssh key failed to be created
1
22,901
3,727,389,436
IssuesEvent
2016-03-06 08:05:05
godfather1103/mentohust
https://api.github.com/repos/godfather1103/mentohust
closed
wrt54g 1.1 ddwrt要怎么操作呢 求指导
auto-migrated Priority-Medium Type-Defect
``` 已经刷好ddwrt了 然后不懂怎么操作了 额 不是集成的我想自己弄个 ``` Original issue reported on code.google.com by `J1140752...@gmail.com` on 19 Apr 2013 at 7:03
1.0
wrt54g 1.1 ddwrt要怎么操作呢 求指导 - ``` 已经刷好ddwrt了 然后不懂怎么操作了 额 不是集成的我想自己弄个 ``` Original issue reported on code.google.com by `J1140752...@gmail.com` on 19 Apr 2013 at 7:03
non_main
ddwrt要怎么操作呢 求指导 已经刷好ddwrt了 然后不懂怎么操作了 额 不是集成的我想自己弄个 original issue reported on code google com by gmail com on apr at
0
4,734
24,444,568,838
IssuesEvent
2022-10-06 16:49:04
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
reopened
No response from invoke container for Lambda inside docker-compose
area/docker type/bug area/local/start-api stage/needs-attention maintainer/need-followup
### Description: Related to https://github.com/aws/aws-sam-cli/issues/2492 When running a docker-compose with an image that is using sam to run a local api, the api can never get a response from the lambda container. Everything works correctly when run locally not in a container. I have read through all the answers in 2492 and tried the various permutations suggested, nothing has worked. There appeared to be no subsequent documentation follow up for getting this setup to work with docker. ### Steps to reproduce: DockerFile ``` FROM python:alpine RUN apk add --no-cache --virtual build-deps build-base=0.5-r2 gcc=10.2.1_pre1-r3 bash=5.1.0-r0 && \ pip install aws-sam-cli==1.23.0 && \ apk del build-deps COPY ./sam_entrypoint.sh /bin/sam_entrypoint.sh RUN chmod +x /bin/sam_entrypoint.sh WORKDIR /app COPY ./node_modules /app/node_modules COPY ./package.json /app/package.json COPY ./template.yml /app/template.yml COPY ./dist /app/dist EXPOSE 3000 ENTRYPOINT ["sh","/bin/sam_entrypoint.sh"] ``` sam_entrypoint.sh ``` #!/bin/bash set -o errexit /usr/local/bin/sam local start-api --debug \ --template ./template.yml \ --host 0.0.0.0 \ --container-host host.docker.internal \ --container-host-interface 127.0.0.1 \ --docker-network host ```` Template.yml ``` AWSTemplateFormatVersion: "2010-09-09" Transform: "AWS::Serverless-2016-10-31" Description: "Automated lambda setup for api gateway" Resources: res0: Type: "AWS::Serverless::Function" Properties: CodeUri: "./" Handler: "dist/index.handler" Runtime: "nodejs12.x" Timeout: 10 Events: CatchAll: Type: "Api" Properties: Path: "/" Method: ANY res1: Type: "AWS::Serverless::Function" Properties: CodeUri: "./" Handler: "dist/index.handler" Runtime: "nodejs12.x" Timeout: 10 Events: CatchAll: Type: "Api" Properties: Path: "/{proxy+}" Method: ANY ``` Docker-compose ``` version: '3.6' services: sam_app: build: context: . dockerfile: Dockerfile.test ports: - "3000:3000" volumes: - /var/run/docker.sock:/var/run/docker.sock extra_hosts: - "host.docker.internal:host-gateway" ``` ### Observed result: ``` sam_app_1 | sam_app_1 | SAM CLI now collects telemetry to better understand customer needs. sam_app_1 | sam_app_1 | You can OPT OUT and disable telemetry collection by setting the sam_app_1 | environment variable SAM_CLI_TELEMETRY=0 in your shell. sam_app_1 | Thanks for your help! sam_app_1 | sam_app_1 | Learn More: https://docs.aws.amazon.com/serverless-application-model/latest/developerguide/serverless-sam-telemetry.html sam_app_1 | sam_app_1 | 2021-04-28 17:40:50,630 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics sam_app_1 | 2021-04-28 17:40:50,630 | Using config file: samconfig.toml, config environment: default sam_app_1 | 2021-04-28 17:40:50,631 | Expand command line arguments to: sam_app_1 | 2021-04-28 17:40:50,631 | --template_file=/app/template.yml --host=0.0.0.0 --port=3000 --static_dir=public --layer_cache_basedir=/root/.aws-sam/layers-pkg --container_host=localhost --container_host_interface=127.0.0.1 sam_app_1 | 2021-04-28 17:40:50,704 | local start-api command is called sam_app_1 | 2021-04-28 17:40:50,709 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,732 | 3 stacks found in the template sam_app_1 | 2021-04-28 17:40:50,732 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,753 | 3 resources found in the stack sam_app_1 | 2021-04-28 17:40:50,753 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,773 | Found Serverless function with name='res0' and CodeUri='./' sam_app_1 | 2021-04-28 17:40:50,773 | --base-dir is not presented, adjusting uri ./ relative to /app/template.yml sam_app_1 | 2021-04-28 17:40:50,773 | Found Serverless function with name='res1' and CodeUri='./' sam_app_1 | 2021-04-28 17:40:50,773 | --base-dir is not presented, adjusting uri ./ relative to /app/template.yml sam_app_1 | 2021-04-28 17:40:50,777 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,798 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,819 | Found '1' API Events in Serverless function with name 'res0' sam_app_1 | 2021-04-28 17:40:50,819 | Found '1' API Events in Serverless function with name 'res1' sam_app_1 | 2021-04-28 17:40:50,819 | Detected Inline Swagger definition sam_app_1 | 2021-04-28 17:40:50,819 | Lambda function integration not found in Swagger document at path='/' method='x-amazon-apigateway-any-method' sam_app_1 | 2021-04-28 17:40:50,819 | Lambda function integration not found in Swagger document at path='/{proxy+}' method='x-amazon-apigateway-any-method' sam_app_1 | 2021-04-28 17:40:50,819 | Found '0' APIs in resource 'ServerlessRestApi' sam_app_1 | 2021-04-28 17:40:50,819 | Removed duplicates from '0' Explicit APIs and '2' Implicit APIs to produce '2' APIs sam_app_1 | 2021-04-28 17:40:50,819 | 2 APIs found in the template sam_app_1 | 2021-04-28 17:40:50,825 | Mounting res0 at http://0.0.0.0:3000/ [DELETE, GET, HEAD, OPTIONS, PATCH, POST, PUT] sam_app_1 | 2021-04-28 17:40:50,825 | Mounting res1 at http://0.0.0.0:3000/{proxy+} [DELETE, GET, HEAD, OPTIONS, PATCH, POST, PUT] sam_app_1 | 2021-04-28 17:40:50,826 | You can now browse to the above endpoints to invoke your functions. You do not need to restart/reload SAM CLI while working on your functions, changes will be reflected instantly/automatically. You only need to restart SAM CLI if you update your AWS SAM template sam_app_1 | 2021-04-28 17:40:50,826 | Localhost server is starting up. Multi-threading = True sam_app_1 | 2021-04-28 17:40:50 * Running on http://0.0.0.0:3000/ (Press CTRL+C to quit) sam_app_1 | 2021-04-28 17:41:20,791 | Constructed String representation of Event to invoke Lambda. Event: {"body": null, "headers": {"Accept": "*/*", "Host": "127.0.0.1:3000", "User-Agent": "insomnia/2021.2.2", "X-Forwarded-Port": "3000", "X-Forwarded-Proto": "http"}, "httpMethod": "GET", "isBase64Encoded": false, "multiValueHeaders": {"Accept": ["*/*"], "Host": ["127.0.0.1:3000"], "User-Agent": ["insomnia/2021.2.2"], "X-Forwarded-Port": ["3000"], "X-Forwarded-Proto": ["http"]}, "multiValueQueryStringParameters": null, "path": "/v1/openapi.json", "pathParameters": {"proxy": "v1/openapi.json"}, "queryStringParameters": null, "requestContext": {"accountId": "123456789012", "apiId": "1234567890", "domainName": "127.0.0.1:3000", "extendedRequestId": null, "httpMethod": "GET", "identity": {"accountId": null, "apiKey": null, "caller": null, "cognitoAuthenticationProvider": null, "cognitoAuthenticationType": null, "cognitoIdentityPoolId": null, "sourceIp": "172.24.0.1", "user": null, "userAgent": "Custom User Agent String", "userArn": null}, "path": "/{proxy+}", "protocol": "HTTP/1.1", "requestId": "d9c5fa20-2328-43f7-b4d2-5a35028dfab1", "requestTime": "28/Apr/2021:17:40:50 +0000", "requestTimeEpoch": 1619631650, "resourceId": "123456", "resourcePath": "/{proxy+}", "stage": "Prod"}, "resource": "/{proxy+}", "stageVariables": null, "version": "1.0"} sam_app_1 | 2021-04-28 17:41:20,791 | Found one Lambda function with name 'res1' sam_app_1 | 2021-04-28 17:41:20,791 | Invoking dist/index.handler (nodejs12.x) sam_app_1 | 2021-04-28 17:41:20,791 | Environment variables overrides data is standard format sam_app_1 | 2021-04-28 17:41:20,791 | Loading AWS credentials from session with profile 'None' sam_app_1 | 2021-04-28 17:41:20,803 | Resolving code path. Cwd=/app, CodeUri=/app sam_app_1 | 2021-04-28 17:41:20,803 | Resolved absolute path to code is /app sam_app_1 | 2021-04-28 17:41:20,803 | Code /app is not a zip/jar file sam_app_1 | 2021-04-28 17:41:20,817 | Skip pulling image and use local one: amazon/aws-sam-cli-emulation-image-nodejs12.x:rapid-1.23.0. sam_app_1 | sam_app_1 | 2021-04-28 17:41:20,817 | Mounting /app as /var/task:ro,delegated inside runtime container sam_app_1 | 2021-04-28 17:41:21,124 | Starting a timer for 10 seconds for function 'res1' sam_app_1 | 2021-04-28 17:41:22,075 | Cleaning all decompressed code dirs sam_app_1 | 2021-04-28 17:41:22,075 | No response from invoke container for res1 sam_app_1 | 2021-04-28 17:41:22,075 | Invalid lambda response received: Lambda response must be valid json sam_app_1 | 2021-04-28 17:41:22 172.24.0.1 - - [28/Apr/2021 17:41:22] "GET /v1/openapi.json HTTP/1.1" 502 - ``` ### Expected result: That the lambda can be executed correctly and a response received. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Windows 10 2. `sam --version`: 1.23.0 3. AWS region: us-east-1
True
No response from invoke container for Lambda inside docker-compose - ### Description: Related to https://github.com/aws/aws-sam-cli/issues/2492 When running a docker-compose with an image that is using sam to run a local api, the api can never get a response from the lambda container. Everything works correctly when run locally not in a container. I have read through all the answers in 2492 and tried the various permutations suggested, nothing has worked. There appeared to be no subsequent documentation follow up for getting this setup to work with docker. ### Steps to reproduce: DockerFile ``` FROM python:alpine RUN apk add --no-cache --virtual build-deps build-base=0.5-r2 gcc=10.2.1_pre1-r3 bash=5.1.0-r0 && \ pip install aws-sam-cli==1.23.0 && \ apk del build-deps COPY ./sam_entrypoint.sh /bin/sam_entrypoint.sh RUN chmod +x /bin/sam_entrypoint.sh WORKDIR /app COPY ./node_modules /app/node_modules COPY ./package.json /app/package.json COPY ./template.yml /app/template.yml COPY ./dist /app/dist EXPOSE 3000 ENTRYPOINT ["sh","/bin/sam_entrypoint.sh"] ``` sam_entrypoint.sh ``` #!/bin/bash set -o errexit /usr/local/bin/sam local start-api --debug \ --template ./template.yml \ --host 0.0.0.0 \ --container-host host.docker.internal \ --container-host-interface 127.0.0.1 \ --docker-network host ```` Template.yml ``` AWSTemplateFormatVersion: "2010-09-09" Transform: "AWS::Serverless-2016-10-31" Description: "Automated lambda setup for api gateway" Resources: res0: Type: "AWS::Serverless::Function" Properties: CodeUri: "./" Handler: "dist/index.handler" Runtime: "nodejs12.x" Timeout: 10 Events: CatchAll: Type: "Api" Properties: Path: "/" Method: ANY res1: Type: "AWS::Serverless::Function" Properties: CodeUri: "./" Handler: "dist/index.handler" Runtime: "nodejs12.x" Timeout: 10 Events: CatchAll: Type: "Api" Properties: Path: "/{proxy+}" Method: ANY ``` Docker-compose ``` version: '3.6' services: sam_app: build: context: . dockerfile: Dockerfile.test ports: - "3000:3000" volumes: - /var/run/docker.sock:/var/run/docker.sock extra_hosts: - "host.docker.internal:host-gateway" ``` ### Observed result: ``` sam_app_1 | sam_app_1 | SAM CLI now collects telemetry to better understand customer needs. sam_app_1 | sam_app_1 | You can OPT OUT and disable telemetry collection by setting the sam_app_1 | environment variable SAM_CLI_TELEMETRY=0 in your shell. sam_app_1 | Thanks for your help! sam_app_1 | sam_app_1 | Learn More: https://docs.aws.amazon.com/serverless-application-model/latest/developerguide/serverless-sam-telemetry.html sam_app_1 | sam_app_1 | 2021-04-28 17:40:50,630 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics sam_app_1 | 2021-04-28 17:40:50,630 | Using config file: samconfig.toml, config environment: default sam_app_1 | 2021-04-28 17:40:50,631 | Expand command line arguments to: sam_app_1 | 2021-04-28 17:40:50,631 | --template_file=/app/template.yml --host=0.0.0.0 --port=3000 --static_dir=public --layer_cache_basedir=/root/.aws-sam/layers-pkg --container_host=localhost --container_host_interface=127.0.0.1 sam_app_1 | 2021-04-28 17:40:50,704 | local start-api command is called sam_app_1 | 2021-04-28 17:40:50,709 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,732 | 3 stacks found in the template sam_app_1 | 2021-04-28 17:40:50,732 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,753 | 3 resources found in the stack sam_app_1 | 2021-04-28 17:40:50,753 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,773 | Found Serverless function with name='res0' and CodeUri='./' sam_app_1 | 2021-04-28 17:40:50,773 | --base-dir is not presented, adjusting uri ./ relative to /app/template.yml sam_app_1 | 2021-04-28 17:40:50,773 | Found Serverless function with name='res1' and CodeUri='./' sam_app_1 | 2021-04-28 17:40:50,773 | --base-dir is not presented, adjusting uri ./ relative to /app/template.yml sam_app_1 | 2021-04-28 17:40:50,777 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,798 | No Parameters detected in the template sam_app_1 | 2021-04-28 17:40:50,819 | Found '1' API Events in Serverless function with name 'res0' sam_app_1 | 2021-04-28 17:40:50,819 | Found '1' API Events in Serverless function with name 'res1' sam_app_1 | 2021-04-28 17:40:50,819 | Detected Inline Swagger definition sam_app_1 | 2021-04-28 17:40:50,819 | Lambda function integration not found in Swagger document at path='/' method='x-amazon-apigateway-any-method' sam_app_1 | 2021-04-28 17:40:50,819 | Lambda function integration not found in Swagger document at path='/{proxy+}' method='x-amazon-apigateway-any-method' sam_app_1 | 2021-04-28 17:40:50,819 | Found '0' APIs in resource 'ServerlessRestApi' sam_app_1 | 2021-04-28 17:40:50,819 | Removed duplicates from '0' Explicit APIs and '2' Implicit APIs to produce '2' APIs sam_app_1 | 2021-04-28 17:40:50,819 | 2 APIs found in the template sam_app_1 | 2021-04-28 17:40:50,825 | Mounting res0 at http://0.0.0.0:3000/ [DELETE, GET, HEAD, OPTIONS, PATCH, POST, PUT] sam_app_1 | 2021-04-28 17:40:50,825 | Mounting res1 at http://0.0.0.0:3000/{proxy+} [DELETE, GET, HEAD, OPTIONS, PATCH, POST, PUT] sam_app_1 | 2021-04-28 17:40:50,826 | You can now browse to the above endpoints to invoke your functions. You do not need to restart/reload SAM CLI while working on your functions, changes will be reflected instantly/automatically. You only need to restart SAM CLI if you update your AWS SAM template sam_app_1 | 2021-04-28 17:40:50,826 | Localhost server is starting up. Multi-threading = True sam_app_1 | 2021-04-28 17:40:50 * Running on http://0.0.0.0:3000/ (Press CTRL+C to quit) sam_app_1 | 2021-04-28 17:41:20,791 | Constructed String representation of Event to invoke Lambda. Event: {"body": null, "headers": {"Accept": "*/*", "Host": "127.0.0.1:3000", "User-Agent": "insomnia/2021.2.2", "X-Forwarded-Port": "3000", "X-Forwarded-Proto": "http"}, "httpMethod": "GET", "isBase64Encoded": false, "multiValueHeaders": {"Accept": ["*/*"], "Host": ["127.0.0.1:3000"], "User-Agent": ["insomnia/2021.2.2"], "X-Forwarded-Port": ["3000"], "X-Forwarded-Proto": ["http"]}, "multiValueQueryStringParameters": null, "path": "/v1/openapi.json", "pathParameters": {"proxy": "v1/openapi.json"}, "queryStringParameters": null, "requestContext": {"accountId": "123456789012", "apiId": "1234567890", "domainName": "127.0.0.1:3000", "extendedRequestId": null, "httpMethod": "GET", "identity": {"accountId": null, "apiKey": null, "caller": null, "cognitoAuthenticationProvider": null, "cognitoAuthenticationType": null, "cognitoIdentityPoolId": null, "sourceIp": "172.24.0.1", "user": null, "userAgent": "Custom User Agent String", "userArn": null}, "path": "/{proxy+}", "protocol": "HTTP/1.1", "requestId": "d9c5fa20-2328-43f7-b4d2-5a35028dfab1", "requestTime": "28/Apr/2021:17:40:50 +0000", "requestTimeEpoch": 1619631650, "resourceId": "123456", "resourcePath": "/{proxy+}", "stage": "Prod"}, "resource": "/{proxy+}", "stageVariables": null, "version": "1.0"} sam_app_1 | 2021-04-28 17:41:20,791 | Found one Lambda function with name 'res1' sam_app_1 | 2021-04-28 17:41:20,791 | Invoking dist/index.handler (nodejs12.x) sam_app_1 | 2021-04-28 17:41:20,791 | Environment variables overrides data is standard format sam_app_1 | 2021-04-28 17:41:20,791 | Loading AWS credentials from session with profile 'None' sam_app_1 | 2021-04-28 17:41:20,803 | Resolving code path. Cwd=/app, CodeUri=/app sam_app_1 | 2021-04-28 17:41:20,803 | Resolved absolute path to code is /app sam_app_1 | 2021-04-28 17:41:20,803 | Code /app is not a zip/jar file sam_app_1 | 2021-04-28 17:41:20,817 | Skip pulling image and use local one: amazon/aws-sam-cli-emulation-image-nodejs12.x:rapid-1.23.0. sam_app_1 | sam_app_1 | 2021-04-28 17:41:20,817 | Mounting /app as /var/task:ro,delegated inside runtime container sam_app_1 | 2021-04-28 17:41:21,124 | Starting a timer for 10 seconds for function 'res1' sam_app_1 | 2021-04-28 17:41:22,075 | Cleaning all decompressed code dirs sam_app_1 | 2021-04-28 17:41:22,075 | No response from invoke container for res1 sam_app_1 | 2021-04-28 17:41:22,075 | Invalid lambda response received: Lambda response must be valid json sam_app_1 | 2021-04-28 17:41:22 172.24.0.1 - - [28/Apr/2021 17:41:22] "GET /v1/openapi.json HTTP/1.1" 502 - ``` ### Expected result: That the lambda can be executed correctly and a response received. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: Windows 10 2. `sam --version`: 1.23.0 3. AWS region: us-east-1
main
no response from invoke container for lambda inside docker compose description related to when running a docker compose with an image that is using sam to run a local api the api can never get a response from the lambda container everything works correctly when run locally not in a container i have read through all the answers in and tried the various permutations suggested nothing has worked there appeared to be no subsequent documentation follow up for getting this setup to work with docker steps to reproduce dockerfile from python alpine run apk add no cache virtual build deps build base gcc bash pip install aws sam cli apk del build deps copy sam entrypoint sh bin sam entrypoint sh run chmod x bin sam entrypoint sh workdir app copy node modules app node modules copy package json app package json copy template yml app template yml copy dist app dist expose entrypoint sam entrypoint sh bin bash set o errexit usr local bin sam local start api debug template template yml host container host host docker internal container host interface docker network host template yml awstemplateformatversion transform aws serverless description automated lambda setup for api gateway resources type aws serverless function properties codeuri handler dist index handler runtime x timeout events catchall type api properties path method any type aws serverless function properties codeuri handler dist index handler runtime x timeout events catchall type api properties path proxy method any docker compose version services sam app build context dockerfile dockerfile test ports volumes var run docker sock var run docker sock extra hosts host docker internal host gateway observed result sam app sam app sam cli now collects telemetry to better understand customer needs sam app sam app you can opt out and disable telemetry collection by setting the sam app environment variable sam cli telemetry in your shell sam app thanks for your help sam app sam app learn more sam app sam app telemetry endpoint configured to be sam app using config file samconfig toml config environment default sam app expand command line arguments to sam app template file app template yml host port static dir public layer cache basedir root aws sam layers pkg container host localhost container host interface sam app local start api command is called sam app no parameters detected in the template sam app stacks found in the template sam app no parameters detected in the template sam app resources found in the stack sam app no parameters detected in the template sam app found serverless function with name and codeuri sam app base dir is not presented adjusting uri relative to app template yml sam app found serverless function with name and codeuri sam app base dir is not presented adjusting uri relative to app template yml sam app no parameters detected in the template sam app no parameters detected in the template sam app found api events in serverless function with name sam app found api events in serverless function with name sam app detected inline swagger definition sam app lambda function integration not found in swagger document at path method x amazon apigateway any method sam app lambda function integration not found in swagger document at path proxy method x amazon apigateway any method sam app found apis in resource serverlessrestapi sam app removed duplicates from explicit apis and implicit apis to produce apis sam app apis found in the template sam app mounting at sam app mounting at sam app you can now browse to the above endpoints to invoke your functions you do not need to restart reload sam cli while working on your functions changes will be reflected instantly automatically you only need to restart sam cli if you update your aws sam template sam app localhost server is starting up multi threading true sam app running on press ctrl c to quit sam app constructed string representation of event to invoke lambda event body null headers accept host user agent insomnia x forwarded port x forwarded proto http httpmethod get false multivalueheaders accept host user agent x forwarded port x forwarded proto multivaluequerystringparameters null path openapi json pathparameters proxy openapi json querystringparameters null requestcontext accountid apiid domainname extendedrequestid null httpmethod get identity accountid null apikey null caller null cognitoauthenticationprovider null cognitoauthenticationtype null cognitoidentitypoolid null sourceip user null useragent custom user agent string userarn null path proxy protocol http requestid requesttime apr requesttimeepoch resourceid resourcepath proxy stage prod resource proxy stagevariables null version sam app found one lambda function with name sam app invoking dist index handler x sam app environment variables overrides data is standard format sam app loading aws credentials from session with profile none sam app resolving code path cwd app codeuri app sam app resolved absolute path to code is app sam app code app is not a zip jar file sam app skip pulling image and use local one amazon aws sam cli emulation image x rapid sam app sam app mounting app as var task ro delegated inside runtime container sam app starting a timer for seconds for function sam app cleaning all decompressed code dirs sam app no response from invoke container for sam app invalid lambda response received lambda response must be valid json sam app get openapi json http expected result that the lambda can be executed correctly and a response received additional environment details ex windows mac amazon linux etc os windows sam version aws region us east
1
3,616
14,619,341,737
IssuesEvent
2020-12-22 17:41:08
casperstorm/ajour
https://api.github.com/repos/casperstorm/ajour
closed
[weakaura] missing field `username`
B - bug C - waiting on maintainer S - weakauras
**Describe the bug** Stuck at parsing weakauras **To Reproduce** Steps to reproduce the behavior: 1. Go to "My Weakauras" 2. Click on dropdown on right side 3. Select my username 4. Get stuck **Expected behavior** Show weakauras **Screenshots** ![image](https://user-images.githubusercontent.com/17293360/102757205-41d86c00-4371-11eb-9b37-27c65c9f3cae.png) **Software involved** Windows 10 20H2 WoW Classic Ajour 0.6.0 **Log Output** ``` 09:37:58.456 [ajour::gui::update][DEBUG] Message::ParsedAddons(Classic, 63 addons) 09:37:58.794 [ajour_core::catalog][DEBUG] Successfully fetched and parsed https://github.com/casperstorm/ajour-catalog/releases/latest/download/wowi.json 09:37:59.195 [ajour_core::catalog][DEBUG] Successfully fetched and parsed https://github.com/casperstorm/ajour-catalog/releases/latest/download/curse.json 09:37:59.242 [ajour::gui::update][DEBUG] Message::CatalogDownloaded(15671 addons in catalog) 09:38:03.107 [ajour::gui::update][DEBUG] Interaction::ModeSelected(MyWeakAuras(Classic)) 09:38:08.333 [ajour::gui::update][DEBUG] Message::WeakAurasAccountSelected(**HIDDEN**) 09:38:08.643 [ajour][ERROR] Failed to parse WeakAuras 09:38:08.643 [ajour][ERROR] caused by: missing field `username` at line 1 column 1868 ```
True
[weakaura] missing field `username` - **Describe the bug** Stuck at parsing weakauras **To Reproduce** Steps to reproduce the behavior: 1. Go to "My Weakauras" 2. Click on dropdown on right side 3. Select my username 4. Get stuck **Expected behavior** Show weakauras **Screenshots** ![image](https://user-images.githubusercontent.com/17293360/102757205-41d86c00-4371-11eb-9b37-27c65c9f3cae.png) **Software involved** Windows 10 20H2 WoW Classic Ajour 0.6.0 **Log Output** ``` 09:37:58.456 [ajour::gui::update][DEBUG] Message::ParsedAddons(Classic, 63 addons) 09:37:58.794 [ajour_core::catalog][DEBUG] Successfully fetched and parsed https://github.com/casperstorm/ajour-catalog/releases/latest/download/wowi.json 09:37:59.195 [ajour_core::catalog][DEBUG] Successfully fetched and parsed https://github.com/casperstorm/ajour-catalog/releases/latest/download/curse.json 09:37:59.242 [ajour::gui::update][DEBUG] Message::CatalogDownloaded(15671 addons in catalog) 09:38:03.107 [ajour::gui::update][DEBUG] Interaction::ModeSelected(MyWeakAuras(Classic)) 09:38:08.333 [ajour::gui::update][DEBUG] Message::WeakAurasAccountSelected(**HIDDEN**) 09:38:08.643 [ajour][ERROR] Failed to parse WeakAuras 09:38:08.643 [ajour][ERROR] caused by: missing field `username` at line 1 column 1868 ```
main
missing field username describe the bug stuck at parsing weakauras to reproduce steps to reproduce the behavior go to my weakauras click on dropdown on right side select my username get stuck expected behavior show weakauras screenshots software involved windows wow classic ajour log output message parsedaddons classic addons successfully fetched and parsed successfully fetched and parsed message catalogdownloaded addons in catalog interaction modeselected myweakauras classic message weakaurasaccountselected hidden failed to parse weakauras caused by missing field username at line column
1
67,808
3,281,961,160
IssuesEvent
2015-10-28 01:46:00
cs2103aug2015-w09-1j/main
https://api.github.com/repos/cs2103aug2015-w09-1j/main
closed
edit function able to update based on variable
priority.medium
example : edit 1 start_time 1800 return the start_time and index
1.0
edit function able to update based on variable - example : edit 1 start_time 1800 return the start_time and index
non_main
edit function able to update based on variable example edit start time return the start time and index
0
3,029
11,206,922,246
IssuesEvent
2020-01-06 00:56:50
javascript-obfuscator/javascript-obfuscator
https://api.github.com/repos/javascript-obfuscator/javascript-obfuscator
closed
@sanex3339 => react-native-obfuscating-transformer maintainership
maintaining
Hi @sanex3339! :wave: Sorry for making this a github issue, I didn't know how else to contact you. I'm no longer actively maintaining react-native-obfuscating-transformer so would like to add some more maintainers who can keep it up to date. Unfortunately I don't have admin rights over that repo after transferring ownership to the javascript-obfuscator org. Can you give me those rights? Or if not, then can you at least add @jesucarr as a collaborator? Thanks :heart:
True
@sanex3339 => react-native-obfuscating-transformer maintainership - Hi @sanex3339! :wave: Sorry for making this a github issue, I didn't know how else to contact you. I'm no longer actively maintaining react-native-obfuscating-transformer so would like to add some more maintainers who can keep it up to date. Unfortunately I don't have admin rights over that repo after transferring ownership to the javascript-obfuscator org. Can you give me those rights? Or if not, then can you at least add @jesucarr as a collaborator? Thanks :heart:
main
react native obfuscating transformer maintainership hi wave sorry for making this a github issue i didn t know how else to contact you i m no longer actively maintaining react native obfuscating transformer so would like to add some more maintainers who can keep it up to date unfortunately i don t have admin rights over that repo after transferring ownership to the javascript obfuscator org can you give me those rights or if not then can you at least add jesucarr as a collaborator thanks heart
1
1,573
6,572,335,784
IssuesEvent
2017-09-11 01:29:41
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
ecs_service module should require an ARN
affects_2.1 aws bug_report cloud waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ecs_service ##### ANSIBLE VERSION ``` # ansible --version ansible 2.1.1.0 config file = /opt/ansible.cfg configured module search path = [] ``` ##### CONFIGURATION nothing relevant ##### OS / ENVIRONMENT N/A ##### SUMMARY The ecs_service module accepts a task definition in the form "family:revision", however AWS always returns an ARN, at least when the revision is specified. Because this module compares strings, it will always try to update the service when the "family:revision" form is used. ##### STEPS TO REPRODUCE Create a service with a specified revision and run it several times: ``` - hosts: localhost tasks: - ecs_cluster: name: cluster state: present - ecs_taskdefinition: family: family state: present containers: - name: web essential: true image: centos memory: 128 cpu: 1 - ecs_service: name: abc cluster: cluster state: present desired_count: 1 task_definition: "family:1" ``` ##### EXPECTED RESULTS nothing changed on second run ##### ACTUAL RESULTS ``` [16:12:29] ansible-playbook tmp.yml -vvv Using /etc/ansible/ansible.cfg as config file [WARNING]: provided hosts list is empty, only localhost is available PLAYBOOK: tmp.yml ************************************************************** 1 plays in tmp.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: dev <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541 `" && echo ansible-tmp-1469801554.65-204441427630541="` echo $HOME/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmp9vyeHf TO /home/dev/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541/setup <127.0.0.1> EXEC /bin/sh -c 'LANG=en_GB.UTF-8 LC_ALL=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8 /usr/bin/python /home/dev/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541/setup; rm -rf "/home/dev/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [ecs_cluster] ************************************************************* task path: /home/dev/portal/tmp.yml:3 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: dev <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642 `" && echo ansible-tmp-1469801555.03-35744438949642="` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpofrmVc TO /home/dev/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642/ecs_cluster <127.0.0.1> EXEC /bin/sh -c 'LANG=en_GB.UTF-8 LC_ALL=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8 /usr/bin/python /home/dev/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642/ecs_cluster; rm -rf "/home/dev/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "cluster": {"activeServicesCount": 1, "clusterArn": "arn:aws:ecs:eu-west-1:626248445012:cluster/cluster", "clusterName": "cluster", "pendingTasksCount": 0, "registeredContainerInstancesCount": 0, "runningTasksCount": 0, "status": "ACTIVE"}, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "delay": 10, "ec2_url": null, "name": "cluster", "profile": null, "region": null, "repeat": 10, "security_token": null, "state": "present", "validate_certs": true}, "module_name": "ecs_cluster"}} TASK [ecs_taskdefinition] ****************************************************** task path: /home/dev/portal/tmp.yml:7 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: dev <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560 `" && echo ansible-tmp-1469801555.46-135858790727560="` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmppKEXY2 TO /home/dev/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560/ecs_taskdefinition <127.0.0.1> EXEC /bin/sh -c 'LANG=en_GB.UTF-8 LC_ALL=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8 /usr/bin/python /home/dev/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560/ecs_taskdefinition; rm -rf "/home/dev/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "invocation": {"module_args": {"arn": null, "aws_access_key": null, "aws_secret_key": null, "containers": [{"cpu": 1, "essential": true, "image": "centos", "memory": 128, "name": "web"}], "ec2_url": null, "family": "family", "profile": null, "region": null, "revision": null, "security_token": null, "state": "present", "validate_certs": true, "volumes": null}, "module_name": "ecs_taskdefinition"}, "taskdefinition": {"containerDefinitions": [{"cpu": 1, "environment": [], "essential": true, "image": "centos", "memory": 128, "mountPoints": [], "name": "web", "portMappings": [], "volumesFrom": []}], "family": "family", "revision": 1, "status": "ACTIVE", "taskDefinitionArn": "arn:aws:ecs:eu-west-1:626248445012:task-definition/family:1", "volumes": []}} TASK [ecs_service] ************************************************************* task path: /home/dev/portal/tmp.yml:17 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: dev <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527 `" && echo ansible-tmp-1469801555.91-67322819554527="` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmptWRjdH TO /home/dev/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527/ecs_service <127.0.0.1> EXEC /bin/sh -c 'LANG=en_GB.UTF-8 LC_ALL=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8 /usr/bin/python /home/dev/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527/ecs_service; rm -rf "/home/dev/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527/" > /dev/null 2>&1 && sleep 0' changed: [localhost] => {"changed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "client_token": null, "cluster": "cluster", "delay": 10, "desired_count": 1, "ec2_url": null, "load_balancers": null, "name": "abc", "profile": null, "region": null, "repeat": 10, "role": null, "security_token": null, "state": "present", "task_definition": "family:1", "validate_certs": true}, "module_name": "ecs_service"}, "service": {"clusterArn": "arn:aws:ecs:eu-west-1:626248445012:cluster/cluster", "createdAt": "2016-07-29T16:12:25.172000+02:00", "deploymentConfiguration": {"maximumPercent": 200, "minimumHealthyPercent": 100}, "deployments": [{"createdAt": "2016-07-29 16:12:25.172000+02:00", "desiredCount": 1, "id": "ecs-svc/9223370567053230635", "pendingCount": 0, "runningCount": 0, "status": "PRIMARY", "taskDefinition": "arn:aws:ecs:eu-west-1:626248445012:task-definition/family:1", "updatedAt": "2016-07-29 16:12:25.172000+02:00"}], "desiredCount": 1, "events": [{"createdAt": "2016-07-29 16:12:29.740000+02:00", "id": "79155548-f7ef-4a9a-92fe-81137463ae5a", "message": "(service abc) was unable to place a task because no container instance met all of its requirements. Reason: No Container Instances were found in your cluster. For more information, see the Troubleshooting section of the Amazon ECS Developer Guide."}], "loadBalancers": [], "pendingCount": 0, "runningCount": 0, "serviceArn": "arn:aws:ecs:eu-west-1:626248445012:service/abc", "serviceName": "abc", "status": "ACTIVE", "taskDefinition": "arn:aws:ecs:eu-west-1:626248445012:task-definition/family:1"}} PLAY RECAP ********************************************************************* localhost : ok=4 changed=1 unreachable=0 failed=0 ```
True
ecs_service module should require an ARN - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME ecs_service ##### ANSIBLE VERSION ``` # ansible --version ansible 2.1.1.0 config file = /opt/ansible.cfg configured module search path = [] ``` ##### CONFIGURATION nothing relevant ##### OS / ENVIRONMENT N/A ##### SUMMARY The ecs_service module accepts a task definition in the form "family:revision", however AWS always returns an ARN, at least when the revision is specified. Because this module compares strings, it will always try to update the service when the "family:revision" form is used. ##### STEPS TO REPRODUCE Create a service with a specified revision and run it several times: ``` - hosts: localhost tasks: - ecs_cluster: name: cluster state: present - ecs_taskdefinition: family: family state: present containers: - name: web essential: true image: centos memory: 128 cpu: 1 - ecs_service: name: abc cluster: cluster state: present desired_count: 1 task_definition: "family:1" ``` ##### EXPECTED RESULTS nothing changed on second run ##### ACTUAL RESULTS ``` [16:12:29] ansible-playbook tmp.yml -vvv Using /etc/ansible/ansible.cfg as config file [WARNING]: provided hosts list is empty, only localhost is available PLAYBOOK: tmp.yml ************************************************************** 1 plays in tmp.yml PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: dev <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541 `" && echo ansible-tmp-1469801554.65-204441427630541="` echo $HOME/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmp9vyeHf TO /home/dev/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541/setup <127.0.0.1> EXEC /bin/sh -c 'LANG=en_GB.UTF-8 LC_ALL=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8 /usr/bin/python /home/dev/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541/setup; rm -rf "/home/dev/.ansible/tmp/ansible-tmp-1469801554.65-204441427630541/" > /dev/null 2>&1 && sleep 0' ok: [localhost] TASK [ecs_cluster] ************************************************************* task path: /home/dev/portal/tmp.yml:3 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: dev <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642 `" && echo ansible-tmp-1469801555.03-35744438949642="` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmpofrmVc TO /home/dev/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642/ecs_cluster <127.0.0.1> EXEC /bin/sh -c 'LANG=en_GB.UTF-8 LC_ALL=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8 /usr/bin/python /home/dev/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642/ecs_cluster; rm -rf "/home/dev/.ansible/tmp/ansible-tmp-1469801555.03-35744438949642/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "cluster": {"activeServicesCount": 1, "clusterArn": "arn:aws:ecs:eu-west-1:626248445012:cluster/cluster", "clusterName": "cluster", "pendingTasksCount": 0, "registeredContainerInstancesCount": 0, "runningTasksCount": 0, "status": "ACTIVE"}, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "delay": 10, "ec2_url": null, "name": "cluster", "profile": null, "region": null, "repeat": 10, "security_token": null, "state": "present", "validate_certs": true}, "module_name": "ecs_cluster"}} TASK [ecs_taskdefinition] ****************************************************** task path: /home/dev/portal/tmp.yml:7 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: dev <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560 `" && echo ansible-tmp-1469801555.46-135858790727560="` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmppKEXY2 TO /home/dev/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560/ecs_taskdefinition <127.0.0.1> EXEC /bin/sh -c 'LANG=en_GB.UTF-8 LC_ALL=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8 /usr/bin/python /home/dev/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560/ecs_taskdefinition; rm -rf "/home/dev/.ansible/tmp/ansible-tmp-1469801555.46-135858790727560/" > /dev/null 2>&1 && sleep 0' ok: [localhost] => {"changed": false, "invocation": {"module_args": {"arn": null, "aws_access_key": null, "aws_secret_key": null, "containers": [{"cpu": 1, "essential": true, "image": "centos", "memory": 128, "name": "web"}], "ec2_url": null, "family": "family", "profile": null, "region": null, "revision": null, "security_token": null, "state": "present", "validate_certs": true, "volumes": null}, "module_name": "ecs_taskdefinition"}, "taskdefinition": {"containerDefinitions": [{"cpu": 1, "environment": [], "essential": true, "image": "centos", "memory": 128, "mountPoints": [], "name": "web", "portMappings": [], "volumesFrom": []}], "family": "family", "revision": 1, "status": "ACTIVE", "taskDefinitionArn": "arn:aws:ecs:eu-west-1:626248445012:task-definition/family:1", "volumes": []}} TASK [ecs_service] ************************************************************* task path: /home/dev/portal/tmp.yml:17 <127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: dev <127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527 `" && echo ansible-tmp-1469801555.91-67322819554527="` echo $HOME/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527 `" ) && sleep 0' <127.0.0.1> PUT /tmp/tmptWRjdH TO /home/dev/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527/ecs_service <127.0.0.1> EXEC /bin/sh -c 'LANG=en_GB.UTF-8 LC_ALL=en_GB.UTF-8 LC_MESSAGES=en_GB.UTF-8 /usr/bin/python /home/dev/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527/ecs_service; rm -rf "/home/dev/.ansible/tmp/ansible-tmp-1469801555.91-67322819554527/" > /dev/null 2>&1 && sleep 0' changed: [localhost] => {"changed": true, "invocation": {"module_args": {"aws_access_key": null, "aws_secret_key": null, "client_token": null, "cluster": "cluster", "delay": 10, "desired_count": 1, "ec2_url": null, "load_balancers": null, "name": "abc", "profile": null, "region": null, "repeat": 10, "role": null, "security_token": null, "state": "present", "task_definition": "family:1", "validate_certs": true}, "module_name": "ecs_service"}, "service": {"clusterArn": "arn:aws:ecs:eu-west-1:626248445012:cluster/cluster", "createdAt": "2016-07-29T16:12:25.172000+02:00", "deploymentConfiguration": {"maximumPercent": 200, "minimumHealthyPercent": 100}, "deployments": [{"createdAt": "2016-07-29 16:12:25.172000+02:00", "desiredCount": 1, "id": "ecs-svc/9223370567053230635", "pendingCount": 0, "runningCount": 0, "status": "PRIMARY", "taskDefinition": "arn:aws:ecs:eu-west-1:626248445012:task-definition/family:1", "updatedAt": "2016-07-29 16:12:25.172000+02:00"}], "desiredCount": 1, "events": [{"createdAt": "2016-07-29 16:12:29.740000+02:00", "id": "79155548-f7ef-4a9a-92fe-81137463ae5a", "message": "(service abc) was unable to place a task because no container instance met all of its requirements. Reason: No Container Instances were found in your cluster. For more information, see the Troubleshooting section of the Amazon ECS Developer Guide."}], "loadBalancers": [], "pendingCount": 0, "runningCount": 0, "serviceArn": "arn:aws:ecs:eu-west-1:626248445012:service/abc", "serviceName": "abc", "status": "ACTIVE", "taskDefinition": "arn:aws:ecs:eu-west-1:626248445012:task-definition/family:1"}} PLAY RECAP ********************************************************************* localhost : ok=4 changed=1 unreachable=0 failed=0 ```
main
ecs service module should require an arn issue type bug report component name ecs service ansible version ansible version ansible config file opt ansible cfg configured module search path configuration nothing relevant os environment n a summary the ecs service module accepts a task definition in the form family revision however aws always returns an arn at least when the revision is specified because this module compares strings it will always try to update the service when the family revision form is used steps to reproduce create a service with a specified revision and run it several times hosts localhost tasks ecs cluster name cluster state present ecs taskdefinition family family state present containers name web essential true image centos memory cpu ecs service name abc cluster cluster state present desired count task definition family expected results nothing changed on second run actual results ansible playbook tmp yml vvv using etc ansible ansible cfg as config file provided hosts list is empty only localhost is available playbook tmp yml plays in tmp yml play task establish local connection for user dev exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home dev ansible tmp ansible tmp setup exec bin sh c lang en gb utf lc all en gb utf lc messages en gb utf usr bin python home dev ansible tmp ansible tmp setup rm rf home dev ansible tmp ansible tmp dev null sleep ok task task path home dev portal tmp yml establish local connection for user dev exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpofrmvc to home dev ansible tmp ansible tmp ecs cluster exec bin sh c lang en gb utf lc all en gb utf lc messages en gb utf usr bin python home dev ansible tmp ansible tmp ecs cluster rm rf home dev ansible tmp ansible tmp dev null sleep ok changed false cluster activeservicescount clusterarn arn aws ecs eu west cluster cluster clustername cluster pendingtaskscount registeredcontainerinstancescount runningtaskscount status active invocation module args aws access key null aws secret key null delay url null name cluster profile null region null repeat security token null state present validate certs true module name ecs cluster task task path home dev portal tmp yml establish local connection for user dev exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home dev ansible tmp ansible tmp ecs taskdefinition exec bin sh c lang en gb utf lc all en gb utf lc messages en gb utf usr bin python home dev ansible tmp ansible tmp ecs taskdefinition rm rf home dev ansible tmp ansible tmp dev null sleep ok changed false invocation module args arn null aws access key null aws secret key null containers url null family family profile null region null revision null security token null state present validate certs true volumes null module name ecs taskdefinition taskdefinition containerdefinitions essential true image centos memory mountpoints name web portmappings volumesfrom family family revision status active taskdefinitionarn arn aws ecs eu west task definition family volumes task task path home dev portal tmp yml establish local connection for user dev exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmptwrjdh to home dev ansible tmp ansible tmp ecs service exec bin sh c lang en gb utf lc all en gb utf lc messages en gb utf usr bin python home dev ansible tmp ansible tmp ecs service rm rf home dev ansible tmp ansible tmp dev null sleep changed changed true invocation module args aws access key null aws secret key null client token null cluster cluster delay desired count url null load balancers null name abc profile null region null repeat role null security token null state present task definition family validate certs true module name ecs service service clusterarn arn aws ecs eu west cluster cluster createdat deploymentconfiguration maximumpercent minimumhealthypercent deployments desiredcount events loadbalancers pendingcount runningcount servicearn arn aws ecs eu west service abc servicename abc status active taskdefinition arn aws ecs eu west task definition family play recap localhost ok changed unreachable failed
1
262,962
8,272,655,735
IssuesEvent
2018-09-16 22:38:32
javaee/glassfish
https://api.github.com/repos/javaee/glassfish
closed
Either PKG_CLIENT_READ_TIMEOUT is too small or Oracle's servers are too slow
3_1-next 3_1-next_release-note 3_1-next_release-note-added 3_1_1-scrubbed 3_1_2-exclude Component: update_center Priority: Major Type: Bug
Installation of update tool fails due to timeout (every time we tried it in the past seven days). Either the timeouts programmed into the installer are too small, or Oracle's servers providing the downloads are too slow. Sometimes no connection at all is possible, but most of the time the download fails: Proxy: Using system proxy settings. Install image: C:\glassfish3 Installing updatetool packages. Downloading 2 packages. Downloading updatetool (564 files, 4.630.541 bytes). File 564/564 Downloading wxpython2.8-minimal (346 files, 13.763.618 bytes). File 46/346Input/output error: Read timed out Could not download application packages. This could be because: * a proxy server is needed to access the internet. Please ensure that the system proxy server settings in your Internet Options control panel (under Connections:LAN Settings) are correct, or set the HTTP_PROXY environment variable to the full URL of the proxy server. * the package server is down or otherwise inaccessible or it is generating invalid data. Please contact the provider of the package server. The negative side effect is that the graphical installer which internally tries to do this download will fail without clearly telling that actual cause of the problem (just says it cannot install the update tool), so people don't know what to do! When increasing the timeout we could install without any problem, so this clearly proofs that either the default timeout is too small or Oracle's servers are just too slow. There is nothing wrong with our internet connection and while the download fails, a test drive proofs a FREE bandwith of 1,5 Mb/s (so it is not related to our infrastructure or net access). C:\>set PKG_CLIENT_CONNECT_TIMEOUT=300 C:\>set PKG_CLIENT_READ_TIMEOUT=300 Proxy: Using system proxy settings. Install image: C:\glassfish3 Installing updatetool packages. Downloading 2 packages. Downloading updatetool (564 files, 4.630.541 bytes). File 564/564 Downloading wxpython2.8-minimal (346 files, 13.763.618 bytes). File 346/346 Executing 1.095 install actions. Registering notifier: Not able to register. Returned exit code 3. Starting notifier. Initialization complete. Please either upgrade your download facilities or increase the default timeouts. But in any case, please provide a clear message to the user that it is not HIS fault, particularly in the graphical installer. We have rated this bug as CRITICAL since it prevents a successful complete default installation using the graphical installer. #### Environment Win7 Pro SP1 64 Bit de_DE #### Affected Versions [3.1]
1.0
Either PKG_CLIENT_READ_TIMEOUT is too small or Oracle's servers are too slow - Installation of update tool fails due to timeout (every time we tried it in the past seven days). Either the timeouts programmed into the installer are too small, or Oracle's servers providing the downloads are too slow. Sometimes no connection at all is possible, but most of the time the download fails: Proxy: Using system proxy settings. Install image: C:\glassfish3 Installing updatetool packages. Downloading 2 packages. Downloading updatetool (564 files, 4.630.541 bytes). File 564/564 Downloading wxpython2.8-minimal (346 files, 13.763.618 bytes). File 46/346Input/output error: Read timed out Could not download application packages. This could be because: * a proxy server is needed to access the internet. Please ensure that the system proxy server settings in your Internet Options control panel (under Connections:LAN Settings) are correct, or set the HTTP_PROXY environment variable to the full URL of the proxy server. * the package server is down or otherwise inaccessible or it is generating invalid data. Please contact the provider of the package server. The negative side effect is that the graphical installer which internally tries to do this download will fail without clearly telling that actual cause of the problem (just says it cannot install the update tool), so people don't know what to do! When increasing the timeout we could install without any problem, so this clearly proofs that either the default timeout is too small or Oracle's servers are just too slow. There is nothing wrong with our internet connection and while the download fails, a test drive proofs a FREE bandwith of 1,5 Mb/s (so it is not related to our infrastructure or net access). C:\>set PKG_CLIENT_CONNECT_TIMEOUT=300 C:\>set PKG_CLIENT_READ_TIMEOUT=300 Proxy: Using system proxy settings. Install image: C:\glassfish3 Installing updatetool packages. Downloading 2 packages. Downloading updatetool (564 files, 4.630.541 bytes). File 564/564 Downloading wxpython2.8-minimal (346 files, 13.763.618 bytes). File 346/346 Executing 1.095 install actions. Registering notifier: Not able to register. Returned exit code 3. Starting notifier. Initialization complete. Please either upgrade your download facilities or increase the default timeouts. But in any case, please provide a clear message to the user that it is not HIS fault, particularly in the graphical installer. We have rated this bug as CRITICAL since it prevents a successful complete default installation using the graphical installer. #### Environment Win7 Pro SP1 64 Bit de_DE #### Affected Versions [3.1]
non_main
either pkg client read timeout is too small or oracle s servers are too slow installation of update tool fails due to timeout every time we tried it in the past seven days either the timeouts programmed into the installer are too small or oracle s servers providing the downloads are too slow sometimes no connection at all is possible but most of the time the download fails proxy using system proxy settings install image c installing updatetool packages downloading packages downloading updatetool files bytes file downloading minimal files bytes file output error read timed out could not download application packages this could be because a proxy server is needed to access the internet please ensure that the system proxy server settings in your internet options control panel under connections lan settings are correct or set the http proxy environment variable to the full url of the proxy server the package server is down or otherwise inaccessible or it is generating invalid data please contact the provider of the package server the negative side effect is that the graphical installer which internally tries to do this download will fail without clearly telling that actual cause of the problem just says it cannot install the update tool so people don t know what to do when increasing the timeout we could install without any problem so this clearly proofs that either the default timeout is too small or oracle s servers are just too slow there is nothing wrong with our internet connection and while the download fails a test drive proofs a free bandwith of mb s so it is not related to our infrastructure or net access c set pkg client connect timeout c set pkg client read timeout proxy using system proxy settings install image c installing updatetool packages downloading packages downloading updatetool files bytes file downloading minimal files bytes file executing install actions registering notifier not able to register returned exit code starting notifier initialization complete please either upgrade your download facilities or increase the default timeouts but in any case please provide a clear message to the user that it is not his fault particularly in the graphical installer we have rated this bug as critical since it prevents a successful complete default installation using the graphical installer environment pro bit de de affected versions
0
1,237
5,268,229,308
IssuesEvent
2017-02-05 08:46:44
viktorradnai/flightgear-ask21
https://api.github.com/repos/viktorradnai/flightgear-ask21
opened
Rework Aircraft Rating?
enhancement maintainability question
In the -set file as well as on the wiki, we still have the following rating **_FDM_**: 2: FDM tuned for cruise configuration. **_Systems_**: 2: Working electrical system, fuel feed cockpit controls, stable autopilot **_Cockpit_**: 2: 2D panel in 3D cockpit, or incomplete 3D panel **_Model_**: 3: Accurate 3D model with animated control surfaces, gear detailing (retraction, rotation), prop In my opinion this is not quite correct, e.g. **Cockpit** should IMO be improved to a 4 as we have " 3D panel and accurately modelled 3D cockpit, plain texturing. Hotspots for majority of controls" **Systems** should be corrected to 4 (note here that the ASK21 has very few systems to model). I have added a limit system which I will push after tooltips are solved. Later we can also add a bit more code to this so that e.g. the spoilers stuck/break if operated above 250km/h or that the wings brake under too heave load. Also, what I've thought of is to add fake spoilers to the wing connected to a rain property decreasing lift and increasing drag when flying in rain :D **Model** could be increased to 4: Accurate 3D model with animated control surfaces, gear, prop, livery support (if applicable). or maybe even 5 as we already have shader effects etc. but I think for a 5 we should improve the exterior model a bit further. About the **FDM** I'm not sure how realistic it is.
True
Rework Aircraft Rating? - In the -set file as well as on the wiki, we still have the following rating **_FDM_**: 2: FDM tuned for cruise configuration. **_Systems_**: 2: Working electrical system, fuel feed cockpit controls, stable autopilot **_Cockpit_**: 2: 2D panel in 3D cockpit, or incomplete 3D panel **_Model_**: 3: Accurate 3D model with animated control surfaces, gear detailing (retraction, rotation), prop In my opinion this is not quite correct, e.g. **Cockpit** should IMO be improved to a 4 as we have " 3D panel and accurately modelled 3D cockpit, plain texturing. Hotspots for majority of controls" **Systems** should be corrected to 4 (note here that the ASK21 has very few systems to model). I have added a limit system which I will push after tooltips are solved. Later we can also add a bit more code to this so that e.g. the spoilers stuck/break if operated above 250km/h or that the wings brake under too heave load. Also, what I've thought of is to add fake spoilers to the wing connected to a rain property decreasing lift and increasing drag when flying in rain :D **Model** could be increased to 4: Accurate 3D model with animated control surfaces, gear, prop, livery support (if applicable). or maybe even 5 as we already have shader effects etc. but I think for a 5 we should improve the exterior model a bit further. About the **FDM** I'm not sure how realistic it is.
main
rework aircraft rating in the set file as well as on the wiki we still have the following rating fdm fdm tuned for cruise configuration systems working electrical system fuel feed cockpit controls stable autopilot cockpit panel in cockpit or incomplete panel model accurate model with animated control surfaces gear detailing retraction rotation prop in my opinion this is not quite correct e g cockpit should imo be improved to a as we have panel and accurately modelled cockpit plain texturing hotspots for majority of controls systems should be corrected to note here that the has very few systems to model i have added a limit system which i will push after tooltips are solved later we can also add a bit more code to this so that e g the spoilers stuck break if operated above h or that the wings brake under too heave load also what i ve thought of is to add fake spoilers to the wing connected to a rain property decreasing lift and increasing drag when flying in rain d model could be increased to accurate model with animated control surfaces gear prop livery support if applicable or maybe even as we already have shader effects etc but i think for a we should improve the exterior model a bit further about the fdm i m not sure how realistic it is
1
125,266
16,750,588,191
IssuesEvent
2021-06-11 22:21:01
tlswg/draft-ietf-tls-esni
https://api.github.com/repos/tlswg/draft-ietf-tls-esni
closed
ECH contradicts second ClientHello consistency requirements in RFC8446
design
I just noticed this as I was pondering our various HRR issues. RFC8446, section 4.1.2 says: > [...] The client will also send a ClientHello when the server has responded to its ClientHello with a HelloRetryRequest. In that case, the client MUST send the same ClientHello without modification, except as follows: > > * [...various cases regarding existing extensions...] > > * Other modifications that may be allowed by an extension defined in the future *and present in the HelloRetryRequest*. https://tools.ietf.org/html/rfc8446#section-4.1.2 The "and present in the HelloRetryRequest" portion is fun. In RFC8446 as written, we're not allowed to change the `encrypted_client_hello` extension on the second ClientHello unless the server sent `encrypted_client_hello` in HelloRetryRequest, which it doesn't. With the protocol as-is, it seems we'll at least need some text to deal with the contradiction. The impact is practice is thankfully limited. If the servers does not enforce consistency between the two ClientHellos, it doesn't matter. (They're not required to, but they're not explicitly forbidden from it either, and a stateless server may need to enforce some minimum consistency to avoid getting into a broken state.) If a server does, ECH's behavior on HRR will break it. Such a server is supposed to handshake with ClientHelloOuter, so it just affects the retry flow. That's maybe okay, but mildly annoying. Administrators using server software that checks need to know to first upgrade to a version that doesn't check before trying to deploy ECH. We can also add an ECH extension to HRR. Then, on ECH-reject + HRR, the server perhaps wouldn't send the extension and the client would be required to replay the extension. This is a waste in many ways, but would work. This would break don't stick out, but some threat models, don't stick out with HRR is hopeless anyway. (See https://github.com/tlswg/draft-ietf-tls-esni/pull/352#issuecomment-721824226.) It would also be weirdly asymmetric between ECH and SH. That said, this constraint is also pretty obnoxious. For instance, #313 currently adds a new padding extension codepoint, because the old one isn't defined for other messages. However, the RFC8446 allowance only works for the existing padding code point, so you don't want to send any other code points in ClientHello, lest the server force you to keep it unchanged in ClientHello2. So maybe we should confirm no servers did the fully strict check and undo that constraint? (Although, for #313, I personally favor the padding message strategy anyway.)
1.0
ECH contradicts second ClientHello consistency requirements in RFC8446 - I just noticed this as I was pondering our various HRR issues. RFC8446, section 4.1.2 says: > [...] The client will also send a ClientHello when the server has responded to its ClientHello with a HelloRetryRequest. In that case, the client MUST send the same ClientHello without modification, except as follows: > > * [...various cases regarding existing extensions...] > > * Other modifications that may be allowed by an extension defined in the future *and present in the HelloRetryRequest*. https://tools.ietf.org/html/rfc8446#section-4.1.2 The "and present in the HelloRetryRequest" portion is fun. In RFC8446 as written, we're not allowed to change the `encrypted_client_hello` extension on the second ClientHello unless the server sent `encrypted_client_hello` in HelloRetryRequest, which it doesn't. With the protocol as-is, it seems we'll at least need some text to deal with the contradiction. The impact is practice is thankfully limited. If the servers does not enforce consistency between the two ClientHellos, it doesn't matter. (They're not required to, but they're not explicitly forbidden from it either, and a stateless server may need to enforce some minimum consistency to avoid getting into a broken state.) If a server does, ECH's behavior on HRR will break it. Such a server is supposed to handshake with ClientHelloOuter, so it just affects the retry flow. That's maybe okay, but mildly annoying. Administrators using server software that checks need to know to first upgrade to a version that doesn't check before trying to deploy ECH. We can also add an ECH extension to HRR. Then, on ECH-reject + HRR, the server perhaps wouldn't send the extension and the client would be required to replay the extension. This is a waste in many ways, but would work. This would break don't stick out, but some threat models, don't stick out with HRR is hopeless anyway. (See https://github.com/tlswg/draft-ietf-tls-esni/pull/352#issuecomment-721824226.) It would also be weirdly asymmetric between ECH and SH. That said, this constraint is also pretty obnoxious. For instance, #313 currently adds a new padding extension codepoint, because the old one isn't defined for other messages. However, the RFC8446 allowance only works for the existing padding code point, so you don't want to send any other code points in ClientHello, lest the server force you to keep it unchanged in ClientHello2. So maybe we should confirm no servers did the fully strict check and undo that constraint? (Although, for #313, I personally favor the padding message strategy anyway.)
non_main
ech contradicts second clienthello consistency requirements in i just noticed this as i was pondering our various hrr issues section says the client will also send a clienthello when the server has responded to its clienthello with a helloretryrequest in that case the client must send the same clienthello without modification except as follows other modifications that may be allowed by an extension defined in the future and present in the helloretryrequest the and present in the helloretryrequest portion is fun in as written we re not allowed to change the encrypted client hello extension on the second clienthello unless the server sent encrypted client hello in helloretryrequest which it doesn t with the protocol as is it seems we ll at least need some text to deal with the contradiction the impact is practice is thankfully limited if the servers does not enforce consistency between the two clienthellos it doesn t matter they re not required to but they re not explicitly forbidden from it either and a stateless server may need to enforce some minimum consistency to avoid getting into a broken state if a server does ech s behavior on hrr will break it such a server is supposed to handshake with clienthelloouter so it just affects the retry flow that s maybe okay but mildly annoying administrators using server software that checks need to know to first upgrade to a version that doesn t check before trying to deploy ech we can also add an ech extension to hrr then on ech reject hrr the server perhaps wouldn t send the extension and the client would be required to replay the extension this is a waste in many ways but would work this would break don t stick out but some threat models don t stick out with hrr is hopeless anyway see it would also be weirdly asymmetric between ech and sh that said this constraint is also pretty obnoxious for instance currently adds a new padding extension codepoint because the old one isn t defined for other messages however the allowance only works for the existing padding code point so you don t want to send any other code points in clienthello lest the server force you to keep it unchanged in so maybe we should confirm no servers did the fully strict check and undo that constraint although for i personally favor the padding message strategy anyway
0
15,093
9,761,316,970
IssuesEvent
2019-06-05 08:28:57
Ericsson/codechecker
https://api.github.com/repos/Ericsson/codechecker
opened
Fail the CI if certain warnings are found
CLI :computer: analyzer :chart_with_upwards_trend: discussion :bulb: enhancement :star2: usability :+1:
**Is your feature request related to a problem? Please describe.** This feature request was mentioned on cfe-dev mailing list and the use-case is already supported by scan-build: http://lists.llvm.org/pipermail/cfe-dev/2019-June/062495.html **Describe the solution you'd like** It should be possible to treat certain checkers or compiler warnings as compilation errors. **Describe alternatives you've considered** Alternatively one could use parse to get the list of warnings from the run and process the result with a script. This is, however, not user friendly at all. We should not require the users to maintain additional scripts for such a simple task.
True
Fail the CI if certain warnings are found - **Is your feature request related to a problem? Please describe.** This feature request was mentioned on cfe-dev mailing list and the use-case is already supported by scan-build: http://lists.llvm.org/pipermail/cfe-dev/2019-June/062495.html **Describe the solution you'd like** It should be possible to treat certain checkers or compiler warnings as compilation errors. **Describe alternatives you've considered** Alternatively one could use parse to get the list of warnings from the run and process the result with a script. This is, however, not user friendly at all. We should not require the users to maintain additional scripts for such a simple task.
non_main
fail the ci if certain warnings are found is your feature request related to a problem please describe this feature request was mentioned on cfe dev mailing list and the use case is already supported by scan build describe the solution you d like it should be possible to treat certain checkers or compiler warnings as compilation errors describe alternatives you ve considered alternatively one could use parse to get the list of warnings from the run and process the result with a script this is however not user friendly at all we should not require the users to maintain additional scripts for such a simple task
0
605,057
18,724,013,872
IssuesEvent
2021-11-03 14:38:53
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
reopened
www.androidcentral.com - Website theme does not change after switching themes
browser-firefox-focus priority-normal severity-important engine-gecko
<!-- @browser: Firefox Mobile 93.0 --> <!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:93.0) Gecko/93.0 Firefox/93.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/92153 --> **URL**: https://www.androidcentral.com/google-pixel-6-screen-flickering-issues **Browser / Version**: Firefox Mobile 93.0 **Operating System**: Android 11 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: when phone is tilted (to landscape mode from portrait), website's theme changes to dark mode and stays in dark mode even when dark mode is not chosen on device or on site. Happens with some other sites too but not in other browsers. **Steps to Reproduce**: When phone is tilted (to landscape mode from portrait), then website's theme suddenly changes to dark mode and stays in dark mode even when dark mode is not chosen on device or on site. Happens with some other sites too which may have dark mode support, but not in other browsers. To see bug, visit reported site on the Focus browser when device is NOT in dark mode, then once page loads, tilt the phone to landscape. Then website suddenly goes into dark mode. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/11/e401e3fb-5cec-4840-a1d8-62a5c63ba65b.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210927210923</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/11/f829b57b-28ae-41a2-9b63-3d5e4fb246df) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.androidcentral.com - Website theme does not change after switching themes - <!-- @browser: Firefox Mobile 93.0 --> <!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:93.0) Gecko/93.0 Firefox/93.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/92153 --> **URL**: https://www.androidcentral.com/google-pixel-6-screen-flickering-issues **Browser / Version**: Firefox Mobile 93.0 **Operating System**: Android 11 **Tested Another Browser**: Yes Chrome **Problem type**: Something else **Description**: when phone is tilted (to landscape mode from portrait), website's theme changes to dark mode and stays in dark mode even when dark mode is not chosen on device or on site. Happens with some other sites too but not in other browsers. **Steps to Reproduce**: When phone is tilted (to landscape mode from portrait), then website's theme suddenly changes to dark mode and stays in dark mode even when dark mode is not chosen on device or on site. Happens with some other sites too which may have dark mode support, but not in other browsers. To see bug, visit reported site on the Focus browser when device is NOT in dark mode, then once page loads, tilt the phone to landscape. Then website suddenly goes into dark mode. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2021/11/e401e3fb-5cec-4840-a1d8-62a5c63ba65b.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20210927210923</li><li>channel: release</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2021/11/f829b57b-28ae-41a2-9b63-3d5e4fb246df) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_main
website theme does not change after switching themes url browser version firefox mobile operating system android tested another browser yes chrome problem type something else description when phone is tilted to landscape mode from portrait website s theme changes to dark mode and stays in dark mode even when dark mode is not chosen on device or on site happens with some other sites too but not in other browsers steps to reproduce when phone is tilted to landscape mode from portrait then website s theme suddenly changes to dark mode and stays in dark mode even when dark mode is not chosen on device or on site happens with some other sites too which may have dark mode support but not in other browsers to see bug visit reported site on the focus browser when device is not in dark mode then once page loads tilt the phone to landscape then website suddenly goes into dark mode view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel release hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
564,077
16,707,174,169
IssuesEvent
2021-06-09 11:25:26
HSLdevcom/bultti
https://api.github.com/repos/HSLdevcom/bultti
opened
Fix downloading of two reports
Priority 3 bug
Fixes to be done: - Fix downloading "Päästöluokkien suoritteet excel raport" - Fix downloading "Ajantasauspysäkkien laiminlyönnit"
1.0
Fix downloading of two reports - Fixes to be done: - Fix downloading "Päästöluokkien suoritteet excel raport" - Fix downloading "Ajantasauspysäkkien laiminlyönnit"
non_main
fix downloading of two reports fixes to be done fix downloading päästöluokkien suoritteet excel raport fix downloading ajantasauspysäkkien laiminlyönnit
0
289,178
8,861,307,631
IssuesEvent
2019-01-10 00:01:14
chef/chef
https://api.github.com/repos/chef/chef
closed
Old-style provider map is breaking with two cookbooks that both use Chef::Platform.set
Component: Chef Core Priority: Medium Type: Bug Type: Regression
I ran across a problem on Friday where two dependencies, of a cookbook I'm working on, both had a file `libraries/z_provider_mapping.rb` that contained calls to `Chef::Set.platform`. One was [mysql2_chef_gem](https://github.com/sinfomicien/mysql2_chef_gem/blob/master/libraries/z_provider_mapping.rb) and the other was [automatic_updates](https://github.com/rackspace-cookbooks/automatic_updates/blob/master/libraries/z_provider_mapping.rb). While I know the method used by the cookbooks isn't preferred in Chef 12.4.x, there's a lot of cookbooks attempting to maintain compatibility with Chef 11.x by doing things this way. However, when two cookbooks both use this method, I was seeing that `automatic_updates` was breaking the provider map entries from `mysql2_chef_gem`. This led to many instances of `wrong number of arguments (2 for 0)` because Chef would fall through to matching providers by class name, and mysql2_chef_gem has an abstract class at the same name and level of the class hierarchy (so the class selected by name was invalid). I believe this is happening to many other folks using mysql2_chef_gem, e.g. https://github.com/sinfomicien/mysql2_chef_gem/issues/6, https://github.com/sinfomicien/mysql2_chef_gem/issues/5, https://github.com/sinfomicien/mysql2_chef_gem/issues/4. While I know that the 'correct' thing to do is update cookbooks to use Chef 12-style `provides:`, I think something is still broken in Chef 12.4.1 when it comes to maintaining compatibility of the platform map behavior. I've created an [example cookbook to demonstrate this problem](https://github.com/martinb3/chef_provider_map_bug). In that example, check out the following [libraries/z_provider_map.rb](https://github.com/martinb3/chef_provider_map_bug/blob/master/libraries/z_provider_map.rb) (I've pasted the output back into the source, when run on chef 12.4.1): ``` pp "Chef::Platform.send(:platforms) = #{Chef::Platform.send(:platforms)}" # Output: "Chef::Platform.send(:platforms) = {:default=>{}}" pp "Setting platform mapping for :foo" Chef::Platform.set platform: :centos, version: '< 7.0', resource: :foo, provider: Chef::Provider::Foo::Beta Chef::Platform.set platform: :centos, version: '>= 7.0', resource: :foo, provider: Chef::Provider::Foo::Alpha Chef::Platform.set platform: :ubuntu, resource: :foo, provider: Chef::Provider::Foo::Beta pp "Chef::Platform.send(:platforms) = #{Chef::Platform.send(:platforms)}" # Output: # Chef::Platform.send(:platforms) = { # :default=>{}, # :centos=>{\"< 7.0\"=>{:foo=>Chef::Provider::Foo::Beta}, \">= 7.0\"=>{:foo=>Chef::Provider::Foo::Alpha}}, # :ubuntu=>{:default=>{:foo=>Chef::Provider::Foo::Beta}} # } pp "Setting platform mapping for :bar" Chef::Platform.set platform: :centos, resource: :bar, provider: Chef::Provider::Bar Chef::Platform.set platform: :ubuntu, resource: :bar, provider: Chef::Provider::Bar pp "Chef::Platform.send(:platforms) = #{Chef::Platform.send(:platforms)}" # Output # Chef::Platform.send(:platforms) = { # :default=>{}, # :centos=>{:default=>{:bar=>Chef::Provider::Bar}}, # :ubuntu=>{:default=>{:foo=>Chef::Provider::Foo::Beta, :bar=>Chef::Provider::Bar}} # } ``` It seems like the second set of `Chef::Platform#set` calls are wiping out map entries created by the first set of calls. I suspect that either the cookbooks I referenced above are 'doing it wrong' for Chef 11.x compatibility, or there is a subtle bug in the (https://github.com/chef/chef/blob/12.4-stable/lib/chef/platform/provider_mapping.rb#L109)[set method] of `platform/provider_mapping.rb`. It's pretty clear that [lines like this one](https://github.com/chef/chef/blob/12.4-stable/lib/chef/platform/provider_mapping.rb#L135) will wipe out any existing map entries when called. Is this a mistake in usage where cookbooks from the Chef-11 era shouldn't be using `Chef::Platform.set` for backwards compatibility in Chef 12? Or is this a case where the z_provider_mapping.rb pattern is okay, but the methods that manipulate the platform map are not supposed to be deleting previous entries? If you read down this far, thank you! /CC some of my colleagues, @jarosser06 /CC some interested parties that are relevant for the commits on this code, @lamont-granquist @jkeiser @danielsdeleo @coderanger
1.0
Old-style provider map is breaking with two cookbooks that both use Chef::Platform.set - I ran across a problem on Friday where two dependencies, of a cookbook I'm working on, both had a file `libraries/z_provider_mapping.rb` that contained calls to `Chef::Set.platform`. One was [mysql2_chef_gem](https://github.com/sinfomicien/mysql2_chef_gem/blob/master/libraries/z_provider_mapping.rb) and the other was [automatic_updates](https://github.com/rackspace-cookbooks/automatic_updates/blob/master/libraries/z_provider_mapping.rb). While I know the method used by the cookbooks isn't preferred in Chef 12.4.x, there's a lot of cookbooks attempting to maintain compatibility with Chef 11.x by doing things this way. However, when two cookbooks both use this method, I was seeing that `automatic_updates` was breaking the provider map entries from `mysql2_chef_gem`. This led to many instances of `wrong number of arguments (2 for 0)` because Chef would fall through to matching providers by class name, and mysql2_chef_gem has an abstract class at the same name and level of the class hierarchy (so the class selected by name was invalid). I believe this is happening to many other folks using mysql2_chef_gem, e.g. https://github.com/sinfomicien/mysql2_chef_gem/issues/6, https://github.com/sinfomicien/mysql2_chef_gem/issues/5, https://github.com/sinfomicien/mysql2_chef_gem/issues/4. While I know that the 'correct' thing to do is update cookbooks to use Chef 12-style `provides:`, I think something is still broken in Chef 12.4.1 when it comes to maintaining compatibility of the platform map behavior. I've created an [example cookbook to demonstrate this problem](https://github.com/martinb3/chef_provider_map_bug). In that example, check out the following [libraries/z_provider_map.rb](https://github.com/martinb3/chef_provider_map_bug/blob/master/libraries/z_provider_map.rb) (I've pasted the output back into the source, when run on chef 12.4.1): ``` pp "Chef::Platform.send(:platforms) = #{Chef::Platform.send(:platforms)}" # Output: "Chef::Platform.send(:platforms) = {:default=>{}}" pp "Setting platform mapping for :foo" Chef::Platform.set platform: :centos, version: '< 7.0', resource: :foo, provider: Chef::Provider::Foo::Beta Chef::Platform.set platform: :centos, version: '>= 7.0', resource: :foo, provider: Chef::Provider::Foo::Alpha Chef::Platform.set platform: :ubuntu, resource: :foo, provider: Chef::Provider::Foo::Beta pp "Chef::Platform.send(:platforms) = #{Chef::Platform.send(:platforms)}" # Output: # Chef::Platform.send(:platforms) = { # :default=>{}, # :centos=>{\"< 7.0\"=>{:foo=>Chef::Provider::Foo::Beta}, \">= 7.0\"=>{:foo=>Chef::Provider::Foo::Alpha}}, # :ubuntu=>{:default=>{:foo=>Chef::Provider::Foo::Beta}} # } pp "Setting platform mapping for :bar" Chef::Platform.set platform: :centos, resource: :bar, provider: Chef::Provider::Bar Chef::Platform.set platform: :ubuntu, resource: :bar, provider: Chef::Provider::Bar pp "Chef::Platform.send(:platforms) = #{Chef::Platform.send(:platforms)}" # Output # Chef::Platform.send(:platforms) = { # :default=>{}, # :centos=>{:default=>{:bar=>Chef::Provider::Bar}}, # :ubuntu=>{:default=>{:foo=>Chef::Provider::Foo::Beta, :bar=>Chef::Provider::Bar}} # } ``` It seems like the second set of `Chef::Platform#set` calls are wiping out map entries created by the first set of calls. I suspect that either the cookbooks I referenced above are 'doing it wrong' for Chef 11.x compatibility, or there is a subtle bug in the (https://github.com/chef/chef/blob/12.4-stable/lib/chef/platform/provider_mapping.rb#L109)[set method] of `platform/provider_mapping.rb`. It's pretty clear that [lines like this one](https://github.com/chef/chef/blob/12.4-stable/lib/chef/platform/provider_mapping.rb#L135) will wipe out any existing map entries when called. Is this a mistake in usage where cookbooks from the Chef-11 era shouldn't be using `Chef::Platform.set` for backwards compatibility in Chef 12? Or is this a case where the z_provider_mapping.rb pattern is okay, but the methods that manipulate the platform map are not supposed to be deleting previous entries? If you read down this far, thank you! /CC some of my colleagues, @jarosser06 /CC some interested parties that are relevant for the commits on this code, @lamont-granquist @jkeiser @danielsdeleo @coderanger
non_main
old style provider map is breaking with two cookbooks that both use chef platform set i ran across a problem on friday where two dependencies of a cookbook i m working on both had a file libraries z provider mapping rb that contained calls to chef set platform one was and the other was while i know the method used by the cookbooks isn t preferred in chef x there s a lot of cookbooks attempting to maintain compatibility with chef x by doing things this way however when two cookbooks both use this method i was seeing that automatic updates was breaking the provider map entries from chef gem this led to many instances of wrong number of arguments for because chef would fall through to matching providers by class name and chef gem has an abstract class at the same name and level of the class hierarchy so the class selected by name was invalid i believe this is happening to many other folks using chef gem e g while i know that the correct thing to do is update cookbooks to use chef style provides i think something is still broken in chef when it comes to maintaining compatibility of the platform map behavior i ve created an in that example check out the following i ve pasted the output back into the source when run on chef pp chef platform send platforms chef platform send platforms output chef platform send platforms default pp setting platform mapping for foo chef platform set platform centos version resource foo provider chef provider foo beta chef platform set platform centos version resource foo provider chef provider foo alpha chef platform set platform ubuntu resource foo provider chef provider foo beta pp chef platform send platforms chef platform send platforms output chef platform send platforms default centos foo chef provider foo beta foo chef provider foo alpha ubuntu default foo chef provider foo beta pp setting platform mapping for bar chef platform set platform centos resource bar provider chef provider bar chef platform set platform ubuntu resource bar provider chef provider bar pp chef platform send platforms chef platform send platforms output chef platform send platforms default centos default bar chef provider bar ubuntu default foo chef provider foo beta bar chef provider bar it seems like the second set of chef platform set calls are wiping out map entries created by the first set of calls i suspect that either the cookbooks i referenced above are doing it wrong for chef x compatibility or there is a subtle bug in the of platform provider mapping rb it s pretty clear that will wipe out any existing map entries when called is this a mistake in usage where cookbooks from the chef era shouldn t be using chef platform set for backwards compatibility in chef or is this a case where the z provider mapping rb pattern is okay but the methods that manipulate the platform map are not supposed to be deleting previous entries if you read down this far thank you cc some of my colleagues cc some interested parties that are relevant for the commits on this code lamont granquist jkeiser danielsdeleo coderanger
0
1,319
5,654,506,361
IssuesEvent
2017-04-09 09:22:01
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
Bug report: google-cloud-sdk fish completion instructions are wrong
awaiting maintainer feedback
#### Description of issue I installed the Google Cloud SDK cask and followed the install instructions for fish completions: ``` set fish_user_paths /usr/local/Caskroom/google-cloud-sdk/latest/google-cloud-sdk/bin set -x MANPATH /usr/local/Caskroom/google-cloud-sdk/latest/google-cloud-sdk/help/man /usr/local/share/man /usr/share/man /opt/x11/share/man ``` then tried running `fish_update_completions`. I got: ``` $ fish_update_completions No paths specified ``` After removing those two lines from my fish config file, I was able to run `fish_update_completions` and got this correct looking result: ``` $ fish_update_completions Parsing man pages and writing completions to /Users/myuser/.local/share/fish/generated_completions/ 3176 / 3176 : zic.8 ``` It looks like the instructions in #9958 aren't valid anymore possibly? I'm pretty sure that MANPATH should be set something like this, although it should also account for `MANPATH` not existing beforehand: https://github.com/fish-shell/fish-shell/issues/2090 ``` set -x MANPATH /usr/local/Caskroom/google-cloud-sdk/latest/google-cloud-sdk/help/man $MANPATH ``` Currently there is no help directory or manpage subfolder at that location, so that won't work at the moment, and this causes the error `No paths specified`. To summarise there are two problems here: 1. There is no manpage folder at the location suggested. 2. If there was, the MANPATH variable should be set in a way that respects the users existing MANPATH, and allows for MANPATH not previously existing.
True
Bug report: google-cloud-sdk fish completion instructions are wrong - #### Description of issue I installed the Google Cloud SDK cask and followed the install instructions for fish completions: ``` set fish_user_paths /usr/local/Caskroom/google-cloud-sdk/latest/google-cloud-sdk/bin set -x MANPATH /usr/local/Caskroom/google-cloud-sdk/latest/google-cloud-sdk/help/man /usr/local/share/man /usr/share/man /opt/x11/share/man ``` then tried running `fish_update_completions`. I got: ``` $ fish_update_completions No paths specified ``` After removing those two lines from my fish config file, I was able to run `fish_update_completions` and got this correct looking result: ``` $ fish_update_completions Parsing man pages and writing completions to /Users/myuser/.local/share/fish/generated_completions/ 3176 / 3176 : zic.8 ``` It looks like the instructions in #9958 aren't valid anymore possibly? I'm pretty sure that MANPATH should be set something like this, although it should also account for `MANPATH` not existing beforehand: https://github.com/fish-shell/fish-shell/issues/2090 ``` set -x MANPATH /usr/local/Caskroom/google-cloud-sdk/latest/google-cloud-sdk/help/man $MANPATH ``` Currently there is no help directory or manpage subfolder at that location, so that won't work at the moment, and this causes the error `No paths specified`. To summarise there are two problems here: 1. There is no manpage folder at the location suggested. 2. If there was, the MANPATH variable should be set in a way that respects the users existing MANPATH, and allows for MANPATH not previously existing.
main
bug report google cloud sdk fish completion instructions are wrong description of issue i installed the google cloud sdk cask and followed the install instructions for fish completions set fish user paths usr local caskroom google cloud sdk latest google cloud sdk bin set x manpath usr local caskroom google cloud sdk latest google cloud sdk help man usr local share man usr share man opt share man then tried running fish update completions i got fish update completions no paths specified after removing those two lines from my fish config file i was able to run fish update completions and got this correct looking result fish update completions parsing man pages and writing completions to users myuser local share fish generated completions zic it looks like the instructions in aren t valid anymore possibly i m pretty sure that manpath should be set something like this although it should also account for manpath not existing beforehand set x manpath usr local caskroom google cloud sdk latest google cloud sdk help man manpath currently there is no help directory or manpage subfolder at that location so that won t work at the moment and this causes the error no paths specified to summarise there are two problems here there is no manpage folder at the location suggested if there was the manpath variable should be set in a way that respects the users existing manpath and allows for manpath not previously existing
1
3,308
12,810,648,277
IssuesEvent
2020-07-03 19:26:23
petl-developers/petl
https://api.github.com/repos/petl-developers/petl
opened
Test a couple of database drivers and fsspec remote sources with docker
Maintainability help wanted
#### Problem description One interesting approach for testing SQLAlchemy drivers would be using docker container for the CI tests. For example, [fsspec](https://filesystem-spec.readthedocs.io/en/latest/index.html) package uses a [container](https://github.com/intake/filesystem_spec/blob/master/fsspec/implementations/tests/test_sftp.py#L44) as remote server for testing it's SFTP filesystem wrapper using [paramiko](http://docs.paramiko.org/en/stable/) package. This can increase the robustness of `petl` CI because after the version [v1.6.0](https://github.com/petl-developers/petl/releases/tag/v1.6.0) reading and writing remote sources are supported. This also can be done with #453 migration.
True
Test a couple of database drivers and fsspec remote sources with docker - #### Problem description One interesting approach for testing SQLAlchemy drivers would be using docker container for the CI tests. For example, [fsspec](https://filesystem-spec.readthedocs.io/en/latest/index.html) package uses a [container](https://github.com/intake/filesystem_spec/blob/master/fsspec/implementations/tests/test_sftp.py#L44) as remote server for testing it's SFTP filesystem wrapper using [paramiko](http://docs.paramiko.org/en/stable/) package. This can increase the robustness of `petl` CI because after the version [v1.6.0](https://github.com/petl-developers/petl/releases/tag/v1.6.0) reading and writing remote sources are supported. This also can be done with #453 migration.
main
test a couple of database drivers and fsspec remote sources with docker problem description one interesting approach for testing sqlalchemy drivers would be using docker container for the ci tests for example package uses a as remote server for testing it s sftp filesystem wrapper using package this can increase the robustness of petl ci because after the version reading and writing remote sources are supported this also can be done with migration
1
1,236
5,267,075,702
IssuesEvent
2017-02-04 18:59:27
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
refactor `uninstall` tests (Homebrew/brew/blob/master/Library/Homebrew/cask/test/cask/artifact/pkg_test.rb)
awaiting maintainer feedback core enhancement
- At least one of the Travis glitches we see most often (false failure) is a problem in `uninstall` tests in `cask/artifact/pkg_test.rb`. - in #4804 @NanoXD pointed out that we should test for multiple `uninstall` stanzas, which, after a quick check, it looks like we do not. - `cask/artifact/pkg_test.rb` is kind of big/messy. The false failure glitch is some kind of order-dependent thing. It might be that it would just go away if we organized `pkg_test.rb` into more separate sections.
True
refactor `uninstall` tests (Homebrew/brew/blob/master/Library/Homebrew/cask/test/cask/artifact/pkg_test.rb) - - At least one of the Travis glitches we see most often (false failure) is a problem in `uninstall` tests in `cask/artifact/pkg_test.rb`. - in #4804 @NanoXD pointed out that we should test for multiple `uninstall` stanzas, which, after a quick check, it looks like we do not. - `cask/artifact/pkg_test.rb` is kind of big/messy. The false failure glitch is some kind of order-dependent thing. It might be that it would just go away if we organized `pkg_test.rb` into more separate sections.
main
refactor uninstall tests homebrew brew blob master library homebrew cask test cask artifact pkg test rb at least one of the travis glitches we see most often false failure is a problem in uninstall tests in cask artifact pkg test rb in nanoxd pointed out that we should test for multiple uninstall stanzas which after a quick check it looks like we do not cask artifact pkg test rb is kind of big messy the false failure glitch is some kind of order dependent thing it might be that it would just go away if we organized pkg test rb into more separate sections
1
4,638
24,021,734,160
IssuesEvent
2022-09-15 08:14:27
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Minified Javascript files in the repository
bug UI maintainability packaging
[Posted by Markus Koschany on the openrefine-dev mailing list](https://groups.google.com/g/openrefine-dev/c/eJbPPYnqJiw): > The webapp includes several minified Javascript files without providing the > original non-minified source files. This is acceptable and in accordance with > the license of those files but it is a reason to reject a package for Debian's > ftp team because Debian requires the original sources. I suggest to include the > corresponding non-minified Javascript files as well. Actually I believe > minified JS files don't improve the performance that much since OpenRefine is > run on the user's local system. You could remove them completely and use the > non-minified versions instead. There are some files like date.js which appear > to be minified Javascript as well despite the missing min.js suffix. We should therefore reevaluate whether we can use non-minified files instead. The real problem here is our lack of a proper dependency management system for our frontend, #2423.
True
Minified Javascript files in the repository - [Posted by Markus Koschany on the openrefine-dev mailing list](https://groups.google.com/g/openrefine-dev/c/eJbPPYnqJiw): > The webapp includes several minified Javascript files without providing the > original non-minified source files. This is acceptable and in accordance with > the license of those files but it is a reason to reject a package for Debian's > ftp team because Debian requires the original sources. I suggest to include the > corresponding non-minified Javascript files as well. Actually I believe > minified JS files don't improve the performance that much since OpenRefine is > run on the user's local system. You could remove them completely and use the > non-minified versions instead. There are some files like date.js which appear > to be minified Javascript as well despite the missing min.js suffix. We should therefore reevaluate whether we can use non-minified files instead. The real problem here is our lack of a proper dependency management system for our frontend, #2423.
main
minified javascript files in the repository the webapp includes several minified javascript files without providing the original non minified source files this is acceptable and in accordance with the license of those files but it is a reason to reject a package for debian s ftp team because debian requires the original sources i suggest to include the corresponding non minified javascript files as well actually i believe minified js files don t improve the performance that much since openrefine is run on the user s local system you could remove them completely and use the non minified versions instead there are some files like date js which appear to be minified javascript as well despite the missing min js suffix we should therefore reevaluate whether we can use non minified files instead the real problem here is our lack of a proper dependency management system for our frontend
1
3,169
12,226,756,390
IssuesEvent
2020-05-03 12:24:34
gfleetwood/asteres
https://api.github.com/repos/gfleetwood/asteres
opened
Zac-HD/escape-from-automanual-testing (182345998)
Python maintain
https://github.com/Zac-HD/escape-from-automanual-testing A three-hour tutorial on property-based testing with https://hypothesis.works
True
Zac-HD/escape-from-automanual-testing (182345998) - https://github.com/Zac-HD/escape-from-automanual-testing A three-hour tutorial on property-based testing with https://hypothesis.works
main
zac hd escape from automanual testing a three hour tutorial on property based testing with
1
200,253
22,739,534,775
IssuesEvent
2022-07-07 01:23:48
valdisiljuconoks/FeatureSwitch
https://api.github.com/repos/valdisiljuconoks/FeatureSwitch
opened
WS-2020-0450 (Medium) detected in handlebars-1.3.0.tgz
security vulnerability
## WS-2020-0450 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-1.3.0.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz</a></p> <p>Path to dependency file: /EpiSample/modules/_protected/Shell/Shell/11.1.0.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: /EpiSample/modules/_protected/Shell/Shell/11.1.0.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - istanbul-0.2.16.tgz - :x: **handlebars-1.3.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars before 4.6.0 vulnerable to Prototype Pollution. Prototype access to the template engine allows for potential code execution, which may lead to Denial Of Service (DoS). <p>Publish Date: 2020-01-09 <p>URL: <a href=https://github.com/handlebars-lang/handlebars.js/commit/33a3b46bc205f768f8edbc67241c68591fe3472c>WS-2020-0450</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-01-09</p> <p>Fix Resolution: handlebars - 4.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2020-0450 (Medium) detected in handlebars-1.3.0.tgz - ## WS-2020-0450 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-1.3.0.tgz</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz">https://registry.npmjs.org/handlebars/-/handlebars-1.3.0.tgz</a></p> <p>Path to dependency file: /EpiSample/modules/_protected/Shell/Shell/11.1.0.0/ClientResources/lib/xstyle/package.json</p> <p>Path to vulnerable library: /EpiSample/modules/_protected/Shell/Shell/11.1.0.0/ClientResources/lib/xstyle/node_modules/handlebars/package.json</p> <p> Dependency Hierarchy: - intern-geezer-2.2.3.tgz (Root Library) - istanbul-0.2.16.tgz - :x: **handlebars-1.3.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars before 4.6.0 vulnerable to Prototype Pollution. Prototype access to the template engine allows for potential code execution, which may lead to Denial Of Service (DoS). <p>Publish Date: 2020-01-09 <p>URL: <a href=https://github.com/handlebars-lang/handlebars.js/commit/33a3b46bc205f768f8edbc67241c68591fe3472c>WS-2020-0450</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2020-01-09</p> <p>Fix Resolution: handlebars - 4.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
ws medium detected in handlebars tgz ws medium severity vulnerability vulnerable library handlebars tgz handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file episample modules protected shell shell clientresources lib xstyle package json path to vulnerable library episample modules protected shell shell clientresources lib xstyle node modules handlebars package json dependency hierarchy intern geezer tgz root library istanbul tgz x handlebars tgz vulnerable library found in base branch master vulnerability details handlebars before vulnerable to prototype pollution prototype access to the template engine allows for potential code execution which may lead to denial of service dos publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution handlebars step up your open source security game with mend
0
4,318
21,719,410,734
IssuesEvent
2022-05-10 21:37:38
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
closed
[MAINTAIN] r-scater: dependency missing
maintain
<!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** [RcppML](https://cran.r-project.org/web/packages/RcppML/index.html) for r-scater <details> ``` put the output here ``` </details> **Packages (please complete the following information):** - Package Name: [e.g. iqtree] **Description** Add any other context about the problem here.
True
[MAINTAIN] r-scater: dependency missing - <!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** [RcppML](https://cran.r-project.org/web/packages/RcppML/index.html) for r-scater <details> ``` put the output here ``` </details> **Packages (please complete the following information):** - Package Name: [e.g. iqtree] **Description** Add any other context about the problem here.
main
r scater dependency missing please report the error of one package in one issue use multi issues to report multi bugs thanks log of the bug for r scater put the output here packages please complete the following information package name description add any other context about the problem here
1
66,583
16,656,208,385
IssuesEvent
2021-06-05 15:17:36
ghdl/ghdl
https://api.github.com/repos/ghdl/ghdl
closed
Allow running the test suite in tree (without install)
Build: Makefile Enhancement
**Is your feature request related to a problem? Please describe.** Yes, when creating a package for GHDL I should also run the tests, this happens before the install step. It seems like the current testsuite runs with the system GHDL instead of the tree GHDL, requiring us to install before running it. **Describe the solution you'd like** I would like to have a new make target, `test`. When invoked it would run the testsuite with the GHDL binary, and the other required files if needed, from the build tree. ```sh make test ``` **Describe alternatives you've considered** I could install GHDL in a temporary destination and run the test suite from there, but this is hacky and not the great solution. I could also try to pass the correct environment to the testsuite but this is not trivial and backend dependent (eg. it would be different between a gcc and a llvm build). **Additional context** It is common practice to have the build system provide a test target, instead of running the tests from the system.
1.0
Allow running the test suite in tree (without install) - **Is your feature request related to a problem? Please describe.** Yes, when creating a package for GHDL I should also run the tests, this happens before the install step. It seems like the current testsuite runs with the system GHDL instead of the tree GHDL, requiring us to install before running it. **Describe the solution you'd like** I would like to have a new make target, `test`. When invoked it would run the testsuite with the GHDL binary, and the other required files if needed, from the build tree. ```sh make test ``` **Describe alternatives you've considered** I could install GHDL in a temporary destination and run the test suite from there, but this is hacky and not the great solution. I could also try to pass the correct environment to the testsuite but this is not trivial and backend dependent (eg. it would be different between a gcc and a llvm build). **Additional context** It is common practice to have the build system provide a test target, instead of running the tests from the system.
non_main
allow running the test suite in tree without install is your feature request related to a problem please describe yes when creating a package for ghdl i should also run the tests this happens before the install step it seems like the current testsuite runs with the system ghdl instead of the tree ghdl requiring us to install before running it describe the solution you d like i would like to have a new make target test when invoked it would run the testsuite with the ghdl binary and the other required files if needed from the build tree sh make test describe alternatives you ve considered i could install ghdl in a temporary destination and run the test suite from there but this is hacky and not the great solution i could also try to pass the correct environment to the testsuite but this is not trivial and backend dependent eg it would be different between a gcc and a llvm build additional context it is common practice to have the build system provide a test target instead of running the tests from the system
0
34,749
16,671,739,939
IssuesEvent
2021-06-07 11:47:45
chakra-ui/chakra-ui
https://api.github.com/repos/chakra-ui/chakra-ui
closed
Static CSS Extraction
Topic: Performance Type: Feature 🚀 styled system 💅🏻
# 🚀 Feature request It'd be amazing to enhance Chakra UI with a zero-runtime CSS in JS library or even Tailwind CSS to improve performance in general. ## 🧱 Problem Statement / Justification The problem is very obvious: https://chakra-ui.com/docs/comparison#the-runtime-trade-off-%EF%B8%8F > Since Chakra UI uses CSS-in-JS under the hood (emotion + styled-system), this flexibility comes with a small price to pay when it comes to runtime. This runtime footprint is caused by style computations by styled-system, and className generation by emotion. > > If your app deals with high, frequently changing data that is performance sensitive, you might notice this footprint as your app grows. ## ✅ Proposed solution or API There was a discussion around it and I want to re-open it because I think this is very important: https://github.com/chakra-ui/chakra-ui/issues/859 ## ↩️ Alternatives I'm considering to use [Linaria](https://linaria.dev/) or [TailwindCSS](https://tailwindcss.com/), but that would require me to create my own UI library on top of it, it'd be just amazing to have Chakra UI replacing Emotion + Styled-System in favor of Linaria/TailwindCSS, but keeping the [ThemeUI specification](https://theme-ui.com/theme-spec/). ## 📝 Additional Information - https://calendar.perfplanet.com/2019/the-unseen-performance-costs-of-css-in-js-in-react-apps - https://daily.dev/blog/why-i-moved-from-styled-components-to-tailwind-css-and-whats-the-future-of-css-in-js - https://leerob.io/blog/tailwind
True
Static CSS Extraction - # 🚀 Feature request It'd be amazing to enhance Chakra UI with a zero-runtime CSS in JS library or even Tailwind CSS to improve performance in general. ## 🧱 Problem Statement / Justification The problem is very obvious: https://chakra-ui.com/docs/comparison#the-runtime-trade-off-%EF%B8%8F > Since Chakra UI uses CSS-in-JS under the hood (emotion + styled-system), this flexibility comes with a small price to pay when it comes to runtime. This runtime footprint is caused by style computations by styled-system, and className generation by emotion. > > If your app deals with high, frequently changing data that is performance sensitive, you might notice this footprint as your app grows. ## ✅ Proposed solution or API There was a discussion around it and I want to re-open it because I think this is very important: https://github.com/chakra-ui/chakra-ui/issues/859 ## ↩️ Alternatives I'm considering to use [Linaria](https://linaria.dev/) or [TailwindCSS](https://tailwindcss.com/), but that would require me to create my own UI library on top of it, it'd be just amazing to have Chakra UI replacing Emotion + Styled-System in favor of Linaria/TailwindCSS, but keeping the [ThemeUI specification](https://theme-ui.com/theme-spec/). ## 📝 Additional Information - https://calendar.perfplanet.com/2019/the-unseen-performance-costs-of-css-in-js-in-react-apps - https://daily.dev/blog/why-i-moved-from-styled-components-to-tailwind-css-and-whats-the-future-of-css-in-js - https://leerob.io/blog/tailwind
non_main
static css extraction 🚀 feature request it d be amazing to enhance chakra ui with a zero runtime css in js library or even tailwind css to improve performance in general 🧱 problem statement justification the problem is very obvious since chakra ui uses css in js under the hood emotion styled system this flexibility comes with a small price to pay when it comes to runtime this runtime footprint is caused by style computations by styled system and classname generation by emotion if your app deals with high frequently changing data that is performance sensitive you might notice this footprint as your app grows ✅ proposed solution or api there was a discussion around it and i want to re open it because i think this is very important ↩️ alternatives i m considering to use or but that would require me to create my own ui library on top of it it d be just amazing to have chakra ui replacing emotion styled system in favor of linaria tailwindcss but keeping the 📝 additional information
0
1,731
6,574,838,367
IssuesEvent
2017-09-11 14:14:47
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Network modules should provide the actual device CLI interaction logging (raw output)
affects_2.3 feature_idea networking waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> network, netcli, ios_command.py, ios_config.py ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 (devel 15ed88547f) last updated 2016/10/25 15:34:39 (GMT -400) lib/ansible/modules/core: (detached HEAD 124bb92416) last updated 2016/10/25 15:34:40 (GMT -400) lib/ansible/modules/extras: (detached HEAD 8ffe314ea5) last updated 2016/10/25 15:34:40 (GMT -400) ``` ##### SUMMARY Currently the network modules do not return any device debugging or raw device output. This is a concern on one hand because network folk are reluctant to run tools on their network without knowing exactly what they are doing on their gear. This is also a challenge when troubleshooting why commands are not working properly with devices that may have multi-level prompts, etc. Being able to see the actual output from the device is key to this. Yes you can later run commands manually to figure it out, but ansible should be able to provide more informative errors. I have started down the path to add this feature to my branch, but as the path to do so is not clear and may require some rework across a number of modules, I would solicit input from the team about how they might like to see this approached. Currently the execute() functions for example only returns the output. Returning a dictionary would offer more flexibility, but there are a lot of places that would need to be changed to support that. There are also two code paths that need to handle this extra data (The normal and the Exceptions path..) I did a proof of concept by adding this to a stderr value when exceptions occur and then had to follow the path from ShellError to NetworkError for example to get this passed back. As execute() currently only returns one output, I simply tacked the debug onto that for a quick test. A couple of examples of the output from my proof are below for added context. Again, I welcome feedback on this feature and can work on it if there is some consensus. Likely I would expect an added parameter in the playbook that turns this feature on, but also perhaps when errors occur the output from the device should be available for troubleshooting. Whether it is called 'stderr' or 'raw' or 'output' etc is also an open question. <!--- Explain the problem briefly --> ##### EXPECTED RESULTS ``` "msg": "timeout waiting for next prompt from remote device", "stderr": "Ciscodev# copy running-config flash:test\r\nDestination filename [test]? ", "stderr_lines": [ [ "ciscodev# copy running-config flash:test\r", "Destination filename [test]?" ] ] ```
True
Network modules should provide the actual device CLI interaction logging (raw output) - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea ##### COMPONENT NAME <!--- Name of the plugin/module/task --> network, netcli, ios_command.py, ios_config.py ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.3.0 (devel 15ed88547f) last updated 2016/10/25 15:34:39 (GMT -400) lib/ansible/modules/core: (detached HEAD 124bb92416) last updated 2016/10/25 15:34:40 (GMT -400) lib/ansible/modules/extras: (detached HEAD 8ffe314ea5) last updated 2016/10/25 15:34:40 (GMT -400) ``` ##### SUMMARY Currently the network modules do not return any device debugging or raw device output. This is a concern on one hand because network folk are reluctant to run tools on their network without knowing exactly what they are doing on their gear. This is also a challenge when troubleshooting why commands are not working properly with devices that may have multi-level prompts, etc. Being able to see the actual output from the device is key to this. Yes you can later run commands manually to figure it out, but ansible should be able to provide more informative errors. I have started down the path to add this feature to my branch, but as the path to do so is not clear and may require some rework across a number of modules, I would solicit input from the team about how they might like to see this approached. Currently the execute() functions for example only returns the output. Returning a dictionary would offer more flexibility, but there are a lot of places that would need to be changed to support that. There are also two code paths that need to handle this extra data (The normal and the Exceptions path..) I did a proof of concept by adding this to a stderr value when exceptions occur and then had to follow the path from ShellError to NetworkError for example to get this passed back. As execute() currently only returns one output, I simply tacked the debug onto that for a quick test. A couple of examples of the output from my proof are below for added context. Again, I welcome feedback on this feature and can work on it if there is some consensus. Likely I would expect an added parameter in the playbook that turns this feature on, but also perhaps when errors occur the output from the device should be available for troubleshooting. Whether it is called 'stderr' or 'raw' or 'output' etc is also an open question. <!--- Explain the problem briefly --> ##### EXPECTED RESULTS ``` "msg": "timeout waiting for next prompt from remote device", "stderr": "Ciscodev# copy running-config flash:test\r\nDestination filename [test]? ", "stderr_lines": [ [ "ciscodev# copy running-config flash:test\r", "Destination filename [test]?" ] ] ```
main
network modules should provide the actual device cli interaction logging raw output issue type feature idea component name network netcli ios command py ios config py ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt summary currently the network modules do not return any device debugging or raw device output this is a concern on one hand because network folk are reluctant to run tools on their network without knowing exactly what they are doing on their gear this is also a challenge when troubleshooting why commands are not working properly with devices that may have multi level prompts etc being able to see the actual output from the device is key to this yes you can later run commands manually to figure it out but ansible should be able to provide more informative errors i have started down the path to add this feature to my branch but as the path to do so is not clear and may require some rework across a number of modules i would solicit input from the team about how they might like to see this approached currently the execute functions for example only returns the output returning a dictionary would offer more flexibility but there are a lot of places that would need to be changed to support that there are also two code paths that need to handle this extra data the normal and the exceptions path i did a proof of concept by adding this to a stderr value when exceptions occur and then had to follow the path from shellerror to networkerror for example to get this passed back as execute currently only returns one output i simply tacked the debug onto that for a quick test a couple of examples of the output from my proof are below for added context again i welcome feedback on this feature and can work on it if there is some consensus likely i would expect an added parameter in the playbook that turns this feature on but also perhaps when errors occur the output from the device should be available for troubleshooting whether it is called stderr or raw or output etc is also an open question expected results msg timeout waiting for next prompt from remote device stderr ciscodev copy running config flash test r ndestination filename stderr lines ciscodev copy running config flash test r destination filename
1
254,302
19,192,123,744
IssuesEvent
2021-12-06 02:51:26
NeuroDesk/neurodesk.github.io
https://api.github.com/repos/NeuroDesk/neurodesk.github.io
closed
add more docker troubleshooting information
documentation help wanted
It is often one of these 4 solutions: Close and open the terminal and retry. Log out and login again, or restart the machine (current user environment doesnt fully know docker is running) Add your user to the OS docker group (current user doesnt have permission to run docker) docker.sock permissions need to be changed (needs some manual intervention to fix) EDIT: Might add these to a troubleshooting section in the docs EDIT: kels's fix: Open docker GUI and accept T&Cs. Check that the docker engine is running To troubleshoot docker, we can try a simpler container first: docker run hello-world
1.0
add more docker troubleshooting information - It is often one of these 4 solutions: Close and open the terminal and retry. Log out and login again, or restart the machine (current user environment doesnt fully know docker is running) Add your user to the OS docker group (current user doesnt have permission to run docker) docker.sock permissions need to be changed (needs some manual intervention to fix) EDIT: Might add these to a troubleshooting section in the docs EDIT: kels's fix: Open docker GUI and accept T&Cs. Check that the docker engine is running To troubleshoot docker, we can try a simpler container first: docker run hello-world
non_main
add more docker troubleshooting information it is often one of these solutions close and open the terminal and retry log out and login again or restart the machine current user environment doesnt fully know docker is running add your user to the os docker group current user doesnt have permission to run docker docker sock permissions need to be changed needs some manual intervention to fix edit might add these to a troubleshooting section in the docs edit kels s fix open docker gui and accept t cs check that the docker engine is running to troubleshoot docker we can try a simpler container first docker run hello world
0