Unnamed: 0
int64
1
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
3
438
labels
stringlengths
4
308
body
stringlengths
7
254k
index
stringclasses
7 values
text_combine
stringlengths
96
254k
label
stringclasses
2 values
text
stringlengths
96
246k
binary_label
int64
0
1
4,702
24,270,811,802
IssuesEvent
2022-09-28 10:06:40
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
closed
SEO Site Audit remediation
needs discussion engineering Epic Maintain
Off the back of the recent Grassriots [SEO audit](https://docs.google.com/spreadsheets/d/15HwgpxSYc4Zl809kcebAhLfLYXFuIk8ZP-Qvk3yVV8Q/edit?usp=sharing), a number of issues have been raised that require resolution.
True
SEO Site Audit remediation - Off the back of the recent Grassriots [SEO audit](https://docs.google.com/spreadsheets/d/15HwgpxSYc4Zl809kcebAhLfLYXFuIk8ZP-Qvk3yVV8Q/edit?usp=sharing), a number of issues have been raised that require resolution.
main
seo site audit remediation off the back of the recent grassriots a number of issues have been raised that require resolution
1
108,189
16,762,812,626
IssuesEvent
2021-06-14 03:15:41
gms-ws-sandbox/nibrs-pr-test
https://api.github.com/repos/gms-ws-sandbox/nibrs-pr-test
opened
CVE-2019-14540 (High) detected in multiple libraries
security vulnerability
## CVE-2019-14540 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.8.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-validate-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.8.0.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: nibrs-pr-test/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.10.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-route/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs-pr-test/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-summary-report-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.5.RELEASE.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs-pr-test/commit/860cc22f54e17594e32e303f0716fb065202fff5">860cc22f54e17594e32e303f0716fb065202fff5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig. <p>Publish Date: 2019-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540>CVE-2019-14540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540</a></p> <p>Release Date: 2019-09-15</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-route/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-validation/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14540","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-14540 (High) detected in multiple libraries - ## CVE-2019-14540 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.6.jar</b>, <b>jackson-databind-2.9.8.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-validate-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.8.0.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: nibrs-pr-test/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.10.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-route/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs-pr-test/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-summary-report-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.5.RELEASE.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs-pr-test/commit/860cc22f54e17594e32e303f0716fb065202fff5">860cc22f54e17594e32e303f0716fb065202fff5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig. <p>Publish Date: 2019-09-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540>CVE-2019-14540</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-14540</a></p> <p>Release Date: 2019-09-15</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-route/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-validation/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.8.11.5,2.9.10,2.10.0.pr3,2.11.0.rc1"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-14540","vulnerabilityDetails":"A Polymorphic Typing issue was discovered in FasterXML jackson-databind before 2.9.10. It is related to com.zaxxer.hikari.HikariConfig.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-14540","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_main
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs validate common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs fbi service pom xml path to vulnerable library nibrs pr test tools nibrs fbi service target nibrs fbi service web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs route pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar nibrs pr test web nibrs web target nibrs web web inf lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs summary report common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikariconfig publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter json release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails a polymorphic typing issue was discovered in fasterxml jackson databind before it is related to com zaxxer hikari hikariconfig vulnerabilityurl
0
221,728
24,657,044,669
IssuesEvent
2022-10-18 01:13:41
jgithaiga/jgithaiga.github.io
https://api.github.com/repos/jgithaiga/jgithaiga.github.io
opened
CVE-2022-37603 (High) detected in loader-utils-2.0.0.tgz
security vulnerability
## CVE-2022-37603 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-2.0.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - gatsby-4.14.0.tgz (Root Library) - raw-loader-4.0.2.tgz - :x: **loader-utils-2.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgithaiga/jgithaiga.github.io/commit/6794c3a4b5723d43cd5e530f985b2dfb944c4c8b">6794c3a4b5723d43cd5e530f985b2dfb944c4c8b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js. <p>Publish Date: 2022-10-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37603>CVE-2022-37603</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-14</p> <p>Fix Resolution (loader-utils): 2.0.1</p> <p>Direct dependency fix Resolution (gatsby): 4.14.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-37603 (High) detected in loader-utils-2.0.0.tgz - ## CVE-2022-37603 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>loader-utils-2.0.0.tgz</b></p></summary> <p>utils for webpack loaders</p> <p>Library home page: <a href="https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz">https://registry.npmjs.org/loader-utils/-/loader-utils-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/loader-utils/package.json</p> <p> Dependency Hierarchy: - gatsby-4.14.0.tgz (Root Library) - raw-loader-4.0.2.tgz - :x: **loader-utils-2.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgithaiga/jgithaiga.github.io/commit/6794c3a4b5723d43cd5e530f985b2dfb944c4c8b">6794c3a4b5723d43cd5e530f985b2dfb944c4c8b</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular expression denial of service (ReDoS) flaw was found in Function interpolateName in interpolateName.js in webpack loader-utils 2.0.0 via the url variable in interpolateName.js. <p>Publish Date: 2022-10-14 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-37603>CVE-2022-37603</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2022-10-14</p> <p>Fix Resolution (loader-utils): 2.0.1</p> <p>Direct dependency fix Resolution (gatsby): 4.14.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in loader utils tgz cve high severity vulnerability vulnerable library loader utils tgz utils for webpack loaders library home page a href path to dependency file package json path to vulnerable library node modules loader utils package json dependency hierarchy gatsby tgz root library raw loader tgz x loader utils tgz vulnerable library found in head commit a href found in base branch master vulnerability details a regular expression denial of service redos flaw was found in function interpolatename in interpolatename js in webpack loader utils via the url variable in interpolatename js publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution loader utils direct dependency fix resolution gatsby step up your open source security game with mend
0
5,872
31,864,561,146
IssuesEvent
2023-09-15 13:18:53
tgstation/tgstation
https://api.github.com/repos/tgstation/tgstation
closed
`COMSIG_HIT_BY_SABOTEUR` returns a boolean which is improper for signals
Maintainability/Hinders improvements
## Reproduction: Signals should not return `TRUE` (or `FALSE`), it should be returning a bitflag.
True
`COMSIG_HIT_BY_SABOTEUR` returns a boolean which is improper for signals - ## Reproduction: Signals should not return `TRUE` (or `FALSE`), it should be returning a bitflag.
main
comsig hit by saboteur returns a boolean which is improper for signals reproduction signals should not return true or false it should be returning a bitflag
1
5,386
27,071,867,029
IssuesEvent
2023-02-14 07:40:43
OpenRefine/OpenRefine
https://api.github.com/repos/OpenRefine/OpenRefine
closed
Unassign contributors automatically after a delay
maintainability
[Our issues labeled "good first issue"](https://github.com/OpenRefine/OpenRefine/issues?q=is%3Aopen+is%3Aissue+label%3A%22good+first+issue%22) tend to attract new contributors, which is great. However, contributors often abandon the issue without unassigning themselves. This means that the pool of good first issues available for prospective contributors shrinks artificially. As a preparation for our participation in Outreachy/GSoC I have cleared assignees of all good first issues today (after checking that they had been assigned for a long time). But I think we should not have to do this manually. There is a GitHub Action which seems to do just that: https://github.com/marketplace/actions/unassign-contributor-after-days-of-inactivity Any resistance to trying this out? As an experiment I would first restrict this to the "good first issue" tag, and set a fairly generous delay - perhaps 3 months?
True
Unassign contributors automatically after a delay - [Our issues labeled "good first issue"](https://github.com/OpenRefine/OpenRefine/issues?q=is%3Aopen+is%3Aissue+label%3A%22good+first+issue%22) tend to attract new contributors, which is great. However, contributors often abandon the issue without unassigning themselves. This means that the pool of good first issues available for prospective contributors shrinks artificially. As a preparation for our participation in Outreachy/GSoC I have cleared assignees of all good first issues today (after checking that they had been assigned for a long time). But I think we should not have to do this manually. There is a GitHub Action which seems to do just that: https://github.com/marketplace/actions/unassign-contributor-after-days-of-inactivity Any resistance to trying this out? As an experiment I would first restrict this to the "good first issue" tag, and set a fairly generous delay - perhaps 3 months?
main
unassign contributors automatically after a delay tend to attract new contributors which is great however contributors often abandon the issue without unassigning themselves this means that the pool of good first issues available for prospective contributors shrinks artificially as a preparation for our participation in outreachy gsoc i have cleared assignees of all good first issues today after checking that they had been assigned for a long time but i think we should not have to do this manually there is a github action which seems to do just that any resistance to trying this out as an experiment i would first restrict this to the good first issue tag and set a fairly generous delay perhaps months
1
256,635
27,561,701,851
IssuesEvent
2023-03-07 22:40:59
samqws-marketing/box_box-ui-elements
https://api.github.com/repos/samqws-marketing/box_box-ui-elements
closed
CVE-2021-29060 (Medium) detected in color-string-1.5.2.tgz - autoclosed
security vulnerability
## CVE-2021-29060 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>color-string-1.5.2.tgz</b></p></summary> <p>Parser and generator for CSS color strings</p> <p>Library home page: <a href="https://registry.npmjs.org/color-string/-/color-string-1.5.2.tgz">https://registry.npmjs.org/color-string/-/color-string-1.5.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/color-string/package.json</p> <p> Dependency Hierarchy: - cssnano-4.1.10.tgz (Root Library) - cssnano-preset-default-4.0.7.tgz - postcss-colormin-4.0.3.tgz - color-3.0.0.tgz - :x: **color-string-1.5.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular Expression Denial of Service (ReDOS) vulnerability was discovered in Color-String version 1.5.5 and below which occurs when the application is provided and checks a crafted invalid HWB string. <p>Publish Date: 2021-06-21 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29060>CVE-2021-29060</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-257v-vj4p-3w2h">https://github.com/advisories/GHSA-257v-vj4p-3w2h</a></p> <p>Release Date: 2021-06-21</p> <p>Fix Resolution (color-string): 1.5.5</p> <p>Direct dependency fix Resolution (cssnano): 4.1.11</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
True
CVE-2021-29060 (Medium) detected in color-string-1.5.2.tgz - autoclosed - ## CVE-2021-29060 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>color-string-1.5.2.tgz</b></p></summary> <p>Parser and generator for CSS color strings</p> <p>Library home page: <a href="https://registry.npmjs.org/color-string/-/color-string-1.5.2.tgz">https://registry.npmjs.org/color-string/-/color-string-1.5.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/color-string/package.json</p> <p> Dependency Hierarchy: - cssnano-4.1.10.tgz (Root Library) - cssnano-preset-default-4.0.7.tgz - postcss-colormin-4.0.3.tgz - color-3.0.0.tgz - :x: **color-string-1.5.2.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_box-ui-elements/commit/4fc776e2b95c8b497f6994cb2165365562ae1f82">4fc776e2b95c8b497f6994cb2165365562ae1f82</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Regular Expression Denial of Service (ReDOS) vulnerability was discovered in Color-String version 1.5.5 and below which occurs when the application is provided and checks a crafted invalid HWB string. <p>Publish Date: 2021-06-21 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-29060>CVE-2021-29060</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-257v-vj4p-3w2h">https://github.com/advisories/GHSA-257v-vj4p-3w2h</a></p> <p>Release Date: 2021-06-21</p> <p>Fix Resolution (color-string): 1.5.5</p> <p>Direct dependency fix Resolution (cssnano): 4.1.11</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END -->
non_main
cve medium detected in color string tgz autoclosed cve medium severity vulnerability vulnerable library color string tgz parser and generator for css color strings library home page a href path to dependency file package json path to vulnerable library node modules color string package json dependency hierarchy cssnano tgz root library cssnano preset default tgz postcss colormin tgz color tgz x color string tgz vulnerable library found in head commit a href found in base branch master vulnerability details a regular expression denial of service redos vulnerability was discovered in color string version and below which occurs when the application is provided and checks a crafted invalid hwb string publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution color string direct dependency fix resolution cssnano check this box to open an automated fix pr
0
489,032
14,100,281,030
IssuesEvent
2020-11-06 03:43:21
PMEAL/OpenPNM
https://api.github.com/repos/PMEAL/OpenPNM
closed
Create new Physics for each algorithm?
enhancement low priority proposal
Like DiffusivePhysics would have all the diffusion models added. This would at least avoid adding a bunch of extra physics to each object that the user doesn't need.
1.0
Create new Physics for each algorithm? - Like DiffusivePhysics would have all the diffusion models added. This would at least avoid adding a bunch of extra physics to each object that the user doesn't need.
non_main
create new physics for each algorithm like diffusivephysics would have all the diffusion models added this would at least avoid adding a bunch of extra physics to each object that the user doesn t need
0
198,028
22,617,851,277
IssuesEvent
2022-06-30 01:15:39
jinhogate/pizza_angular
https://api.github.com/repos/jinhogate/pizza_angular
opened
CVE-2021-23440 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz
security vulnerability
## CVE-2021-23440 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary> <p> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - build-angular-0.6.8.tgz (Root Library) - webpack-4.8.3.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/union-value/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - build-angular-0.6.8.tgz (Root Library) - webpack-4.8.3.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/jinhogate/pizza_angular/commit/353ead6609cbe90c767728d4050746689c22c532">353ead6609cbe90c767728d4050746689c22c532</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays. <p>Publish Date: 2021-09-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23440>CVE-2021-23440</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23440">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23440</a></p> <p>Release Date: 2021-09-12</p> <p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (@angular-devkit/build-angular): 0.7.0</p><p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (@angular-devkit/build-angular): 0.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23440 (High) detected in set-value-2.0.0.tgz, set-value-0.4.3.tgz - ## CVE-2021-23440 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-2.0.0.tgz</b>, <b>set-value-0.4.3.tgz</b></p></summary> <p> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - build-angular-0.6.8.tgz (Root Library) - webpack-4.8.3.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/union-value/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - build-angular-0.6.8.tgz (Root Library) - webpack-4.8.3.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/jinhogate/pizza_angular/commit/353ead6609cbe90c767728d4050746689c22c532">353ead6609cbe90c767728d4050746689c22c532</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> This affects the package set-value before <2.0.1, >=3.0.0 <4.0.1. A type confusion vulnerability can lead to a bypass of CVE-2019-10747 when the user-provided keys used in the path parameter are arrays. <p>Publish Date: 2021-09-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23440>CVE-2021-23440</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23440">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23440</a></p> <p>Release Date: 2021-09-12</p> <p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (@angular-devkit/build-angular): 0.7.0</p><p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (@angular-devkit/build-angular): 0.7.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in set value tgz set value tgz cve high severity vulnerability vulnerable libraries set value tgz set value tgz set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file package json path to vulnerable library node modules set value package json dependency hierarchy build angular tgz root library webpack tgz micromatch tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file package json path to vulnerable library node modules union value node modules set value package json dependency hierarchy build angular tgz root library webpack tgz micromatch tgz snapdragon tgz base tgz cache base tgz union value tgz x set value tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects the package set value before a type confusion vulnerability can lead to a bypass of cve when the user provided keys used in the path parameter are arrays publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution set value direct dependency fix resolution angular devkit build angular fix resolution set value direct dependency fix resolution angular devkit build angular step up your open source security game with mend
0
2,758
9,872,909,715
IssuesEvent
2019-06-22 09:21:12
arcticicestudio/snowsaw
https://api.github.com/repos/arcticicestudio/snowsaw
opened
EditorConfig
context-workflow scope-maintainability type-task
<p align="center"><img src="https://editorconfig.org/logo.png" /></p> Update the [EditorConfig][] file to match _Arctic Ice Studio's_ consistent coding styles between different editors and IDEs. [editorconfig]: https://editorconfig.org
True
EditorConfig - <p align="center"><img src="https://editorconfig.org/logo.png" /></p> Update the [EditorConfig][] file to match _Arctic Ice Studio's_ consistent coding styles between different editors and IDEs. [editorconfig]: https://editorconfig.org
main
editorconfig update the file to match arctic ice studio s consistent coding styles between different editors and ides
1
5,821
30,815,855,016
IssuesEvent
2023-08-01 13:26:21
precice/precice
https://api.github.com/repos/precice/precice
closed
Cleanup preallocation options for PETSc RBF Mappings
maintainability breaking change
We currently have a quite low test coverage in the [PETSc radial-basis function mapping](https://github.com/precice/precice/blob/develop/src/mapping/PetRadialBasisFctMapping.hpp). To be more precise, we only test the `Tree` preallocation strategy. All remaining strategies `Compute`, `Save`, `Estimate`, `None` are not tested at all. See the file on [CodeCov](https://codecov.io/gh/precice/precice/src/develop/src/mapping/PetRadialBasisFctMapping.hpp)
True
Cleanup preallocation options for PETSc RBF Mappings - We currently have a quite low test coverage in the [PETSc radial-basis function mapping](https://github.com/precice/precice/blob/develop/src/mapping/PetRadialBasisFctMapping.hpp). To be more precise, we only test the `Tree` preallocation strategy. All remaining strategies `Compute`, `Save`, `Estimate`, `None` are not tested at all. See the file on [CodeCov](https://codecov.io/gh/precice/precice/src/develop/src/mapping/PetRadialBasisFctMapping.hpp)
main
cleanup preallocation options for petsc rbf mappings we currently have a quite low test coverage in the to be more precise we only test the tree preallocation strategy all remaining strategies compute save estimate none are not tested at all see the file on
1
24,760
17,694,755,195
IssuesEvent
2021-08-24 14:11:42
youngeek-0410/SekigaeService
https://api.github.com/repos/youngeek-0410/SekigaeService
closed
less command in container
type:enhancement task:infrastructure
# 概要 <!-- どこに / どんな改善を施したいのか --> コンテナ内で`less`コマンドが使用できないのでbuild時にinstallするようにする # 改善の背景 <!-- 改善が必要とされる背景 --> logを`cat`で見るのはつらい # 要件 <!-- 期待される動作や満たされるべき内容を具体的に --> # 検討事項 <!-- 改善を施すにあたり検討が必要な事項(あれば) --> - # 実装方針 <!-- どのように実装するのか --> - # テスト項目 <!--改善が正しく実装されたかどうかを確認する方法 --> - <!-- 全て埋めなくても良いが、必要な情報をわかりやすく書く。 -->
1.0
less command in container - # 概要 <!-- どこに / どんな改善を施したいのか --> コンテナ内で`less`コマンドが使用できないのでbuild時にinstallするようにする # 改善の背景 <!-- 改善が必要とされる背景 --> logを`cat`で見るのはつらい # 要件 <!-- 期待される動作や満たされるべき内容を具体的に --> # 検討事項 <!-- 改善を施すにあたり検討が必要な事項(あれば) --> - # 実装方針 <!-- どのように実装するのか --> - # テスト項目 <!--改善が正しく実装されたかどうかを確認する方法 --> - <!-- 全て埋めなくても良いが、必要な情報をわかりやすく書く。 -->
non_main
less command in container 概要 コンテナ内で less コマンドが使用できないのでbuild時にinstallするようにする 改善の背景 logを cat で見るのはつらい 要件 検討事項 実装方針 テスト項目
0
2,871
10,276,010,371
IssuesEvent
2019-08-24 13:33:42
arcticicestudio/arctic
https://api.github.com/repos/arcticicestudio/arctic
closed
Prettier
context-workflow scope-dx scope-maintainability scope-quality type-feature
<p align="center"><img src="https://user-images.githubusercontent.com/7836623/63637792-4dcef380-c681-11e9-9252-f2fb22499985.png" width="30%" /></p> Integrate [Prettier][], the opinionated code formatter with support for many languages and integrations with most editors. It ensures that all outputted code conforms to a consistent style. ### Configuration This is one of the main features of Prettier: It already provides the best and recommended style configurations of-out-the-box™. The only option we will change is the [print width][prettier-docs-pwidth]. It is set to 80 by default which not up-to-date for modern screens (might only be relevant when working in terminals only like e.g. with Vim). It'll be changed to 120 used by all of Arctic Ice Studio's style guides. The `prettier.config.js` configuration file will be placed in the project root as well as the `.prettierignore` file to also define ignore pattern. ### ESLint Compatibility To be fully compatible with ESLint, [eslint-plugin-prettier][gh-eslint-plugin-prettier] has already been included in #30 as well as the [set of recommended rules][gh-eslint-config-prettier] via the [`@arcticicestudio/eslint-config/prettier`][stg-js-esl-c#ep] and [`@arcticicestudio/eslint-config-typescript/prettier`][stg-js-esl-c-ts#ep] extension entry points. ### Package Script To allow to format all sources a `format:pretty` package script will be added that'll also run in the main `format` script flow. A new `lint:pretty` script will also allow to check if all supported files are correct formatted. It will be included in the main `lint` script flow. ## Tasks - ~~Install [prettier][npm-prettier].~~ Resolved in #30 - [x] Implement `prettier.config.js` configuration file. - [x] Implement `.prettierignore` ignore pattern file. - [x] Implement `format:pretty` package script and add to main `format` script flow. - [x] Format current code base for the first time and fix possible style guide violations using the configured linters of the project. [gh-eslint-config-prettier]: https://github.com/prettier/eslint-config-prettier [gh-eslint-plugin-prettier]: https://github.com/prettier/eslint-plugin-prettier [npm-eslint-plugin-prettier]: https://www.npmjs.com/package/eslint-plugin-prettier [npm-prettier]: https://www.npmjs.com/package/prettier [prettier-blog-1.15-mdx]: https://prettier.io/blog/2018/11/07/1.15.0.html#mdx [prettier-docs-pwidth]: https://prettier.io/docs/en/options.html#print-width [prettier]: https://prettier.io [stg-js-esl-c#ep]: https://github.com/arcticicestudio/styleguide-javascript/blob/develop/packages/%40arcticicestudio/eslint-config/README.md#entry-points [stg-js-esl-c-ts#ep]: https://github.com/arcticicestudio/styleguide-javascript/blob/develop/packages/%40arcticicestudio/eslint-config-typescript/README.md#entry-points
True
Prettier - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/63637792-4dcef380-c681-11e9-9252-f2fb22499985.png" width="30%" /></p> Integrate [Prettier][], the opinionated code formatter with support for many languages and integrations with most editors. It ensures that all outputted code conforms to a consistent style. ### Configuration This is one of the main features of Prettier: It already provides the best and recommended style configurations of-out-the-box™. The only option we will change is the [print width][prettier-docs-pwidth]. It is set to 80 by default which not up-to-date for modern screens (might only be relevant when working in terminals only like e.g. with Vim). It'll be changed to 120 used by all of Arctic Ice Studio's style guides. The `prettier.config.js` configuration file will be placed in the project root as well as the `.prettierignore` file to also define ignore pattern. ### ESLint Compatibility To be fully compatible with ESLint, [eslint-plugin-prettier][gh-eslint-plugin-prettier] has already been included in #30 as well as the [set of recommended rules][gh-eslint-config-prettier] via the [`@arcticicestudio/eslint-config/prettier`][stg-js-esl-c#ep] and [`@arcticicestudio/eslint-config-typescript/prettier`][stg-js-esl-c-ts#ep] extension entry points. ### Package Script To allow to format all sources a `format:pretty` package script will be added that'll also run in the main `format` script flow. A new `lint:pretty` script will also allow to check if all supported files are correct formatted. It will be included in the main `lint` script flow. ## Tasks - ~~Install [prettier][npm-prettier].~~ Resolved in #30 - [x] Implement `prettier.config.js` configuration file. - [x] Implement `.prettierignore` ignore pattern file. - [x] Implement `format:pretty` package script and add to main `format` script flow. - [x] Format current code base for the first time and fix possible style guide violations using the configured linters of the project. [gh-eslint-config-prettier]: https://github.com/prettier/eslint-config-prettier [gh-eslint-plugin-prettier]: https://github.com/prettier/eslint-plugin-prettier [npm-eslint-plugin-prettier]: https://www.npmjs.com/package/eslint-plugin-prettier [npm-prettier]: https://www.npmjs.com/package/prettier [prettier-blog-1.15-mdx]: https://prettier.io/blog/2018/11/07/1.15.0.html#mdx [prettier-docs-pwidth]: https://prettier.io/docs/en/options.html#print-width [prettier]: https://prettier.io [stg-js-esl-c#ep]: https://github.com/arcticicestudio/styleguide-javascript/blob/develop/packages/%40arcticicestudio/eslint-config/README.md#entry-points [stg-js-esl-c-ts#ep]: https://github.com/arcticicestudio/styleguide-javascript/blob/develop/packages/%40arcticicestudio/eslint-config-typescript/README.md#entry-points
main
prettier integrate the opinionated code formatter with support for many languages and integrations with most editors it ensures that all outputted code conforms to a consistent style configuration this is one of the main features of prettier it already provides the best and recommended style configurations of out the box™ the only option we will change is the it is set to by default which not up to date for modern screens might only be relevant when working in terminals only like e g with vim it ll be changed to used by all of arctic ice studio s style guides the prettier config js configuration file will be placed in the project root as well as the prettierignore file to also define ignore pattern eslint compatibility to be fully compatible with eslint has already been included in as well as the via the and extension entry points package script to allow to format all sources a format pretty package script will be added that ll also run in the main format script flow a new lint pretty script will also allow to check if all supported files are correct formatted it will be included in the main lint script flow tasks install resolved in implement prettier config js configuration file implement prettierignore ignore pattern file implement format pretty package script and add to main format script flow format current code base for the first time and fix possible style guide violations using the configured linters of the project
1
39
2,587,569,757
IssuesEvent
2015-02-17 19:18:13
spyder-ide/spyder
https://api.github.com/repos/spyder-ide/spyder
opened
Registry for all ready signals from widgets
1 star enhancement imported Maintainability Miscelleneous
_From [techtonik@gmail.com](https://code.google.com/u/techtonik@gmail.com/) on 2013-06-02T02:00:50Z_ Partially inspired by e75a44b270ba, the idea is to gather all ready signals into one container, which can provide real-time data about state of Spyder components and operations. I can already see this implemented as a LED bar (dashboard) for all ready signals. Some implementation details. 1. Helper to create signals and add them to registry. As far as I know QT4 doesn't allow to track when signals are created or query them. 2. When signal is added, it is given a "grey" placeholder. 3. When signal if fired, this "gray" changes state to "green" 4. A dashboard page accessible from some `?' submenus, maybe even from About screen _Original issue: http://code.google.com/p/spyderlib/issues/detail?id=1421_
True
Registry for all ready signals from widgets - _From [techtonik@gmail.com](https://code.google.com/u/techtonik@gmail.com/) on 2013-06-02T02:00:50Z_ Partially inspired by e75a44b270ba, the idea is to gather all ready signals into one container, which can provide real-time data about state of Spyder components and operations. I can already see this implemented as a LED bar (dashboard) for all ready signals. Some implementation details. 1. Helper to create signals and add them to registry. As far as I know QT4 doesn't allow to track when signals are created or query them. 2. When signal is added, it is given a "grey" placeholder. 3. When signal if fired, this "gray" changes state to "green" 4. A dashboard page accessible from some `?' submenus, maybe even from About screen _Original issue: http://code.google.com/p/spyderlib/issues/detail?id=1421_
main
registry for all ready signals from widgets from on partially inspired by the idea is to gather all ready signals into one container which can provide real time data about state of spyder components and operations i can already see this implemented as a led bar dashboard for all ready signals some implementation details helper to create signals and add them to registry as far as i know doesn t allow to track when signals are created or query them when signal is added it is given a grey placeholder when signal if fired this gray changes state to green a dashboard page accessible from some submenus maybe even from about screen original issue
1
2,751
9,828,363,468
IssuesEvent
2019-06-15 10:51:25
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFM - Scilab
Status: Available For Maintainer(s)
<!-- If you want to request a new maintainer for a package that you DO NOT MAINTAIN, please ensure you have followed the Package Triage Process - https://chocolatey.org/docs/package-triage-process - specifically you have contacted the maintainer using the 'Contact Maintainer' link on the package page. If you have followed the Package Triage Process above and want to request to become the maintainer of a package that you DO NOT MAINTAIN, please go to the package page and click the 'Contact Site Admins' link and complete the details. If you have followed the Package Triage Process above and do not want to request to become the maintainer of a package that you DO NOT MAINTAIN, please continue. Please ensure the issue title starts with 'RFM - ' - for example 'RFM - Adobe Reader' Please ensure you have the package URL from https://chocolatey.org/packages before continuing. NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM. ## Current Maintainer - [ ] I am the maintainer of the package and wish to pass it to someone else; --> ## I DON'T Want To Become The Maintainer - [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package; - [x] There is no existing open maintainer request for this package; ## Checklist - [x] Issue title starts with 'RFM - ' ## Existing Package Details Package URL: https://chocolatey.org/packages/SciLab Package source URL: https://www.scilab.org/download/6.0.2 <!-- please remove the following section if you are the current maintainer of the package, otherwise fill out the details --> Date the maintainer was contacted: two months ago How the maintainer was contacted: via [this ](http://disq.us/p/20r3btn) Disqus comment and [this ](https://github.com/dtgm/chocolatey-packages/issues/452) GitHub issue ## Other: According to [this comment](http://disq.us/p/229jlbn), this formula should be auto-updatable. However, it doesn't seem to be the case.
True
RFM - Scilab - <!-- If you want to request a new maintainer for a package that you DO NOT MAINTAIN, please ensure you have followed the Package Triage Process - https://chocolatey.org/docs/package-triage-process - specifically you have contacted the maintainer using the 'Contact Maintainer' link on the package page. If you have followed the Package Triage Process above and want to request to become the maintainer of a package that you DO NOT MAINTAIN, please go to the package page and click the 'Contact Site Admins' link and complete the details. If you have followed the Package Triage Process above and do not want to request to become the maintainer of a package that you DO NOT MAINTAIN, please continue. Please ensure the issue title starts with 'RFM - ' - for example 'RFM - Adobe Reader' Please ensure you have the package URL from https://chocolatey.org/packages before continuing. NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM. ## Current Maintainer - [ ] I am the maintainer of the package and wish to pass it to someone else; --> ## I DON'T Want To Become The Maintainer - [x] I have followed the Package Triage Process and I do NOT want to become maintainer of the package; - [x] There is no existing open maintainer request for this package; ## Checklist - [x] Issue title starts with 'RFM - ' ## Existing Package Details Package URL: https://chocolatey.org/packages/SciLab Package source URL: https://www.scilab.org/download/6.0.2 <!-- please remove the following section if you are the current maintainer of the package, otherwise fill out the details --> Date the maintainer was contacted: two months ago How the maintainer was contacted: via [this ](http://disq.us/p/20r3btn) Disqus comment and [this ](https://github.com/dtgm/chocolatey-packages/issues/452) GitHub issue ## Other: According to [this comment](http://disq.us/p/229jlbn), this formula should be auto-updatable. However, it doesn't seem to be the case.
main
rfm scilab if you want to request a new maintainer for a package that you do not maintain please ensure you have followed the package triage process specifically you have contacted the maintainer using the contact maintainer link on the package page if you have followed the package triage process above and want to request to become the maintainer of a package that you do not maintain please go to the package page and click the contact site admins link and complete the details if you have followed the package triage process above and do not want to request to become the maintainer of a package that you do not maintain please continue please ensure the issue title starts with rfm for example rfm adobe reader please ensure you have the package url from before continuing note keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the chocolatey community please remove all comments once you have read them current maintainer i am the maintainer of the package and wish to pass it to someone else i don t want to become the maintainer i have followed the package triage process and i do not want to become maintainer of the package there is no existing open maintainer request for this package checklist issue title starts with rfm existing package details package url package source url date the maintainer was contacted two months ago how the maintainer was contacted via disqus comment and github issue other according to this formula should be auto updatable however it doesn t seem to be the case
1
5,276
26,671,533,866
IssuesEvent
2023-01-26 10:42:30
beyarkay/eskom-calendar
https://api.github.com/repos/beyarkay/eskom-calendar
closed
Schedule Missing for Eersterivier, Eastern Cape
bug good first issue waiting-on-maintainer missing-area-schedule
hallo, pls will you add the schedule for Eersterivier in Eastern Cape? thanks!!!
True
Schedule Missing for Eersterivier, Eastern Cape - hallo, pls will you add the schedule for Eersterivier in Eastern Cape? thanks!!!
main
schedule missing for eersterivier eastern cape hallo pls will you add the schedule for eersterivier in eastern cape thanks
1
4,001
18,672,517,355
IssuesEvent
2021-10-31 00:43:08
exercism/python
https://api.github.com/repos/exercism/python
closed
[Tooling] Add GH Action to Greet New Contributors and Explain Submission Guidlines.
claimed 🐾 in-progress 🌿 maintainer chore 🔧
Similar to this one from Elixir (_and probably based off it...._) https://github.com/exercism/elixir/blob/main/.github/pr-commenter.yml
True
[Tooling] Add GH Action to Greet New Contributors and Explain Submission Guidlines. - Similar to this one from Elixir (_and probably based off it...._) https://github.com/exercism/elixir/blob/main/.github/pr-commenter.yml
main
add gh action to greet new contributors and explain submission guidlines similar to this one from elixir and probably based off it
1
4,384
22,310,480,704
IssuesEvent
2022-06-13 16:31:00
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
Investigate pins for various CI components (pytest-cov, coverage, msmb_theme)
maintainability Component-Docs Continuous Integration
Follow up from #3369 Related to #3224 Our docs build infrastructure is starting to be a bit flaky, particularly in #3369 we've had to pin pytest-cov due to some dependency resolution issues we were having when installing it at the same time as the sphinx components. We should aim to check a) if all the pins are still necessary, b) what steps we should take to remove these pins. My understanding is that some of this is related to the fact that we can't upgrade to the latest sphinx (see #3224), so we probably will want to fix that first.
True
Investigate pins for various CI components (pytest-cov, coverage, msmb_theme) - Follow up from #3369 Related to #3224 Our docs build infrastructure is starting to be a bit flaky, particularly in #3369 we've had to pin pytest-cov due to some dependency resolution issues we were having when installing it at the same time as the sphinx components. We should aim to check a) if all the pins are still necessary, b) what steps we should take to remove these pins. My understanding is that some of this is related to the fact that we can't upgrade to the latest sphinx (see #3224), so we probably will want to fix that first.
main
investigate pins for various ci components pytest cov coverage msmb theme follow up from related to our docs build infrastructure is starting to be a bit flaky particularly in we ve had to pin pytest cov due to some dependency resolution issues we were having when installing it at the same time as the sphinx components we should aim to check a if all the pins are still necessary b what steps we should take to remove these pins my understanding is that some of this is related to the fact that we can t upgrade to the latest sphinx see so we probably will want to fix that first
1
128,749
12,390,790,003
IssuesEvent
2020-05-20 11:19:41
opendatacam/opendatacam
https://api.github.com/repos/opendatacam/opendatacam
opened
JSON Specification for using custom Neural Network
documentation question
As we spoke about maybe splitting the Tracking/Counting part from the neural network, I had the Idea of writing down the specification the neural network part would have to have to be able to communicate with the ODC component. From my understanding the JSON stream would need something to sync itself with the video stream (e.g. framecount) the dimensions and the tracking information. Is there already something like this?
1.0
JSON Specification for using custom Neural Network - As we spoke about maybe splitting the Tracking/Counting part from the neural network, I had the Idea of writing down the specification the neural network part would have to have to be able to communicate with the ODC component. From my understanding the JSON stream would need something to sync itself with the video stream (e.g. framecount) the dimensions and the tracking information. Is there already something like this?
non_main
json specification for using custom neural network as we spoke about maybe splitting the tracking counting part from the neural network i had the idea of writing down the specification the neural network part would have to have to be able to communicate with the odc component from my understanding the json stream would need something to sync itself with the video stream e g framecount the dimensions and the tracking information is there already something like this
0
5,129
26,142,738,609
IssuesEvent
2022-12-29 21:17:05
darekkay/what-the-commit
https://api.github.com/repos/darekkay/what-the-commit
closed
Unreleased version on plugins.jetbrains.com
Type: Maintainance
Hey there, huge fan of the extension 🙏 Could you do a release on [plugins.jetbrains.com](https://plugins.jetbrains.com/plugin/7536-what-the-commit) as the current extension is not working anymore without #3 This tool is the best way to get the most of the Git experience, and I would be saddened if peoples couldn't easily use it 😢
True
Unreleased version on plugins.jetbrains.com - Hey there, huge fan of the extension 🙏 Could you do a release on [plugins.jetbrains.com](https://plugins.jetbrains.com/plugin/7536-what-the-commit) as the current extension is not working anymore without #3 This tool is the best way to get the most of the Git experience, and I would be saddened if peoples couldn't easily use it 😢
main
unreleased version on plugins jetbrains com hey there huge fan of the extension 🙏 could you do a release on as the current extension is not working anymore without this tool is the best way to get the most of the git experience and i would be saddened if peoples couldn t easily use it 😢
1
373
3,368,098,195
IssuesEvent
2015-11-22 18:35:51
jenkinsci/slack-plugin
https://api.github.com/repos/jenkinsci/slack-plugin
opened
Don't store global config per job
contributions welcome enhancement maintainer communication
I feel a major source of issues this plugin faces is how it stores global config in every job. What it should do instead is reference the global config in the Jenkins runtime. Job config should only contain settings related to the job. This has caused issues in the past like updating global config not properly propagating to all jobs. As we move to slack-2.0 and beyond I could see this causing problems even more. As issues are opened related to this I'll link them to this issue.
True
Don't store global config per job - I feel a major source of issues this plugin faces is how it stores global config in every job. What it should do instead is reference the global config in the Jenkins runtime. Job config should only contain settings related to the job. This has caused issues in the past like updating global config not properly propagating to all jobs. As we move to slack-2.0 and beyond I could see this causing problems even more. As issues are opened related to this I'll link them to this issue.
main
don t store global config per job i feel a major source of issues this plugin faces is how it stores global config in every job what it should do instead is reference the global config in the jenkins runtime job config should only contain settings related to the job this has caused issues in the past like updating global config not properly propagating to all jobs as we move to slack and beyond i could see this causing problems even more as issues are opened related to this i ll link them to this issue
1
48,309
20,103,551,110
IssuesEvent
2022-02-07 08:12:34
ballerina-platform/openapi-tools
https://api.github.com/repos/ballerina-platform/openapi-tools
closed
Path parameter is not generated for path name `key`
Type/Bug Points/1 Service OpenAPIToBallerina
**Description:** When a path parameter is given in the name of `key`, that path parameter is not generated in the service resource function. **Steps to reproduce:** Generate Ballerina service for following openapi ```yaml openapi: 3.0.1 info: title: Storage Space version: 1.0.0 paths: /storage-spaces/{storageSpaceName}/keys/{key}: get: summary: Get key value operationId: getKey parameters: - name: storageSpaceName in: path description: name of the storage space required: true schema: type: string - name: key in: path description: name of the key to retrieve the value required: true schema: type: string responses: 200: description: successful operation content: text/plain: schema: type: string 400: description: Invalid key supplied security: - authorization: [] components: securitySchemes: authorization: type: apiKey description: Represents API Key `authorization` name: Authorization in: header ``` Generated code ```ballerina import ballerina/http; listener http:Listener ep0 = new (9090, config = {host: "localhost"}); service / on ep0 { resource function get 'storage\-spaces/[string storageSpaceName]/keys() returns string|http:BadRequest { } } ``` **Affected Versions:** SL Beta6
1.0
Path parameter is not generated for path name `key` - **Description:** When a path parameter is given in the name of `key`, that path parameter is not generated in the service resource function. **Steps to reproduce:** Generate Ballerina service for following openapi ```yaml openapi: 3.0.1 info: title: Storage Space version: 1.0.0 paths: /storage-spaces/{storageSpaceName}/keys/{key}: get: summary: Get key value operationId: getKey parameters: - name: storageSpaceName in: path description: name of the storage space required: true schema: type: string - name: key in: path description: name of the key to retrieve the value required: true schema: type: string responses: 200: description: successful operation content: text/plain: schema: type: string 400: description: Invalid key supplied security: - authorization: [] components: securitySchemes: authorization: type: apiKey description: Represents API Key `authorization` name: Authorization in: header ``` Generated code ```ballerina import ballerina/http; listener http:Listener ep0 = new (9090, config = {host: "localhost"}); service / on ep0 { resource function get 'storage\-spaces/[string storageSpaceName]/keys() returns string|http:BadRequest { } } ``` **Affected Versions:** SL Beta6
non_main
path parameter is not generated for path name key description when a path parameter is given in the name of key that path parameter is not generated in the service resource function steps to reproduce generate ballerina service for following openapi yaml openapi info title storage space version paths storage spaces storagespacename keys key get summary get key value operationid getkey parameters name storagespacename in path description name of the storage space required true schema type string name key in path description name of the key to retrieve the value required true schema type string responses description successful operation content text plain schema type string description invalid key supplied security authorization components securityschemes authorization type apikey description represents api key authorization name authorization in header generated code ballerina import ballerina http listener http listener new config host localhost service on resource function get storage spaces keys returns string http badrequest affected versions sl
0
375,135
26,148,812,399
IssuesEvent
2022-12-30 10:14:49
Avaiga/taipy-core
https://api.github.com/repos/Avaiga/taipy-core
closed
Missing tooltips in IDE when using Config class for configuring data nodes
Core: ⚙️ Configuration ❌ Blocked 📄 Documentation 🟧 Priority: High
**Description** In Taipy 2.0, I noticed that methods such as `configure_data_node` are assigned to the `Config` class in `taipy.core.config.__init__.py`, rather than originally being in the class definition. A side effect of this change appears to be that tooltips // type hints are no longer showing up for me when I use `Config.configure_data_node` in my IDE. I'm using VS Code with Pylance.
1.0
Missing tooltips in IDE when using Config class for configuring data nodes - **Description** In Taipy 2.0, I noticed that methods such as `configure_data_node` are assigned to the `Config` class in `taipy.core.config.__init__.py`, rather than originally being in the class definition. A side effect of this change appears to be that tooltips // type hints are no longer showing up for me when I use `Config.configure_data_node` in my IDE. I'm using VS Code with Pylance.
non_main
missing tooltips in ide when using config class for configuring data nodes description in taipy i noticed that methods such as configure data node are assigned to the config class in taipy core config init py rather than originally being in the class definition a side effect of this change appears to be that tooltips type hints are no longer showing up for me when i use config configure data node in my ide i m using vs code with pylance
0
3,091
11,741,740,297
IssuesEvent
2020-03-11 22:32:02
alacritty/alacritty
https://api.github.com/repos/alacritty/alacritty
closed
Crashes on start on Windows
A - deps B - crash C - waiting on maintainer H - windows S - winit/glutin
After installing the latest version, Alacritty crashes on start with an error pop-up. I've tried running it with and without a config file but no luck. Any ideas? ### System OS: Windows 10 Version: 0.4.1 (installed via msi installer) Windows: ConPTY ### Error output #### Alacritty: Runtime Error panicked at 'assertion failed: `(left == right)` left: `1`, right: `0`', C:\Users\travis\.cargo\git\checkouts\winit-c2fdb27092aba5a7\d1c6506\src\platform_impl\windows\dark_mode.rs:135:13
True
Crashes on start on Windows - After installing the latest version, Alacritty crashes on start with an error pop-up. I've tried running it with and without a config file but no luck. Any ideas? ### System OS: Windows 10 Version: 0.4.1 (installed via msi installer) Windows: ConPTY ### Error output #### Alacritty: Runtime Error panicked at 'assertion failed: `(left == right)` left: `1`, right: `0`', C:\Users\travis\.cargo\git\checkouts\winit-c2fdb27092aba5a7\d1c6506\src\platform_impl\windows\dark_mode.rs:135:13
main
crashes on start on windows after installing the latest version alacritty crashes on start with an error pop up i ve tried running it with and without a config file but no luck any ideas system os windows version installed via msi installer windows conpty error output alacritty runtime error panicked at assertion failed left right left right c users travis cargo git checkouts winit src platform impl windows dark mode rs
1
20,373
10,513,236,454
IssuesEvent
2019-09-27 20:01:35
nextcheckbox/nextcheckbox
https://api.github.com/repos/nextcheckbox/nextcheckbox
closed
Sanitize service content
security
Currently we're using `v-html` in the `service/:id` route...let's not do that
True
Sanitize service content - Currently we're using `v-html` in the `service/:id` route...let's not do that
non_main
sanitize service content currently we re using v html in the service id route let s not do that
0
19
2,515,398,693
IssuesEvent
2015-01-15 18:22:10
simplesamlphp/simplesamlphp
https://api.github.com/repos/simplesamlphp/simplesamlphp
closed
Remove www/example-simple/
enhancement maintainability started
The `www/example-simple/hostnames.php` file should be moved to the `www/admin/` folder. The rest of files must go away.
True
Remove www/example-simple/ - The `www/example-simple/hostnames.php` file should be moved to the `www/admin/` folder. The rest of files must go away.
main
remove www example simple the www example simple hostnames php file should be moved to the www admin folder the rest of files must go away
1
784,478
27,572,648,923
IssuesEvent
2023-03-08 10:28:59
inverse-inc/packetfence
https://api.github.com/repos/inverse-inc/packetfence
closed
v12.2: RADIUS source+pfconnector is not working in admin context
Type: Bug Priority: High
**Describe the bug** If you use a RADIUS source with authorization rules (admin context) and your source is configured with "Use connector" **enabled**, RADIUS response will be ignored. If you disable "Use Connector", everything works as expected. I tried to use same source (with "Use connector" enabled) on captive portal and it works. **To Reproduce** Steps to reproduce the behavior: 1. Configure a RADIUS server 2. Create your RADIUS source with "Use Connector" enabled 3. Create an administration rule on that source 4. Try to authenticate on webadmin using an account on your RADIUS server => Following message will appear in logs: ``` Feb 23 11:20:14 pfdeb11localdev pfperl-api-docker-wrapper[2050]: pfperl-api(16) ERROR: [mac:[undef]] Unable to perform RADIUS authentication on any server: ETIMEOUT (pf::Authentication::Source::RADIUSSource::_handle_radius_request) ``` **Expected behavior** Access to web admin granted **Additional context** I captured traffic on port use by pfconnector and I can see that RADIUS server is returning an Access-Accept.
1.0
v12.2: RADIUS source+pfconnector is not working in admin context - **Describe the bug** If you use a RADIUS source with authorization rules (admin context) and your source is configured with "Use connector" **enabled**, RADIUS response will be ignored. If you disable "Use Connector", everything works as expected. I tried to use same source (with "Use connector" enabled) on captive portal and it works. **To Reproduce** Steps to reproduce the behavior: 1. Configure a RADIUS server 2. Create your RADIUS source with "Use Connector" enabled 3. Create an administration rule on that source 4. Try to authenticate on webadmin using an account on your RADIUS server => Following message will appear in logs: ``` Feb 23 11:20:14 pfdeb11localdev pfperl-api-docker-wrapper[2050]: pfperl-api(16) ERROR: [mac:[undef]] Unable to perform RADIUS authentication on any server: ETIMEOUT (pf::Authentication::Source::RADIUSSource::_handle_radius_request) ``` **Expected behavior** Access to web admin granted **Additional context** I captured traffic on port use by pfconnector and I can see that RADIUS server is returning an Access-Accept.
non_main
radius source pfconnector is not working in admin context describe the bug if you use a radius source with authorization rules admin context and your source is configured with use connector enabled radius response will be ignored if you disable use connector everything works as expected i tried to use same source with use connector enabled on captive portal and it works to reproduce steps to reproduce the behavior configure a radius server create your radius source with use connector enabled create an administration rule on that source try to authenticate on webadmin using an account on your radius server following message will appear in logs feb pfperl api docker wrapper pfperl api error unable to perform radius authentication on any server etimeout pf authentication source radiussource handle radius request expected behavior access to web admin granted additional context i captured traffic on port use by pfconnector and i can see that radius server is returning an access accept
0
516,299
14,978,934,285
IssuesEvent
2021-01-28 11:31:26
HabitRPG/habitica-android
https://api.github.com/repos/HabitRPG/habitica-android
closed
Tasks are difficult to edit if they have a link in them
Accessibility Help wanted Priority: medium Type: Bug
User report: "In the screenshot with the red spot on the "End Break it Down! Challenge" To Do you see the spot (red) where I click when I want to edit the task, but instead of opening the editor, it opens the link to the Break it Down challenge. This is not only true for this To Do but for To Do's (and other task types?) in general. It's really hard to edit them if there is a link in the title." Device: Sony F5321 Android Version: 25 AppVersion: Version 3.1.1 (2691) BETA User ID: 33bb14bd-814d-40cb-98a4-7b76a752761c Level: 84 Class: rogue Is in Inn: false Uses Costume: true Custom Day Start: 16 Timezone Offset: -60 ![_20201212_165158](https://user-images.githubusercontent.com/8144640/105178466-f952d080-5b1f-11eb-9722-dcbde53190dc.JPG)
1.0
Tasks are difficult to edit if they have a link in them - User report: "In the screenshot with the red spot on the "End Break it Down! Challenge" To Do you see the spot (red) where I click when I want to edit the task, but instead of opening the editor, it opens the link to the Break it Down challenge. This is not only true for this To Do but for To Do's (and other task types?) in general. It's really hard to edit them if there is a link in the title." Device: Sony F5321 Android Version: 25 AppVersion: Version 3.1.1 (2691) BETA User ID: 33bb14bd-814d-40cb-98a4-7b76a752761c Level: 84 Class: rogue Is in Inn: false Uses Costume: true Custom Day Start: 16 Timezone Offset: -60 ![_20201212_165158](https://user-images.githubusercontent.com/8144640/105178466-f952d080-5b1f-11eb-9722-dcbde53190dc.JPG)
non_main
tasks are difficult to edit if they have a link in them user report in the screenshot with the red spot on the end break it down challenge to do you see the spot red where i click when i want to edit the task but instead of opening the editor it opens the link to the break it down challenge this is not only true for this to do but for to do s and other task types in general it s really hard to edit them if there is a link in the title device sony android version appversion version beta user id level class rogue is in inn false uses costume true custom day start timezone offset
0
223,015
7,445,636,328
IssuesEvent
2018-03-28 05:54:11
cilium/cilium
https://api.github.com/repos/cilium/cilium
closed
CIDR egress policy doesn't restrict access to world
area/datapath kind/bug kind/community-report priority/1.0-blocker
When I set up the following policy on egress for a container via k8s networkpolicy, it could contact IPs outside of the specified CIDRs (via ICMP and HTTP). ### cilium policy list ``` ENDPOINT POLICY (ingress) POLICY (egress) IDENTITY LABELS (source:key[=value]) IPv6 IPv4 STATUS ENFORCEMENT ENFORCEMENT 3365 Enabled Disabled 23405 k8s:id=app1 f00d::a0f:0:0:d25 10.15.191.0 ready k8s:io.kubernetes.pod.namespace=default 11004 Disabled Disabled 4636 k8s:id=app3 f00d::a0f:0:0:2afc 10.15.54.125 ready k8s:io.kubernetes.pod.namespace=default 19080 Disabled Disabled 39700 k8s:io.kubernetes.pod.namespace=kube-system f00d::a0f:0:0:4a88 10.15.236.148 ready k8s:run=microscope 29898 Disabled Disabled 41522 reserved:health f00d::a0f:0:0:74ca 10.15.242.54 ready 42910 Enabled Disabled 23405 k8s:id=app1 f00d::a0f:0:0:a79e 10.15.236.254 ready k8s:io.kubernetes.pod.namespace=default 49508 Disabled Enabled 9087 k8s:id=app2 f00d::a0f:0:0:c164 10.15.114.197 ready k8s:io.kubernetes.pod.namespace=default ``` ### cilium policy get (I removed the policy for app1 in the below output) ``` # cilium policy get [ { "endpointSelector": { "matchLabels": { "k8s:id": "app2", "k8s:io.kubernetes.pod.namespace": "default" } }, "egress": [ { "toCIDRSet": [ { "cidr": "172.16.0.0/14", "except": [ "172.18.1.0/24" ] } ] } ], "labels": [ { "key": "io.cilium.k8s.policy.name", "value": "test-network-policy-egress", "source": "unspec" }, { "key": "io.cilium.k8s.policy.namespace", "value": "default", "source": "unspec" } ] } ] Revision: 54 ``` ### cilium monitor ``` ------------------------------------------------------------------------------ CPU 01: MARK 0x34196f76 FROM 49508 from-endpoint: 74 bytes, state new, identity 9087->0 Ethernet {Contents=[..14..] Payload=[..62..] SrcMAC=c6:b2:48:c2:75:59 DstMAC=b2:f1:e4:65:db:13 EthernetType=IPv4 Length=0} IPv4 {Contents=[..20..] Payload=[..40..] Version=4 IHL=5 TOS=0 Length=60 Id=45882 Flags=DF FragOffset=0 TTL=64 Protocol=TCP Checksum=23270 SrcIP=10.15.114.197 DstIP=172.217.2.238 Options=[] Padding=[]} TCP {Contents=[..40..] Payload=[] SrcPort=40330 DstPort=80(http) Seq=907956180 Ack=0 DataOffset=10 FIN=false SYN=true RST=false PSH=false ACK=false URG=false ECE=false CWR=false NS=false Window=28200 Checksum=11466 Urgent=0 Options=[..5..] Padding=[]} CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Conntrack lookup 1/2: src=10.15.114.197:40330 dst=172.217.2.238:80 CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Conntrack lookup 2/2: nexthdr=6 flags=1 CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: CT verdict: New, proxy_port=0 revnat=0 CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Failed to map daddr=3137322e3231372e322e323338 to identity CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Resolved L4 policy to: 0 / egress CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Conntrack create: proxy-port=0 revnat=0 src-identity=9087 lb=0.0.0.0 CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Going to the stack, policy-skip=0 ------------------------------------------------------------------------------ CPU 01: MARK 0x34196f76 FROM 49508 to-stack: 74 bytes, state new, identity 9087->2 Ethernet {Contents=[..14..] Payload=[..62..] SrcMAC=c6:b2:48:c2:75:59 DstMAC=b2:f1:e4:65:db:13 EthernetType=IPv4 Length=0} IPv4 {Contents=[..20..] Payload=[..40..] Version=4 IHL=5 TOS=0 Length=60 Id=45882 Flags=DF FragOffset=0 TTL=63 Protocol=TCP Checksum=23526 SrcIP=10.15.114.197 DstIP=172.217.2.238 Options=[] Padding=[]} TCP {Contents=[..40..] Payload=[] SrcPort=40330 DstPort=80(http) Seq=907956180 Ack=0 DataOffset=10 FIN=false SYN=true RST=false PSH=false ACK=false URG=false ECE=false CWR=false NS=false Window=28200 Checksum=11466 Urgent=0 Options=[..5..] Padding=[]} CPU 01: MARK 0x79cf9f06 FROM 49508 DEBUG: Conntrack lookup 1/2: src=172.217.2.238:80 dst=10.15.114.197:40330 CPU 01: MARK 0x79cf9f06 FROM 49508 DEBUG: Conntrack lookup 2/2: nexthdr=6 flags=0 CPU 01: MARK 0x79cf9f06 FROM 49508 DEBUG: CT entry found lifetime=4662, proxy_port=0 revnat=0 CPU 01: MARK 0x79cf9f06 FROM 49508 DEBUG: CT verdict: Reply, proxy_port=0 revnat=0 ------------------------------------------------------------------------------ CPU 01: MARK 0x79cf9f06 FROM 49508 to-endpoint: 58 bytes, state reply, interface lxcb8a77, identity 1->9087, to endpoint 49508 Ethernet {Contents=[..14..] Payload=[..46..] SrcMAC=b2:f1:e4:65:db:13 DstMAC=c6:b2:48:c2:75:59 EthernetType=IPv4 Length=0} IPv4 {Contents=[..20..] Payload=[..24..] Version=4 IHL=5 TOS=0 Length=44 Id=17288 Flags= FragOffset=0 TTL=62 Protocol=TCP Checksum=3241 SrcIP=172.217.2.238 DstIP=10.15.114.197 Options=[] Padding=[]} TCP {Contents=[..24..] Payload=[] SrcPort=80(http) DstPort=40330 Seq=560832001 Ack=907956181 DataOffset=6 FIN=false SYN=true RST=false PSH=false ACK=true URG=false ECE=false CWR=false NS=false Window=65535 Checksum=34878 Urgent=0 Options=[TCPOption(MS S:1460 0x05b4)] Padding=[]} ```
1.0
CIDR egress policy doesn't restrict access to world - When I set up the following policy on egress for a container via k8s networkpolicy, it could contact IPs outside of the specified CIDRs (via ICMP and HTTP). ### cilium policy list ``` ENDPOINT POLICY (ingress) POLICY (egress) IDENTITY LABELS (source:key[=value]) IPv6 IPv4 STATUS ENFORCEMENT ENFORCEMENT 3365 Enabled Disabled 23405 k8s:id=app1 f00d::a0f:0:0:d25 10.15.191.0 ready k8s:io.kubernetes.pod.namespace=default 11004 Disabled Disabled 4636 k8s:id=app3 f00d::a0f:0:0:2afc 10.15.54.125 ready k8s:io.kubernetes.pod.namespace=default 19080 Disabled Disabled 39700 k8s:io.kubernetes.pod.namespace=kube-system f00d::a0f:0:0:4a88 10.15.236.148 ready k8s:run=microscope 29898 Disabled Disabled 41522 reserved:health f00d::a0f:0:0:74ca 10.15.242.54 ready 42910 Enabled Disabled 23405 k8s:id=app1 f00d::a0f:0:0:a79e 10.15.236.254 ready k8s:io.kubernetes.pod.namespace=default 49508 Disabled Enabled 9087 k8s:id=app2 f00d::a0f:0:0:c164 10.15.114.197 ready k8s:io.kubernetes.pod.namespace=default ``` ### cilium policy get (I removed the policy for app1 in the below output) ``` # cilium policy get [ { "endpointSelector": { "matchLabels": { "k8s:id": "app2", "k8s:io.kubernetes.pod.namespace": "default" } }, "egress": [ { "toCIDRSet": [ { "cidr": "172.16.0.0/14", "except": [ "172.18.1.0/24" ] } ] } ], "labels": [ { "key": "io.cilium.k8s.policy.name", "value": "test-network-policy-egress", "source": "unspec" }, { "key": "io.cilium.k8s.policy.namespace", "value": "default", "source": "unspec" } ] } ] Revision: 54 ``` ### cilium monitor ``` ------------------------------------------------------------------------------ CPU 01: MARK 0x34196f76 FROM 49508 from-endpoint: 74 bytes, state new, identity 9087->0 Ethernet {Contents=[..14..] Payload=[..62..] SrcMAC=c6:b2:48:c2:75:59 DstMAC=b2:f1:e4:65:db:13 EthernetType=IPv4 Length=0} IPv4 {Contents=[..20..] Payload=[..40..] Version=4 IHL=5 TOS=0 Length=60 Id=45882 Flags=DF FragOffset=0 TTL=64 Protocol=TCP Checksum=23270 SrcIP=10.15.114.197 DstIP=172.217.2.238 Options=[] Padding=[]} TCP {Contents=[..40..] Payload=[] SrcPort=40330 DstPort=80(http) Seq=907956180 Ack=0 DataOffset=10 FIN=false SYN=true RST=false PSH=false ACK=false URG=false ECE=false CWR=false NS=false Window=28200 Checksum=11466 Urgent=0 Options=[..5..] Padding=[]} CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Conntrack lookup 1/2: src=10.15.114.197:40330 dst=172.217.2.238:80 CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Conntrack lookup 2/2: nexthdr=6 flags=1 CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: CT verdict: New, proxy_port=0 revnat=0 CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Failed to map daddr=3137322e3231372e322e323338 to identity CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Resolved L4 policy to: 0 / egress CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Conntrack create: proxy-port=0 revnat=0 src-identity=9087 lb=0.0.0.0 CPU 01: MARK 0x34196f76 FROM 49508 DEBUG: Going to the stack, policy-skip=0 ------------------------------------------------------------------------------ CPU 01: MARK 0x34196f76 FROM 49508 to-stack: 74 bytes, state new, identity 9087->2 Ethernet {Contents=[..14..] Payload=[..62..] SrcMAC=c6:b2:48:c2:75:59 DstMAC=b2:f1:e4:65:db:13 EthernetType=IPv4 Length=0} IPv4 {Contents=[..20..] Payload=[..40..] Version=4 IHL=5 TOS=0 Length=60 Id=45882 Flags=DF FragOffset=0 TTL=63 Protocol=TCP Checksum=23526 SrcIP=10.15.114.197 DstIP=172.217.2.238 Options=[] Padding=[]} TCP {Contents=[..40..] Payload=[] SrcPort=40330 DstPort=80(http) Seq=907956180 Ack=0 DataOffset=10 FIN=false SYN=true RST=false PSH=false ACK=false URG=false ECE=false CWR=false NS=false Window=28200 Checksum=11466 Urgent=0 Options=[..5..] Padding=[]} CPU 01: MARK 0x79cf9f06 FROM 49508 DEBUG: Conntrack lookup 1/2: src=172.217.2.238:80 dst=10.15.114.197:40330 CPU 01: MARK 0x79cf9f06 FROM 49508 DEBUG: Conntrack lookup 2/2: nexthdr=6 flags=0 CPU 01: MARK 0x79cf9f06 FROM 49508 DEBUG: CT entry found lifetime=4662, proxy_port=0 revnat=0 CPU 01: MARK 0x79cf9f06 FROM 49508 DEBUG: CT verdict: Reply, proxy_port=0 revnat=0 ------------------------------------------------------------------------------ CPU 01: MARK 0x79cf9f06 FROM 49508 to-endpoint: 58 bytes, state reply, interface lxcb8a77, identity 1->9087, to endpoint 49508 Ethernet {Contents=[..14..] Payload=[..46..] SrcMAC=b2:f1:e4:65:db:13 DstMAC=c6:b2:48:c2:75:59 EthernetType=IPv4 Length=0} IPv4 {Contents=[..20..] Payload=[..24..] Version=4 IHL=5 TOS=0 Length=44 Id=17288 Flags= FragOffset=0 TTL=62 Protocol=TCP Checksum=3241 SrcIP=172.217.2.238 DstIP=10.15.114.197 Options=[] Padding=[]} TCP {Contents=[..24..] Payload=[] SrcPort=80(http) DstPort=40330 Seq=560832001 Ack=907956181 DataOffset=6 FIN=false SYN=true RST=false PSH=false ACK=true URG=false ECE=false CWR=false NS=false Window=65535 Checksum=34878 Urgent=0 Options=[TCPOption(MS S:1460 0x05b4)] Padding=[]} ```
non_main
cidr egress policy doesn t restrict access to world when i set up the following policy on egress for a container via networkpolicy it could contact ips outside of the specified cidrs via icmp and http cilium policy list endpoint policy ingress policy egress identity labels source key status enforcement enforcement enabled disabled id ready io kubernetes pod namespace default disabled disabled id ready io kubernetes pod namespace default disabled disabled io kubernetes pod namespace kube system ready run microscope disabled disabled reserved health ready enabled disabled id ready io kubernetes pod namespace default disabled enabled id ready io kubernetes pod namespace default cilium policy get i removed the policy for in the below output cilium policy get endpointselector matchlabels id io kubernetes pod namespace default egress tocidrset cidr except labels key io cilium policy name value test network policy egress source unspec key io cilium policy namespace value default source unspec revision cilium monitor cpu mark from from endpoint bytes state new identity ethernet contents payload srcmac dstmac db ethernettype length contents payload version ihl tos length id flags df fragoffset ttl protocol tcp checksum srcip dstip options padding tcp contents payload srcport dstport http seq ack dataoffset fin false syn true rst false psh false ack false urg false ece false cwr false ns false window checksum urgent options padding cpu mark from debug conntrack lookup src dst cpu mark from debug conntrack lookup nexthdr flags cpu mark from debug ct verdict new proxy port revnat cpu mark from debug failed to map daddr to identity cpu mark from debug resolved policy to egress cpu mark from debug conntrack create proxy port revnat src identity lb cpu mark from debug going to the stack policy skip cpu mark from to stack bytes state new identity ethernet contents payload srcmac dstmac db ethernettype length contents payload version ihl tos length id flags df fragoffset ttl protocol tcp checksum srcip dstip options padding tcp contents payload srcport dstport http seq ack dataoffset fin false syn true rst false psh false ack false urg false ece false cwr false ns false window checksum urgent options padding cpu mark from debug conntrack lookup src dst cpu mark from debug conntrack lookup nexthdr flags cpu mark from debug ct entry found lifetime proxy port revnat cpu mark from debug ct verdict reply proxy port revnat cpu mark from to endpoint bytes state reply interface identity to endpoint ethernet contents payload srcmac db dstmac ethernettype length contents payload version ihl tos length id flags fragoffset ttl protocol tcp checksum srcip dstip options padding tcp contents payload srcport http dstport seq ack dataoffset fin false syn true rst false psh false ack true urg false ece false cwr false ns false window checksum urgent options tcpoption ms s padding
0
2,016
6,749,905,498
IssuesEvent
2017-10-23 00:03:37
dgets/DANT2a
https://api.github.com/repos/dgets/DANT2a
closed
Spaghetti detected in HeadsUp/EntryType display & update code
duplicate maintainability
I think that there are duplicate method names between `HeadsUp` & `EntryType`'s definitions, specifically the _update*()_ ones. This is making it rather confusing to trace the code execution through these bits as invoked by the __Tick_ code; this should probably be traced out on a printout, paper, or a better screen than just an rdp window, in order to see how to untangle things best here.
True
Spaghetti detected in HeadsUp/EntryType display & update code - I think that there are duplicate method names between `HeadsUp` & `EntryType`'s definitions, specifically the _update*()_ ones. This is making it rather confusing to trace the code execution through these bits as invoked by the __Tick_ code; this should probably be traced out on a printout, paper, or a better screen than just an rdp window, in order to see how to untangle things best here.
main
spaghetti detected in headsup entrytype display update code i think that there are duplicate method names between headsup entrytype s definitions specifically the update ones this is making it rather confusing to trace the code execution through these bits as invoked by the tick code this should probably be traced out on a printout paper or a better screen than just an rdp window in order to see how to untangle things best here
1
3,224
12,368,705,969
IssuesEvent
2020-05-18 14:13:27
Kashdeya/Tiny-Progressions
https://api.github.com/repos/Kashdeya/Tiny-Progressions
closed
[Request]Higher tier cobble gen
Version not Maintainted
I would like to request a higher tier cobble gen that can generate a stack of cobble a tick. Possibly a netherstar tier or a block with no recipe that can be enabled by pack makers. I'm using the cobble stone generators to create octuple compressed cobble and for an efficient setup I have over 200 top tier cobble gens running. I would love a way to compress this to reduce lag and space. Thanks in advance!
True
[Request]Higher tier cobble gen - I would like to request a higher tier cobble gen that can generate a stack of cobble a tick. Possibly a netherstar tier or a block with no recipe that can be enabled by pack makers. I'm using the cobble stone generators to create octuple compressed cobble and for an efficient setup I have over 200 top tier cobble gens running. I would love a way to compress this to reduce lag and space. Thanks in advance!
main
higher tier cobble gen i would like to request a higher tier cobble gen that can generate a stack of cobble a tick possibly a netherstar tier or a block with no recipe that can be enabled by pack makers i m using the cobble stone generators to create octuple compressed cobble and for an efficient setup i have over top tier cobble gens running i would love a way to compress this to reduce lag and space thanks in advance
1
509,116
14,713,238,278
IssuesEvent
2021-01-05 10:04:43
NodeFactoryIo/ChainGuardian
https://api.github.com/repos/NodeFactoryIo/ChainGuardian
closed
missing docker feedback
priority: P3 important
if user do not have docker installed need to have feedback while trying to create local beacon node
1.0
missing docker feedback - if user do not have docker installed need to have feedback while trying to create local beacon node
non_main
missing docker feedback if user do not have docker installed need to have feedback while trying to create local beacon node
0
4,169
19,985,352,443
IssuesEvent
2022-01-30 15:23:59
BioArchLinux/Packages
https://api.github.com/repos/BioArchLinux/Packages
opened
[MAINTAIN] r-mfa
maintain
<!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** <details> ``` ==> Starting build()... * installing *source* package ‘mfa’ ... ** using staged installation ** libs g++ -std=gnu++14 -I"/usr/include/R/" -DNDEBUG -I'/usr/lib/R/library/Rcpp/include' -D_FORTIFY_SOURCE=2 -fpic -march=x86-64 -mtune=generic -O2 -pipe -fno-plt -c RcppExports.cpp -o RcppExports.o g++ -std=gnu++14 -I"/usr/include/R/" -DNDEBUG -I'/usr/lib/R/library/Rcpp/include' -D_FORTIFY_SOURCE=2 -fpic -march=x86-64 -mtune=generic -O2 -pipe -fno-plt -c gibbs.cpp -o gibbs.o gibbs.cpp: In function ‘double log_d_norm(double, double, double)’: gibbs.cpp:13:31: error: ‘PI’ was not declared in this scope 13 | double ldn = -0.5 * log(2 * PI); | ^~ make: *** [/usr/lib64/R/etc/Makeconf:177: gibbs.o] Error 1 make: *** Waiting for unfinished jobs.... ERROR: compilation failed for package ‘mfa’ * removing ‘/build/r-mfa/src/mfa’ * restoring previous ‘/build/r-mfa/src/mfa’ ==> ERROR: A failure occurred in build(). Aborting... ==> ERROR: Build failed, check /var/lib/archbuild/BioArchLinux-x86_64/bio/build ``` </details> **Packages (please complete the following information):** - Package Name: r-mfa **Description** https://log.bioarchlinux.org/2022-01-29T15%3A51%3A16/r-mfa.log
True
[MAINTAIN] r-mfa - <!-- Please report the error of one package in one issue! Use multi issues to report multi bugs. Thanks! --> **Log of the bug** <details> ``` ==> Starting build()... * installing *source* package ‘mfa’ ... ** using staged installation ** libs g++ -std=gnu++14 -I"/usr/include/R/" -DNDEBUG -I'/usr/lib/R/library/Rcpp/include' -D_FORTIFY_SOURCE=2 -fpic -march=x86-64 -mtune=generic -O2 -pipe -fno-plt -c RcppExports.cpp -o RcppExports.o g++ -std=gnu++14 -I"/usr/include/R/" -DNDEBUG -I'/usr/lib/R/library/Rcpp/include' -D_FORTIFY_SOURCE=2 -fpic -march=x86-64 -mtune=generic -O2 -pipe -fno-plt -c gibbs.cpp -o gibbs.o gibbs.cpp: In function ‘double log_d_norm(double, double, double)’: gibbs.cpp:13:31: error: ‘PI’ was not declared in this scope 13 | double ldn = -0.5 * log(2 * PI); | ^~ make: *** [/usr/lib64/R/etc/Makeconf:177: gibbs.o] Error 1 make: *** Waiting for unfinished jobs.... ERROR: compilation failed for package ‘mfa’ * removing ‘/build/r-mfa/src/mfa’ * restoring previous ‘/build/r-mfa/src/mfa’ ==> ERROR: A failure occurred in build(). Aborting... ==> ERROR: Build failed, check /var/lib/archbuild/BioArchLinux-x86_64/bio/build ``` </details> **Packages (please complete the following information):** - Package Name: r-mfa **Description** https://log.bioarchlinux.org/2022-01-29T15%3A51%3A16/r-mfa.log
main
r mfa please report the error of one package in one issue use multi issues to report multi bugs thanks log of the bug starting build installing source package ‘mfa’ using staged installation libs g std gnu i usr include r dndebug i usr lib r library rcpp include d fortify source fpic march mtune generic pipe fno plt c rcppexports cpp o rcppexports o g std gnu i usr include r dndebug i usr lib r library rcpp include d fortify source fpic march mtune generic pipe fno plt c gibbs cpp o gibbs o gibbs cpp in function ‘double log d norm double double double ’ gibbs cpp error ‘pi’ was not declared in this scope double ldn log pi make error make waiting for unfinished jobs error compilation failed for package ‘mfa’ removing ‘ build r mfa src mfa’ restoring previous ‘ build r mfa src mfa’ error a failure occurred in build aborting error build failed check var lib archbuild bioarchlinux bio build packages please complete the following information package name r mfa description
1
5,363
26,982,805,397
IssuesEvent
2023-02-09 14:16:07
precice/precice
https://api.github.com/repos/precice/precice
reopened
Discussion: Restructuring the Integration Tests
maintainability
As our codebase and the application areas getting bigger, our integration tests are also getting bigger, which is really nice. However, currently, it is getting dirtier and dirtier both in terms of configuration files and the longer `SerialTests.cpp` and `ParallelTests.cpp` files. **Describe the solution you propose.** As the first step, I suggest to move the integration tests into a separate folder, such as `src/precice/tests/config-files`. While doing so, we can fix a naming scheme for the configuration files as well. In addition, we can even remove duplicated configuration files and group them into seperate subfolders even. As the second step, which would be more involved, we can try to seperate the integration tests into seperate files under the same directory. It would be not as trivial as the unit tests due to the nature of the integration tests but it should be doable. **Describe alternatives you've considered** Do not touch at all, everyone already hates implementing tests. **Additional context** We can check other projects about how they are approaching this topic.
True
Discussion: Restructuring the Integration Tests - As our codebase and the application areas getting bigger, our integration tests are also getting bigger, which is really nice. However, currently, it is getting dirtier and dirtier both in terms of configuration files and the longer `SerialTests.cpp` and `ParallelTests.cpp` files. **Describe the solution you propose.** As the first step, I suggest to move the integration tests into a separate folder, such as `src/precice/tests/config-files`. While doing so, we can fix a naming scheme for the configuration files as well. In addition, we can even remove duplicated configuration files and group them into seperate subfolders even. As the second step, which would be more involved, we can try to seperate the integration tests into seperate files under the same directory. It would be not as trivial as the unit tests due to the nature of the integration tests but it should be doable. **Describe alternatives you've considered** Do not touch at all, everyone already hates implementing tests. **Additional context** We can check other projects about how they are approaching this topic.
main
discussion restructuring the integration tests as our codebase and the application areas getting bigger our integration tests are also getting bigger which is really nice however currently it is getting dirtier and dirtier both in terms of configuration files and the longer serialtests cpp and paralleltests cpp files describe the solution you propose as the first step i suggest to move the integration tests into a separate folder such as src precice tests config files while doing so we can fix a naming scheme for the configuration files as well in addition we can even remove duplicated configuration files and group them into seperate subfolders even as the second step which would be more involved we can try to seperate the integration tests into seperate files under the same directory it would be not as trivial as the unit tests due to the nature of the integration tests but it should be doable describe alternatives you ve considered do not touch at all everyone already hates implementing tests additional context we can check other projects about how they are approaching this topic
1
3,168
12,226,717,516
IssuesEvent
2020-05-03 12:14:34
gfleetwood/asteres
https://api.github.com/repos/gfleetwood/asteres
opened
timothycrosley/hypothesis-auto (208208579)
Python maintain
https://github.com/timothycrosley/hypothesis-auto An extensions for Hypothesis that provides fully automatic testing for type hinted functions
True
timothycrosley/hypothesis-auto (208208579) - https://github.com/timothycrosley/hypothesis-auto An extensions for Hypothesis that provides fully automatic testing for type hinted functions
main
timothycrosley hypothesis auto an extensions for hypothesis that provides fully automatic testing for type hinted functions
1
742
4,349,450,299
IssuesEvent
2016-07-30 15:37:41
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
apt_rpm.py does not support installation using "/usr/bin/rpm" as mentioned
feature_idea waiting_on_maintainer
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea apt_rpm does not support managing packages with rpm on linux distros apt_rpm - apt_rpm package manager ##### COMPONENT NAME <!--- Name of the plugin/module/task --> apt_rpm ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> NO but may be required if to support additinal capabilities of RPM command like non root installation ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Running Ansible from and to : RHEL 7.2 ##### SUMMARY <!--- Explain the problem briefly --> As written for package apt_rpm apt_rpm documentation says it manages installation,update and removal using low-level (rpm) and high-level (apt-get) package manager binaries however if i have to use this package for my linux distro like fedora/RHEL then i dont have binary apt-get installed on my machine , but i do have "/usr/bin/rpm" so apt_rpm module is supposed to work "- apt_rpm: pkg=foo state=present" but it throws error saying {"changed": false, "cmd": "/usr/bin/apt-get -y install /home/xyz.rpm '>' /dev/null", "failed": true, "msg": "[Errno 2] No such file or directory", "rc": 2} this is becasue there is bug in code line numbers 127 and 153 https://github.com/ansible/ansible-modules-core/blob/devel/packaging/os/apt_rpm.py apt_rpm.py is supporting only APT_PATH and not RPM_PATH (line no 73) line number 153 should be fixed and "or" condition should be changed to "and" also apt_rpm.py needs to support RPM_PATH since this package uses RPM method to manage packages so facility to support RPM installation upgrade removal via non root user should also be added in apt_rpm.py code because rpm supports non root root installation using options like --dbpath --relocate http://docs.ansible.com/ansible/apt_rpm_module.html ##### STEPS TO REPRODUCE <!--- --> use below playbook on any fedora like OS and try to install wget package but it gives error : {"changed": false, "cmd": "/usr/bin/apt-get -y install /home/xyz.rpm '>' /dev/null", "failed": true, "msg": "[Errno 2] No such file or directory", "rc": 2} <!--- Paste example playbooks or commands between quotes below --> ``` --- - hosts: RMP tasks: - apt_rpm: pkg=wget state=present - shell: echo "hello world" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> i was expecting rpm to get installted on my RHEL machine using "- apt_rpm: pkg=foo state=present" ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> Yes <!--- Paste verbatim command output between quotes below --> ``` {"changed": false, "cmd": "/usr/bin/apt-get -y install /home/xyz.rpm '>' /dev/null", "failed": true, "msg": "[Errno 2] No such file or directory", "rc": 2} ```
True
apt_rpm.py does not support installation using "/usr/bin/rpm" as mentioned - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Feature Idea apt_rpm does not support managing packages with rpm on linux distros apt_rpm - apt_rpm package manager ##### COMPONENT NAME <!--- Name of the plugin/module/task --> apt_rpm ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> NO but may be required if to support additinal capabilities of RPM command like non root installation ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> Running Ansible from and to : RHEL 7.2 ##### SUMMARY <!--- Explain the problem briefly --> As written for package apt_rpm apt_rpm documentation says it manages installation,update and removal using low-level (rpm) and high-level (apt-get) package manager binaries however if i have to use this package for my linux distro like fedora/RHEL then i dont have binary apt-get installed on my machine , but i do have "/usr/bin/rpm" so apt_rpm module is supposed to work "- apt_rpm: pkg=foo state=present" but it throws error saying {"changed": false, "cmd": "/usr/bin/apt-get -y install /home/xyz.rpm '>' /dev/null", "failed": true, "msg": "[Errno 2] No such file or directory", "rc": 2} this is becasue there is bug in code line numbers 127 and 153 https://github.com/ansible/ansible-modules-core/blob/devel/packaging/os/apt_rpm.py apt_rpm.py is supporting only APT_PATH and not RPM_PATH (line no 73) line number 153 should be fixed and "or" condition should be changed to "and" also apt_rpm.py needs to support RPM_PATH since this package uses RPM method to manage packages so facility to support RPM installation upgrade removal via non root user should also be added in apt_rpm.py code because rpm supports non root root installation using options like --dbpath --relocate http://docs.ansible.com/ansible/apt_rpm_module.html ##### STEPS TO REPRODUCE <!--- --> use below playbook on any fedora like OS and try to install wget package but it gives error : {"changed": false, "cmd": "/usr/bin/apt-get -y install /home/xyz.rpm '>' /dev/null", "failed": true, "msg": "[Errno 2] No such file or directory", "rc": 2} <!--- Paste example playbooks or commands between quotes below --> ``` --- - hosts: RMP tasks: - apt_rpm: pkg=wget state=present - shell: echo "hello world" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> i was expecting rpm to get installted on my RHEL machine using "- apt_rpm: pkg=foo state=present" ##### ACTUAL RESULTS <!--- What actually happened? If possible run with extra verbosity (-vvvv) --> Yes <!--- Paste verbatim command output between quotes below --> ``` {"changed": false, "cmd": "/usr/bin/apt-get -y install /home/xyz.rpm '>' /dev/null", "failed": true, "msg": "[Errno 2] No such file or directory", "rc": 2} ```
main
apt rpm py does not support installation using usr bin rpm as mentioned issue type feature idea apt rpm does not support managing packages with rpm on linux distros apt rpm apt rpm package manager component name apt rpm ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables no but may be required if to support additinal capabilities of rpm command like non root installation os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific running ansible from and to rhel summary as written for package apt rpm apt rpm documentation says it manages installation update and removal using low level rpm and high level apt get package manager binaries however if i have to use this package for my linux distro like fedora rhel then i dont have binary apt get installed on my machine but i do have usr bin rpm so apt rpm module is supposed to work apt rpm pkg foo state present but it throws error saying changed false cmd usr bin apt get y install home xyz rpm dev null failed true msg no such file or directory rc this is becasue there is bug in code line numbers and apt rpm py is supporting only apt path and not rpm path line no line number should be fixed and or condition should be changed to and also apt rpm py needs to support rpm path since this package uses rpm method to manage packages so facility to support rpm installation upgrade removal via non root user should also be added in apt rpm py code because rpm supports non root root installation using options like dbpath relocate steps to reproduce use below playbook on any fedora like os and try to install wget package but it gives error changed false cmd usr bin apt get y install home xyz rpm dev null failed true msg no such file or directory rc hosts rmp tasks apt rpm pkg wget state present shell echo hello world expected results i was expecting rpm to get installted on my rhel machine using apt rpm pkg foo state present actual results yes changed false cmd usr bin apt get y install home xyz rpm dev null failed true msg no such file or directory rc
1
168,162
26,607,572,084
IssuesEvent
2023-01-23 20:42:32
flutter/flutter
https://api.github.com/repos/flutter/flutter
closed
Dragging the iOS text selection handles is jumpy
a: text input severe: regression platform-ios framework f: material design
Dragging the handles horizontally in a multi-line text field on iOS is really inaccurate and not smooth. ![out](https://user-images.githubusercontent.com/389558/176048578-fbf12e06-4826-4b6f-9615-ac4df62fc714.gif) This definitely didn't happen before, so I've marked it as a regression. Android doesn't seem to have the same problem. It also doesn't seem to happen on the first line of text. ### Steps to reproduce 1. Run any app with a multiline text field on iOS, such as the one below. 2. Enter a few lines of text. 3. Select some text. 4. Drag either handle in a horizontal direction back and forth a few times Expected: The selection and handle follow the pointer smoothly. Actual: The selection and handle pause and jump around. <details> <summary>Code</summary> ```dart import 'package:flutter/material.dart'; void main() { runApp(const MyApp()); } class MyApp extends StatelessWidget { const MyApp({Key? key}) : super(key: key); @override Widget build(BuildContext context) { return MaterialApp( title: 'Flutter Demo', theme: ThemeData( primarySwatch: Colors.blue, ), home: const MyHomePage(title: 'Flutter Demo Home Page'), ); } } class MyHomePage extends StatelessWidget { const MyHomePage({Key? key, required this.title}) : super(key: key); final String title; @override Widget build(BuildContext context) { return Scaffold( appBar: AppBar( title: Text(title), ), body: Center( child: Column( mainAxisAlignment: MainAxisAlignment.center, children: <Widget>[ TextField(minLines: 1, maxLines: 4), ], ), ), ); } } ``` </details> CC @antholeole who discovered this.
1.0
Dragging the iOS text selection handles is jumpy - Dragging the handles horizontally in a multi-line text field on iOS is really inaccurate and not smooth. ![out](https://user-images.githubusercontent.com/389558/176048578-fbf12e06-4826-4b6f-9615-ac4df62fc714.gif) This definitely didn't happen before, so I've marked it as a regression. Android doesn't seem to have the same problem. It also doesn't seem to happen on the first line of text. ### Steps to reproduce 1. Run any app with a multiline text field on iOS, such as the one below. 2. Enter a few lines of text. 3. Select some text. 4. Drag either handle in a horizontal direction back and forth a few times Expected: The selection and handle follow the pointer smoothly. Actual: The selection and handle pause and jump around. <details> <summary>Code</summary> ```dart import 'package:flutter/material.dart'; void main() { runApp(const MyApp()); } class MyApp extends StatelessWidget { const MyApp({Key? key}) : super(key: key); @override Widget build(BuildContext context) { return MaterialApp( title: 'Flutter Demo', theme: ThemeData( primarySwatch: Colors.blue, ), home: const MyHomePage(title: 'Flutter Demo Home Page'), ); } } class MyHomePage extends StatelessWidget { const MyHomePage({Key? key, required this.title}) : super(key: key); final String title; @override Widget build(BuildContext context) { return Scaffold( appBar: AppBar( title: Text(title), ), body: Center( child: Column( mainAxisAlignment: MainAxisAlignment.center, children: <Widget>[ TextField(minLines: 1, maxLines: 4), ], ), ), ); } } ``` </details> CC @antholeole who discovered this.
non_main
dragging the ios text selection handles is jumpy dragging the handles horizontally in a multi line text field on ios is really inaccurate and not smooth this definitely didn t happen before so i ve marked it as a regression android doesn t seem to have the same problem it also doesn t seem to happen on the first line of text steps to reproduce run any app with a multiline text field on ios such as the one below enter a few lines of text select some text drag either handle in a horizontal direction back and forth a few times expected the selection and handle follow the pointer smoothly actual the selection and handle pause and jump around code dart import package flutter material dart void main runapp const myapp class myapp extends statelesswidget const myapp key key super key key override widget build buildcontext context return materialapp title flutter demo theme themedata primaryswatch colors blue home const myhomepage title flutter demo home page class myhomepage extends statelesswidget const myhomepage key key required this title super key key final string title override widget build buildcontext context return scaffold appbar appbar title text title body center child column mainaxisalignment mainaxisalignment center children textfield minlines maxlines cc antholeole who discovered this
0
94,229
3,923,201,240
IssuesEvent
2016-04-22 10:10:37
SpongePowered/SpongeDocs
https://api.github.com/repos/SpongePowered/SpongeDocs
closed
Inconsistency of SpongeAPI
low priority
It is done ```SpongeAPI``` in some parts of the documentation and as ```Sponge API``` in others
1.0
Inconsistency of SpongeAPI - It is done ```SpongeAPI``` in some parts of the documentation and as ```Sponge API``` in others
non_main
inconsistency of spongeapi it is done spongeapi in some parts of the documentation and as sponge api in others
0
279,227
30,702,476,757
IssuesEvent
2023-07-27 01:33:26
nidhi7598/linux-3.0.35_CVE-2018-13405
https://api.github.com/repos/nidhi7598/linux-3.0.35_CVE-2018-13405
closed
CVE-2019-11811 (High) detected in linux-stable-rtv3.8.6 - autoclosed
Mend: dependency security vulnerability
## CVE-2019-11811 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_si_intf.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_si_intf.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_si_intf.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.0.4. There is a use-after-free upon attempted read access to /proc/ioports after the ipmi_si module is removed, related to drivers/char/ipmi/ipmi_si_intf.c, drivers/char/ipmi/ipmi_si_mem_io.c, and drivers/char/ipmi/ipmi_si_port_io.c. <p>Publish Date: 2019-05-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11811>CVE-2019-11811</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11811">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11811</a></p> <p>Release Date: 2019-05-07</p> <p>Fix Resolution: v5.1-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-11811 (High) detected in linux-stable-rtv3.8.6 - autoclosed - ## CVE-2019-11811 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv3.8.6</b></p></summary> <p> <p>Julia Cartwright's fork of linux-stable-rt.git</p> <p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-3.0.35_CVE-2018-13405/commit/662fbf6e1ed61fd353add2f52e2dd27e990364c7">662fbf6e1ed61fd353add2f52e2dd27e990364c7</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_si_intf.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_si_intf.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/char/ipmi/ipmi_si_intf.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.0.4. There is a use-after-free upon attempted read access to /proc/ioports after the ipmi_si module is removed, related to drivers/char/ipmi/ipmi_si_intf.c, drivers/char/ipmi/ipmi_si_mem_io.c, and drivers/char/ipmi/ipmi_si_port_io.c. <p>Publish Date: 2019-05-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-11811>CVE-2019-11811</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: High - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11811">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11811</a></p> <p>Release Date: 2019-05-07</p> <p>Fix Resolution: v5.1-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve high detected in linux stable autoclosed cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files drivers char ipmi ipmi si intf c drivers char ipmi ipmi si intf c drivers char ipmi ipmi si intf c vulnerability details an issue was discovered in the linux kernel before there is a use after free upon attempted read access to proc ioports after the ipmi si module is removed related to drivers char ipmi ipmi si intf c drivers char ipmi ipmi si mem io c and drivers char ipmi ipmi si port io c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
1,380
5,988,926,177
IssuesEvent
2017-06-02 06:58:05
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_updates expansion
affects_2.1 bot_broken feature_idea waiting_on_maintainer windows
##### ISSUE TYPE Feature Idea ##### COMPONENT NAME win_updates module ##### ANSIBLE VERSION N/A ##### SUMMARY To mitigate the risk of pushing out security updates to prod that may not have been tested on dev or staging, it would be nice to have an additional option in this module wherein updates could be applied based on KB#.
True
win_updates expansion - ##### ISSUE TYPE Feature Idea ##### COMPONENT NAME win_updates module ##### ANSIBLE VERSION N/A ##### SUMMARY To mitigate the risk of pushing out security updates to prod that may not have been tested on dev or staging, it would be nice to have an additional option in this module wherein updates could be applied based on KB#.
main
win updates expansion issue type feature idea component name win updates module ansible version n a summary to mitigate the risk of pushing out security updates to prod that may not have been tested on dev or staging it would be nice to have an additional option in this module wherein updates could be applied based on kb
1
4,583
23,811,233,837
IssuesEvent
2022-09-04 19:48:29
chocolatey-community/chocolatey-package-requests
https://api.github.com/repos/chocolatey-community/chocolatey-package-requests
closed
RFP - nheko-reborn
Status: Available For Maintainer(s) Embeddable
## Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL : https://github.com/Nheko-Reborn/nheko Direct download URL for the software / installer : https://github.com/Nheko-Reborn/nheko/releases/download/v0.8.2/nheko-v0.8.2-installer.exe Software summary / short description: live updated fork for https://community.chocolatey.org/packages/nheko/0.1.0.484-nightly
True
RFP - nheko-reborn - ## Checklist - [x] The package I am requesting does not already exist on https://chocolatey.org/packages; - [x] There is no open issue for this package; - [x] The issue title starts with 'RFP - '; - [x] The download URL is public and not locked behind a paywall / login; ## Package Details Software project URL : https://github.com/Nheko-Reborn/nheko Direct download URL for the software / installer : https://github.com/Nheko-Reborn/nheko/releases/download/v0.8.2/nheko-v0.8.2-installer.exe Software summary / short description: live updated fork for https://community.chocolatey.org/packages/nheko/0.1.0.484-nightly
main
rfp nheko reborn checklist the package i am requesting does not already exist on there is no open issue for this package the issue title starts with rfp the download url is public and not locked behind a paywall login package details software project url direct download url for the software installer software summary short description live updated fork for
1
282,447
30,889,335,888
IssuesEvent
2023-08-04 02:34:29
maddyCode23/linux-4.1.15
https://api.github.com/repos/maddyCode23/linux-4.1.15
reopened
CVE-2020-9391 (Medium) detected in linux-stable-rtv4.1.33, linux-stable-rtv4.1.33
Mend: dependency security vulnerability
## CVE-2020-9391 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linux-stable-rtv4.1.33</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel 5.4 and 5.5 through 5.5.6 on the AArch64 architecture. It ignores the top byte in the address passed to the brk system call, potentially moving the memory break downwards when the application expects it to move upwards, aka CID-dcde237319e6. This has been observed to cause heap corruption with the GNU C Library malloc implementation. <p>Publish Date: 2020-02-25 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9391>CVE-2020-9391</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-9391">https://www.linuxkernelcves.com/cves/CVE-2020-9391</a></p> <p>Release Date: 2020-02-25</p> <p>Fix Resolution: v5.6-rc3,v5.4.23,v5.5.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-9391 (Medium) detected in linux-stable-rtv4.1.33, linux-stable-rtv4.1.33 - ## CVE-2020-9391 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>linux-stable-rtv4.1.33</b>, <b>linux-stable-rtv4.1.33</b></p></summary> <p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel 5.4 and 5.5 through 5.5.6 on the AArch64 architecture. It ignores the top byte in the address passed to the brk system call, potentially moving the memory break downwards when the application expects it to move upwards, aka CID-dcde237319e6. This has been observed to cause heap corruption with the GNU C Library malloc implementation. <p>Publish Date: 2020-02-25 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-9391>CVE-2020-9391</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-9391">https://www.linuxkernelcves.com/cves/CVE-2020-9391</a></p> <p>Release Date: 2020-02-25</p> <p>Fix Resolution: v5.6-rc3,v5.4.23,v5.5.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in linux stable linux stable cve medium severity vulnerability vulnerable libraries linux stable linux stable vulnerability details an issue was discovered in the linux kernel and through on the architecture it ignores the top byte in the address passed to the brk system call potentially moving the memory break downwards when the application expects it to move upwards aka cid this has been observed to cause heap corruption with the gnu c library malloc implementation publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
30,922
2,729,656,404
IssuesEvent
2015-04-16 09:59:29
jkall/qgis-midvatten-plugin
https://api.github.com/repos/jkall/qgis-midvatten-plugin
opened
allow obsid dubplicates by introducing another ID as primary key
enhancement Priority-High
Introduce a unique observation id that is independent of obsid and name. This will allow the existance of duplicates among obsid. Major code revisions are needed. Several security checks are needed during imports and also, when obsid duplicates are found, user interaction to distinguish between observations of same obsid. Probably time to introduce a main table with id for each observation as the primary key. This table will hold all observations, no matter if they are points or lines.
1.0
allow obsid dubplicates by introducing another ID as primary key - Introduce a unique observation id that is independent of obsid and name. This will allow the existance of duplicates among obsid. Major code revisions are needed. Several security checks are needed during imports and also, when obsid duplicates are found, user interaction to distinguish between observations of same obsid. Probably time to introduce a main table with id for each observation as the primary key. This table will hold all observations, no matter if they are points or lines.
non_main
allow obsid dubplicates by introducing another id as primary key introduce a unique observation id that is independent of obsid and name this will allow the existance of duplicates among obsid major code revisions are needed several security checks are needed during imports and also when obsid duplicates are found user interaction to distinguish between observations of same obsid probably time to introduce a main table with id for each observation as the primary key this table will hold all observations no matter if they are points or lines
0
2,920
10,418,388,070
IssuesEvent
2019-09-15 08:09:19
NaluKit/nalu
https://api.github.com/repos/NaluKit/nalu
opened
Remove errorRoute implementation
maintainance
With the new upcoming error event feature, it gets clearer, that supporting both features will end up in a lot of more code. Some methods gets inconsistent and make the framework harder to use. Due to this fact, the errorRoute implementation gets removed. The error implementation can be easily reused. There will be an example how to do this.
True
Remove errorRoute implementation - With the new upcoming error event feature, it gets clearer, that supporting both features will end up in a lot of more code. Some methods gets inconsistent and make the framework harder to use. Due to this fact, the errorRoute implementation gets removed. The error implementation can be easily reused. There will be an example how to do this.
main
remove errorroute implementation with the new upcoming error event feature it gets clearer that supporting both features will end up in a lot of more code some methods gets inconsistent and make the framework harder to use due to this fact the errorroute implementation gets removed the error implementation can be easily reused there will be an example how to do this
1
4,860
25,012,288,884
IssuesEvent
2022-11-03 16:04:42
centerofci/mathesar
https://api.github.com/repos/centerofci/mathesar
opened
Re-focus table cell after opening record selector on that cell
type: bug work: frontend status: ready restricted: maintainers
## Steps to reproduce 1. Go to the Table Page for a table that has a foreign key column. 1. Open the Record Selector from an FK cell. 1. Close the Record Selector, either via Esc, the close button, or by submitting a value. 1. Expect the cell within the table to remain selected and active. 1. Instead observe that after closing the Record Selector, the cell is no longer active.
True
Re-focus table cell after opening record selector on that cell - ## Steps to reproduce 1. Go to the Table Page for a table that has a foreign key column. 1. Open the Record Selector from an FK cell. 1. Close the Record Selector, either via Esc, the close button, or by submitting a value. 1. Expect the cell within the table to remain selected and active. 1. Instead observe that after closing the Record Selector, the cell is no longer active.
main
re focus table cell after opening record selector on that cell steps to reproduce go to the table page for a table that has a foreign key column open the record selector from an fk cell close the record selector either via esc the close button or by submitting a value expect the cell within the table to remain selected and active instead observe that after closing the record selector the cell is no longer active
1
641,040
20,816,049,162
IssuesEvent
2022-03-18 10:23:37
googleapis/java-analytics-data
https://api.github.com/repos/googleapis/java-analytics-data
reopened
com.example.analytics.QuickstartJsonCredentialsSampleTest: testQuickstart failed
type: bug priority: p1 :rotating_light: flakybot: issue flakybot: flaky api: analyticsdata
Note: #347 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: b38fe49352a865b22884949a465e2526f3abb425 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/4251a2d0-376c-4859-8abf-7c6b97a2502e), [Sponge](http://sponge2/4251a2d0-376c-4859-8abf-7c6b97a2502e) status: failed <details><summary>Test output</summary><br><pre>java.lang.NoClassDefFoundError: com/google/auth/Retryable at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:756) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:473) at java.net.URLClassLoader.access$100(URLClassLoader.java:74) at java.net.URLClassLoader$1.run(URLClassLoader.java:369) at java.net.URLClassLoader$1.run(URLClassLoader.java:363) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:362) at java.lang.ClassLoader.loadClass(ClassLoader.java:418) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) at java.lang.ClassLoader.loadClass(ClassLoader.java:351) at com.google.auth.oauth2.GoogleCredentials.fromStream(GoogleCredentials.java:172) at com.google.auth.oauth2.GoogleCredentials.fromStream(GoogleCredentials.java:139) at com.example.analytics.QuickstartJsonCredentialsSample.sampleRunReport(QuickstartJsonCredentialsSample.java:71) at com.example.analytics.QuickstartJsonCredentialsSampleTest.runSample(QuickstartJsonCredentialsSampleTest.java:43) at com.example.analytics.QuickstartJsonCredentialsSampleTest.testQuickstart(QuickstartJsonCredentialsSampleTest.java:52) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: java.lang.ClassNotFoundException: com.google.auth.Retryable at java.net.URLClassLoader.findClass(URLClassLoader.java:387) at java.lang.ClassLoader.loadClass(ClassLoader.java:418) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ... 45 more </pre></details>
1.0
com.example.analytics.QuickstartJsonCredentialsSampleTest: testQuickstart failed - Note: #347 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: b38fe49352a865b22884949a465e2526f3abb425 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/4251a2d0-376c-4859-8abf-7c6b97a2502e), [Sponge](http://sponge2/4251a2d0-376c-4859-8abf-7c6b97a2502e) status: failed <details><summary>Test output</summary><br><pre>java.lang.NoClassDefFoundError: com/google/auth/Retryable at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClass(ClassLoader.java:756) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:142) at java.net.URLClassLoader.defineClass(URLClassLoader.java:473) at java.net.URLClassLoader.access$100(URLClassLoader.java:74) at java.net.URLClassLoader$1.run(URLClassLoader.java:369) at java.net.URLClassLoader$1.run(URLClassLoader.java:363) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:362) at java.lang.ClassLoader.loadClass(ClassLoader.java:418) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) at java.lang.ClassLoader.loadClass(ClassLoader.java:351) at com.google.auth.oauth2.GoogleCredentials.fromStream(GoogleCredentials.java:172) at com.google.auth.oauth2.GoogleCredentials.fromStream(GoogleCredentials.java:139) at com.example.analytics.QuickstartJsonCredentialsSample.sampleRunReport(QuickstartJsonCredentialsSample.java:71) at com.example.analytics.QuickstartJsonCredentialsSampleTest.runSample(QuickstartJsonCredentialsSampleTest.java:43) at com.example.analytics.QuickstartJsonCredentialsSampleTest.testQuickstart(QuickstartJsonCredentialsSampleTest.java:52) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.BlockJUnit4ClassRunner$1.evaluate(BlockJUnit4ClassRunner.java:100) at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:366) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:103) at org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:63) at org.junit.runners.ParentRunner$4.run(ParentRunner.java:331) at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:79) at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:329) at org.junit.runners.ParentRunner.access$100(ParentRunner.java:66) at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:293) at org.junit.runners.ParentRunner$3.evaluate(ParentRunner.java:306) at org.junit.runners.ParentRunner.run(ParentRunner.java:413) at org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:364) at org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:272) at org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:237) at org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:158) at org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:428) at org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:162) at org.apache.maven.surefire.booter.ForkedBooter.run(ForkedBooter.java:562) at org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:548) Caused by: java.lang.ClassNotFoundException: com.google.auth.Retryable at java.net.URLClassLoader.findClass(URLClassLoader.java:387) at java.lang.ClassLoader.loadClass(ClassLoader.java:418) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:352) at java.lang.ClassLoader.loadClass(ClassLoader.java:351) ... 45 more </pre></details>
non_main
com example analytics quickstartjsoncredentialssampletest testquickstart failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output java lang noclassdeffounderror com google auth retryable at java lang classloader native method at java lang classloader defineclass classloader java at java security secureclassloader defineclass secureclassloader java at java net urlclassloader defineclass urlclassloader java at java net urlclassloader access urlclassloader java at java net urlclassloader run urlclassloader java at java net urlclassloader run urlclassloader java at java security accesscontroller doprivileged native method at java net urlclassloader findclass urlclassloader java at java lang classloader loadclass classloader java at sun misc launcher appclassloader loadclass launcher java at java lang classloader loadclass classloader java at com google auth googlecredentials fromstream googlecredentials java at com google auth googlecredentials fromstream googlecredentials java at com example analytics quickstartjsoncredentialssample samplerunreport quickstartjsoncredentialssample java at com example analytics quickstartjsoncredentialssampletest runsample quickstartjsoncredentialssampletest java at com example analytics quickstartjsoncredentialssampletest testquickstart quickstartjsoncredentialssampletest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at org junit runners parentrunner evaluate parentrunner java at org junit runners evaluate java at org junit runners parentrunner runleaf parentrunner java at org junit runners runchild java at org junit runners runchild java at org junit runners parentrunner run parentrunner java at org junit runners parentrunner schedule parentrunner java at org junit runners parentrunner runchildren parentrunner java at org junit runners parentrunner access parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner evaluate parentrunner java at org junit runners parentrunner run parentrunner java at org apache maven surefire execute java at org apache maven surefire executewithrerun java at org apache maven surefire executetestset java at org apache maven surefire invoke java at org apache maven surefire booter forkedbooter runsuitesinprocess forkedbooter java at org apache maven surefire booter forkedbooter execute forkedbooter java at org apache maven surefire booter forkedbooter run forkedbooter java at org apache maven surefire booter forkedbooter main forkedbooter java caused by java lang classnotfoundexception com google auth retryable at java net urlclassloader findclass urlclassloader java at java lang classloader loadclass classloader java at sun misc launcher appclassloader loadclass launcher java at java lang classloader loadclass classloader java more
0
126,825
17,111,127,957
IssuesEvent
2021-07-10 10:14:34
nextcloud/server
https://api.github.com/repos/nextcloud/server
closed
File list scroll bar spanning whole page and triggered too early
0. Needs triage design enhancement needs info stale
Just updated to 16.0.4. Debian 10, PHP 7.3, Chrome browser. Two issues with the scroll bar. Firstly, when the file list length triggers the need for a scroll bar, it spans the whole screen (including the header bar): ![image](https://user-images.githubusercontent.com/665111/63153891-50897300-c007-11e9-9b4e-6cca54786468.png) The tfoot element also appears to be oversized, resulting in the scroll bar being required prior to the list length exceeding available room: ![image](https://user-images.githubusercontent.com/665111/63153856-38195880-c007-11e9-99f9-441685482e42.png)
1.0
File list scroll bar spanning whole page and triggered too early - Just updated to 16.0.4. Debian 10, PHP 7.3, Chrome browser. Two issues with the scroll bar. Firstly, when the file list length triggers the need for a scroll bar, it spans the whole screen (including the header bar): ![image](https://user-images.githubusercontent.com/665111/63153891-50897300-c007-11e9-9b4e-6cca54786468.png) The tfoot element also appears to be oversized, resulting in the scroll bar being required prior to the list length exceeding available room: ![image](https://user-images.githubusercontent.com/665111/63153856-38195880-c007-11e9-99f9-441685482e42.png)
non_main
file list scroll bar spanning whole page and triggered too early just updated to debian php chrome browser two issues with the scroll bar firstly when the file list length triggers the need for a scroll bar it spans the whole screen including the header bar the tfoot element also appears to be oversized resulting in the scroll bar being required prior to the list length exceeding available room
0
385,865
26,657,202,812
IssuesEvent
2023-01-25 17:47:23
redpanda-data/documentation
https://api.github.com/repos/redpanda-data/documentation
closed
Update default segment size
documentation P2 Feature Q1 23.1 Feature
### Describe the Issue The default value of `log_segment_size` will be 128MiB as of 23.1 See https://github.com/redpanda-data/redpanda/pull/8180 <!-- What problem does this issue solve for customers? Who is the audience for this update? For example, Infosec admins, cluster admins, or developers. Do you know the context in which users will likely search for this content? If so, share it. If this is a new feature, add a label for the version in which the update is expected. --> ### Updates to existing documentation <!-- Provide the URL of the page(s) to which the updates apply. Which topic(s) should be updated? What is the requested fix? Describe what is wrong in the existing doc and include screenshots if possible. Then provide the correct information. Is this request to document an existing Redpanda feature that is not currently documented? --> ### New feature or content gap requests <!-- If this is a new feature, provide material that will help the writer draft the new or updated content. If an issue only includes the name of a feature and release, it will be marked “More Info Needed” or “Blocked”. Ideally, the issue will provide: - Links to internal wikis, PRDs, design docs, KIPs, training material, or slide decks. - Links to engineering PRs that implement the update. - Links to blog posts, if applicable. - Identify specific Slack channels that the writer can join to get more details about the feature. - Links to Slack conversations (RP and RP Community) about the issue. - Names of SMEs who can answer questions and act as reviewers (at least one engineer and one PM). - Any additional information that will be of use to the writer. --> ### If new feature, in which release is it included or expected? ### If the requested update is from customer feedback or a Community slack conversation, provide a link: ### Additional notes <!-- Documentation Issues must be distinct and not overlap. If filing an Issue for a feature that spans platform and cloud, you must file two separate Issues. Each Issue will drive the new content in GitHub for the respective platform or cloud doc. Include: - Relevant GitHub issues and pull requests - Dependencies on other features or components - Specific Redpanda contributors to notify -->
1.0
Update default segment size - ### Describe the Issue The default value of `log_segment_size` will be 128MiB as of 23.1 See https://github.com/redpanda-data/redpanda/pull/8180 <!-- What problem does this issue solve for customers? Who is the audience for this update? For example, Infosec admins, cluster admins, or developers. Do you know the context in which users will likely search for this content? If so, share it. If this is a new feature, add a label for the version in which the update is expected. --> ### Updates to existing documentation <!-- Provide the URL of the page(s) to which the updates apply. Which topic(s) should be updated? What is the requested fix? Describe what is wrong in the existing doc and include screenshots if possible. Then provide the correct information. Is this request to document an existing Redpanda feature that is not currently documented? --> ### New feature or content gap requests <!-- If this is a new feature, provide material that will help the writer draft the new or updated content. If an issue only includes the name of a feature and release, it will be marked “More Info Needed” or “Blocked”. Ideally, the issue will provide: - Links to internal wikis, PRDs, design docs, KIPs, training material, or slide decks. - Links to engineering PRs that implement the update. - Links to blog posts, if applicable. - Identify specific Slack channels that the writer can join to get more details about the feature. - Links to Slack conversations (RP and RP Community) about the issue. - Names of SMEs who can answer questions and act as reviewers (at least one engineer and one PM). - Any additional information that will be of use to the writer. --> ### If new feature, in which release is it included or expected? ### If the requested update is from customer feedback or a Community slack conversation, provide a link: ### Additional notes <!-- Documentation Issues must be distinct and not overlap. If filing an Issue for a feature that spans platform and cloud, you must file two separate Issues. Each Issue will drive the new content in GitHub for the respective platform or cloud doc. Include: - Relevant GitHub issues and pull requests - Dependencies on other features or components - Specific Redpanda contributors to notify -->
non_main
update default segment size describe the issue the default value of log segment size will be as of see what problem does this issue solve for customers who is the audience for this update for example infosec admins cluster admins or developers do you know the context in which users will likely search for this content if so share it if this is a new feature add a label for the version in which the update is expected updates to existing documentation provide the url of the page s to which the updates apply which topic s should be updated what is the requested fix describe what is wrong in the existing doc and include screenshots if possible then provide the correct information is this request to document an existing redpanda feature that is not currently documented new feature or content gap requests if this is a new feature provide material that will help the writer draft the new or updated content if an issue only includes the name of a feature and release it will be marked “more info needed” or “blocked” ideally the issue will provide links to internal wikis prds design docs kips training material or slide decks links to engineering prs that implement the update links to blog posts if applicable identify specific slack channels that the writer can join to get more details about the feature links to slack conversations rp and rp community about the issue names of smes who can answer questions and act as reviewers at least one engineer and one pm any additional information that will be of use to the writer if new feature in which release is it included or expected if the requested update is from customer feedback or a community slack conversation provide a link additional notes documentation issues must be distinct and not overlap if filing an issue for a feature that spans platform and cloud you must file two separate issues each issue will drive the new content in github for the respective platform or cloud doc include relevant github issues and pull requests dependencies on other features or components specific redpanda contributors to notify
0
138,134
18,771,460,741
IssuesEvent
2021-11-06 22:48:48
samqws-marketing/box_mojito
https://api.github.com/repos/samqws-marketing/box_mojito
opened
CVE-2020-28498 (Medium) detected in elliptic-6.5.3.tgz
security vulnerability
## CVE-2020-28498 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.3.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.3.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.3.tgz</a></p> <p> Dependency Hierarchy: - webpack-2.7.0.tgz (Root Library) - node-libs-browser-2.2.0.tgz - crypto-browserify-3.12.0.tgz - create-ecdh-4.0.3.tgz - :x: **elliptic-6.5.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_mojito/commit/65290aeb818102fa2443a637efdccebebfed1eb9">65290aeb818102fa2443a637efdccebebfed1eb9</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed. <p>Publish Date: 2021-02-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498>CVE-2020-28498</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498</a></p> <p>Release Date: 2021-02-02</p> <p>Fix Resolution: v6.5.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.5.3","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"webpack:2.7.0;node-libs-browser:2.2.0;crypto-browserify:3.12.0;create-ecdh:4.0.3;elliptic:6.5.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v6.5.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28498","vulnerabilityDetails":"The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498","cvss3Severity":"medium","cvss3Score":"6.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Changed","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-28498 (Medium) detected in elliptic-6.5.3.tgz - ## CVE-2020-28498 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>elliptic-6.5.3.tgz</b></p></summary> <p>EC cryptography</p> <p>Library home page: <a href="https://registry.npmjs.org/elliptic/-/elliptic-6.5.3.tgz">https://registry.npmjs.org/elliptic/-/elliptic-6.5.3.tgz</a></p> <p> Dependency Hierarchy: - webpack-2.7.0.tgz (Root Library) - node-libs-browser-2.2.0.tgz - crypto-browserify-3.12.0.tgz - create-ecdh-4.0.3.tgz - :x: **elliptic-6.5.3.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/samqws-marketing/box_mojito/commit/65290aeb818102fa2443a637efdccebebfed1eb9">65290aeb818102fa2443a637efdccebebfed1eb9</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed. <p>Publish Date: 2021-02-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498>CVE-2020-28498</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Changed - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-28498</a></p> <p>Release Date: 2021-02-02</p> <p>Fix Resolution: v6.5.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"elliptic","packageVersion":"6.5.3","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"webpack:2.7.0;node-libs-browser:2.2.0;crypto-browserify:3.12.0;create-ecdh:4.0.3;elliptic:6.5.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v6.5.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-28498","vulnerabilityDetails":"The package elliptic before 6.5.4 are vulnerable to Cryptographic Issues via the secp256k1 implementation in elliptic/ec/key.js. There is no check to confirm that the public key point passed into the derive function actually exists on the secp256k1 curve. This results in the potential for the private key used in this implementation to be revealed after a number of ECDH operations are performed.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-28498","cvss3Severity":"medium","cvss3Score":"6.8","cvss3Metrics":{"A":"None","AC":"High","PR":"None","S":"Changed","C":"High","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> -->
non_main
cve medium detected in elliptic tgz cve medium severity vulnerability vulnerable library elliptic tgz ec cryptography library home page a href dependency hierarchy webpack tgz root library node libs browser tgz crypto browserify tgz create ecdh tgz x elliptic tgz vulnerable library found in head commit a href found in base branch master vulnerability details the package elliptic before are vulnerable to cryptographic issues via the implementation in elliptic ec key js there is no check to confirm that the public key point passed into the derive function actually exists on the curve this results in the potential for the private key used in this implementation to be revealed after a number of ecdh operations are performed publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope changed impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree webpack node libs browser crypto browserify create ecdh elliptic isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the package elliptic before are vulnerable to cryptographic issues via the implementation in elliptic ec key js there is no check to confirm that the public key point passed into the derive function actually exists on the curve this results in the potential for the private key used in this implementation to be revealed after a number of ecdh operations are performed vulnerabilityurl
0
4,899
25,162,496,800
IssuesEvent
2022-11-10 17:52:54
mozilla/foundation.mozilla.org
https://api.github.com/repos/mozilla/foundation.mozilla.org
opened
Setup `django-pattern-library`
engineering Maintain
# Description > The [django-pattern-library](https://pypi.org/project/django-pattern-library/) package automates the maintenance of UI pattern libraries or styleguides for Django projects, and allows developers to experiment with Django templates without having to create Django views and models. > * Create reusable patterns by creating Django templates files as usual. > * All patterns automatically show up in the pattern library’s interface. > * Define data as YAML files for the templates to render with the relevant Django context. > * Override Django templates tags as needed to mock the template’s dependencies. > * Document your patterns with Markdown. This allows us to decouple the development of frontend and backend. It supports us in keeping the templates structured and organized. It also serves as a library of existing template components, which should improve the reuse of templates. # Acceptance criteria - [ ] Django pattern library is installed and configured in the project. - [ ] A couple of templates are added as examples to the pattern library. - [ ] The dev team had a small workshop to get introduced to working with the pattern library.
True
Setup `django-pattern-library` - # Description > The [django-pattern-library](https://pypi.org/project/django-pattern-library/) package automates the maintenance of UI pattern libraries or styleguides for Django projects, and allows developers to experiment with Django templates without having to create Django views and models. > * Create reusable patterns by creating Django templates files as usual. > * All patterns automatically show up in the pattern library’s interface. > * Define data as YAML files for the templates to render with the relevant Django context. > * Override Django templates tags as needed to mock the template’s dependencies. > * Document your patterns with Markdown. This allows us to decouple the development of frontend and backend. It supports us in keeping the templates structured and organized. It also serves as a library of existing template components, which should improve the reuse of templates. # Acceptance criteria - [ ] Django pattern library is installed and configured in the project. - [ ] A couple of templates are added as examples to the pattern library. - [ ] The dev team had a small workshop to get introduced to working with the pattern library.
main
setup django pattern library description the package automates the maintenance of ui pattern libraries or styleguides for django projects and allows developers to experiment with django templates without having to create django views and models create reusable patterns by creating django templates files as usual all patterns automatically show up in the pattern library’s interface define data as yaml files for the templates to render with the relevant django context override django templates tags as needed to mock the template’s dependencies document your patterns with markdown this allows us to decouple the development of frontend and backend it supports us in keeping the templates structured and organized it also serves as a library of existing template components which should improve the reuse of templates acceptance criteria django pattern library is installed and configured in the project a couple of templates are added as examples to the pattern library the dev team had a small workshop to get introduced to working with the pattern library
1
915
4,613,786,430
IssuesEvent
2016-09-25 06:54:21
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
mongodb_user isn't idempotent for mongo 2.4
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME mongodb_user ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### OS / ENVIRONMENT CentOS 7.2.1511 relevant packages installed: mongo-10gen-server-2.4.14-mongodb_1.x86_64 mongo-10gen-2.4.14-mongodb_1.x86_64 python-pymongo-2.5.2-4.el7.x86_64 mongodb repository used: baseurl = http://downloads-distro.mongodb.org/repo/redhat/os/x86_64/ exclude = mongodb-org mongodb-org-server gpgcheck = 0 name = MongoDB Respositry ##### SUMMARY A user is added just fine the first time the playbook is run. On the second run, mongodb_user spits out a stack trace. Here's the the line of code in the module that causes the stack trace: `if mongo_user['user'] == user and mongo_user['db'] == db_name:` This is a problem because there is no 'db' field in the user object in mongo 2.4 Here's some example output for a find command: `> db.system.users.find() { "_id" : ObjectId("57b23c29da0d7b110931329c"), "pwd" : "<pwd hash>", "readOnly" : false, "user" : "admin_user" }` ##### STEPS TO REPRODUCE Here's a play to reproduce the problem: `mongodb_user: database=admin name=admin_user password=12345` It will work on the first run, but fail on the second run because it can't find a db field. Verbose Stack Trace: task path: /home/centos/installer/roles/mongo/tasks/main.yml:28 <cent7min-1-1> ESTABLISH SSH CONNECTION FOR USER: None <cent7min-1-1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/centos/.ansible/cp/ansible-ssh-%h-%p-%r cent7min-1-1 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619 `" && echo ansible-tmp-1471299402.49-91867766701619="` echo $HOME/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619 `" ) && sleep 0'"'"'' <cent7min-1-1> PUT /tmp/tmpWGQr4X TO /home/centos/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619/mongodb_user <cent7min-1-1> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/centos/.ansible/cp/ansible-ssh-%h-%p-%r '[cent7min-1-1]' <cent7min-1-1> ESTABLISH SSH CONNECTION FOR USER: None <cent7min-1-1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/centos/.ansible/cp/ansible-ssh-%h-%p-%r -tt cent7min-1-1 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-kapjnzmgxofqvhcqnynztglnmuxjhqup; LANG=en_US LC_ALL=en_US LC_MESSAGES=en_US /usr/bin/python /home/centos/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619/mongodb_user; rm -rf "/home/centos/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"'' fatal: [cent7min-1-1]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "mongodb_user"}, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_ARr9BX/ansible_module_mongodb_user.py\", line 352, in <module>\r\n main()\r\n File \"/tmp/ansible_ARr9BX/ansible_module_mongodb_user.py\", line 324, in main\r\n uinfo = user_find(client, user, db_name)\r\n File \"/tmp/ansible_ARr9BX/ansible_module_mongodb_user.py\", line 170, in user_find\r\n if mongo_user['user'] == user and mongo_user['db'] == db_name:\r\nKeyError: 'db'\r\n", "msg": "MODULE FAILURE", "parsed": false} ##### EXPECTED BEHAVIOR Don't check for the 'db' field if the mongo version is 2.4
True
mongodb_user isn't idempotent for mongo 2.4 - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME mongodb_user ##### ANSIBLE VERSION ``` ansible 2.1.1.0 ``` ##### OS / ENVIRONMENT CentOS 7.2.1511 relevant packages installed: mongo-10gen-server-2.4.14-mongodb_1.x86_64 mongo-10gen-2.4.14-mongodb_1.x86_64 python-pymongo-2.5.2-4.el7.x86_64 mongodb repository used: baseurl = http://downloads-distro.mongodb.org/repo/redhat/os/x86_64/ exclude = mongodb-org mongodb-org-server gpgcheck = 0 name = MongoDB Respositry ##### SUMMARY A user is added just fine the first time the playbook is run. On the second run, mongodb_user spits out a stack trace. Here's the the line of code in the module that causes the stack trace: `if mongo_user['user'] == user and mongo_user['db'] == db_name:` This is a problem because there is no 'db' field in the user object in mongo 2.4 Here's some example output for a find command: `> db.system.users.find() { "_id" : ObjectId("57b23c29da0d7b110931329c"), "pwd" : "<pwd hash>", "readOnly" : false, "user" : "admin_user" }` ##### STEPS TO REPRODUCE Here's a play to reproduce the problem: `mongodb_user: database=admin name=admin_user password=12345` It will work on the first run, but fail on the second run because it can't find a db field. Verbose Stack Trace: task path: /home/centos/installer/roles/mongo/tasks/main.yml:28 <cent7min-1-1> ESTABLISH SSH CONNECTION FOR USER: None <cent7min-1-1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/centos/.ansible/cp/ansible-ssh-%h-%p-%r cent7min-1-1 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619 `" && echo ansible-tmp-1471299402.49-91867766701619="` echo $HOME/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619 `" ) && sleep 0'"'"'' <cent7min-1-1> PUT /tmp/tmpWGQr4X TO /home/centos/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619/mongodb_user <cent7min-1-1> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/centos/.ansible/cp/ansible-ssh-%h-%p-%r '[cent7min-1-1]' <cent7min-1-1> ESTABLISH SSH CONNECTION FOR USER: None <cent7min-1-1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o ConnectTimeout=10 -o ControlPath=/home/centos/.ansible/cp/ansible-ssh-%h-%p-%r -tt cent7min-1-1 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-kapjnzmgxofqvhcqnynztglnmuxjhqup; LANG=en_US LC_ALL=en_US LC_MESSAGES=en_US /usr/bin/python /home/centos/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619/mongodb_user; rm -rf "/home/centos/.ansible/tmp/ansible-tmp-1471299402.49-91867766701619/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"'' fatal: [cent7min-1-1]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "mongodb_user"}, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_ARr9BX/ansible_module_mongodb_user.py\", line 352, in <module>\r\n main()\r\n File \"/tmp/ansible_ARr9BX/ansible_module_mongodb_user.py\", line 324, in main\r\n uinfo = user_find(client, user, db_name)\r\n File \"/tmp/ansible_ARr9BX/ansible_module_mongodb_user.py\", line 170, in user_find\r\n if mongo_user['user'] == user and mongo_user['db'] == db_name:\r\nKeyError: 'db'\r\n", "msg": "MODULE FAILURE", "parsed": false} ##### EXPECTED BEHAVIOR Don't check for the 'db' field if the mongo version is 2.4
main
mongodb user isn t idempotent for mongo issue type bug report component name mongodb user ansible version ansible os environment centos relevant packages installed mongo server mongodb mongo mongodb python pymongo mongodb repository used baseurl exclude mongodb org mongodb org server gpgcheck name mongodb respositry summary a user is added just fine the first time the playbook is run on the second run mongodb user spits out a stack trace here s the the line of code in the module that causes the stack trace if mongo user user and mongo user db name this is a problem because there is no db field in the user object in mongo here s some example output for a find command db system users find id objectid pwd readonly false user admin user steps to reproduce here s a play to reproduce the problem mongodb user database admin name admin user password it will work on the first run but fail on the second run because it can t find a db field verbose stack trace task path home centos installer roles mongo tasks main yml establish ssh connection for user none ssh exec ssh c q o controlmaster auto o controlpersist o stricthostkeychecking no o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o connecttimeout o controlpath home centos ansible cp ansible ssh h p r bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home centos ansible tmp ansible tmp mongodb user ssh exec sftp b c o controlmaster auto o controlpersist o stricthostkeychecking no o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o connecttimeout o controlpath home centos ansible cp ansible ssh h p r establish ssh connection for user none ssh exec ssh c q o controlmaster auto o controlpersist o stricthostkeychecking no o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o connecttimeout o controlpath home centos ansible cp ansible ssh h p r tt bin sh c sudo h s n u root bin sh c echo become success kapjnzmgxofqvhcqnynztglnmuxjhqup lang en us lc all en us lc messages en us usr bin python home centos ansible tmp ansible tmp mongodb user rm rf home centos ansible tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module name mongodb user module stderr module stdout traceback most recent call last r n file tmp ansible ansible module mongodb user py line in r n main r n file tmp ansible ansible module mongodb user py line in main r n uinfo user find client user db name r n file tmp ansible ansible module mongodb user py line in user find r n if mongo user user and mongo user db name r nkeyerror db r n msg module failure parsed false expected behavior don t check for the db field if the mongo version is
1
5,008
25,733,017,122
IssuesEvent
2022-12-07 21:54:25
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[Bug]: when the table column name contains id and the id value is the same, we couldn't identify this two columns
type: question ❓ component: data-table status: waiting for maintainer response 💬 package: @carbon/react
### Package carbon-components-react ### Browser Chrome ### Package version 10.40.0 ### React version 16.13.1 ### Description This issue happens when we using carbon DataTable component. When the table column name contains `id` and the id value is the same, we couldn't identify this two columns, that will cause the other columns value will be overlapped by the first one. For example: create two tables, one has column `id` and the other one doesn't have this column. And the id value is the same. postgres=# select * from public.a2 ; a1 | a2 | a3 | a4 | a5 | a6 ----+----------------------+-------+----+----+---- 1 | C1-sjis-ok | A0_*& | | | 1 | C2-sjis-ok | abc | | | (2 rows) postgres=# select * from public.a3 ; id | a2 | a3 | a4 | a5 | a6 ----+----------------------+-------+----+----+---- 1 | C1-sjis-ok | A0_*& | | | 1 | C2-sjis-ok | abc | | | (2 rows) When we use carbon DataTable to render this two tables, we can see table a3 returned two rows and all the data is the same. ### Reproduction/example no need ### Steps to reproduce no need ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [ ] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
True
[Bug]: when the table column name contains id and the id value is the same, we couldn't identify this two columns - ### Package carbon-components-react ### Browser Chrome ### Package version 10.40.0 ### React version 16.13.1 ### Description This issue happens when we using carbon DataTable component. When the table column name contains `id` and the id value is the same, we couldn't identify this two columns, that will cause the other columns value will be overlapped by the first one. For example: create two tables, one has column `id` and the other one doesn't have this column. And the id value is the same. postgres=# select * from public.a2 ; a1 | a2 | a3 | a4 | a5 | a6 ----+----------------------+-------+----+----+---- 1 | C1-sjis-ok | A0_*& | | | 1 | C2-sjis-ok | abc | | | (2 rows) postgres=# select * from public.a3 ; id | a2 | a3 | a4 | a5 | a6 ----+----------------------+-------+----+----+---- 1 | C1-sjis-ok | A0_*& | | | 1 | C2-sjis-ok | abc | | | (2 rows) When we use carbon DataTable to render this two tables, we can see table a3 returned two rows and all the data is the same. ### Reproduction/example no need ### Steps to reproduce no need ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [ ] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
main
when the table column name contains id and the id value is the same we couldn t identify this two columns package carbon components react browser chrome package version react version description this issue happens when we using carbon datatable component when the table column name contains id and the id value is the same we couldn t identify this two columns that will cause the other columns value will be overlapped by the first one for example create two tables one has column id and the other one doesn t have this column and the id value is the same postgres select from public sjis ok sjis ok abc rows postgres select from public id sjis ok sjis ok abc rows when we use carbon datatable to render this two tables we can see table returned two rows and all the data is the same reproduction example no need steps to reproduce no need code of conduct i agree to follow this project s i checked the for duplicate problems
1
216,548
16,770,112,418
IssuesEvent
2021-06-14 13:54:07
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
opened
Use Visual Testing Tool to Monitor Production
VSP-testing-team
# Description Platform needs a solution for monitoring production for visual differences. Platform needs to detect issues in production before other stakeholders receive complaints. # Assumptions * [43,800 minutes in a month](http://www.easysurf.cc/monthtable.htm) # Questions * How frequently should this monitoring occur? * Where are issues reported to? (#oncall???)
1.0
Use Visual Testing Tool to Monitor Production - # Description Platform needs a solution for monitoring production for visual differences. Platform needs to detect issues in production before other stakeholders receive complaints. # Assumptions * [43,800 minutes in a month](http://www.easysurf.cc/monthtable.htm) # Questions * How frequently should this monitoring occur? * Where are issues reported to? (#oncall???)
non_main
use visual testing tool to monitor production description platform needs a solution for monitoring production for visual differences platform needs to detect issues in production before other stakeholders receive complaints assumptions questions how frequently should this monitoring occur where are issues reported to oncall
0
66,114
6,989,196,901
IssuesEvent
2017-12-14 15:28:24
edenlabllc/ehealth.api
https://api.github.com/repos/edenlabllc/ehealth.api
closed
Implement password expiration & rotation policy
BE epic/Auth kind/task project/CR status/test
Also, implement password expiration period: enforce password change every `PASSWORD_EXPIRATION_DAYS` (env variable) days. if password is expired - do not allow user to login untill new password is set, expiry all the refresh tokens do not allow to use 3 previously used passwords - save history - [x] update .erd - [x] update Mithrill apiary with new error codes related to #1556
1.0
Implement password expiration & rotation policy - Also, implement password expiration period: enforce password change every `PASSWORD_EXPIRATION_DAYS` (env variable) days. if password is expired - do not allow user to login untill new password is set, expiry all the refresh tokens do not allow to use 3 previously used passwords - save history - [x] update .erd - [x] update Mithrill apiary with new error codes related to #1556
non_main
implement password expiration rotation policy also implement password expiration period enforce password change every password expiration days env variable days if password is expired do not allow user to login untill new password is set expiry all the refresh tokens do not allow to use previously used passwords save history update erd update mithrill apiary with new error codes related to
0
1,736
6,574,864,289
IssuesEvent
2017-09-11 14:19:45
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
Document Git module's result
affects_2.1 docs_report waiting_on_maintainer
##### ISSUE TYPE Documentation Report ##### COMPONENT NAME git ##### ANSIBLE VERSION ``` ansible 2.1.2.0 config file = configured module search path = Default w/o overrides ``` ##### SUMMARY The Git module docs don't indicate that any result is available. In fact, the git module provides "after" and "before" results which are the commit IDs before and after the git command was run (I presume). This should be documented.
True
Document Git module's result - ##### ISSUE TYPE Documentation Report ##### COMPONENT NAME git ##### ANSIBLE VERSION ``` ansible 2.1.2.0 config file = configured module search path = Default w/o overrides ``` ##### SUMMARY The Git module docs don't indicate that any result is available. In fact, the git module provides "after" and "before" results which are the commit IDs before and after the git command was run (I presume). This should be documented.
main
document git module s result issue type documentation report component name git ansible version ansible config file configured module search path default w o overrides summary the git module docs don t indicate that any result is available in fact the git module provides after and before results which are the commit ids before and after the git command was run i presume this should be documented
1
5,728
30,285,775,716
IssuesEvent
2023-07-08 17:01:48
0ptim/JellyChat
https://api.github.com/repos/0ptim/JellyChat
closed
Directory cleanup
docs/maintainance area:general
Move code into `/src`. - [ ] `/job` - [ ] `/backend` Dependent on #31 because of possible merge conflicts.
True
Directory cleanup - Move code into `/src`. - [ ] `/job` - [ ] `/backend` Dependent on #31 because of possible merge conflicts.
main
directory cleanup move code into src job backend dependent on because of possible merge conflicts
1
1,861
6,577,413,389
IssuesEvent
2017-09-12 00:44:24
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
command module does not return stderr of `postfix check`
affects_2.1 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME command ##### ANSIBLE VERSION ``` ansible 2.1.0 config file = /home/jooadam/.ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ``` [ssh_connection] ssh_args = '' ``` ##### OS / ENVIRONMENT Ubuntu 12.04 / Ubuntu 15.10 ##### SUMMARY Running `postfix check` using the command module on error the stderr attribute is empty. ##### STEPS TO REPRODUCE ``` # /etc/postfix/main.cf jakjsdkjfksdfjskdjfkdj ``` ``` { "name": "validate postfix configuration", "action": "command postfix check" } ``` ##### EXPECTED RESULTS ``` fatal: [192.0.2.0]: FAILED! => {"changed": true, "cmd": ["postfix", "check"], "delta": "0:00:01.004179", "end": "2016-04-18 21:01:41.040310", "failed": true, "rc": 1, "start": "2016-04-18 21:01:40.036131", "stderr": "postfix: fatal: /etc/postfix/main.cf, line 1: missing '=' after attribute name: \"jakjsdkjfksdfjskdjfkdj\"", "stdout": "", "stdout_lines": [], "warnings": []} ``` ##### ACTUAL RESULTS ``` fatal: [192.0.2.0]: FAILED! => {"changed": true, "cmd": ["postfix", "check"], "delta": "0:00:01.004179", "end": "2016-04-18 21:01:41.040310", "failed": true, "rc": 1, "start": "2016-04-18 21:01:40.036131", "stderr": "", "stdout": "", "stdout_lines": [], "warnings": []} ```
True
command module does not return stderr of `postfix check` - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME command ##### ANSIBLE VERSION ``` ansible 2.1.0 config file = /home/jooadam/.ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION ``` [ssh_connection] ssh_args = '' ``` ##### OS / ENVIRONMENT Ubuntu 12.04 / Ubuntu 15.10 ##### SUMMARY Running `postfix check` using the command module on error the stderr attribute is empty. ##### STEPS TO REPRODUCE ``` # /etc/postfix/main.cf jakjsdkjfksdfjskdjfkdj ``` ``` { "name": "validate postfix configuration", "action": "command postfix check" } ``` ##### EXPECTED RESULTS ``` fatal: [192.0.2.0]: FAILED! => {"changed": true, "cmd": ["postfix", "check"], "delta": "0:00:01.004179", "end": "2016-04-18 21:01:41.040310", "failed": true, "rc": 1, "start": "2016-04-18 21:01:40.036131", "stderr": "postfix: fatal: /etc/postfix/main.cf, line 1: missing '=' after attribute name: \"jakjsdkjfksdfjskdjfkdj\"", "stdout": "", "stdout_lines": [], "warnings": []} ``` ##### ACTUAL RESULTS ``` fatal: [192.0.2.0]: FAILED! => {"changed": true, "cmd": ["postfix", "check"], "delta": "0:00:01.004179", "end": "2016-04-18 21:01:41.040310", "failed": true, "rc": 1, "start": "2016-04-18 21:01:40.036131", "stderr": "", "stdout": "", "stdout_lines": [], "warnings": []} ```
main
command module does not return stderr of postfix check issue type bug report component name command ansible version ansible config file home jooadam ansible cfg configured module search path default w o overrides configuration ssh args os environment ubuntu ubuntu summary running postfix check using the command module on error the stderr attribute is empty steps to reproduce etc postfix main cf jakjsdkjfksdfjskdjfkdj name validate postfix configuration action command postfix check expected results fatal failed changed true cmd delta end failed true rc start stderr postfix fatal etc postfix main cf line missing after attribute name jakjsdkjfksdfjskdjfkdj stdout stdout lines warnings actual results fatal failed changed true cmd delta end failed true rc start stderr stdout stdout lines warnings
1
7,004
2,596,348,029
IssuesEvent
2015-02-20 20:03:31
TheLens/land-records
https://api.github.com/repos/TheLens/land-records
closed
Strange behavior on map move
bug High priority
Search for "neighborhood:leonidas". Map shows ~10 results. Then click "redo search on map move". Move away from that neighborhood. Properties disappear on map. that seems right. however, move back to leonidas. The properties in leonidas are gone except for two.
1.0
Strange behavior on map move - Search for "neighborhood:leonidas". Map shows ~10 results. Then click "redo search on map move". Move away from that neighborhood. Properties disappear on map. that seems right. however, move back to leonidas. The properties in leonidas are gone except for two.
non_main
strange behavior on map move search for neighborhood leonidas map shows results then click redo search on map move move away from that neighborhood properties disappear on map that seems right however move back to leonidas the properties in leonidas are gone except for two
0
5,370
26,997,152,898
IssuesEvent
2023-02-10 02:42:50
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
Validation error with more than 50 parameters
stage/needs-feedback stage/bug-repro maintainer/need-followup
I didn't find this issue in the backlog. ### Description: My template has 53 parameters. `sam validate` throws this error: ``` % sam validate 2022-03-18 12:13:00 Loading policies from IAM... 2022-03-18 12:13:03 Finished loading policies from IAM. 2022-03-18 12:13:03 Template schema validation reported the following errors: [, P, a, r, a, m, e, t, e, r, s, ], , O, r, d, e, r, e, d, D, i, c, t, (, [, (, ', A, w, s, R, e, g, i, o, n, ', ,, , O, r, d, e, r, e, d, D, i, c, t, (, [, (, ', T, y, p, e, ', ,, , ', S, t, r, i, n, g, ', ), ,, , (, ', D, e, s, c, r, i, p, t, i, o, n, ', ,, , ', A, W, S, , R, e, g, i, o, n, , b, e, i, n, g, , u, s, e, d, ', ), ,, , (, ', D, e, f, a, u, l, t, ', ,, , ', u, s, -, e, a, s, t, -, 1, ', ), ], ), )... ``` I know the spaces in the error message are due to #3493 (still waiting on aws/serverless-application-model#2245). That's not the issue. This is what it looks like without the spaces (and with added line breaks): ``` [Parameters] OrderedDict( [ ( 'AwsRegion', OrderedDict( [ ('Type', 'String'), ('Description', 'AWS Region being used'), ('Default', 'us-east-1') ] ) ), [a bunch of other properties here...], ] ) has too many properties ``` The error at the bottom is the relevant part. Removing any three of the parameters makes this error go away. ### Steps to reproduce: 1. Run `sam validate` on a template with more than 50 parameters. ### Observed result: Validation error: `[Parameters] OrderedDict... has too many properties` ### Expected result: CloudFormation supposedly supports up to 200 parameters, so it should validate. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: MacOS 11.6.4 2. `sam --version`: 1.40.1 3. AWS region: us-east-1 `Add --debug flag to command you are running` ``` % sam validate --debug 2022-03-18 12:45:56,928 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics 2022-03-18 12:45:56,928 | Using config file: samconfig.toml, config environment: default 2022-03-18 12:45:56,928 | Expand command line arguments to: 2022-03-18 12:45:56,928 | --template_file=/Users/dave.kennedy/sam-api/template.yaml 2022-03-18 12:45:57 Loading policies from IAM... 2022-03-18 12:45:59 Finished loading policies from IAM. 2022-03-18 12:45:59 Template schema validation reported the following errors: [, P, a, r, a, m, e, t, e, r, s, ], , O, r, d, e, r, e, d, D, i, c, t, (, [, (, ', A, w, s, R, e, g, i, o, n, ', ,, , O, r, d, e, r, e, d, D, i, c, t, (, [, (, ', T, y, p, e, ', ,, , ', S, t, r, i, n, g, ', ), ,, , (, ', D, e, s, c, r, i, p, t, i, o, n, ', ,, , ', A, W, S, , R, e, g, i, o, n, , b, e, i, n, g, , u, s, e, d, ', )... 2022-03-18 12:45:59,425 | Translated template is: AWSTemplateFormatVersion: '2010-09-09' Description: SAM API template Parameters: AwsRegion: Type: String Description: AWS Region being used Default: us-east-1 ... /Users/dave.kennedy/sam-api/template.yaml is a valid SAM Template 2022-03-18 12:45:59 There are unpublished metrics. Please make sure you call publish after you record all metrics. 2022-03-18 12:45:59 There are unpublished metrics. Please make sure you call publish after you record all metrics. ``` I can't disclose the entire template, hope that isn't a problem.
True
Validation error with more than 50 parameters - I didn't find this issue in the backlog. ### Description: My template has 53 parameters. `sam validate` throws this error: ``` % sam validate 2022-03-18 12:13:00 Loading policies from IAM... 2022-03-18 12:13:03 Finished loading policies from IAM. 2022-03-18 12:13:03 Template schema validation reported the following errors: [, P, a, r, a, m, e, t, e, r, s, ], , O, r, d, e, r, e, d, D, i, c, t, (, [, (, ', A, w, s, R, e, g, i, o, n, ', ,, , O, r, d, e, r, e, d, D, i, c, t, (, [, (, ', T, y, p, e, ', ,, , ', S, t, r, i, n, g, ', ), ,, , (, ', D, e, s, c, r, i, p, t, i, o, n, ', ,, , ', A, W, S, , R, e, g, i, o, n, , b, e, i, n, g, , u, s, e, d, ', ), ,, , (, ', D, e, f, a, u, l, t, ', ,, , ', u, s, -, e, a, s, t, -, 1, ', ), ], ), )... ``` I know the spaces in the error message are due to #3493 (still waiting on aws/serverless-application-model#2245). That's not the issue. This is what it looks like without the spaces (and with added line breaks): ``` [Parameters] OrderedDict( [ ( 'AwsRegion', OrderedDict( [ ('Type', 'String'), ('Description', 'AWS Region being used'), ('Default', 'us-east-1') ] ) ), [a bunch of other properties here...], ] ) has too many properties ``` The error at the bottom is the relevant part. Removing any three of the parameters makes this error go away. ### Steps to reproduce: 1. Run `sam validate` on a template with more than 50 parameters. ### Observed result: Validation error: `[Parameters] OrderedDict... has too many properties` ### Expected result: CloudFormation supposedly supports up to 200 parameters, so it should validate. ### Additional environment details (Ex: Windows, Mac, Amazon Linux etc) 1. OS: MacOS 11.6.4 2. `sam --version`: 1.40.1 3. AWS region: us-east-1 `Add --debug flag to command you are running` ``` % sam validate --debug 2022-03-18 12:45:56,928 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics 2022-03-18 12:45:56,928 | Using config file: samconfig.toml, config environment: default 2022-03-18 12:45:56,928 | Expand command line arguments to: 2022-03-18 12:45:56,928 | --template_file=/Users/dave.kennedy/sam-api/template.yaml 2022-03-18 12:45:57 Loading policies from IAM... 2022-03-18 12:45:59 Finished loading policies from IAM. 2022-03-18 12:45:59 Template schema validation reported the following errors: [, P, a, r, a, m, e, t, e, r, s, ], , O, r, d, e, r, e, d, D, i, c, t, (, [, (, ', A, w, s, R, e, g, i, o, n, ', ,, , O, r, d, e, r, e, d, D, i, c, t, (, [, (, ', T, y, p, e, ', ,, , ', S, t, r, i, n, g, ', ), ,, , (, ', D, e, s, c, r, i, p, t, i, o, n, ', ,, , ', A, W, S, , R, e, g, i, o, n, , b, e, i, n, g, , u, s, e, d, ', )... 2022-03-18 12:45:59,425 | Translated template is: AWSTemplateFormatVersion: '2010-09-09' Description: SAM API template Parameters: AwsRegion: Type: String Description: AWS Region being used Default: us-east-1 ... /Users/dave.kennedy/sam-api/template.yaml is a valid SAM Template 2022-03-18 12:45:59 There are unpublished metrics. Please make sure you call publish after you record all metrics. 2022-03-18 12:45:59 There are unpublished metrics. Please make sure you call publish after you record all metrics. ``` I can't disclose the entire template, hope that isn't a problem.
main
validation error with more than parameters i didn t find this issue in the backlog description my template has parameters sam validate throws this error sam validate loading policies from iam finished loading policies from iam template schema validation reported the following errors o r d e r e d d i c t i know the spaces in the error message are due to still waiting on aws serverless application model that s not the issue this is what it looks like without the spaces and with added line breaks ordereddict awsregion ordereddict type string description aws region being used default us east has too many properties the error at the bottom is the relevant part removing any three of the parameters makes this error go away steps to reproduce run sam validate on a template with more than parameters observed result validation error ordereddict has too many properties expected result cloudformation supposedly supports up to parameters so it should validate additional environment details ex windows mac amazon linux etc os macos sam version aws region us east add debug flag to command you are running sam validate debug telemetry endpoint configured to be using config file samconfig toml config environment default expand command line arguments to template file users dave kennedy sam api template yaml loading policies from iam finished loading policies from iam template schema validation reported the following errors o r d e r e d d i c t a w s r e g i o n o r d e r e d d i c t t y p e s t r i n g d e s c r i p t i o n a w s r e g i o n b e i n g u s e d translated template is awstemplateformatversion description sam api template parameters awsregion type string description aws region being used default us east users dave kennedy sam api template yaml is a valid sam template there are unpublished metrics please make sure you call publish after you record all metrics there are unpublished metrics please make sure you call publish after you record all metrics i can t disclose the entire template hope that isn t a problem
1
5,866
31,821,151,970
IssuesEvent
2023-09-14 02:28:59
halcyonproject/.github
https://api.github.com/repos/halcyonproject/.github
closed
New maintainer
maintainer-req approved
### Device To maintain Mi 11X Pro (haydn) ### Device Tree Sources Device Tree: https://github.com/DigiGoon/android_device_xiaomi_haydn Common Device Tree: https://github.com/DigiGoon/android_device_xiaomi_sm8350-common Kernel Source: https://github.com/DigiGoon/android_kernel_xiaomi_sm8350 Device Vendor Tree: https://gitlab.com/DigiGoon/proprietary_vendor_xiaomi_haydn Common Vendor Tree: https://gitlab.com/DigiGoon/proprietary_vendor_xiaomi_sm8350-common ### Number of ROM's you maintain officially 0 ### Any Exceptions/special concern? No ### Telegram username DigiGoon ### GitHub username DigiGoon ### Logcat [haydn_halcyon.log](https://github.com/halcyonproject/.github/files/12598015/haydn_halcyon.log)
True
New maintainer - ### Device To maintain Mi 11X Pro (haydn) ### Device Tree Sources Device Tree: https://github.com/DigiGoon/android_device_xiaomi_haydn Common Device Tree: https://github.com/DigiGoon/android_device_xiaomi_sm8350-common Kernel Source: https://github.com/DigiGoon/android_kernel_xiaomi_sm8350 Device Vendor Tree: https://gitlab.com/DigiGoon/proprietary_vendor_xiaomi_haydn Common Vendor Tree: https://gitlab.com/DigiGoon/proprietary_vendor_xiaomi_sm8350-common ### Number of ROM's you maintain officially 0 ### Any Exceptions/special concern? No ### Telegram username DigiGoon ### GitHub username DigiGoon ### Logcat [haydn_halcyon.log](https://github.com/halcyonproject/.github/files/12598015/haydn_halcyon.log)
main
new maintainer device to maintain mi pro haydn device tree sources device tree common device tree kernel source device vendor tree common vendor tree number of rom s you maintain officially any exceptions special concern no telegram username digigoon github username digigoon logcat
1
2,405
8,528,812,834
IssuesEvent
2018-11-03 03:51:33
TabbycatDebate/tabbycat
https://api.github.com/repos/TabbycatDebate/tabbycat
opened
CSV api for adjudicators
awaiting maintainer feature wudc2019
To support #910 there should be a list of basic adj information in terms of name, institution etc. This would allow the feedback list to refer to primary keys.
True
CSV api for adjudicators - To support #910 there should be a list of basic adj information in terms of name, institution etc. This would allow the feedback list to refer to primary keys.
main
csv api for adjudicators to support there should be a list of basic adj information in terms of name institution etc this would allow the feedback list to refer to primary keys
1
4,070
19,180,659,087
IssuesEvent
2021-12-04 10:28:15
RalfKoban/MiKo-Analyzers
https://api.github.com/repos/RalfKoban/MiKo-Analyzers
closed
Do not use magic numbers on assignments
feature Area: analyzer Area: maintainability feasability unclear
Often, numbers are assigned and a comment is made why the number is as it is. Maybe the same number is assigned at different places which will make it error prone to adjust all of them. It's better to have a constant that has a proper name. So you can avoid the comment (that would describe the value of the number) as well. Note: For tests it's ok to use numbers as these are probably specific to the test.
True
Do not use magic numbers on assignments - Often, numbers are assigned and a comment is made why the number is as it is. Maybe the same number is assigned at different places which will make it error prone to adjust all of them. It's better to have a constant that has a proper name. So you can avoid the comment (that would describe the value of the number) as well. Note: For tests it's ok to use numbers as these are probably specific to the test.
main
do not use magic numbers on assignments often numbers are assigned and a comment is made why the number is as it is maybe the same number is assigned at different places which will make it error prone to adjust all of them it s better to have a constant that has a proper name so you can avoid the comment that would describe the value of the number as well note for tests it s ok to use numbers as these are probably specific to the test
1
1,908
6,577,567,505
IssuesEvent
2017-09-12 01:48:53
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
ec2_vpc remove doesn't work as documented
affects_2.0 aws bug_report cloud docs_report waiting_on_maintainer
##### Issue Type: bug report Module: ec2_vpc ##### Ansible Version: ansible 2.0.0.2 ##### Summary: Error in documentation: http://docs.ansible.com/ansible/ec2_vpc_module.html#examples ##### Steps To Reproduce: Documentation states: ``` # Removal of a VPC by id ec2_vpc: state: absent vpc_id: vpc-aaaaaaa region: us-west-2 ``` However, when run, there is an error: ``` TASK [delete vpc] ************************************************************** fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "missing required arguments: resource_tags"} ``` So it isn't possible to remove a vpc just by stating its id.
True
ec2_vpc remove doesn't work as documented - ##### Issue Type: bug report Module: ec2_vpc ##### Ansible Version: ansible 2.0.0.2 ##### Summary: Error in documentation: http://docs.ansible.com/ansible/ec2_vpc_module.html#examples ##### Steps To Reproduce: Documentation states: ``` # Removal of a VPC by id ec2_vpc: state: absent vpc_id: vpc-aaaaaaa region: us-west-2 ``` However, when run, there is an error: ``` TASK [delete vpc] ************************************************************** fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "msg": "missing required arguments: resource_tags"} ``` So it isn't possible to remove a vpc just by stating its id.
main
vpc remove doesn t work as documented issue type bug report module vpc ansible version ansible summary error in documentation steps to reproduce documentation states removal of a vpc by id vpc state absent vpc id vpc aaaaaaa region us west however when run there is an error task fatal failed changed false failed true msg missing required arguments resource tags so it isn t possible to remove a vpc just by stating its id
1
7,620
2,917,972,416
IssuesEvent
2015-06-24 03:26:06
jasonzhu1991/BillSharer
https://api.github.com/repos/jasonzhu1991/BillSharer
closed
QueryParser unit tests
test
As a developer, I would like the QueryParser class to be fully covered by unit tests, so that I can verify this class is working. QueryParser should resolve a "query" which is a JS object, and call underlying persistence layer or balance calculator to get/update/add/delete data (currently known as BillPersistence) or calculate balances, then return response or throw error.
1.0
QueryParser unit tests - As a developer, I would like the QueryParser class to be fully covered by unit tests, so that I can verify this class is working. QueryParser should resolve a "query" which is a JS object, and call underlying persistence layer or balance calculator to get/update/add/delete data (currently known as BillPersistence) or calculate balances, then return response or throw error.
non_main
queryparser unit tests as a developer i would like the queryparser class to be fully covered by unit tests so that i can verify this class is working queryparser should resolve a query which is a js object and call underlying persistence layer or balance calculator to get update add delete data currently known as billpersistence or calculate balances then return response or throw error
0
3,545
14,006,017,235
IssuesEvent
2020-10-28 19:19:10
tensorflow/models
https://api.github.com/repos/tensorflow/models
closed
How print the prediction probabilities in `eval_image_classifier.py`
stat:awaiting maintainer type:feature
Currently, I followed tutorial in [https://github.com/tensorflow/models/tree/master/slim](https://github.com/tensorflow/models/tree/master/slim) to fine-tune the inception_v3 model on pathological images. Image data was converted and stored as TFRecord. When finished fine-tuning, I run `eval_image_classifier.py` on validation set with the new model. The `eval_image_classifier.py` just print accuracy. However, what I want is to print all probabilities for every class. Can anyone help me address it? Thanks in advance for any advise.
True
How print the prediction probabilities in `eval_image_classifier.py` - Currently, I followed tutorial in [https://github.com/tensorflow/models/tree/master/slim](https://github.com/tensorflow/models/tree/master/slim) to fine-tune the inception_v3 model on pathological images. Image data was converted and stored as TFRecord. When finished fine-tuning, I run `eval_image_classifier.py` on validation set with the new model. The `eval_image_classifier.py` just print accuracy. However, what I want is to print all probabilities for every class. Can anyone help me address it? Thanks in advance for any advise.
main
how print the prediction probabilities in eval image classifier py currently i followed tutorial in to fine tune the inception model on pathological images image data was converted and stored as tfrecord when finished fine tuning i run eval image classifier py on validation set with the new model the eval image classifier py just print accuracy however what i want is to print all probabilities for every class can anyone help me address it thanks in advance for any advise
1
5,126
26,131,650,955
IssuesEvent
2022-12-29 06:14:39
camunda/zeebe
https://api.github.com/repos/camunda/zeebe
closed
In the stream processor, avoid sending no response when adding side-effects
kind/toil scope/broker severity/mid area/maintainability
**Description** A stream processor can implement the following method to add side-effects (e.g. like scheduling a timer). ``` @Override public void processRecord( final TypedRecord<DeploymentRecord> command, final TypedResponseWriter responseWriter, final TypedStreamWriter streamWriter, final Consumer<SideEffectProducer> sideEffect) ``` By default, the response is written as a side-effect. If the side-effects are overridden by using `sideEffects::accept` then no response is sent implicitly. The response needs to be added explicitly. It is easy to fall into this trap and forget to add the response as a side-effect. Making the usage more safe could save us from bugs (like #6918).
True
In the stream processor, avoid sending no response when adding side-effects - **Description** A stream processor can implement the following method to add side-effects (e.g. like scheduling a timer). ``` @Override public void processRecord( final TypedRecord<DeploymentRecord> command, final TypedResponseWriter responseWriter, final TypedStreamWriter streamWriter, final Consumer<SideEffectProducer> sideEffect) ``` By default, the response is written as a side-effect. If the side-effects are overridden by using `sideEffects::accept` then no response is sent implicitly. The response needs to be added explicitly. It is easy to fall into this trap and forget to add the response as a side-effect. Making the usage more safe could save us from bugs (like #6918).
main
in the stream processor avoid sending no response when adding side effects description a stream processor can implement the following method to add side effects e g like scheduling a timer override public void processrecord final typedrecord command final typedresponsewriter responsewriter final typedstreamwriter streamwriter final consumer sideeffect by default the response is written as a side effect if the side effects are overridden by using sideeffects accept then no response is sent implicitly the response needs to be added explicitly it is easy to fall into this trap and forget to add the response as a side effect making the usage more safe could save us from bugs like
1
363,223
25,414,017,390
IssuesEvent
2022-11-22 21:47:39
WordPress/Documentation-Issue-Tracker
https://api.github.com/repos/WordPress/Documentation-Issue-Tracker
closed
[HelpHub] Comments Title block
user documentation new document block editor medium priority 6.1
## New page - [ ] Create a new article for the Comments Title block - [ ] Make sure the URL slug under Permalink is: comments-title-block - [ ] At the top of the page, under the title, add the  ‘Link back to blocks’ reusable block. ## Update for 6.1 - [ ] Try a single icon for title blocks [40596](https://github.com/WordPress/gutenberg/pull/40596) - [ ] Create comments title with simple styling [40419](https://github.com/WordPress/gutenberg/pull/40419) ## General - [ ] Make sure all screenshots are relevant to the latest version - [ ] Make sure videos are up to date - [ ] Add ALT tags for the images - [ ] Make sure the headings are in sentence case - [ ] Convert all reusable blocks to a ‘regular block’. - [ ] Update the changelog at the end of the article
1.0
[HelpHub] Comments Title block - ## New page - [ ] Create a new article for the Comments Title block - [ ] Make sure the URL slug under Permalink is: comments-title-block - [ ] At the top of the page, under the title, add the  ‘Link back to blocks’ reusable block. ## Update for 6.1 - [ ] Try a single icon for title blocks [40596](https://github.com/WordPress/gutenberg/pull/40596) - [ ] Create comments title with simple styling [40419](https://github.com/WordPress/gutenberg/pull/40419) ## General - [ ] Make sure all screenshots are relevant to the latest version - [ ] Make sure videos are up to date - [ ] Add ALT tags for the images - [ ] Make sure the headings are in sentence case - [ ] Convert all reusable blocks to a ‘regular block’. - [ ] Update the changelog at the end of the article
non_main
comments title block new page create a new article for the comments title block make sure the url slug under permalink is comments title block at the top of the page under the title add the  ‘link back to blocks’ reusable block update for try a single icon for title blocks create comments title with simple styling general make sure all screenshots are relevant to the latest version make sure videos are up to date add alt tags for the images make sure the headings are in sentence case convert all reusable blocks to a ‘regular block’ update the changelog at the end of the article
0
1,740
6,574,888,853
IssuesEvent
2017-09-11 14:24:12
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
systemd module docs poorly formatted
affects_2.3 docs_report waiting_on_maintainer
##### ISSUE TYPE - Documentation Report ##### COMPONENT NAME system/systemd.py ##### ANSIBLE VERSION ``` public docs as of 2016-10-24 ``` ##### SUMMARY status in 'return values' is illegible in the docs at: https://docs.ansible.com/ansible/systemd_module.html ##### STEPS TO REPRODUCE Go to page ##### EXPECTED RESULTS Not what I see. ##### ACTUAL RESULTS ![screenshot from 2016-10-24 10-55-46](https://cloud.githubusercontent.com/assets/5208768/19650781/7e0726e0-99d8-11e6-9d34-3074d63e2d89.png)
True
systemd module docs poorly formatted - ##### ISSUE TYPE - Documentation Report ##### COMPONENT NAME system/systemd.py ##### ANSIBLE VERSION ``` public docs as of 2016-10-24 ``` ##### SUMMARY status in 'return values' is illegible in the docs at: https://docs.ansible.com/ansible/systemd_module.html ##### STEPS TO REPRODUCE Go to page ##### EXPECTED RESULTS Not what I see. ##### ACTUAL RESULTS ![screenshot from 2016-10-24 10-55-46](https://cloud.githubusercontent.com/assets/5208768/19650781/7e0726e0-99d8-11e6-9d34-3074d63e2d89.png)
main
systemd module docs poorly formatted issue type documentation report component name system systemd py ansible version public docs as of summary status in return values is illegible in the docs at steps to reproduce go to page expected results not what i see actual results
1
365,397
10,781,160,354
IssuesEvent
2019-11-04 14:25:24
salesagility/SuiteCRM
https://api.github.com/repos/salesagility/SuiteCRM
closed
PHP Warning in 7.3 with SubPanelDynamic.tpl
Bug Fix Proposed Low Priority Resolved: Next Release
<!--- Provide a general summary of the issue in the **Title** above --> <!--- Before you open an issue, please check if a similar issue already exists or has been closed before. ---> <!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. ---> #### Issue <!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug --> Running SuiteCRM with PHP 7.3 we are having a lot warnings in php log: ```php PHP Warning: count(): Parameter must be an array or an object that implements Countable in /home/jose/trabajos/gca-suitecrm7.10x/cache/smarty/templates_c/%%EA^EA9^EA90381D%%SubPanelDynamic.tpl.php on line 67 PHP Stack trace: PHP 1. {main}() /home/jose/trabajos/gca-suitecrm7.10x/index.php:0 PHP 2. SugarApplication->execute() /home/jose/trabajos/gca-suitecrm7.10x/index.php:80 PHP 3. CustomContactsController->execute() /home/jose/trabajos/gca-suitecrm7.10x/include/MVC/SugarApplication.php:113 PHP 4. CustomContactsController->processView() /home/jose/trabajos/gca-suitecrm7.10x/include/MVC/Controller/SugarController.php:375 PHP 5. ViewClassic->process() /home/jose/trabajos/gca-suitecrm7.10x/include/MVC/Controller/SugarController.php:435 PHP 6. ViewClassic->display() /home/jose/trabajos/gca-suitecrm7.10x/custom/include/MVC/View/SugarView.php:108 PHP 7. ViewClassic->includeClassicFile() /home/jose/trabajos/gca-suitecrm7.10x/custom/include/MVC/View/views/view.classic.php:32 PHP 8. include_once() /home/jose/trabajos/gca-suitecrm7.10x/include/MVC/View/SugarView.php:818 PHP 9. NonDBSubPanel->display() /home/jose/trabajos/gca-suitecrm7.10x/custom/modules/Contacts/SubPanelViewer.php:69 PHP 10. NonDBSubPanel->ProcessSubPanelListView() /home/jose/trabajos/gca-suitecrm7.10x/include/SubPanel/SubPanel.php:238 PHP 11. NonDBListViewSubPanel->process_dynamic_listview() /home/jose/trabajos/gca-suitecrm7.10x/include/SubPanel/SubPanel.php:222 PHP 12. Sugar_Smarty->display() /home/jose/trabajos/gca-suitecrm7.10x/include/ListView/ListViewSubPanel.php:162 PHP 13. Sugar_Smarty->fetch() /home/jose/trabajos/gca-suitecrm7.10x/vendor/smarty/smarty/libs/Smarty.class.php:1113 PHP 14. Sugar_Smarty->fetch() /home/jose/trabajos/gca-suitecrm7.10x/include/Sugar_Smarty.php:164 PHP 15. include() /home/jose/trabajos/gca-suitecrm7.10x/vendor/smarty/smarty/libs/Smarty.class.php:1263 ``` AFAIK the problem is in the [line 93 the file themes/SuiteP/include/SubPanel/tpls/SubPanelDynamic.tpl](https://github.com/salesagility/SuiteCRM/blob/master/themes/SuiteP/include/SubPanel/tpls/SubPanelDynamic.tpl#L93) ```smarty 93: {if $ROWS_BUTTONS.$rowHeader|@count gt 0} 94: {sugar_action_menu id="$rowHeader" buttons=$ROWS_BUTTONS.$rowHeader class="clickMenu subpanel records fancymenu button" flat=false} 95: {/if} ``` That piece of Smarty code is 'transformed' into a php file in `cache/smarty/templates_c` ```php <?php if (count($this->_tpl_vars['ROWS_BUTTONS'][$this->_tpl_vars['rowHeader']]) > 0): ?> ``` Sometimes `$this->_tpl_vars['ROWS_BUTTONS'][$this->_tpl_vars['rowHeader']]` is `null` and if you execute `count()` (with PHP 7.3) in a non `countable` object (like null) the Warning is raised #### Expected Behavior <!--- Tell us what should happen --> No Warnings in php log while running SuiteCRM with 7.3 #### Actual Behavior <!--- Tell us what happens instead --> <!--- Also please check relevant logs (suitecrm.log, php error.log etc.) --> A lot of warnings in PHP log ;-( #### Possible Fix <!--- Not obligatory, but suggest a fix or reason for the bug --> Verify if `$this->_tpl_vars['ROWS_BUTTONS'][$this->_tpl_vars['rowHeader']]` is an array, so we can avoid counting it. #### Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug include code to reproduce, if relevant --> 1. Run SuiteCRM with PHP 7.3 2. Load a DetailView with subpanels 3. Reload a Subpanle 4. Check PHP error log #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: 7.10.20 * Environment name and version (e.g. MySQL, PHP 7): PHP 7.3 * Operating System and version (e.g Ubuntu 16.04): Ubuntu 19.04
1.0
PHP Warning in 7.3 with SubPanelDynamic.tpl - <!--- Provide a general summary of the issue in the **Title** above --> <!--- Before you open an issue, please check if a similar issue already exists or has been closed before. ---> <!--- If you have discovered a security risk please report it by emailing security@suitecrm.com. This will be delivered to the product team who handle security issues. Please don't disclose security bugs publicly until they have been handled by the security team. ---> #### Issue <!--- Provide a more detailed introduction to the issue itself, and why you consider it to be a bug --> Running SuiteCRM with PHP 7.3 we are having a lot warnings in php log: ```php PHP Warning: count(): Parameter must be an array or an object that implements Countable in /home/jose/trabajos/gca-suitecrm7.10x/cache/smarty/templates_c/%%EA^EA9^EA90381D%%SubPanelDynamic.tpl.php on line 67 PHP Stack trace: PHP 1. {main}() /home/jose/trabajos/gca-suitecrm7.10x/index.php:0 PHP 2. SugarApplication->execute() /home/jose/trabajos/gca-suitecrm7.10x/index.php:80 PHP 3. CustomContactsController->execute() /home/jose/trabajos/gca-suitecrm7.10x/include/MVC/SugarApplication.php:113 PHP 4. CustomContactsController->processView() /home/jose/trabajos/gca-suitecrm7.10x/include/MVC/Controller/SugarController.php:375 PHP 5. ViewClassic->process() /home/jose/trabajos/gca-suitecrm7.10x/include/MVC/Controller/SugarController.php:435 PHP 6. ViewClassic->display() /home/jose/trabajos/gca-suitecrm7.10x/custom/include/MVC/View/SugarView.php:108 PHP 7. ViewClassic->includeClassicFile() /home/jose/trabajos/gca-suitecrm7.10x/custom/include/MVC/View/views/view.classic.php:32 PHP 8. include_once() /home/jose/trabajos/gca-suitecrm7.10x/include/MVC/View/SugarView.php:818 PHP 9. NonDBSubPanel->display() /home/jose/trabajos/gca-suitecrm7.10x/custom/modules/Contacts/SubPanelViewer.php:69 PHP 10. NonDBSubPanel->ProcessSubPanelListView() /home/jose/trabajos/gca-suitecrm7.10x/include/SubPanel/SubPanel.php:238 PHP 11. NonDBListViewSubPanel->process_dynamic_listview() /home/jose/trabajos/gca-suitecrm7.10x/include/SubPanel/SubPanel.php:222 PHP 12. Sugar_Smarty->display() /home/jose/trabajos/gca-suitecrm7.10x/include/ListView/ListViewSubPanel.php:162 PHP 13. Sugar_Smarty->fetch() /home/jose/trabajos/gca-suitecrm7.10x/vendor/smarty/smarty/libs/Smarty.class.php:1113 PHP 14. Sugar_Smarty->fetch() /home/jose/trabajos/gca-suitecrm7.10x/include/Sugar_Smarty.php:164 PHP 15. include() /home/jose/trabajos/gca-suitecrm7.10x/vendor/smarty/smarty/libs/Smarty.class.php:1263 ``` AFAIK the problem is in the [line 93 the file themes/SuiteP/include/SubPanel/tpls/SubPanelDynamic.tpl](https://github.com/salesagility/SuiteCRM/blob/master/themes/SuiteP/include/SubPanel/tpls/SubPanelDynamic.tpl#L93) ```smarty 93: {if $ROWS_BUTTONS.$rowHeader|@count gt 0} 94: {sugar_action_menu id="$rowHeader" buttons=$ROWS_BUTTONS.$rowHeader class="clickMenu subpanel records fancymenu button" flat=false} 95: {/if} ``` That piece of Smarty code is 'transformed' into a php file in `cache/smarty/templates_c` ```php <?php if (count($this->_tpl_vars['ROWS_BUTTONS'][$this->_tpl_vars['rowHeader']]) > 0): ?> ``` Sometimes `$this->_tpl_vars['ROWS_BUTTONS'][$this->_tpl_vars['rowHeader']]` is `null` and if you execute `count()` (with PHP 7.3) in a non `countable` object (like null) the Warning is raised #### Expected Behavior <!--- Tell us what should happen --> No Warnings in php log while running SuiteCRM with 7.3 #### Actual Behavior <!--- Tell us what happens instead --> <!--- Also please check relevant logs (suitecrm.log, php error.log etc.) --> A lot of warnings in PHP log ;-( #### Possible Fix <!--- Not obligatory, but suggest a fix or reason for the bug --> Verify if `$this->_tpl_vars['ROWS_BUTTONS'][$this->_tpl_vars['rowHeader']]` is an array, so we can avoid counting it. #### Steps to Reproduce <!--- Provide a link to a live example, or an unambiguous set of steps to --> <!--- reproduce this bug include code to reproduce, if relevant --> 1. Run SuiteCRM with PHP 7.3 2. Load a DetailView with subpanels 3. Reload a Subpanle 4. Check PHP error log #### Your Environment <!--- Include as many relevant details about the environment you experienced the bug in --> * SuiteCRM Version used: 7.10.20 * Environment name and version (e.g. MySQL, PHP 7): PHP 7.3 * Operating System and version (e.g Ubuntu 16.04): Ubuntu 19.04
non_main
php warning in with subpaneldynamic tpl issue running suitecrm with php we are having a lot warnings in php log php php warning count parameter must be an array or an object that implements countable in home jose trabajos gca cache smarty templates c ea subpaneldynamic tpl php on line php stack trace php main home jose trabajos gca index php php sugarapplication execute home jose trabajos gca index php php customcontactscontroller execute home jose trabajos gca include mvc sugarapplication php php customcontactscontroller processview home jose trabajos gca include mvc controller sugarcontroller php php viewclassic process home jose trabajos gca include mvc controller sugarcontroller php php viewclassic display home jose trabajos gca custom include mvc view sugarview php php viewclassic includeclassicfile home jose trabajos gca custom include mvc view views view classic php php include once home jose trabajos gca include mvc view sugarview php php nondbsubpanel display home jose trabajos gca custom modules contacts subpanelviewer php php nondbsubpanel processsubpanellistview home jose trabajos gca include subpanel subpanel php php nondblistviewsubpanel process dynamic listview home jose trabajos gca include subpanel subpanel php php sugar smarty display home jose trabajos gca include listview listviewsubpanel php php sugar smarty fetch home jose trabajos gca vendor smarty smarty libs smarty class php php sugar smarty fetch home jose trabajos gca include sugar smarty php php include home jose trabajos gca vendor smarty smarty libs smarty class php afaik the problem is in the smarty if rows buttons rowheader count gt sugar action menu id rowheader buttons rows buttons rowheader class clickmenu subpanel records fancymenu button flat false if that piece of smarty code is transformed into a php file in cache smarty templates c php tpl vars sometimes this tpl vars is null and if you execute count with php in a non countable object like null the warning is raised expected behavior no warnings in php log while running suitecrm with actual behavior a lot of warnings in php log possible fix verify if this tpl vars is an array so we can avoid counting it steps to reproduce run suitecrm with php load a detailview with subpanels reload a subpanle check php error log your environment suitecrm version used environment name and version e g mysql php php operating system and version e g ubuntu ubuntu
0
60,962
17,023,567,482
IssuesEvent
2021-07-03 02:41:42
tomhughes/trac-tickets
https://api.github.com/repos/tomhughes/trac-tickets
closed
Loading images suggests adding to deleted layer
Component: merkaartor Priority: major Resolution: fixed Type: defect
**[Submitted to the original trac issue database at 5.54pm, Wednesday, 24th March 2010]** Open one or more GPS tracks. Delete a few of the track layers. Load images - observe that the selection includes all the deleted layers.
1.0
Loading images suggests adding to deleted layer - **[Submitted to the original trac issue database at 5.54pm, Wednesday, 24th March 2010]** Open one or more GPS tracks. Delete a few of the track layers. Load images - observe that the selection includes all the deleted layers.
non_main
loading images suggests adding to deleted layer open one or more gps tracks delete a few of the track layers load images observe that the selection includes all the deleted layers
0
2,273
8,047,754,569
IssuesEvent
2018-08-01 02:37:55
ansible/ansible
https://api.github.com/repos/ansible/ansible
closed
New module request: azure_rm_virtualmachine_facts
affects_2.5 azure cloud easyfix feature module needs_maintainer support:community
##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME azure_rm_virtualmachine_facts ##### ANSIBLE VERSION 2.5 ##### CONFIGURATION ##### OS / ENVIRONMENT any ##### SUMMARY There's no module to query VM facts. Several people already complained about this. New facts module should return facts in 2 formats: (1) raw -- as returned by Azure REST API (2) ansible -- in format accepted by azure_rm_virtualmachine
True
New module request: azure_rm_virtualmachine_facts - ##### ISSUE TYPE - Feature Idea ##### COMPONENT NAME azure_rm_virtualmachine_facts ##### ANSIBLE VERSION 2.5 ##### CONFIGURATION ##### OS / ENVIRONMENT any ##### SUMMARY There's no module to query VM facts. Several people already complained about this. New facts module should return facts in 2 formats: (1) raw -- as returned by Azure REST API (2) ansible -- in format accepted by azure_rm_virtualmachine
main
new module request azure rm virtualmachine facts issue type feature idea component name azure rm virtualmachine facts ansible version configuration os environment any summary there s no module to query vm facts several people already complained about this new facts module should return facts in formats raw as returned by azure rest api ansible in format accepted by azure rm virtualmachine
1
18,826
3,089,087,127
IssuesEvent
2015-08-25 19:47:58
dart-lang/sdk
https://api.github.com/repos/dart-lang/sdk
reopened
Observatory's debugger can't scroll source code if wider than 80 characters
Area-Observatory Type-Defect
If source code is greater than 80 characters wide, I'm unable to scroll to see the remaining characters. ![screen shot 2015-08-03 at 10 22 15 pm](https://cloud.githubusercontent.com/assets/5479/9053748/39069e5a-3a2e-11e5-966b-687507b16697.png)
1.0
Observatory's debugger can't scroll source code if wider than 80 characters - If source code is greater than 80 characters wide, I'm unable to scroll to see the remaining characters. ![screen shot 2015-08-03 at 10 22 15 pm](https://cloud.githubusercontent.com/assets/5479/9053748/39069e5a-3a2e-11e5-966b-687507b16697.png)
non_main
observatory s debugger can t scroll source code if wider than characters if source code is greater than characters wide i m unable to scroll to see the remaining characters
0
1,396
6,025,336,082
IssuesEvent
2017-06-08 08:25:59
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
win_iis_webbinding references property 'sslFlags' which did not exist before IIS 8
affects_2.1 bug_report waiting_on_maintainer windows
<!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> win_iis_webbinding ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.0.0 config file = /home/ubuntu/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> N/A ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> IIS 7.5 on Windows Server 2008R2 ##### SUMMARY <!--- Explain the problem briefly --> Adding bindings to a website in IIS 7.5 (and I suspect earlier versions) throws the following error: "msg": "The property 'sslFlags' cannot be found on this object. Verify that the property exists." This seems to be because win_iis_webbinding.ps1 attempts to access the sslFlags property of the current website before it proceeds with the addition, but this property was only added in IIS 8 (ref: https://www.iis.net/configreference/system.applicationhost/sites/site/bindings/binding) It's worth noting, the binding does get created, despite the error. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: Add additional bindings win_iis_webbinding: name: "{{ site_name }}" state: present protocol: http ip: '*' port: 80 host_header: "{{ item }}" with_items: "{{ site_additional_bindings | default([]) }}" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> ``` TASK [Add additional bindings] ************************************************* ok: [x.x.x.x] ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` TASK [Add additional bindings] ************************************************* failed: [x.x.x.x] (item=testbinding.com) => {"failed": true, "item": "testbinding.com", "msg": "The property 'sslFlags' cannot be found on this object. Verify that the property exists."} ```
True
win_iis_webbinding references property 'sslFlags' which did not exist before IIS 8 - <!--- Verify first that your issue/request is not already reported in GitHub --> ##### ISSUE TYPE <!--- Pick one below and delete the rest: --> - Bug Report ##### COMPONENT NAME <!--- Name of the plugin/module/task --> win_iis_webbinding ##### ANSIBLE VERSION <!--- Paste verbatim output from “ansible --version” between quotes below --> ``` ansible 2.1.0.0 config file = /home/ubuntu/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION <!--- Mention any settings you have changed/added/removed in ansible.cfg (or using the ANSIBLE_* environment variables). --> N/A ##### OS / ENVIRONMENT <!--- Mention the OS you are running Ansible from, and the OS you are managing, or say “N/A” for anything that is not platform-specific. --> IIS 7.5 on Windows Server 2008R2 ##### SUMMARY <!--- Explain the problem briefly --> Adding bindings to a website in IIS 7.5 (and I suspect earlier versions) throws the following error: "msg": "The property 'sslFlags' cannot be found on this object. Verify that the property exists." This seems to be because win_iis_webbinding.ps1 attempts to access the sslFlags property of the current website before it proceeds with the addition, but this property was only added in IIS 8 (ref: https://www.iis.net/configreference/system.applicationhost/sites/site/bindings/binding) It's worth noting, the binding does get created, despite the error. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> <!--- Paste example playbooks or commands between quotes below --> ``` - name: Add additional bindings win_iis_webbinding: name: "{{ site_name }}" state: present protocol: http ip: '*' port: 80 host_header: "{{ item }}" with_items: "{{ site_additional_bindings | default([]) }}" ``` <!--- You can also paste gist.github.com links for larger files --> ##### EXPECTED RESULTS <!--- What did you expect to happen when running the steps above? --> ``` TASK [Add additional bindings] ************************************************* ok: [x.x.x.x] ``` ##### ACTUAL RESULTS <!--- What actually happened? If possible run with high verbosity (-vvvv) --> <!--- Paste verbatim command output between quotes below --> ``` TASK [Add additional bindings] ************************************************* failed: [x.x.x.x] (item=testbinding.com) => {"failed": true, "item": "testbinding.com", "msg": "The property 'sslFlags' cannot be found on this object. Verify that the property exists."} ```
main
win iis webbinding references property sslflags which did not exist before iis issue type bug report component name win iis webbinding ansible version ansible config file home ubuntu ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables n a os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific iis on windows server summary adding bindings to a website in iis and i suspect earlier versions throws the following error msg the property sslflags cannot be found on this object verify that the property exists this seems to be because win iis webbinding attempts to access the sslflags property of the current website before it proceeds with the addition but this property was only added in iis ref it s worth noting the binding does get created despite the error steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name add additional bindings win iis webbinding name site name state present protocol http ip port host header item with items site additional bindings default expected results task ok actual results task failed item testbinding com failed true item testbinding com msg the property sslflags cannot be found on this object verify that the property exists
1
64,951
18,961,041,621
IssuesEvent
2021-11-19 04:56:33
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
Spinner for a pending sync has gone missing
T-Defect P1 X-Cannot-Reproduce X-Regression
I suspect this is what is causing https://github.com/vector-im/riot-web/issues/10513 Steps: 1. Have a large account with saved sync data 2. Send some messages before the app saves sync data again (about every 5min) 3. Refresh the app 4. See the app restore the last saved sync 5. **Expected**: A spinner to show that a `/sync` request is in flight. This used to happen, but went away recently. Last time I remember seeing it was ~4-6 days ago.
1.0
Spinner for a pending sync has gone missing - I suspect this is what is causing https://github.com/vector-im/riot-web/issues/10513 Steps: 1. Have a large account with saved sync data 2. Send some messages before the app saves sync data again (about every 5min) 3. Refresh the app 4. See the app restore the last saved sync 5. **Expected**: A spinner to show that a `/sync` request is in flight. This used to happen, but went away recently. Last time I remember seeing it was ~4-6 days ago.
non_main
spinner for a pending sync has gone missing i suspect this is what is causing steps have a large account with saved sync data send some messages before the app saves sync data again about every refresh the app see the app restore the last saved sync expected a spinner to show that a sync request is in flight this used to happen but went away recently last time i remember seeing it was days ago
0
25,005
12,474,470,329
IssuesEvent
2020-05-29 09:42:38
returntocorp/semgrep
https://api.github.com/repos/returntocorp/semgrep
closed
Semgrep consume all of the CPU, Possible Regex DoS inside the engine.
bug external-user feature:ellipsis lang:javascript performance
**Describe the bug** Certain files and rules cause semgrep to consume all CPU and hangs the process. **To Reproduce** `$ semgrep -f foo.yaml ~/Downloads/juice-shop-master/frontend/src/assets/private/three.js` File: https://github.com/bkimminich/juice-shop/blob/master/frontend/src/assets/private/three.js Rule: foo.yaml ```yaml rules: - id: cpu_kill patterns: - pattern-either: - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.parseString(..., <... $REQ.$QUERY ...>, ...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.foo(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.bar(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.again(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.again(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.blaa(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.fo(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.ell(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.nothing(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.again(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.fance(...); message: CPU gone! languages: - javascript severity: ERROR ``` This issue is present with rules that have ellipsis `...` used to match eveything in between, which makes me think that there is Regex dos happening inside semgrep core. ``` - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.fance(...); ``` **Expected behavior** semgrep should have checks against this or timeout on the file if the analysis is taking more than X time. **Screenshots** <img width="638" alt="Screenshot 2020-05-23 at 5 48 52 PM" src="https://user-images.githubusercontent.com/4301109/82741260-5b9fcb00-9d1e-11ea-91fb-04672f9844ad.png"> **Environment** semgrep 0.8.0 installed from PyPi ``` $ semgrep --version 0.8.0 ```
True
Semgrep consume all of the CPU, Possible Regex DoS inside the engine. - **Describe the bug** Certain files and rules cause semgrep to consume all CPU and hangs the process. **To Reproduce** `$ semgrep -f foo.yaml ~/Downloads/juice-shop-master/frontend/src/assets/private/three.js` File: https://github.com/bkimminich/juice-shop/blob/master/frontend/src/assets/private/three.js Rule: foo.yaml ```yaml rules: - id: cpu_kill patterns: - pattern-either: - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.parseString(..., <... $REQ.$QUERY ...>, ...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.foo(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.bar(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.again(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.again(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.blaa(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.fo(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.ell(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.nothing(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.again(...); - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.fance(...); message: CPU gone! languages: - javascript severity: ERROR ``` This issue is present with rules that have ellipsis `...` used to match eveything in between, which makes me think that there is Regex dos happening inside semgrep core. ``` - pattern: | var $PARSER = new libxmljs.SaxParser(); ... $PARSER.fance(...); ``` **Expected behavior** semgrep should have checks against this or timeout on the file if the analysis is taking more than X time. **Screenshots** <img width="638" alt="Screenshot 2020-05-23 at 5 48 52 PM" src="https://user-images.githubusercontent.com/4301109/82741260-5b9fcb00-9d1e-11ea-91fb-04672f9844ad.png"> **Environment** semgrep 0.8.0 installed from PyPi ``` $ semgrep --version 0.8.0 ```
non_main
semgrep consume all of the cpu possible regex dos inside the engine describe the bug certain files and rules cause semgrep to consume all cpu and hangs the process to reproduce semgrep f foo yaml downloads juice shop master frontend src assets private three js file rule foo yaml yaml rules id cpu kill patterns pattern either pattern var parser new libxmljs saxparser parser parsestring pattern var parser new libxmljs saxparser parser foo pattern var parser new libxmljs saxparser parser bar pattern var parser new libxmljs saxparser parser again pattern var parser new libxmljs saxparser parser again pattern var parser new libxmljs saxparser parser blaa pattern var parser new libxmljs saxparser parser fo pattern var parser new libxmljs saxparser parser ell pattern var parser new libxmljs saxparser parser nothing pattern var parser new libxmljs saxparser parser again pattern var parser new libxmljs saxparser parser fance message cpu gone languages javascript severity error this issue is present with rules that have ellipsis used to match eveything in between which makes me think that there is regex dos happening inside semgrep core pattern var parser new libxmljs saxparser parser fance expected behavior semgrep should have checks against this or timeout on the file if the analysis is taking more than x time screenshots img width alt screenshot at pm src environment semgrep installed from pypi semgrep version
0
153,445
19,706,434,918
IssuesEvent
2022-01-12 22:40:28
KaterinaOrg/maven-modular
https://api.github.com/repos/KaterinaOrg/maven-modular
opened
CVE-2020-36187 (High) detected in jackson-databind-2.9.6.jar
security vulnerability
## CVE-2020-36187 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /module2/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - jackson-module-kotlin-2.9.6.jar (Root Library) - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/KaterinaOrg/maven-modular/commit/5316d1e17d60b08f67a1c0f5526eeffbf1f3103a">5316d1e17d60b08f67a1c0f5526eeffbf1f3103a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36187>CVE-2020-36187</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2997">https://github.com/FasterXML/jackson-databind/issues/2997</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/module2/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.module:jackson-module-kotlin:2.9.6;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36187","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36187","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-36187 (High) detected in jackson-databind-2.9.6.jar - ## CVE-2020-36187 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /module2/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - jackson-module-kotlin-2.9.6.jar (Root Library) - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/KaterinaOrg/maven-modular/commit/5316d1e17d60b08f67a1c0f5526eeffbf1f3103a">5316d1e17d60b08f67a1c0f5526eeffbf1f3103a</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource. <p>Publish Date: 2021-01-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36187>CVE-2020-36187</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2997">https://github.com/FasterXML/jackson-databind/issues/2997</a></p> <p>Release Date: 2021-01-06</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.8</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/module2/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"com.fasterxml.jackson.module:jackson-module-kotlin:2.9.6;com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.8","isBinary":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36187","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.8 mishandles the interaction between serialization gadgets and typing, related to org.apache.tomcat.dbcp.dbcp.datasources.SharedPoolDataSource.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36187","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_main
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy jackson module kotlin jar root library x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources sharedpooldatasource publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree com fasterxml jackson module jackson module kotlin com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind isbinary false basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache tomcat dbcp dbcp datasources sharedpooldatasource vulnerabilityurl
0
159,037
12,452,712,341
IssuesEvent
2020-05-27 12:45:47
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
opened
Click on `I Understand` button should show text for downloading the component instead of a blank spinner
QA/Test-Plan-Specified QA/Yes feature/crypto-wallets suggestion
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Click on `I Understand` button should show text for downloading the component instead of a blank spinner ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile 1.10.74 2. Enable Crypto wallet via hamburger menu 3. Click on `I Understand` 4. Click on `I Understand` button is showing a blank spinner ## Actual result: <!--Please add screenshots if needed--> Click on `I Understand` button is showing a blank spinner ![image](https://user-images.githubusercontent.com/38657976/83019904-5b1c7280-a045-11ea-92ba-73afcdd06fab.png) ## Expected result: Click on `I Understand` button should show text for downloading the component instead of a blank spinner ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easy ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.10.74 Chromium: 81.0.4044.138 (Official Build) dev (64-bit) -- | -- Revision | 8c6c7ba89cc9453625af54f11fd83179e23450fa-refs/branch-heads/4044@{#999} OS | Windows 10 OS Version 1803 (Build 17134.1006) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? No - Can you reproduce this issue with the dev channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/legacy_qa @ryanml @rebron
1.0
Click on `I Understand` button should show text for downloading the component instead of a blank spinner - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Click on `I Understand` button should show text for downloading the component instead of a blank spinner ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Clean profile 1.10.74 2. Enable Crypto wallet via hamburger menu 3. Click on `I Understand` 4. Click on `I Understand` button is showing a blank spinner ## Actual result: <!--Please add screenshots if needed--> Click on `I Understand` button is showing a blank spinner ![image](https://user-images.githubusercontent.com/38657976/83019904-5b1c7280-a045-11ea-92ba-73afcdd06fab.png) ## Expected result: Click on `I Understand` button should show text for downloading the component instead of a blank spinner ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easy ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.10.74 Chromium: 81.0.4044.138 (Official Build) dev (64-bit) -- | -- Revision | 8c6c7ba89cc9453625af54f11fd83179e23450fa-refs/branch-heads/4044@{#999} OS | Windows 10 OS Version 1803 (Build 17134.1006) ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? No - Can you reproduce this issue with the beta channel? No - Can you reproduce this issue with the dev channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/legacy_qa @ryanml @rebron
non_main
click on i understand button should show text for downloading the component instead of a blank spinner have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description click on i understand button should show text for downloading the component instead of a blank spinner steps to reproduce clean profile enable crypto wallet via hamburger menu click on i understand click on i understand button is showing a blank spinner actual result click on i understand button is showing a blank spinner expected result click on i understand button should show text for downloading the component instead of a blank spinner reproduces how often easy brave version brave version info brave chromium   official build  dev  bit revision refs branch heads os windows  os version build version channel information can you reproduce this issue with the current release no can you reproduce this issue with the beta channel no can you reproduce this issue with the dev channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc brave legacy qa ryanml rebron
0
195,439
15,528,682,528
IssuesEvent
2021-03-13 12:03:43
inception-project/inception
https://api.github.com/repos/inception-project/inception
closed
Markdown table broken in developers guide
🐛Bug 📖Documentation
This screenshot is taken is from the current 0.18.3 documentation. ![image](https://user-images.githubusercontent.com/345385/110784580-268f4580-826a-11eb-87d0-19ebc5b8da7e.png)
1.0
Markdown table broken in developers guide - This screenshot is taken is from the current 0.18.3 documentation. ![image](https://user-images.githubusercontent.com/345385/110784580-268f4580-826a-11eb-87d0-19ebc5b8da7e.png)
non_main
markdown table broken in developers guide this screenshot is taken is from the current documentation
0
775
4,381,858,728
IssuesEvent
2016-08-06 14:23:30
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
reopened
Ascii to Binary: handle phrase/sentences
Improvement Maintainer Input Requested Suggestion
What about handling longer input so we can essentially reverse answers like [01001101 01110101 01100011 01101000 01100001 01110011 00100000 01000111 01110010 01100001 01100011 01101001 01100001 01110011 00100000 01000011 01101111 01101101 01110000 01100001 01110011 0100110111110000](https://duckduckgo.com/?q=01001101+01110101+01100011+01101000+01100001+01110011+00100000+01000111+01110010+01100001+01100011+01101001+01100001+01110011+00100000+01000011+01101111+01101101+01110000+01100001+01110011+0100110111110000&ia=answer)? ------ IA Page: http://duck.co/ia/view/binary [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @hunterlang
True
Ascii to Binary: handle phrase/sentences - What about handling longer input so we can essentially reverse answers like [01001101 01110101 01100011 01101000 01100001 01110011 00100000 01000111 01110010 01100001 01100011 01101001 01100001 01110011 00100000 01000011 01101111 01101101 01110000 01100001 01110011 0100110111110000](https://duckduckgo.com/?q=01001101+01110101+01100011+01101000+01100001+01110011+00100000+01000111+01110010+01100001+01100011+01101001+01100001+01110011+00100000+01000011+01101111+01101101+01110000+01100001+01110011+0100110111110000&ia=answer)? ------ IA Page: http://duck.co/ia/view/binary [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @hunterlang
main
ascii to binary handle phrase sentences what about handling longer input so we can essentially reverse answers like ia page hunterlang
1
2,804
10,050,312,045
IssuesEvent
2019-07-20 09:59:25
arcticicestudio/snowsaw
https://api.github.com/repos/arcticicestudio/snowsaw
closed
Handle problems detected by used linters
context-pkg scope-maintainability scope-quality scope-stability type-improvement
The problems in the current code base detected by the linters that have been integrated in #62 through _GolangCI_ should be handled by refactoring the affected implementations. This helps to improve the overall code quality and prevents possible errors. 1. Remove unused function parameters detected by [unparam][]. 1. `(*cmdOptions).prepare` - `cmd` is unused: [cmd/snowsaw/bootstrap/bootstrap.go][cmd/bootstrap-unparam]:51:30 ```go func (o *cmdOptions) prepare(cmd *cobra.Command, args []string) { ^ ``` 2. `(*cmdOptions).run` - `cmd` is unused: [cmd/snowsaw/bootstrap/bootstrap.go][cmd/bootstrap-unparam]:100:26 ```go func (o *cmdOptions) run(cmd *cobra.Command, args []string) { ^ ``` 3. `(*cmdOptions).run` - `args` is unused: [cmd/snowsaw/bootstrap/bootstrap.go][cmd/bootstrap-unparam]:100:46 ```go func (o *cmdOptions) run(cmd *cobra.Command, args []string) { ^ ``` 2. Improve function names and code flows detected by [golint][]. 1. func `NewJsonEncoder` should be `NewJSONEncoder`: [pkg/config/encoder/json/json.go][]:34:6 ```go func NewJsonEncoder() Encoder { ^ ``` 2. var `ExtensionsJson` should be `ExtensionsJSON`: [pkg/config/encoder/constants.go][]:26:2 ```go ExtensionsJson = "json" ^ ``` 3. if block ends with a return statement, so drop this else and outdent its block (move short variable declaration to its own line if necessary): [pkg/prt/printer.go][printer-golint-1]:121:9 ```go } else { ^ ``` 4. exported func Load returns unexported type \*builder.builder, which can be annoying to use: [pkg/config/builder/builder.go][]:39:32 ```go func Load(files ...*file.File) *builder { ^ ``` 3. Improve code style smells detected by [gocritic][]. 1. assignOp: replace `format = format + "\n"` with `format += "\n"`: [pkg/prt/printer.go][printer-gocritic-1]:179:4 ```go format = format + "\n" ^ ``` 2. paramTypeCombine: `func(v Verbosity, w io.Writer, prefix string, format string, args ...interface{})` could be replaced with `func(v Verbosity, w io.Writer, prefix, format string, args ...interface{})`: [pkg/prt/printer.go][printer-gocritic-2]:176:1 ```go func (p *printerConfig) withNewLine(v Verbosity, w io.Writer, prefix string, format string, args ...interface{}) { ^ ``` 3. emptyStringTest: replace `len(parts[0]) == 0` with `parts[0] == ""`: [pkg/snowblock/task/shell/shell.go][]:165:5 ```go if len(parts[0]) == 0 { ^ ``` 4. elseif: can replace 'else {if cond {}}' with 'else if cond {}': [cmd/snowsaw/bootstrap/bootstrap.go][cmd/bootstrap-gocritic]:57:9 ```go } else { ^ ``` 4. Remove unnecessary type conversions detected by [unconvert][]. 1. unnecessary conversion: [pkg/prt/printer.go][printer-unconvert-1]:132:16 ```go *v = Verbosity(l) ^ ``` [cmd/bootstrap-gocritic]: https://github.com/go-critic/go-critic/blob/9366c4a9c6d59dd0fccad12fbc413842ea751fa6/cmd/snowsaw/bootstrap/bootstrap.go#L57 [cmd/bootstrap-unparam]: https://github.com/arcticicestudio/snowsaw/blob/9366c4a9c6d59dd0fccad12fbc413842ea751fa6/cmd/snowsaw/bootstrap/bootstrap.go#L51 [gocritic]: https://github.com/go-critic/go-critic [golint]: https://github.com/golang/lint [pkg/config/builder/builder.go]: https://github.com/arcticicestudio/snowsaw/blob/dea6ab56b7410a8cbc8901818703d5ab1ace5c87/pkg/config/builder/builder.go#L39 [pkg/config/encoder/constants.go]: https://github.com/arcticicestudio/snowsaw/blob/008edbcb509af2cb5ced942d679fa3845a4ec1e1/pkg/config/encoder/constants.go#L26 [pkg/config/encoder/json/json.go]: https://github.com/arcticicestudio/snowsaw/blob/5aa483e7e5e45888254aa4d0143d2afb898b4332/pkg/config/encoder/json/json.go#L34 [pkg/snowblock/task/shell/shell.go]: https://github.com/arcticicestudio/snowsaw/blob/a78810b7ccb5ddb8e80929d54fb7c461a1b80a1c/pkg/snowblock/task/shell/shell.go#L165 [printer-gocritic-1]: https://github.com/arcticicestudio/snowsaw/blob/79afc12ebc15620fd94e78416e0b49a68bbf2eb6/pkg/prt/printer.go#L179 [printer-gocritic-2]: https://github.com/arcticicestudio/snowsaw/blob/79afc12ebc15620fd94e78416e0b49a68bbf2eb6/pkg/prt/printer.go#L176 [printer-golint-1]: https://github.com/arcticicestudio/snowsaw/blob/79afc12ebc15620fd94e78416e0b49a68bbf2eb6/pkg/prt/printer.go#L121 [printer-unconvert-1]: https://github.com/arcticicestudio/snowsaw/blob/79afc12ebc15620fd94e78416e0b49a68bbf2eb6/pkg/prt/printer.go#L132 [unconvert]: https://github.com/mdempsky/unconvert [unparam]: https://github.com/mvdan/unparam
True
Handle problems detected by used linters - The problems in the current code base detected by the linters that have been integrated in #62 through _GolangCI_ should be handled by refactoring the affected implementations. This helps to improve the overall code quality and prevents possible errors. 1. Remove unused function parameters detected by [unparam][]. 1. `(*cmdOptions).prepare` - `cmd` is unused: [cmd/snowsaw/bootstrap/bootstrap.go][cmd/bootstrap-unparam]:51:30 ```go func (o *cmdOptions) prepare(cmd *cobra.Command, args []string) { ^ ``` 2. `(*cmdOptions).run` - `cmd` is unused: [cmd/snowsaw/bootstrap/bootstrap.go][cmd/bootstrap-unparam]:100:26 ```go func (o *cmdOptions) run(cmd *cobra.Command, args []string) { ^ ``` 3. `(*cmdOptions).run` - `args` is unused: [cmd/snowsaw/bootstrap/bootstrap.go][cmd/bootstrap-unparam]:100:46 ```go func (o *cmdOptions) run(cmd *cobra.Command, args []string) { ^ ``` 2. Improve function names and code flows detected by [golint][]. 1. func `NewJsonEncoder` should be `NewJSONEncoder`: [pkg/config/encoder/json/json.go][]:34:6 ```go func NewJsonEncoder() Encoder { ^ ``` 2. var `ExtensionsJson` should be `ExtensionsJSON`: [pkg/config/encoder/constants.go][]:26:2 ```go ExtensionsJson = "json" ^ ``` 3. if block ends with a return statement, so drop this else and outdent its block (move short variable declaration to its own line if necessary): [pkg/prt/printer.go][printer-golint-1]:121:9 ```go } else { ^ ``` 4. exported func Load returns unexported type \*builder.builder, which can be annoying to use: [pkg/config/builder/builder.go][]:39:32 ```go func Load(files ...*file.File) *builder { ^ ``` 3. Improve code style smells detected by [gocritic][]. 1. assignOp: replace `format = format + "\n"` with `format += "\n"`: [pkg/prt/printer.go][printer-gocritic-1]:179:4 ```go format = format + "\n" ^ ``` 2. paramTypeCombine: `func(v Verbosity, w io.Writer, prefix string, format string, args ...interface{})` could be replaced with `func(v Verbosity, w io.Writer, prefix, format string, args ...interface{})`: [pkg/prt/printer.go][printer-gocritic-2]:176:1 ```go func (p *printerConfig) withNewLine(v Verbosity, w io.Writer, prefix string, format string, args ...interface{}) { ^ ``` 3. emptyStringTest: replace `len(parts[0]) == 0` with `parts[0] == ""`: [pkg/snowblock/task/shell/shell.go][]:165:5 ```go if len(parts[0]) == 0 { ^ ``` 4. elseif: can replace 'else {if cond {}}' with 'else if cond {}': [cmd/snowsaw/bootstrap/bootstrap.go][cmd/bootstrap-gocritic]:57:9 ```go } else { ^ ``` 4. Remove unnecessary type conversions detected by [unconvert][]. 1. unnecessary conversion: [pkg/prt/printer.go][printer-unconvert-1]:132:16 ```go *v = Verbosity(l) ^ ``` [cmd/bootstrap-gocritic]: https://github.com/go-critic/go-critic/blob/9366c4a9c6d59dd0fccad12fbc413842ea751fa6/cmd/snowsaw/bootstrap/bootstrap.go#L57 [cmd/bootstrap-unparam]: https://github.com/arcticicestudio/snowsaw/blob/9366c4a9c6d59dd0fccad12fbc413842ea751fa6/cmd/snowsaw/bootstrap/bootstrap.go#L51 [gocritic]: https://github.com/go-critic/go-critic [golint]: https://github.com/golang/lint [pkg/config/builder/builder.go]: https://github.com/arcticicestudio/snowsaw/blob/dea6ab56b7410a8cbc8901818703d5ab1ace5c87/pkg/config/builder/builder.go#L39 [pkg/config/encoder/constants.go]: https://github.com/arcticicestudio/snowsaw/blob/008edbcb509af2cb5ced942d679fa3845a4ec1e1/pkg/config/encoder/constants.go#L26 [pkg/config/encoder/json/json.go]: https://github.com/arcticicestudio/snowsaw/blob/5aa483e7e5e45888254aa4d0143d2afb898b4332/pkg/config/encoder/json/json.go#L34 [pkg/snowblock/task/shell/shell.go]: https://github.com/arcticicestudio/snowsaw/blob/a78810b7ccb5ddb8e80929d54fb7c461a1b80a1c/pkg/snowblock/task/shell/shell.go#L165 [printer-gocritic-1]: https://github.com/arcticicestudio/snowsaw/blob/79afc12ebc15620fd94e78416e0b49a68bbf2eb6/pkg/prt/printer.go#L179 [printer-gocritic-2]: https://github.com/arcticicestudio/snowsaw/blob/79afc12ebc15620fd94e78416e0b49a68bbf2eb6/pkg/prt/printer.go#L176 [printer-golint-1]: https://github.com/arcticicestudio/snowsaw/blob/79afc12ebc15620fd94e78416e0b49a68bbf2eb6/pkg/prt/printer.go#L121 [printer-unconvert-1]: https://github.com/arcticicestudio/snowsaw/blob/79afc12ebc15620fd94e78416e0b49a68bbf2eb6/pkg/prt/printer.go#L132 [unconvert]: https://github.com/mdempsky/unconvert [unparam]: https://github.com/mvdan/unparam
main
handle problems detected by used linters the problems in the current code base detected by the linters that have been integrated in through golangci should be handled by refactoring the affected implementations this helps to improve the overall code quality and prevents possible errors remove unused function parameters detected by cmdoptions prepare cmd is unused go func o cmdoptions prepare cmd cobra command args string cmdoptions run cmd is unused go func o cmdoptions run cmd cobra command args string cmdoptions run args is unused go func o cmdoptions run cmd cobra command args string improve function names and code flows detected by func newjsonencoder should be newjsonencoder go func newjsonencoder encoder var extensionsjson should be extensionsjson go extensionsjson json if block ends with a return statement so drop this else and outdent its block move short variable declaration to its own line if necessary go else exported func load returns unexported type builder builder which can be annoying to use go func load files file file builder improve code style smells detected by assignop replace format format n with format n go format format n paramtypecombine func v verbosity w io writer prefix string format string args interface could be replaced with func v verbosity w io writer prefix format string args interface go func p printerconfig withnewline v verbosity w io writer prefix string format string args interface emptystringtest replace len parts with parts go if len parts elseif can replace else if cond with else if cond go else remove unnecessary type conversions detected by unnecessary conversion go v verbosity l
1
209,610
16,044,924,246
IssuesEvent
2021-04-22 12:37:29
ceph/ceph-csi
https://api.github.com/repos/ceph/ceph-csi
opened
Enable more tests for the k8s-e2e-external-storage job
component/testing
The `ci/centos/k8s-e2e-external-storage` runs the [Kubernetes end-to-end external storage tests](https://github.com/kubernetes/kubernetes/tree/master/test/e2e/storage/external), based on the [different `driver-*.yaml` files](https://github.com/ceph/ceph-csi/tree/devel/scripts/k8s-storage). These driver files describe the capabilities of the provisioner. Currently many features are not enabled in the drivers-test configuration. We ate aiming to have have all features that the provisioners support tested by the Kubernetes external storage tests. Each option that gets enabled should ideally be submitted by its own PR, referencing this Issue.
1.0
Enable more tests for the k8s-e2e-external-storage job - The `ci/centos/k8s-e2e-external-storage` runs the [Kubernetes end-to-end external storage tests](https://github.com/kubernetes/kubernetes/tree/master/test/e2e/storage/external), based on the [different `driver-*.yaml` files](https://github.com/ceph/ceph-csi/tree/devel/scripts/k8s-storage). These driver files describe the capabilities of the provisioner. Currently many features are not enabled in the drivers-test configuration. We ate aiming to have have all features that the provisioners support tested by the Kubernetes external storage tests. Each option that gets enabled should ideally be submitted by its own PR, referencing this Issue.
non_main
enable more tests for the external storage job the ci centos external storage runs the based on the these driver files describe the capabilities of the provisioner currently many features are not enabled in the drivers test configuration we ate aiming to have have all features that the provisioners support tested by the kubernetes external storage tests each option that gets enabled should ideally be submitted by its own pr referencing this issue
0
2,657
9,101,888,394
IssuesEvent
2019-02-20 12:21:39
influxdata/influxdb-python
https://api.github.com/repos/influxdata/influxdb-python
closed
Hosted documentation is out of date
pending maintainer
The docs that are hosted as 'latest' are for 4.1.1 while the latest version is currently 5.0.0 https://influxdb-python.readthedocs.io/en/latest/
True
Hosted documentation is out of date - The docs that are hosted as 'latest' are for 4.1.1 while the latest version is currently 5.0.0 https://influxdb-python.readthedocs.io/en/latest/
main
hosted documentation is out of date the docs that are hosted as latest are for while the latest version is currently
1
419
3,493,537,770
IssuesEvent
2016-01-05 02:59:14
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
#wkhtmltopdf Issues, “brew install Caskroom/cask/wkhtmltopdf” Changes Ownership Permissions
awaiting maintainer feedback
`brew install Caskroom/cask/wkhtmltopdf` Changes Ownership Permissions Installing wkhtmltopdf changes the owner of /usr/local See, https://www.evernote.com/l/AAckZvkLB_NE-rYNfJ5CY4C6gckLkiVHqqo E.g., ``` Warning: /usr/local/bin isn't writable. You should probably change the ownership and permissions of /usr/local/bin
 back to your user account.
sudo chown -R $(whoami) /usr/local/bin ``` Etc, It appears to be a recurring issue, see https://github.com/caskroom/homebrew-cask/issues/12093 I used the above and brew doctor to uninstall and clean up. Save 40 minutes, do not do this install at present.
True
#wkhtmltopdf Issues, “brew install Caskroom/cask/wkhtmltopdf” Changes Ownership Permissions - `brew install Caskroom/cask/wkhtmltopdf` Changes Ownership Permissions Installing wkhtmltopdf changes the owner of /usr/local See, https://www.evernote.com/l/AAckZvkLB_NE-rYNfJ5CY4C6gckLkiVHqqo E.g., ``` Warning: /usr/local/bin isn't writable. You should probably change the ownership and permissions of /usr/local/bin
 back to your user account.
sudo chown -R $(whoami) /usr/local/bin ``` Etc, It appears to be a recurring issue, see https://github.com/caskroom/homebrew-cask/issues/12093 I used the above and brew doctor to uninstall and clean up. Save 40 minutes, do not do this install at present.
main
wkhtmltopdf issues “brew install caskroom cask wkhtmltopdf” changes ownership permissions brew install caskroom cask wkhtmltopdf changes ownership permissions installing wkhtmltopdf changes the owner of usr local see e g warning usr local bin isn t writable you should probably change the ownership and permissions of usr local bin
 back to your user account 
sudo chown r whoami usr local bin etc it appears to be a recurring issue see i used the above and brew doctor to uninstall and clean up save minutes do not do this install at present
1
65,130
12,531,830,285
IssuesEvent
2020-06-04 15:04:17
eclipse-theia/theia
https://api.github.com/repos/eclipse-theia/theia
closed
vscode.window.createTerminal is not fully implemented.
bug help wanted vscode
<!-- Please fill out the following content for a feature request. --> <!-- Please provide a clear description of the feature and any relevant information. --> ### Feature Description: vscode plugin's `window.createTerminal` API support "Creates a Terminal where an extension controls its input and output" by the option [ExtensionTerminalOptions](https://code.visualstudio.com/api/references/vscode-api#ExtensionTerminalOptions) which is not implemented in [theia.window.createTerminal](https://github.com/eclipse-theia/theia/blob/master/packages/plugin/src/theia.d.ts#L3511)
1.0
vscode.window.createTerminal is not fully implemented. - <!-- Please fill out the following content for a feature request. --> <!-- Please provide a clear description of the feature and any relevant information. --> ### Feature Description: vscode plugin's `window.createTerminal` API support "Creates a Terminal where an extension controls its input and output" by the option [ExtensionTerminalOptions](https://code.visualstudio.com/api/references/vscode-api#ExtensionTerminalOptions) which is not implemented in [theia.window.createTerminal](https://github.com/eclipse-theia/theia/blob/master/packages/plugin/src/theia.d.ts#L3511)
non_main
vscode window createterminal is not fully implemented feature description vscode plugin s window createterminal api support creates a terminal where an extension controls its input and output by the option which is not implemented in
0
5,221
26,486,831,171
IssuesEvent
2023-01-17 18:46:25
deislabs/spiderlightning
https://api.github.com/repos/deislabs/spiderlightning
closed
address aws dynamo db test taking too long to run in some Windows builds
🐛 bug 🚧 maintainer issue
**Description of the bug** see title **To Reproduce** n/a **Additional context** Here's a run where it happened: https://dev.azure.com/spiderlightning/slight/_build/results?buildId=94&view=logs&j=8c465dcb-f145-5732-cd9e-733cf3dc7b9c&t=45966b96-b6f1-5d78-ac3a-7f0e40cc01d5
True
address aws dynamo db test taking too long to run in some Windows builds - **Description of the bug** see title **To Reproduce** n/a **Additional context** Here's a run where it happened: https://dev.azure.com/spiderlightning/slight/_build/results?buildId=94&view=logs&j=8c465dcb-f145-5732-cd9e-733cf3dc7b9c&t=45966b96-b6f1-5d78-ac3a-7f0e40cc01d5
main
address aws dynamo db test taking too long to run in some windows builds description of the bug see title to reproduce n a additional context here s a run where it happened
1
542,500
15,861,599,357
IssuesEvent
2021-04-08 10:31:11
HYPERNETS/hypernets_processor
https://api.github.com/repos/HYPERNETS/hypernets_processor
closed
Rename image during L0 reading + add image in filename utils
priority: high
- one image per series,add time, theta phi - add within metadata add list of images - add image naming in the writer
1.0
Rename image during L0 reading + add image in filename utils - - one image per series,add time, theta phi - add within metadata add list of images - add image naming in the writer
non_main
rename image during reading add image in filename utils one image per series add time theta phi add within metadata add list of images add image naming in the writer
0
1,334
5,718,731,703
IssuesEvent
2017-04-19 20:14:25
caskroom/homebrew-cask
https://api.github.com/repos/caskroom/homebrew-cask
closed
virtualbox-extension-pack 5.1.20-114628 will not install
awaiting maintainer feedback
#### General troubleshooting steps - [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue. - [X] None of the templates was appropriate for my issue, or I’m not sure. - [X] I ran `brew update-reset && brew update` and retried my command. - [X] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue It appears that the latest version of VBoxManage required a response to the license question even in non-interactive mode and so it fails to install because the license hasn't been accepted. Appears to be an issue after the latest update #32528. Would fix it, but can't seem to find an accept option. Might need to be an expect script or something? Any ideas? #### Output of your command with `--verbose --debug` ``` $ brew cask install virtualbox-extension-pack --verbose --debug ==> Hbc::Installer#install ==> Printing caveats ==> Hbc::Installer#fetch ==> Satisfying dependencies ==> Installing Cask dependencies: virtualbox virtualbox ... already installed complete ==> Downloading ==> Downloading http://download.virtualbox.org/virtualbox/5.1.20/Oracle_VM_VirtualBox_Extension_Pack-5.1.20-114628.vbox-extpack Already downloaded: /Users/dunnc/Library/Caches/Homebrew/Cask/virtualbox-extension-pack--5.1.20-114628.vbox-extpack ==> Downloaded to -> /Users/dunnc/Library/Caches/Homebrew/Cask/virtualbox-extension-pack--5.1.20-114628.vbox-extpack ==> Verifying download ==> Determining which verifications to run for Cask virtualbox-extension-pack ==> Checking for verification class Hbc::Verify::Checksum ==> 1 verifications defined Hbc::Verify::Checksum ==> Running verification of class Hbc::Verify::Checksum ==> Verifying checksum for Cask virtualbox-extension-pack ==> SHA256 checksums match ==> Hbc::Installer#stage ==> Extracting primary container ==> Determining which containers to use based on 'container :type' ==> Using container class Hbc::Container::Naked for /Users/dunnc/Library/Caches/Homebrew/Cask/virtualbox-extension-pack--5.1.20-114628.vbox-extpack ==> Executing: ["/usr/bin/ditto", "--", "/Users/dunnc/Library/Caches/Homebrew/Cask/virtualbox-extension-pack--5.1.20-114628.vbox-extpack", "/usr/local/Caskroom/virtualbox-extension-pack/5.1.20-114628/Oracle_VM_VirtualBox_Extension_Pack-5.1.20-114628.vbox-extpack"] ==> Creating metadata directory /usr/local/Caskroom/virtualbox-extension-pack/.metadata/5.1.20-114628/20170419044053.974 ==> Creating metadata subdirectory /usr/local/Caskroom/virtualbox-extension-pack/.metadata/5.1.20-114628/20170419044053.974/Casks ==> Installing artifacts ==> Determining which artifacts are present in Cask virtualbox-extension-pack ==> 2 artifact/s defined #<Hbc::Artifact::StageOnly:0x007f9378131218> #<Hbc::Artifact::PostflightBlock:0x007f9378130fe8> ==> Installing artifact of class Hbc::Artifact::PostflightBlock ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/local/bin/VBoxManage", "extpack", "install", "--replace", "/usr/local/Caskroom/virtualbox-extension-pack/5.1.20-114628/Oracle_VM_VirtualBox_Extension_Pack-5.1.20-114628.vbox-extpack"] ==> Purging files for version 5.1.20-114628 of Cask virtualbox-extension-pack Error: Command failed to execute! ==> Failed command: /usr/bin/sudo -E -- /usr/local/bin/VBoxManage extpack install --replace /usr/local/Caskroom/virtualbox-extension-pack/5.1.20-114628/Oracle_VM_VirtualBox_Extension_Pack-5.1.20-114628.vbox-extpack ==> Standard Output of failed command: VirtualBox Personal Use and Evaluation License (PUEL) License version 9, 17 November 2016 ORACLE AMERICA, INC. ("ORACLE") IS WILLING TO LICENSE THE PRODUCT DEFINED IN SECTION 1 BELOW ONLY ON THE CONDITION THAT YOU ACCEPT ALL OF THE TERMS CONTAINED IN THIS VIRTUALBOX PERSONAL USE AND EVALUATION LICENSE AGREEMENT ("AGREEMENT"). IF YOU ARE AGREEING TO THIS LICENSE ON BEHALF OF AN ENTITY (RATHER THAN AS AN INDIVIDUAL HUMAN BEING), YOU REPRESENT THAT YOU HAVE THE APPROPRIATE AUTHORITY TO ACCEPT THESE TERMS AND CONDITIONS ON BEHALF OF SUCH ENTITY. 1 SUBJECT OF AGREEMENT. This Agreement governs your use of the binary software package called "Oracle VM VirtualBox Extension Pack" (the "Product"), which contains a set of Enterprise Features for "Oracle VM VirtualBox" that allows multiple virtual computers ("Guest Computers") to exist on a single physical computer ("Host Computer"). The Product consists of executable files in machine code, data files, and all documentation and updates provided to You by Oracle. 2 GRANT OF LICENSE. Oracle grants you a personal, non-exclusive, non-transferable, limited license without fees to reproduce, install, execute, and use internally the Product on Host Computers for your Personal Use, Educational Use, or Evaluation. "Personal Use" is use solely by the person downloading the Product from Oracle on a single Host Computer, provided that no more than one client or remote computer is connected to that Host Computer and that client or remote computer is used solely to remotely view the Guest Computers. "Educational Use" is any use by teachers or students in an academic institution (schools, colleges and universities) as part of the institutions educational curriculum. "Evaluation" means testing the Product for up to thirty (30) days; after expiry of that term, you are no longer permitted to use the Product. All rights not expressly licensed by Oracle are hereby reserved. 3 RESTRICTIONS AND RESERVATION OF RIGHTS. (1) The Product and copies thereof provided to you under this Agreement are copyrighted and licensed, not sold, to you by Oracle. (2) You may not do any of the following: (a) modify any part of the Product, except to the extent allowed in the documentation accompanying the Product; (b) rent, lease, lend or encumber the Product; (c) remove or alter any proprietary legends or notices contained in the Product; or (d) decompile, or reverse engineer the Product (except to the extent permitted by applicable law). (3) The Product is not designed, licensed or intended for use in the design, construction, operation or maintenance of any nuclear facility and Oracle and its licensors disclaim any express or implied warranty of fitness for such uses. (4) No right, title or interest in or to any trademark, service mark, logo or trade name of Oracle or its licensors is granted under this Agreement. 4 TERMINATION. The Agreement is effective on the date you receive the Product and remains effective until terminated. Your rights under this Agreement will terminate immediately without notice from Oracle if you materially breach it or take any action in derogation of Oracle's and/or its licensors' rights to Product. Oracle may terminate this Agreement should any part of the Product become or in Oracle's reasonable opinion likely to become the subject of a claim of intellectual property infringement or trade secret misappropriation. Upon termination, you will cease use of and destroy all copies of the Product under your control and confirm compliance in writing to Oracle. Sections 3-9, inclusive, will survive termination of the Agreement. 5 DISCLAIMER OF WARRANTY. TO THE EXTENT NOT PROHIBITED BY APPLICABLE LAW, ORACLE PROVIDES THE PRODUCT "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT, EXCEPT TO THE EXTENT THAT THESE DISCLAIMERS ARE HELD TO BE LEGALLY INVALID. The entire risk as to the quality and performance of the Product is with you. Should it prove defective, you assume the cost of all necessary servicing, repair, or correction. In addition, Oracle shall be allowed to provide updates to the Product in urgent cases. You are then obliged to install such updates. Such an urgent case includes, but is not limited to, a claim of rights to the Product by a third party. 6 LIMITATION OF LIABILITY. TO THE EXTENT NOT PROHIBITED BY APPLICABLE LAW, IN NO EVENT WILL ORACLE OR ITS LICENSORS BE LIABLE FOR ANY LOST REVENUE, PROFIT OR DATA, OR FOR SPECIAL, INDIRECT, CONSEQUENTIAL, INCIDENTAL OR PUNITIVE DAMAGES, HOWEVER CAUSED REGARDLESS OF THE THEORY OF LIABILITY, ARISING OUT OF OR RELATED TO THE USE OF OR INABILITY TO USE PRODUCT, EVEN IF ORACLE HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. In no event will Oracle's liability to you, whether in contract, tort (including negligence), or otherwise, exceed the amount paid by you for Product under this Agreement. Some states do not allow the exclusion of incidental or consequential damages, so some of the terms above may not be applicable to you. 7 THIRD PARTY CODE. Portions of Product may be provided with notices and open source licenses from communities and third parties that govern the use of those portions, and any licenses granted hereunder do not alter any rights and obligations You may have under such open source licenses, however, the disclaimer of warranty and limitation of liability provisions in this Agreement will apply to all of the Product. 8 EXPORT REGULATIONS. All Product, documents, technical data, and any other materials delivered under this Agreement are subject to U.S. export control laws and may be subject to export or import regulations in other countries. You agree to comply strictly with these laws and regulations and acknowledge that you have the responsibility to obtain any licenses to export, re-export, or import as may be required after delivery to you. 9 U.S. GOVERNMENT END USERS. If the Product or related documentation is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice applies: Oracle programs, including any operating system, integrated software, any programs installed on hardware, and/or documentation, delivered to U.S. Government end users are "commercial computer software" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, use, duplication, disclosure, modification, and adaptation of the programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, shall be subject to license terms and license restrictions applicable to the programs. No other rights are granted to the U.S. Government. 10 MISCELLANEOUS. This Agreement is the entire agreement between you and Oracle relating to its subject matter. It supersedes all prior or contemporaneous oral or written communications, proposals, representations and warranties and prevails over any conflicting or additional terms of any quote, order, acknowledgment, or other communication between the parties relating to its subject matter during the term of this Agreement. No modification of this Agreement will be binding, unless in writing and signed by an authorized representative of each party. If any provision of this Agreement is held to be unenforceable, this Agreement will remain in effect with the provision omitted, unless omission would frustrate the intent of the parties, in which case this Agreement will immediately terminate. Course of dealing and other standard business conditions of the parties or the industry shall not apply. This Agreement is governed by the substantive and procedural laws of the State of California, USA, and you and Oracle agree to submit to the exclusive jurisdiction of, and venue in, the courts of San Francisco or Santa Clara counties in California in any dispute arising out of or relating to this Agreement. Do you agree to these license terms and conditions (y/n)? Installation of "Oracle VM VirtualBox Extension Pack" aborted. ==> Standard Error of failed command: ==> Exit status of failed command: #<Process::Status: pid 45582 exit 1> Error: nothing to install/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/install.rb:13:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:115:in `run_command' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:158:in `process' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:91:in `<main>' Error: Kernel.exit /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:163:in `exit' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:163:in `rescue in process' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:149:in `process' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:91:in `<main>' ``` #### Output of `brew cask doctor` ``` $ brew doctor Your system is ready to brew. ```
True
virtualbox-extension-pack 5.1.20-114628 will not install - #### General troubleshooting steps - [X] I have checked the instructions for [reporting bugs](https://github.com/caskroom/homebrew-cask#reporting-bugs) (or [making requests](https://github.com/caskroom/homebrew-cask#requests)) before opening the issue. - [X] None of the templates was appropriate for my issue, or I’m not sure. - [X] I ran `brew update-reset && brew update` and retried my command. - [X] I ran `brew doctor`, fixed as many issues as possible and retried my command. - [X] I understand that [if I ignore these instructions, my issue may be closed without review](https://github.com/caskroom/homebrew-cask/blob/master/doc/faq/closing_issues_without_review.md). #### Description of issue It appears that the latest version of VBoxManage required a response to the license question even in non-interactive mode and so it fails to install because the license hasn't been accepted. Appears to be an issue after the latest update #32528. Would fix it, but can't seem to find an accept option. Might need to be an expect script or something? Any ideas? #### Output of your command with `--verbose --debug` ``` $ brew cask install virtualbox-extension-pack --verbose --debug ==> Hbc::Installer#install ==> Printing caveats ==> Hbc::Installer#fetch ==> Satisfying dependencies ==> Installing Cask dependencies: virtualbox virtualbox ... already installed complete ==> Downloading ==> Downloading http://download.virtualbox.org/virtualbox/5.1.20/Oracle_VM_VirtualBox_Extension_Pack-5.1.20-114628.vbox-extpack Already downloaded: /Users/dunnc/Library/Caches/Homebrew/Cask/virtualbox-extension-pack--5.1.20-114628.vbox-extpack ==> Downloaded to -> /Users/dunnc/Library/Caches/Homebrew/Cask/virtualbox-extension-pack--5.1.20-114628.vbox-extpack ==> Verifying download ==> Determining which verifications to run for Cask virtualbox-extension-pack ==> Checking for verification class Hbc::Verify::Checksum ==> 1 verifications defined Hbc::Verify::Checksum ==> Running verification of class Hbc::Verify::Checksum ==> Verifying checksum for Cask virtualbox-extension-pack ==> SHA256 checksums match ==> Hbc::Installer#stage ==> Extracting primary container ==> Determining which containers to use based on 'container :type' ==> Using container class Hbc::Container::Naked for /Users/dunnc/Library/Caches/Homebrew/Cask/virtualbox-extension-pack--5.1.20-114628.vbox-extpack ==> Executing: ["/usr/bin/ditto", "--", "/Users/dunnc/Library/Caches/Homebrew/Cask/virtualbox-extension-pack--5.1.20-114628.vbox-extpack", "/usr/local/Caskroom/virtualbox-extension-pack/5.1.20-114628/Oracle_VM_VirtualBox_Extension_Pack-5.1.20-114628.vbox-extpack"] ==> Creating metadata directory /usr/local/Caskroom/virtualbox-extension-pack/.metadata/5.1.20-114628/20170419044053.974 ==> Creating metadata subdirectory /usr/local/Caskroom/virtualbox-extension-pack/.metadata/5.1.20-114628/20170419044053.974/Casks ==> Installing artifacts ==> Determining which artifacts are present in Cask virtualbox-extension-pack ==> 2 artifact/s defined #<Hbc::Artifact::StageOnly:0x007f9378131218> #<Hbc::Artifact::PostflightBlock:0x007f9378130fe8> ==> Installing artifact of class Hbc::Artifact::PostflightBlock ==> Executing: ["/usr/bin/sudo", "-E", "--", "/usr/local/bin/VBoxManage", "extpack", "install", "--replace", "/usr/local/Caskroom/virtualbox-extension-pack/5.1.20-114628/Oracle_VM_VirtualBox_Extension_Pack-5.1.20-114628.vbox-extpack"] ==> Purging files for version 5.1.20-114628 of Cask virtualbox-extension-pack Error: Command failed to execute! ==> Failed command: /usr/bin/sudo -E -- /usr/local/bin/VBoxManage extpack install --replace /usr/local/Caskroom/virtualbox-extension-pack/5.1.20-114628/Oracle_VM_VirtualBox_Extension_Pack-5.1.20-114628.vbox-extpack ==> Standard Output of failed command: VirtualBox Personal Use and Evaluation License (PUEL) License version 9, 17 November 2016 ORACLE AMERICA, INC. ("ORACLE") IS WILLING TO LICENSE THE PRODUCT DEFINED IN SECTION 1 BELOW ONLY ON THE CONDITION THAT YOU ACCEPT ALL OF THE TERMS CONTAINED IN THIS VIRTUALBOX PERSONAL USE AND EVALUATION LICENSE AGREEMENT ("AGREEMENT"). IF YOU ARE AGREEING TO THIS LICENSE ON BEHALF OF AN ENTITY (RATHER THAN AS AN INDIVIDUAL HUMAN BEING), YOU REPRESENT THAT YOU HAVE THE APPROPRIATE AUTHORITY TO ACCEPT THESE TERMS AND CONDITIONS ON BEHALF OF SUCH ENTITY. 1 SUBJECT OF AGREEMENT. This Agreement governs your use of the binary software package called "Oracle VM VirtualBox Extension Pack" (the "Product"), which contains a set of Enterprise Features for "Oracle VM VirtualBox" that allows multiple virtual computers ("Guest Computers") to exist on a single physical computer ("Host Computer"). The Product consists of executable files in machine code, data files, and all documentation and updates provided to You by Oracle. 2 GRANT OF LICENSE. Oracle grants you a personal, non-exclusive, non-transferable, limited license without fees to reproduce, install, execute, and use internally the Product on Host Computers for your Personal Use, Educational Use, or Evaluation. "Personal Use" is use solely by the person downloading the Product from Oracle on a single Host Computer, provided that no more than one client or remote computer is connected to that Host Computer and that client or remote computer is used solely to remotely view the Guest Computers. "Educational Use" is any use by teachers or students in an academic institution (schools, colleges and universities) as part of the institutions educational curriculum. "Evaluation" means testing the Product for up to thirty (30) days; after expiry of that term, you are no longer permitted to use the Product. All rights not expressly licensed by Oracle are hereby reserved. 3 RESTRICTIONS AND RESERVATION OF RIGHTS. (1) The Product and copies thereof provided to you under this Agreement are copyrighted and licensed, not sold, to you by Oracle. (2) You may not do any of the following: (a) modify any part of the Product, except to the extent allowed in the documentation accompanying the Product; (b) rent, lease, lend or encumber the Product; (c) remove or alter any proprietary legends or notices contained in the Product; or (d) decompile, or reverse engineer the Product (except to the extent permitted by applicable law). (3) The Product is not designed, licensed or intended for use in the design, construction, operation or maintenance of any nuclear facility and Oracle and its licensors disclaim any express or implied warranty of fitness for such uses. (4) No right, title or interest in or to any trademark, service mark, logo or trade name of Oracle or its licensors is granted under this Agreement. 4 TERMINATION. The Agreement is effective on the date you receive the Product and remains effective until terminated. Your rights under this Agreement will terminate immediately without notice from Oracle if you materially breach it or take any action in derogation of Oracle's and/or its licensors' rights to Product. Oracle may terminate this Agreement should any part of the Product become or in Oracle's reasonable opinion likely to become the subject of a claim of intellectual property infringement or trade secret misappropriation. Upon termination, you will cease use of and destroy all copies of the Product under your control and confirm compliance in writing to Oracle. Sections 3-9, inclusive, will survive termination of the Agreement. 5 DISCLAIMER OF WARRANTY. TO THE EXTENT NOT PROHIBITED BY APPLICABLE LAW, ORACLE PROVIDES THE PRODUCT "AS IS" WITHOUT WARRANTY OF ANY KIND, EITHER EXPRESSED OR IMPLIED, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE OR NON-INFRINGEMENT, EXCEPT TO THE EXTENT THAT THESE DISCLAIMERS ARE HELD TO BE LEGALLY INVALID. The entire risk as to the quality and performance of the Product is with you. Should it prove defective, you assume the cost of all necessary servicing, repair, or correction. In addition, Oracle shall be allowed to provide updates to the Product in urgent cases. You are then obliged to install such updates. Such an urgent case includes, but is not limited to, a claim of rights to the Product by a third party. 6 LIMITATION OF LIABILITY. TO THE EXTENT NOT PROHIBITED BY APPLICABLE LAW, IN NO EVENT WILL ORACLE OR ITS LICENSORS BE LIABLE FOR ANY LOST REVENUE, PROFIT OR DATA, OR FOR SPECIAL, INDIRECT, CONSEQUENTIAL, INCIDENTAL OR PUNITIVE DAMAGES, HOWEVER CAUSED REGARDLESS OF THE THEORY OF LIABILITY, ARISING OUT OF OR RELATED TO THE USE OF OR INABILITY TO USE PRODUCT, EVEN IF ORACLE HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. In no event will Oracle's liability to you, whether in contract, tort (including negligence), or otherwise, exceed the amount paid by you for Product under this Agreement. Some states do not allow the exclusion of incidental or consequential damages, so some of the terms above may not be applicable to you. 7 THIRD PARTY CODE. Portions of Product may be provided with notices and open source licenses from communities and third parties that govern the use of those portions, and any licenses granted hereunder do not alter any rights and obligations You may have under such open source licenses, however, the disclaimer of warranty and limitation of liability provisions in this Agreement will apply to all of the Product. 8 EXPORT REGULATIONS. All Product, documents, technical data, and any other materials delivered under this Agreement are subject to U.S. export control laws and may be subject to export or import regulations in other countries. You agree to comply strictly with these laws and regulations and acknowledge that you have the responsibility to obtain any licenses to export, re-export, or import as may be required after delivery to you. 9 U.S. GOVERNMENT END USERS. If the Product or related documentation is delivered to the U.S. Government or anyone licensing it on behalf of the U.S. Government, then the following notice applies: Oracle programs, including any operating system, integrated software, any programs installed on hardware, and/or documentation, delivered to U.S. Government end users are "commercial computer software" pursuant to the applicable Federal Acquisition Regulation and agency-specific supplemental regulations. As such, use, duplication, disclosure, modification, and adaptation of the programs, including any operating system, integrated software, any programs installed on the hardware, and/or documentation, shall be subject to license terms and license restrictions applicable to the programs. No other rights are granted to the U.S. Government. 10 MISCELLANEOUS. This Agreement is the entire agreement between you and Oracle relating to its subject matter. It supersedes all prior or contemporaneous oral or written communications, proposals, representations and warranties and prevails over any conflicting or additional terms of any quote, order, acknowledgment, or other communication between the parties relating to its subject matter during the term of this Agreement. No modification of this Agreement will be binding, unless in writing and signed by an authorized representative of each party. If any provision of this Agreement is held to be unenforceable, this Agreement will remain in effect with the provision omitted, unless omission would frustrate the intent of the parties, in which case this Agreement will immediately terminate. Course of dealing and other standard business conditions of the parties or the industry shall not apply. This Agreement is governed by the substantive and procedural laws of the State of California, USA, and you and Oracle agree to submit to the exclusive jurisdiction of, and venue in, the courts of San Francisco or Santa Clara counties in California in any dispute arising out of or relating to this Agreement. Do you agree to these license terms and conditions (y/n)? Installation of "Oracle VM VirtualBox Extension Pack" aborted. ==> Standard Error of failed command: ==> Exit status of failed command: #<Process::Status: pid 45582 exit 1> Error: nothing to install/usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli/install.rb:13:in `run' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:115:in `run_command' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:158:in `process' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:91:in `<main>' Error: Kernel.exit /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:163:in `exit' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:163:in `rescue in process' /usr/local/Homebrew/Library/Homebrew/cask/lib/hbc/cli.rb:149:in `process' /usr/local/Homebrew/Library/Homebrew/cmd/cask.rb:8:in `cask' /usr/local/Homebrew/Library/Homebrew/brew.rb:91:in `<main>' ``` #### Output of `brew cask doctor` ``` $ brew doctor Your system is ready to brew. ```
main
virtualbox extension pack will not install general troubleshooting steps i have checked the instructions for or before opening the issue none of the templates was appropriate for my issue or i’m not sure i ran brew update reset brew update and retried my command i ran brew doctor fixed as many issues as possible and retried my command i understand that description of issue it appears that the latest version of vboxmanage required a response to the license question even in non interactive mode and so it fails to install because the license hasn t been accepted appears to be an issue after the latest update would fix it but can t seem to find an accept option might need to be an expect script or something any ideas output of your command with verbose debug brew cask install virtualbox extension pack verbose debug hbc installer install printing caveats hbc installer fetch satisfying dependencies installing cask dependencies virtualbox virtualbox already installed complete downloading downloading already downloaded users dunnc library caches homebrew cask virtualbox extension pack vbox extpack downloaded to users dunnc library caches homebrew cask virtualbox extension pack vbox extpack verifying download determining which verifications to run for cask virtualbox extension pack checking for verification class hbc verify checksum verifications defined hbc verify checksum running verification of class hbc verify checksum verifying checksum for cask virtualbox extension pack checksums match hbc installer stage extracting primary container determining which containers to use based on container type using container class hbc container naked for users dunnc library caches homebrew cask virtualbox extension pack vbox extpack executing creating metadata directory usr local caskroom virtualbox extension pack metadata creating metadata subdirectory usr local caskroom virtualbox extension pack metadata casks installing artifacts determining which artifacts are present in cask virtualbox extension pack artifact s defined installing artifact of class hbc artifact postflightblock executing purging files for version of cask virtualbox extension pack error command failed to execute failed command usr bin sudo e usr local bin vboxmanage extpack install replace usr local caskroom virtualbox extension pack oracle vm virtualbox extension pack vbox extpack standard output of failed command virtualbox personal use and evaluation license puel license version november oracle america inc oracle is willing to license the product defined in section below only on the condition that you accept all of the terms contained in this virtualbox personal use and evaluation license agreement agreement if you are agreeing to this license on behalf of an entity rather than as an individual human being you represent that you have the appropriate authority to accept these terms and conditions on behalf of such entity subject of agreement this agreement governs your use of the binary software package called oracle vm virtualbox extension pack the product which contains a set of enterprise features for oracle vm virtualbox that allows multiple virtual computers guest computers to exist on a single physical computer host computer the product consists of executable files in machine code data files and all documentation and updates provided to you by oracle grant of license oracle grants you a personal non exclusive non transferable limited license without fees to reproduce install execute and use internally the product on host computers for your personal use educational use or evaluation personal use is use solely by the person downloading the product from oracle on a single host computer provided that no more than one client or remote computer is connected to that host computer and that client or remote computer is used solely to remotely view the guest computers educational use is any use by teachers or students in an academic institution schools colleges and universities as part of the institutions educational curriculum evaluation means testing the product for up to thirty days after expiry of that term you are no longer permitted to use the product all rights not expressly licensed by oracle are hereby reserved restrictions and reservation of rights the product and copies thereof provided to you under this agreement are copyrighted and licensed not sold to you by oracle you may not do any of the following a modify any part of the product except to the extent allowed in the documentation accompanying the product b rent lease lend or encumber the product c remove or alter any proprietary legends or notices contained in the product or d decompile or reverse engineer the product except to the extent permitted by applicable law the product is not designed licensed or intended for use in the design construction operation or maintenance of any nuclear facility and oracle and its licensors disclaim any express or implied warranty of fitness for such uses no right title or interest in or to any trademark service mark logo or trade name of oracle or its licensors is granted under this agreement termination the agreement is effective on the date you receive the product and remains effective until terminated your rights under this agreement will terminate immediately without notice from oracle if you materially breach it or take any action in derogation of oracle s and or its licensors rights to product oracle may terminate this agreement should any part of the product become or in oracle s reasonable opinion likely to become the subject of a claim of intellectual property infringement or trade secret misappropriation upon termination you will cease use of and destroy all copies of the product under your control and confirm compliance in writing to oracle sections inclusive will survive termination of the agreement disclaimer of warranty to the extent not prohibited by applicable law oracle provides the product as is without warranty of any kind either expressed or implied including but not limited to the implied warranties of merchantability fitness for a particular purpose or non infringement except to the extent that these disclaimers are held to be legally invalid the entire risk as to the quality and performance of the product is with you should it prove defective you assume the cost of all necessary servicing repair or correction in addition oracle shall be allowed to provide updates to the product in urgent cases you are then obliged to install such updates such an urgent case includes but is not limited to a claim of rights to the product by a third party limitation of liability to the extent not prohibited by applicable law in no event will oracle or its licensors be liable for any lost revenue profit or data or for special indirect consequential incidental or punitive damages however caused regardless of the theory of liability arising out of or related to the use of or inability to use product even if oracle has been advised of the possibility of such damages in no event will oracle s liability to you whether in contract tort including negligence or otherwise exceed the amount paid by you for product under this agreement some states do not allow the exclusion of incidental or consequential damages so some of the terms above may not be applicable to you third party code portions of product may be provided with notices and open source licenses from communities and third parties that govern the use of those portions and any licenses granted hereunder do not alter any rights and obligations you may have under such open source licenses however the disclaimer of warranty and limitation of liability provisions in this agreement will apply to all of the product export regulations all product documents technical data and any other materials delivered under this agreement are subject to u s export control laws and may be subject to export or import regulations in other countries you agree to comply strictly with these laws and regulations and acknowledge that you have the responsibility to obtain any licenses to export re export or import as may be required after delivery to you u s government end users if the product or related documentation is delivered to the u s government or anyone licensing it on behalf of the u s government then the following notice applies oracle programs including any operating system integrated software any programs installed on hardware and or documentation delivered to u s government end users are commercial computer software pursuant to the applicable federal acquisition regulation and agency specific supplemental regulations as such use duplication disclosure modification and adaptation of the programs including any operating system integrated software any programs installed on the hardware and or documentation shall be subject to license terms and license restrictions applicable to the programs no other rights are granted to the u s government miscellaneous this agreement is the entire agreement between you and oracle relating to its subject matter it supersedes all prior or contemporaneous oral or written communications proposals representations and warranties and prevails over any conflicting or additional terms of any quote order acknowledgment or other communication between the parties relating to its subject matter during the term of this agreement no modification of this agreement will be binding unless in writing and signed by an authorized representative of each party if any provision of this agreement is held to be unenforceable this agreement will remain in effect with the provision omitted unless omission would frustrate the intent of the parties in which case this agreement will immediately terminate course of dealing and other standard business conditions of the parties or the industry shall not apply this agreement is governed by the substantive and procedural laws of the state of california usa and you and oracle agree to submit to the exclusive jurisdiction of and venue in the courts of san francisco or santa clara counties in california in any dispute arising out of or relating to this agreement do you agree to these license terms and conditions y n installation of oracle vm virtualbox extension pack aborted standard error of failed command exit status of failed command error nothing to install usr local homebrew library homebrew cask lib hbc cli install rb in run usr local homebrew library homebrew cask lib hbc cli rb in run command usr local homebrew library homebrew cask lib hbc cli rb in process usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in error kernel exit usr local homebrew library homebrew cask lib hbc cli rb in exit usr local homebrew library homebrew cask lib hbc cli rb in rescue in process usr local homebrew library homebrew cask lib hbc cli rb in process usr local homebrew library homebrew cmd cask rb in cask usr local homebrew library homebrew brew rb in output of brew cask doctor brew doctor your system is ready to brew
1
716
4,309,112,231
IssuesEvent
2016-07-21 15:03:48
duckduckgo/zeroclickinfo-goodies
https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies
closed
URL Decode: Trigger is capitalization dependent
Bug Maintainer Approved
Does not trigger when trigger word URL is written in caps eg. "url decode <link>" triggers the IA but "URL decode <link>" does not. ------ IA Page: http://duck.co/ia/view/urldecode [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mintsoft
True
URL Decode: Trigger is capitalization dependent - Does not trigger when trigger word URL is written in caps eg. "url decode <link>" triggers the IA but "URL decode <link>" does not. ------ IA Page: http://duck.co/ia/view/urldecode [Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mintsoft
main
url decode trigger is capitalization dependent does not trigger when trigger word url is written in caps eg url decode triggers the ia but url decode does not ia page mintsoft
1
490,223
14,117,003,719
IssuesEvent
2020-11-08 06:52:26
Avi-Kenny/simba
https://api.github.com/repos/Avi-Kenny/simba
opened
Need a system to allow multiple methods to be executed on a single dataset
Priority: mid enhancement
Sometimes, a user wants to generate one dataset and execute, for example, three different methods using the single dataset. A system is needed (possibly via seeds) to manage this process.
1.0
Need a system to allow multiple methods to be executed on a single dataset - Sometimes, a user wants to generate one dataset and execute, for example, three different methods using the single dataset. A system is needed (possibly via seeds) to manage this process.
non_main
need a system to allow multiple methods to be executed on a single dataset sometimes a user wants to generate one dataset and execute for example three different methods using the single dataset a system is needed possibly via seeds to manage this process
0
11,400
3,202,471,790
IssuesEvent
2015-10-02 14:14:11
adminspotter/r9
https://api.github.com/repos/adminspotter/r9
opened
Add tests for the server console
server testing
There are currently no tests at all. We've got another issue ( #5 ) for the main socket objects, so maybe that's a better one to do first. One specific item we need to test is the libwrap functionality. Docu is pretty hazy for that, so I mostly just guessed.
1.0
Add tests for the server console - There are currently no tests at all. We've got another issue ( #5 ) for the main socket objects, so maybe that's a better one to do first. One specific item we need to test is the libwrap functionality. Docu is pretty hazy for that, so I mostly just guessed.
non_main
add tests for the server console there are currently no tests at all we ve got another issue for the main socket objects so maybe that s a better one to do first one specific item we need to test is the libwrap functionality docu is pretty hazy for that so i mostly just guessed
0
253,679
27,300,789,403
IssuesEvent
2023-02-24 01:38:11
panasalap/linux-4.19.72_1
https://api.github.com/repos/panasalap/linux-4.19.72_1
closed
CVE-2020-15436 (Medium) detected in linux-yoctov5.4.51 - autoclosed
security vulnerability
## CVE-2020-15436 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/block_dev.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/block_dev.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Use-after-free vulnerability in fs/block_dev.c in the Linux kernel before 5.8 allows local users to gain privileges or cause a denial of service by leveraging improper access to a certain error field. Mend Note: After conducting further research, Mend has determined that versions v2.6.12-rc2 through v4.4.228, v4.5-rc1 through v4.9.228, v4.10-rc1 through v4.14.185, v4.15-rc1 through v4.19.129, v5.0-rc1 through v5.4.48, v5.5-rc1 through v5.7.5 and v5.8-rc1 of Linux Kernel are vulnerable to CVE-2020-15436. <p>Publish Date: 2020-11-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15436>CVE-2020-15436</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-15436">https://www.linuxkernelcves.com/cves/CVE-2020-15436</a></p> <p>Release Date: 2020-11-23</p> <p>Fix Resolution: v4.4.229, v4.9.229, v4.14.186, v4.19.130, v5.4.49, v5.7.6, v5.8-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-15436 (Medium) detected in linux-yoctov5.4.51 - autoclosed - ## CVE-2020-15436 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary> <p> <p>Yocto Linux Embedded kernel</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p> <p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/c5a08fe8179013aad614165d792bc5b436591df6">c5a08fe8179013aad614165d792bc5b436591df6</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/block_dev.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/fs/block_dev.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Use-after-free vulnerability in fs/block_dev.c in the Linux kernel before 5.8 allows local users to gain privileges or cause a denial of service by leveraging improper access to a certain error field. Mend Note: After conducting further research, Mend has determined that versions v2.6.12-rc2 through v4.4.228, v4.5-rc1 through v4.9.228, v4.10-rc1 through v4.14.185, v4.15-rc1 through v4.19.129, v5.0-rc1 through v5.4.48, v5.5-rc1 through v5.7.5 and v5.8-rc1 of Linux Kernel are vulnerable to CVE-2020-15436. <p>Publish Date: 2020-11-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2020-15436>CVE-2020-15436</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.7</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-15436">https://www.linuxkernelcves.com/cves/CVE-2020-15436</a></p> <p>Release Date: 2020-11-23</p> <p>Fix Resolution: v4.4.229, v4.9.229, v4.14.186, v4.19.130, v5.4.49, v5.7.6, v5.8-rc2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_main
cve medium detected in linux autoclosed cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files fs block dev c fs block dev c vulnerability details use after free vulnerability in fs block dev c in the linux kernel before allows local users to gain privileges or cause a denial of service by leveraging improper access to a certain error field mend note after conducting further research mend has determined that versions through through through through through through and of linux kernel are vulnerable to cve publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
13,794
5,451,853,962
IssuesEvent
2017-03-08 00:39:23
istio/manager
https://api.github.com/repos/istio/manager
closed
Include GIT SHA in the binary builds
build & test infrastructure
Sometimes, its hard to identify if Bazel built the binary or not.. It would be nice to include all the git commit versions into the binary so that we can be sure that it is using the current version. For example, in amalgam8 we use something like this: ```makefile BUILD_SYM := github.com/amalgam8/amalgam8/pkg/version LDFLAGS += -X $(BUILD_SYM).version=$(APP_VER) LDFLAGS += -X $(BUILD_SYM).gitRevision=$(shell git rev-parse --short HEAD 2> /dev/null || echo unknown) LDFLAGS += -X $(BUILD_SYM).branch=$(shell git rev-parse --abbrev-ref HEAD 2> /dev/null || echo unknown) LDFLAGS += -X $(BUILD_SYM).buildUser=$(shell whoami || echo nobody)@$(shell hostname -f || echo builder) LDFLAGS += -X $(BUILD_SYM).buildDate=$(shell date +%Y-%m-%dT%H:%M:%S%:z) LDFLAGS += -X $(BUILD_SYM).goVersion=$(word 3,$(shell go version)) ``` We need an equivalent in bazel. Could someone help with this please?
1.0
Include GIT SHA in the binary builds - Sometimes, its hard to identify if Bazel built the binary or not.. It would be nice to include all the git commit versions into the binary so that we can be sure that it is using the current version. For example, in amalgam8 we use something like this: ```makefile BUILD_SYM := github.com/amalgam8/amalgam8/pkg/version LDFLAGS += -X $(BUILD_SYM).version=$(APP_VER) LDFLAGS += -X $(BUILD_SYM).gitRevision=$(shell git rev-parse --short HEAD 2> /dev/null || echo unknown) LDFLAGS += -X $(BUILD_SYM).branch=$(shell git rev-parse --abbrev-ref HEAD 2> /dev/null || echo unknown) LDFLAGS += -X $(BUILD_SYM).buildUser=$(shell whoami || echo nobody)@$(shell hostname -f || echo builder) LDFLAGS += -X $(BUILD_SYM).buildDate=$(shell date +%Y-%m-%dT%H:%M:%S%:z) LDFLAGS += -X $(BUILD_SYM).goVersion=$(word 3,$(shell go version)) ``` We need an equivalent in bazel. Could someone help with this please?
non_main
include git sha in the binary builds sometimes its hard to identify if bazel built the binary or not it would be nice to include all the git commit versions into the binary so that we can be sure that it is using the current version for example in we use something like this makefile build sym github com pkg version ldflags x build sym version app ver ldflags x build sym gitrevision shell git rev parse short head dev null echo unknown ldflags x build sym branch shell git rev parse abbrev ref head dev null echo unknown ldflags x build sym builduser shell whoami echo nobody shell hostname f echo builder ldflags x build sym builddate shell date y m dt h m s z ldflags x build sym goversion word shell go version we need an equivalent in bazel could someone help with this please
0
5,537
27,715,997,335
IssuesEvent
2023-03-14 16:58:32
MozillaFoundation/foundation.mozilla.org
https://api.github.com/repos/MozillaFoundation/foundation.mozilla.org
opened
Assess Percy false positive issue
engineering qa maintain
Placeholder ticket for now. Will add more description here soon. ## Related tickets - https://github.com/MozillaFoundation/foundation.mozilla.org/issues/10086 - https://github.com/MozillaFoundation/foundation.mozilla.org/issues/9440 ## Dev Tasks - [ ] identify pages that are currently included in Percy - [ ] identify pages that often give false-positives - [ ] assess if we want to include those pages in Percy
True
Assess Percy false positive issue - Placeholder ticket for now. Will add more description here soon. ## Related tickets - https://github.com/MozillaFoundation/foundation.mozilla.org/issues/10086 - https://github.com/MozillaFoundation/foundation.mozilla.org/issues/9440 ## Dev Tasks - [ ] identify pages that are currently included in Percy - [ ] identify pages that often give false-positives - [ ] assess if we want to include those pages in Percy
main
assess percy false positive issue placeholder ticket for now will add more description here soon related tickets dev tasks identify pages that are currently included in percy identify pages that often give false positives assess if we want to include those pages in percy
1
1,156
5,047,421,279
IssuesEvent
2016-12-20 09:22:42
ansible/ansible-modules-core
https://api.github.com/repos/ansible/ansible-modules-core
closed
docker_container with pull: true will always report changed
affects_2.2 bug_report cloud docker waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION *none* ##### OS / ENVIRONMENT Ubuntu 16.04 ##### SUMMARY When using `pull: yes`, the action always reports changed, even when the image is not changed (and so the container is not restarted). The old `docker` module would report ok in this case. ##### STEPS TO REPRODUCE Run this playbook twice: <!--- Paste example playbooks or commands between quotes below --> ``` - hosts: localhost tasks: - docker_container: name: test image: alpine pull: yes command: sleep 10000 ``` ##### EXPECTED RESULTS I would expect the second run to report *ok* for the task. ##### ACTUAL RESULTS The task reports *changed* on the second run. ``` $ ansible-playbook test.yml [WARNING]: provided hosts list is empty, only localhost is available PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [docker_container] ******************************************************** changed: [localhost] PLAY RECAP ********************************************************************* localhost : ok=2 changed=1 unreachable=0 failed=0 $ ansible-playbook test.yml [WARNING]: provided hosts list is empty, only localhost is available PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [docker_container] ******************************************************** changed: [localhost] PLAY RECAP ********************************************************************* localhost : ok=2 changed=1 unreachable=0 failed=0 ```
True
docker_container with pull: true will always report changed - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME docker_container ##### ANSIBLE VERSION ``` ansible 2.2.0.0 config file = /etc/ansible/ansible.cfg configured module search path = Default w/o overrides ``` ##### CONFIGURATION *none* ##### OS / ENVIRONMENT Ubuntu 16.04 ##### SUMMARY When using `pull: yes`, the action always reports changed, even when the image is not changed (and so the container is not restarted). The old `docker` module would report ok in this case. ##### STEPS TO REPRODUCE Run this playbook twice: <!--- Paste example playbooks or commands between quotes below --> ``` - hosts: localhost tasks: - docker_container: name: test image: alpine pull: yes command: sleep 10000 ``` ##### EXPECTED RESULTS I would expect the second run to report *ok* for the task. ##### ACTUAL RESULTS The task reports *changed* on the second run. ``` $ ansible-playbook test.yml [WARNING]: provided hosts list is empty, only localhost is available PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [docker_container] ******************************************************** changed: [localhost] PLAY RECAP ********************************************************************* localhost : ok=2 changed=1 unreachable=0 failed=0 $ ansible-playbook test.yml [WARNING]: provided hosts list is empty, only localhost is available PLAY [localhost] *************************************************************** TASK [setup] ******************************************************************* ok: [localhost] TASK [docker_container] ******************************************************** changed: [localhost] PLAY RECAP ********************************************************************* localhost : ok=2 changed=1 unreachable=0 failed=0 ```
main
docker container with pull true will always report changed issue type bug report component name docker container ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration none os environment ubuntu summary when using pull yes the action always reports changed even when the image is not changed and so the container is not restarted the old docker module would report ok in this case steps to reproduce run this playbook twice hosts localhost tasks docker container name test image alpine pull yes command sleep expected results i would expect the second run to report ok for the task actual results the task reports changed on the second run ansible playbook test yml provided hosts list is empty only localhost is available play task ok task changed play recap localhost ok changed unreachable failed ansible playbook test yml provided hosts list is empty only localhost is available play task ok task changed play recap localhost ok changed unreachable failed
1
440,366
30,743,683,918
IssuesEvent
2023-07-28 13:29:24
taiki-e/cargo-llvm-cov
https://api.github.com/repos/taiki-e/cargo-llvm-cov
opened
docs: cross-testing
C-documentation
When used together with [setup-cross-toolchain-action](https://github.com/taiki-e/setup-cross-toolchain-action), we can easily get coverage of the target different from the host. It would be nice to document this in readme.
1.0
docs: cross-testing - When used together with [setup-cross-toolchain-action](https://github.com/taiki-e/setup-cross-toolchain-action), we can easily get coverage of the target different from the host. It would be nice to document this in readme.
non_main
docs cross testing when used together with we can easily get coverage of the target different from the host it would be nice to document this in readme
0
4,060
18,983,452,008
IssuesEvent
2021-11-21 09:55:35
svengreb/wand
https://api.github.com/repos/svengreb/wand
opened
Insufficient repository fetch-depth for action workflows
type-bug context-workflow scope-maintainability scope-quality
The [GitHub action workflows][1] using the [`actions/checkout` action][2] to fetch the repository that triggered the workflow. However, by default only the history of the latest commit is fetched which results in errors when the _wand_ tries to extract repository metadata information like the amount of commits ahead of the latest commit. As an example this can be seen when [running the `bootstrap` command in the `test` job of the `ci-go` workflow][5] which fails with an `object not found` error because the history only contains a single commit. To fiy this problem `action/checkout` provides an option to [fetch all history for all tags and branches][3] which will be used to prevent errors like this in the pipeline. [1]: https://github.com/svengreb/wand/tree/9caf10f9d3b0c97e1f6c18b29c175e71764b0ece/.github/workflows [2]: https://github.com/actions/checkout [3]: https://github.com/actions/checkout#Fetch-all-history-for-all-tags-and-branches [4]: https://github.com/svengreb/wand/blob/cabd635c4ec73680b1776e7c536feca16643b00b/magefile.go#L136 [5]: https://github.com/svengreb/wand/runs/4275275079?check_suite_focus=true
True
Insufficient repository fetch-depth for action workflows - The [GitHub action workflows][1] using the [`actions/checkout` action][2] to fetch the repository that triggered the workflow. However, by default only the history of the latest commit is fetched which results in errors when the _wand_ tries to extract repository metadata information like the amount of commits ahead of the latest commit. As an example this can be seen when [running the `bootstrap` command in the `test` job of the `ci-go` workflow][5] which fails with an `object not found` error because the history only contains a single commit. To fiy this problem `action/checkout` provides an option to [fetch all history for all tags and branches][3] which will be used to prevent errors like this in the pipeline. [1]: https://github.com/svengreb/wand/tree/9caf10f9d3b0c97e1f6c18b29c175e71764b0ece/.github/workflows [2]: https://github.com/actions/checkout [3]: https://github.com/actions/checkout#Fetch-all-history-for-all-tags-and-branches [4]: https://github.com/svengreb/wand/blob/cabd635c4ec73680b1776e7c536feca16643b00b/magefile.go#L136 [5]: https://github.com/svengreb/wand/runs/4275275079?check_suite_focus=true
main
insufficient repository fetch depth for action workflows the using the to fetch the repository that triggered the workflow however by default only the history of the latest commit is fetched which results in errors when the wand tries to extract repository metadata information like the amount of commits ahead of the latest commit as an example this can be seen when which fails with an object not found error because the history only contains a single commit to fiy this problem action checkout provides an option to which will be used to prevent errors like this in the pipeline
1
1,549
6,572,241,282
IssuesEvent
2017-09-11 00:29:38
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
pam_limits comment problem
affects_2.0 bug_report waiting_on_maintainer
##### ISSUE TYPE - Bug Report ##### COMPONENT NAME pam_limits ##### ANSIBLE VERSION ``` ansible 2.0.2.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION none ##### OS / ENVIRONMENT Controller: MacOSX 10.11.5 Target: CentOS/7 ##### SUMMARY pam_limits misformats comments, first comment has no "#" (and not even space) prefix, third comment is indented twice. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> ``` - pam_limits: domain={{item.domain}} limit_type={{item.limit_type}} limit_item={{item.limit_item}} value={{item.value}} comment={{item.comment}} with_items: - { domain: 'user1', limit_type: 'soft', limit_item: 'nofile', value: '4096', comment: 'Comment 1' } - { domain: 'user2', limit_type: 'soft', limit_item: 'nofile', value: '4096', comment: 'Comment 2' } - { domain: 'user3', limit_type: 'soft', limit_item: 'nofile', value: '4096', comment: 'Comment 3' } ``` ##### EXPECTED RESULTS ``` user1 soft nofile 4096 #Comment 1 user2 soft nofile 4096 #Comment 2 user3 soft nofile 4096 #Comment 3 # End of file ``` ##### ACTUAL RESULTS ``` # End of file user1 soft nofile 4096Comment 1 user2 soft nofile 4096 #Comment 2 user3 soft nofile 4096 # #Comment 3 ```
True
pam_limits comment problem - ##### ISSUE TYPE - Bug Report ##### COMPONENT NAME pam_limits ##### ANSIBLE VERSION ``` ansible 2.0.2.0 config file = configured module search path = Default w/o overrides ``` ##### CONFIGURATION none ##### OS / ENVIRONMENT Controller: MacOSX 10.11.5 Target: CentOS/7 ##### SUMMARY pam_limits misformats comments, first comment has no "#" (and not even space) prefix, third comment is indented twice. ##### STEPS TO REPRODUCE <!--- For bugs, show exactly how to reproduce the problem. For new features, show how the feature would be used. --> ``` - pam_limits: domain={{item.domain}} limit_type={{item.limit_type}} limit_item={{item.limit_item}} value={{item.value}} comment={{item.comment}} with_items: - { domain: 'user1', limit_type: 'soft', limit_item: 'nofile', value: '4096', comment: 'Comment 1' } - { domain: 'user2', limit_type: 'soft', limit_item: 'nofile', value: '4096', comment: 'Comment 2' } - { domain: 'user3', limit_type: 'soft', limit_item: 'nofile', value: '4096', comment: 'Comment 3' } ``` ##### EXPECTED RESULTS ``` user1 soft nofile 4096 #Comment 1 user2 soft nofile 4096 #Comment 2 user3 soft nofile 4096 #Comment 3 # End of file ``` ##### ACTUAL RESULTS ``` # End of file user1 soft nofile 4096Comment 1 user2 soft nofile 4096 #Comment 2 user3 soft nofile 4096 # #Comment 3 ```
main
pam limits comment problem issue type bug report component name pam limits ansible version ansible config file configured module search path default w o overrides configuration none os environment controller macosx target centos summary pam limits misformats comments first comment has no and not even space prefix third comment is indented twice steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used pam limits domain item domain limit type item limit type limit item item limit item value item value comment item comment with items domain limit type soft limit item nofile value comment comment domain limit type soft limit item nofile value comment comment domain limit type soft limit item nofile value comment comment expected results soft nofile comment soft nofile comment soft nofile comment end of file actual results end of file soft nofile soft nofile comment soft nofile comment
1
3,975
18,278,496,173
IssuesEvent
2021-10-04 22:10:27
carbon-design-system/carbon
https://api.github.com/repos/carbon-design-system/carbon
closed
[Bug]: Dropdown menu should overlay Modal
type: bug 🐛 status: needs triage 🕵️‍♀️ status: waiting for maintainer response 💬
### Package carbon-components-react ### Browser Chrome ### Package version 7.43.0 ### Description I think that would be cool if dropdown menu will overflow modal so that you don't have to scroll to see the menu. ` .bx--modal-container, bx--modal-content { overflow: visible; }` helps to solve the problem for modal that don't have scrolling content. But I cannot find a good solution for the case when you do have scrolling content (like in my example in codesandbox). That's how it looks now ![Screenshot 2021-09-29 at 10 52 33](https://user-images.githubusercontent.com/21106246/135237410-210f36bc-0ea7-46af-98d9-e1585c517a63.png) That's how it should look like ![Screenshot 2021-09-29 at 10 56 43](https://user-images.githubusercontent.com/21106246/135237438-6e84928e-6cdb-490d-9f82-799d5de42ba1.png) ### CodeSandbox example https://codesandbox.io/s/epic-cookies-emg02?file=/src/index.js ### Steps to reproduce 1) Create a modal 3) Create dropdown inside modal body 4) Click on dropdown to open menu ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [ ] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
True
[Bug]: Dropdown menu should overlay Modal - ### Package carbon-components-react ### Browser Chrome ### Package version 7.43.0 ### Description I think that would be cool if dropdown menu will overflow modal so that you don't have to scroll to see the menu. ` .bx--modal-container, bx--modal-content { overflow: visible; }` helps to solve the problem for modal that don't have scrolling content. But I cannot find a good solution for the case when you do have scrolling content (like in my example in codesandbox). That's how it looks now ![Screenshot 2021-09-29 at 10 52 33](https://user-images.githubusercontent.com/21106246/135237410-210f36bc-0ea7-46af-98d9-e1585c517a63.png) That's how it should look like ![Screenshot 2021-09-29 at 10 56 43](https://user-images.githubusercontent.com/21106246/135237438-6e84928e-6cdb-490d-9f82-799d5de42ba1.png) ### CodeSandbox example https://codesandbox.io/s/epic-cookies-emg02?file=/src/index.js ### Steps to reproduce 1) Create a modal 3) Create dropdown inside modal body 4) Click on dropdown to open menu ### Code of Conduct - [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md) - [ ] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems
main
dropdown menu should overlay modal package carbon components react browser chrome package version description i think that would be cool if dropdown menu will overflow modal so that you don t have to scroll to see the menu bx modal container bx modal content overflow visible helps to solve the problem for modal that don t have scrolling content but i cannot find a good solution for the case when you do have scrolling content like in my example in codesandbox that s how it looks now that s how it should look like codesandbox example steps to reproduce create a modal create dropdown inside modal body click on dropdown to open menu code of conduct i agree to follow this project s i checked the for duplicate problems
1
669
4,211,235,605
IssuesEvent
2016-06-29 12:55:30
Particular/NServiceBus
https://api.github.com/repos/Particular/NServiceBus
closed
V6 - Prevent users from injecting IEndpointInstance and using it inside of Message handlers to send / publish messages
Project: V6 Launch State: In Progress - Maintainer Prio Tag: Maintainer Prio
Relates to: https://github.com/Particular/V6Launch/issues/27 https://github.com/Particular/PlatformDevelopment/issues/819 ### Problem Often times customers have old / common libraries where a IBus was injected into the library. And the IBus was used to send messages / publish events from a self host or a NServiceBus.Host either in bootstrap scenarios or in message handlers. However, `IEndpointInstance/MessageSession` is not safe to use from message handlers. Some example of issues with sending messages using IEndpointInstance/IMessageSession within a message handler are: - those messages will not participate in the same transaction scope as that of the message handler. This could result in messages dispatched or events published via the `IEndpointInstance` interface even if the message handler resulted in an exception and the operation was rolled back. - those messages will not be part of the [batching operation](/nservicebus/messaging/batched-dispatch.md) and therefore will not have the benefits that Version 6 provides. - those messages will not contain any important message header information that is available via the `IHandlerMessageContext` interface parameter, e.g., CorrelationId. How can we stop new and existing users from writing "invalid" code from either compiling or throwing a run time warning / exception? Because without it and just relying on guidance, users can still run into issues with code that compile and seems like it runs fine on the surface while leading to undesirable run time downsides, especially when a message handler transaction fails. ### Proposed solution **Type checking on handlers at startup** To prevent users from wrongly using IEndpointInstance and IMessageSession on their handlers we should check on startup and throw an exception if any handler has a ctor/property dependency in said types. The exception should tell them to use the handler context instead.
True
V6 - Prevent users from injecting IEndpointInstance and using it inside of Message handlers to send / publish messages - Relates to: https://github.com/Particular/V6Launch/issues/27 https://github.com/Particular/PlatformDevelopment/issues/819 ### Problem Often times customers have old / common libraries where a IBus was injected into the library. And the IBus was used to send messages / publish events from a self host or a NServiceBus.Host either in bootstrap scenarios or in message handlers. However, `IEndpointInstance/MessageSession` is not safe to use from message handlers. Some example of issues with sending messages using IEndpointInstance/IMessageSession within a message handler are: - those messages will not participate in the same transaction scope as that of the message handler. This could result in messages dispatched or events published via the `IEndpointInstance` interface even if the message handler resulted in an exception and the operation was rolled back. - those messages will not be part of the [batching operation](/nservicebus/messaging/batched-dispatch.md) and therefore will not have the benefits that Version 6 provides. - those messages will not contain any important message header information that is available via the `IHandlerMessageContext` interface parameter, e.g., CorrelationId. How can we stop new and existing users from writing "invalid" code from either compiling or throwing a run time warning / exception? Because without it and just relying on guidance, users can still run into issues with code that compile and seems like it runs fine on the surface while leading to undesirable run time downsides, especially when a message handler transaction fails. ### Proposed solution **Type checking on handlers at startup** To prevent users from wrongly using IEndpointInstance and IMessageSession on their handlers we should check on startup and throw an exception if any handler has a ctor/property dependency in said types. The exception should tell them to use the handler context instead.
main
prevent users from injecting iendpointinstance and using it inside of message handlers to send publish messages relates to problem often times customers have old common libraries where a ibus was injected into the library and the ibus was used to send messages publish events from a self host or a nservicebus host either in bootstrap scenarios or in message handlers however iendpointinstance messagesession is not safe to use from message handlers some example of issues with sending messages using iendpointinstance imessagesession within a message handler are those messages will not participate in the same transaction scope as that of the message handler this could result in messages dispatched or events published via the iendpointinstance interface even if the message handler resulted in an exception and the operation was rolled back those messages will not be part of the nservicebus messaging batched dispatch md and therefore will not have the benefits that version provides those messages will not contain any important message header information that is available via the ihandlermessagecontext interface parameter e g correlationid how can we stop new and existing users from writing invalid code from either compiling or throwing a run time warning exception because without it and just relying on guidance users can still run into issues with code that compile and seems like it runs fine on the surface while leading to undesirable run time downsides especially when a message handler transaction fails proposed solution type checking on handlers at startup to prevent users from wrongly using iendpointinstance and imessagesession on their handlers we should check on startup and throw an exception if any handler has a ctor property dependency in said types the exception should tell them to use the handler context instead
1
496,570
14,349,811,008
IssuesEvent
2020-11-29 18:09:51
Poobslag/turbofat
https://api.github.com/repos/Poobslag/turbofat
closed
Add environment details: static objects like trees and bushes which get in your way
priority-3
These should be used for decoration, and maybe occasionally as obstacles.
1.0
Add environment details: static objects like trees and bushes which get in your way - These should be used for decoration, and maybe occasionally as obstacles.
non_main
add environment details static objects like trees and bushes which get in your way these should be used for decoration and maybe occasionally as obstacles
0
1,061
4,877,033,163
IssuesEvent
2016-11-16 14:38:06
ansible/ansible-modules-extras
https://api.github.com/repos/ansible/ansible-modules-extras
closed
rabbitmq_user module assumes users have only one set of permissions
affects_2.1 feature_idea waiting_on_maintainer
##### ISSUE TYPE Feature Idea ##### COMPONENT NAME rabbitmq_user module ##### ANSIBLE VERSION N/A ##### SUMMARY Hi there! The current implementation of `rabbitmq_user` manages both user creation, tagging and permission setting. Unfortunately users do not have a 1-1 relationship with vhost permissions. I created a local module to manage permissions separately, allowing the correct 1-N permissions, but since the `rabbitmq_user` module still tries to manage those, I have to create one set of permissions on the `rabbitmq_user` and the rest on my custom resource `rabbitmq_permission`. In my opinion it would be best if we added an option to the `rabbitmq_user` to not manage the permissions. This way I could provide a PR to add this new module, keep the existing functionality working and allowing multiple vhost permissions to be properly configured. Another thing is the naming. The current naming suggests **privileges**, but RabbitMQ docs all use the word **permission**. For consistency's sake, I think we should follow RMQ's wording and deprecate the old wording. Any feedback is appreciated!
True
rabbitmq_user module assumes users have only one set of permissions - ##### ISSUE TYPE Feature Idea ##### COMPONENT NAME rabbitmq_user module ##### ANSIBLE VERSION N/A ##### SUMMARY Hi there! The current implementation of `rabbitmq_user` manages both user creation, tagging and permission setting. Unfortunately users do not have a 1-1 relationship with vhost permissions. I created a local module to manage permissions separately, allowing the correct 1-N permissions, but since the `rabbitmq_user` module still tries to manage those, I have to create one set of permissions on the `rabbitmq_user` and the rest on my custom resource `rabbitmq_permission`. In my opinion it would be best if we added an option to the `rabbitmq_user` to not manage the permissions. This way I could provide a PR to add this new module, keep the existing functionality working and allowing multiple vhost permissions to be properly configured. Another thing is the naming. The current naming suggests **privileges**, but RabbitMQ docs all use the word **permission**. For consistency's sake, I think we should follow RMQ's wording and deprecate the old wording. Any feedback is appreciated!
main
rabbitmq user module assumes users have only one set of permissions issue type feature idea component name rabbitmq user module ansible version n a summary hi there the current implementation of rabbitmq user manages both user creation tagging and permission setting unfortunately users do not have a relationship with vhost permissions i created a local module to manage permissions separately allowing the correct n permissions but since the rabbitmq user module still tries to manage those i have to create one set of permissions on the rabbitmq user and the rest on my custom resource rabbitmq permission in my opinion it would be best if we added an option to the rabbitmq user to not manage the permissions this way i could provide a pr to add this new module keep the existing functionality working and allowing multiple vhost permissions to be properly configured another thing is the naming the current naming suggests privileges but rabbitmq docs all use the word permission for consistency s sake i think we should follow rmq s wording and deprecate the old wording any feedback is appreciated
1