Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
2,938 | 10,526,825,036 | IssuesEvent | 2019-09-30 17:57:15 | 18F/cg-product | https://api.github.com/repos/18F/cg-product | closed | [research: X days] Figure our easiest path to making k8s stemcells updates trivial | contractor-3-maintainability operations | In order to make K8S stemcell updates sustainable and learnable, we need to learn from Jon's experiences leading the updates for the K8S stemcells that are time sensitive for 7/24 (for compliance deadline reasons).
This card is a placeholder for the work to actually figure out how to put those ideas and work into practice. | True | [research: X days] Figure our easiest path to making k8s stemcells updates trivial - In order to make K8S stemcell updates sustainable and learnable, we need to learn from Jon's experiences leading the updates for the K8S stemcells that are time sensitive for 7/24 (for compliance deadline reasons).
This card is a placeholder for the work to actually figure out how to put those ideas and work into practice. | main | figure our easiest path to making stemcells updates trivial in order to make stemcell updates sustainable and learnable we need to learn from jon s experiences leading the updates for the stemcells that are time sensitive for for compliance deadline reasons this card is a placeholder for the work to actually figure out how to put those ideas and work into practice | 1 |
1,086 | 4,934,170,651 | IssuesEvent | 2016-11-28 18:18:48 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | copy: dest=filename fails with "Destination directory does not exist" | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
copy
##### ANSIBLE VERSION
```
ansible 2.1.2.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
##### OS / ENVIRONMENT
Linux
##### SUMMARY
When I try to copy a file into the current working directory on the remote end, Ansible fails because it tries to check for the existence of the (empty string) directory.
##### STEPS TO REPRODUCE
```
$ ansible localhost -m copy -a "src=/dev/null dest=file.txt"
```
##### EXPECTED RESULTS
I expect an empty file `file.txt` to show up in my home directory
##### ACTUAL RESULTS
```
localhost | FAILED! => {
"changed": false,
"checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
"failed": true,
"msg": "Destination directory does not exist"
}
```
| True | copy: dest=filename fails with "Destination directory does not exist" - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
copy
##### ANSIBLE VERSION
```
ansible 2.1.2.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
##### OS / ENVIRONMENT
Linux
##### SUMMARY
When I try to copy a file into the current working directory on the remote end, Ansible fails because it tries to check for the existence of the (empty string) directory.
##### STEPS TO REPRODUCE
```
$ ansible localhost -m copy -a "src=/dev/null dest=file.txt"
```
##### EXPECTED RESULTS
I expect an empty file `file.txt` to show up in my home directory
##### ACTUAL RESULTS
```
localhost | FAILED! => {
"changed": false,
"checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709",
"failed": true,
"msg": "Destination directory does not exist"
}
```
| main | copy dest filename fails with destination directory does not exist issue type bug report component name copy ansible version ansible config file configured module search path default w o overrides configuration os environment linux summary when i try to copy a file into the current working directory on the remote end ansible fails because it tries to check for the existence of the empty string directory steps to reproduce ansible localhost m copy a src dev null dest file txt expected results i expect an empty file file txt to show up in my home directory actual results localhost failed changed false checksum failed true msg destination directory does not exist | 1 |
251,228 | 21,447,894,476 | IssuesEvent | 2022-04-25 08:25:27 | MohistMC/Mohist | https://api.github.com/repos/MohistMC/Mohist | closed | [1.16.5] Plugin ItemAdder doesn't work | 1.16.5 Wait Needs Testing | <!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.-->
<!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.-->
<!-- If you don't know where to upload your logs and crash reports, you can use these websites : -->
<!-- https://paste.ubuntu.com/ (recommended) -->
<!-- https://mclo.gs -->
<!-- https://haste.mohistmc.com -->
<!-- https://pastebin.com -->
<!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT -->
**Minecraft Version :** 1.16.5
**Mohist Version :** 811
**Operating System :** Linux
**Concerned mod / plugin** : ItemAdder (https://www.spigotmc.org/resources/%E2%9C%85must-have%E2%9C%85-itemsadder%E2%9C%A8custom-items-huds-guis-mobs-3dmodels-emojis-blocks-wings-hats-liquids.73355/)
**Logs :** https://pastebin.com/5wwNeTws
**Steps to Reproduce :**
1. Instal ItemAdder
2. Start Server
**Description of issue :** There are a lot of reports in the console. It is not possible to log in to the server again without restarting. After the restart - it is impossible to move
| 1.0 | [1.16.5] Plugin ItemAdder doesn't work - <!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.-->
<!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.-->
<!-- If you don't know where to upload your logs and crash reports, you can use these websites : -->
<!-- https://paste.ubuntu.com/ (recommended) -->
<!-- https://mclo.gs -->
<!-- https://haste.mohistmc.com -->
<!-- https://pastebin.com -->
<!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT -->
**Minecraft Version :** 1.16.5
**Mohist Version :** 811
**Operating System :** Linux
**Concerned mod / plugin** : ItemAdder (https://www.spigotmc.org/resources/%E2%9C%85must-have%E2%9C%85-itemsadder%E2%9C%A8custom-items-huds-guis-mobs-3dmodels-emojis-blocks-wings-hats-liquids.73355/)
**Logs :** https://pastebin.com/5wwNeTws
**Steps to Reproduce :**
1. Instal ItemAdder
2. Start Server
**Description of issue :** There are a lot of reports in the console. It is not possible to log in to the server again without restarting. After the restart - it is impossible to move
| non_main | plugin itemadder doesn t work important do not delete this line minecraft version mohist version operating system linux concerned mod plugin itemadder logs steps to reproduce instal itemadder start server description of issue there are a lot of reports in the console it is not possible to log in to the server again without restarting after the restart it is impossible to move | 0 |
4,426 | 22,796,849,113 | IssuesEvent | 2022-07-10 21:05:49 | jesus2099/konami-command | https://api.github.com/repos/jesus2099/konami-command | closed | COOL_SEARCH_LINKS: Drop (now redundant) All edits and Open edits Cool links | mb_SUPER-MIND-CONTROL-II-X-TURBO minor maintainability | I see now:
- **Open edits for this entity** in https://musicbrainz.org/artist/88d8f38f-adb4-48a0-8c1f-ec34f2a675ff/edits
- **All edits for this entity** in https://musicbrainz.org/artist/88d8f38f-adb4-48a0-8c1f-ec34f2a675ff/open_edits
- **Open edits for jesus2099** in https://musicbrainz.org/user/jesus2099/edits
- **All edits for jesus2099** in https://musicbrainz.org/user/jesus2099/edits/open
So my **All edits** and **Open edits** _Cool links_ are now redundant.
Check that it is not missing anywhere, first. | True | COOL_SEARCH_LINKS: Drop (now redundant) All edits and Open edits Cool links - I see now:
- **Open edits for this entity** in https://musicbrainz.org/artist/88d8f38f-adb4-48a0-8c1f-ec34f2a675ff/edits
- **All edits for this entity** in https://musicbrainz.org/artist/88d8f38f-adb4-48a0-8c1f-ec34f2a675ff/open_edits
- **Open edits for jesus2099** in https://musicbrainz.org/user/jesus2099/edits
- **All edits for jesus2099** in https://musicbrainz.org/user/jesus2099/edits/open
So my **All edits** and **Open edits** _Cool links_ are now redundant.
Check that it is not missing anywhere, first. | main | cool search links drop now redundant all edits and open edits cool links i see now open edits for this entity in all edits for this entity in open edits for in all edits for in so my all edits and open edits cool links are now redundant check that it is not missing anywhere first | 1 |
272,289 | 29,795,002,450 | IssuesEvent | 2023-06-16 01:03:22 | billmcchesney1/flowgate | https://api.github.com/repos/billmcchesney1/flowgate | closed | CVE-2022-22965 (Critical) detected in spring-beans-4.3.9.RELEASE.jar, spring-beans-5.2.6.RELEASE.jar - autoclosed | Mend: dependency security vulnerability | ## CVE-2022-22965 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-beans-4.3.9.RELEASE.jar</b>, <b>spring-beans-5.2.6.RELEASE.jar</b></p></summary>
<p>
<details><summary><b>spring-beans-4.3.9.RELEASE.jar</b></p></summary>
<p>Spring Beans</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /operation-expert/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-beans/4.3.9.RELEASE/spring-beans-4.3.9.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-mongodb-1.4.7.RELEASE.jar (Root Library)
- spring-data-mongodb-1.9.11.RELEASE.jar
- :x: **spring-beans-4.3.9.RELEASE.jar** (Vulnerable Library)
</details>
<details><summary><b>spring-beans-5.2.6.RELEASE.jar</b></p></summary>
<p>Spring Beans</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /infoblox-worker/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-hateoas-2.2.7.RELEASE.jar (Root Library)
- spring-hateoas-1.0.5.RELEASE.jar
- :x: **spring-beans-5.2.6.RELEASE.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flowgate/commit/dd01a1d4381c7a3b94ba25748c015a094c33088e">dd01a1d4381c7a3b94ba25748c015a094c33088e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A Spring MVC or Spring WebFlux application running on JDK 9+ may be vulnerable to remote code execution (RCE) via data binding. The specific exploit requires the application to run on Tomcat as a WAR deployment. If the application is deployed as a Spring Boot executable jar, i.e. the default, it is not vulnerable to the exploit. However, the nature of the vulnerability is more general, and there may be other ways to exploit it.
Mend Note: Converted from WS-2022-0107, on 2022-11-07.
<p>Publish Date: 2022-04-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22965>CVE-2022-22965</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://spring.io/blog/2022/03/31/spring-framework-rce-early-announcement">https://spring.io/blog/2022/03/31/spring-framework-rce-early-announcement</a></p>
<p>Release Date: 2022-04-01</p>
<p>Fix Resolution: org.springframework:spring-beans:5.2.20.RELEASE,5.3.18</p>
</p>
</details>
<p></p>
| True | CVE-2022-22965 (Critical) detected in spring-beans-4.3.9.RELEASE.jar, spring-beans-5.2.6.RELEASE.jar - autoclosed - ## CVE-2022-22965 - Critical Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>spring-beans-4.3.9.RELEASE.jar</b>, <b>spring-beans-5.2.6.RELEASE.jar</b></p></summary>
<p>
<details><summary><b>spring-beans-4.3.9.RELEASE.jar</b></p></summary>
<p>Spring Beans</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /operation-expert/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-beans/4.3.9.RELEASE/spring-beans-4.3.9.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-data-mongodb-1.4.7.RELEASE.jar (Root Library)
- spring-data-mongodb-1.9.11.RELEASE.jar
- :x: **spring-beans-4.3.9.RELEASE.jar** (Vulnerable Library)
</details>
<details><summary><b>spring-beans-5.2.6.RELEASE.jar</b></p></summary>
<p>Spring Beans</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: /infoblox-worker/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar,/home/wss-scanner/.m2/repository/org/springframework/spring-beans/5.2.6.RELEASE/spring-beans-5.2.6.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-hateoas-2.2.7.RELEASE.jar (Root Library)
- spring-hateoas-1.0.5.RELEASE.jar
- :x: **spring-beans-5.2.6.RELEASE.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flowgate/commit/dd01a1d4381c7a3b94ba25748c015a094c33088e">dd01a1d4381c7a3b94ba25748c015a094c33088e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/critical_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
A Spring MVC or Spring WebFlux application running on JDK 9+ may be vulnerable to remote code execution (RCE) via data binding. The specific exploit requires the application to run on Tomcat as a WAR deployment. If the application is deployed as a Spring Boot executable jar, i.e. the default, it is not vulnerable to the exploit. However, the nature of the vulnerability is more general, and there may be other ways to exploit it.
Mend Note: Converted from WS-2022-0107, on 2022-11-07.
<p>Publish Date: 2022-04-01
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-22965>CVE-2022-22965</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://spring.io/blog/2022/03/31/spring-framework-rce-early-announcement">https://spring.io/blog/2022/03/31/spring-framework-rce-early-announcement</a></p>
<p>Release Date: 2022-04-01</p>
<p>Fix Resolution: org.springframework:spring-beans:5.2.20.RELEASE,5.3.18</p>
</p>
</details>
<p></p>
| non_main | cve critical detected in spring beans release jar spring beans release jar autoclosed cve critical severity vulnerability vulnerable libraries spring beans release jar spring beans release jar spring beans release jar spring beans library home page a href path to dependency file operation expert pom xml path to vulnerable library home wss scanner repository org springframework spring beans release spring beans release jar dependency hierarchy spring boot starter data mongodb release jar root library spring data mongodb release jar x spring beans release jar vulnerable library spring beans release jar spring beans library home page a href path to dependency file infoblox worker pom xml path to vulnerable library home wss scanner repository org springframework spring beans release spring beans release jar home wss scanner repository org springframework spring beans release spring beans release jar home wss scanner repository org springframework spring beans release spring beans release jar home wss scanner repository org springframework spring beans release spring beans release jar home wss scanner repository org springframework spring beans release spring beans release jar home wss scanner repository org springframework spring beans release spring beans release jar home wss scanner repository org springframework spring beans release spring beans release jar home wss scanner repository org springframework spring beans release spring beans release jar dependency hierarchy spring boot starter hateoas release jar root library spring hateoas release jar x spring beans release jar vulnerable library found in head commit a href found in base branch master vulnerability details a spring mvc or spring webflux application running on jdk may be vulnerable to remote code execution rce via data binding the specific exploit requires the application to run on tomcat as a war deployment if the application is deployed as a spring boot executable jar i e the default it is not vulnerable to the exploit however the nature of the vulnerability is more general and there may be other ways to exploit it mend note converted from ws on publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring beans release | 0 |
131,934 | 18,262,001,351 | IssuesEvent | 2021-10-04 00:41:00 | ghc-dev/Sierra-Randall | https://api.github.com/repos/ghc-dev/Sierra-Randall | opened | CVE-2017-16138 (High) detected in mime-1.2.11.tgz | security vulnerability | ## CVE-2017-16138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.2.11.tgz</b></p></summary>
<p>A comprehensive library for mime-type mapping</p>
<p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.2.11.tgz">https://registry.npmjs.org/mime/-/mime-1.2.11.tgz</a></p>
<p>Path to dependency file: Sierra-Randall/package.json</p>
<p>Path to vulnerable library: Sierra-Randall/node_modules/mime/package.json</p>
<p>
Dependency Hierarchy:
- send-0.11.1.tgz (Root Library)
- :x: **mime-1.2.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Sierra-Randall/commit/770a54c5cebfd1b24a2c0a8eb9d2c47fc00c0270">770a54c5cebfd1b24a2c0a8eb9d2c47fc00c0270</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 1.4.1,2.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.2.11","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"send:0.11.1;mime:1.2.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2017-16138 (High) detected in mime-1.2.11.tgz - ## CVE-2017-16138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>mime-1.2.11.tgz</b></p></summary>
<p>A comprehensive library for mime-type mapping</p>
<p>Library home page: <a href="https://registry.npmjs.org/mime/-/mime-1.2.11.tgz">https://registry.npmjs.org/mime/-/mime-1.2.11.tgz</a></p>
<p>Path to dependency file: Sierra-Randall/package.json</p>
<p>Path to vulnerable library: Sierra-Randall/node_modules/mime/package.json</p>
<p>
Dependency Hierarchy:
- send-0.11.1.tgz (Root Library)
- :x: **mime-1.2.11.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ghc-dev/Sierra-Randall/commit/770a54c5cebfd1b24a2c0a8eb9d2c47fc00c0270">770a54c5cebfd1b24a2c0a8eb9d2c47fc00c0270</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The mime module < 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.
<p>Publish Date: 2018-06-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138>CVE-2017-16138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2017-16138</a></p>
<p>Release Date: 2018-06-07</p>
<p>Fix Resolution: 1.4.1,2.0.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"mime","packageVersion":"1.2.11","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"send:0.11.1;mime:1.2.11","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.1,2.0.3"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2017-16138","vulnerabilityDetails":"The mime module \u003c 1.4.1, 2.0.1, 2.0.2 is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2017-16138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_main | cve high detected in mime tgz cve high severity vulnerability vulnerable library mime tgz a comprehensive library for mime type mapping library home page a href path to dependency file sierra randall package json path to vulnerable library sierra randall node modules mime package json dependency hierarchy send tgz root library x mime tgz vulnerable library found in head commit a href found in base branch master vulnerability details the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree send mime isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails the mime module is vulnerable to regular expression denial of service when a mime lookup is performed on untrusted user input vulnerabilityurl | 0 |
5,465 | 27,323,202,538 | IssuesEvent | 2023-02-24 22:06:32 | sipb/uplink | https://api.github.com/repos/sipb/uplink | opened | Optimize postgres | P2 dev/maintainability | > The default settings should be fine for most deployments. For larger scale deployments tuning some of the settings is recommended
https://matrix-org.github.io/synapse/latest/postgres.html#tuning-postgres | True | Optimize postgres - > The default settings should be fine for most deployments. For larger scale deployments tuning some of the settings is recommended
https://matrix-org.github.io/synapse/latest/postgres.html#tuning-postgres | main | optimize postgres the default settings should be fine for most deployments for larger scale deployments tuning some of the settings is recommended | 1 |
3,828 | 16,654,473,475 | IssuesEvent | 2021-06-05 09:06:38 | travisbrown/dhallj | https://api.github.com/repos/travisbrown/dhallj | closed | Clean up toString implementation | enhancement maintainability | This is a follow-up to #7. I've now got `toString` for `Expr` working well enough that it can be used to round-trip the unnormalised prelude, which was my goal for 0.1.0, but the implementation is still a disaster. I'd originally been using it for debugging and didn't really care about producing valid Dhall code, and I just threw together the current version in the past couple of days. It parenthesises unnecessarily, probably still gets precedence wrong in some cases, is an unmaintainable mess, etc.
We'll also probably want some kind of Dhall code pretty-printing at some point, but I'll open a separate issue for that. | True | Clean up toString implementation - This is a follow-up to #7. I've now got `toString` for `Expr` working well enough that it can be used to round-trip the unnormalised prelude, which was my goal for 0.1.0, but the implementation is still a disaster. I'd originally been using it for debugging and didn't really care about producing valid Dhall code, and I just threw together the current version in the past couple of days. It parenthesises unnecessarily, probably still gets precedence wrong in some cases, is an unmaintainable mess, etc.
We'll also probably want some kind of Dhall code pretty-printing at some point, but I'll open a separate issue for that. | main | clean up tostring implementation this is a follow up to i ve now got tostring for expr working well enough that it can be used to round trip the unnormalised prelude which was my goal for but the implementation is still a disaster i d originally been using it for debugging and didn t really care about producing valid dhall code and i just threw together the current version in the past couple of days it parenthesises unnecessarily probably still gets precedence wrong in some cases is an unmaintainable mess etc we ll also probably want some kind of dhall code pretty printing at some point but i ll open a separate issue for that | 1 |
4,987 | 25,596,395,758 | IssuesEvent | 2022-12-01 16:30:04 | mozilla/foundation.mozilla.org | https://api.github.com/repos/mozilla/foundation.mozilla.org | closed | Sort Python imports consistently with `isort` | engineering Maintain unplanned | # Description
> isort is a Python utility / library to sort imports alphabetically, and automatically separated into sections and by type. It provides a command line utility, Python library and [plugins for various editors](https://github.com/pycqa/isort/wiki/isort-Plugins) to quickly sort all your imports. It requires Python 3.6+ to run but supports formatting Python 2 code too.
Without a defined sort order it can be hard to see which objects are already imported.
Defining a sort order and checking that code adheres to it manually is pretty laborious.
A linter can check the import sorting and a formatter can do the sorting for us.
`isort` combines both.
# Dev tasks
- [x] `isort` is installed and configured for the project
- [x] There is an `inv` command to check the import sorting
- [x] There is an `inv` command to apply the import sorting
- [x] CI checks that all imports are sorted correctly with `isort`
- [x] All files are sorted correctly with `isort`
# Acceptance criteria
- [x] `isort` is installed and configured for the project
- [x] There is an `inv` command to check the import sorting
- [x] There is an `inv` command to apply the import sorting
- [x] CI checks that all imports are sorted correctly with `isort`
- [x] All files are sorted correctly with `isort` | True | Sort Python imports consistently with `isort` - # Description
> isort is a Python utility / library to sort imports alphabetically, and automatically separated into sections and by type. It provides a command line utility, Python library and [plugins for various editors](https://github.com/pycqa/isort/wiki/isort-Plugins) to quickly sort all your imports. It requires Python 3.6+ to run but supports formatting Python 2 code too.
Without a defined sort order it can be hard to see which objects are already imported.
Defining a sort order and checking that code adheres to it manually is pretty laborious.
A linter can check the import sorting and a formatter can do the sorting for us.
`isort` combines both.
# Dev tasks
- [x] `isort` is installed and configured for the project
- [x] There is an `inv` command to check the import sorting
- [x] There is an `inv` command to apply the import sorting
- [x] CI checks that all imports are sorted correctly with `isort`
- [x] All files are sorted correctly with `isort`
# Acceptance criteria
- [x] `isort` is installed and configured for the project
- [x] There is an `inv` command to check the import sorting
- [x] There is an `inv` command to apply the import sorting
- [x] CI checks that all imports are sorted correctly with `isort`
- [x] All files are sorted correctly with `isort` | main | sort python imports consistently with isort description isort is a python utility library to sort imports alphabetically and automatically separated into sections and by type it provides a command line utility python library and to quickly sort all your imports it requires python to run but supports formatting python code too without a defined sort order it can be hard to see which objects are already imported defining a sort order and checking that code adheres to it manually is pretty laborious a linter can check the import sorting and a formatter can do the sorting for us isort combines both dev tasks isort is installed and configured for the project there is an inv command to check the import sorting there is an inv command to apply the import sorting ci checks that all imports are sorted correctly with isort all files are sorted correctly with isort acceptance criteria isort is installed and configured for the project there is an inv command to check the import sorting there is an inv command to apply the import sorting ci checks that all imports are sorted correctly with isort all files are sorted correctly with isort | 1 |
68,806 | 14,953,057,331 | IssuesEvent | 2021-01-26 16:14:36 | RG4421/SACM-Architecture | https://api.github.com/repos/RG4421/SACM-Architecture | opened | CVE-2020-27218 (Medium) detected in jetty-server-9.4.19.v20190610.jar | security vulnerability | ## CVE-2020-27218 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jetty-server-9.4.19.v20190610.jar</b></p></summary>
<p>The core jetty server artifact.</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: SACM-Architecture/build.gradle</p>
<p>Path to vulnerable library: 20210126161333_YSPMLO/downloadResource_VBSWSF/20210126161351/jetty-server-9.4.19.v20190610.jar,/tmp/ws-ua_20210126161333_YSPMLO/downloadResource_VBSWSF/20210126161351/jetty-server-9.4.19.v20190610.jar</p>
<p>
Dependency Hierarchy:
- :x: **jetty-server-9.4.19.v20190610.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RG4421/SACM-Architecture/commit/fa0694422097590d3033c2972d17949341886e01">fa0694422097590d3033c2972d17949341886e01</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Eclipse Jetty version 9.4.0.RC0 to 9.4.34.v20201102, 10.0.0.alpha0 to 10.0.0.beta2, and 11.0.0.alpha0 to 11.0.0.beta2, if GZIP request body inflation is enabled and requests from different clients are multiplexed onto a single connection, and if an attacker can send a request with a body that is received entirely but not consumed by the application, then a subsequent request on the same connection will see that body prepended to its body. The attacker will not see any data but may inject data into the body of the subsequent request.
<p>Publish Date: 2020-11-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27218>CVE-2020-27218</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-86wm-rrjm-8wh8">https://github.com/eclipse/jetty.project/security/advisories/GHSA-86wm-rrjm-8wh8</a></p>
<p>Release Date: 2020-11-28</p>
<p>Fix Resolution: org.eclipse.jetty:jetty-server:9.4.35.v20201120, 10.0.0.beta3, 11.0.0.beta3</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-server","packageVersion":"9.4.19.v20190610","isTransitiveDependency":false,"dependencyTree":"org.eclipse.jetty:jetty-server:9.4.19.v20190610","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-server:9.4.35.v20201120, 10.0.0.beta3, 11.0.0.beta3"}],"vulnerabilityIdentifier":"CVE-2020-27218","vulnerabilityDetails":"In Eclipse Jetty version 9.4.0.RC0 to 9.4.34.v20201102, 10.0.0.alpha0 to 10.0.0.beta2, and 11.0.0.alpha0 to 11.0.0.beta2, if GZIP request body inflation is enabled and requests from different clients are multiplexed onto a single connection, and if an attacker can send a request with a body that is received entirely but not consumed by the application, then a subsequent request on the same connection will see that body prepended to its body. The attacker will not see any data but may inject data into the body of the subsequent request.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27218","cvss3Severity":"medium","cvss3Score":"4.8","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-27218 (Medium) detected in jetty-server-9.4.19.v20190610.jar - ## CVE-2020-27218 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jetty-server-9.4.19.v20190610.jar</b></p></summary>
<p>The core jetty server artifact.</p>
<p>Library home page: <a href="http://www.eclipse.org/jetty">http://www.eclipse.org/jetty</a></p>
<p>Path to dependency file: SACM-Architecture/build.gradle</p>
<p>Path to vulnerable library: 20210126161333_YSPMLO/downloadResource_VBSWSF/20210126161351/jetty-server-9.4.19.v20190610.jar,/tmp/ws-ua_20210126161333_YSPMLO/downloadResource_VBSWSF/20210126161351/jetty-server-9.4.19.v20190610.jar</p>
<p>
Dependency Hierarchy:
- :x: **jetty-server-9.4.19.v20190610.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/RG4421/SACM-Architecture/commit/fa0694422097590d3033c2972d17949341886e01">fa0694422097590d3033c2972d17949341886e01</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Eclipse Jetty version 9.4.0.RC0 to 9.4.34.v20201102, 10.0.0.alpha0 to 10.0.0.beta2, and 11.0.0.alpha0 to 11.0.0.beta2, if GZIP request body inflation is enabled and requests from different clients are multiplexed onto a single connection, and if an attacker can send a request with a body that is received entirely but not consumed by the application, then a subsequent request on the same connection will see that body prepended to its body. The attacker will not see any data but may inject data into the body of the subsequent request.
<p>Publish Date: 2020-11-28
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27218>CVE-2020-27218</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/eclipse/jetty.project/security/advisories/GHSA-86wm-rrjm-8wh8">https://github.com/eclipse/jetty.project/security/advisories/GHSA-86wm-rrjm-8wh8</a></p>
<p>Release Date: 2020-11-28</p>
<p>Fix Resolution: org.eclipse.jetty:jetty-server:9.4.35.v20201120, 10.0.0.beta3, 11.0.0.beta3</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.eclipse.jetty","packageName":"jetty-server","packageVersion":"9.4.19.v20190610","isTransitiveDependency":false,"dependencyTree":"org.eclipse.jetty:jetty-server:9.4.19.v20190610","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.eclipse.jetty:jetty-server:9.4.35.v20201120, 10.0.0.beta3, 11.0.0.beta3"}],"vulnerabilityIdentifier":"CVE-2020-27218","vulnerabilityDetails":"In Eclipse Jetty version 9.4.0.RC0 to 9.4.34.v20201102, 10.0.0.alpha0 to 10.0.0.beta2, and 11.0.0.alpha0 to 11.0.0.beta2, if GZIP request body inflation is enabled and requests from different clients are multiplexed onto a single connection, and if an attacker can send a request with a body that is received entirely but not consumed by the application, then a subsequent request on the same connection will see that body prepended to its body. The attacker will not see any data but may inject data into the body of the subsequent request.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27218","cvss3Severity":"medium","cvss3Score":"4.8","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_main | cve medium detected in jetty server jar cve medium severity vulnerability vulnerable library jetty server jar the core jetty server artifact library home page a href path to dependency file sacm architecture build gradle path to vulnerable library yspmlo downloadresource vbswsf jetty server jar tmp ws ua yspmlo downloadresource vbswsf jetty server jar dependency hierarchy x jetty server jar vulnerable library found in head commit a href found in base branch master vulnerability details in eclipse jetty version to to and to if gzip request body inflation is enabled and requests from different clients are multiplexed onto a single connection and if an attacker can send a request with a body that is received entirely but not consumed by the application then a subsequent request on the same connection will see that body prepended to its body the attacker will not see any data but may inject data into the body of the subsequent request publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org eclipse jetty jetty server rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in eclipse jetty version to to and to if gzip request body inflation is enabled and requests from different clients are multiplexed onto a single connection and if an attacker can send a request with a body that is received entirely but not consumed by the application then a subsequent request on the same connection will see that body prepended to its body the attacker will not see any data but may inject data into the body of the subsequent request vulnerabilityurl | 0 |
110,147 | 9,437,111,051 | IssuesEvent | 2019-04-13 12:45:51 | zim-desktop-wiki/zim-desktop-wiki | https://api.github.com/repos/zim-desktop-wiki/zim-desktop-wiki | closed | Create table error | bug needs testing | Xubuntu 18.10
Zim 0.70
Menu: Insert - Table
In table-window add 3 columns and give it`s names
Then in terminal:
DEBUG: Opening dialog "Вставить таблицу"
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/zim/plugins/tableeditor.py", line 1041, in on_add_new_column
self.treeview.set_cursor_on_cell(model.get_path(newiter), self.treeview.get_column(0), start_editing=True)
TypeError: Gtk.TreeView.set_cursor_on_cell() takes exactly 5 non-keyword arguments (3 given)
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/zim/plugins/tableeditor.py", line 1041, in on_add_new_column
self.treeview.set_cursor_on_cell(model.get_path(newiter), self.treeview.get_column(0), start_editing=True)
TypeError: Gtk.TreeView.set_cursor_on_cell() takes exactly 5 non-keyword arguments (3 given)
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/zim/plugins/tableeditor.py", line 1041, in on_add_new_column
self.treeview.set_cursor_on_cell(model.get_path(newiter), self.treeview.get_column(0), start_editing=True)
TypeError: Gtk.TreeView.set_cursor_on_cell() takes exactly 5 non-keyword arguments (3 given)
Error-window:
Package: zim 0.70-ubuntu18.10.1 [origin: LP-PPA-jaap.karssenberg-zim]
PackageArchitecture: all
ProcCpuinfoMinimal:
processor : 3
vendor_id : GenuineIntel
cpu family : 6
model : 30
model name : Intel(R) Core(TM) i5 CPU 760 @ 2.80GHz
stepping : 5
microcode : 0xa
cpu MHz : 2707.123
cache size : 8192 KB
physical id : 0
siblings : 4
core id : 3
cpu cores : 4
apicid : 6
initial apicid : 6
fpu : yes
fpu_exception : yes
cpuid level : 11
wp : yes
flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni dtes64 monitor ds_cpl vmx smx est tm2 ssse3 cx16 xtpr pdcm sse4_1 sse4_2 popcnt lahf_lm pti ssbd ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid dtherm ida flush_l1d
bugs : cpu_meltdown spectre_v1 spectre_v2 spec_store_bypass l1tf
bogomips : 5619.54
clflush size : 64
cache_alignment : 64
address sizes : 36 bits physical, 48 bits virtual
power management:
ProcVersionSignature: Ubuntu 4.18.0-17.18-generic 4.18.20
Python3Details: /usr/bin/python3.6, Python 3.6.7, python3-minimal, 3.6.7-1~18.10
PythonDetails: /usr/bin/python2.7, Python 2.7.15+, python-minimal, 2.7.15-3
SourcePackage: zim
Tags: third-party-packages cosmic
Title: zim crashed with TypeError in on_add_new_column(): Gtk.TreeView.set_cursor_on_cell() takes exactly 5 non-keyword arguments (3 given)
Uname: Linux 4.18.0-17-generic x86_64
UnreportableReason: Этот пакет не является официальным пакетом Ubuntu. Пожалуйста, удалите все сторонние пакеты и попробуйте снова.
UpgradeStatus: Upgraded to cosmic on 2018-11-13 (145 days ago)
_MarkForUpload: True | 1.0 | Create table error - Xubuntu 18.10
Zim 0.70
Menu: Insert - Table
In table-window add 3 columns and give it`s names
Then in terminal:
DEBUG: Opening dialog "Вставить таблицу"
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/zim/plugins/tableeditor.py", line 1041, in on_add_new_column
self.treeview.set_cursor_on_cell(model.get_path(newiter), self.treeview.get_column(0), start_editing=True)
TypeError: Gtk.TreeView.set_cursor_on_cell() takes exactly 5 non-keyword arguments (3 given)
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/zim/plugins/tableeditor.py", line 1041, in on_add_new_column
self.treeview.set_cursor_on_cell(model.get_path(newiter), self.treeview.get_column(0), start_editing=True)
TypeError: Gtk.TreeView.set_cursor_on_cell() takes exactly 5 non-keyword arguments (3 given)
Traceback (most recent call last):
File "/usr/lib/python3/dist-packages/zim/plugins/tableeditor.py", line 1041, in on_add_new_column
self.treeview.set_cursor_on_cell(model.get_path(newiter), self.treeview.get_column(0), start_editing=True)
TypeError: Gtk.TreeView.set_cursor_on_cell() takes exactly 5 non-keyword arguments (3 given)
Error-window:
Package: zim 0.70-ubuntu18.10.1 [origin: LP-PPA-jaap.karssenberg-zim]
PackageArchitecture: all
ProcCpuinfoMinimal:
processor : 3
vendor_id : GenuineIntel
cpu family : 6
model : 30
model name : Intel(R) Core(TM) i5 CPU 760 @ 2.80GHz
stepping : 5
microcode : 0xa
cpu MHz : 2707.123
cache size : 8192 KB
physical id : 0
siblings : 4
core id : 3
cpu cores : 4
apicid : 6
initial apicid : 6
fpu : yes
fpu_exception : yes
cpuid level : 11
wp : yes
flags : fpu vme de pse tsc msr pae mce cx8 apic sep mtrr pge mca cmov pat pse36 clflush dts acpi mmx fxsr sse sse2 ss ht tm pbe syscall nx rdtscp lm constant_tsc arch_perfmon pebs bts rep_good nopl xtopology nonstop_tsc cpuid aperfmperf pni dtes64 monitor ds_cpl vmx smx est tm2 ssse3 cx16 xtpr pdcm sse4_1 sse4_2 popcnt lahf_lm pti ssbd ibrs ibpb stibp tpr_shadow vnmi flexpriority ept vpid dtherm ida flush_l1d
bugs : cpu_meltdown spectre_v1 spectre_v2 spec_store_bypass l1tf
bogomips : 5619.54
clflush size : 64
cache_alignment : 64
address sizes : 36 bits physical, 48 bits virtual
power management:
ProcVersionSignature: Ubuntu 4.18.0-17.18-generic 4.18.20
Python3Details: /usr/bin/python3.6, Python 3.6.7, python3-minimal, 3.6.7-1~18.10
PythonDetails: /usr/bin/python2.7, Python 2.7.15+, python-minimal, 2.7.15-3
SourcePackage: zim
Tags: third-party-packages cosmic
Title: zim crashed with TypeError in on_add_new_column(): Gtk.TreeView.set_cursor_on_cell() takes exactly 5 non-keyword arguments (3 given)
Uname: Linux 4.18.0-17-generic x86_64
UnreportableReason: Этот пакет не является официальным пакетом Ubuntu. Пожалуйста, удалите все сторонние пакеты и попробуйте снова.
UpgradeStatus: Upgraded to cosmic on 2018-11-13 (145 days ago)
_MarkForUpload: True | non_main | create table error xubuntu zim menu insert table in table window add columns and give it s names then in terminal debug opening dialog вставить таблицу traceback most recent call last file usr lib dist packages zim plugins tableeditor py line in on add new column self treeview set cursor on cell model get path newiter self treeview get column start editing true typeerror gtk treeview set cursor on cell takes exactly non keyword arguments given traceback most recent call last file usr lib dist packages zim plugins tableeditor py line in on add new column self treeview set cursor on cell model get path newiter self treeview get column start editing true typeerror gtk treeview set cursor on cell takes exactly non keyword arguments given traceback most recent call last file usr lib dist packages zim plugins tableeditor py line in on add new column self treeview set cursor on cell model get path newiter self treeview get column start editing true typeerror gtk treeview set cursor on cell takes exactly non keyword arguments given error window package zim packagearchitecture all proccpuinfominimal processor vendor id genuineintel cpu family model model name intel r core tm cpu stepping microcode cpu mhz cache size kb physical id siblings core id cpu cores apicid initial apicid fpu yes fpu exception yes cpuid level wp yes flags fpu vme de pse tsc msr pae mce apic sep mtrr pge mca cmov pat clflush dts acpi mmx fxsr sse ss ht tm pbe syscall nx rdtscp lm constant tsc arch perfmon pebs bts rep good nopl xtopology nonstop tsc cpuid aperfmperf pni monitor ds cpl vmx smx est xtpr pdcm popcnt lahf lm pti ssbd ibrs ibpb stibp tpr shadow vnmi flexpriority ept vpid dtherm ida flush bugs cpu meltdown spectre spectre spec store bypass bogomips clflush size cache alignment address sizes bits physical bits virtual power management procversionsignature ubuntu generic usr bin python minimal pythondetails usr bin python python minimal sourcepackage zim tags third party packages cosmic title zim crashed with typeerror in on add new column gtk treeview set cursor on cell takes exactly non keyword arguments given uname linux generic unreportablereason этот пакет не является официальным пакетом ubuntu пожалуйста удалите все сторонние пакеты и попробуйте снова upgradestatus upgraded to cosmic on days ago markforupload true | 0 |
618 | 3,851,660,157 | IssuesEvent | 2016-04-06 03:43:01 | MovingBlocks/Terasology | https://api.github.com/repos/MovingBlocks/Terasology | closed | When life-cycle events called for block entities - getting a block from that chunk returns "Air" | API Architecture Bug | When block entities have their lifecycle events called upon loading chunk, it's impossible to get any block from that chunk.
Blocks should probably be set in the WorldProvider, before the life-cycle events are called.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/21234166-when-life-cycle-events-called-for-block-entities-getting-a-block-from-that-chunk-returns-air?utm_campaign=plugin&utm_content=tracker%2F5279&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F5279&utm_medium=issues&utm_source=github).
</bountysource-plugin> | 1.0 | When life-cycle events called for block entities - getting a block from that chunk returns "Air" - When block entities have their lifecycle events called upon loading chunk, it's impossible to get any block from that chunk.
Blocks should probably be set in the WorldProvider, before the life-cycle events are called.
<bountysource-plugin>
---
Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/21234166-when-life-cycle-events-called-for-block-entities-getting-a-block-from-that-chunk-returns-air?utm_campaign=plugin&utm_content=tracker%2F5279&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F5279&utm_medium=issues&utm_source=github).
</bountysource-plugin> | non_main | when life cycle events called for block entities getting a block from that chunk returns air when block entities have their lifecycle events called upon loading chunk it s impossible to get any block from that chunk blocks should probably be set in the worldprovider before the life cycle events are called want to back this issue we accept bounties via | 0 |
132,168 | 12,499,828,819 | IssuesEvent | 2020-06-01 20:55:46 | ColdIV/math-tool | https://api.github.com/repos/ColdIV/math-tool | closed | UML | documentation help wanted | All classes should be displayed in an UML diagram.
If someone wants to do it in advance, you're welcome!
That's how one _should_ do it...
If not, then we can just build it after all classes were finished.
That's how I would do it.
Our classes:
- [x] ./gui/CalculatorInput.h
- [x] ./gui/MenuInput.h
- [x] ./gui/GeometryApp.h
- [x] ./gui/MenuApp.h
- [x] ./gui/LineInput.h
- [x] ./gui/FunctionPlotterInput.h
- [x] ./gui/FunctionPlotterApp.h
- [x] ./gui/Graph.h
- [x] ./gui/App.h
- [x] ./gui/CalculatorApp.h
- [x] ./gui/GeometryInput.h
- [x] ./gui/Widget.h
- [x] ./gui/TextOutput.h
- [x] ./backend/Line.h
- [x] ./backend/Circle.h
- [x] ./backend/TokenStream.h
- [x] ./backend/Point.h
- [x] ./backend/Object.h
- [x] ./backend/Triangle.h
- [x] ./backend/Rectangle.h
- [x] ./backend/Parser.h
- [x] ./backend/FunctionPlotter.h
- [x] ./backend/Square.h
- [x] ./backend/Calculator.h
Our structs:
- [x] ./backend/error.h
- [x] ./backend/Token.h | 1.0 | UML - All classes should be displayed in an UML diagram.
If someone wants to do it in advance, you're welcome!
That's how one _should_ do it...
If not, then we can just build it after all classes were finished.
That's how I would do it.
Our classes:
- [x] ./gui/CalculatorInput.h
- [x] ./gui/MenuInput.h
- [x] ./gui/GeometryApp.h
- [x] ./gui/MenuApp.h
- [x] ./gui/LineInput.h
- [x] ./gui/FunctionPlotterInput.h
- [x] ./gui/FunctionPlotterApp.h
- [x] ./gui/Graph.h
- [x] ./gui/App.h
- [x] ./gui/CalculatorApp.h
- [x] ./gui/GeometryInput.h
- [x] ./gui/Widget.h
- [x] ./gui/TextOutput.h
- [x] ./backend/Line.h
- [x] ./backend/Circle.h
- [x] ./backend/TokenStream.h
- [x] ./backend/Point.h
- [x] ./backend/Object.h
- [x] ./backend/Triangle.h
- [x] ./backend/Rectangle.h
- [x] ./backend/Parser.h
- [x] ./backend/FunctionPlotter.h
- [x] ./backend/Square.h
- [x] ./backend/Calculator.h
Our structs:
- [x] ./backend/error.h
- [x] ./backend/Token.h | non_main | uml all classes should be displayed in an uml diagram if someone wants to do it in advance you re welcome that s how one should do it if not then we can just build it after all classes were finished that s how i would do it our classes gui calculatorinput h gui menuinput h gui geometryapp h gui menuapp h gui lineinput h gui functionplotterinput h gui functionplotterapp h gui graph h gui app h gui calculatorapp h gui geometryinput h gui widget h gui textoutput h backend line h backend circle h backend tokenstream h backend point h backend object h backend triangle h backend rectangle h backend parser h backend functionplotter h backend square h backend calculator h our structs backend error h backend token h | 0 |
5,405 | 27,123,454,522 | IssuesEvent | 2023-02-16 01:55:31 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | opened | `playsound` should have better defaults | Maintainability/Hinders improvements | `playsound` is a very common proc with bad defaults. Nearly every consumer wants `vary = TRUE` (usually copied and pasted as just "TRUE", making it hard to read), and a "normal" volume, though what volume you should pick isn't clear or readable. This makes the code noisier to read and harder to write. This may also apply to `playsound_local`.
We should do two things.
## 1. Make `vary` TRUE by default
We should statically (please not regex) find every current usage of `playsound` that does NOT set `vary = TRUE`, and make them explicitly `vary = FALSE`.
## 2. Introduce human readable volumes, and pick a sane one as the default
I don't know what the values should be, but I'm thinking a straight forward `NORMAL`, `QUIET`, `LOUD`, with a step in between for something like `HIGHER_NORMAL` or something. You should have a nice variety but it should be easy for us to see what the volume is. Anything that truly wants a custom value for flavor can do so, but should likely do them as some variation of these (like `NORMAL * 2` or something 🤷♀️ ) | True | `playsound` should have better defaults - `playsound` is a very common proc with bad defaults. Nearly every consumer wants `vary = TRUE` (usually copied and pasted as just "TRUE", making it hard to read), and a "normal" volume, though what volume you should pick isn't clear or readable. This makes the code noisier to read and harder to write. This may also apply to `playsound_local`.
We should do two things.
## 1. Make `vary` TRUE by default
We should statically (please not regex) find every current usage of `playsound` that does NOT set `vary = TRUE`, and make them explicitly `vary = FALSE`.
## 2. Introduce human readable volumes, and pick a sane one as the default
I don't know what the values should be, but I'm thinking a straight forward `NORMAL`, `QUIET`, `LOUD`, with a step in between for something like `HIGHER_NORMAL` or something. You should have a nice variety but it should be easy for us to see what the volume is. Anything that truly wants a custom value for flavor can do so, but should likely do them as some variation of these (like `NORMAL * 2` or something 🤷♀️ ) | main | playsound should have better defaults playsound is a very common proc with bad defaults nearly every consumer wants vary true usually copied and pasted as just true making it hard to read and a normal volume though what volume you should pick isn t clear or readable this makes the code noisier to read and harder to write this may also apply to playsound local we should do two things make vary true by default we should statically please not regex find every current usage of playsound that does not set vary true and make them explicitly vary false introduce human readable volumes and pick a sane one as the default i don t know what the values should be but i m thinking a straight forward normal quiet loud with a step in between for something like higher normal or something you should have a nice variety but it should be easy for us to see what the volume is anything that truly wants a custom value for flavor can do so but should likely do them as some variation of these like normal or something 🤷♀️ | 1 |
3,036 | 11,258,967,893 | IssuesEvent | 2020-01-13 06:53:49 | microsoft/DirectXTex | https://api.github.com/repos/microsoft/DirectXTex | closed | Retire support for VS 2015 | maintainence | In 2020, I plan to retire support for VS 2015. The following projects will be removed, and the NuGet ``directxtex_desktop_2015`` package will be deprecated in favor of one built with VS 2017 and/or VS 2019:
DirectXTex_Desktop_2015
DirectXTex_Desktop_2015_Win10
DirectXTex_Windows10_2015
DirectXTex_XboxOneXDK_2015
DDSView_Desktop_2015
Texassemble_Desktop_2015
Texconv_Desktop_2015
texdiag_Desktop_2015
Please put any requests for continued support for one or more of these here. | True | Retire support for VS 2015 - In 2020, I plan to retire support for VS 2015. The following projects will be removed, and the NuGet ``directxtex_desktop_2015`` package will be deprecated in favor of one built with VS 2017 and/or VS 2019:
DirectXTex_Desktop_2015
DirectXTex_Desktop_2015_Win10
DirectXTex_Windows10_2015
DirectXTex_XboxOneXDK_2015
DDSView_Desktop_2015
Texassemble_Desktop_2015
Texconv_Desktop_2015
texdiag_Desktop_2015
Please put any requests for continued support for one or more of these here. | main | retire support for vs in i plan to retire support for vs the following projects will be removed and the nuget directxtex desktop package will be deprecated in favor of one built with vs and or vs directxtex desktop directxtex desktop directxtex directxtex xboxonexdk ddsview desktop texassemble desktop texconv desktop texdiag desktop please put any requests for continued support for one or more of these here | 1 |
119,204 | 25,490,282,880 | IssuesEvent | 2022-11-27 00:43:29 | wraiford/ibgib | https://api.github.com/repos/wraiford/ibgib | reopened | Resources for learning ibgib tech stack | good first issue question learning vscode ts-js css-scss html ng-ionic node-npm-nvm git | ## don't learn ibgib itself?
ibgib is a technology being built for distributed collaboration, distributed computation & distributed learning (all the same thing) in a new way. In today's parlance, this equates to web3 technology. As such it is admittedly a speculative time investment.
But it's being built with technologies that are not speculative, and anyone should be able to transfer this learning to your future programming domains whatever they may be.
## tech stack
* [vscode](https://code.visualstudio.com/)
* [node](https://nodejs.org/en/) (& [npm](https://www.npmjs.com/), [nvm](https://github.com/nvm-sh/nvm))
* [TypeScript](https://www.typescriptlang.org/) (& [JavaScript](https://www.javascript.com/))
* [Ionic](https://ionicframework.com/) (& [Angular](https://angular.io/))
* [SCSS](https://sass-lang.com/documentation/syntax#scss) (& [CSS3](https://developer.mozilla.org/en-US/docs/Web/CSS))
* [git](https://git-scm.com/) + [GitHub](https://github.com/) + [GitLab](https://gitlab.com/)
* [web-ext](https://github.com/mozilla/web-ext)
We'll use this issue tracker similarly to a forum, wiki, or slack/irc channel for learning collaboration. I'll create labels for each of the above technologies and we'll get by until we can [dogfood](https://en.wikipedia.org/wiki/Eating_your_own_dog_food) this learning with ibgib itself.
## so...
Just ask questions, post them here in this issue and/or a new issue with other labels and we'll see what we can do. | 1.0 | Resources for learning ibgib tech stack - ## don't learn ibgib itself?
ibgib is a technology being built for distributed collaboration, distributed computation & distributed learning (all the same thing) in a new way. In today's parlance, this equates to web3 technology. As such it is admittedly a speculative time investment.
But it's being built with technologies that are not speculative, and anyone should be able to transfer this learning to your future programming domains whatever they may be.
## tech stack
* [vscode](https://code.visualstudio.com/)
* [node](https://nodejs.org/en/) (& [npm](https://www.npmjs.com/), [nvm](https://github.com/nvm-sh/nvm))
* [TypeScript](https://www.typescriptlang.org/) (& [JavaScript](https://www.javascript.com/))
* [Ionic](https://ionicframework.com/) (& [Angular](https://angular.io/))
* [SCSS](https://sass-lang.com/documentation/syntax#scss) (& [CSS3](https://developer.mozilla.org/en-US/docs/Web/CSS))
* [git](https://git-scm.com/) + [GitHub](https://github.com/) + [GitLab](https://gitlab.com/)
* [web-ext](https://github.com/mozilla/web-ext)
We'll use this issue tracker similarly to a forum, wiki, or slack/irc channel for learning collaboration. I'll create labels for each of the above technologies and we'll get by until we can [dogfood](https://en.wikipedia.org/wiki/Eating_your_own_dog_food) this learning with ibgib itself.
## so...
Just ask questions, post them here in this issue and/or a new issue with other labels and we'll see what we can do. | non_main | resources for learning ibgib tech stack don t learn ibgib itself ibgib is a technology being built for distributed collaboration distributed computation distributed learning all the same thing in a new way in today s parlance this equates to technology as such it is admittedly a speculative time investment but it s being built with technologies that are not speculative and anyone should be able to transfer this learning to your future programming domains whatever they may be tech stack we ll use this issue tracker similarly to a forum wiki or slack irc channel for learning collaboration i ll create labels for each of the above technologies and we ll get by until we can this learning with ibgib itself so just ask questions post them here in this issue and or a new issue with other labels and we ll see what we can do | 0 |
3,143 | 12,057,923,338 | IssuesEvent | 2020-04-15 16:35:39 | short-d/short | https://api.github.com/repos/short-d/short | closed | [Refactor] Replace hard-coded auth token valid duration with value from env variable | maintainability p0 | Currently, auth token valid duration is hard-coded in [wire.go](https://github.com/short-d/short/blob/master/backend/dep/wire.go). This value should be extracted from environmental variable to avoid unnecessary code changes in the future. | True | [Refactor] Replace hard-coded auth token valid duration with value from env variable - Currently, auth token valid duration is hard-coded in [wire.go](https://github.com/short-d/short/blob/master/backend/dep/wire.go). This value should be extracted from environmental variable to avoid unnecessary code changes in the future. | main | replace hard coded auth token valid duration with value from env variable currently auth token valid duration is hard coded in this value should be extracted from environmental variable to avoid unnecessary code changes in the future | 1 |
85,150 | 7,962,617,885 | IssuesEvent | 2018-07-13 14:50:39 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: kv95/nodes=1 failed on release-2.0 | A-testing C-test-failure O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/62ab23d62538ebdccde0d1e73a2af241b363b1ce
Parameters:
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=673547&tab=buildLog
```
cluster.go:699,kv.go:27,kv.go:62: /home/agent/work/.go/bin/roachprod run teamcity-673547-kv95-nodes-1:1-2 -- sudo umount /mnt/data1 ; sudo mount -o discard,defaults,nobarrier /dev/disk/by-id/google-local-ssd-0 /mnt/data1: exit status 1
``` | 2.0 | roachtest: kv95/nodes=1 failed on release-2.0 - SHA: https://github.com/cockroachdb/cockroach/commits/62ab23d62538ebdccde0d1e73a2af241b363b1ce
Parameters:
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=673547&tab=buildLog
```
cluster.go:699,kv.go:27,kv.go:62: /home/agent/work/.go/bin/roachprod run teamcity-673547-kv95-nodes-1:1-2 -- sudo umount /mnt/data1 ; sudo mount -o discard,defaults,nobarrier /dev/disk/by-id/google-local-ssd-0 /mnt/data1: exit status 1
``` | non_main | roachtest nodes failed on release sha parameters failed test cluster go kv go kv go home agent work go bin roachprod run teamcity nodes sudo umount mnt sudo mount o discard defaults nobarrier dev disk by id google local ssd mnt exit status | 0 |
907 | 4,569,335,430 | IssuesEvent | 2016-09-15 16:55:29 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | cloudformation module fails with | affects_2.2 aws bug_report cloud waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
cloudformation
##### ANSIBLE VERSION
```
ansible 2.2.0
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
After a successful stack operation the module always fails when calling `exit_json(**results)` due to a bad result variable type. The cloudformation module tries to include stack events in module output. The stack events are an `itertools.imap` iterator which ansible doesn't know how to deal with when calling `exit_json()` and fails with a `Value of unknown type: <type 'itertools.imap'>`. The fix here can be as simple as wrapping the stack events variable in a call to `list()`. The more general fix would be to make ansible's `remove_values()` function accept `itertools.imap` objects.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
**ansible-cfn-bug.yml**
```yaml
---
- hosts: localhost
tasks:
- name: deploy stack
cloudformation:
stack_name: ansible-cfn-bug
state: present
template: stack.json
```
**stack.json**
```json
{
"Resources": {
"S3B3QFCX": {
"Type": "AWS::S3::Bucket",
"Properties": {}
}
}
}
```
```bash
ansible-playbook -vvvv -c local ansible-cfn-bug.yml
```
##### EXPECTED RESULTS
Stack should be deployed. Deploy task should successfully complete.
##### ACTUAL RESULTS
Deploy task fails for reasons listed above.
```
TASK [deploy stack] ************************************************************
task path: [redacted]/ansible-cfn-bug.yml:4
Using module file [redacted]/venv/local/lib/python2.7/site-packages/ansible/modules/core/cloud/amazon/cloudformation.py
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: [redacted]
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975 `" && echo ansible-tmp-1473951676.35-213586073691975="` echo $HOME/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975 `" ) && sleep 0'
<127.0.0.1> PUT /tmp/tmpGPzHre TO [redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/cloudformation.py
<127.0.0.1> EXEC /bin/sh -c 'chmod u+x [redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/ [redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/cloudformation.py && sleep 0'
<127.0.0.1> EXEC /bin/sh -c '[redacted]/venv/bin/python2 [redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/cloudformation.py; rm -rf "[redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/" > /dev/null 2>&1 && sleep 0'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_rX9aA1/ansible_module_cloudformation.py", line 402, in <module>
main()
File "/tmp/ansible_rX9aA1/ansible_module_cloudformation.py", line 395, in main
module.exit_json(**result)
File "/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py", line 1798, in exit_json
File "/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py", line 387, in remove_values
File "/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py", line 387, in <genexpr>
File "/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py", line 398, in remove_values
TypeError: Value of unknown type: <type 'itertools.imap'>, <itertools.imap object at 0x7f300e408950>
fatal: [localhost]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "cloudformation"
},
"module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_rX9aA1/ansible_module_cloudformation.py\", line 402, in <module>\n main()\n File \"/tmp/ansible_rX9aA1/ansible_module_cloudformation.py\", line 395, in main\n module.exit_json(**result)\n File \"/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py\", line 1798, in exit_json\n File \"/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py\", line 387, in remove_values\n File \"/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py\", line 387, in <genexpr>\n File \"/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py\", line 398, in remove_values\nTypeError: Value of unknown type: <type 'itertools.imap'>, <itertools.imap object at 0x7f300e408950>\n",
```
| True | cloudformation module fails with - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
cloudformation
##### ANSIBLE VERSION
```
ansible 2.2.0
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
After a successful stack operation the module always fails when calling `exit_json(**results)` due to a bad result variable type. The cloudformation module tries to include stack events in module output. The stack events are an `itertools.imap` iterator which ansible doesn't know how to deal with when calling `exit_json()` and fails with a `Value of unknown type: <type 'itertools.imap'>`. The fix here can be as simple as wrapping the stack events variable in a call to `list()`. The more general fix would be to make ansible's `remove_values()` function accept `itertools.imap` objects.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
**ansible-cfn-bug.yml**
```yaml
---
- hosts: localhost
tasks:
- name: deploy stack
cloudformation:
stack_name: ansible-cfn-bug
state: present
template: stack.json
```
**stack.json**
```json
{
"Resources": {
"S3B3QFCX": {
"Type": "AWS::S3::Bucket",
"Properties": {}
}
}
}
```
```bash
ansible-playbook -vvvv -c local ansible-cfn-bug.yml
```
##### EXPECTED RESULTS
Stack should be deployed. Deploy task should successfully complete.
##### ACTUAL RESULTS
Deploy task fails for reasons listed above.
```
TASK [deploy stack] ************************************************************
task path: [redacted]/ansible-cfn-bug.yml:4
Using module file [redacted]/venv/local/lib/python2.7/site-packages/ansible/modules/core/cloud/amazon/cloudformation.py
<127.0.0.1> ESTABLISH LOCAL CONNECTION FOR USER: [redacted]
<127.0.0.1> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975 `" && echo ansible-tmp-1473951676.35-213586073691975="` echo $HOME/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975 `" ) && sleep 0'
<127.0.0.1> PUT /tmp/tmpGPzHre TO [redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/cloudformation.py
<127.0.0.1> EXEC /bin/sh -c 'chmod u+x [redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/ [redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/cloudformation.py && sleep 0'
<127.0.0.1> EXEC /bin/sh -c '[redacted]/venv/bin/python2 [redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/cloudformation.py; rm -rf "[redacted]/.ansible/tmp/ansible-tmp-1473951676.35-213586073691975/" > /dev/null 2>&1 && sleep 0'
An exception occurred during task execution. The full traceback is:
Traceback (most recent call last):
File "/tmp/ansible_rX9aA1/ansible_module_cloudformation.py", line 402, in <module>
main()
File "/tmp/ansible_rX9aA1/ansible_module_cloudformation.py", line 395, in main
module.exit_json(**result)
File "/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py", line 1798, in exit_json
File "/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py", line 387, in remove_values
File "/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py", line 387, in <genexpr>
File "/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py", line 398, in remove_values
TypeError: Value of unknown type: <type 'itertools.imap'>, <itertools.imap object at 0x7f300e408950>
fatal: [localhost]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "cloudformation"
},
"module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_rX9aA1/ansible_module_cloudformation.py\", line 402, in <module>\n main()\n File \"/tmp/ansible_rX9aA1/ansible_module_cloudformation.py\", line 395, in main\n module.exit_json(**result)\n File \"/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py\", line 1798, in exit_json\n File \"/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py\", line 387, in remove_values\n File \"/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py\", line 387, in <genexpr>\n File \"/tmp/ansible_rX9aA1/ansible_modlib.zip/ansible/module_utils/basic.py\", line 398, in remove_values\nTypeError: Value of unknown type: <type 'itertools.imap'>, <itertools.imap object at 0x7f300e408950>\n",
```
| main | cloudformation module fails with issue type bug report component name cloudformation ansible version ansible configuration n a os environment n a summary after a successful stack operation the module always fails when calling exit json results due to a bad result variable type the cloudformation module tries to include stack events in module output the stack events are an itertools imap iterator which ansible doesn t know how to deal with when calling exit json and fails with a value of unknown type the fix here can be as simple as wrapping the stack events variable in a call to list the more general fix would be to make ansible s remove values function accept itertools imap objects steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used ansible cfn bug yml yaml hosts localhost tasks name deploy stack cloudformation stack name ansible cfn bug state present template stack json stack json json resources type aws bucket properties bash ansible playbook vvvv c local ansible cfn bug yml expected results stack should be deployed deploy task should successfully complete actual results deploy task fails for reasons listed above task task path ansible cfn bug yml using module file venv local lib site packages ansible modules core cloud amazon cloudformation py establish local connection for user exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpgpzhre to ansible tmp ansible tmp cloudformation py exec bin sh c chmod u x ansible tmp ansible tmp ansible tmp ansible tmp cloudformation py sleep exec bin sh c venv bin ansible tmp ansible tmp cloudformation py rm rf ansible tmp ansible tmp dev null sleep an exception occurred during task execution the full traceback is traceback most recent call last file tmp ansible ansible module cloudformation py line in main file tmp ansible ansible module cloudformation py line in main module exit json result file tmp ansible ansible modlib zip ansible module utils basic py line in exit json file tmp ansible ansible modlib zip ansible module utils basic py line in remove values file tmp ansible ansible modlib zip ansible module utils basic py line in file tmp ansible ansible modlib zip ansible module utils basic py line in remove values typeerror value of unknown type fatal failed changed false failed true invocation module name cloudformation module stderr traceback most recent call last n file tmp ansible ansible module cloudformation py line in n main n file tmp ansible ansible module cloudformation py line in main n module exit json result n file tmp ansible ansible modlib zip ansible module utils basic py line in exit json n file tmp ansible ansible modlib zip ansible module utils basic py line in remove values n file tmp ansible ansible modlib zip ansible module utils basic py line in n file tmp ansible ansible modlib zip ansible module utils basic py line in remove values ntypeerror value of unknown type n | 1 |
908 | 4,577,132,249 | IssuesEvent | 2016-09-17 01:43:50 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Guest Customization template in vSphere_Guest module | affects_2.1 cloud feature_idea vmware waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
vsphere_guest
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
2.1.1
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
##### SUMMARY
<!--- Explain the problem briefly -->
Does the vsphere_guest module support applying a guest customization template while deploying a VM from template?
| True | Guest Customization template in vSphere_Guest module - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
vsphere_guest
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
2.1.1
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
##### SUMMARY
<!--- Explain the problem briefly -->
Does the vsphere_guest module support applying a guest customization template while deploying a VM from template?
| main | guest customization template in vsphere guest module issue type feature idea component name vsphere guest ansible version configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific summary does the vsphere guest module support applying a guest customization template while deploying a vm from template | 1 |
270,156 | 20,593,353,845 | IssuesEvent | 2022-03-05 05:11:13 | atharva3428/Titanic_EDA | https://api.github.com/repos/atharva3428/Titanic_EDA | opened | Improvement in documentation | documentation | - [ ] Documentation of ipynb file
- [ ] Documentation of Readme file | 1.0 | Improvement in documentation - - [ ] Documentation of ipynb file
- [ ] Documentation of Readme file | non_main | improvement in documentation documentation of ipynb file documentation of readme file | 0 |
177,987 | 29,448,107,393 | IssuesEvent | 2023-06-01 00:25:22 | status-im/status-mobile | https://api.github.com/repos/status-im/status-mobile | closed | Improve UX when you tap "Mark all notifications as read" in AC and there is no unread notifications | design feature UI/UX improvements Activity center | # Feature Issue
Improve UX in Activity Center, when the user taps on `Mark all notifications as read` and there is *no unread notifications to mark as read*.

### Possible Improvements
- Disable the button (`action-drawer`) component. We need designs for **disabled** state for `action-drawer` component.
### Notes
Currently, we display a JS alert with a text `"No unread notifications to mark as read"`, if there is no notifications to mark as read.
| 1.0 | Improve UX when you tap "Mark all notifications as read" in AC and there is no unread notifications - # Feature Issue
Improve UX in Activity Center, when the user taps on `Mark all notifications as read` and there is *no unread notifications to mark as read*.

### Possible Improvements
- Disable the button (`action-drawer`) component. We need designs for **disabled** state for `action-drawer` component.
### Notes
Currently, we display a JS alert with a text `"No unread notifications to mark as read"`, if there is no notifications to mark as read.
| non_main | improve ux when you tap mark all notifications as read in ac and there is no unread notifications feature issue improve ux in activity center when the user taps on mark all notifications as read and there is no unread notifications to mark as read possible improvements disable the button action drawer component we need designs for disabled state for action drawer component notes currently we display a js alert with a text no unread notifications to mark as read if there is no notifications to mark as read | 0 |
212,687 | 16,473,230,060 | IssuesEvent | 2021-05-23 20:43:18 | angular/angular | https://api.github.com/repos/angular/angular | closed | Spy variable not initialised in Testing documentation, section "#spy-stub" | P4 comp: docs comp: testing freq2: medium type: bug/fix | ## I'm submitting a...
<pre><code>
[ ] Regression (a behavior that used to work and stopped working in a new release)
[ ] Bug report <!-- Please search GitHub for a similar issue or PR before submitting -->
[ ] Feature request
[x] Documentation issue or request
[ ] Support request => Please do not submit support request here, instead see https://github.com/angular/angular/blob/master/CONTRIBUTING.md#question
</code></pre>
## Current behavior
<!-- Describe how the issue manifests. -->
In Testing, at https://angular.io/guide/testing#spy-stub the 'spy' variable isn't initialised in the code example.
## Expected behavior
<!-- Describe what the desired behavior would be. -->
The 'spy' variable should be initialised to prevent compile errors.
## What is the motivation / use case for changing the behavior?
While it may seem obvious to many, for new coders being able to study working code is absolutely crucial and this code won't work from the snippet. | 1.0 | Spy variable not initialised in Testing documentation, section "#spy-stub" - ## I'm submitting a...
<pre><code>
[ ] Regression (a behavior that used to work and stopped working in a new release)
[ ] Bug report <!-- Please search GitHub for a similar issue or PR before submitting -->
[ ] Feature request
[x] Documentation issue or request
[ ] Support request => Please do not submit support request here, instead see https://github.com/angular/angular/blob/master/CONTRIBUTING.md#question
</code></pre>
## Current behavior
<!-- Describe how the issue manifests. -->
In Testing, at https://angular.io/guide/testing#spy-stub the 'spy' variable isn't initialised in the code example.
## Expected behavior
<!-- Describe what the desired behavior would be. -->
The 'spy' variable should be initialised to prevent compile errors.
## What is the motivation / use case for changing the behavior?
While it may seem obvious to many, for new coders being able to study working code is absolutely crucial and this code won't work from the snippet. | non_main | spy variable not initialised in testing documentation section spy stub i m submitting a regression a behavior that used to work and stopped working in a new release bug report feature request documentation issue or request support request please do not submit support request here instead see current behavior in testing at the spy variable isn t initialised in the code example expected behavior the spy variable should be initialised to prevent compile errors what is the motivation use case for changing the behavior while it may seem obvious to many for new coders being able to study working code is absolutely crucial and this code won t work from the snippet | 0 |
340,502 | 24,657,135,552 | IssuesEvent | 2022-10-18 01:22:06 | Paulac23/DW-FS-2022 | https://api.github.com/repos/Paulac23/DW-FS-2022 | opened | Creacion diagrama de clase UML | documentation | - [ ] Planificación de casos de usos y análisis de los casos
- [ ] Creacion de diagrama | 1.0 | Creacion diagrama de clase UML - - [ ] Planificación de casos de usos y análisis de los casos
- [ ] Creacion de diagrama | non_main | creacion diagrama de clase uml planificación de casos de usos y análisis de los casos creacion de diagrama | 0 |
127,318 | 12,312,215,188 | IssuesEvent | 2020-05-12 13:37:31 | virtualsatellite/VirtualSatellite4-FDIR | https://api.github.com/repos/virtualsatellite/VirtualSatellite4-FDIR | closed | Write manual chapter on FDIR Report Generation | documentation | Write manual chapter on FDIR Report generation. | 1.0 | Write manual chapter on FDIR Report Generation - Write manual chapter on FDIR Report generation. | non_main | write manual chapter on fdir report generation write manual chapter on fdir report generation | 0 |
1,186 | 5,102,867,002 | IssuesEvent | 2017-01-04 19:37:16 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | ec2_group.py and --diff | affects_2.1 aws cloud feature_idea waiting_on_maintainer | ##### Issue Type:
- Feature Idea
##### Plugin Name:
ec2_group.py
##### Ansible Version:
```
ansible 2.1.0 (devel 33f96edcd0) last updated 2016/03/10 14:52:58 (GMT -400)
lib/ansible/modules/core: (detached HEAD c86a0ef84a) last updated 2016/03/10 14:53:03 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 33a557cc59) last updated 2016/03/10 14:53:03 (GMT -400)
config file =
configured module search path = Default w/o overrides
```
##### Ansible Configuration:
none
##### Environment:
N/A
##### Summary:
ec2_group.py doesn't do anything useful with --diff, and it'd be great if it did (with or without --check as well); without it, we have to resort to before-and-after captures of 'aws ec2 describe-security-groups' output, and even that only tells what changed after the change was already made.
##### Steps To Reproduce:
Adding --diff to a playbook that uses ec2_group.py would show output indicating what was changed (or would be changed, with --check) -- perhaps 'diff' style, perhaps some other format if 'diff' style doesn't really make sense. (I don't have a specific idea in mind, but would be happy to help come up with something and/or comment on something that others come up with.)
##### Expected Results:
Some sort of output indicating what had change (or would changed).
##### Actual Results:
No change in output from when the playbook is run without --diff.
| True | ec2_group.py and --diff - ##### Issue Type:
- Feature Idea
##### Plugin Name:
ec2_group.py
##### Ansible Version:
```
ansible 2.1.0 (devel 33f96edcd0) last updated 2016/03/10 14:52:58 (GMT -400)
lib/ansible/modules/core: (detached HEAD c86a0ef84a) last updated 2016/03/10 14:53:03 (GMT -400)
lib/ansible/modules/extras: (detached HEAD 33a557cc59) last updated 2016/03/10 14:53:03 (GMT -400)
config file =
configured module search path = Default w/o overrides
```
##### Ansible Configuration:
none
##### Environment:
N/A
##### Summary:
ec2_group.py doesn't do anything useful with --diff, and it'd be great if it did (with or without --check as well); without it, we have to resort to before-and-after captures of 'aws ec2 describe-security-groups' output, and even that only tells what changed after the change was already made.
##### Steps To Reproduce:
Adding --diff to a playbook that uses ec2_group.py would show output indicating what was changed (or would be changed, with --check) -- perhaps 'diff' style, perhaps some other format if 'diff' style doesn't really make sense. (I don't have a specific idea in mind, but would be happy to help come up with something and/or comment on something that others come up with.)
##### Expected Results:
Some sort of output indicating what had change (or would changed).
##### Actual Results:
No change in output from when the playbook is run without --diff.
| main | group py and diff issue type feature idea plugin name group py ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file configured module search path default w o overrides ansible configuration none environment n a summary group py doesn t do anything useful with diff and it d be great if it did with or without check as well without it we have to resort to before and after captures of aws describe security groups output and even that only tells what changed after the change was already made steps to reproduce adding diff to a playbook that uses group py would show output indicating what was changed or would be changed with check perhaps diff style perhaps some other format if diff style doesn t really make sense i don t have a specific idea in mind but would be happy to help come up with something and or comment on something that others come up with expected results some sort of output indicating what had change or would changed actual results no change in output from when the playbook is run without diff | 1 |
216,079 | 16,628,255,373 | IssuesEvent | 2021-06-03 12:32:55 | clastix/capsule | https://api.github.com/repos/clastix/capsule | closed | Fix typo in docs/contributing | documentation | In section: [Setting up the development environment](https://github.com/clastix/capsule/blob/master/docs/operator/contributing.md#setting-up-the-development-environment), requisite **OperatorSDK 1.9** needs to be fixed in **OperatorSDK 1.7.2**. | 1.0 | Fix typo in docs/contributing - In section: [Setting up the development environment](https://github.com/clastix/capsule/blob/master/docs/operator/contributing.md#setting-up-the-development-environment), requisite **OperatorSDK 1.9** needs to be fixed in **OperatorSDK 1.7.2**. | non_main | fix typo in docs contributing in section requisite operatorsdk needs to be fixed in operatorsdk | 0 |
9,502 | 2,906,220,699 | IssuesEvent | 2015-06-19 08:34:17 | ramu2016/SXUMX357FFAWEJHBMNM54FBQ | https://api.github.com/repos/ramu2016/SXUMX357FFAWEJHBMNM54FBQ | closed | nbO5nPMF2BQa7xO42AYsrTnhM8TONnmZ0vwW2R6w9yxyk+6iztWB3LznV77rpTImjfdejgAb4aqT54uNeVAI9VGxh1JBgeiIwehOXoCy134zrjif3F5s83dOXnlKVIsRw76SZ6PQnEH/WE+b90MINd4o3r53trz9kTLo8drkO7o= | design | kg5fMHBHZ/CaD01dLDTPOsEXIosYbQ+CFtIPspgSD2rgV48UwfIVoeTWKzbma3onZBwR4spcATLgKIZNyA6wkmuolV9FuqjBfbW7uEfwikHc2jSUbEdzLtFWiMcmeYNAuj5yuKoV07obPRtn9HgXHg/9VUb49MrZ87yV0bE7tABBRvT42n0rXCrSsBlAL3WQzt6mR6EaDz6sq2BW2TK0E+mnR7eSaFAqDFFOuZsRjhhDK/Gj4EuxgxAuHtO+3vfv5zp6rlAhp6jUr4izRvRaZwERmX+9wLYlxJJ5nng9j854AHtWWJSSBGFgrbmSZ/b+OmHNgER0ZlMeivICDEq1l8Gz1VmuPXBGjlVvdS59fIbiapZNEGIc9ZYQ8zFwnqrFLiIO5xVUSTNKdoWZ+Hf9VfrjyFwCZ6uz3o5t8wxSxNlHsE7fHE30aC9STOV697sUli4N/amKjc5bFi/xKV6Z//rjyFwCZ6uz3o5t8wxSxNnEBz8ao8CR/zgHtnTa3HgCmCC3CH/jkMHVmhWvqTN0xH9KIpA0HWBG2x1dewF4OTQj3k4Qyr3wyI6WohycYl0DTn2WxO+ZdHXvC6jnM19K+Q4k+93437kivtgrhTyoNWdddQMo/Im/0eJLyxTdrjVVc09qygIrCKKv/9r1vhJTdiijfLUsbjKQ5kJlX6llYkxr7UrX91CZx9ab/npbRzLvnKxKk1glIG+D8ZOgkhc0c39+mkAdiVaYrX/YZ6jcKhk4aA/qhKaHopNrZ35m5l+XecL+hSwmHVl+47kF44MOEKpkpck7MqqMoR7UuLYwFuBsCy1qq2KIdX0rYmtyLFIW2qe0NEPcdPlBbdVA2GRndD074+jGt1jE4cs4JsNSYcE3qiPFU3d5vsU0Ox0kDK6gISgKGbePALfI5TvWawP9qCk05cIrfw9ZHhn4ZuaUTZd4y+X3PdcEI4CqckvDHaRmjc4mDMY554lfhh4vCED/Ulz04ST87Nr4HBaOaAlNeveA4DHWQ/QftH1rfLQYstawX1B15EcakCqdiUVmOMtmoy5wyc/+AiD5nsZuO0dfZ2ArrqFWEjrGWxN05HxpGEfT54Eroi+Br3ujZBxoppgJ9ebWI9WSHCG9WAbsUfAYhip2rP8KMBUNM8Z/1Yn8FeKgxEwjymP9Oh+pKezs/ghQT/DQ3Fgqsh+cZk/i94+ZY3vrxhklmig0gwxtm6Nhj5g0egNmjos1e937Q4dsuCzp5bD6Y+JpZyux0PZrxi/1SHLvoaB2TMft+hgy5slEaKluVV7iEd7w9YHZZJOhWH4qLvE9zTNksVFt8w1MZj0VS0pZlzd9gSNShduKWz8j1yr/g6lFVUUzPPN1OhRydf5nyNT2XghDW6t7JzX0BjoSKVSZ34Z2Vu6D2+ebRe7Vf4ewF0ThRXfwhcQvNAHN9rD1MYJyEGdoH1a9n8sGns/ELOsZaUpxNYGZz49cU6PTY0QQql/opYiWvvpgNyWiAlc4Qle/M+RJsZjQkIACqmqVFNLd7sPH1v5entABLcxvlnXcyV5uBz6NdsrGhjkqgmqp83cLBAnC7HDcuqzH+tgiQxqqO5Y571zaW25Gc3QK7pwADO3U34ang60JFe75IQ352PdZsUNZhE2evgUkCUChQDL9fnwWOy/HgK/x/Kxynty87Xc6jO8ZxVyP2UGczdFiRpiGp9+aK9DKUxyWTEO124PwUmQiguag9cIvtUan6DHNAJTqbeCIIkAe3E+unny/P+uV01K0y02Vro99uLZazBlK0nXcKwyqfMNS2u1eG77MEuLLyAwD0u1IR60py6VnohIX5ApGZeapZf77OjW5Npg= | 1.0 | nbO5nPMF2BQa7xO42AYsrTnhM8TONnmZ0vwW2R6w9yxyk+6iztWB3LznV77rpTImjfdejgAb4aqT54uNeVAI9VGxh1JBgeiIwehOXoCy134zrjif3F5s83dOXnlKVIsRw76SZ6PQnEH/WE+b90MINd4o3r53trz9kTLo8drkO7o= - kg5fMHBHZ/CaD01dLDTPOsEXIosYbQ+CFtIPspgSD2rgV48UwfIVoeTWKzbma3onZBwR4spcATLgKIZNyA6wkmuolV9FuqjBfbW7uEfwikHc2jSUbEdzLtFWiMcmeYNAuj5yuKoV07obPRtn9HgXHg/9VUb49MrZ87yV0bE7tABBRvT42n0rXCrSsBlAL3WQzt6mR6EaDz6sq2BW2TK0E+mnR7eSaFAqDFFOuZsRjhhDK/Gj4EuxgxAuHtO+3vfv5zp6rlAhp6jUr4izRvRaZwERmX+9wLYlxJJ5nng9j854AHtWWJSSBGFgrbmSZ/b+OmHNgER0ZlMeivICDEq1l8Gz1VmuPXBGjlVvdS59fIbiapZNEGIc9ZYQ8zFwnqrFLiIO5xVUSTNKdoWZ+Hf9VfrjyFwCZ6uz3o5t8wxSxNlHsE7fHE30aC9STOV697sUli4N/amKjc5bFi/xKV6Z//rjyFwCZ6uz3o5t8wxSxNnEBz8ao8CR/zgHtnTa3HgCmCC3CH/jkMHVmhWvqTN0xH9KIpA0HWBG2x1dewF4OTQj3k4Qyr3wyI6WohycYl0DTn2WxO+ZdHXvC6jnM19K+Q4k+93437kivtgrhTyoNWdddQMo/Im/0eJLyxTdrjVVc09qygIrCKKv/9r1vhJTdiijfLUsbjKQ5kJlX6llYkxr7UrX91CZx9ab/npbRzLvnKxKk1glIG+D8ZOgkhc0c39+mkAdiVaYrX/YZ6jcKhk4aA/qhKaHopNrZ35m5l+XecL+hSwmHVl+47kF44MOEKpkpck7MqqMoR7UuLYwFuBsCy1qq2KIdX0rYmtyLFIW2qe0NEPcdPlBbdVA2GRndD074+jGt1jE4cs4JsNSYcE3qiPFU3d5vsU0Ox0kDK6gISgKGbePALfI5TvWawP9qCk05cIrfw9ZHhn4ZuaUTZd4y+X3PdcEI4CqckvDHaRmjc4mDMY554lfhh4vCED/Ulz04ST87Nr4HBaOaAlNeveA4DHWQ/QftH1rfLQYstawX1B15EcakCqdiUVmOMtmoy5wyc/+AiD5nsZuO0dfZ2ArrqFWEjrGWxN05HxpGEfT54Eroi+Br3ujZBxoppgJ9ebWI9WSHCG9WAbsUfAYhip2rP8KMBUNM8Z/1Yn8FeKgxEwjymP9Oh+pKezs/ghQT/DQ3Fgqsh+cZk/i94+ZY3vrxhklmig0gwxtm6Nhj5g0egNmjos1e937Q4dsuCzp5bD6Y+JpZyux0PZrxi/1SHLvoaB2TMft+hgy5slEaKluVV7iEd7w9YHZZJOhWH4qLvE9zTNksVFt8w1MZj0VS0pZlzd9gSNShduKWz8j1yr/g6lFVUUzPPN1OhRydf5nyNT2XghDW6t7JzX0BjoSKVSZ34Z2Vu6D2+ebRe7Vf4ewF0ThRXfwhcQvNAHN9rD1MYJyEGdoH1a9n8sGns/ELOsZaUpxNYGZz49cU6PTY0QQql/opYiWvvpgNyWiAlc4Qle/M+RJsZjQkIACqmqVFNLd7sPH1v5entABLcxvlnXcyV5uBz6NdsrGhjkqgmqp83cLBAnC7HDcuqzH+tgiQxqqO5Y571zaW25Gc3QK7pwADO3U34ang60JFe75IQ352PdZsUNZhE2evgUkCUChQDL9fnwWOy/HgK/x/Kxynty87Xc6jO8ZxVyP2UGczdFiRpiGp9+aK9DKUxyWTEO124PwUmQiguag9cIvtUan6DHNAJTqbeCIIkAe3E+unny/P+uV01K0y02Vro99uLZazBlK0nXcKwyqfMNS2u1eG77MEuLLyAwD0u1IR60py6VnohIX5ApGZeapZf77OjW5Npg= | non_main | we b im mkadivayrx xecl hswmhvl pkezs ghqt czk m hgk x unny p | 0 |
778,487 | 27,318,474,125 | IssuesEvent | 2023-02-24 17:35:12 | AY2223S2-CS2103T-W11-2/tp | https://api.github.com/repos/AY2223S2-CS2103T-W11-2/tp | opened | Have a calendar view of all application deadlines | type.Story priority.Medium | as an Intermediate user so that I can schedule a new interview | 1.0 | Have a calendar view of all application deadlines - as an Intermediate user so that I can schedule a new interview | non_main | have a calendar view of all application deadlines as an intermediate user so that i can schedule a new interview | 0 |
288,958 | 21,729,005,663 | IssuesEvent | 2022-05-11 10:13:44 | reliatec-gmbh/LibreClinica | https://api.github.com/repos/reliatec-gmbh/LibreClinica | closed | Merge developer documentation into the main repository and use Markdown markup | documentation | To unify the documentation effort it was recommended to merge the developer documentation into main docs. RestructuredText will be converted to Markdown markup. | 1.0 | Merge developer documentation into the main repository and use Markdown markup - To unify the documentation effort it was recommended to merge the developer documentation into main docs. RestructuredText will be converted to Markdown markup. | non_main | merge developer documentation into the main repository and use markdown markup to unify the documentation effort it was recommended to merge the developer documentation into main docs restructuredtext will be converted to markdown markup | 0 |
1,050 | 4,863,110,733 | IssuesEvent | 2016-11-14 14:37:04 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | vmware_guest - set static IP for virtual machine | affects_2.3 cloud feature_idea vmware waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
vmware_guest
##### SUMMARY
<!--- Explain the problem briefly -->
Would like to be able to set the IP address of a VM when cloning from a template. Is this already possible? How are other people solving this??
| True | vmware_guest - set static IP for virtual machine - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
vmware_guest
##### SUMMARY
<!--- Explain the problem briefly -->
Would like to be able to set the IP address of a VM when cloning from a template. Is this already possible? How are other people solving this??
| main | vmware guest set static ip for virtual machine issue type feature idea component name vmware guest summary would like to be able to set the ip address of a vm when cloning from a template is this already possible how are other people solving this | 1 |
625,730 | 19,761,898,790 | IssuesEvent | 2022-01-16 14:55:30 | openvk/openvk | https://api.github.com/repos/openvk/openvk | opened | SEO-оптимизация | type:enhancement priority:normal language:html complexity:easy scope:locales | - [ ] meta-теги на самых популярных страничках
- [ ] описание соцсети
- [ ] каталог юзеров, как в старом вк (для улучшения индексации)
Ещё было бы круто иметь sitemap.xml, но задача с картой сайта весьма... интересная и сложная, чтобы валяться тут. | 1.0 | SEO-оптимизация - - [ ] meta-теги на самых популярных страничках
- [ ] описание соцсети
- [ ] каталог юзеров, как в старом вк (для улучшения индексации)
Ещё было бы круто иметь sitemap.xml, но задача с картой сайта весьма... интересная и сложная, чтобы валяться тут. | non_main | seo оптимизация meta теги на самых популярных страничках описание соцсети каталог юзеров как в старом вк для улучшения индексации ещё было бы круто иметь sitemap xml но задача с картой сайта весьма интересная и сложная чтобы валяться тут | 0 |
2,988 | 10,790,895,495 | IssuesEvent | 2019-11-05 15:45:49 | ansible/ansible | https://api.github.com/repos/ansible/ansible | opened | Allow the ability to replace/restore a device configuration. | affects_2.10 feature module needs_maintainer needs_triage support:community | <!--- Verify first that your feature was not already discussed on GitHub -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
Currently we have the ability to push configurations to a device but not completely replace a running/startup configuration. It would be nice to have the ability to replace the device config with a backup configuration from a custom database.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
Network Modules
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
Assuming we have saved configurations in our own database, we would like to replace the running config with the ones from our collection.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
```
<!--- HINT: You can also paste gist.github.com links for larger files -->
| True | Allow the ability to replace/restore a device configuration. - <!--- Verify first that your feature was not already discussed on GitHub -->
<!--- Complete *all* sections as described, this form is processed automatically -->
##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
Currently we have the ability to push configurations to a device but not completely replace a running/startup configuration. It would be nice to have the ability to replace the device config with a backup configuration from a custom database.
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the module, plugin, task or feature below, use your best guess if unsure -->
Network Modules
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
Assuming we have saved configurations in our own database, we would like to replace the running config with the ones from our collection.
<!--- Paste example playbooks or commands between quotes below -->
```yaml
```
<!--- HINT: You can also paste gist.github.com links for larger files -->
| main | allow the ability to replace restore a device configuration summary currently we have the ability to push configurations to a device but not completely replace a running startup configuration it would be nice to have the ability to replace the device config with a backup configuration from a custom database issue type feature idea component name network modules additional information assuming we have saved configurations in our own database we would like to replace the running config with the ones from our collection yaml | 1 |
131,109 | 18,235,702,149 | IssuesEvent | 2021-10-01 06:32:08 | reearth/reearth | https://api.github.com/repos/reearth/reearth | closed | feat: refine notification system | enhancement feature request design SP: 5 | # Leader
@KaWaite
# User story
We don't have enough notification feedback in our system, and also not easy to develop. Please update a little bit for our team and our users.
# Design
please check [here](https://www.figma.com/file/bdnfDaGXGoNfUws4y7JItl/Re-Earth-UI-Gamma?node-id=8461%3A158978)
| 1.0 | feat: refine notification system - # Leader
@KaWaite
# User story
We don't have enough notification feedback in our system, and also not easy to develop. Please update a little bit for our team and our users.
# Design
please check [here](https://www.figma.com/file/bdnfDaGXGoNfUws4y7JItl/Re-Earth-UI-Gamma?node-id=8461%3A158978)
| non_main | feat refine notification system leader kawaite user story we don t have enough notification feedback in our system and also not easy to develop please update a little bit for our team and our users design please check | 0 |
736 | 4,329,594,735 | IssuesEvent | 2016-07-26 17:19:21 | duckduckgo/zeroclickinfo-spice | https://api.github.com/repos/duckduckgo/zeroclickinfo-spice | closed | Whois: JS Error | Bug Maintainer Input Requested | Need to gracefully handle this and not throw an error:
```
TypeError: undefined is not an object (evaluating 'api_result.registryData.registrant)
```
------
IA Page: http://duck.co/ia/view/whois
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @b1ake | True | Whois: JS Error - Need to gracefully handle this and not throw an error:
```
TypeError: undefined is not an object (evaluating 'api_result.registryData.registrant)
```
------
IA Page: http://duck.co/ia/view/whois
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @b1ake | main | whois js error need to gracefully handle this and not throw an error typeerror undefined is not an object evaluating api result registrydata registrant ia page | 1 |
73,560 | 9,670,583,942 | IssuesEvent | 2019-05-21 20:17:27 | microsoft/msbuild | https://api.github.com/repos/microsoft/msbuild | closed | Add GetFileHash and VerifyFileHash tasks to XSD (for IntelliSense) | Documentation | _Reported by @japj in https://github.com/Microsoft/msbuild/pull/3999#issuecomment-458141311._
#3398 exposed two new tasks, but they're not exposed to Visual Studio's autocompletion because they're not mentioned in the XSD used for that.
They need an entry like
https://github.com/Microsoft/msbuild/blob/966cdf2ac602e359530ad63f608e6cbba9c9fd0f/src/MSBuild/Microsoft.Build.CommonTypes.xsd#L2198-L2211
but with the right inputs/outputs. | 1.0 | Add GetFileHash and VerifyFileHash tasks to XSD (for IntelliSense) - _Reported by @japj in https://github.com/Microsoft/msbuild/pull/3999#issuecomment-458141311._
#3398 exposed two new tasks, but they're not exposed to Visual Studio's autocompletion because they're not mentioned in the XSD used for that.
They need an entry like
https://github.com/Microsoft/msbuild/blob/966cdf2ac602e359530ad63f608e6cbba9c9fd0f/src/MSBuild/Microsoft.Build.CommonTypes.xsd#L2198-L2211
but with the right inputs/outputs. | non_main | add getfilehash and verifyfilehash tasks to xsd for intellisense reported by japj in exposed two new tasks but they re not exposed to visual studio s autocompletion because they re not mentioned in the xsd used for that they need an entry like but with the right inputs outputs | 0 |
5,644 | 28,369,926,189 | IssuesEvent | 2023-04-12 16:12:44 | deislabs/spiderlightning | https://api.github.com/repos/deislabs/spiderlightning | opened | slight.exe doesn't run on WS2019 or WS2022 due to missing dlls | 🐛 bug 🚧 maintainer issue | **Description of the bug**
When running on Windows Server 2019/2022 slight doesn't run:
```
PS C:\Program Files\containerd\release> .\slight.exe -h
<no output>
```
**To Reproduce**
Run either WS2022 and WS2019
```
cmd /c ver
Microsoft Windows [Version 10.0.20348.1607]
```
install:
```
curl.exe -LO https://github.com/deislabs/spiderlightning/releases/download/v0.4.1/slight-windows-x86_64.tar.gz
tar.exe xvf .\slight-windows-x86_64.tar.gz
x release/slight.exe
```

**Additional context**
We may want to link the runtime for windows statically as many of the binaries. info in https://stackoverflow.com/a/44387312/697126 | True | slight.exe doesn't run on WS2019 or WS2022 due to missing dlls - **Description of the bug**
When running on Windows Server 2019/2022 slight doesn't run:
```
PS C:\Program Files\containerd\release> .\slight.exe -h
<no output>
```
**To Reproduce**
Run either WS2022 and WS2019
```
cmd /c ver
Microsoft Windows [Version 10.0.20348.1607]
```
install:
```
curl.exe -LO https://github.com/deislabs/spiderlightning/releases/download/v0.4.1/slight-windows-x86_64.tar.gz
tar.exe xvf .\slight-windows-x86_64.tar.gz
x release/slight.exe
```

**Additional context**
We may want to link the runtime for windows statically as many of the binaries. info in https://stackoverflow.com/a/44387312/697126 | main | slight exe doesn t run on or due to missing dlls description of the bug when running on windows server slight doesn t run ps c program files containerd release slight exe h to reproduce run either and cmd c ver microsoft windows install curl exe lo tar exe xvf slight windows tar gz x release slight exe additional context we may want to link the runtime for windows statically as many of the binaries info in | 1 |
745 | 4,350,929,598 | IssuesEvent | 2016-07-31 15:23:33 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Apt module - the possibilty to know if a debian package is present or not | feature_idea waiting_on_maintainer | ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
apt
##### ANSIBLE VERSION
```
ansible --version
ansible 2.0.2.0
config file =
configured module search path = Default w/o override
```
##### CONFIGURATION
No files and no env
##### OS / ENVIRONMENT
```
lsb_release -a
No LSB modules are available.
Distributor ID: Debian
Description: Debian GNU/Linux 8.4 (jessie)
Release: 8.4
Codename: jessie
```
##### SUMMARY
Just for obtain an little feature : the possibilty to know if a debian package is present or not .
After, if the condition is true, we can register the package version.
Because, using the shell module is dirty :+1:
```
- name: test version
shell: haproxy -v | awk '$0 ~ /HA-Proxy/ {print$3}'
register: haproxyversion
tags:
- status
- name: status of backends
shell: echo "show servers state" | nc localhost 666 | grep -Ev "^1|^#|^$" | awk '{print"frontend:"" "$2" ""backend:"" "$4" ""ip:"" "$5" ""status:"" "$6}'
register: haproxyout
when: haproxyversion.stdout.find('1.6') != -1
tags:
- status
```
##### STEPS TO REPRODUCE
It's not a bug
##### EXPECTED RESULTS
It's not a bug
##### ACTUAL RESULTS
It's not a bug
| True | Apt module - the possibilty to know if a debian package is present or not - ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
apt
##### ANSIBLE VERSION
```
ansible --version
ansible 2.0.2.0
config file =
configured module search path = Default w/o override
```
##### CONFIGURATION
No files and no env
##### OS / ENVIRONMENT
```
lsb_release -a
No LSB modules are available.
Distributor ID: Debian
Description: Debian GNU/Linux 8.4 (jessie)
Release: 8.4
Codename: jessie
```
##### SUMMARY
Just for obtain an little feature : the possibilty to know if a debian package is present or not .
After, if the condition is true, we can register the package version.
Because, using the shell module is dirty :+1:
```
- name: test version
shell: haproxy -v | awk '$0 ~ /HA-Proxy/ {print$3}'
register: haproxyversion
tags:
- status
- name: status of backends
shell: echo "show servers state" | nc localhost 666 | grep -Ev "^1|^#|^$" | awk '{print"frontend:"" "$2" ""backend:"" "$4" ""ip:"" "$5" ""status:"" "$6}'
register: haproxyout
when: haproxyversion.stdout.find('1.6') != -1
tags:
- status
```
##### STEPS TO REPRODUCE
It's not a bug
##### EXPECTED RESULTS
It's not a bug
##### ACTUAL RESULTS
It's not a bug
| main | apt module the possibilty to know if a debian package is present or not issue type feature idea component name apt ansible version ansible version ansible config file configured module search path default w o override configuration no files and no env os environment lsb release a no lsb modules are available distributor id debian description debian gnu linux jessie release codename jessie summary just for obtain an little feature the possibilty to know if a debian package is present or not after if the condition is true we can register the package version because using the shell module is dirty name test version shell haproxy v awk ha proxy print register haproxyversion tags status name status of backends shell echo show servers state nc localhost grep ev awk print frontend backend ip status register haproxyout when haproxyversion stdout find tags status steps to reproduce it s not a bug expected results it s not a bug actual results it s not a bug | 1 |
15,501 | 19,703,263,239 | IssuesEvent | 2022-01-12 18:52:05 | googleapis/java-analytics-data | https://api.github.com/repos/googleapis/java-analytics-data | opened | Your .repo-metadata.json file has a problem 🤒 | type: process repo-metadata: lint | You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'analytics-data' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions. | 1.0 | Your .repo-metadata.json file has a problem 🤒 - You have a problem with your .repo-metadata.json file:
Result of scan 📈:
* release_level must be equal to one of the allowed values in .repo-metadata.json
* api_shortname 'analytics-data' invalid in .repo-metadata.json
☝️ Once you correct these problems, you can close this issue.
Reach out to **go/github-automation** if you have any questions. | non_main | your repo metadata json file has a problem 🤒 you have a problem with your repo metadata json file result of scan 📈 release level must be equal to one of the allowed values in repo metadata json api shortname analytics data invalid in repo metadata json ☝️ once you correct these problems you can close this issue reach out to go github automation if you have any questions | 0 |
3,016 | 11,184,945,983 | IssuesEvent | 2019-12-31 21:14:47 | laminas/laminas-code | https://api.github.com/repos/laminas/laminas-code | opened | MethodReflection::getPrototype() crashes when method has no return | Awaiting Maintainer Response Bug | When I use this method to grab the protype of a method of my class, it crashes with this error:
> **"Call to a member function getTypes() on boolean"**
It happens because I have a method without return. My opinion is: the return type should be "void" on prototype. Issue #84 adresses the support for void. I've modified the code to make it return void. I could create a pull request if you guys want me to do so.
It was like this:
<pre>
$return = $docBlock->getTag('return');
$returnTypes = $return->getTypes();
$returnType = count($returnTypes) > 1 ? implode('|', $returnTypes) : $returnTypes[0];
</pre>
And I've changed to this:
<pre>
$return = $docBlock->getTag('return');
if($return === false)
{
$returnType = 'void';
}
else
{
$returnTypes = $return->getTypes();
$returnType = count($returnTypes) > 1 ? implode('|', $returnTypes) : $returnTypes[0];
}
</pre>
If you don't think it should be void, at least set <b>$returnType</b> variable to empty string or null.
I've read the CONTRIBUTING.md, but I'm still pretty new to contributing on github, so fell free to get this code and push it into repository.
Regards
---
Originally posted by @ericklima-comp at https://github.com/zendframework/zend-code/issues/154 | True | MethodReflection::getPrototype() crashes when method has no return - When I use this method to grab the protype of a method of my class, it crashes with this error:
> **"Call to a member function getTypes() on boolean"**
It happens because I have a method without return. My opinion is: the return type should be "void" on prototype. Issue #84 adresses the support for void. I've modified the code to make it return void. I could create a pull request if you guys want me to do so.
It was like this:
<pre>
$return = $docBlock->getTag('return');
$returnTypes = $return->getTypes();
$returnType = count($returnTypes) > 1 ? implode('|', $returnTypes) : $returnTypes[0];
</pre>
And I've changed to this:
<pre>
$return = $docBlock->getTag('return');
if($return === false)
{
$returnType = 'void';
}
else
{
$returnTypes = $return->getTypes();
$returnType = count($returnTypes) > 1 ? implode('|', $returnTypes) : $returnTypes[0];
}
</pre>
If you don't think it should be void, at least set <b>$returnType</b> variable to empty string or null.
I've read the CONTRIBUTING.md, but I'm still pretty new to contributing on github, so fell free to get this code and push it into repository.
Regards
---
Originally posted by @ericklima-comp at https://github.com/zendframework/zend-code/issues/154 | main | methodreflection getprototype crashes when method has no return when i use this method to grab the protype of a method of my class it crashes with this error call to a member function gettypes on boolean it happens because i have a method without return my opinion is the return type should be void on prototype issue adresses the support for void i ve modified the code to make it return void i could create a pull request if you guys want me to do so it was like this return docblock gettag return returntypes return gettypes returntype count returntypes implode returntypes returntypes and i ve changed to this return docblock gettag return if return false returntype void else returntypes return gettypes returntype count returntypes implode returntypes returntypes if you don t think it should be void at least set returntype variable to empty string or null i ve read the contributing md but i m still pretty new to contributing on github so fell free to get this code and push it into repository regards originally posted by ericklima comp at | 1 |
157,330 | 5,997,301,280 | IssuesEvent | 2017-06-03 22:35:50 | ncssar/sartopo-feature-requests | https://api.github.com/repos/ncssar/sartopo-feature-requests | opened | bulk ops select all / unselect all buttons | Priority:Medium | seems silly, but, would be a good time-saver for big maps with lots of objects. | 1.0 | bulk ops select all / unselect all buttons - seems silly, but, would be a good time-saver for big maps with lots of objects. | non_main | bulk ops select all unselect all buttons seems silly but would be a good time saver for big maps with lots of objects | 0 |
2,137 | 7,346,312,294 | IssuesEvent | 2018-03-07 20:17:17 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Technical question about Linux Server | BYOND Issue Bug Maintainability/Hinders improvements Needs Reproducing | Is that normal, if, on 512.1407, Linux distro, I get very often server crashes? If yes, is there any way to avoid them, or to log them, so contributors could bug them out? | True | Technical question about Linux Server - Is that normal, if, on 512.1407, Linux distro, I get very often server crashes? If yes, is there any way to avoid them, or to log them, so contributors could bug them out? | main | technical question about linux server is that normal if on linux distro i get very often server crashes if yes is there any way to avoid them or to log them so contributors could bug them out | 1 |
2,446 | 8,639,857,682 | IssuesEvent | 2018-11-23 22:07:57 | F5OEO/rpitx | https://api.github.com/repos/F5OEO/rpitx | closed | rpitx does not transmit anything | V1 related (not maintained) | I've been playing with `pifm` and `PiFmRds` a lot on my Raspberry Pi and everything works fine.
Today I tried using `rpitx`, but it does not transmit anything. Here is an example output from `testssb.sh`:
```
Reading file : ./sampleaudio.wav
Sample Rate : 48000
Channels : 1
Writing file : ./ssbIQ.wav
Channels : 2
rpitx Version 0.2 compiled Apr 4 2018 (F5OEO Evariste) running on Frequency=0.000000 SampleRate=48000 Master PLL = 1000000000
Warning : Using harmonic 3
memory: 1024 MB
processor: Broadcom BCM2837
i2cDevice: /dev/i2c-1
model: Model B Pi 3
manufacturer: Embest
pcb revision: 2
warranty void: no
revision: a22082
peripheral base: 0x3f000000
Jessie
Using mbox device /dev/vcio.
3616000 Size NUM PAGES 883 PAGE_SIZE 4096
MASH 1 Freq PLL# 6
Calibrate : ppm=-6 DMA 166ns:1277ns
WaitNano=20833 F1=96426385.423043 TuneFrequency 96428222.222222 F2=96428655.507687 Initial Resolution(Hz)=2270.084644 ResolutionPWMF 18.160677 NbStep=125 DELAYStep=7
****** STARTING TRANSMIT ********
```
I have a RTL-SDR dongle plugged in my desktop computer, and I get no signal on 933.9MHz.
Here is my `uname -a`:
```Linux raspberrypi 4.14.30-v7+ #1102 SMP Mon Mar 26 16:45:49 BST 2018 armv7l GNU/Linux``` | True | rpitx does not transmit anything - I've been playing with `pifm` and `PiFmRds` a lot on my Raspberry Pi and everything works fine.
Today I tried using `rpitx`, but it does not transmit anything. Here is an example output from `testssb.sh`:
```
Reading file : ./sampleaudio.wav
Sample Rate : 48000
Channels : 1
Writing file : ./ssbIQ.wav
Channels : 2
rpitx Version 0.2 compiled Apr 4 2018 (F5OEO Evariste) running on Frequency=0.000000 SampleRate=48000 Master PLL = 1000000000
Warning : Using harmonic 3
memory: 1024 MB
processor: Broadcom BCM2837
i2cDevice: /dev/i2c-1
model: Model B Pi 3
manufacturer: Embest
pcb revision: 2
warranty void: no
revision: a22082
peripheral base: 0x3f000000
Jessie
Using mbox device /dev/vcio.
3616000 Size NUM PAGES 883 PAGE_SIZE 4096
MASH 1 Freq PLL# 6
Calibrate : ppm=-6 DMA 166ns:1277ns
WaitNano=20833 F1=96426385.423043 TuneFrequency 96428222.222222 F2=96428655.507687 Initial Resolution(Hz)=2270.084644 ResolutionPWMF 18.160677 NbStep=125 DELAYStep=7
****** STARTING TRANSMIT ********
```
I have a RTL-SDR dongle plugged in my desktop computer, and I get no signal on 933.9MHz.
Here is my `uname -a`:
```Linux raspberrypi 4.14.30-v7+ #1102 SMP Mon Mar 26 16:45:49 BST 2018 armv7l GNU/Linux``` | main | rpitx does not transmit anything i ve been playing with pifm and pifmrds a lot on my raspberry pi and everything works fine today i tried using rpitx but it does not transmit anything here is an example output from testssb sh reading file sampleaudio wav sample rate channels writing file ssbiq wav channels rpitx version compiled apr evariste running on frequency samplerate master pll warning using harmonic memory mb processor broadcom dev model model b pi manufacturer embest pcb revision warranty void no revision peripheral base jessie using mbox device dev vcio size num pages page size mash freq pll calibrate ppm dma waitnano tunefrequency initial resolution hz resolutionpwmf nbstep delaystep starting transmit i have a rtl sdr dongle plugged in my desktop computer and i get no signal on here is my uname a linux raspberrypi smp mon mar bst gnu linux | 1 |
4,910 | 25,257,604,391 | IssuesEvent | 2022-11-15 19:32:15 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | Allow node prop-types in SideNavMenu & TooltipDefinition | type: a11y ♿ status: waiting for maintainer response 💬 | Would you see any problem with allowing `React.ReacNode` (in typescript sense) to be passed into these?
- https://github.com/carbon-design-system/carbon/blob/master/packages/react/src/components/UIShell/SideNavMenu.js#L69
- https://github.com/carbon-design-system/carbon/blob/master/packages/react/src/components/TooltipDefinition/TooltipDefinition.js#L104
With the `SideNavMenu` it's about our design team wanting to put little charts into menu items & `TooltipDefinition` change is about being able to render icon as the "trigger text". Both of these work for us just fine. It's just about the annoying development mode warnings caused by `prop-types`.
## Detailed description
> What version of the Carbon Design System are you using?
`carbon-components-react@7.22.0` | True | Allow node prop-types in SideNavMenu & TooltipDefinition - Would you see any problem with allowing `React.ReacNode` (in typescript sense) to be passed into these?
- https://github.com/carbon-design-system/carbon/blob/master/packages/react/src/components/UIShell/SideNavMenu.js#L69
- https://github.com/carbon-design-system/carbon/blob/master/packages/react/src/components/TooltipDefinition/TooltipDefinition.js#L104
With the `SideNavMenu` it's about our design team wanting to put little charts into menu items & `TooltipDefinition` change is about being able to render icon as the "trigger text". Both of these work for us just fine. It's just about the annoying development mode warnings caused by `prop-types`.
## Detailed description
> What version of the Carbon Design System are you using?
`carbon-components-react@7.22.0` | main | allow node prop types in sidenavmenu tooltipdefinition would you see any problem with allowing react reacnode in typescript sense to be passed into these with the sidenavmenu it s about our design team wanting to put little charts into menu items tooltipdefinition change is about being able to render icon as the trigger text both of these work for us just fine it s just about the annoying development mode warnings caused by prop types detailed description what version of the carbon design system are you using carbon components react | 1 |
1,403 | 6,025,502,809 | IssuesEvent | 2017-06-08 08:52:18 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | Disable existing Scheduled Task with win_scheduled_task | affects_2.1 bug_report waiting_on_maintainer windows | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_scheduled_task
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
no changes to ansible.cfg, or ansible variables set.
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ansible running from = RHEL 6.7 (Santiago)
OS i'm managing = Windows 2008 r2
##### SUMMARY
<!--- Explain the problem briefly -->
trying to use win_scheduled_task to disable an existing scheduled task, using;
win_scheduled_task: name="ServerManager" enable=no state=present returns,
"Missing required argument: execute"
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
as above, simply using
win_scheduled_task: name="ServerManager" enable=no state=present
results in an failed message, stating "Missing required argument: execute" even though execute is not a required option according to http://docs.ansible.com/ansible/win_scheduled_task_module.html.
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Windows Build Test
hosts: all
tasks:
- name: Disable Server Manager Scheduled Task
win_scheduled_task: name="ServerManager" enable=no state=present
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
I would have expected the existing scheduled task to be disabled, and even then not error asking for an execute option, when its not required.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
server1.domain.com]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "win_scheduled_task"}, "msg": "Missing required argument: execute"}
``` | True | Disable existing Scheduled Task with win_scheduled_task - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_scheduled_task
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
no changes to ansible.cfg, or ansible variables set.
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ansible running from = RHEL 6.7 (Santiago)
OS i'm managing = Windows 2008 r2
##### SUMMARY
<!--- Explain the problem briefly -->
trying to use win_scheduled_task to disable an existing scheduled task, using;
win_scheduled_task: name="ServerManager" enable=no state=present returns,
"Missing required argument: execute"
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
as above, simply using
win_scheduled_task: name="ServerManager" enable=no state=present
results in an failed message, stating "Missing required argument: execute" even though execute is not a required option according to http://docs.ansible.com/ansible/win_scheduled_task_module.html.
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Windows Build Test
hosts: all
tasks:
- name: Disable Server Manager Scheduled Task
win_scheduled_task: name="ServerManager" enable=no state=present
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
I would have expected the existing scheduled task to be disabled, and even then not error asking for an execute option, when its not required.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
server1.domain.com]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "win_scheduled_task"}, "msg": "Missing required argument: execute"}
``` | main | disable existing scheduled task with win scheduled task issue type bug report component name win scheduled task ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables no changes to ansible cfg or ansible variables set os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific ansible running from rhel santiago os i m managing windows summary trying to use win scheduled task to disable an existing scheduled task using win scheduled task name servermanager enable no state present returns missing required argument execute steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used as above simply using win scheduled task name servermanager enable no state present results in an failed message stating missing required argument execute even though execute is not a required option according to name windows build test hosts all tasks name disable server manager scheduled task win scheduled task name servermanager enable no state present expected results i would have expected the existing scheduled task to be disabled and even then not error asking for an execute option when its not required actual results domain com failed changed false failed true invocation module name win scheduled task msg missing required argument execute | 1 |
4,085 | 19,292,420,627 | IssuesEvent | 2021-12-12 02:03:20 | cloverhearts/quilljs-markdown | https://api.github.com/repos/cloverhearts/quilljs-markdown | closed | Disable markdown on-the-fly | NICE IDEA Saw with Maintainer | Hi, I'd like to give my users the ability to enable/disable markdown support on-the-fly.
When markdown is enabled, I just call `new QuillMarkdown(quill)`, which works fine.
But I haven't found a way to disable it on the fly. `delete`ing the `QuillMarkdown` instance doesn't seem to work. We'd probably have to remove the event listeners added in the constructor.
Is there any support for that? Thanks! | True | Disable markdown on-the-fly - Hi, I'd like to give my users the ability to enable/disable markdown support on-the-fly.
When markdown is enabled, I just call `new QuillMarkdown(quill)`, which works fine.
But I haven't found a way to disable it on the fly. `delete`ing the `QuillMarkdown` instance doesn't seem to work. We'd probably have to remove the event listeners added in the constructor.
Is there any support for that? Thanks! | main | disable markdown on the fly hi i d like to give my users the ability to enable disable markdown support on the fly when markdown is enabled i just call new quillmarkdown quill which works fine but i haven t found a way to disable it on the fly delete ing the quillmarkdown instance doesn t seem to work we d probably have to remove the event listeners added in the constructor is there any support for that thanks | 1 |
54,976 | 6,885,967,144 | IssuesEvent | 2017-11-21 17:45:20 | EnginyeriaSofwareUB/ES2017A3 | https://api.github.com/repos/EnginyeriaSofwareUB/ES2017A3 | closed | Disseny, efectes de partícules i efectes de so dels objectes de defensa | Design | Escut: modelat (fet durant l'Sprint 2) + so d'impacte de projectil
Escut doble: modelat + so d'impacte de projectil
Iglú: modelat + so d'impacte de projectil
Àngel de la guarda: sprite + efecte de recuperació + so de recuperació
Temps estimat: 5h
Temps de realització: 4h
Resultat visual:




| 1.0 | Disseny, efectes de partícules i efectes de so dels objectes de defensa - Escut: modelat (fet durant l'Sprint 2) + so d'impacte de projectil
Escut doble: modelat + so d'impacte de projectil
Iglú: modelat + so d'impacte de projectil
Àngel de la guarda: sprite + efecte de recuperació + so de recuperació
Temps estimat: 5h
Temps de realització: 4h
Resultat visual:




| non_main | disseny efectes de partícules i efectes de so dels objectes de defensa escut modelat fet durant l sprint so d impacte de projectil escut doble modelat so d impacte de projectil iglú modelat so d impacte de projectil àngel de la guarda sprite efecte de recuperació so de recuperació temps estimat temps de realització resultat visual | 0 |
568,584 | 16,983,499,572 | IssuesEvent | 2021-06-30 11:52:22 | Hayao-H/Niconicome | https://api.github.com/repos/Hayao-H/Niconicome | closed | [バグ] 起動時に実行ファイルがあるディレクトリにchromeディレクトリがないとエラーが発生する | High Priority bug | **バグの説明**
起動時に実行ファイルがあるディレクトリにchromeディレクトリがないとエラーが発生する
**再現方法**
**期待する挙動**
本来ならばどのような挙動が望ましいのか
**スクリーンショット**
可能であれば
**Desktop (下記の情報の記載をお願いします):**
- OS: 最低限OSバージョン・64bitか32bitかなどをお願いします
- Version: アプリケーションのバージョン
**ログファイル**
https://gist.githubusercontent.com/sh4869221b/4b23d46c1d9fecdbb287914d04645057/raw/8a5dced74e49027698ecb31f07e3399d436bd38c/niconicome-log
**備考**
その他
| 1.0 | [バグ] 起動時に実行ファイルがあるディレクトリにchromeディレクトリがないとエラーが発生する - **バグの説明**
起動時に実行ファイルがあるディレクトリにchromeディレクトリがないとエラーが発生する
**再現方法**
**期待する挙動**
本来ならばどのような挙動が望ましいのか
**スクリーンショット**
可能であれば
**Desktop (下記の情報の記載をお願いします):**
- OS: 最低限OSバージョン・64bitか32bitかなどをお願いします
- Version: アプリケーションのバージョン
**ログファイル**
https://gist.githubusercontent.com/sh4869221b/4b23d46c1d9fecdbb287914d04645057/raw/8a5dced74e49027698ecb31f07e3399d436bd38c/niconicome-log
**備考**
その他
| non_main | 起動時に実行ファイルがあるディレクトリにchromeディレクトリがないとエラーが発生する バグの説明 起動時に実行ファイルがあるディレクトリにchromeディレクトリがないとエラーが発生する 再現方法 期待する挙動 本来ならばどのような挙動が望ましいのか スクリーンショット 可能であれば desktop 下記の情報の記載をお願いします os 最低限osバージョン・ version アプリケーションのバージョン ログファイル 備考 その他 | 0 |
703,568 | 24,166,267,785 | IssuesEvent | 2022-09-22 15:15:51 | Fiserv/Support | https://api.github.com/repos/Fiserv/Support | closed | To enable the Preview feature for FirstVision-apac product | enhancement Priority - Medium FirstVisionAPAC | **Region** (optional)
The region of a user requesting the feature.
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
<!-- ** Recommended **
- Assign a Feature
On you right hand side, you can set an Assignees by Clicking setting button next to it.
- Subscribe to a Team
At the end of your Feature template, please mention team you would like to subscribe for this Issue for future updates. You can mention teams in Github
Example: @Fiserv/fiserv-developer)
-->
| 1.0 | To enable the Preview feature for FirstVision-apac product - **Region** (optional)
The region of a user requesting the feature.
**Is your feature request related to a problem? Please describe.**
A clear and concise description of what the problem is. Ex. I'm always frustrated when [...]
**Describe the solution you'd like**
A clear and concise description of what you want to happen.
**Describe alternatives you've considered**
A clear and concise description of any alternative solutions or features you've considered.
**Additional context**
Add any other context or screenshots about the feature request here.
<!-- ** Recommended **
- Assign a Feature
On you right hand side, you can set an Assignees by Clicking setting button next to it.
- Subscribe to a Team
At the end of your Feature template, please mention team you would like to subscribe for this Issue for future updates. You can mention teams in Github
Example: @Fiserv/fiserv-developer)
-->
| non_main | to enable the preview feature for firstvision apac product region optional the region of a user requesting the feature is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when describe the solution you d like a clear and concise description of what you want to happen describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here recommended assign a feature on you right hand side you can set an assignees by clicking setting button next to it subscribe to a team at the end of your feature template please mention team you would like to subscribe for this issue for future updates you can mention teams in github example fiserv fiserv developer | 0 |
724 | 4,318,959,122 | IssuesEvent | 2016-07-24 11:05:43 | gogits/gogs | https://api.github.com/repos/gogits/gogs | closed | AdminCreateOrg API ignores FullName property | kind/bug status/assigned to maintainer status/needs feedback | - Gogs version (or commit ref):
- 0.9.22.0425
- 7049cb9d975505a6f1fa3fad809dd92c1078360f
- Database:
- [x] PostgreSQL
- Can you reproduce the bug at http://try.gogs.io:
- [ ] Yes (provide example URL)
- [x] No... because I don't known how to create a new organization in my account with `client.AdminCreateOrg`...
- [ ] Not relevant
## Description
The `FullName` property from bellow seems to be ignored by the gogs API endpoint. It seem to end up with the same value as the `UserName` property.
```go
org, err := client.AdminCreateOrg(
"gogs",
gogs.CreateOrgOption{
UserName: "acmeOrganization",
FullName: "acmeOrganization Full Name", // XXX this is ignored.
Description: "Description",
Website: "http://acme.example.com",
Location: "Lisbon, Portugal",
})
```
| True | AdminCreateOrg API ignores FullName property - - Gogs version (or commit ref):
- 0.9.22.0425
- 7049cb9d975505a6f1fa3fad809dd92c1078360f
- Database:
- [x] PostgreSQL
- Can you reproduce the bug at http://try.gogs.io:
- [ ] Yes (provide example URL)
- [x] No... because I don't known how to create a new organization in my account with `client.AdminCreateOrg`...
- [ ] Not relevant
## Description
The `FullName` property from bellow seems to be ignored by the gogs API endpoint. It seem to end up with the same value as the `UserName` property.
```go
org, err := client.AdminCreateOrg(
"gogs",
gogs.CreateOrgOption{
UserName: "acmeOrganization",
FullName: "acmeOrganization Full Name", // XXX this is ignored.
Description: "Description",
Website: "http://acme.example.com",
Location: "Lisbon, Portugal",
})
```
| main | admincreateorg api ignores fullname property gogs version or commit ref database postgresql can you reproduce the bug at yes provide example url no because i don t known how to create a new organization in my account with client admincreateorg not relevant description the fullname property from bellow seems to be ignored by the gogs api endpoint it seem to end up with the same value as the username property go org err client admincreateorg gogs gogs createorgoption username acmeorganization fullname acmeorganization full name xxx this is ignored description description website location lisbon portugal | 1 |
5,157 | 26,269,262,630 | IssuesEvent | 2023-01-06 15:27:18 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Support descriptions for tables on the frontend | type: enhancement work: frontend status: ready restricted: maintainers | * This issue is to track the implementation of descriptions for tables on the frontend. | True | Support descriptions for tables on the frontend - * This issue is to track the implementation of descriptions for tables on the frontend. | main | support descriptions for tables on the frontend this issue is to track the implementation of descriptions for tables on the frontend | 1 |
77,036 | 7,552,537,940 | IssuesEvent | 2018-04-19 00:53:55 | trilinos/Trilinos | https://api.github.com/repos/trilinos/Trilinos | closed | Build error in MueLu unit tests | MueLu tests | In the Albany CDash, I noticed Trilinos failed to build on one platform, Cori, due to MueLu tests:
https://my.cdash.org/viewBuildError.php?buildid=1373917
I can just turn this off for Albany b/c we do not really need them, but I wanted to report the problem.
@trilinos/muelu | 1.0 | Build error in MueLu unit tests - In the Albany CDash, I noticed Trilinos failed to build on one platform, Cori, due to MueLu tests:
https://my.cdash.org/viewBuildError.php?buildid=1373917
I can just turn this off for Albany b/c we do not really need them, but I wanted to report the problem.
@trilinos/muelu | non_main | build error in muelu unit tests in the albany cdash i noticed trilinos failed to build on one platform cori due to muelu tests i can just turn this off for albany b c we do not really need them but i wanted to report the problem trilinos muelu | 0 |
1,713 | 6,574,460,288 | IssuesEvent | 2017-09-11 12:58:46 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | docker_image returns changed even if the image is in the latest version | affects_2.2 bug_report cloud docker waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`docker_image`
##### ANSIBLE VERSION
```
ansible 2.2.0.0
config file = /home/albertom/ciao/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
None
##### OS / ENVIRONMENT
N/A
##### SUMMARY
Pulling a docker image with `force=yes` always returns `changed` even if the image is in its latest version.
##### STEPS TO REPRODUCE
The following task always return changed even when the image is in its latest version.
```
tasks:
- name: Download busybox image
docker_image: name=busybox state=present force=yes
```
But running `docker pull` from shell you could tell if the image was updated or not
```
# docker pull busybox
Using default tag: latest
latest: Pulling from library/busybox
56bec22e3559: Pull complete
Digest: sha256:29f5d56d12684887bdfa50dcd29fc31eea4aaf4ad3bec43daf19026a7ce69912
Status: Downloaded newer image for busybox:latest
# docker pull busybox
Using default tag: latest
latest: Pulling from library/busybox
Digest: sha256:29f5d56d12684887bdfa50dcd29fc31eea4aaf4ad3bec43daf19026a7ce69912
Status: Image is up to date for busybox:latest
```
##### EXPECTED RESULTS
It should return unchanged if the image was not updated.
##### ACTUAL RESULTS
```
TASK [Download Download busybox image] *********************************
Thursday 03 November 2016 11:03:34 -0500 (0:00:00.567) 0:00:00.580 *****
changed: [wonderwoman.intel.com] => {"actions": ["Pulled image busybox:latest"], "changed": true, "image": {"Architecture": "amd64", "Author": "", "Comment": "", "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["sh"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "4a74292706a0", "Image": "sha256:1679bae2167496818312013654f5c66a16e185d0a0f6b762b53c8558014457c6", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Container": "8bb318a3b4672c53a1747991c95fff3306eea13ec308740ebe0c81b56ece530f", "ContainerConfig": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/sh", "-c", "#(nop) ", "CMD [\"sh\"]"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "4a74292706a0", "Image": "sha256:1679bae2167496818312013654f5c66a16e185d0a0f6b762b53c8558014457c6", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-10-07T21:03:58.469866982Z", "DockerVersion": "1.12.1", "GraphDriver": {"Data": null, "Name": "aufs"}, "Id": "sha256:e02e811dd08fd49e7f6032625495118e63f597eb150403d02e3238af1df240ba", "Os": "linux", "Parent": "", "RepoDigests": ["busybox@sha256:29f5d56d12684887bdfa50dcd29fc31eea4aaf4ad3bec43daf19026a7ce69912"], "RepoTags": ["busybox:latest"], "RootFS": {"Layers": ["sha256:e88b3f82283bc59d5e0df427c824e9f95557e661fcb0ea15fb0fb6f97760f9d9"], "Type": "layers"}, "Size": 1093484, "VirtualSize": 1093484}}
```
| True | docker_image returns changed even if the image is in the latest version - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`docker_image`
##### ANSIBLE VERSION
```
ansible 2.2.0.0
config file = /home/albertom/ciao/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
None
##### OS / ENVIRONMENT
N/A
##### SUMMARY
Pulling a docker image with `force=yes` always returns `changed` even if the image is in its latest version.
##### STEPS TO REPRODUCE
The following task always return changed even when the image is in its latest version.
```
tasks:
- name: Download busybox image
docker_image: name=busybox state=present force=yes
```
But running `docker pull` from shell you could tell if the image was updated or not
```
# docker pull busybox
Using default tag: latest
latest: Pulling from library/busybox
56bec22e3559: Pull complete
Digest: sha256:29f5d56d12684887bdfa50dcd29fc31eea4aaf4ad3bec43daf19026a7ce69912
Status: Downloaded newer image for busybox:latest
# docker pull busybox
Using default tag: latest
latest: Pulling from library/busybox
Digest: sha256:29f5d56d12684887bdfa50dcd29fc31eea4aaf4ad3bec43daf19026a7ce69912
Status: Image is up to date for busybox:latest
```
##### EXPECTED RESULTS
It should return unchanged if the image was not updated.
##### ACTUAL RESULTS
```
TASK [Download Download busybox image] *********************************
Thursday 03 November 2016 11:03:34 -0500 (0:00:00.567) 0:00:00.580 *****
changed: [wonderwoman.intel.com] => {"actions": ["Pulled image busybox:latest"], "changed": true, "image": {"Architecture": "amd64", "Author": "", "Comment": "", "Config": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["sh"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "4a74292706a0", "Image": "sha256:1679bae2167496818312013654f5c66a16e185d0a0f6b762b53c8558014457c6", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Container": "8bb318a3b4672c53a1747991c95fff3306eea13ec308740ebe0c81b56ece530f", "ContainerConfig": {"AttachStderr": false, "AttachStdin": false, "AttachStdout": false, "Cmd": ["/bin/sh", "-c", "#(nop) ", "CMD [\"sh\"]"], "Domainname": "", "Entrypoint": null, "Env": ["PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin"], "Hostname": "4a74292706a0", "Image": "sha256:1679bae2167496818312013654f5c66a16e185d0a0f6b762b53c8558014457c6", "Labels": {}, "OnBuild": null, "OpenStdin": false, "StdinOnce": false, "Tty": false, "User": "", "Volumes": null, "WorkingDir": ""}, "Created": "2016-10-07T21:03:58.469866982Z", "DockerVersion": "1.12.1", "GraphDriver": {"Data": null, "Name": "aufs"}, "Id": "sha256:e02e811dd08fd49e7f6032625495118e63f597eb150403d02e3238af1df240ba", "Os": "linux", "Parent": "", "RepoDigests": ["busybox@sha256:29f5d56d12684887bdfa50dcd29fc31eea4aaf4ad3bec43daf19026a7ce69912"], "RepoTags": ["busybox:latest"], "RootFS": {"Layers": ["sha256:e88b3f82283bc59d5e0df427c824e9f95557e661fcb0ea15fb0fb6f97760f9d9"], "Type": "layers"}, "Size": 1093484, "VirtualSize": 1093484}}
```
| main | docker image returns changed even if the image is in the latest version issue type bug report component name docker image ansible version ansible config file home albertom ciao ansible cfg configured module search path default w o overrides configuration none os environment n a summary pulling a docker image with force yes always returns changed even if the image is in its latest version steps to reproduce the following task always return changed even when the image is in its latest version tasks name download busybox image docker image name busybox state present force yes but running docker pull from shell you could tell if the image was updated or not docker pull busybox using default tag latest latest pulling from library busybox pull complete digest status downloaded newer image for busybox latest docker pull busybox using default tag latest latest pulling from library busybox digest status image is up to date for busybox latest expected results it should return unchanged if the image was not updated actual results task thursday november changed actions changed true image architecture author comment config attachstderr false attachstdin false attachstdout false cmd domainname entrypoint null env hostname image labels onbuild null openstdin false stdinonce false tty false user volumes null workingdir container containerconfig attachstderr false attachstdin false attachstdout false cmd domainname entrypoint null env hostname image labels onbuild null openstdin false stdinonce false tty false user volumes null workingdir created dockerversion graphdriver data null name aufs id os linux parent repodigests repotags rootfs layers type layers size virtualsize | 1 |
364,164 | 25,482,895,362 | IssuesEvent | 2022-11-26 01:53:01 | Jovenasso/SistemaJudocas2022 | https://api.github.com/repos/Jovenasso/SistemaJudocas2022 | closed | Documentação - Especificação de requisitos - 4.5 - Requisitos detalhados de casos de uso - não coerente | documentation Ambiguidade | Tabela de descrição do item 4.5 - Requisitos detalhados de caso de uso incompleto

| 1.0 | Documentação - Especificação de requisitos - 4.5 - Requisitos detalhados de casos de uso - não coerente - Tabela de descrição do item 4.5 - Requisitos detalhados de caso de uso incompleto

| non_main | documentação especificação de requisitos requisitos detalhados de casos de uso não coerente tabela de descrição do item requisitos detalhados de caso de uso incompleto | 0 |
139,996 | 18,869,741,378 | IssuesEvent | 2021-11-13 01:18:50 | gms-ws-sandbox/nibrs | https://api.github.com/repos/gms-ws-sandbox/nibrs | closed | CVE-2019-17571 (High) detected in log4j-core-2.7.jar, log4j-core-2.6.jar - autoclosed | security vulnerability | ## CVE-2019-17571 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>log4j-core-2.7.jar</b>, <b>log4j-core-2.6.jar</b></p></summary>
<p>
<details><summary><b>log4j-core-2.7.jar</b></p></summary>
<p>The Apache Log4j Implementation</p>
<p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/log4j-core-2.7.jar,/home/wss-scanner/.m2/repository/org/apache/logging/log4j/log4j-core/2.7/log4j-core-2.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-core-2.7.jar** (Vulnerable Library)
</details>
<details><summary><b>log4j-core-2.6.jar</b></p></summary>
<p>The Apache Log4j Implementation</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/2.x/">http://logging.apache.org/log4j/2.x/</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-common/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/org/apache/logging/log4j/log4j-core/2.6/log4j-core-2.6.jar,/home/wss-scanner/.m2/repository/org/apache/logging/log4j/log4j-core/2.6/log4j-core-2.6.jar,canner/.m2/repository/org/apache/logging/log4j/log4j-core/2.6/log4j-core-2.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-core-2.6.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.7","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.7","isMinimumFixVersionAvailable":false},{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.6","packageFilePaths":["/tools/nibrs-common/pom.xml","/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.6","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-17571","vulnerabilityDetails":"Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-17571 (High) detected in log4j-core-2.7.jar, log4j-core-2.6.jar - autoclosed - ## CVE-2019-17571 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>log4j-core-2.7.jar</b>, <b>log4j-core-2.6.jar</b></p></summary>
<p>
<details><summary><b>log4j-core-2.7.jar</b></p></summary>
<p>The Apache Log4j Implementation</p>
<p>Path to dependency file: nibrs/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: nibrs/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/log4j-core-2.7.jar,/home/wss-scanner/.m2/repository/org/apache/logging/log4j/log4j-core/2.7/log4j-core-2.7.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-core-2.7.jar** (Vulnerable Library)
</details>
<details><summary><b>log4j-core-2.6.jar</b></p></summary>
<p>The Apache Log4j Implementation</p>
<p>Library home page: <a href="http://logging.apache.org/log4j/2.x/">http://logging.apache.org/log4j/2.x/</a></p>
<p>Path to dependency file: nibrs/tools/nibrs-common/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/org/apache/logging/log4j/log4j-core/2.6/log4j-core-2.6.jar,/home/wss-scanner/.m2/repository/org/apache/logging/log4j/log4j-core/2.6/log4j-core-2.6.jar,canner/.m2/repository/org/apache/logging/log4j/log4j-core/2.6/log4j-core-2.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **log4j-core-2.6.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.
<p>Publish Date: 2019-12-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571>CVE-2019-17571</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.7","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.7","isMinimumFixVersionAvailable":false},{"packageType":"Java","groupId":"org.apache.logging.log4j","packageName":"log4j-core","packageVersion":"2.6","packageFilePaths":["/tools/nibrs-common/pom.xml","/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"org.apache.logging.log4j:log4j-core:2.6","isMinimumFixVersionAvailable":false}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-17571","vulnerabilityDetails":"Included in Log4j 1.2 is a SocketServer class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data. This affects Log4j versions up to 1.2 up to 1.2.17.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-17571","cvss3Severity":"high","cvss3Score":"9.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_main | cve high detected in core jar core jar autoclosed cve high severity vulnerability vulnerable libraries core jar core jar core jar the apache implementation path to dependency file nibrs tools nibrs fbi service pom xml path to vulnerable library nibrs tools nibrs fbi service target nibrs fbi service web inf lib core jar home wss scanner repository org apache logging core core jar dependency hierarchy x core jar vulnerable library core jar the apache implementation library home page a href path to dependency file nibrs tools nibrs common pom xml path to vulnerable library canner repository org apache logging core core jar home wss scanner repository org apache logging core core jar canner repository org apache logging core core jar dependency hierarchy x core jar vulnerable library found in base branch master vulnerability details included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree org apache logging core isminimumfixversionavailable false packagetype java groupid org apache logging packagename core packageversion packagefilepaths istransitivedependency false dependencytree org apache logging core isminimumfixversionavailable false basebranches vulnerabilityidentifier cve vulnerabilitydetails included in is a socketserver class that is vulnerable to deserialization of untrusted data which can be exploited to remotely execute arbitrary code when combined with a deserialization gadget when listening to untrusted network traffic for log data this affects versions up to up to vulnerabilityurl | 0 |
4,984 | 25,589,398,107 | IssuesEvent | 2022-12-01 11:50:59 | libp2p/js-libp2p | https://api.github.com/repos/libp2p/js-libp2p | closed | Connection drop causes pubsub to stop working. | need/maintainer-input | <!--
Thank you for reporting an issue.
This issue tracker is for bugs found within the JavaScript implementation of libp2p.
If you are asking a question about how to use libp2p, please ask on https://discuss.libp2p.io
Otherwise please fill in as much of the template below as possible.
-->
- **Version**: 0.40.0
<!--
Check package.json version
-->
- **Platform**: MacOS, Node v16.14.0
<!--
Output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows). If using in a Browser, please share the browser version as well
-->
- **Subsystem**: Pubsub, Relay, ConnectionManager
<!--
If known, please specify affected core module name (e.g Dialer, Pubsub, Relay etc)
-->
#### Severity: High
<!--
One of following:
Critical - System crash, application panic.
High - The main functionality of the application does not work, API breakage, repo format breakage, etc.
Medium - A non-essential functionality does not work, performance issues, etc.
Low - An optional functionality does not work.
Very Low - Translation or documentation mistake. Something that won't give anyone a bad day.
-->
#### Description:
<!--
- What you did
- What happened
- What you expected to happen
-->
I have two peers - Alice and Bob, both connected to the same relay server and subscribed to the same topic.
I need to find all peers subscribed to a given topic, e.g. I call `libp2p.pubsub.getSubscribers("my_topic")` from the Alice side and expect to see Bob's peer id in response.
When we run both peers the first time, everything works as expected - peers sees each other in topic subscribers. But when some of the peers reconnects (doesn't matter Alice or Bob), they does not see each other anymore. But if I restart one of the peers manually, everything works as expected until next reconnect.
I am using `gossibsub` for pubsub, but also tried `floodsub` with the same result.
My issue is very similar to #658.
#### Steps to reproduce the error:
<!--
If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able
-->
I created [reproduction repository](https://github.com/AlexMesser/pubsub-issue-reproduction) for better explanation.
| True | Connection drop causes pubsub to stop working. - <!--
Thank you for reporting an issue.
This issue tracker is for bugs found within the JavaScript implementation of libp2p.
If you are asking a question about how to use libp2p, please ask on https://discuss.libp2p.io
Otherwise please fill in as much of the template below as possible.
-->
- **Version**: 0.40.0
<!--
Check package.json version
-->
- **Platform**: MacOS, Node v16.14.0
<!--
Output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows). If using in a Browser, please share the browser version as well
-->
- **Subsystem**: Pubsub, Relay, ConnectionManager
<!--
If known, please specify affected core module name (e.g Dialer, Pubsub, Relay etc)
-->
#### Severity: High
<!--
One of following:
Critical - System crash, application panic.
High - The main functionality of the application does not work, API breakage, repo format breakage, etc.
Medium - A non-essential functionality does not work, performance issues, etc.
Low - An optional functionality does not work.
Very Low - Translation or documentation mistake. Something that won't give anyone a bad day.
-->
#### Description:
<!--
- What you did
- What happened
- What you expected to happen
-->
I have two peers - Alice and Bob, both connected to the same relay server and subscribed to the same topic.
I need to find all peers subscribed to a given topic, e.g. I call `libp2p.pubsub.getSubscribers("my_topic")` from the Alice side and expect to see Bob's peer id in response.
When we run both peers the first time, everything works as expected - peers sees each other in topic subscribers. But when some of the peers reconnects (doesn't matter Alice or Bob), they does not see each other anymore. But if I restart one of the peers manually, everything works as expected until next reconnect.
I am using `gossibsub` for pubsub, but also tried `floodsub` with the same result.
My issue is very similar to #658.
#### Steps to reproduce the error:
<!--
If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able
-->
I created [reproduction repository](https://github.com/AlexMesser/pubsub-issue-reproduction) for better explanation.
| main | connection drop causes pubsub to stop working thank you for reporting an issue this issue tracker is for bugs found within the javascript implementation of if you are asking a question about how to use please ask on otherwise please fill in as much of the template below as possible version check package json version platform macos node output of uname a unix or version and or bit windows if using in a browser please share the browser version as well subsystem pubsub relay connectionmanager if known please specify affected core module name e g dialer pubsub relay etc severity high one of following critical system crash application panic high the main functionality of the application does not work api breakage repo format breakage etc medium a non essential functionality does not work performance issues etc low an optional functionality does not work very low translation or documentation mistake something that won t give anyone a bad day description what you did what happened what you expected to happen i have two peers alice and bob both connected to the same relay server and subscribed to the same topic i need to find all peers subscribed to a given topic e g i call pubsub getsubscribers my topic from the alice side and expect to see bob s peer id in response when we run both peers the first time everything works as expected peers sees each other in topic subscribers but when some of the peers reconnects doesn t matter alice or bob they does not see each other anymore but if i restart one of the peers manually everything works as expected until next reconnect i am using gossibsub for pubsub but also tried floodsub with the same result my issue is very similar to steps to reproduce the error if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able i created for better explanation | 1 |
104,636 | 16,619,976,175 | IssuesEvent | 2021-06-02 22:32:20 | jtimberlake/tds-core | https://api.github.com/repos/jtimberlake/tds-core | opened | CVE-2020-15138 (High) detected in prismjs-1.6.0.tgz, prismjs-1.19.0.tgz | security vulnerability | ## CVE-2020-15138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>prismjs-1.6.0.tgz</b>, <b>prismjs-1.19.0.tgz</b></p></summary>
<p>
<details><summary><b>prismjs-1.6.0.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.6.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.6.0.tgz</a></p>
<p>Path to dependency file: tds-core/package.json</p>
<p>Path to vulnerable library: tds-core/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- kit-1.0.47.tgz (Root Library)
- react-live-1.12.0.tgz
- :x: **prismjs-1.6.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>prismjs-1.19.0.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.19.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.19.0.tgz</a></p>
<p>Path to dependency file: tds-core/package.json</p>
<p>Path to vulnerable library: tds-core/node_modules/react-styleguidist/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- react-styleguidist-10.6.1.tgz (Root Library)
- :x: **prismjs-1.19.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/tds-core/commit/dea0d6818322df63aea1fb950dce6e8cb94c53ba">dea0d6818322df63aea1fb950dce6e8cb94c53ba</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prism is vulnerable to Cross-Site Scripting. The easing preview of the Previewers plugin has an XSS vulnerability that allows attackers to execute arbitrary code in Safari and Internet Explorer. This impacts all Safari and Internet Explorer users of Prism >=v1.1.0 that use the _Previewers_ plugin (>=v1.10.0) or the _Previewer: Easing_ plugin (v1.1.0 to v1.9.0). This problem is fixed in version 1.21.0. To workaround the issue without upgrading, disable the easing preview on all impacted code blocks. You need Prism v1.10.0 or newer to apply this workaround.
<p>Publish Date: 2020-08-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15138>CVE-2020-15138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/PrismJS/prism/v1.21.0">https://github.com/PrismJS/prism/v1.21.0</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: v1.21.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"prismjs","packageVersion":"1.6.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@compositor/kit:1.0.47;react-live:1.12.0;prismjs:1.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.21.0"},{"packageType":"javascript/Node.js","packageName":"prismjs","packageVersion":"1.19.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-styleguidist:10.6.1;prismjs:1.19.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.21.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-15138","vulnerabilityDetails":"Prism is vulnerable to Cross-Site Scripting. The easing preview of the Previewers plugin has an XSS vulnerability that allows attackers to execute arbitrary code in Safari and Internet Explorer. This impacts all Safari and Internet Explorer users of Prism \u003e\u003dv1.1.0 that use the _Previewers_ plugin (\u003e\u003dv1.10.0) or the _Previewer: Easing_ plugin (v1.1.0 to v1.9.0). This problem is fixed in version 1.21.0. To workaround the issue without upgrading, disable the easing preview on all impacted code blocks. You need Prism v1.10.0 or newer to apply this workaround.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Changed","C":"High","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-15138 (High) detected in prismjs-1.6.0.tgz, prismjs-1.19.0.tgz - ## CVE-2020-15138 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>prismjs-1.6.0.tgz</b>, <b>prismjs-1.19.0.tgz</b></p></summary>
<p>
<details><summary><b>prismjs-1.6.0.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.6.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.6.0.tgz</a></p>
<p>Path to dependency file: tds-core/package.json</p>
<p>Path to vulnerable library: tds-core/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- kit-1.0.47.tgz (Root Library)
- react-live-1.12.0.tgz
- :x: **prismjs-1.6.0.tgz** (Vulnerable Library)
</details>
<details><summary><b>prismjs-1.19.0.tgz</b></p></summary>
<p>Lightweight, robust, elegant syntax highlighting. A spin-off project from Dabblet.</p>
<p>Library home page: <a href="https://registry.npmjs.org/prismjs/-/prismjs-1.19.0.tgz">https://registry.npmjs.org/prismjs/-/prismjs-1.19.0.tgz</a></p>
<p>Path to dependency file: tds-core/package.json</p>
<p>Path to vulnerable library: tds-core/node_modules/react-styleguidist/node_modules/prismjs/package.json</p>
<p>
Dependency Hierarchy:
- react-styleguidist-10.6.1.tgz (Root Library)
- :x: **prismjs-1.19.0.tgz** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/jtimberlake/tds-core/commit/dea0d6818322df63aea1fb950dce6e8cb94c53ba">dea0d6818322df63aea1fb950dce6e8cb94c53ba</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Prism is vulnerable to Cross-Site Scripting. The easing preview of the Previewers plugin has an XSS vulnerability that allows attackers to execute arbitrary code in Safari and Internet Explorer. This impacts all Safari and Internet Explorer users of Prism >=v1.1.0 that use the _Previewers_ plugin (>=v1.10.0) or the _Previewer: Easing_ plugin (v1.1.0 to v1.9.0). This problem is fixed in version 1.21.0. To workaround the issue without upgrading, disable the easing preview on all impacted code blocks. You need Prism v1.10.0 or newer to apply this workaround.
<p>Publish Date: 2020-08-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15138>CVE-2020-15138</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: Low
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/PrismJS/prism/v1.21.0">https://github.com/PrismJS/prism/v1.21.0</a></p>
<p>Release Date: 2020-07-21</p>
<p>Fix Resolution: v1.21.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"prismjs","packageVersion":"1.6.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"@compositor/kit:1.0.47;react-live:1.12.0;prismjs:1.6.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.21.0"},{"packageType":"javascript/Node.js","packageName":"prismjs","packageVersion":"1.19.0","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"react-styleguidist:10.6.1;prismjs:1.19.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"v1.21.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-15138","vulnerabilityDetails":"Prism is vulnerable to Cross-Site Scripting. The easing preview of the Previewers plugin has an XSS vulnerability that allows attackers to execute arbitrary code in Safari and Internet Explorer. This impacts all Safari and Internet Explorer users of Prism \u003e\u003dv1.1.0 that use the _Previewers_ plugin (\u003e\u003dv1.10.0) or the _Previewer: Easing_ plugin (v1.1.0 to v1.9.0). This problem is fixed in version 1.21.0. To workaround the issue without upgrading, disable the easing preview on all impacted code blocks. You need Prism v1.10.0 or newer to apply this workaround.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-15138","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"Low","AC":"High","PR":"None","S":"Changed","C":"High","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_main | cve high detected in prismjs tgz prismjs tgz cve high severity vulnerability vulnerable libraries prismjs tgz prismjs tgz prismjs tgz lightweight robust elegant syntax highlighting a spin off project from dabblet library home page a href path to dependency file tds core package json path to vulnerable library tds core node modules prismjs package json dependency hierarchy kit tgz root library react live tgz x prismjs tgz vulnerable library prismjs tgz lightweight robust elegant syntax highlighting a spin off project from dabblet library home page a href path to dependency file tds core package json path to vulnerable library tds core node modules react styleguidist node modules prismjs package json dependency hierarchy react styleguidist tgz root library x prismjs tgz vulnerable library found in head commit a href found in base branch master vulnerability details prism is vulnerable to cross site scripting the easing preview of the previewers plugin has an xss vulnerability that allows attackers to execute arbitrary code in safari and internet explorer this impacts all safari and internet explorer users of prism that use the previewers plugin or the previewer easing plugin to this problem is fixed in version to workaround the issue without upgrading disable the easing preview on all impacted code blocks you need prism or newer to apply this workaround publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope changed impact metrics confidentiality impact high integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree compositor kit react live prismjs isminimumfixversionavailable true minimumfixversion packagetype javascript node js packagename prismjs packageversion packagefilepaths istransitivedependency true dependencytree react styleguidist prismjs isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails prism is vulnerable to cross site scripting the easing preview of the previewers plugin has an xss vulnerability that allows attackers to execute arbitrary code in safari and internet explorer this impacts all safari and internet explorer users of prism that use the previewers plugin or the previewer easing plugin to this problem is fixed in version to workaround the issue without upgrading disable the easing preview on all impacted code blocks you need prism or newer to apply this workaround vulnerabilityurl | 0 |
569 | 7,929,781,797 | IssuesEvent | 2018-07-06 16:13:24 | chapel-lang/chapel | https://api.github.com/repos/chapel-lang/chapel | closed | Mac OS X: gasnet w segment=fast segfaults. Segment=everything works. | area: Compiler type: Portability | ### Summary of Problem
On a macbook pro running Sierra, Chapel w CHPL_COMM=gasnet and CHPL_GASNET_SEGMENT=fast, segfaults when you run the hellos, e.g. "make check"
```
== Actual Test Output (raw, with verbose) ==
/Users/bogus/workspace/correctness-test-gasnet.fast.darwin/third-party/gasnet/install/darwin-clang-none/seg-fast/nodbg/bin/amudprun -np 4 ./hello6-taskpar-dist_real -nl4 --printLocaleName=false -v
QTHREADS: Using 4 Shepherds
QTHREADS: Using 4 Shepherds
QTHREADS: Using 4 Shepherds
QTHREADS: Using 1 Workers per Shepherd
QTHREADS: Using 4 Shepherds
QTHREADS: Using 1 Workers per Shepherd
QTHREADS: Using 1 Workers per Shepherd
QTHREADS: Using 1 Workers per Shepherd
QTHREADS: Guard Pages Enabled
QTHREADS: Guard Pages Enabled
QTHREADS: Guard Pages Enabled
QTHREADS: Using 8388608 byte stack size.
QTHREADS: Guard Pages Enabled
QTHREADS: Using 8388608 byte stack size.
QTHREADS: Using 8388608 byte stack size.
QTHREADS: Using 8388608 byte stack size.
executing on node 2 of 4 node(s): mbp-awallace.local
executing on node 1 of 4 node(s): mbp-awallace.local
executing on node 3 of 4 node(s): mbp-awallace.local
executing on node 0 of 4 node(s): mbp-awallace.local
*** Caught a fatal signal: SIGSEGV(11) on node 0/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
*** Caught a fatal signal: SIGSEGV(11) on node 3/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
*** Caught a fatal signal: SIGSEGV(11) on node 2/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
*** Caught a fatal signal: SIGSEGV(11) on node 1/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
make: *** [check] Error 20
```
Change CHPL_GASNET_SEGMENT=everything (or just unset the variable), holding everything else the same, rebuild Chapel from "git clean", and the "hellos" work.
### Steps to Reproduce
**Source Code:**
Chapel master branch, Git rev 4bd4e822f
Chapel release/1.16 branch, Git rev f0bd20812 (not shown)
**Compile command:**
`make`
**Execution command:**
`make check`
### Configuration Information
- Output of `chpl --version`:
`chpl Version 1.17.0 pre-release (4bd4e822f1)`
- Output of `$CHPL_HOME/util/printchplenv --anonymize`:
```
CHPL_TARGET_PLATFORM: darwin
CHPL_TARGET_COMPILER: clang
CHPL_TARGET_ARCH: none *
CHPL_LOCALE_MODEL: flat
CHPL_COMM: gasnet *
CHPL_COMM_SUBSTRATE: udp
CHPL_GASNET_SEGMENT: fast *
CHPL_TASKS: qthreads
CHPL_LAUNCHER: amudprun
CHPL_TIMERS: generic
CHPL_UNWIND: none
CHPL_MEM: jemalloc
CHPL_MAKE: make
CHPL_ATOMICS: intrinsics
CHPL_NETWORK_ATOMICS: none
CHPL_GMP: gmp
CHPL_HWLOC: hwloc
CHPL_REGEXP: re2
CHPL_AUX_FILESYS: none
```
- environment
```bash
CHPL_COMM=gasnet
CHPL_GASNET_CFG_OPTIONS=--disable-ibv
CHPL_GASNET_SEGMENT=fast
CHPL_HOME=/Users/bogus/workspace/correctness-test-gasnet.fast.darwin
CHPL_HOST_PLATFORM=darwin
CHPL_QTHREAD_ENABLE_OVERSUBSCRIPTION=1
CHPL_TARGET_ARCH=none
GASNET_ROUTE_OUTPUT=0
GASNET_SPAWNFN=L
MANPATH=/Users/bogus/workspace/correctness-test-gasnet.fast.darwin/man:
PATH=/Users/bogus/workspace/correctness-test-gasnet.fast.darwin/bin/darwin:/Users/bogus/workspace/correctness-test-gasnet.fast.darwin/util:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin
QT_AFFINITY=no
```
- Back-end compiler and version, `clang --version`:
```
Apple LLVM version 9.0.0 (clang-900.0.38)
Target: x86_64-apple-darwin16.7.0
```
## Mac Firewall: DISABLED
Otherwise, you get pop-ups asking to allow "amudprun" to accept incoming connections
## VPN: DISCONNECTED
Otherwise, you get errors like this:
```
*** FATAL ERROR(Node 0): An active message was returned to sender,
and trapped by the default returned message handler (handler 0):
Message type: AM_REQUEST_IM
Error Code: ECONGESTION: Congestion at destination endpoint
...
*** Caught a fatal signal: SIGABRT(6) on node 0/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
```
| True | Mac OS X: gasnet w segment=fast segfaults. Segment=everything works. - ### Summary of Problem
On a macbook pro running Sierra, Chapel w CHPL_COMM=gasnet and CHPL_GASNET_SEGMENT=fast, segfaults when you run the hellos, e.g. "make check"
```
== Actual Test Output (raw, with verbose) ==
/Users/bogus/workspace/correctness-test-gasnet.fast.darwin/third-party/gasnet/install/darwin-clang-none/seg-fast/nodbg/bin/amudprun -np 4 ./hello6-taskpar-dist_real -nl4 --printLocaleName=false -v
QTHREADS: Using 4 Shepherds
QTHREADS: Using 4 Shepherds
QTHREADS: Using 4 Shepherds
QTHREADS: Using 1 Workers per Shepherd
QTHREADS: Using 4 Shepherds
QTHREADS: Using 1 Workers per Shepherd
QTHREADS: Using 1 Workers per Shepherd
QTHREADS: Using 1 Workers per Shepherd
QTHREADS: Guard Pages Enabled
QTHREADS: Guard Pages Enabled
QTHREADS: Guard Pages Enabled
QTHREADS: Using 8388608 byte stack size.
QTHREADS: Guard Pages Enabled
QTHREADS: Using 8388608 byte stack size.
QTHREADS: Using 8388608 byte stack size.
QTHREADS: Using 8388608 byte stack size.
executing on node 2 of 4 node(s): mbp-awallace.local
executing on node 1 of 4 node(s): mbp-awallace.local
executing on node 3 of 4 node(s): mbp-awallace.local
executing on node 0 of 4 node(s): mbp-awallace.local
*** Caught a fatal signal: SIGSEGV(11) on node 0/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
*** Caught a fatal signal: SIGSEGV(11) on node 3/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
*** Caught a fatal signal: SIGSEGV(11) on node 2/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
*** Caught a fatal signal: SIGSEGV(11) on node 1/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
make: *** [check] Error 20
```
Change CHPL_GASNET_SEGMENT=everything (or just unset the variable), holding everything else the same, rebuild Chapel from "git clean", and the "hellos" work.
### Steps to Reproduce
**Source Code:**
Chapel master branch, Git rev 4bd4e822f
Chapel release/1.16 branch, Git rev f0bd20812 (not shown)
**Compile command:**
`make`
**Execution command:**
`make check`
### Configuration Information
- Output of `chpl --version`:
`chpl Version 1.17.0 pre-release (4bd4e822f1)`
- Output of `$CHPL_HOME/util/printchplenv --anonymize`:
```
CHPL_TARGET_PLATFORM: darwin
CHPL_TARGET_COMPILER: clang
CHPL_TARGET_ARCH: none *
CHPL_LOCALE_MODEL: flat
CHPL_COMM: gasnet *
CHPL_COMM_SUBSTRATE: udp
CHPL_GASNET_SEGMENT: fast *
CHPL_TASKS: qthreads
CHPL_LAUNCHER: amudprun
CHPL_TIMERS: generic
CHPL_UNWIND: none
CHPL_MEM: jemalloc
CHPL_MAKE: make
CHPL_ATOMICS: intrinsics
CHPL_NETWORK_ATOMICS: none
CHPL_GMP: gmp
CHPL_HWLOC: hwloc
CHPL_REGEXP: re2
CHPL_AUX_FILESYS: none
```
- environment
```bash
CHPL_COMM=gasnet
CHPL_GASNET_CFG_OPTIONS=--disable-ibv
CHPL_GASNET_SEGMENT=fast
CHPL_HOME=/Users/bogus/workspace/correctness-test-gasnet.fast.darwin
CHPL_HOST_PLATFORM=darwin
CHPL_QTHREAD_ENABLE_OVERSUBSCRIPTION=1
CHPL_TARGET_ARCH=none
GASNET_ROUTE_OUTPUT=0
GASNET_SPAWNFN=L
MANPATH=/Users/bogus/workspace/correctness-test-gasnet.fast.darwin/man:
PATH=/Users/bogus/workspace/correctness-test-gasnet.fast.darwin/bin/darwin:/Users/bogus/workspace/correctness-test-gasnet.fast.darwin/util:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin
QT_AFFINITY=no
```
- Back-end compiler and version, `clang --version`:
```
Apple LLVM version 9.0.0 (clang-900.0.38)
Target: x86_64-apple-darwin16.7.0
```
## Mac Firewall: DISABLED
Otherwise, you get pop-ups asking to allow "amudprun" to accept incoming connections
## VPN: DISCONNECTED
Otherwise, you get errors like this:
```
*** FATAL ERROR(Node 0): An active message was returned to sender,
and trapped by the default returned message handler (handler 0):
Message type: AM_REQUEST_IM
Error Code: ECONGESTION: Congestion at destination endpoint
...
*** Caught a fatal signal: SIGABRT(6) on node 0/4
NOTICE: Before reporting bugs, run with GASNET_BACKTRACE=1 in the environment to generate a backtrace.
```
| non_main | mac os x gasnet w segment fast segfaults segment everything works summary of problem on a macbook pro running sierra chapel w chpl comm gasnet and chpl gasnet segment fast segfaults when you run the hellos e g make check actual test output raw with verbose users bogus workspace correctness test gasnet fast darwin third party gasnet install darwin clang none seg fast nodbg bin amudprun np taskpar dist real printlocalename false v qthreads using shepherds qthreads using shepherds qthreads using shepherds qthreads using workers per shepherd qthreads using shepherds qthreads using workers per shepherd qthreads using workers per shepherd qthreads using workers per shepherd qthreads guard pages enabled qthreads guard pages enabled qthreads guard pages enabled qthreads using byte stack size qthreads guard pages enabled qthreads using byte stack size qthreads using byte stack size qthreads using byte stack size executing on node of node s mbp awallace local executing on node of node s mbp awallace local executing on node of node s mbp awallace local executing on node of node s mbp awallace local caught a fatal signal sigsegv on node notice before reporting bugs run with gasnet backtrace in the environment to generate a backtrace caught a fatal signal sigsegv on node notice before reporting bugs run with gasnet backtrace in the environment to generate a backtrace caught a fatal signal sigsegv on node notice before reporting bugs run with gasnet backtrace in the environment to generate a backtrace caught a fatal signal sigsegv on node notice before reporting bugs run with gasnet backtrace in the environment to generate a backtrace make error change chpl gasnet segment everything or just unset the variable holding everything else the same rebuild chapel from git clean and the hellos work steps to reproduce source code chapel master branch git rev chapel release branch git rev not shown compile command make execution command make check configuration information output of chpl version chpl version pre release output of chpl home util printchplenv anonymize chpl target platform darwin chpl target compiler clang chpl target arch none chpl locale model flat chpl comm gasnet chpl comm substrate udp chpl gasnet segment fast chpl tasks qthreads chpl launcher amudprun chpl timers generic chpl unwind none chpl mem jemalloc chpl make make chpl atomics intrinsics chpl network atomics none chpl gmp gmp chpl hwloc hwloc chpl regexp chpl aux filesys none environment bash chpl comm gasnet chpl gasnet cfg options disable ibv chpl gasnet segment fast chpl home users bogus workspace correctness test gasnet fast darwin chpl host platform darwin chpl qthread enable oversubscription chpl target arch none gasnet route output gasnet spawnfn l manpath users bogus workspace correctness test gasnet fast darwin man path users bogus workspace correctness test gasnet fast darwin bin darwin users bogus workspace correctness test gasnet fast darwin util usr bin bin usr sbin sbin opt bin qt affinity no back end compiler and version clang version apple llvm version clang target apple mac firewall disabled otherwise you get pop ups asking to allow amudprun to accept incoming connections vpn disconnected otherwise you get errors like this fatal error node an active message was returned to sender and trapped by the default returned message handler handler message type am request im error code econgestion congestion at destination endpoint caught a fatal signal sigabrt on node notice before reporting bugs run with gasnet backtrace in the environment to generate a backtrace | 0 |
3,474 | 13,339,094,692 | IssuesEvent | 2020-08-28 12:17:36 | libgdx/libgdx | https://api.github.com/repos/libgdx/libgdx | closed | G3D models mostly white because of wrong emissive color | 3D maintainer needed | #### Issue details
Since #5077 emissive is now implemented (which is good) but there is a serious drawback. Several devs have an issue with their 3D model rendering looking brighter than before because of emissive information in their model files.
First of all, this change breaks in some way and i think it would be great to add something like this in CHANGES.MD (kind of migration guide) :
```
[1.9.9]
- Emissive texture and color are now implemented in the default 3D shader. 3D models files(g3d, gdb, obj)
may have unwanted emissive information and have to be migrated in order to look as before. In Blender, any material
with "shadeless" option, non-zero "Emit." value or "emit texture" will look brighter or totally white. If so, you
have to export them again without these emissive information. A workaround is to remove all emissive information
programmatically in your game : material.remove(ColorAttribute.Emissive).
```
I investigated and it appears that migration guide above is not enough. I'm pretty sure this is a bug related to fbx-conv. I'm posting here to centralize information and make it easily available to libgdx users. Here is what i found :
* Blender FBX plugin exports both Emissive color and Emissive factor. As you can see [here](https://github.com/sobotka/blender-addons/blob/blender-v2.79b-release/io_scene_fbx/export_fbx_bin.py#L1239) Emissive color is just a copy of Diffuse color and Emissive factor is the "Emit." factor from Blender.
* FBX-CONV only use Emissive color which is wrong IMO, see [here](https://github.com/libgdx/fbx-conv/blob/2be464f21823d2ea918b7b6587c6715076ddafdb/src/readers/FbxConverter.h#L563).
* I think it should be multiplied by Emissive factor. For [reference](http://docs.autodesk.com/FBX/2014/ENU/FBX-SDK-Documentation/cpp_ref/class_fbx_surface_lambert.html), FBX SDK provides both values.
If fbx-conv has to be fixed, it would be great to add a note about it in changes.md or another appropriate place to inform users about fbx-conv / libgdx version compatibility.
#### Reproduction steps/code
* open Blender with its default template (a box with a white material)
* add a dark diffuse texture (optional step to easily see the issue)
* export as FBX
* convert it with fbx-conv (json or binary)
* render that model in libgdx
* model will be mostly white
#### Version of LibGDX and/or relevant dependencies
* libgdx 1.9.9+
* fbx-conv 0.01.0055 x64 , FBXSDK 2014.02
* blender 2.79b
#### Stacktrace
N/A
#### Please select the affected platforms
- [x] Android
- [x] iOS (robovm)
- [x] iOS (MOE)
- [x] HTML/GWT
- [x] Windows
- [x] Linux
- [x] MacOS
| True | G3D models mostly white because of wrong emissive color - #### Issue details
Since #5077 emissive is now implemented (which is good) but there is a serious drawback. Several devs have an issue with their 3D model rendering looking brighter than before because of emissive information in their model files.
First of all, this change breaks in some way and i think it would be great to add something like this in CHANGES.MD (kind of migration guide) :
```
[1.9.9]
- Emissive texture and color are now implemented in the default 3D shader. 3D models files(g3d, gdb, obj)
may have unwanted emissive information and have to be migrated in order to look as before. In Blender, any material
with "shadeless" option, non-zero "Emit." value or "emit texture" will look brighter or totally white. If so, you
have to export them again without these emissive information. A workaround is to remove all emissive information
programmatically in your game : material.remove(ColorAttribute.Emissive).
```
I investigated and it appears that migration guide above is not enough. I'm pretty sure this is a bug related to fbx-conv. I'm posting here to centralize information and make it easily available to libgdx users. Here is what i found :
* Blender FBX plugin exports both Emissive color and Emissive factor. As you can see [here](https://github.com/sobotka/blender-addons/blob/blender-v2.79b-release/io_scene_fbx/export_fbx_bin.py#L1239) Emissive color is just a copy of Diffuse color and Emissive factor is the "Emit." factor from Blender.
* FBX-CONV only use Emissive color which is wrong IMO, see [here](https://github.com/libgdx/fbx-conv/blob/2be464f21823d2ea918b7b6587c6715076ddafdb/src/readers/FbxConverter.h#L563).
* I think it should be multiplied by Emissive factor. For [reference](http://docs.autodesk.com/FBX/2014/ENU/FBX-SDK-Documentation/cpp_ref/class_fbx_surface_lambert.html), FBX SDK provides both values.
If fbx-conv has to be fixed, it would be great to add a note about it in changes.md or another appropriate place to inform users about fbx-conv / libgdx version compatibility.
#### Reproduction steps/code
* open Blender with its default template (a box with a white material)
* add a dark diffuse texture (optional step to easily see the issue)
* export as FBX
* convert it with fbx-conv (json or binary)
* render that model in libgdx
* model will be mostly white
#### Version of LibGDX and/or relevant dependencies
* libgdx 1.9.9+
* fbx-conv 0.01.0055 x64 , FBXSDK 2014.02
* blender 2.79b
#### Stacktrace
N/A
#### Please select the affected platforms
- [x] Android
- [x] iOS (robovm)
- [x] iOS (MOE)
- [x] HTML/GWT
- [x] Windows
- [x] Linux
- [x] MacOS
| main | models mostly white because of wrong emissive color issue details since emissive is now implemented which is good but there is a serious drawback several devs have an issue with their model rendering looking brighter than before because of emissive information in their model files first of all this change breaks in some way and i think it would be great to add something like this in changes md kind of migration guide emissive texture and color are now implemented in the default shader models files gdb obj may have unwanted emissive information and have to be migrated in order to look as before in blender any material with shadeless option non zero emit value or emit texture will look brighter or totally white if so you have to export them again without these emissive information a workaround is to remove all emissive information programmatically in your game material remove colorattribute emissive i investigated and it appears that migration guide above is not enough i m pretty sure this is a bug related to fbx conv i m posting here to centralize information and make it easily available to libgdx users here is what i found blender fbx plugin exports both emissive color and emissive factor as you can see emissive color is just a copy of diffuse color and emissive factor is the emit factor from blender fbx conv only use emissive color which is wrong imo see i think it should be multiplied by emissive factor for fbx sdk provides both values if fbx conv has to be fixed it would be great to add a note about it in changes md or another appropriate place to inform users about fbx conv libgdx version compatibility reproduction steps code open blender with its default template a box with a white material add a dark diffuse texture optional step to easily see the issue export as fbx convert it with fbx conv json or binary render that model in libgdx model will be mostly white version of libgdx and or relevant dependencies libgdx fbx conv fbxsdk blender stacktrace n a please select the affected platforms android ios robovm ios moe html gwt windows linux macos | 1 |
4,648 | 24,075,815,836 | IssuesEvent | 2022-09-18 19:36:02 | microsoft/DirectXTex | https://api.github.com/repos/microsoft/DirectXTex | closed | Clang Warning Diagnostic: definition of implicit copy assignment operator for 'HDRColorA/INTColor' is deprecated because it has a user-provided copy constructor | maintainence | I ran into this error while compiling my project with with clang using `-Wall -Werror` and wanted to bring it to your attention.
An easy fix would be adding the following pragma to DirectXTexP.h:
```cpp
#pragma clang diagnostic ignored "-Wdeprecated-copy-with-user-provided-copy"
```
Or you can use default for the copy constructors, or remove them entirely, both types are POD so there's no need for a user-defined copy constructor:
```cpp
class HDRColorA
{
HDRColorA(const HDRColorA& c) noexcept = default;
};
class INTColor
{
INTColor(const INTColor& c) noexcept = default;
};
``` | True | Clang Warning Diagnostic: definition of implicit copy assignment operator for 'HDRColorA/INTColor' is deprecated because it has a user-provided copy constructor - I ran into this error while compiling my project with with clang using `-Wall -Werror` and wanted to bring it to your attention.
An easy fix would be adding the following pragma to DirectXTexP.h:
```cpp
#pragma clang diagnostic ignored "-Wdeprecated-copy-with-user-provided-copy"
```
Or you can use default for the copy constructors, or remove them entirely, both types are POD so there's no need for a user-defined copy constructor:
```cpp
class HDRColorA
{
HDRColorA(const HDRColorA& c) noexcept = default;
};
class INTColor
{
INTColor(const INTColor& c) noexcept = default;
};
``` | main | clang warning diagnostic definition of implicit copy assignment operator for hdrcolora intcolor is deprecated because it has a user provided copy constructor i ran into this error while compiling my project with with clang using wall werror and wanted to bring it to your attention an easy fix would be adding the following pragma to directxtexp h cpp pragma clang diagnostic ignored wdeprecated copy with user provided copy or you can use default for the copy constructors or remove them entirely both types are pod so there s no need for a user defined copy constructor cpp class hdrcolora hdrcolora const hdrcolora c noexcept default class intcolor intcolor const intcolor c noexcept default | 1 |
2,747 | 9,792,338,501 | IssuesEvent | 2019-06-10 17:09:25 | react-navigation/react-navigation | https://api.github.com/repos/react-navigation/react-navigation | closed | Modal dismissal in stack navigators with dynamic transitions behave incorrectly | needs response from maintainer | ### Current Behavior
- For screens specified as a modal in stack navigators with dynamic transition configurations, the gesture recognizer uses a push style (swipe left/right) as opposed as a modal style (swipe down) for dismissal. The screen itself still presents and dismisses from the bottom of the screen as expected.
### Expected Behavior
- I should not be able to swipe right to dismiss a modal
- I should be able to swipe down to dismiss a modal
### How to reproduce
This behavior can be reproduced by applying the [example dynamic config](https://reactnavigation.org/docs/en/stack-navigator.html#specifying-the-transition-mode-for-a-stack-s-screens-explicitly) to the [playground app](https://github.com/react-navigation/react-navigation/tree/master/examples/NavigationPlayground):
https://github.com/react-navigation/react-navigation/compare/master...alexpyoung:master
### Your Environment
```
React Native Environment Info:
System:
OS: macOS High Sierra 10.13.6
CPU: (8) x64 Intel(R) Core(TM) i7-4770HQ CPU @ 2.20GHz
Memory: 166.48 MB / 16.00 GB
Shell: 5.3 - /bin/zsh
Binaries:
Node: 8.15.1 - ~/.nvm/versions/node/v8.15.1/bin/node
Yarn: 1.15.2 - /usr/local/bin/yarn
npm: 6.4.1 - ~/.nvm/versions/node/v8.15.1/bin/npm
Watchman: 4.9.0 - /usr/local/bin/watchman
SDKs:
iOS SDK:
Platforms: iOS 12.1, macOS 10.14, tvOS 12.1, watchOS 5.1
IDEs:
Android Studio: 3.4 AI-183.5429.30.34.5452501
Xcode: 10.1/10B61 - /usr/bin/xcodebuild
npmPackages:
react: 16.5.0 => 16.5.0
react-native: https://github.com/expo/react-native/archive/sdk-32.0.0.tar.gz => 0.57.1
```
| True | Modal dismissal in stack navigators with dynamic transitions behave incorrectly - ### Current Behavior
- For screens specified as a modal in stack navigators with dynamic transition configurations, the gesture recognizer uses a push style (swipe left/right) as opposed as a modal style (swipe down) for dismissal. The screen itself still presents and dismisses from the bottom of the screen as expected.
### Expected Behavior
- I should not be able to swipe right to dismiss a modal
- I should be able to swipe down to dismiss a modal
### How to reproduce
This behavior can be reproduced by applying the [example dynamic config](https://reactnavigation.org/docs/en/stack-navigator.html#specifying-the-transition-mode-for-a-stack-s-screens-explicitly) to the [playground app](https://github.com/react-navigation/react-navigation/tree/master/examples/NavigationPlayground):
https://github.com/react-navigation/react-navigation/compare/master...alexpyoung:master
### Your Environment
```
React Native Environment Info:
System:
OS: macOS High Sierra 10.13.6
CPU: (8) x64 Intel(R) Core(TM) i7-4770HQ CPU @ 2.20GHz
Memory: 166.48 MB / 16.00 GB
Shell: 5.3 - /bin/zsh
Binaries:
Node: 8.15.1 - ~/.nvm/versions/node/v8.15.1/bin/node
Yarn: 1.15.2 - /usr/local/bin/yarn
npm: 6.4.1 - ~/.nvm/versions/node/v8.15.1/bin/npm
Watchman: 4.9.0 - /usr/local/bin/watchman
SDKs:
iOS SDK:
Platforms: iOS 12.1, macOS 10.14, tvOS 12.1, watchOS 5.1
IDEs:
Android Studio: 3.4 AI-183.5429.30.34.5452501
Xcode: 10.1/10B61 - /usr/bin/xcodebuild
npmPackages:
react: 16.5.0 => 16.5.0
react-native: https://github.com/expo/react-native/archive/sdk-32.0.0.tar.gz => 0.57.1
```
| main | modal dismissal in stack navigators with dynamic transitions behave incorrectly current behavior for screens specified as a modal in stack navigators with dynamic transition configurations the gesture recognizer uses a push style swipe left right as opposed as a modal style swipe down for dismissal the screen itself still presents and dismisses from the bottom of the screen as expected expected behavior i should not be able to swipe right to dismiss a modal i should be able to swipe down to dismiss a modal how to reproduce this behavior can be reproduced by applying the to the your environment react native environment info system os macos high sierra cpu intel r core tm cpu memory mb gb shell bin zsh binaries node nvm versions node bin node yarn usr local bin yarn npm nvm versions node bin npm watchman usr local bin watchman sdks ios sdk platforms ios macos tvos watchos ides android studio ai xcode usr bin xcodebuild npmpackages react react native | 1 |
1,178 | 5,096,335,992 | IssuesEvent | 2017-01-03 17:52:22 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | service module: ubuntu xenial on travis: failure 1 running systemctl show for 'apache2': Failed to connect to bus: No such file or directory | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
service module
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
No extra
##### OS / ENVIRONMENT
Docker in Travis, mapping ubuntu xenial in local connection
##### SUMMARY
when using service task to check one is started, task is failed with above message
##### STEPS TO REPRODUCE
See
https://travis-ci.org/juju4/ansible-icinga2/jobs/150094107
https://travis-ci.org/juju4/ansible-bro-ids/jobs/150093516
https://travis-ci.org/juju4/ansible-mhn/jobs/150159300
(only on target xenial, working on trusty)
On a local Vagrantfile with xenial (as found in github/#role#/test/vagrant): it's working normally...
So there is something with travis environment
##### EXPECTED RESULTS
test/vagrant local execution on xenial
```
TASK [bro-ids : check that mysql is running] ***********************************
task path: /home/julien/Documents/script/homelab/roles/bro-ids/tasks/main.yml:80
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -vvv -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2202 -o 'IdentityFile="/home/julien/Documents/script/homelab/roles/bro-ids/test/vagrant/.vagrant/machines/broids/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=30 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r 127.0.0.1 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936 `" && echo ansible-tmp-1470432978.69-241202336337936="` echo $HOME/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936 `" ) && sleep 0'"'"''
<127.0.0.1> PUT /tmp/tmp2IUx42 TO /home/vagrant/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936/service
<127.0.0.1> SSH: EXEC sftp -b - -C -vvv -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2202 -o 'IdentityFile="/home/julien/Documents/script/homelab/roles/bro-ids/test/vagrant/.vagrant/machines/broids/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=30 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r '[127.0.0.1]'
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -vvv -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2202 -o 'IdentityFile="/home/julien/Documents/script/homelab/roles/bro-ids/test/vagrant/.vagrant/machines/broids/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=30 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-drceqxkhrlhnomixtamaaklasdzascol; LANG=fr_FR.UTF-8 LC_ALL=fr_FR.UTF-8 LC_MESSAGES=fr_FR.UTF-8 /usr/bin/python /home/vagrant/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936/service; rm -rf "/home/vagrant/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
ok: [broids] => {"changed": false, "invocation": {"module_args": {"arguments": "", "enabled": null, "name": "mysql", "pattern": null, "runlevel": "default", "sleep": null, "state": "started"}, "module_name": "service"}, "name": "mysql", "state": "started"}
```
##### ACTUAL RESULTS
https://travis-ci.org/juju4/ansible-bro-ids/jobs/150166292
```
TASK [bro-ids : check that mysql is running] ***********************************
task path: /etc/ansible/roles/bro-ids/tasks/main.yml:80
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: root
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016 `" && echo ansible-tmp-1470433523.11-95534498356016="` echo $HOME/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016 `" ) && sleep 0'
<localhost> PUT /tmp/tmpMC1xcR TO /root/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016/service
<localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016/service; rm -rf "/root/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016/" > /dev/null 2>&1 && sleep 0'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"arguments": "", "enabled": null, "name": "mysql", "pattern": null, "runlevel": "default", "sleep": null, "state": "started"}, "module_name": "service"}, "msg": "failure 1 running systemctl show for 'mysql': Failed to connect to bus: No such file or directory\n"}
```
| True | service module: ubuntu xenial on travis: failure 1 running systemctl show for 'apache2': Failed to connect to bus: No such file or directory - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
service module
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
No extra
##### OS / ENVIRONMENT
Docker in Travis, mapping ubuntu xenial in local connection
##### SUMMARY
when using service task to check one is started, task is failed with above message
##### STEPS TO REPRODUCE
See
https://travis-ci.org/juju4/ansible-icinga2/jobs/150094107
https://travis-ci.org/juju4/ansible-bro-ids/jobs/150093516
https://travis-ci.org/juju4/ansible-mhn/jobs/150159300
(only on target xenial, working on trusty)
On a local Vagrantfile with xenial (as found in github/#role#/test/vagrant): it's working normally...
So there is something with travis environment
##### EXPECTED RESULTS
test/vagrant local execution on xenial
```
TASK [bro-ids : check that mysql is running] ***********************************
task path: /home/julien/Documents/script/homelab/roles/bro-ids/tasks/main.yml:80
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -vvv -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2202 -o 'IdentityFile="/home/julien/Documents/script/homelab/roles/bro-ids/test/vagrant/.vagrant/machines/broids/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=30 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r 127.0.0.1 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936 `" && echo ansible-tmp-1470432978.69-241202336337936="` echo $HOME/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936 `" ) && sleep 0'"'"''
<127.0.0.1> PUT /tmp/tmp2IUx42 TO /home/vagrant/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936/service
<127.0.0.1> SSH: EXEC sftp -b - -C -vvv -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2202 -o 'IdentityFile="/home/julien/Documents/script/homelab/roles/bro-ids/test/vagrant/.vagrant/machines/broids/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=30 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r '[127.0.0.1]'
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -vvv -o UserKnownHostsFile=/dev/null -o IdentitiesOnly=yes -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2202 -o 'IdentityFile="/home/julien/Documents/script/homelab/roles/bro-ids/test/vagrant/.vagrant/machines/broids/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=30 -o ControlPath=/home/julien/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-drceqxkhrlhnomixtamaaklasdzascol; LANG=fr_FR.UTF-8 LC_ALL=fr_FR.UTF-8 LC_MESSAGES=fr_FR.UTF-8 /usr/bin/python /home/vagrant/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936/service; rm -rf "/home/vagrant/.ansible/tmp/ansible-tmp-1470432978.69-241202336337936/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
ok: [broids] => {"changed": false, "invocation": {"module_args": {"arguments": "", "enabled": null, "name": "mysql", "pattern": null, "runlevel": "default", "sleep": null, "state": "started"}, "module_name": "service"}, "name": "mysql", "state": "started"}
```
##### ACTUAL RESULTS
https://travis-ci.org/juju4/ansible-bro-ids/jobs/150166292
```
TASK [bro-ids : check that mysql is running] ***********************************
task path: /etc/ansible/roles/bro-ids/tasks/main.yml:80
<localhost> ESTABLISH LOCAL CONNECTION FOR USER: root
<localhost> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016 `" && echo ansible-tmp-1470433523.11-95534498356016="` echo $HOME/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016 `" ) && sleep 0'
<localhost> PUT /tmp/tmpMC1xcR TO /root/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016/service
<localhost> EXEC /bin/sh -c 'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /root/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016/service; rm -rf "/root/.ansible/tmp/ansible-tmp-1470433523.11-95534498356016/" > /dev/null 2>&1 && sleep 0'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_args": {"arguments": "", "enabled": null, "name": "mysql", "pattern": null, "runlevel": "default", "sleep": null, "state": "started"}, "module_name": "service"}, "msg": "failure 1 running systemctl show for 'mysql': Failed to connect to bus: No such file or directory\n"}
```
| main | service module ubuntu xenial on travis failure running systemctl show for failed to connect to bus no such file or directory issue type bug report component name service module ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration no extra os environment docker in travis mapping ubuntu xenial in local connection summary when using service task to check one is started task is failed with above message steps to reproduce see only on target xenial working on trusty on a local vagrantfile with xenial as found in github role test vagrant it s working normally so there is something with travis environment expected results test vagrant local execution on xenial task task path home julien documents script homelab roles bro ids tasks main yml establish ssh connection for user vagrant ssh exec ssh c vvv o userknownhostsfile dev null o identitiesonly yes o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile home julien documents script homelab roles bro ids test vagrant vagrant machines broids virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home julien ansible cp ansible ssh h p r bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to home vagrant ansible tmp ansible tmp service ssh exec sftp b c vvv o userknownhostsfile dev null o identitiesonly yes o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile home julien documents script homelab roles bro ids test vagrant vagrant machines broids virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home julien ansible cp ansible ssh h p r establish ssh connection for user vagrant ssh exec ssh c vvv o userknownhostsfile dev null o identitiesonly yes o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile home julien documents script homelab roles bro ids test vagrant vagrant machines broids virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home julien ansible cp ansible ssh h p r tt bin sh c sudo h s n u root bin sh c echo become success drceqxkhrlhnomixtamaaklasdzascol lang fr fr utf lc all fr fr utf lc messages fr fr utf usr bin python home vagrant ansible tmp ansible tmp service rm rf home vagrant ansible tmp ansible tmp dev null sleep ok changed false invocation module args arguments enabled null name mysql pattern null runlevel default sleep null state started module name service name mysql state started actual results task task path etc ansible roles bro ids tasks main yml establish local connection for user root exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp to root ansible tmp ansible tmp service exec bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python root ansible tmp ansible tmp service rm rf root ansible tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module args arguments enabled null name mysql pattern null runlevel default sleep null state started module name service msg failure running systemctl show for mysql failed to connect to bus no such file or directory n | 1 |
26,401 | 7,834,906,642 | IssuesEvent | 2018-06-16 20:10:43 | scalameta/scalameta | https://api.github.com/repos/scalameta/scalameta | closed | Publish sources generated for the protobuf schema | Build | That would help with navigating code that uses `semanticdb3`. Due to some reason, this confuses Intellij. | 1.0 | Publish sources generated for the protobuf schema - That would help with navigating code that uses `semanticdb3`. Due to some reason, this confuses Intellij. | non_main | publish sources generated for the protobuf schema that would help with navigating code that uses due to some reason this confuses intellij | 0 |
17,002 | 5,316,846,249 | IssuesEvent | 2017-02-13 20:58:28 | WayofTime/BloodMagic | https://api.github.com/repos/WayofTime/BloodMagic | closed | Living armor self sacrifice update not working | 1.10 1.11 bug code complete | #### Issue Description:
No matter how many self sacrifices I make living armor is not getting Tough Palms upgrade.
#### Steps to reproduce:
1. Craft Living Armor set (with available upgrade points)
2. Self sacrifice on blood altar
#### Affected Versions
- BloodMagic: 1.9.4-2.1.0-69
- Minecraft: 1.10.2
- Forge: 12.18.3.2221
#### Technical details
ItemSacrificialDagger.onItemRightClick() seems to have a lot of obsolete code (all sacrifical logic is done by onPlayerStopperUsing() method).
In normal flow execution can't get past line 110 so SacrificeKnieUsedEvent is never called. | 1.0 | Living armor self sacrifice update not working - #### Issue Description:
No matter how many self sacrifices I make living armor is not getting Tough Palms upgrade.
#### Steps to reproduce:
1. Craft Living Armor set (with available upgrade points)
2. Self sacrifice on blood altar
#### Affected Versions
- BloodMagic: 1.9.4-2.1.0-69
- Minecraft: 1.10.2
- Forge: 12.18.3.2221
#### Technical details
ItemSacrificialDagger.onItemRightClick() seems to have a lot of obsolete code (all sacrifical logic is done by onPlayerStopperUsing() method).
In normal flow execution can't get past line 110 so SacrificeKnieUsedEvent is never called. | non_main | living armor self sacrifice update not working issue description no matter how many self sacrifices i make living armor is not getting tough palms upgrade steps to reproduce craft living armor set with available upgrade points self sacrifice on blood altar affected versions bloodmagic minecraft forge technical details itemsacrificialdagger onitemrightclick seems to have a lot of obsolete code all sacrifical logic is done by onplayerstopperusing method in normal flow execution can t get past line so sacrificeknieusedevent is never called | 0 |
4,821 | 24,849,173,721 | IssuesEvent | 2022-10-26 18:27:27 | ocsf/ocsf-schema | https://api.github.com/repos/ocsf/ocsf-schema | closed | Determine the reserved attributes | maintainers | ### Discussed in https://github.com/ocsf/ocsf-schema/discussions/84
<div type='discussions-op-text'>
<sup>Originally posted by **rroupski** July 19, 2022</sup>
Attributes that are either generated or derived by the collection, post-collection processing, or storage systems other than the mapping process are designated **Reserved**. The current list of the **reserved** attributes is:
- _log_time (from metadata), rename as `_logged_time`
- _type_uid
- _type_name
- _time
- _observables
- _raw_data
- _unmapped
This discussion is about whether the last 3 attributes should be reserved or not.
- _observables
- _raw_data
- _unmapped
- The `observables` should generated based the input data and the schema. In other words, the `observables` data should not be manually added by the source that generated the event.
- The `raw_data` is attribute that contains the original data as generated by the source. If the event source creates events in the OCSF Schema, then the `raw_data` should not be used.
- The `unmapped` is attribute that contains the attributes, which are not defined by the OCSF Schema. If the event source creates events in the OCSF Schema, then the `unmapped` attribute could be used to add additional attribute, which are not defined by the schema.
</div> | True | Determine the reserved attributes - ### Discussed in https://github.com/ocsf/ocsf-schema/discussions/84
<div type='discussions-op-text'>
<sup>Originally posted by **rroupski** July 19, 2022</sup>
Attributes that are either generated or derived by the collection, post-collection processing, or storage systems other than the mapping process are designated **Reserved**. The current list of the **reserved** attributes is:
- _log_time (from metadata), rename as `_logged_time`
- _type_uid
- _type_name
- _time
- _observables
- _raw_data
- _unmapped
This discussion is about whether the last 3 attributes should be reserved or not.
- _observables
- _raw_data
- _unmapped
- The `observables` should generated based the input data and the schema. In other words, the `observables` data should not be manually added by the source that generated the event.
- The `raw_data` is attribute that contains the original data as generated by the source. If the event source creates events in the OCSF Schema, then the `raw_data` should not be used.
- The `unmapped` is attribute that contains the attributes, which are not defined by the OCSF Schema. If the event source creates events in the OCSF Schema, then the `unmapped` attribute could be used to add additional attribute, which are not defined by the schema.
</div> | main | determine the reserved attributes discussed in originally posted by rroupski july attributes that are either generated or derived by the collection post collection processing or storage systems other than the mapping process are designated reserved the current list of the reserved attributes is log time from metadata rename as logged time type uid type name time observables raw data unmapped this discussion is about whether the last attributes should be reserved or not observables raw data unmapped the observables should generated based the input data and the schema in other words the observables data should not be manually added by the source that generated the event the raw data is attribute that contains the original data as generated by the source if the event source creates events in the ocsf schema then the raw data should not be used the unmapped is attribute that contains the attributes which are not defined by the ocsf schema if the event source creates events in the ocsf schema then the unmapped attribute could be used to add additional attribute which are not defined by the schema | 1 |
955 | 4,699,765,651 | IssuesEvent | 2016-10-12 16:34:41 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | rpm_key is not accepting a list of key-urls | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
rpm_key
##### ANSIBLE VERSION
```
ansible 2.1.2.0
```
##### CONFIGURATION
nothing changed in /etc/ansible.cfg
##### OS / ENVIRONMENT
ubuntu 16.04
##### SUMMARY
Inside a role i got:
```
…
- name: Install gpg keys for HP repo
rpm_key: state=present key={{ item }}
with_items:
- "{{hprepokeys}}"
…
```
and in my vars.yml i got
```
hprepokeys:
- 'http://downloads.linux.hpe.com/SDR/hpPublicKey1024.pub'
- 'http://downloads.linux.hpe.com/SDR/hpPublicKey2048.pub'
- 'http://downloads.linux.hpe.com/SDR/hpPublicKey2048_key1.pub'
- 'http://downloads.linux.hpe.com/SDR/hpePublicKey2048_key1.pub'
```
Running the playbook is telling
```
TASK [common : Install gpg keys for HP repo] ***********************************
fatal: [192.168.0.168]: FAILED! => {"failed": true, "msg": "the field 'args' has an invalid value, which appears to include a variable that is undefined. The error was: 'hprepokeys' is undefined\n\nThe error appears to have been in '/foobar/roles/common/tasks/main.yml': line 39, column 3, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n\n- name: Install gpg keys for HP repo\n ^ here\n"}
```
##### STEPS TO REPRODUCE
Have a look at the SUMMARY
##### EXPECTED RESULTS
rpm_key is handling the list of vars
##### ACTUAL RESULTS
A fatal error, see SUMMARY
| True | rpm_key is not accepting a list of key-urls - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
rpm_key
##### ANSIBLE VERSION
```
ansible 2.1.2.0
```
##### CONFIGURATION
nothing changed in /etc/ansible.cfg
##### OS / ENVIRONMENT
ubuntu 16.04
##### SUMMARY
Inside a role i got:
```
…
- name: Install gpg keys for HP repo
rpm_key: state=present key={{ item }}
with_items:
- "{{hprepokeys}}"
…
```
and in my vars.yml i got
```
hprepokeys:
- 'http://downloads.linux.hpe.com/SDR/hpPublicKey1024.pub'
- 'http://downloads.linux.hpe.com/SDR/hpPublicKey2048.pub'
- 'http://downloads.linux.hpe.com/SDR/hpPublicKey2048_key1.pub'
- 'http://downloads.linux.hpe.com/SDR/hpePublicKey2048_key1.pub'
```
Running the playbook is telling
```
TASK [common : Install gpg keys for HP repo] ***********************************
fatal: [192.168.0.168]: FAILED! => {"failed": true, "msg": "the field 'args' has an invalid value, which appears to include a variable that is undefined. The error was: 'hprepokeys' is undefined\n\nThe error appears to have been in '/foobar/roles/common/tasks/main.yml': line 39, column 3, but may\nbe elsewhere in the file depending on the exact syntax problem.\n\nThe offending line appears to be:\n\n\n- name: Install gpg keys for HP repo\n ^ here\n"}
```
##### STEPS TO REPRODUCE
Have a look at the SUMMARY
##### EXPECTED RESULTS
rpm_key is handling the list of vars
##### ACTUAL RESULTS
A fatal error, see SUMMARY
| main | rpm key is not accepting a list of key urls issue type bug report component name rpm key ansible version ansible configuration nothing changed in etc ansible cfg os environment ubuntu summary inside a role i got … name install gpg keys for hp repo rpm key state present key item with items hprepokeys … and in my vars yml i got hprepokeys running the playbook is telling task fatal failed failed true msg the field args has an invalid value which appears to include a variable that is undefined the error was hprepokeys is undefined n nthe error appears to have been in foobar roles common tasks main yml line column but may nbe elsewhere in the file depending on the exact syntax problem n nthe offending line appears to be n n n name install gpg keys for hp repo n here n steps to reproduce have a look at the summary expected results rpm key is handling the list of vars actual results a fatal error see summary | 1 |
3,448 | 13,212,427,276 | IssuesEvent | 2020-08-16 06:56:55 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | Parameters not saved in Analysis classes | Component-Analysis Difficulty-easy close? deprecation maintainability usability | In all analysis classes the parameters `start`, `stop`, `step` and `verbose` are not saved if they are changed after the object was run once.
After looking into the [code](https://github.com/MDAnalysis/mdanalysis/blob/develop/package/MDAnalysis/analysis/base.py#L137-L139) I found that for the `start`, `stop` and `step` parameters this is an issue due to the fix of the old deprecated syntax. For the `verbose` one just have to save this in the `_verbose` attribute of the object. All these things are easy to fix and I could provide a patch.
**Code to reproduce the behavior**
``` python
from MDAnalysis.analysis.lineardensity import LinearDensity
from MDAnalysis.tests.datafiles import PSF_TRICLINIC, DCD_TRICLINIC
u = mda.Universe(PSF_TRICLINIC, DCD_TRICLINIC)
ldens = LinearDensity(u.atoms, verbose=True)
ldens.run(start=0, stop=10)
print(ldens.start, ldens.stop)
ldens.run(start=0, stop=5, verbose=False)
print(ldens.start, ldens.stop)
```
**Currently version of MDAnalysis**
Python 3.6, mda 0.19.3, MacOS | True | Parameters not saved in Analysis classes - In all analysis classes the parameters `start`, `stop`, `step` and `verbose` are not saved if they are changed after the object was run once.
After looking into the [code](https://github.com/MDAnalysis/mdanalysis/blob/develop/package/MDAnalysis/analysis/base.py#L137-L139) I found that for the `start`, `stop` and `step` parameters this is an issue due to the fix of the old deprecated syntax. For the `verbose` one just have to save this in the `_verbose` attribute of the object. All these things are easy to fix and I could provide a patch.
**Code to reproduce the behavior**
``` python
from MDAnalysis.analysis.lineardensity import LinearDensity
from MDAnalysis.tests.datafiles import PSF_TRICLINIC, DCD_TRICLINIC
u = mda.Universe(PSF_TRICLINIC, DCD_TRICLINIC)
ldens = LinearDensity(u.atoms, verbose=True)
ldens.run(start=0, stop=10)
print(ldens.start, ldens.stop)
ldens.run(start=0, stop=5, verbose=False)
print(ldens.start, ldens.stop)
```
**Currently version of MDAnalysis**
Python 3.6, mda 0.19.3, MacOS | main | parameters not saved in analysis classes in all analysis classes the parameters start stop step and verbose are not saved if they are changed after the object was run once after looking into the i found that for the start stop and step parameters this is an issue due to the fix of the old deprecated syntax for the verbose one just have to save this in the verbose attribute of the object all these things are easy to fix and i could provide a patch code to reproduce the behavior python from mdanalysis analysis lineardensity import lineardensity from mdanalysis tests datafiles import psf triclinic dcd triclinic u mda universe psf triclinic dcd triclinic ldens lineardensity u atoms verbose true ldens run start stop print ldens start ldens stop ldens run start stop verbose false print ldens start ldens stop currently version of mdanalysis python mda macos | 1 |
325,904 | 27,969,237,828 | IssuesEvent | 2023-03-24 23:11:44 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | pkg/sql/opt/exec/execbuilder/tests/local/local_test: TestExecBuild_explain_analyze failed | C-test-failure O-robot branch-release-23.1 | pkg/sql/opt/exec/execbuilder/tests/local/local_test.TestExecBuild_explain_analyze [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9242191?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9242191?buildTab=artifacts#/) on release-23.1 @ [6be5755ca62b8506743b0a0adde19b64135a435f](https://github.com/cockroachdb/cockroach/commits/6be5755ca62b8506743b0a0adde19b64135a435f):
```
=== RUN TestExecBuild_explain_analyze
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/ec6a1341baa02e55a69448c66fcc4415/logTestExecBuild_explain_analyze1346695959
test_log_scope.go:79: use -show-logs to present logs inline
logic.go:2841:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/6275/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/opt/exec/execbuilder/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/opt/exec/execbuilder/testdata/explain_analyze:127: EXPLAIN (VERBOSE) SELECT * FROM system.privileges WHERE path = 'vtable/crdb_internal/tables'
expected:
distribution: local
vectorized: true
·
• scan
columns: (username, path, privileges, grant_options, user_id)
estimated row count: 10 (missing stats)
table: privileges@privileges_path_user_id_key
spans: /"vtable/crdb_internal/tables"-/"vtable/crdb_internal/tables"/PrefixEnd
but found (query options: "") :
distribution: local
vectorized: true
·
• scan
columns: (username, path, privileges, grant_options, user_id)
estimated row count: 1 (100% of the table; stats collected <hidden> ago)
table: privileges@privileges_path_user_id_key
spans: /"vtable/crdb_internal/tables"-/"vtable/crdb_internal/tables"/PrefixEnd
[22:20:46] --- done: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/6275/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/opt/exec/execbuilder/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/opt/exec/execbuilder/testdata/explain_analyze with config local: 8 tests, 1 failures
[22:20:46] --- total progress: 8 statements
--- total: 8 tests, 1 failures
logic.go:4121: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/ec6a1341baa02e55a69448c66fcc4415/logTestExecBuild_explain_analyze1346695959
--- FAIL: TestExecBuild_explain_analyze (3.08s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestExecBuild_explain_analyze.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 1.0 | pkg/sql/opt/exec/execbuilder/tests/local/local_test: TestExecBuild_explain_analyze failed - pkg/sql/opt/exec/execbuilder/tests/local/local_test.TestExecBuild_explain_analyze [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9242191?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Ci_TestsAwsLinuxArm64_UnitTests/9242191?buildTab=artifacts#/) on release-23.1 @ [6be5755ca62b8506743b0a0adde19b64135a435f](https://github.com/cockroachdb/cockroach/commits/6be5755ca62b8506743b0a0adde19b64135a435f):
```
=== RUN TestExecBuild_explain_analyze
test_log_scope.go:161: test logs captured to: /artifacts/tmp/_tmp/ec6a1341baa02e55a69448c66fcc4415/logTestExecBuild_explain_analyze1346695959
test_log_scope.go:79: use -show-logs to present logs inline
logic.go:2841:
/home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/6275/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/opt/exec/execbuilder/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/opt/exec/execbuilder/testdata/explain_analyze:127: EXPLAIN (VERBOSE) SELECT * FROM system.privileges WHERE path = 'vtable/crdb_internal/tables'
expected:
distribution: local
vectorized: true
·
• scan
columns: (username, path, privileges, grant_options, user_id)
estimated row count: 10 (missing stats)
table: privileges@privileges_path_user_id_key
spans: /"vtable/crdb_internal/tables"-/"vtable/crdb_internal/tables"/PrefixEnd
but found (query options: "") :
distribution: local
vectorized: true
·
• scan
columns: (username, path, privileges, grant_options, user_id)
estimated row count: 1 (100% of the table; stats collected <hidden> ago)
table: privileges@privileges_path_user_id_key
spans: /"vtable/crdb_internal/tables"-/"vtable/crdb_internal/tables"/PrefixEnd
[22:20:46] --- done: /home/roach/.cache/bazel/_bazel_roach/c5a4e7d36696d9cd970af2045211a7df/sandbox/processwrapper-sandbox/6275/execroot/com_github_cockroachdb_cockroach/bazel-out/aarch64-fastbuild/bin/pkg/sql/opt/exec/execbuilder/tests/local/local_test_/local_test.runfiles/com_github_cockroachdb_cockroach/pkg/sql/opt/exec/execbuilder/testdata/explain_analyze with config local: 8 tests, 1 failures
[22:20:46] --- total progress: 8 statements
--- total: 8 tests, 1 failures
logic.go:4121: -- test log scope end --
test logs left over in: /artifacts/tmp/_tmp/ec6a1341baa02e55a69448c66fcc4415/logTestExecBuild_explain_analyze1346695959
--- FAIL: TestExecBuild_explain_analyze (3.08s)
```
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestExecBuild_explain_analyze.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_main | pkg sql opt exec execbuilder tests local local test testexecbuild explain analyze failed pkg sql opt exec execbuilder tests local local test testexecbuild explain analyze with on release run testexecbuild explain analyze test log scope go test logs captured to artifacts tmp tmp logtestexecbuild explain test log scope go use show logs to present logs inline logic go home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql opt exec execbuilder tests local local test local test runfiles com github cockroachdb cockroach pkg sql opt exec execbuilder testdata explain analyze explain verbose select from system privileges where path vtable crdb internal tables expected distribution local vectorized true · • scan columns username path privileges grant options user id estimated row count missing stats table privileges privileges path user id key spans vtable crdb internal tables vtable crdb internal tables prefixend but found query options distribution local vectorized true · • scan columns username path privileges grant options user id estimated row count of the table stats collected ago table privileges privileges path user id key spans vtable crdb internal tables vtable crdb internal tables prefixend done home roach cache bazel bazel roach sandbox processwrapper sandbox execroot com github cockroachdb cockroach bazel out fastbuild bin pkg sql opt exec execbuilder tests local local test local test runfiles com github cockroachdb cockroach pkg sql opt exec execbuilder testdata explain analyze with config local tests failures total progress statements total tests failures logic go test log scope end test logs left over in artifacts tmp tmp logtestexecbuild explain fail testexecbuild explain analyze help see also cc cockroachdb sql queries | 0 |
242,996 | 26,277,893,699 | IssuesEvent | 2023-01-07 01:26:05 | vincenzodistasio97/kiptab | https://api.github.com/repos/vincenzodistasio97/kiptab | opened | WS-2017-3738 (Medium) detected in vue-1.0.26.js | security vulnerability | ## WS-2017-3738 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>vue-1.0.26.js</b></p></summary>
<p>Simple, Fast & Composable MVVM for building interactive interfaces</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/vue/1.0.26/vue.js">https://cdnjs.cloudflare.com/ajax/libs/vue/1.0.26/vue.js</a></p>
<p>Path to vulnerable library: /client/scripts/vue.js</p>
<p>
Dependency Hierarchy:
- :x: **vue-1.0.26.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/kiptab/commit/fd7b8d927ccfb80e005e1a4504293529e563bdba">fd7b8d927ccfb80e005e1a4504293529e563bdba</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vue-Project before version 2.3.0-beta.1 has a possible xss vulnerability.
<p>Publish Date: 2017-04-01
<p>URL: <a href=https://github.com/vuejs/vue/commit/1e37633567f5d015db24ae0210b1adb4b1c3d355>WS-2017-3738</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2017-04-01</p>
<p>Fix Resolution: 2.3.0-beta.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2017-3738 (Medium) detected in vue-1.0.26.js - ## WS-2017-3738 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>vue-1.0.26.js</b></p></summary>
<p>Simple, Fast & Composable MVVM for building interactive interfaces</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/vue/1.0.26/vue.js">https://cdnjs.cloudflare.com/ajax/libs/vue/1.0.26/vue.js</a></p>
<p>Path to vulnerable library: /client/scripts/vue.js</p>
<p>
Dependency Hierarchy:
- :x: **vue-1.0.26.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/vincenzodistasio97/kiptab/commit/fd7b8d927ccfb80e005e1a4504293529e563bdba">fd7b8d927ccfb80e005e1a4504293529e563bdba</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Vue-Project before version 2.3.0-beta.1 has a possible xss vulnerability.
<p>Publish Date: 2017-04-01
<p>URL: <a href=https://github.com/vuejs/vue/commit/1e37633567f5d015db24ae0210b1adb4b1c3d355>WS-2017-3738</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2017-04-01</p>
<p>Fix Resolution: 2.3.0-beta.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | ws medium detected in vue js ws medium severity vulnerability vulnerable library vue js simple fast composable mvvm for building interactive interfaces library home page a href path to vulnerable library client scripts vue js dependency hierarchy x vue js vulnerable library found in head commit a href found in base branch master vulnerability details vue project before version beta has a possible xss vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution beta step up your open source security game with mend | 0 |
143,925 | 5,532,867,046 | IssuesEvent | 2017-03-21 11:49:06 | CS2103JAN2017-T09-B4/main | https://api.github.com/repos/CS2103JAN2017-T09-B4/main | closed | Mark a task as completed and filter tasks that are not completed | priority.high status.ongoing type.story | So that I can keep track of tasks that I have to complete | 1.0 | Mark a task as completed and filter tasks that are not completed - So that I can keep track of tasks that I have to complete | non_main | mark a task as completed and filter tasks that are not completed so that i can keep track of tasks that i have to complete | 0 |
127,338 | 18,010,411,284 | IssuesEvent | 2021-09-16 07:56:57 | maddyCode23/linux-4.1.15 | https://api.github.com/repos/maddyCode23/linux-4.1.15 | opened | CVE-2020-25669 (High) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2020-25669 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/keyboard/sunkbd.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/keyboard/sunkbd.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/keyboard/sunkbd.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the Linux Kernel where the function sunkbd_reinit having been scheduled by sunkbd_interrupt before sunkbd being freed. Though the dangling pointer is set to NULL in sunkbd_disconnect, there is still an alias in sunkbd_reinit causing Use After Free.
<p>Publish Date: 2021-05-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25669>CVE-2020-25669</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25669">https://www.linuxkernelcves.com/cves/CVE-2020-25669</a></p>
<p>Release Date: 2020-09-17</p>
<p>Fix Resolution: v4.4.245, v4.9.245, v4.14.208, v4.19.159, v5.4.79, v5.9.10, v5.10-rc5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-25669 (High) detected in linux-stable-rtv4.1.33 - ## CVE-2020-25669 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/maddyCode23/linux-4.1.15/commit/f1f3d2b150be669390b32dfea28e773471bdd6e7">f1f3d2b150be669390b32dfea28e773471bdd6e7</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/keyboard/sunkbd.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/keyboard/sunkbd.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/input/keyboard/sunkbd.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A vulnerability was found in the Linux Kernel where the function sunkbd_reinit having been scheduled by sunkbd_interrupt before sunkbd being freed. Though the dangling pointer is set to NULL in sunkbd_disconnect, there is still an alias in sunkbd_reinit causing Use After Free.
<p>Publish Date: 2021-05-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25669>CVE-2020-25669</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-25669">https://www.linuxkernelcves.com/cves/CVE-2020-25669</a></p>
<p>Release Date: 2020-09-17</p>
<p>Fix Resolution: v4.4.245, v4.9.245, v4.14.208, v4.19.159, v5.4.79, v5.9.10, v5.10-rc5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in linux stable cve high severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href vulnerable source files drivers input keyboard sunkbd c drivers input keyboard sunkbd c drivers input keyboard sunkbd c vulnerability details a vulnerability was found in the linux kernel where the function sunkbd reinit having been scheduled by sunkbd interrupt before sunkbd being freed though the dangling pointer is set to null in sunkbd disconnect there is still an alias in sunkbd reinit causing use after free publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
2,217 | 7,835,039,385 | IssuesEvent | 2018-06-16 22:02:49 | MarcusWolschon/osmeditor4android | https://api.github.com/repos/MarcusWolschon/osmeditor4android | opened | Refactor file handling in PresetItem | Maintainability Usability | Move the fields to own objects and maintain them in a list so that order is preserved. | True | Refactor file handling in PresetItem - Move the fields to own objects and maintain them in a list so that order is preserved. | main | refactor file handling in presetitem move the fields to own objects and maintain them in a list so that order is preserved | 1 |
730 | 4,321,190,024 | IssuesEvent | 2016-07-25 09:13:16 | Particular/NServiceBus.AzureServiceBus | https://api.github.com/repos/Particular/NServiceBus.AzureServiceBus | closed | Can't set custom implementation of IConvertBrokeredMessagesToIncomingMessages and IConvertOutgoingMessagesToBrokeredMessages | Tag: Maintainer Prio Type: Refactoring | It's not possible to provide a custom implementation of `IConvertBrokeredMessagesToIncomingMessages` or `IConvertOutgoingMessagesToBrokeredMessages` through configuration API.
Topologies register default implementations directly into the container. | True | Can't set custom implementation of IConvertBrokeredMessagesToIncomingMessages and IConvertOutgoingMessagesToBrokeredMessages - It's not possible to provide a custom implementation of `IConvertBrokeredMessagesToIncomingMessages` or `IConvertOutgoingMessagesToBrokeredMessages` through configuration API.
Topologies register default implementations directly into the container. | main | can t set custom implementation of iconvertbrokeredmessagestoincomingmessages and iconvertoutgoingmessagestobrokeredmessages it s not possible to provide a custom implementation of iconvertbrokeredmessagestoincomingmessages or iconvertoutgoingmessagestobrokeredmessages through configuration api topologies register default implementations directly into the container | 1 |
487,909 | 14,061,197,161 | IssuesEvent | 2020-11-03 07:41:58 | aau-giraf/wiki | https://api.github.com/repos/aau-giraf/wiki | closed | The textfont size for the app, is described wrongly in the designguide | group 6 point: 2 priority: medium type: refactor | **Is your feature request related to a problem? Please describe.**
The designguide is currently referering to a file with font sizes, but this files does not exist.
**Describe the solution you'd like**
The file should be made with the font sizes for the different screens and widgets in the weekplanner.
**Describe alternatives you've considered**
None at the moment
**Additional context**
This is closely related to aau-giraf/weekplanner#316 and the file made from this issue should be implemented in the weekplanner.
| 1.0 | The textfont size for the app, is described wrongly in the designguide - **Is your feature request related to a problem? Please describe.**
The designguide is currently referering to a file with font sizes, but this files does not exist.
**Describe the solution you'd like**
The file should be made with the font sizes for the different screens and widgets in the weekplanner.
**Describe alternatives you've considered**
None at the moment
**Additional context**
This is closely related to aau-giraf/weekplanner#316 and the file made from this issue should be implemented in the weekplanner.
| non_main | the textfont size for the app is described wrongly in the designguide is your feature request related to a problem please describe the designguide is currently referering to a file with font sizes but this files does not exist describe the solution you d like the file should be made with the font sizes for the different screens and widgets in the weekplanner describe alternatives you ve considered none at the moment additional context this is closely related to aau giraf weekplanner and the file made from this issue should be implemented in the weekplanner | 0 |
124,847 | 12,240,091,220 | IssuesEvent | 2020-05-04 23:14:58 | NathanPB/Mysticis | https://api.github.com/repos/NathanPB/Mysticis | opened | Mysticis Mod Initial Concept | documentation | I'm writing this down right now because I got some inspiration about what to do with this mod. Since I paused the development AGAIN due creativity crisis. I won't let this idea go.
# Inspirations
Mysticis is meant to be inspired in games and Minecraft mods that I like the most and have this idea in mind, the "magic" thing. They are mainly Ars Magica 2 and Thaumcraft 4 (Minecraft mods) and Magicka. The concepts of the mods/games described above are:
- **Thaumcraft 4**
You have 5 types of mana that are called _vis_ ingame, the vis you get comes from mystical nodes that are all around the world and are not supposed to regenerate passively in your items as a main form of acquiring them.
On Thaumcraft you can also build staffs and wands - which the vis is stored in - by combining the rod, the caps and so on. Different materials will give the staff/wand different buffs. Usually its materials rank is pretty linear, and I don't like that.
You can also attach focuses to the wands, and allow it to make magical effects in exchange of the vis that is stored in the wand. The focuses can be upgraded.
Thaumcraft also has a burnout system in long-time terms that your char starts getting "crazy" if it studies too much or use too much magic. Knowledge will drive you crazy if you don't know how to deal with it.
Thaumcraft also has a research system that I don't wanna touch right now.
- **Ars Magica 2**
In Ars Magica you have three trees of different skills you can follow, being them attack, defense and utility. The greater the skills, harder to acquire are the points you need to unlock them.
There are casting skills, effect skills, modifier skills and so on, the meta of the mod is to combine the casts and the effects in order to create a completely custom spell. E.g combine an explosion effect with a projectile cast type, so you can throw projectiles that explodes.
In Ars Magica 2 the mana is stored in the player (not in the items) and passively generates depending of your magic level. There is also a burnout system that prevents you from spamming spells for too long.
- **Magicka**
In the Magicka game you have 8 primary elements (water, heal, shield, frost, lighting, death, rocks and fire). Opposite elements cancel each other (e.g. death cancels heal, fire cancels frost and so on). You can also combine two elements in order to create a compost one (e.g fire + water = steam).
The thing here is that in order to cast the spells you want to combine two elements, not exactly to create a compost element, but to combine its different powers. E.g. cast fire + death, that creates a beam of fire (death casts in beam form, and fire as a firethrower), or lightning + shield that creates a shield around you that protects the player over lightning damage.
In Magicka you can also cast the spells in three ways: Projectile, AoE and Self. They are pretty self-explanatory.
And another functionality is that you can collect different staffs or weapons dropped by bosses or something to make an element in specific more powerful. E.g. staffs dropped by fire bosses will make your fire spells stronger, but will make the water or frost ones weaker.
And the final concept, you can **again** combine various elements in a specific order in order to cast a Magicka, that is basically a special effect. You learn different Magickas as the story goes on.
# Concepts gathered by the Inspirations
- Mysticis will have mainly 5 types of mana, and the opposite types will cancel each other (Thaumcraft, Magicka)
- Mana will be per-player and not per-item (Ars Magica, Magicka)
- Each mana element type will have its own crystal shard, that will be the main resource of the mod (Thaumcraft)
- Crystals spawn in the world and can be found in dungeon (Thaumcraft)
- The wands will be made by combining different parts of different materials. Each material will have its buffs or debuffs (Thaumcraft, a bit of Magicka too)
- Spells are made by combining different elements (Ars Magica, Magicka)
- Spells are focis to be put in the wands/staffs (Thaumcraft)
- The wands will cast spells (Thaumcraft, Magicka)
- Each mana type will represent a force of nature (air, water, fire, nature), named the Primary ones (Thaumcraft)
# Original Concepts
- Mysticis is meant to make the player master a few mana element types, **but not all of them**
- Players will also have affinity with each type of mana element
- If the mana of a player is high and its affinity is low he will receive debuffs
- Higher affinities will give the player properly buffs and debuffs
- Two of the mana element types will be end-game, named the Primordial ones (magic and dark)
| 1.0 | Mysticis Mod Initial Concept - I'm writing this down right now because I got some inspiration about what to do with this mod. Since I paused the development AGAIN due creativity crisis. I won't let this idea go.
# Inspirations
Mysticis is meant to be inspired in games and Minecraft mods that I like the most and have this idea in mind, the "magic" thing. They are mainly Ars Magica 2 and Thaumcraft 4 (Minecraft mods) and Magicka. The concepts of the mods/games described above are:
- **Thaumcraft 4**
You have 5 types of mana that are called _vis_ ingame, the vis you get comes from mystical nodes that are all around the world and are not supposed to regenerate passively in your items as a main form of acquiring them.
On Thaumcraft you can also build staffs and wands - which the vis is stored in - by combining the rod, the caps and so on. Different materials will give the staff/wand different buffs. Usually its materials rank is pretty linear, and I don't like that.
You can also attach focuses to the wands, and allow it to make magical effects in exchange of the vis that is stored in the wand. The focuses can be upgraded.
Thaumcraft also has a burnout system in long-time terms that your char starts getting "crazy" if it studies too much or use too much magic. Knowledge will drive you crazy if you don't know how to deal with it.
Thaumcraft also has a research system that I don't wanna touch right now.
- **Ars Magica 2**
In Ars Magica you have three trees of different skills you can follow, being them attack, defense and utility. The greater the skills, harder to acquire are the points you need to unlock them.
There are casting skills, effect skills, modifier skills and so on, the meta of the mod is to combine the casts and the effects in order to create a completely custom spell. E.g combine an explosion effect with a projectile cast type, so you can throw projectiles that explodes.
In Ars Magica 2 the mana is stored in the player (not in the items) and passively generates depending of your magic level. There is also a burnout system that prevents you from spamming spells for too long.
- **Magicka**
In the Magicka game you have 8 primary elements (water, heal, shield, frost, lighting, death, rocks and fire). Opposite elements cancel each other (e.g. death cancels heal, fire cancels frost and so on). You can also combine two elements in order to create a compost one (e.g fire + water = steam).
The thing here is that in order to cast the spells you want to combine two elements, not exactly to create a compost element, but to combine its different powers. E.g. cast fire + death, that creates a beam of fire (death casts in beam form, and fire as a firethrower), or lightning + shield that creates a shield around you that protects the player over lightning damage.
In Magicka you can also cast the spells in three ways: Projectile, AoE and Self. They are pretty self-explanatory.
And another functionality is that you can collect different staffs or weapons dropped by bosses or something to make an element in specific more powerful. E.g. staffs dropped by fire bosses will make your fire spells stronger, but will make the water or frost ones weaker.
And the final concept, you can **again** combine various elements in a specific order in order to cast a Magicka, that is basically a special effect. You learn different Magickas as the story goes on.
# Concepts gathered by the Inspirations
- Mysticis will have mainly 5 types of mana, and the opposite types will cancel each other (Thaumcraft, Magicka)
- Mana will be per-player and not per-item (Ars Magica, Magicka)
- Each mana element type will have its own crystal shard, that will be the main resource of the mod (Thaumcraft)
- Crystals spawn in the world and can be found in dungeon (Thaumcraft)
- The wands will be made by combining different parts of different materials. Each material will have its buffs or debuffs (Thaumcraft, a bit of Magicka too)
- Spells are made by combining different elements (Ars Magica, Magicka)
- Spells are focis to be put in the wands/staffs (Thaumcraft)
- The wands will cast spells (Thaumcraft, Magicka)
- Each mana type will represent a force of nature (air, water, fire, nature), named the Primary ones (Thaumcraft)
# Original Concepts
- Mysticis is meant to make the player master a few mana element types, **but not all of them**
- Players will also have affinity with each type of mana element
- If the mana of a player is high and its affinity is low he will receive debuffs
- Higher affinities will give the player properly buffs and debuffs
- Two of the mana element types will be end-game, named the Primordial ones (magic and dark)
| non_main | mysticis mod initial concept i m writing this down right now because i got some inspiration about what to do with this mod since i paused the development again due creativity crisis i won t let this idea go inspirations mysticis is meant to be inspired in games and minecraft mods that i like the most and have this idea in mind the magic thing they are mainly ars magica and thaumcraft minecraft mods and magicka the concepts of the mods games described above are thaumcraft you have types of mana that are called vis ingame the vis you get comes from mystical nodes that are all around the world and are not supposed to regenerate passively in your items as a main form of acquiring them on thaumcraft you can also build staffs and wands which the vis is stored in by combining the rod the caps and so on different materials will give the staff wand different buffs usually its materials rank is pretty linear and i don t like that you can also attach focuses to the wands and allow it to make magical effects in exchange of the vis that is stored in the wand the focuses can be upgraded thaumcraft also has a burnout system in long time terms that your char starts getting crazy if it studies too much or use too much magic knowledge will drive you crazy if you don t know how to deal with it thaumcraft also has a research system that i don t wanna touch right now ars magica in ars magica you have three trees of different skills you can follow being them attack defense and utility the greater the skills harder to acquire are the points you need to unlock them there are casting skills effect skills modifier skills and so on the meta of the mod is to combine the casts and the effects in order to create a completely custom spell e g combine an explosion effect with a projectile cast type so you can throw projectiles that explodes in ars magica the mana is stored in the player not in the items and passively generates depending of your magic level there is also a burnout system that prevents you from spamming spells for too long magicka in the magicka game you have primary elements water heal shield frost lighting death rocks and fire opposite elements cancel each other e g death cancels heal fire cancels frost and so on you can also combine two elements in order to create a compost one e g fire water steam the thing here is that in order to cast the spells you want to combine two elements not exactly to create a compost element but to combine its different powers e g cast fire death that creates a beam of fire death casts in beam form and fire as a firethrower or lightning shield that creates a shield around you that protects the player over lightning damage in magicka you can also cast the spells in three ways projectile aoe and self they are pretty self explanatory and another functionality is that you can collect different staffs or weapons dropped by bosses or something to make an element in specific more powerful e g staffs dropped by fire bosses will make your fire spells stronger but will make the water or frost ones weaker and the final concept you can again combine various elements in a specific order in order to cast a magicka that is basically a special effect you learn different magickas as the story goes on concepts gathered by the inspirations mysticis will have mainly types of mana and the opposite types will cancel each other thaumcraft magicka mana will be per player and not per item ars magica magicka each mana element type will have its own crystal shard that will be the main resource of the mod thaumcraft crystals spawn in the world and can be found in dungeon thaumcraft the wands will be made by combining different parts of different materials each material will have its buffs or debuffs thaumcraft a bit of magicka too spells are made by combining different elements ars magica magicka spells are focis to be put in the wands staffs thaumcraft the wands will cast spells thaumcraft magicka each mana type will represent a force of nature air water fire nature named the primary ones thaumcraft original concepts mysticis is meant to make the player master a few mana element types but not all of them players will also have affinity with each type of mana element if the mana of a player is high and its affinity is low he will receive debuffs higher affinities will give the player properly buffs and debuffs two of the mana element types will be end game named the primordial ones magic and dark | 0 |
5,788 | 30,657,878,380 | IssuesEvent | 2023-07-25 13:18:31 | toolbx-images/images | https://api.github.com/repos/toolbx-images/images | closed | Add distribution: Debian 12 (bookworm) | new-image-request maintainers-wanted | ### Distribution name and versions requested
Debian 12 (Bookworm)
### Where are the official container images from the distribution published?
https://hub.docker.com/_/debian
### Will you be interested in maintaining this image?
Yes, but I'm using distrobox | True | Add distribution: Debian 12 (bookworm) - ### Distribution name and versions requested
Debian 12 (Bookworm)
### Where are the official container images from the distribution published?
https://hub.docker.com/_/debian
### Will you be interested in maintaining this image?
Yes, but I'm using distrobox | main | add distribution debian bookworm distribution name and versions requested debian bookworm where are the official container images from the distribution published will you be interested in maintaining this image yes but i m using distrobox | 1 |
401 | 3,455,708,038 | IssuesEvent | 2015-12-17 21:22:05 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | opened | Streamlining Updates | awaiting maintainer feedback discussion enhancement | Some thoughts and a tentative roadmap towards minimizing both the time and effort to keep Casks up to date in `homebrew-cask`
Currently: Cask updates are submitted as PRs by (our awesome) contributors. A new PR must be made for each version, and then merged into `caskroom/homebrew-cask`
We have some pieces (ie. `appcast` URLs and [`cask-repair`](https://github.com/vitorgalvao/tiny-scripts/blob/master/cask-repair) to push towards a better process, but there's nothing streamlined yet.
Proposal:
- [ ] Update [`cask-repair`](https://github.com/vitorgalvao/tiny-scripts/blob/master/cask-repair) to also update the `appcast` stanza.
- [ ] Instead of trying to be "smart" and parse different types of `appcast`s, go back to having just the stanza, with a required `shasum`. The type could be left in for future work, but it's not currently being used anywhere.
- [ ] Write a script to iterate through the Casks with an `appcast` - updates would clearly show up as having a changed `shasum`.
- [ ] These Casks could then be fed into `cask-repair` (or similar) to make the appropriate PR.
On the user side (keeping Casks up to date), I've had great success with the [`brew-cask-upgrade`](https://github.com/buo/brew-cask-upgrade) gem. Would integrating something like this as the actual `brew cask upgrade` command make sense? (as there's clearly a demand for it). | True | Streamlining Updates - Some thoughts and a tentative roadmap towards minimizing both the time and effort to keep Casks up to date in `homebrew-cask`
Currently: Cask updates are submitted as PRs by (our awesome) contributors. A new PR must be made for each version, and then merged into `caskroom/homebrew-cask`
We have some pieces (ie. `appcast` URLs and [`cask-repair`](https://github.com/vitorgalvao/tiny-scripts/blob/master/cask-repair) to push towards a better process, but there's nothing streamlined yet.
Proposal:
- [ ] Update [`cask-repair`](https://github.com/vitorgalvao/tiny-scripts/blob/master/cask-repair) to also update the `appcast` stanza.
- [ ] Instead of trying to be "smart" and parse different types of `appcast`s, go back to having just the stanza, with a required `shasum`. The type could be left in for future work, but it's not currently being used anywhere.
- [ ] Write a script to iterate through the Casks with an `appcast` - updates would clearly show up as having a changed `shasum`.
- [ ] These Casks could then be fed into `cask-repair` (or similar) to make the appropriate PR.
On the user side (keeping Casks up to date), I've had great success with the [`brew-cask-upgrade`](https://github.com/buo/brew-cask-upgrade) gem. Would integrating something like this as the actual `brew cask upgrade` command make sense? (as there's clearly a demand for it). | main | streamlining updates some thoughts and a tentative roadmap towards minimizing both the time and effort to keep casks up to date in homebrew cask currently cask updates are submitted as prs by our awesome contributors a new pr must be made for each version and then merged into caskroom homebrew cask we have some pieces ie appcast urls and to push towards a better process but there s nothing streamlined yet proposal update to also update the appcast stanza instead of trying to be smart and parse different types of appcast s go back to having just the stanza with a required shasum the type could be left in for future work but it s not currently being used anywhere write a script to iterate through the casks with an appcast updates would clearly show up as having a changed shasum these casks could then be fed into cask repair or similar to make the appropriate pr on the user side keeping casks up to date i ve had great success with the gem would integrating something like this as the actual brew cask upgrade command make sense as there s clearly a demand for it | 1 |
2,206 | 7,788,438,304 | IssuesEvent | 2018-06-07 04:41:46 | chef/chef | https://api.github.com/repos/chef/chef | closed | windows_service resource's delayed_start property doesn't configure anything | Area: Windows Status: Pending Maintainer Review | <!---
!!!!!! NOTE: CHEF CLIENT BUGS ONLY !!!!!!
This issue tracker is for the code contained within this repo -- `chef-client`, base `knife` functionality (not
plugins), `chef-apply`, `chef-solo`, `chef-client -z`, etc.
* Requests for new or alternative functionality should be made to [feedback.chef.io](https://feedback.chef.io/forums/301644-chef-product-feedback/category/110832-chef-client)
* [Chef Server issues](https://github.com/chef/chef-server/issues/new)
* [ChefDK issues](https://github.com/chef/chef-dk/issues/new)
* Cookbook Issues (see the https://github.com/chef-cookbooks repos or search [Supermarket](https://supermarket.chef.io) or GitHub/Google)
-->
## Description
When attempting to use the `delayed_start` property that was added for `windows_service` in Chef Client 14 it doesn't seem to do anything. The `windows_service` resource will neither enable or disable the flag for delayed start.
## Chef Version
14.1.1
## Platform Version
Windows Server 2012 R2
## Replication Case
- Pick a service that is either `Disabled` or `Automatic` start up type and set its startup type to `Disabled`.
- Attempt to set the service as `Automatic (Delayed)` startup type in Chef, like this example:
```
windows_service 'SQL Server Service' do
service_name 'MSSQLSERVER'
action :configure_startup
startup_type :automatic
delayed_start true
end
```
- Notice it will flag the service to start automatically, but not delay the start.
- Alternatively, enable the delay manually. Notice the resource also won't turn off the delayed flag if you do:
```
windows_service 'SQL Server Service' do
service_name 'MSSQLSERVER'
action :configure_startup
startup_type :automatic
delayed_start false
end
```
## Client Output
## Stacktrace
| True | windows_service resource's delayed_start property doesn't configure anything - <!---
!!!!!! NOTE: CHEF CLIENT BUGS ONLY !!!!!!
This issue tracker is for the code contained within this repo -- `chef-client`, base `knife` functionality (not
plugins), `chef-apply`, `chef-solo`, `chef-client -z`, etc.
* Requests for new or alternative functionality should be made to [feedback.chef.io](https://feedback.chef.io/forums/301644-chef-product-feedback/category/110832-chef-client)
* [Chef Server issues](https://github.com/chef/chef-server/issues/new)
* [ChefDK issues](https://github.com/chef/chef-dk/issues/new)
* Cookbook Issues (see the https://github.com/chef-cookbooks repos or search [Supermarket](https://supermarket.chef.io) or GitHub/Google)
-->
## Description
When attempting to use the `delayed_start` property that was added for `windows_service` in Chef Client 14 it doesn't seem to do anything. The `windows_service` resource will neither enable or disable the flag for delayed start.
## Chef Version
14.1.1
## Platform Version
Windows Server 2012 R2
## Replication Case
- Pick a service that is either `Disabled` or `Automatic` start up type and set its startup type to `Disabled`.
- Attempt to set the service as `Automatic (Delayed)` startup type in Chef, like this example:
```
windows_service 'SQL Server Service' do
service_name 'MSSQLSERVER'
action :configure_startup
startup_type :automatic
delayed_start true
end
```
- Notice it will flag the service to start automatically, but not delay the start.
- Alternatively, enable the delay manually. Notice the resource also won't turn off the delayed flag if you do:
```
windows_service 'SQL Server Service' do
service_name 'MSSQLSERVER'
action :configure_startup
startup_type :automatic
delayed_start false
end
```
## Client Output
## Stacktrace
| main | windows service resource s delayed start property doesn t configure anything note chef client bugs only this issue tracker is for the code contained within this repo chef client base knife functionality not plugins chef apply chef solo chef client z etc requests for new or alternative functionality should be made to cookbook issues see the repos or search or github google description when attempting to use the delayed start property that was added for windows service in chef client it doesn t seem to do anything the windows service resource will neither enable or disable the flag for delayed start chef version platform version windows server replication case pick a service that is either disabled or automatic start up type and set its startup type to disabled attempt to set the service as automatic delayed startup type in chef like this example windows service sql server service do service name mssqlserver action configure startup startup type automatic delayed start true end notice it will flag the service to start automatically but not delay the start alternatively enable the delay manually notice the resource also won t turn off the delayed flag if you do windows service sql server service do service name mssqlserver action configure startup startup type automatic delayed start false end client output stacktrace | 1 |
4,214 | 20,825,590,665 | IssuesEvent | 2022-03-18 20:24:51 | HPCL/code-analysis | https://api.github.com/repos/HPCL/code-analysis | closed | CWE-1079 Parent Class without Virtual Destructor Method | IN-PROGRESS CLAIMED ISO/IEC 5055:2021 ParentClass WEAKNESS CATEGORY: MAINTAINABILITY | **Usage Name**
Parent class missing virtual destructor
**Reference**
[https://cwe.mitre.org/data/definitions/1079](https://cwe.mitre.org/data/definitions/1079)
**Roles**
- the *ParentClass*
**Detection Patterns**
- 8.2.36 ASCQM Implement Virtual Destructor for Parent Classes | True | CWE-1079 Parent Class without Virtual Destructor Method - **Usage Name**
Parent class missing virtual destructor
**Reference**
[https://cwe.mitre.org/data/definitions/1079](https://cwe.mitre.org/data/definitions/1079)
**Roles**
- the *ParentClass*
**Detection Patterns**
- 8.2.36 ASCQM Implement Virtual Destructor for Parent Classes | main | cwe parent class without virtual destructor method usage name parent class missing virtual destructor reference roles the parentclass detection patterns ascqm implement virtual destructor for parent classes | 1 |
86,712 | 8,048,204,028 | IssuesEvent | 2018-08-01 05:31:04 | Microsoft/ptvsd | https://api.github.com/repos/Microsoft/ptvsd | opened | Create a whole suite of tests for enable_attach without using wait_for_attach | priority:P1 test issue | Currently almost all system tests use `wait_for_attach` due to a bug in `enable_attach` the simulated the same behavior of `wait_for_attach`. | 1.0 | Create a whole suite of tests for enable_attach without using wait_for_attach - Currently almost all system tests use `wait_for_attach` due to a bug in `enable_attach` the simulated the same behavior of `wait_for_attach`. | non_main | create a whole suite of tests for enable attach without using wait for attach currently almost all system tests use wait for attach due to a bug in enable attach the simulated the same behavior of wait for attach | 0 |
780 | 4,386,285,886 | IssuesEvent | 2016-08-08 12:17:32 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | win_feature source parameter escapes incorrectly | bug_report waiting_on_maintainer windows | ##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_feature
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
-->
##### OS / ENVIRONMENT
Ubuntu Xenial 64
##### SUMMARY
<!--- Explain the problem briefly -->
When using an UNC-Path for the Source-Parameter of the win_feature module escaping doesn't work correctly. Any other combination of escape characters doesn't work.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
---
# tasks file for vx_dotnet35
- name: Ensure .NET 3.5 feature is present
win_feature: name="NET-Framework-Features" state=present restart=true source='\\defrasadm01\windvd\sources\sxs'
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
Installation of Feature
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
fatal: [172.29.84.91]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "win_feature"}, "msg": "Failed to find source path \\\\defrasadm01\\windvd\\sources\\sxs"}
```
| True | win_feature source parameter escapes incorrectly - ##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_feature
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
-->
##### OS / ENVIRONMENT
Ubuntu Xenial 64
##### SUMMARY
<!--- Explain the problem briefly -->
When using an UNC-Path for the Source-Parameter of the win_feature module escaping doesn't work correctly. Any other combination of escape characters doesn't work.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
---
# tasks file for vx_dotnet35
- name: Ensure .NET 3.5 feature is present
win_feature: name="NET-Framework-Features" state=present restart=true source='\\defrasadm01\windvd\sources\sxs'
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
Installation of Feature
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
fatal: [172.29.84.91]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "win_feature"}, "msg": "Failed to find source path \\\\defrasadm01\\windvd\\sources\\sxs"}
```
| main | win feature source parameter escapes incorrectly issue type bug report component name win feature ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration os environment ubuntu xenial summary when using an unc path for the source parameter of the win feature module escaping doesn t work correctly any other combination of escape characters doesn t work steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used tasks file for vx name ensure net feature is present win feature name net framework features state present restart true source windvd sources sxs expected results installation of feature actual results fatal failed changed false failed true invocation module name win feature msg failed to find source path windvd sources sxs | 1 |
272,088 | 29,794,980,192 | IssuesEvent | 2023-06-16 01:01:37 | billmcchesney1/flowgate | https://api.github.com/repos/billmcchesney1/flowgate | closed | CVE-2018-5968 (High) detected in jackson-databind-2.8.8.jar - autoclosed | Mend: dependency security vulnerability | ## CVE-2018-5968 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /operation-expert/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-hateoas-1.4.7.RELEASE.jar (Root Library)
- spring-boot-starter-web-1.4.7.RELEASE.jar
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flowgate/commit/dd01a1d4381c7a3b94ba25748c015a094c33088e">dd01a1d4381c7a3b94ba25748c015a094c33088e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind through 2.8.11 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 and CVE-2017-17485 deserialization flaws. This is exploitable via two different gadgets that bypass a blacklist.
<p>Publish Date: 2018-01-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-5968>CVE-2018-5968</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968</a></p>
<p>Release Date: 2018-01-22</p>
<p>Fix Resolution: 2.8.11.1, 2.9.4</p>
</p>
</details>
<p></p>
| True | CVE-2018-5968 (High) detected in jackson-databind-2.8.8.jar - autoclosed - ## CVE-2018-5968 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.8.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /operation-expert/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.8/jackson-databind-2.8.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-hateoas-1.4.7.RELEASE.jar (Root Library)
- spring-boot-starter-web-1.4.7.RELEASE.jar
- :x: **jackson-databind-2.8.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/flowgate/commit/dd01a1d4381c7a3b94ba25748c015a094c33088e">dd01a1d4381c7a3b94ba25748c015a094c33088e</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind through 2.8.11 and 2.9.x through 2.9.3 allows unauthenticated remote code execution because of an incomplete fix for the CVE-2017-7525 and CVE-2017-17485 deserialization flaws. This is exploitable via two different gadgets that bypass a blacklist.
<p>Publish Date: 2018-01-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-5968>CVE-2018-5968</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968">http://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-5968</a></p>
<p>Release Date: 2018-01-22</p>
<p>Fix Resolution: 2.8.11.1, 2.9.4</p>
</p>
</details>
<p></p>
| non_main | cve high detected in jackson databind jar autoclosed cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file operation expert pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter hateoas release jar root library spring boot starter web release jar x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind through and x through allows unauthenticated remote code execution because of an incomplete fix for the cve and cve deserialization flaws this is exploitable via two different gadgets that bypass a blacklist publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
4,689 | 24,207,845,239 | IssuesEvent | 2022-09-25 13:41:21 | beekama/NutritionApp | https://api.github.com/repos/beekama/NutritionApp | closed | Commented out Code | maintainability | There is considerable commented out code of which I'm not sure if it's obsolete or still relevant to development. Most notably:
WeightTracking: L98, L124
FoodOverviewAdapter: L237, L252, L260 | True | Commented out Code - There is considerable commented out code of which I'm not sure if it's obsolete or still relevant to development. Most notably:
WeightTracking: L98, L124
FoodOverviewAdapter: L237, L252, L260 | main | commented out code there is considerable commented out code of which i m not sure if it s obsolete or still relevant to development most notably weighttracking foodoverviewadapter | 1 |
5,605 | 28,056,628,652 | IssuesEvent | 2023-03-29 09:47:54 | conbench/conbench | https://api.github.com/repos/conbench/conbench | closed | tests: cover /benchmarks/<benchmark-result-id> | maintainability | The testsuite does not cover fetching the `/benchmarks/<benchmark-result-id>` UI route; we only cover this in ci/screenshots. Noted that before, forgot to create an issue. This is important test coverage that should be part of the regular test suite. | True | tests: cover /benchmarks/<benchmark-result-id> - The testsuite does not cover fetching the `/benchmarks/<benchmark-result-id>` UI route; we only cover this in ci/screenshots. Noted that before, forgot to create an issue. This is important test coverage that should be part of the regular test suite. | main | tests cover benchmarks the testsuite does not cover fetching the benchmarks ui route we only cover this in ci screenshots noted that before forgot to create an issue this is important test coverage that should be part of the regular test suite | 1 |
4,837 | 24,937,856,901 | IssuesEvent | 2022-10-31 16:26:48 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | Tabs component scroll button on mobile scrolling very little | role: ux 🍿 type: bug 🐛 component: tabs proposal: needs more research 🕵️♀️ status: needs triage 🕵️♀️ status: waiting for maintainer response 💬 | <!-- Feel free to remove sections that aren't relevant.
## Title line template: [Title]: Brief description
-->
## What package(s) are you using?
<!--
Add an x in one of the options below, for example:
- [x] package name
-->
- [ ] `carbon-components`
- [x] `carbon-components-react`
## Detailed description
> Describe in detail the issue you're having.
When on mobile, the scroll buttons only scroll by about 15 pixels when clicked on. Holding does not work on mobile as it brings up the item menu.
> Is this issue related to a specific component?
Yes, Tabs
> What did you expect to happen? What happened instead? What would you like to
> see changed?
Matching scroll distance/behaviour on mobile/desktop.
> What browser are you working in?
any mobile, or alternatively, chrome responsive mode
> What version of the Carbon Design System are you using?
Latest
> What offering/product do you work on? Any pressing ship or release dates we
> should be aware of?
IBM Cloud docs
## Steps to reproduce the issue
1. Go to https://react.carbondesignsystem.com/?path=/story/components-tabs--default
2. In desktop chrome, toggle on `device toolbar`
3. From the drop down, select Iphone X
4. click on the scroll buttons
- Screenshots or code

| True | Tabs component scroll button on mobile scrolling very little - <!-- Feel free to remove sections that aren't relevant.
## Title line template: [Title]: Brief description
-->
## What package(s) are you using?
<!--
Add an x in one of the options below, for example:
- [x] package name
-->
- [ ] `carbon-components`
- [x] `carbon-components-react`
## Detailed description
> Describe in detail the issue you're having.
When on mobile, the scroll buttons only scroll by about 15 pixels when clicked on. Holding does not work on mobile as it brings up the item menu.
> Is this issue related to a specific component?
Yes, Tabs
> What did you expect to happen? What happened instead? What would you like to
> see changed?
Matching scroll distance/behaviour on mobile/desktop.
> What browser are you working in?
any mobile, or alternatively, chrome responsive mode
> What version of the Carbon Design System are you using?
Latest
> What offering/product do you work on? Any pressing ship or release dates we
> should be aware of?
IBM Cloud docs
## Steps to reproduce the issue
1. Go to https://react.carbondesignsystem.com/?path=/story/components-tabs--default
2. In desktop chrome, toggle on `device toolbar`
3. From the drop down, select Iphone X
4. click on the scroll buttons
- Screenshots or code

| main | tabs component scroll button on mobile scrolling very little feel free to remove sections that aren t relevant title line template brief description what package s are you using add an x in one of the options below for example package name carbon components carbon components react detailed description describe in detail the issue you re having when on mobile the scroll buttons only scroll by about pixels when clicked on holding does not work on mobile as it brings up the item menu is this issue related to a specific component yes tabs what did you expect to happen what happened instead what would you like to see changed matching scroll distance behaviour on mobile desktop what browser are you working in any mobile or alternatively chrome responsive mode what version of the carbon design system are you using latest what offering product do you work on any pressing ship or release dates we should be aware of ibm cloud docs steps to reproduce the issue go to in desktop chrome toggle on device toolbar from the drop down select iphone x click on the scroll buttons screenshots or code | 1 |
4,465 | 23,268,566,939 | IssuesEvent | 2022-08-04 20:02:20 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Refactor storage of icon definitions | type: enhancement work: frontend status: ready restricted: new maintainers | ## Current code
- In `mathesar_ui/src/sections/table-view/actions-pane/ActionsPane.svelte` we have:
```svelte
<MenuItem on:click={handleDeleteTable} icon={{ data: faTrashAlt }}>
Delete
</MenuItem>
```
Here, `{ data: faTrashAlt }` is an object of type `IconProps`. Search the codebase for the literal string `{ data: fa` and you'll see more objects like this.
- And we have some plain `<Icon />` component instances like this code in `mathesar_ui/src/component-library/help/Help.svelte`:
```svelte
<Icon data={faQuestionCircle} />
```
Search the codebase for for the regex `<Icon[^/]*?data` and you'll see more like this.
## Desired code
- I'd like to move all that icon data to a central files.
- We'd have `mathesar_ui/src/component-library/common/icons.ts` for all icons used by the component library (any files within `mathesar_ui/src/component-library`).
- And we have `mathesar_ui/src/icons.ts` for all the icons used by Mathesar (any files within `mathesar_ui/src/` and _not_ in the component library).
- That `ActionsPane.svelte` example above would become:
```svelte
<MenuItem on:click={handleDeleteTable} icon={iconDelete}>
Delete
</MenuItem>
```
and `mathesar_ui/src/icons.ts` would include:
```ts
export const iconDelete = { data: faTrashAlt };
```
- That `Help.svelte` example above would become:
```svelte
<Icon {...iconHelp} />
```
and `mathesar_ui/src/component-library/common/icons.ts` would include:
```ts
export const iconHelp = { data: faQuestionCircle };
```
- The FontAwesome icons describe the _appearance_ of the icon. In contrast, our `icons.ts` files would describe the _action_ or _concept_ within the context of either the component library or Mathesar. These files would provide a centralized place for us to keep track of the mapping between app concepts and icons.
## Rationale
- It help us keep our icons consistent. (E.g. helping to answer the question "Should I use `faTrash` or `faTrashAlt`"?)
- It makes it easier to pick out icons for concepts we've already defined. (E.g. Helping to speed up the process of figuring out that we use `faThList` for the concept of "grouping".)
- It allows us to add aliases when we'd like to designate the same icon for different actions.
- It helps us move closer supporting custom icons eventually.
@seancolsen said:
> @pavish This is a proposal for a very small refactor that I'd like to run by you. If this sounds good, you can mark the ticket as ready and label it as a good first issue.
Update from @pavish: The proposal sounds good and the issue is ready for implementation. | True | Refactor storage of icon definitions - ## Current code
- In `mathesar_ui/src/sections/table-view/actions-pane/ActionsPane.svelte` we have:
```svelte
<MenuItem on:click={handleDeleteTable} icon={{ data: faTrashAlt }}>
Delete
</MenuItem>
```
Here, `{ data: faTrashAlt }` is an object of type `IconProps`. Search the codebase for the literal string `{ data: fa` and you'll see more objects like this.
- And we have some plain `<Icon />` component instances like this code in `mathesar_ui/src/component-library/help/Help.svelte`:
```svelte
<Icon data={faQuestionCircle} />
```
Search the codebase for for the regex `<Icon[^/]*?data` and you'll see more like this.
## Desired code
- I'd like to move all that icon data to a central files.
- We'd have `mathesar_ui/src/component-library/common/icons.ts` for all icons used by the component library (any files within `mathesar_ui/src/component-library`).
- And we have `mathesar_ui/src/icons.ts` for all the icons used by Mathesar (any files within `mathesar_ui/src/` and _not_ in the component library).
- That `ActionsPane.svelte` example above would become:
```svelte
<MenuItem on:click={handleDeleteTable} icon={iconDelete}>
Delete
</MenuItem>
```
and `mathesar_ui/src/icons.ts` would include:
```ts
export const iconDelete = { data: faTrashAlt };
```
- That `Help.svelte` example above would become:
```svelte
<Icon {...iconHelp} />
```
and `mathesar_ui/src/component-library/common/icons.ts` would include:
```ts
export const iconHelp = { data: faQuestionCircle };
```
- The FontAwesome icons describe the _appearance_ of the icon. In contrast, our `icons.ts` files would describe the _action_ or _concept_ within the context of either the component library or Mathesar. These files would provide a centralized place for us to keep track of the mapping between app concepts and icons.
## Rationale
- It help us keep our icons consistent. (E.g. helping to answer the question "Should I use `faTrash` or `faTrashAlt`"?)
- It makes it easier to pick out icons for concepts we've already defined. (E.g. Helping to speed up the process of figuring out that we use `faThList` for the concept of "grouping".)
- It allows us to add aliases when we'd like to designate the same icon for different actions.
- It helps us move closer supporting custom icons eventually.
@seancolsen said:
> @pavish This is a proposal for a very small refactor that I'd like to run by you. If this sounds good, you can mark the ticket as ready and label it as a good first issue.
Update from @pavish: The proposal sounds good and the issue is ready for implementation. | main | refactor storage of icon definitions current code in mathesar ui src sections table view actions pane actionspane svelte we have svelte delete here data fatrashalt is an object of type iconprops search the codebase for the literal string data fa and you ll see more objects like this and we have some plain component instances like this code in mathesar ui src component library help help svelte svelte search the codebase for for the regex icon data and you ll see more like this desired code i d like to move all that icon data to a central files we d have mathesar ui src component library common icons ts for all icons used by the component library any files within mathesar ui src component library and we have mathesar ui src icons ts for all the icons used by mathesar any files within mathesar ui src and not in the component library that actionspane svelte example above would become svelte delete and mathesar ui src icons ts would include ts export const icondelete data fatrashalt that help svelte example above would become svelte and mathesar ui src component library common icons ts would include ts export const iconhelp data faquestioncircle the fontawesome icons describe the appearance of the icon in contrast our icons ts files would describe the action or concept within the context of either the component library or mathesar these files would provide a centralized place for us to keep track of the mapping between app concepts and icons rationale it help us keep our icons consistent e g helping to answer the question should i use fatrash or fatrashalt it makes it easier to pick out icons for concepts we ve already defined e g helping to speed up the process of figuring out that we use fathlist for the concept of grouping it allows us to add aliases when we d like to designate the same icon for different actions it helps us move closer supporting custom icons eventually seancolsen said pavish this is a proposal for a very small refactor that i d like to run by you if this sounds good you can mark the ticket as ready and label it as a good first issue update from pavish the proposal sounds good and the issue is ready for implementation | 1 |
21,449 | 3,711,441,020 | IssuesEvent | 2016-03-02 10:20:59 | owncloud/core | https://api.github.com/repos/owncloud/core | closed | Sharing dialog difference file / folder | design enhancement feature:sharing | Referencing to the design forward share dialog #3224 I have grabbed here a specific point.
The Dialog for the sharing user should present for files just the possibilities to give the permissions: "can edit" and "can share". The Options: create and delete are not useful there. Create isn't there at the moment, so thats not the issue. The Option delete is as @DeepDiver1975 agreed at #5836, maybe not really needed in the ui.
So the dialog can be shortened to "can edit" and "can share".
The sharing Dialog for folders should contain the "can edit" with a drop-down, as @jancborchardt suggested in his mockup #3224. Is the delete function for folders useful? or should it be the same as with files, merged with "can edit"? | 1.0 | Sharing dialog difference file / folder - Referencing to the design forward share dialog #3224 I have grabbed here a specific point.
The Dialog for the sharing user should present for files just the possibilities to give the permissions: "can edit" and "can share". The Options: create and delete are not useful there. Create isn't there at the moment, so thats not the issue. The Option delete is as @DeepDiver1975 agreed at #5836, maybe not really needed in the ui.
So the dialog can be shortened to "can edit" and "can share".
The sharing Dialog for folders should contain the "can edit" with a drop-down, as @jancborchardt suggested in his mockup #3224. Is the delete function for folders useful? or should it be the same as with files, merged with "can edit"? | non_main | sharing dialog difference file folder referencing to the design forward share dialog i have grabbed here a specific point the dialog for the sharing user should present for files just the possibilities to give the permissions can edit and can share the options create and delete are not useful there create isn t there at the moment so thats not the issue the option delete is as agreed at maybe not really needed in the ui so the dialog can be shortened to can edit and can share the sharing dialog for folders should contain the can edit with a drop down as jancborchardt suggested in his mockup is the delete function for folders useful or should it be the same as with files merged with can edit | 0 |
1,492 | 6,444,249,156 | IssuesEvent | 2017-08-12 08:38:54 | OpenRefine/OpenRefine | https://api.github.com/repos/OpenRefine/OpenRefine | closed | Localization support: use Weblate? | maintainability | It seems that some people are keen to contribute to OpenRefine's localization. It is a bit sad to see PRs like #924, where a lot of work was done but wasn't eventually delivered to the users because the translation is not complete.
I have come across [Weblate](https://weblate.org/), a nice web interface to help translators localize the interface. It would both decrease the technical requirements to start translating, and make it more of a collaborative process. It is tightly integrated with git, which means translations can be pushed to the git repository by Weblate, or proposed as pull requests.
It seems that it is possible to get free hosted accounts for open source projects. If the community is interested, I can ask them for one and do the initial set up. | True | Localization support: use Weblate? - It seems that some people are keen to contribute to OpenRefine's localization. It is a bit sad to see PRs like #924, where a lot of work was done but wasn't eventually delivered to the users because the translation is not complete.
I have come across [Weblate](https://weblate.org/), a nice web interface to help translators localize the interface. It would both decrease the technical requirements to start translating, and make it more of a collaborative process. It is tightly integrated with git, which means translations can be pushed to the git repository by Weblate, or proposed as pull requests.
It seems that it is possible to get free hosted accounts for open source projects. If the community is interested, I can ask them for one and do the initial set up. | main | localization support use weblate it seems that some people are keen to contribute to openrefine s localization it is a bit sad to see prs like where a lot of work was done but wasn t eventually delivered to the users because the translation is not complete i have come across a nice web interface to help translators localize the interface it would both decrease the technical requirements to start translating and make it more of a collaborative process it is tightly integrated with git which means translations can be pushed to the git repository by weblate or proposed as pull requests it seems that it is possible to get free hosted accounts for open source projects if the community is interested i can ask them for one and do the initial set up | 1 |
10,120 | 4,007,728,445 | IssuesEvent | 2016-05-12 19:08:20 | rubberduck-vba/Rubberduck | https://api.github.com/repos/rubberduck-vba/Rubberduck | closed | Parser gets stuck on "Loading references" with certain references | bug code-parsing | I have installed the v2 Alpha 1 version (build 2.0.0.2650) and when I open a project that contains certain references, or add a reference and then click the "Parse all opened projects" button, it gets stuck at "Loading references".
To reproduce:
If I open a new workbook, open the VBE, add a new module to Book1 (the new workbook), and click on the parse all button then it will parse ok and say "Ready". If I then, in this same workbook, add a reference to the Microsoft Scripting Runtime, for example, and then click on the parse button, it stops at "Loading references".
For reference I am on Windows 7 running 32-bit Excel 2010. | 1.0 | Parser gets stuck on "Loading references" with certain references - I have installed the v2 Alpha 1 version (build 2.0.0.2650) and when I open a project that contains certain references, or add a reference and then click the "Parse all opened projects" button, it gets stuck at "Loading references".
To reproduce:
If I open a new workbook, open the VBE, add a new module to Book1 (the new workbook), and click on the parse all button then it will parse ok and say "Ready". If I then, in this same workbook, add a reference to the Microsoft Scripting Runtime, for example, and then click on the parse button, it stops at "Loading references".
For reference I am on Windows 7 running 32-bit Excel 2010. | non_main | parser gets stuck on loading references with certain references i have installed the alpha version build and when i open a project that contains certain references or add a reference and then click the parse all opened projects button it gets stuck at loading references to reproduce if i open a new workbook open the vbe add a new module to the new workbook and click on the parse all button then it will parse ok and say ready if i then in this same workbook add a reference to the microsoft scripting runtime for example and then click on the parse button it stops at loading references for reference i am on windows running bit excel | 0 |
441,584 | 30,790,815,517 | IssuesEvent | 2023-07-31 16:02:13 | Azure-Samples/app-templates-java-openai-springapps | https://api.github.com/repos/Azure-Samples/app-templates-java-openai-springapps | opened | Update the documentation to explain how Azure SDK for Java is used | documentation enhancement | Update the documentation to explain how Azure SDK for Java is used
> Please provide us with the following information:
> ---------------------------------------------------------------
### This issue is for a: (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x] feature request
- [ ] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)
```
### Minimal steps to reproduce
> N/.A
### Any log messages given by the failure
> N/A
### Expected/desired behavior
> More details on Azure OpenAI
### Versions
> 1.0.0
### Mention any other details that might be useful
> ---------------------------------------------------------------
> Thanks! We'll be in touch soon.
| 1.0 | Update the documentation to explain how Azure SDK for Java is used - Update the documentation to explain how Azure SDK for Java is used
> Please provide us with the following information:
> ---------------------------------------------------------------
### This issue is for a: (mark with an `x`)
```
- [ ] bug report -> please search issues before submitting
- [x] feature request
- [ ] documentation issue or request
- [ ] regression (a behavior that used to work and stopped in a new release)
```
### Minimal steps to reproduce
> N/.A
### Any log messages given by the failure
> N/A
### Expected/desired behavior
> More details on Azure OpenAI
### Versions
> 1.0.0
### Mention any other details that might be useful
> ---------------------------------------------------------------
> Thanks! We'll be in touch soon.
| non_main | update the documentation to explain how azure sdk for java is used update the documentation to explain how azure sdk for java is used please provide us with the following information this issue is for a mark with an x bug report please search issues before submitting feature request documentation issue or request regression a behavior that used to work and stopped in a new release minimal steps to reproduce n a any log messages given by the failure n a expected desired behavior more details on azure openai versions mention any other details that might be useful thanks we ll be in touch soon | 0 |
1,282 | 5,412,022,697 | IssuesEvent | 2017-03-01 13:29:34 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Force user creation to be local | affects_2.3 feature_idea waiting_on_maintainer | Sometimes LDAP/NIS user names conflict with names we want to have local (like mysql, postgres). I'm worked around it other ways, but was told recently about luseradd.
I wonder if there would be interest in adding a force=yes to the user module such that is uses luseradd instead of useradd.
| True | Force user creation to be local - Sometimes LDAP/NIS user names conflict with names we want to have local (like mysql, postgres). I'm worked around it other ways, but was told recently about luseradd.
I wonder if there would be interest in adding a force=yes to the user module such that is uses luseradd instead of useradd.
| main | force user creation to be local sometimes ldap nis user names conflict with names we want to have local like mysql postgres i m worked around it other ways but was told recently about luseradd i wonder if there would be interest in adding a force yes to the user module such that is uses luseradd instead of useradd | 1 |
5,271 | 26,637,420,965 | IssuesEvent | 2023-01-24 23:32:43 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | Proposal: AWS::Serverless::Role | type/feature maintainer/need-response area/new-resource | **Description:**
The AWS::Serverless::Lambda Role system with AssumeRolePolicyDocument and his Policy templates is very handy, I would like to get this power to even more use cases.
**Example before:**
```yaml
ServiceRole:
Type: AWS::Serverless::Role
Properties:
AssumeRolePolicyDocument:
Version: 2012-10-17
Statement:
- Effect: Allow
Principal:
Service:
- textract.amazonaws.com
Action:
- sts:AssumeRole
Path: /
Policies:
- PolicyName: ReadS3
PolicyDocument:
Version: 2012-10-17
Statement:
- Effect: Allow
Action: s3:GetObject
Resource:
- !Sub ${TempBucket.Arn}/*
- PolicyName: PublishSNS
PolicyDocument:
Version: 2012-10-17
Statement:
- Effect: Allow
Action: sns:Publish
Resource:
- !Ref Topic
```
**After example:**
```yaml
ServiceRole:
Type: AWS::Serverless::Role
Properties:
AssumeRolePolicyDocument:
Version: 2012-10-17
Statement:
- Effect: Allow
Principal:
Service:
- textract.amazonaws.com
Action:
- sts:AssumeRole
Policies:
- S3ReadPolicy:
Bucket: !Ref Bucket
- SNSPublishMessagePolicy: PublishSNS
TopicName: !Ref Topic
```
Combined with #1573 it could look like:
```yaml
ServiceRole:
Type: AWS::Serverless::Role
Properties:
Service:
- textract.amazonaws.com
Policies:
- S3ReadPolicy:
Bucket: !Ref Bucket
- SNSPublishMessagePolicy: PublishSNS
TopicName: !Ref Topic
``` | True | Proposal: AWS::Serverless::Role - **Description:**
The AWS::Serverless::Lambda Role system with AssumeRolePolicyDocument and his Policy templates is very handy, I would like to get this power to even more use cases.
**Example before:**
```yaml
ServiceRole:
Type: AWS::Serverless::Role
Properties:
AssumeRolePolicyDocument:
Version: 2012-10-17
Statement:
- Effect: Allow
Principal:
Service:
- textract.amazonaws.com
Action:
- sts:AssumeRole
Path: /
Policies:
- PolicyName: ReadS3
PolicyDocument:
Version: 2012-10-17
Statement:
- Effect: Allow
Action: s3:GetObject
Resource:
- !Sub ${TempBucket.Arn}/*
- PolicyName: PublishSNS
PolicyDocument:
Version: 2012-10-17
Statement:
- Effect: Allow
Action: sns:Publish
Resource:
- !Ref Topic
```
**After example:**
```yaml
ServiceRole:
Type: AWS::Serverless::Role
Properties:
AssumeRolePolicyDocument:
Version: 2012-10-17
Statement:
- Effect: Allow
Principal:
Service:
- textract.amazonaws.com
Action:
- sts:AssumeRole
Policies:
- S3ReadPolicy:
Bucket: !Ref Bucket
- SNSPublishMessagePolicy: PublishSNS
TopicName: !Ref Topic
```
Combined with #1573 it could look like:
```yaml
ServiceRole:
Type: AWS::Serverless::Role
Properties:
Service:
- textract.amazonaws.com
Policies:
- S3ReadPolicy:
Bucket: !Ref Bucket
- SNSPublishMessagePolicy: PublishSNS
TopicName: !Ref Topic
``` | main | proposal aws serverless role description the aws serverless lambda role system with assumerolepolicydocument and his policy templates is very handy i would like to get this power to even more use cases example before yaml servicerole type aws serverless role properties assumerolepolicydocument version statement effect allow principal service textract amazonaws com action sts assumerole path policies policyname policydocument version statement effect allow action getobject resource sub tempbucket arn policyname publishsns policydocument version statement effect allow action sns publish resource ref topic after example yaml servicerole type aws serverless role properties assumerolepolicydocument version statement effect allow principal service textract amazonaws com action sts assumerole policies bucket ref bucket snspublishmessagepolicy publishsns topicname ref topic combined with it could look like yaml servicerole type aws serverless role properties service textract amazonaws com policies bucket ref bucket snspublishmessagepolicy publishsns topicname ref topic | 1 |
251 | 3,000,429,917 | IssuesEvent | 2015-07-24 01:33:21 | Homebrew/homebrew | https://api.github.com/repos/Homebrew/homebrew | closed | Error: invalid byte sequence in UTF-8 | maintainer feedback | Hello,
I'm sorry to open an issue for what looks like a common problem but I have been cracking my teeth on this for days now.
Brew doctor throws the
```
Error: invalid byte sequence in UTF-8 with:
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:44:in `=~'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:44:in `chop_basename'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:338:in `plus'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:331:in `+'
/usr/local/Library/Homebrew/extend/pathname.rb:295:in `resolved_path_exists?'
/usr/local/Library/Homebrew/cmd/doctor.rb:241:in `block (2 levels) in check_for_broken_symlinks'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:543:in `block in find'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/find.rb:43:in `block in find'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/find.rb:42:in `catch'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/find.rb:42:in `find'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:543:in `find'
/usr/local/Library/Homebrew/cmd/doctor.rb:240:in `block in check_for_broken_symlinks'
/usr/local/Library/Homebrew/cmd/doctor.rb:239:in `each'
/usr/local/Library/Homebrew/cmd/doctor.rb:239:in `check_for_broken_symlinks'
/usr/local/Library/Homebrew/cmd/doctor.rb:1258:in `block in doctor'
/usr/local/Library/Homebrew/cmd/doctor.rb:1256:in `each'
/usr/local/Library/Homebrew/cmd/doctor.rb:1256:in `doctor'
/usr/local/Library/brew.rb:124:in `<main>'
```
First, I checked the path looking for strange characters. This was the output:
/usr/local/bin:/usr/local/sbin:/usr/local/heroku/bin:/opt/local/bin:/opt/local/sbin:/opt/local/bin:/opt/local/sbin:/usr/local/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/git/bin:/usr/local/MacGPG2/bin
I found a broken symlink in my ruby.frameworks from Current to 1.8 and fixed that, but the problem persisted. I've gone through every issue and attempted many fixes, but none seem to work. I also fully reinstalled the essential.pkg from the Yosemite installer, hoping to replace ruby and any other development-related changes that could have been made in the last few years, but that didn't help either.
Here is the output of brew --config:
HOMEBREW_VERSION: 0.9.5
ORIGIN: https://github.com/Homebrew/homebrew
HEAD: a9d599bf799bb898150d13b073541a2821409b49
Last commit: 2 hours ago
HOMEBREW_PREFIX: /usr/local
HOMEBREW_CELLAR: /usr/local/Cellar
HOMEBREW_BOTTLE_DOMAIN: https://homebrew.bintray.com
CPU: 8-core 64-bit ivybridge
OS X: 10.10.4-x86_64
Xcode: 6.4
CLT: 6.4.0.0.1.1435007323
Clang: 6.1 build 602
MacPorts/Fink: /opt/local/bin/port
X11: 2.7.7 => /opt/X11
System Ruby: 2.0.0-p481
Perl: /opt/local/bin/perl => /opt/local/bin/perl5.16
Python: /usr/bin/python => /System/Library/Frameworks/Python.framework/Versions/2.7/bin/python2.7
Ruby: /usr/bin/ruby
Java: 1.8.0_45
And my environment:
TERM_PROGRAM=Apple_Terminal
SHELL=/bin/bash
TERM=xterm-256color
TMPDIR=/var/folders/l4/lgvz2_w56ml52f35zhmlhgwh0000gn/T/
Apple_PubSub_Socket_Render=/private/tmp/com.apple.launchd.5tDsmG0iTy/Render
TERM_PROGRAM_VERSION=343.7
TERM_SESSION_ID=0727783D-6A7C-490B-A334-80B231E8E916
USER=xxxx
SSH_AUTH_SOCK=/private/tmp/com.apple.launchd.5GlRyo6mnC/Listeners
__CF_USER_TEXT_ENCODING=0x1F5:0x0:0x0
PATH=/usr/local/bin:/usr/local/sbin:/usr/local/heroku/bin:/opt/local/bin:/opt/local/sbin:/opt/local/bin:/opt/local/sbin:/usr/local/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/git/bin:/usr/local/MacGPG2/bin
PWD=/usr/local
LANG=en_US.UTF-8
XPC_FLAGS=0x0
XPC_SERVICE_NAME=0
SHLVL=1
HOME=/Users/xxxxxxx
LOGNAME=xxxxxx
DISPLAY=/private/tmp/com.apple.launchd.wFAxs8HiE6/org.macosforge.xquartz:0
_=/usr/bin/env
OLDPWD=/usr/local/Library/Homebrew
Thank you for your help!! It is much appreciated. | True | Error: invalid byte sequence in UTF-8 - Hello,
I'm sorry to open an issue for what looks like a common problem but I have been cracking my teeth on this for days now.
Brew doctor throws the
```
Error: invalid byte sequence in UTF-8 with:
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:44:in `=~'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:44:in `chop_basename'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:338:in `plus'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:331:in `+'
/usr/local/Library/Homebrew/extend/pathname.rb:295:in `resolved_path_exists?'
/usr/local/Library/Homebrew/cmd/doctor.rb:241:in `block (2 levels) in check_for_broken_symlinks'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:543:in `block in find'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/find.rb:43:in `block in find'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/find.rb:42:in `catch'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/find.rb:42:in `find'
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/pathname.rb:543:in `find'
/usr/local/Library/Homebrew/cmd/doctor.rb:240:in `block in check_for_broken_symlinks'
/usr/local/Library/Homebrew/cmd/doctor.rb:239:in `each'
/usr/local/Library/Homebrew/cmd/doctor.rb:239:in `check_for_broken_symlinks'
/usr/local/Library/Homebrew/cmd/doctor.rb:1258:in `block in doctor'
/usr/local/Library/Homebrew/cmd/doctor.rb:1256:in `each'
/usr/local/Library/Homebrew/cmd/doctor.rb:1256:in `doctor'
/usr/local/Library/brew.rb:124:in `<main>'
```
First, I checked the path looking for strange characters. This was the output:
/usr/local/bin:/usr/local/sbin:/usr/local/heroku/bin:/opt/local/bin:/opt/local/sbin:/opt/local/bin:/opt/local/sbin:/usr/local/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/git/bin:/usr/local/MacGPG2/bin
I found a broken symlink in my ruby.frameworks from Current to 1.8 and fixed that, but the problem persisted. I've gone through every issue and attempted many fixes, but none seem to work. I also fully reinstalled the essential.pkg from the Yosemite installer, hoping to replace ruby and any other development-related changes that could have been made in the last few years, but that didn't help either.
Here is the output of brew --config:
HOMEBREW_VERSION: 0.9.5
ORIGIN: https://github.com/Homebrew/homebrew
HEAD: a9d599bf799bb898150d13b073541a2821409b49
Last commit: 2 hours ago
HOMEBREW_PREFIX: /usr/local
HOMEBREW_CELLAR: /usr/local/Cellar
HOMEBREW_BOTTLE_DOMAIN: https://homebrew.bintray.com
CPU: 8-core 64-bit ivybridge
OS X: 10.10.4-x86_64
Xcode: 6.4
CLT: 6.4.0.0.1.1435007323
Clang: 6.1 build 602
MacPorts/Fink: /opt/local/bin/port
X11: 2.7.7 => /opt/X11
System Ruby: 2.0.0-p481
Perl: /opt/local/bin/perl => /opt/local/bin/perl5.16
Python: /usr/bin/python => /System/Library/Frameworks/Python.framework/Versions/2.7/bin/python2.7
Ruby: /usr/bin/ruby
Java: 1.8.0_45
And my environment:
TERM_PROGRAM=Apple_Terminal
SHELL=/bin/bash
TERM=xterm-256color
TMPDIR=/var/folders/l4/lgvz2_w56ml52f35zhmlhgwh0000gn/T/
Apple_PubSub_Socket_Render=/private/tmp/com.apple.launchd.5tDsmG0iTy/Render
TERM_PROGRAM_VERSION=343.7
TERM_SESSION_ID=0727783D-6A7C-490B-A334-80B231E8E916
USER=xxxx
SSH_AUTH_SOCK=/private/tmp/com.apple.launchd.5GlRyo6mnC/Listeners
__CF_USER_TEXT_ENCODING=0x1F5:0x0:0x0
PATH=/usr/local/bin:/usr/local/sbin:/usr/local/heroku/bin:/opt/local/bin:/opt/local/sbin:/opt/local/bin:/opt/local/sbin:/usr/local/bin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/git/bin:/usr/local/MacGPG2/bin
PWD=/usr/local
LANG=en_US.UTF-8
XPC_FLAGS=0x0
XPC_SERVICE_NAME=0
SHLVL=1
HOME=/Users/xxxxxxx
LOGNAME=xxxxxx
DISPLAY=/private/tmp/com.apple.launchd.wFAxs8HiE6/org.macosforge.xquartz:0
_=/usr/bin/env
OLDPWD=/usr/local/Library/Homebrew
Thank you for your help!! It is much appreciated. | main | error invalid byte sequence in utf hello i m sorry to open an issue for what looks like a common problem but i have been cracking my teeth on this for days now brew doctor throws the error invalid byte sequence in utf with system library frameworks ruby framework versions usr lib ruby pathname rb in system library frameworks ruby framework versions usr lib ruby pathname rb in chop basename system library frameworks ruby framework versions usr lib ruby pathname rb in plus system library frameworks ruby framework versions usr lib ruby pathname rb in usr local library homebrew extend pathname rb in resolved path exists usr local library homebrew cmd doctor rb in block levels in check for broken symlinks system library frameworks ruby framework versions usr lib ruby pathname rb in block in find system library frameworks ruby framework versions usr lib ruby find rb in block in find system library frameworks ruby framework versions usr lib ruby find rb in catch system library frameworks ruby framework versions usr lib ruby find rb in find system library frameworks ruby framework versions usr lib ruby pathname rb in find usr local library homebrew cmd doctor rb in block in check for broken symlinks usr local library homebrew cmd doctor rb in each usr local library homebrew cmd doctor rb in check for broken symlinks usr local library homebrew cmd doctor rb in block in doctor usr local library homebrew cmd doctor rb in each usr local library homebrew cmd doctor rb in doctor usr local library brew rb in first i checked the path looking for strange characters this was the output usr local bin usr local sbin usr local heroku bin opt local bin opt local sbin opt local bin opt local sbin usr local bin usr local bin usr bin bin usr sbin sbin opt bin usr local git bin usr local bin i found a broken symlink in my ruby frameworks from current to and fixed that but the problem persisted i ve gone through every issue and attempted many fixes but none seem to work i also fully reinstalled the essential pkg from the yosemite installer hoping to replace ruby and any other development related changes that could have been made in the last few years but that didn t help either here is the output of brew config homebrew version origin head last commit hours ago homebrew prefix usr local homebrew cellar usr local cellar homebrew bottle domain cpu core bit ivybridge os x xcode clt clang build macports fink opt local bin port opt system ruby perl opt local bin perl opt local bin python usr bin python system library frameworks python framework versions bin ruby usr bin ruby java and my environment term program apple terminal shell bin bash term xterm tmpdir var folders t apple pubsub socket render private tmp com apple launchd render term program version term session id user xxxx ssh auth sock private tmp com apple launchd listeners cf user text encoding path usr local bin usr local sbin usr local heroku bin opt local bin opt local sbin opt local bin opt local sbin usr local bin usr local bin usr bin bin usr sbin sbin opt bin usr local git bin usr local bin pwd usr local lang en us utf xpc flags xpc service name shlvl home users xxxxxxx logname xxxxxx display private tmp com apple launchd org macosforge xquartz usr bin env oldpwd usr local library homebrew thank you for your help it is much appreciated | 1 |
3,056 | 11,440,985,849 | IssuesEvent | 2020-02-05 10:43:47 | precice/precice | https://api.github.com/repos/precice/precice | opened | Why do we run our unit tests against the latest release, if our baseline is PETSc 3.6.2? | bug maintainability | See:
https://github.com/precice/precice/blob/4f03e9464280670f3445e13eec587dd444af90e5/tools/ci/travis-install-dependencies.sh#L51
Problem with this approach: We only see it very late, if the our baseline verison of PETSc does not support the latests revision of preCICE anymore. | True | Why do we run our unit tests against the latest release, if our baseline is PETSc 3.6.2? - See:
https://github.com/precice/precice/blob/4f03e9464280670f3445e13eec587dd444af90e5/tools/ci/travis-install-dependencies.sh#L51
Problem with this approach: We only see it very late, if the our baseline verison of PETSc does not support the latests revision of preCICE anymore. | main | why do we run our unit tests against the latest release if our baseline is petsc see problem with this approach we only see it very late if the our baseline verison of petsc does not support the latests revision of precice anymore | 1 |
64,658 | 3,213,854,169 | IssuesEvent | 2015-10-06 21:46:11 | YetiForceCompany/YetiForceCRM | https://api.github.com/repos/YetiForceCompany/YetiForceCRM | closed | [improvement] Possibility to add field to user | Label::Core Priority::#2 Normal Type::Discussion | If you could add field to user, e.g. phone number, it could be used in quotes pdf-templates, etc. | 1.0 | [improvement] Possibility to add field to user - If you could add field to user, e.g. phone number, it could be used in quotes pdf-templates, etc. | non_main | possibility to add field to user if you could add field to user e g phone number it could be used in quotes pdf templates etc | 0 |
1,431 | 2,545,072,047 | IssuesEvent | 2015-01-29 15:02:07 | AAndharia/PSB | https://api.github.com/repos/AAndharia/PSB | closed | File Report - UI & Layout must be exact same as old - see attached | New Ready For Testing | Task1
BM File Report.pdf - it is Report UI for previous version
KBS File Report.pdf - it is current Report UI, we need to make it same as BM File Report.pdf
Task2
Additionally, we'll also need footer to appear on each page of pdf. | 1.0 | File Report - UI & Layout must be exact same as old - see attached - Task1
BM File Report.pdf - it is Report UI for previous version
KBS File Report.pdf - it is current Report UI, we need to make it same as BM File Report.pdf
Task2
Additionally, we'll also need footer to appear on each page of pdf. | non_main | file report ui layout must be exact same as old see attached bm file report pdf it is report ui for previous version kbs file report pdf it is current report ui we need to make it same as bm file report pdf additionally we ll also need footer to appear on each page of pdf | 0 |
364,136 | 25,481,819,219 | IssuesEvent | 2022-11-25 22:45:52 | GuiChagas/Sistema-de-Judocas | https://api.github.com/repos/GuiChagas/Sistema-de-Judocas | closed | Inconsistência dos atores na especificação de requisitos | documentation invalid | 
Segundo a especificação de requisitos, alunos e professores não irão interagir com o sistema, portanto eles não deveriam ser considerados atores do sistema. | 1.0 | Inconsistência dos atores na especificação de requisitos - 
Segundo a especificação de requisitos, alunos e professores não irão interagir com o sistema, portanto eles não deveriam ser considerados atores do sistema. | non_main | inconsistência dos atores na especificação de requisitos segundo a especificação de requisitos alunos e professores não irão interagir com o sistema portanto eles não deveriam ser considerados atores do sistema | 0 |
121,512 | 17,658,738,466 | IssuesEvent | 2021-08-21 03:59:15 | hygieia/hygieia-apiaudit | https://api.github.com/repos/hygieia/hygieia-apiaudit | closed | CVE-2020-25649 (High) detected in jackson-databind-2.10.5.jar | wontfix security vulnerability | ## CVE-2020-25649 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.10.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: hygieia-apiaudit/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.5/jackson-databind-2.10.5.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.9.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.2.9.RELEASE.jar
- :x: **jackson-databind-2.10.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/hygieia/hygieia-apiaudit/commits/eb22f7ef148688a43fc0af0d80e807da2cc3ab48">eb22f7ef148688a43fc0af0d80e807da2cc3ab48</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in FasterXML Jackson Databind, where it did not have entity expansion secured properly. This flaw allows vulnerability to XML external entity (XXE) attacks. The highest threat from this vulnerability is data integrity.
<p>Publish Date: 2020-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25649>CVE-2020-25649</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2589">https://github.com/FasterXML/jackson-databind/issues/2589</a></p>
<p>Release Date: 2020-12-03</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.4,2.9.10.7,2.10.5.1,2.11.0.rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-25649 (High) detected in jackson-databind-2.10.5.jar - ## CVE-2020-25649 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.10.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: hygieia-apiaudit/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.10.5/jackson-databind-2.10.5.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.9.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.2.9.RELEASE.jar
- :x: **jackson-databind-2.10.5.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://api.github.com/repos/hygieia/hygieia-apiaudit/commits/eb22f7ef148688a43fc0af0d80e807da2cc3ab48">eb22f7ef148688a43fc0af0d80e807da2cc3ab48</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A flaw was found in FasterXML Jackson Databind, where it did not have entity expansion secured properly. This flaw allows vulnerability to XML external entity (XXE) attacks. The highest threat from this vulnerability is data integrity.
<p>Publish Date: 2020-12-03
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-25649>CVE-2020-25649</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2589">https://github.com/FasterXML/jackson-databind/issues/2589</a></p>
<p>Release Date: 2020-12-03</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.6.7.4,2.9.10.7,2.10.5.1,2.11.0.rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file hygieia apiaudit pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library found in head commit a href found in base branch main vulnerability details a flaw was found in fasterxml jackson databind where it did not have entity expansion secured properly this flaw allows vulnerability to xml external entity xxe attacks the highest threat from this vulnerability is data integrity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind step up your open source security game with whitesource | 0 |
247,699 | 20,987,715,553 | IssuesEvent | 2022-03-29 06:07:43 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: kv95/enc=true/nodes=3 failed | C-test-failure O-robot O-roachtest branch-master release-blocker | roachtest.kv95/enc=true/nodes=3 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713655&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713655&tab=artifacts#/kv95/enc=true/nodes=3) on master @ [29716850b181718594663889ddb5f479fef7a305](https://github.com/cockroachdb/cockroach/commits/29716850b181718594663889ddb5f479fef7a305):
```
The test failed on branch=master, cloud=aws:
test artifacts and logs in: /artifacts/kv95/enc=true/nodes=3/run_1
cluster.go:1868,kv.go:85,kv.go:270,test_runner.go:875: one or more parallel execution failure
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167
| github.com/cockroachdb/cockroach/pkg/roachprod.Start
| github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660
| main.(*clusterImpl).StartE
| main/pkg/cmd/roachtest/cluster.go:1826
| main.(*clusterImpl).Start
| main/pkg/cmd/roachtest/cluster.go:1867
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func2
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:85
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func3
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:270
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:875
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (2) one or more parallel execution failure
Error types: (1) *withstack.withStack (2) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv95/enc=true/nodes=3.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: kv95/enc=true/nodes=3 failed - roachtest.kv95/enc=true/nodes=3 [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713655&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4713655&tab=artifacts#/kv95/enc=true/nodes=3) on master @ [29716850b181718594663889ddb5f479fef7a305](https://github.com/cockroachdb/cockroach/commits/29716850b181718594663889ddb5f479fef7a305):
```
The test failed on branch=master, cloud=aws:
test artifacts and logs in: /artifacts/kv95/enc=true/nodes=3/run_1
cluster.go:1868,kv.go:85,kv.go:270,test_runner.go:875: one or more parallel execution failure
(1) attached stack trace
-- stack trace:
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2042
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:1923
| github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start
| github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:167
| github.com/cockroachdb/cockroach/pkg/roachprod.Start
| github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:660
| main.(*clusterImpl).StartE
| main/pkg/cmd/roachtest/cluster.go:1826
| main.(*clusterImpl).Start
| main/pkg/cmd/roachtest/cluster.go:1867
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func2
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:85
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerKV.func3
| github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/kv.go:270
| main.(*testRunner).runTest.func2
| main/pkg/cmd/roachtest/test_runner.go:875
| runtime.goexit
| GOROOT/src/runtime/asm_amd64.s:1581
Wraps: (2) one or more parallel execution failure
Error types: (1) *withstack.withStack (2) *errutil.leafError
```
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/kv-triage
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*kv95/enc=true/nodes=3.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| non_main | roachtest enc true nodes failed roachtest enc true nodes with on master the test failed on branch master cloud aws test artifacts and logs in artifacts enc true nodes run cluster go kv go kv go test runner go one or more parallel execution failure attached stack trace stack trace github com cockroachdb cockroach pkg roachprod install syncedcluster parallele github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster parallel github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster start github com cockroachdb cockroach pkg roachprod install cockroach go github com cockroachdb cockroach pkg roachprod start github com cockroachdb cockroach pkg roachprod roachprod go main clusterimpl starte main pkg cmd roachtest cluster go main clusterimpl start main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests registerkv github com cockroachdb cockroach pkg cmd roachtest tests kv go github com cockroachdb cockroach pkg cmd roachtest tests registerkv github com cockroachdb cockroach pkg cmd roachtest tests kv go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps one or more parallel execution failure error types withstack withstack errutil leaferror help see see cc cockroachdb kv triage | 0 |
1,195 | 5,116,662,611 | IssuesEvent | 2017-01-07 06:38:30 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | There is no restore organs function | Maintainability/Hinders improvements | Turns out issue #21408 is way more general than originally thought. There is NO restore organs function anywhere in the code. This means no type of healing restores organs and the only way to regain organs is through surgery.
This means organs are not restored for:
Changelings using stasis or regenerate.
healing staffs.
probably admin heals as well?
Anything else you would expect to heal organs.
| True | There is no restore organs function - Turns out issue #21408 is way more general than originally thought. There is NO restore organs function anywhere in the code. This means no type of healing restores organs and the only way to regain organs is through surgery.
This means organs are not restored for:
Changelings using stasis or regenerate.
healing staffs.
probably admin heals as well?
Anything else you would expect to heal organs.
| main | there is no restore organs function turns out issue is way more general than originally thought there is no restore organs function anywhere in the code this means no type of healing restores organs and the only way to regain organs is through surgery this means organs are not restored for changelings using stasis or regenerate healing staffs probably admin heals as well anything else you would expect to heal organs | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.