Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1,824 | 6,577,335,030 | IssuesEvent | 2017-09-12 00:11:21 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | postgresql_db owner for tables and views | affects_2.0 bug_report feature_idea waiting_on_maintainer | ##### ISSUE TYPE
- Bug
- Feature Idea
##### COMPONENT NAME
postgresql_db
##### ANSIBLE VERSION
```
ansible 2.0.2.0
config file = /home/toga/PycharmProjects/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
ansible.cfg
```
[defaults]
inventory = hosts
ask_vault_pass=True
[privilege_escalation]
become=True
become_ask_pass=True
```
##### OS / ENVIRONMENT
```
Ubuntu 14.04.4 LTS
virtualenv with ansible
PostgreSQL 9.3
```
##### SUMMARY
The module `postgresql_db` should recursively set the database owner for all included tables, views and sequences.
`ALTER DATABASE %s OWNER TO %s` only affects the database but not the tables in the database.
After restoring a dump with the postgres user, the owner of all tables where set to postgres and won't be reassigned to the user named in the task.
##### STEPS TO REPRODUCE
Connect to psql
```
CREATE USER test;
CREATE DATABASE test OWNER postgres;
\c test
CREATE TABLE films (
code char(5) CONSTRAINT firstkey PRIMARY KEY,
title varchar(40) NOT NULL,
did integer NOT NULL,
date_prod date,
kind varchar(10),
len interval hour to minute
);
```
User ansible
```
ansible localhost --connection=local -b --become-user postgres -m postgresql_db -a "name=test owner=test"
```
##### EXPECTED RESULTS
```
test=# \l
Name | Owner | Encoding | Collate | Ctype | Access privileges
-----------+----------+----------+-------------+-------------+-----------------------
test | test | UTF8 | de_DE.UTF-8 | de_DE.UTF-8 | =Tc/test2 +
| | | | | test2=CTc/test2
test=# \d
List of relations
Schema | Name | Type | Owner
--------+-------+-------+----------
public | films | table | test
(1 row)
```
##### ACTUAL RESULTS
```
test=# \l
Name | Owner | Encoding | Collate | Ctype | Access privileges
-----------+----------+----------+-------------+-------------+-----------------------
test | test | UTF8 | de_DE.UTF-8 | de_DE.UTF-8 | =Tc/test2 +
| | | | | test2=CTc/test2
test=# \d
List of relations
Schema | Name | Type | Owner
--------+-------+-------+----------
public | films | table | postgres
(1 row)
```
| True | postgresql_db owner for tables and views - ##### ISSUE TYPE
- Bug
- Feature Idea
##### COMPONENT NAME
postgresql_db
##### ANSIBLE VERSION
```
ansible 2.0.2.0
config file = /home/toga/PycharmProjects/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
ansible.cfg
```
[defaults]
inventory = hosts
ask_vault_pass=True
[privilege_escalation]
become=True
become_ask_pass=True
```
##### OS / ENVIRONMENT
```
Ubuntu 14.04.4 LTS
virtualenv with ansible
PostgreSQL 9.3
```
##### SUMMARY
The module `postgresql_db` should recursively set the database owner for all included tables, views and sequences.
`ALTER DATABASE %s OWNER TO %s` only affects the database but not the tables in the database.
After restoring a dump with the postgres user, the owner of all tables where set to postgres and won't be reassigned to the user named in the task.
##### STEPS TO REPRODUCE
Connect to psql
```
CREATE USER test;
CREATE DATABASE test OWNER postgres;
\c test
CREATE TABLE films (
code char(5) CONSTRAINT firstkey PRIMARY KEY,
title varchar(40) NOT NULL,
did integer NOT NULL,
date_prod date,
kind varchar(10),
len interval hour to minute
);
```
User ansible
```
ansible localhost --connection=local -b --become-user postgres -m postgresql_db -a "name=test owner=test"
```
##### EXPECTED RESULTS
```
test=# \l
Name | Owner | Encoding | Collate | Ctype | Access privileges
-----------+----------+----------+-------------+-------------+-----------------------
test | test | UTF8 | de_DE.UTF-8 | de_DE.UTF-8 | =Tc/test2 +
| | | | | test2=CTc/test2
test=# \d
List of relations
Schema | Name | Type | Owner
--------+-------+-------+----------
public | films | table | test
(1 row)
```
##### ACTUAL RESULTS
```
test=# \l
Name | Owner | Encoding | Collate | Ctype | Access privileges
-----------+----------+----------+-------------+-------------+-----------------------
test | test | UTF8 | de_DE.UTF-8 | de_DE.UTF-8 | =Tc/test2 +
| | | | | test2=CTc/test2
test=# \d
List of relations
Schema | Name | Type | Owner
--------+-------+-------+----------
public | films | table | postgres
(1 row)
```
| main | postgresql db owner for tables and views issue type bug feature idea component name postgresql db ansible version ansible config file home toga pycharmprojects ansible ansible cfg configured module search path default w o overrides configuration ansible cfg inventory hosts ask vault pass true become true become ask pass true os environment ubuntu lts virtualenv with ansible postgresql summary the module postgresql db should recursively set the database owner for all included tables views and sequences alter database s owner to s only affects the database but not the tables in the database after restoring a dump with the postgres user the owner of all tables where set to postgres and won t be reassigned to the user named in the task steps to reproduce connect to psql create user test create database test owner postgres c test create table films code char constraint firstkey primary key title varchar not null did integer not null date prod date kind varchar len interval hour to minute user ansible ansible localhost connection local b become user postgres m postgresql db a name test owner test expected results test l name owner encoding collate ctype access privileges test test de de utf de de utf tc ctc test d list of relations schema name type owner public films table test row actual results test l name owner encoding collate ctype access privileges test test de de utf de de utf tc ctc test d list of relations schema name type owner public films table postgres row | 1 |
19,853 | 10,428,319,387 | IssuesEvent | 2019-09-16 22:11:53 | gate5/test2 | https://api.github.com/repos/gate5/test2 | opened | CVE-2018-10054 (High) detected in h2-1.4.187.jar | security vulnerability | ## CVE-2018-10054 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.187.jar</b></p></summary>
<p>null</p>
<p>Path to dependency file: /tmp/ws-scm/test2/pom.xml</p>
<p>Path to vulnerable library: epository/com/h2database/h2/1.4.187/h2-1.4.187.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.187.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gate5/test2/commit/9caf3d214cc15f500423a2e431ea111cf9526739">9caf3d214cc15f500423a2e431ea111cf9526739</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
H2 1.4.197, as used in Datomic before 0.9.5697 and other products, allows remote code execution because CREATE ALIAS can execute arbitrary Java code.
<p>Publish Date: 2018-04-11
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054>CVE-2018-10054</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054</a></p>
<p>Release Date: 2018-04-11</p>
<p>Fix Resolution: 1.4.198</p>
</p>
</details>
<p></p>
| True | CVE-2018-10054 (High) detected in h2-1.4.187.jar - ## CVE-2018-10054 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>h2-1.4.187.jar</b></p></summary>
<p>null</p>
<p>Path to dependency file: /tmp/ws-scm/test2/pom.xml</p>
<p>Path to vulnerable library: epository/com/h2database/h2/1.4.187/h2-1.4.187.jar</p>
<p>
Dependency Hierarchy:
- :x: **h2-1.4.187.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/gate5/test2/commit/9caf3d214cc15f500423a2e431ea111cf9526739">9caf3d214cc15f500423a2e431ea111cf9526739</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
H2 1.4.197, as used in Datomic before 0.9.5697 and other products, allows remote code execution because CREATE ALIAS can execute arbitrary Java code.
<p>Publish Date: 2018-04-11
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054>CVE-2018-10054</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-10054</a></p>
<p>Release Date: 2018-04-11</p>
<p>Fix Resolution: 1.4.198</p>
</p>
</details>
<p></p>
| non_main | cve high detected in jar cve high severity vulnerability vulnerable library jar null path to dependency file tmp ws scm pom xml path to vulnerable library epository com jar dependency hierarchy x jar vulnerable library found in head commit a href vulnerability details as used in datomic before and other products allows remote code execution because create alias can execute arbitrary java code publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
342,333 | 24,738,950,544 | IssuesEvent | 2022-10-21 02:10:16 | AY2223S1-CS2103T-W16-2/tp | https://api.github.com/repos/AY2223S1-CS2103T-W16-2/tp | closed | Remove hard line breaks from markdown files | priority.High documentation.UserGuide documentation.DeveloperGuide documentation bug | * They make it harder to change text
* They don't follow our module's markdown coding style
* They cause rendering issues as different parsers interpret hard breaks differently
We need to wrap this up quickly as we have a draft UG/DG submission due. | 3.0 | Remove hard line breaks from markdown files - * They make it harder to change text
* They don't follow our module's markdown coding style
* They cause rendering issues as different parsers interpret hard breaks differently
We need to wrap this up quickly as we have a draft UG/DG submission due. | non_main | remove hard line breaks from markdown files they make it harder to change text they don t follow our module s markdown coding style they cause rendering issues as different parsers interpret hard breaks differently we need to wrap this up quickly as we have a draft ug dg submission due | 0 |
501,797 | 14,533,719,975 | IssuesEvent | 2020-12-15 01:14:30 | goharbor/harbor-operator | https://api.github.com/repos/goharbor/harbor-operator | closed | harbor-operator harbor-cluster controller reconcile dead loop. | area/reconciler kind/bug priority/high reconciler/harbor-cluster release/1.0 | What can we help you?

All service is ready, but harbor-cluster is always reconciled into infinite cycles. | 1.0 | harbor-operator harbor-cluster controller reconcile dead loop. - What can we help you?

All service is ready, but harbor-cluster is always reconciled into infinite cycles. | non_main | harbor operator harbor cluster controller reconcile dead loop what can we help you all service is ready but harbor cluster is always reconciled into infinite cycles | 0 |
5,280 | 26,682,333,414 | IssuesEvent | 2023-01-26 18:45:34 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Persist visibility of table inspector and its sections within localStorage | type: enhancement work: frontend status: ready restricted: maintainers | ## Current behavior
- The table inspector is enabled when loading a page. Users can hide it, but it always re-appears when reloading.
- The sections within the table inspector work in a similar manner, though some are set to hidden by default.
## Desired behavior
- The visibility of the table inspector and each of its sections is synchronized with localStorage.
- We'll need to figure out a proper schema for storing this data in localStorage, taking into account the likelihood of eventually putting more data in localStorage and possibly needing to migrate existing localStorage data.
This is marked as restricted because we're not yet persisting anything to localStorage and I want to make sure this we set this up in a manner that provides a clean integration with our Svelte stores.
| True | Persist visibility of table inspector and its sections within localStorage - ## Current behavior
- The table inspector is enabled when loading a page. Users can hide it, but it always re-appears when reloading.
- The sections within the table inspector work in a similar manner, though some are set to hidden by default.
## Desired behavior
- The visibility of the table inspector and each of its sections is synchronized with localStorage.
- We'll need to figure out a proper schema for storing this data in localStorage, taking into account the likelihood of eventually putting more data in localStorage and possibly needing to migrate existing localStorage data.
This is marked as restricted because we're not yet persisting anything to localStorage and I want to make sure this we set this up in a manner that provides a clean integration with our Svelte stores.
| main | persist visibility of table inspector and its sections within localstorage current behavior the table inspector is enabled when loading a page users can hide it but it always re appears when reloading the sections within the table inspector work in a similar manner though some are set to hidden by default desired behavior the visibility of the table inspector and each of its sections is synchronized with localstorage we ll need to figure out a proper schema for storing this data in localstorage taking into account the likelihood of eventually putting more data in localstorage and possibly needing to migrate existing localstorage data this is marked as restricted because we re not yet persisting anything to localstorage and i want to make sure this we set this up in a manner that provides a clean integration with our svelte stores | 1 |
322,071 | 27,579,121,405 | IssuesEvent | 2023-03-08 15:03:50 | onc-healthit/onc-certification-g10-test-kit | https://api.github.com/repos/onc-healthit/onc-certification-g10-test-kit | closed | Check bulk export status of cancelled job in bulk v2 | g10-test-kit add constraint v3.5.0 | > Following the delete request, when subsequent requests are made to the polling location, the server SHALL return a 404 Not Found error and an associated FHIR OperationOutcome in JSON format.
http://hl7.org/fhir/uv/bulkdata/STU2/export.html#bulk-data-delete-request
A test needs to be added for these requirements. | 1.0 | Check bulk export status of cancelled job in bulk v2 - > Following the delete request, when subsequent requests are made to the polling location, the server SHALL return a 404 Not Found error and an associated FHIR OperationOutcome in JSON format.
http://hl7.org/fhir/uv/bulkdata/STU2/export.html#bulk-data-delete-request
A test needs to be added for these requirements. | non_main | check bulk export status of cancelled job in bulk following the delete request when subsequent requests are made to the polling location the server shall return a not found error and an associated fhir operationoutcome in json format a test needs to be added for these requirements | 0 |
5,863 | 31,768,815,833 | IssuesEvent | 2023-09-12 10:25:17 | markus-wa/demoinfocs-golang | https://api.github.com/repos/markus-wa/demoinfocs-golang | closed | CS2 corrupt input | needs: investigation (maintainer) | **Describe the bug**
This demo causes a parser error:
```
panic: snappy: corrupt input [recovered]
panic: snappy: corrupt input
goroutine 1 [running]:
github.com/markus-wa/demoinfocs-golang/v4/pkg/demoinfocs.recoverFromUnexpectedEOF({0x81a3e0, 0xc0000482e0})
C:/Users/runie/go/pkg/mod/github.com/markus-wa/demoinfocs-golang/v4@v4.0.0-beta.0/pkg/demoinfocs/parsing.go:171 +0x274
github.com/markus-wa/demoinfocs-golang/v4/pkg/demoinfocs.(*parser).ParseToEnd.func1()
C:/Users/runie/go/pkg/mod/github.com/markus-wa/demoinfocs-golang/v4@v4.0.0-beta.0/pkg/demoinfocs/parsing.go:124 +0x9f
panic({0x81a3e0?, 0xc0000482e0?})
C:/Program Files/Go/src/runtime/panic.go:920 +0x290
github.com/markus-wa/demoinfocs-golang/v4/pkg/demoinfocs.(*parser).parseFrameS2(0xc000069a00)
C:/Users/runie/go/pkg/mod/github.com/markus-wa/demoinfocs-golang/v4@v4.0.0-beta.0/pkg/demoinfocs/parsing.go:349 +0x905
github.com/markus-wa/demoinfocs-golang/v4/pkg/demoinfocs.(*parser).ParseToEnd(0xc000069a00)
C:/Users/runie/go/pkg/mod/github.com/markus-wa/demoinfocs-golang/v4@v4.0.0-beta.0/pkg/demoinfocs/parsing.go:147 +0x210
```
Not familiar enough with the library to tell if this is a bugged demo or not.
**To Reproduce**
Replay: http://replay183.valve.net/730/003636925493187444888_0599554593.dem.bz2
**Library version**
v4.0.0-beta.0
| True | CS2 corrupt input - **Describe the bug**
This demo causes a parser error:
```
panic: snappy: corrupt input [recovered]
panic: snappy: corrupt input
goroutine 1 [running]:
github.com/markus-wa/demoinfocs-golang/v4/pkg/demoinfocs.recoverFromUnexpectedEOF({0x81a3e0, 0xc0000482e0})
C:/Users/runie/go/pkg/mod/github.com/markus-wa/demoinfocs-golang/v4@v4.0.0-beta.0/pkg/demoinfocs/parsing.go:171 +0x274
github.com/markus-wa/demoinfocs-golang/v4/pkg/demoinfocs.(*parser).ParseToEnd.func1()
C:/Users/runie/go/pkg/mod/github.com/markus-wa/demoinfocs-golang/v4@v4.0.0-beta.0/pkg/demoinfocs/parsing.go:124 +0x9f
panic({0x81a3e0?, 0xc0000482e0?})
C:/Program Files/Go/src/runtime/panic.go:920 +0x290
github.com/markus-wa/demoinfocs-golang/v4/pkg/demoinfocs.(*parser).parseFrameS2(0xc000069a00)
C:/Users/runie/go/pkg/mod/github.com/markus-wa/demoinfocs-golang/v4@v4.0.0-beta.0/pkg/demoinfocs/parsing.go:349 +0x905
github.com/markus-wa/demoinfocs-golang/v4/pkg/demoinfocs.(*parser).ParseToEnd(0xc000069a00)
C:/Users/runie/go/pkg/mod/github.com/markus-wa/demoinfocs-golang/v4@v4.0.0-beta.0/pkg/demoinfocs/parsing.go:147 +0x210
```
Not familiar enough with the library to tell if this is a bugged demo or not.
**To Reproduce**
Replay: http://replay183.valve.net/730/003636925493187444888_0599554593.dem.bz2
**Library version**
v4.0.0-beta.0
| main | corrupt input describe the bug this demo causes a parser error panic snappy corrupt input panic snappy corrupt input goroutine github com markus wa demoinfocs golang pkg demoinfocs recoverfromunexpectedeof c users runie go pkg mod github com markus wa demoinfocs golang beta pkg demoinfocs parsing go github com markus wa demoinfocs golang pkg demoinfocs parser parsetoend c users runie go pkg mod github com markus wa demoinfocs golang beta pkg demoinfocs parsing go panic c program files go src runtime panic go github com markus wa demoinfocs golang pkg demoinfocs parser c users runie go pkg mod github com markus wa demoinfocs golang beta pkg demoinfocs parsing go github com markus wa demoinfocs golang pkg demoinfocs parser parsetoend c users runie go pkg mod github com markus wa demoinfocs golang beta pkg demoinfocs parsing go not familiar enough with the library to tell if this is a bugged demo or not to reproduce replay library version beta | 1 |
104,324 | 16,613,611,299 | IssuesEvent | 2021-06-02 14:18:10 | Thanraj/linux-4.1.15 | https://api.github.com/repos/Thanraj/linux-4.1.15 | opened | CVE-2020-8992 (Medium) detected in linux-stable-rtv4.1.33 | security vulnerability | ## CVE-2020-8992 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-4.1.15/commits/5e3fb3e332499e1ad10a0969e55582af1027b085">5e3fb3e332499e1ad10a0969e55582af1027b085</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/fs/ext4/block_validity.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/fs/ext4/block_validity.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ext4_protect_reserved_inode in fs/ext4/block_validity.c in the Linux kernel through 5.5.3 allows attackers to cause a denial of service (soft lockup) via a crafted journal size.
<p>Publish Date: 2020-02-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8992>CVE-2020-8992</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-8992">https://www.linuxkernelcves.com/cves/CVE-2020-8992</a></p>
<p>Release Date: 2020-07-22</p>
<p>Fix Resolution: v5.6-rc2,v5.4.21,v5.5.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2020-8992 (Medium) detected in linux-stable-rtv4.1.33 - ## CVE-2020-8992 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-stable-rtv4.1.33</b></p></summary>
<p>
<p>Julia Cartwright's fork of linux-stable-rt.git</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git>https://git.kernel.org/pub/scm/linux/kernel/git/julia/linux-stable-rt.git</a></p>
<p>Found in HEAD commit: <a href="https://api.github.com/repos/Thanraj/linux-4.1.15/commits/5e3fb3e332499e1ad10a0969e55582af1027b085">5e3fb3e332499e1ad10a0969e55582af1027b085</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/fs/ext4/block_validity.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>linux-4.1.15/fs/ext4/block_validity.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
ext4_protect_reserved_inode in fs/ext4/block_validity.c in the Linux kernel through 5.5.3 allows attackers to cause a denial of service (soft lockup) via a crafted journal size.
<p>Publish Date: 2020-02-14
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8992>CVE-2020-8992</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2020-8992">https://www.linuxkernelcves.com/cves/CVE-2020-8992</a></p>
<p>Release Date: 2020-07-22</p>
<p>Fix Resolution: v5.6-rc2,v5.4.21,v5.5.5</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in linux stable cve medium severity vulnerability vulnerable library linux stable julia cartwright s fork of linux stable rt git library home page a href found in head commit a href found in base branch master vulnerable source files linux fs block validity c linux fs block validity c vulnerability details protect reserved inode in fs block validity c in the linux kernel through allows attackers to cause a denial of service soft lockup via a crafted journal size publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
5,653 | 28,961,043,329 | IssuesEvent | 2023-05-10 02:31:19 | MozillaFoundation/foundation.mozilla.org | https://api.github.com/repos/MozillaFoundation/foundation.mozilla.org | opened | Investigate `test_library_page.py` | engineering backend maintain nice to have needs grooming | While working on #10488, the test `test_filter_multiple_author_profiles` was giving me some trouble whenever I tried to update its implementation from using a .get() request to instead using the `_get_research_detail_pages()` method directly, as all the other tests do.
What I found was that using `_get_research_detail_pages()` directly would return an empty queryset of no pages, even though using the test client and a .get() request would still technically send the same author ID's to the method in the end.
I also found that both using the method directly and using the test client would return the same Author Profiles using the same translation keys when using print statements [here](https://github.com/MozillaFoundation/foundation.mozilla.org/blob/96654723c6a3e82e209f14bfe84db651ec4d6d03/network-api/networkapi/wagtailpages/pagemodels/research_hub/library_page.py#L126-L135).
However, it seems that this may part of a bigger issue as if you rename this function, other tests begin to fail as well.
I think we should investigate this further in its own ticket to see what is going on here.
| True | Investigate `test_library_page.py` - While working on #10488, the test `test_filter_multiple_author_profiles` was giving me some trouble whenever I tried to update its implementation from using a .get() request to instead using the `_get_research_detail_pages()` method directly, as all the other tests do.
What I found was that using `_get_research_detail_pages()` directly would return an empty queryset of no pages, even though using the test client and a .get() request would still technically send the same author ID's to the method in the end.
I also found that both using the method directly and using the test client would return the same Author Profiles using the same translation keys when using print statements [here](https://github.com/MozillaFoundation/foundation.mozilla.org/blob/96654723c6a3e82e209f14bfe84db651ec4d6d03/network-api/networkapi/wagtailpages/pagemodels/research_hub/library_page.py#L126-L135).
However, it seems that this may part of a bigger issue as if you rename this function, other tests begin to fail as well.
I think we should investigate this further in its own ticket to see what is going on here.
| main | investigate test library page py while working on the test test filter multiple author profiles was giving me some trouble whenever i tried to update its implementation from using a get request to instead using the get research detail pages method directly as all the other tests do what i found was that using get research detail pages directly would return an empty queryset of no pages even though using the test client and a get request would still technically send the same author id s to the method in the end i also found that both using the method directly and using the test client would return the same author profiles using the same translation keys when using print statements however it seems that this may part of a bigger issue as if you rename this function other tests begin to fail as well i think we should investigate this further in its own ticket to see what is going on here | 1 |
2,132 | 7,302,364,406 | IssuesEvent | 2018-02-27 09:28:36 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | reopened | Test classes should contain tests | Area: analyzer Area: maintainability feature | Classes that are marked as tests should include tests. Otherwise it does not make sense to mark them as test classes. | True | Test classes should contain tests - Classes that are marked as tests should include tests. Otherwise it does not make sense to mark them as test classes. | main | test classes should contain tests classes that are marked as tests should include tests otherwise it does not make sense to mark them as test classes | 1 |
4,613 | 23,879,134,688 | IssuesEvent | 2022-09-07 22:28:52 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | "sam local start-api" should use HTTP/1.1 | contributors/good-first-issue type/feature area/local/start-api maintainer/need-followup | ### Describe your idea/feature/enhancement
It would be nice if `sam local start-api` responses used HTTP/1.1. I ran into this idea while trying to test my API locally with Postman. Requests keep failing there with the console message `Error: Parse Error: Expected HTTP/`, even though the local server output shows a successful response and `curl` requests also work correctly. After some googling, it appears that Postman cannot handle HTTP/1.0 responses. Frankly, I'm annoyed with Postman for not being able to support different HTTP versions (and [others](https://github.com/postmanlabs/postman-app-support/issues/6242) [have](https://stackoverflow.com/questions/20359149/how-can-you-set-the-http-protocol-version-in-postman) already opened tickets with Postman to resolve that), but I see no reason why SAM CLI should use such an ancient version of the protocol either. HTTP/1.1 or /2 would be preferrable (though I think /2 would also break Postman 😬). I believe this is happening because `sam local start-api` uses Flask internally (v1.1.2, at time of this writing), which apparently defaults to using HTTP/1.0
### Proposal
The `sam local start-api` command handler should configure its internal Flask instance to use HTTP/1.1. Based on [this SO post](https://stackoverflow.com/questions/59027260/make-flask-return-response-header-http1-1-instead-http1-0) and a quick search through SAM CLI's source, I think you would just need to add the following line to the [`LocalApigwService.create`](https://github.com/aws/aws-sam-cli/blob/2b2a921172a22f43ef8580eb8ad0be1558e9cb7d/samcli/local/apigw/local_apigw_service.py#L154) method:
```python
WSGIRequestHandler.protocol_version = "HTTP/1.1"
```
I would have opened a PR for this, but I'm not very familiar with Python.
This would not require any updates to the SAM Spec, though it wouldn't hurt to add a warning to the docs about what HTTP version `sam local start-api` uses. | True | "sam local start-api" should use HTTP/1.1 - ### Describe your idea/feature/enhancement
It would be nice if `sam local start-api` responses used HTTP/1.1. I ran into this idea while trying to test my API locally with Postman. Requests keep failing there with the console message `Error: Parse Error: Expected HTTP/`, even though the local server output shows a successful response and `curl` requests also work correctly. After some googling, it appears that Postman cannot handle HTTP/1.0 responses. Frankly, I'm annoyed with Postman for not being able to support different HTTP versions (and [others](https://github.com/postmanlabs/postman-app-support/issues/6242) [have](https://stackoverflow.com/questions/20359149/how-can-you-set-the-http-protocol-version-in-postman) already opened tickets with Postman to resolve that), but I see no reason why SAM CLI should use such an ancient version of the protocol either. HTTP/1.1 or /2 would be preferrable (though I think /2 would also break Postman 😬). I believe this is happening because `sam local start-api` uses Flask internally (v1.1.2, at time of this writing), which apparently defaults to using HTTP/1.0
### Proposal
The `sam local start-api` command handler should configure its internal Flask instance to use HTTP/1.1. Based on [this SO post](https://stackoverflow.com/questions/59027260/make-flask-return-response-header-http1-1-instead-http1-0) and a quick search through SAM CLI's source, I think you would just need to add the following line to the [`LocalApigwService.create`](https://github.com/aws/aws-sam-cli/blob/2b2a921172a22f43ef8580eb8ad0be1558e9cb7d/samcli/local/apigw/local_apigw_service.py#L154) method:
```python
WSGIRequestHandler.protocol_version = "HTTP/1.1"
```
I would have opened a PR for this, but I'm not very familiar with Python.
This would not require any updates to the SAM Spec, though it wouldn't hurt to add a warning to the docs about what HTTP version `sam local start-api` uses. | main | sam local start api should use http describe your idea feature enhancement it would be nice if sam local start api responses used http i ran into this idea while trying to test my api locally with postman requests keep failing there with the console message error parse error expected http even though the local server output shows a successful response and curl requests also work correctly after some googling it appears that postman cannot handle http responses frankly i m annoyed with postman for not being able to support different http versions and already opened tickets with postman to resolve that but i see no reason why sam cli should use such an ancient version of the protocol either http or would be preferrable though i think would also break postman 😬 i believe this is happening because sam local start api uses flask internally at time of this writing which apparently defaults to using http proposal the sam local start api command handler should configure its internal flask instance to use http based on and a quick search through sam cli s source i think you would just need to add the following line to the method python wsgirequesthandler protocol version http i would have opened a pr for this but i m not very familiar with python this would not require any updates to the sam spec though it wouldn t hurt to add a warning to the docs about what http version sam local start api uses | 1 |
877 | 4,541,075,367 | IssuesEvent | 2016-09-09 16:34:13 | coniks-sys/coniks-go | https://api.github.com/repos/coniks-sys/coniks-go | closed | Use script to gather all package test coverage profiles | maintainability | Instead of having to add each new package and subpackage to our .travis.yml file manually to collect the test coverage profiles, we should use a script like https://github.com/dedis/cosi/blob/master/coveralls.sh to handle this automatically. | True | Use script to gather all package test coverage profiles - Instead of having to add each new package and subpackage to our .travis.yml file manually to collect the test coverage profiles, we should use a script like https://github.com/dedis/cosi/blob/master/coveralls.sh to handle this automatically. | main | use script to gather all package test coverage profiles instead of having to add each new package and subpackage to our travis yml file manually to collect the test coverage profiles we should use a script like to handle this automatically | 1 |
276,908 | 8,614,579,025 | IssuesEvent | 2018-11-19 17:53:36 | Qiskit/qiskit-terra | https://api.github.com/repos/Qiskit/qiskit-terra | closed | No method to override use of interactive plot visualizations | priority: medium | <!-- ⚠️ If you do not respect this template, your issue will be closed -->
<!-- ⚠️ Make sure to browse the opened and closed issues to confirm this idea does not exist. -->
### What is the expected enhancement?
When running visualizations that have an interactive mode if you're system meets the requirements we unconditionally use the interactive visualizations. We need to provide an option to let users override this and use the non-interactive visualizations.
| 1.0 | No method to override use of interactive plot visualizations - <!-- ⚠️ If you do not respect this template, your issue will be closed -->
<!-- ⚠️ Make sure to browse the opened and closed issues to confirm this idea does not exist. -->
### What is the expected enhancement?
When running visualizations that have an interactive mode if you're system meets the requirements we unconditionally use the interactive visualizations. We need to provide an option to let users override this and use the non-interactive visualizations.
| non_main | no method to override use of interactive plot visualizations what is the expected enhancement when running visualizations that have an interactive mode if you re system meets the requirements we unconditionally use the interactive visualizations we need to provide an option to let users override this and use the non interactive visualizations | 0 |
640,442 | 20,789,895,821 | IssuesEvent | 2022-03-17 00:04:37 | JeffreyCHChan/SOEN390 | https://api.github.com/repos/JeffreyCHChan/SOEN390 | closed | As a patient, I want to be able to get an 8-digit ID so that I can better share information with my doctor | Priority 4 Back End USP 1 | - [ ] Store 8-digit ID in patients database. | 1.0 | As a patient, I want to be able to get an 8-digit ID so that I can better share information with my doctor - - [ ] Store 8-digit ID in patients database. | non_main | as a patient i want to be able to get an digit id so that i can better share information with my doctor store digit id in patients database | 0 |
117,055 | 11,945,666,626 | IssuesEvent | 2020-04-03 06:25:53 | grrrrnt/ped | https://api.github.com/repos/grrrrnt/ped | opened | 3.1. preface missing | severity.Low type.DocumentationBug | Will be better to have a preface for 3.1. Viewing help
e.g. This command will allow you to view a list of help options.
| 1.0 | 3.1. preface missing - Will be better to have a preface for 3.1. Viewing help
e.g. This command will allow you to view a list of help options.
| non_main | preface missing will be better to have a preface for viewing help e g this command will allow you to view a list of help options | 0 |
3,153 | 12,179,471,794 | IssuesEvent | 2020-04-28 10:44:29 | ipfs/js-ipfs | https://api.github.com/repos/ipfs/js-ipfs | opened | Run E2E tests against version at API port's /webui | need/analysis need/maintainers-input need/triage | > Needs #2984 to land first
Thanks to https://github.com/ipfs/js-ipfs/pull/2706 we have E2E tests that check if js-ipfs is compatible with ipfs-webui's `master` branch, but we don't run those tests against the webui version aliased at `/webui` on API port.
Over time, support for older version at `/webui` may break, while the `master` of webui continues to work with js-ipfs `master`, which means we miss some regressions.
A fix here would be to run ipfs-webui E2E test suite against version at `/webui` at well:
- extract CID from `packages/ipfs/src/http/api/routes/webui.js`
- prefetch it and cache so CI is not slowed down
- make it a separate task in CI with distinct name (eg. ipfs-webui(/webui))
- no need to run it twice
| True | Run E2E tests against version at API port's /webui - > Needs #2984 to land first
Thanks to https://github.com/ipfs/js-ipfs/pull/2706 we have E2E tests that check if js-ipfs is compatible with ipfs-webui's `master` branch, but we don't run those tests against the webui version aliased at `/webui` on API port.
Over time, support for older version at `/webui` may break, while the `master` of webui continues to work with js-ipfs `master`, which means we miss some regressions.
A fix here would be to run ipfs-webui E2E test suite against version at `/webui` at well:
- extract CID from `packages/ipfs/src/http/api/routes/webui.js`
- prefetch it and cache so CI is not slowed down
- make it a separate task in CI with distinct name (eg. ipfs-webui(/webui))
- no need to run it twice
| main | run tests against version at api port s webui needs to land first thanks to we have tests that check if js ipfs is compatible with ipfs webui s master branch but we don t run those tests against the webui version aliased at webui on api port over time support for older version at webui may break while the master of webui continues to work with js ipfs master which means we miss some regressions a fix here would be to run ipfs webui test suite against version at webui at well extract cid from packages ipfs src http api routes webui js prefetch it and cache so ci is not slowed down make it a separate task in ci with distinct name eg ipfs webui webui no need to run it twice | 1 |
41 | 2,588,002,698 | IssuesEvent | 2015-02-17 22:00:24 | jenkinsci/slack-plugin | https://api.github.com/repos/jenkinsci/slack-plugin | closed | Having trouble performing plugin release | maintainer communication | I'm currently having issues releasing the plugin. Here's an excerpt from slack...
> 21:01 < sag47> kohsuke: I'm having trouble releasing a plugin. My GitHub username is samrocketman and my jenkins-ci.org ID is sag47. I have added my credentials to ~/.m2/settings.xml from this article
https://wiki.jenkins-ci.org/display/JENKINS/Hosting+Plugins
21:02 < sag47> kohsuke: Do I have the capability to release if I'm able to be a contributor for a plugin? If not, what do I need to do in order to be able to release?
21:02 < sag47> kohsuke: I have taken over maintaining this as it was previously abandoned.
21:03 < sag47> kohsuke: Plugin in question: https://github.com/jenkinsci/slack-plugin
I'll be working through this and keep any followers for the next release updated via this issue. | True | Having trouble performing plugin release - I'm currently having issues releasing the plugin. Here's an excerpt from slack...
> 21:01 < sag47> kohsuke: I'm having trouble releasing a plugin. My GitHub username is samrocketman and my jenkins-ci.org ID is sag47. I have added my credentials to ~/.m2/settings.xml from this article
https://wiki.jenkins-ci.org/display/JENKINS/Hosting+Plugins
21:02 < sag47> kohsuke: Do I have the capability to release if I'm able to be a contributor for a plugin? If not, what do I need to do in order to be able to release?
21:02 < sag47> kohsuke: I have taken over maintaining this as it was previously abandoned.
21:03 < sag47> kohsuke: Plugin in question: https://github.com/jenkinsci/slack-plugin
I'll be working through this and keep any followers for the next release updated via this issue. | main | having trouble performing plugin release i m currently having issues releasing the plugin here s an excerpt from slack kohsuke i m having trouble releasing a plugin my github username is samrocketman and my jenkins ci org id is i have added my credentials to settings xml from this article kohsuke do i have the capability to release if i m able to be a contributor for a plugin if not what do i need to do in order to be able to release kohsuke i have taken over maintaining this as it was previously abandoned kohsuke plugin in question i ll be working through this and keep any followers for the next release updated via this issue | 1 |
550,507 | 16,114,391,082 | IssuesEvent | 2021-04-28 04:38:26 | calyco-yale/calyco | https://api.github.com/repos/calyco-yale/calyco | closed | Add Push Notifications | priority: high | Notify users whenever an invite or friend request is sent to them, and (by default) 10 minutes prior to the event. | 1.0 | Add Push Notifications - Notify users whenever an invite or friend request is sent to them, and (by default) 10 minutes prior to the event. | non_main | add push notifications notify users whenever an invite or friend request is sent to them and by default minutes prior to the event | 0 |
329,517 | 28,264,540,636 | IssuesEvent | 2023-04-07 05:01:15 | EunChanNam/We-Share-Wish-Hair | https://api.github.com/repos/EunChanNam/We-Share-Wish-Hair | closed | MyPage 서비스 테스트 및 Point 서비스 추가테스트 | test | ## 🤷 구현할 기능
* 리뷰 테스트를 마친 후 MyPageService 테스트
* PointSearchService - 정렬 테스팅 추가
## 📄 참고 사항
| 1.0 | MyPage 서비스 테스트 및 Point 서비스 추가테스트 - ## 🤷 구현할 기능
* 리뷰 테스트를 마친 후 MyPageService 테스트
* PointSearchService - 정렬 테스팅 추가
## 📄 참고 사항
| non_main | mypage 서비스 테스트 및 point 서비스 추가테스트 🤷 구현할 기능 리뷰 테스트를 마친 후 mypageservice 테스트 pointsearchservice 정렬 테스팅 추가 📄 참고 사항 | 0 |
343,031 | 10,324,603,565 | IssuesEvent | 2019-09-01 10:40:43 | weaveworks/ignite | https://api.github.com/repos/weaveworks/ignite | closed | Implement "docker attach" using custom Go code instead of execing | help wanted kind/enhancement priority/backlog | This is a follow-up from https://github.com/weaveworks/ignite/issues/67, that was left as a non-goal for v0.5.0. This should be completely doable, and we have to do the same for containerd anyways (#209). So we'll target this at the same time as containerd support, v0.6.0
This would be a really nice community contribution. | 1.0 | Implement "docker attach" using custom Go code instead of execing - This is a follow-up from https://github.com/weaveworks/ignite/issues/67, that was left as a non-goal for v0.5.0. This should be completely doable, and we have to do the same for containerd anyways (#209). So we'll target this at the same time as containerd support, v0.6.0
This would be a really nice community contribution. | non_main | implement docker attach using custom go code instead of execing this is a follow up from that was left as a non goal for this should be completely doable and we have to do the same for containerd anyways so we ll target this at the same time as containerd support this would be a really nice community contribution | 0 |
4,259 | 21,259,081,395 | IssuesEvent | 2022-04-13 00:43:57 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | Lambda SAM build (ARM64 archi) - Python libs fails | blocked/more-info-needed stage/bug-repro maintainer/need-followup platform/linux/arm | I am trying to build a **Lambda** function using **sam** (1.36.0) within **CodeBuild**, the Lambda has **arm64** architecture as well as the CodeBuild configuration (**aws/codebuild/amazonlinux2-aarch64-standard:2.0**).
Using Python 3.8.10
I have in my _requirements.txt_ the dependencies below:
```
PyYAML
pandas
s3fs
xlsxwriter
```
CodeBuild throws error:
```
Build Failed
Error: PythonPipBuilder:ResolveDependencies - {wrapt==1.13.3(sdist)}
```
My build details:
```
version: 0.2
phases:
install:
runtime-versions:
python: 3.8
commands:
- which python
- python --version
- pip install aws-sam-cli --upgrade
- sam build --template $(pwd)/template.yaml
- aws cloudformation package --template $(pwd)/.aws-sam/build/template.yaml --s3-bucket XXXXX --s3-prefix PREFIX_PATH_DUMMY --output-template-file product.template-eu-west-1.yaml
- sam deploy --template-file ./product.template-eu-west-1.yaml --stack-name DUMMY --capabilities CAPABILITY_IAM
```
I also manually tested this on **t4g Graviton instance**, doing pip install requirements.txt, there it works. But when using **CodeBuild it fails**. | True | Lambda SAM build (ARM64 archi) - Python libs fails - I am trying to build a **Lambda** function using **sam** (1.36.0) within **CodeBuild**, the Lambda has **arm64** architecture as well as the CodeBuild configuration (**aws/codebuild/amazonlinux2-aarch64-standard:2.0**).
Using Python 3.8.10
I have in my _requirements.txt_ the dependencies below:
```
PyYAML
pandas
s3fs
xlsxwriter
```
CodeBuild throws error:
```
Build Failed
Error: PythonPipBuilder:ResolveDependencies - {wrapt==1.13.3(sdist)}
```
My build details:
```
version: 0.2
phases:
install:
runtime-versions:
python: 3.8
commands:
- which python
- python --version
- pip install aws-sam-cli --upgrade
- sam build --template $(pwd)/template.yaml
- aws cloudformation package --template $(pwd)/.aws-sam/build/template.yaml --s3-bucket XXXXX --s3-prefix PREFIX_PATH_DUMMY --output-template-file product.template-eu-west-1.yaml
- sam deploy --template-file ./product.template-eu-west-1.yaml --stack-name DUMMY --capabilities CAPABILITY_IAM
```
I also manually tested this on **t4g Graviton instance**, doing pip install requirements.txt, there it works. But when using **CodeBuild it fails**. | main | lambda sam build archi python libs fails i am trying to build a lambda function using sam within codebuild the lambda has architecture as well as the codebuild configuration aws codebuild standard using python i have in my requirements txt the dependencies below pyyaml pandas xlsxwriter codebuild throws error build failed error pythonpipbuilder resolvedependencies wrapt sdist my build details version phases install runtime versions python commands which python python version pip install aws sam cli upgrade sam build template pwd template yaml aws cloudformation package template pwd aws sam build template yaml bucket xxxxx prefix prefix path dummy output template file product template eu west yaml sam deploy template file product template eu west yaml stack name dummy capabilities capability iam i also manually tested this on graviton instance doing pip install requirements txt there it works but when using codebuild it fails | 1 |
1,559 | 6,572,254,405 | IssuesEvent | 2017-09-11 00:39:41 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | unable to set mysql root password the first time | affects_1.9 bug_report waiting_on_maintainer | Issue Type:
Bug Report
Component Name:
mysql_user
Ansible Version:
```
ansible 1.9.4 (stable-1.9 2d914d4b1e) last updated 2015/11/16 21:35:26 (GMT +100)
lib/ansible/modules/core: (stable-1.9 4b65a4a8b5) last updated 2015/11/16 21:35:36 (GMT +100)
lib/ansible/modules/extras: (stable-1.9 29c3e31a92) last updated 2015/11/16 21:35:50 (GMT +100)
configured module search path = None
```
Ansible Configuration:
```
[defaults]
inventory = ./hosts
host_key_checking=False
forks = 40
display_skipped_hosts=False
retry_files_save_path = ~/.ansible-retry
log_path=/var/log/ansible.log
gathering = smart
# pipelining=True
```
Environment:
```
lsb_release -a
No LSB modules are available.
Distributor ID: Ubuntu
Description: Ubuntu 14.04.3 LTS
Release: 14.04
Codename: trusty
```
Summary:
Im unable to set the root password for mysql when deploying it with ansible.
Steps To Reproduce:
1. run [this](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-mysq-yml) playbook once and get [this](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-first_run-txt) result.
2. ssh to the server and try mysql without any luck. ([log](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-first_sql_attempt-txt))
3. re-run playbook [log](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-second_run-txt)
4. Run sql with success ([log](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-second_sql-attempt-txt))
Expected Results:
Be able to start a mysql root shell at step 2 (and not at step 4).
Actual Results:
Mysql password was not set until i applied the same playbook twice.
| True | unable to set mysql root password the first time - Issue Type:
Bug Report
Component Name:
mysql_user
Ansible Version:
```
ansible 1.9.4 (stable-1.9 2d914d4b1e) last updated 2015/11/16 21:35:26 (GMT +100)
lib/ansible/modules/core: (stable-1.9 4b65a4a8b5) last updated 2015/11/16 21:35:36 (GMT +100)
lib/ansible/modules/extras: (stable-1.9 29c3e31a92) last updated 2015/11/16 21:35:50 (GMT +100)
configured module search path = None
```
Ansible Configuration:
```
[defaults]
inventory = ./hosts
host_key_checking=False
forks = 40
display_skipped_hosts=False
retry_files_save_path = ~/.ansible-retry
log_path=/var/log/ansible.log
gathering = smart
# pipelining=True
```
Environment:
```
lsb_release -a
No LSB modules are available.
Distributor ID: Ubuntu
Description: Ubuntu 14.04.3 LTS
Release: 14.04
Codename: trusty
```
Summary:
Im unable to set the root password for mysql when deploying it with ansible.
Steps To Reproduce:
1. run [this](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-mysq-yml) playbook once and get [this](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-first_run-txt) result.
2. ssh to the server and try mysql without any luck. ([log](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-first_sql_attempt-txt))
3. re-run playbook [log](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-second_run-txt)
4. Run sql with success ([log](https://gist.github.com/serialdoom/427b246ab52b229da52b#file-second_sql-attempt-txt))
Expected Results:
Be able to start a mysql root shell at step 2 (and not at step 4).
Actual Results:
Mysql password was not set until i applied the same playbook twice.
| main | unable to set mysql root password the first time issue type bug report component name mysql user ansible version ansible stable last updated gmt lib ansible modules core stable last updated gmt lib ansible modules extras stable last updated gmt configured module search path none ansible configuration inventory hosts host key checking false forks display skipped hosts false retry files save path ansible retry log path var log ansible log gathering smart pipelining true environment lsb release a no lsb modules are available distributor id ubuntu description ubuntu lts release codename trusty summary im unable to set the root password for mysql when deploying it with ansible steps to reproduce run playbook once and get result ssh to the server and try mysql without any luck re run playbook run sql with success expected results be able to start a mysql root shell at step and not at step actual results mysql password was not set until i applied the same playbook twice | 1 |
366 | 3,355,023,564 | IssuesEvent | 2015-11-18 14:57:20 | Homebrew/homebrew | https://api.github.com/repos/Homebrew/homebrew | closed | brew install --HEAD ghostscript fails | maintainer feedback | command fails with an error:
```
make: *** No rule to make target `src/png.h', needed by `obj/xpspng.o'. Stop.
make: *** Waiting for unfinished jobs....
```
I suppose this is a mixup in using system vs local libpng (autogen.sh finds system wide libpng):
```
checking for local zlib source... yes
checking for local png library source... no
checking for png_create_write_struct in -lpng... yes
checking png.h usability... yes
checking png.h presence... yes
```
Logs:
/.../Library/Logs/Homebrew/ghostscript/01.autogen.sh
/.../Library/Logs/Homebrew/ghostscript/01.autogen.sh.cc
/.../Library/Logs/Homebrew/ghostscript/02.make
/.../Library/Logs/Homebrew/ghostscript/02.make.cc
/.../Library/Logs/Homebrew/ghostscript/config.log
uploaded to: https://gist.github.com/a0cfb6f3ab1bd4e3429b
| True | brew install --HEAD ghostscript fails - command fails with an error:
```
make: *** No rule to make target `src/png.h', needed by `obj/xpspng.o'. Stop.
make: *** Waiting for unfinished jobs....
```
I suppose this is a mixup in using system vs local libpng (autogen.sh finds system wide libpng):
```
checking for local zlib source... yes
checking for local png library source... no
checking for png_create_write_struct in -lpng... yes
checking png.h usability... yes
checking png.h presence... yes
```
Logs:
/.../Library/Logs/Homebrew/ghostscript/01.autogen.sh
/.../Library/Logs/Homebrew/ghostscript/01.autogen.sh.cc
/.../Library/Logs/Homebrew/ghostscript/02.make
/.../Library/Logs/Homebrew/ghostscript/02.make.cc
/.../Library/Logs/Homebrew/ghostscript/config.log
uploaded to: https://gist.github.com/a0cfb6f3ab1bd4e3429b
| main | brew install head ghostscript fails command fails with an error make no rule to make target src png h needed by obj xpspng o stop make waiting for unfinished jobs i suppose this is a mixup in using system vs local libpng autogen sh finds system wide libpng checking for local zlib source yes checking for local png library source no checking for png create write struct in lpng yes checking png h usability yes checking png h presence yes logs library logs homebrew ghostscript autogen sh library logs homebrew ghostscript autogen sh cc library logs homebrew ghostscript make library logs homebrew ghostscript make cc library logs homebrew ghostscript config log uploaded to | 1 |
259,446 | 22,475,117,487 | IssuesEvent | 2022-06-22 11:36:22 | FOLIO-FSE/folio_migration_tools | https://api.github.com/repos/FOLIO-FSE/folio_migration_tools | closed | Add test files for all migration tasks | improve_test_coverage | In an effort to up the coverage, I am working on creating tests and setup methods that allows everyone to write tests for exactly what they need to test. | 1.0 | Add test files for all migration tasks - In an effort to up the coverage, I am working on creating tests and setup methods that allows everyone to write tests for exactly what they need to test. | non_main | add test files for all migration tasks in an effort to up the coverage i am working on creating tests and setup methods that allows everyone to write tests for exactly what they need to test | 0 |
196,696 | 22,498,602,407 | IssuesEvent | 2022-06-23 09:44:43 | jonazbot/IPington | https://api.github.com/repos/jonazbot/IPington | closed | Allow bot to share source-code on demand | enhancement security | Open source-code should be easily and readily available to end-users who want it.
### Options:
- **Post link to repo**
- Post source to channel
- Upload and post source-files | True | Allow bot to share source-code on demand - Open source-code should be easily and readily available to end-users who want it.
### Options:
- **Post link to repo**
- Post source to channel
- Upload and post source-files | non_main | allow bot to share source code on demand open source code should be easily and readily available to end users who want it options post link to repo post source to channel upload and post source files | 0 |
5,512 | 27,558,554,087 | IssuesEvent | 2023-03-07 19:57:59 | cosmos/ibc-rs | https://api.github.com/repos/cosmos/ibc-rs | closed | Investigate moving `verify_delay_passed` to a more appropriate section | A: breaking I: logic O: maintainability | <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Please also ensure that this is not a duplicate issue :)
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
Context: https://github.com/cosmos/ibc-rs/pull/393#discussion_r1097850720
The implemented `verify_delay_passed` method inside the `07_tendermint` doesn't seem a client-specific process and it should be placed somewhere in the "connection" part of the stack. The issue arose because we wanted to stop passing the `ctx` field into some of the verification methods like `verify_packet_acknowledgement`
| True | Investigate moving `verify_delay_passed` to a more appropriate section - <!-- < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < < ☺
v ✰ Thanks for opening an issue! ✰
v Before smashing the submit button please review the template.
v Please also ensure that this is not a duplicate issue :)
☺ > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > -->
## Summary
Context: https://github.com/cosmos/ibc-rs/pull/393#discussion_r1097850720
The implemented `verify_delay_passed` method inside the `07_tendermint` doesn't seem a client-specific process and it should be placed somewhere in the "connection" part of the stack. The issue arose because we wanted to stop passing the `ctx` field into some of the verification methods like `verify_packet_acknowledgement`
| main | investigate moving verify delay passed to a more appropriate section ☺ v ✰ thanks for opening an issue ✰ v before smashing the submit button please review the template v please also ensure that this is not a duplicate issue ☺ summary context the implemented verify delay passed method inside the tendermint doesn t seem a client specific process and it should be placed somewhere in the connection part of the stack the issue arose because we wanted to stop passing the ctx field into some of the verification methods like verify packet acknowledgement | 1 |
95,971 | 16,112,999,992 | IssuesEvent | 2021-04-28 01:20:29 | RG4421/developers | https://api.github.com/repos/RG4421/developers | opened | CVE-2021-23382 (Medium) detected in multiple libraries | security vulnerability | ## CVE-2021-23382 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-5.2.18.tgz</b>, <b>postcss-7.0.35.tgz</b>, <b>postcss-6.0.23.tgz</b></p></summary>
<p>
<details><summary><b>postcss-5.2.18.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz">https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz</a></p>
<p>Path to dependency file: developers/package.json</p>
<p>Path to vulnerable library: developers/node_modules/perfectionist/node_modules/postcss/package.json,developers/node_modules/postcss-scss/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- tailwindcss-0.7.4.tgz (Root Library)
- perfectionist-2.4.0.tgz
- postcss-scss-0.3.1.tgz
- :x: **postcss-5.2.18.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.35.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz</a></p>
<p>Path to dependency file: developers/package.json</p>
<p>Path to vulnerable library: developers/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- tailwindcss-0.7.4.tgz (Root Library)
- :x: **postcss-7.0.35.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-6.0.23.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz">https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz</a></p>
<p>Path to dependency file: developers/package.json</p>
<p>Path to vulnerable library: developers/node_modules/postcss-modules-values/node_modules/postcss/package.json,developers/node_modules/css-loader/node_modules/postcss/package.json,developers/node_modules/postcss-functions/node_modules/postcss/package.json,developers/node_modules/postcss-easy-import/node_modules/postcss/package.json,developers/node_modules/postcss-modules-local-by-default/node_modules/postcss/package.json,developers/node_modules/postcss-modules-extract-imports/node_modules/postcss/package.json,developers/node_modules/postcss-modules-scope/node_modules/postcss/package.json,developers/node_modules/icss-utils/node_modules/postcss/package.json,developers/node_modules/postcss-import/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- postcss-easy-import-3.0.0.tgz (Root Library)
- postcss-import-10.0.0.tgz
- :x: **postcss-6.0.23.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>development</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution: postcss - 8.2.13</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"5.2.18","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"tailwindcss:0.7.4;perfectionist:2.4.0;postcss-scss:0.3.1;postcss:5.2.18","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"7.0.35","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"tailwindcss:0.7.4;postcss:7.0.35","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"6.0.23","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"postcss-easy-import:3.0.0;postcss-import:10.0.0;postcss:6.0.23","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"}],"baseBranches":["development"],"vulnerabilityIdentifier":"CVE-2021-23382","vulnerabilityDetails":"The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \\/\\*\\s* sourceMappingURL\u003d(.*).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | True | CVE-2021-23382 (Medium) detected in multiple libraries - ## CVE-2021-23382 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>postcss-5.2.18.tgz</b>, <b>postcss-7.0.35.tgz</b>, <b>postcss-6.0.23.tgz</b></p></summary>
<p>
<details><summary><b>postcss-5.2.18.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz">https://registry.npmjs.org/postcss/-/postcss-5.2.18.tgz</a></p>
<p>Path to dependency file: developers/package.json</p>
<p>Path to vulnerable library: developers/node_modules/perfectionist/node_modules/postcss/package.json,developers/node_modules/postcss-scss/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- tailwindcss-0.7.4.tgz (Root Library)
- perfectionist-2.4.0.tgz
- postcss-scss-0.3.1.tgz
- :x: **postcss-5.2.18.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-7.0.35.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz">https://registry.npmjs.org/postcss/-/postcss-7.0.35.tgz</a></p>
<p>Path to dependency file: developers/package.json</p>
<p>Path to vulnerable library: developers/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- tailwindcss-0.7.4.tgz (Root Library)
- :x: **postcss-7.0.35.tgz** (Vulnerable Library)
</details>
<details><summary><b>postcss-6.0.23.tgz</b></p></summary>
<p>Tool for transforming styles with JS plugins</p>
<p>Library home page: <a href="https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz">https://registry.npmjs.org/postcss/-/postcss-6.0.23.tgz</a></p>
<p>Path to dependency file: developers/package.json</p>
<p>Path to vulnerable library: developers/node_modules/postcss-modules-values/node_modules/postcss/package.json,developers/node_modules/css-loader/node_modules/postcss/package.json,developers/node_modules/postcss-functions/node_modules/postcss/package.json,developers/node_modules/postcss-easy-import/node_modules/postcss/package.json,developers/node_modules/postcss-modules-local-by-default/node_modules/postcss/package.json,developers/node_modules/postcss-modules-extract-imports/node_modules/postcss/package.json,developers/node_modules/postcss-modules-scope/node_modules/postcss/package.json,developers/node_modules/icss-utils/node_modules/postcss/package.json,developers/node_modules/postcss-import/node_modules/postcss/package.json</p>
<p>
Dependency Hierarchy:
- postcss-easy-import-3.0.0.tgz (Root Library)
- postcss-import-10.0.0.tgz
- :x: **postcss-6.0.23.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>development</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \/\*\s* sourceMappingURL=(.*).
<p>Publish Date: 2021-04-26
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382>CVE-2021-23382</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: Low
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-23382</a></p>
<p>Release Date: 2021-04-26</p>
<p>Fix Resolution: postcss - 8.2.13</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"5.2.18","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"tailwindcss:0.7.4;perfectionist:2.4.0;postcss-scss:0.3.1;postcss:5.2.18","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"7.0.35","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"tailwindcss:0.7.4;postcss:7.0.35","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"},{"packageType":"javascript/Node.js","packageName":"postcss","packageVersion":"6.0.23","packageFilePaths":["/package.json"],"isTransitiveDependency":true,"dependencyTree":"postcss-easy-import:3.0.0;postcss-import:10.0.0;postcss:6.0.23","isMinimumFixVersionAvailable":true,"minimumFixVersion":"postcss - 8.2.13"}],"baseBranches":["development"],"vulnerabilityIdentifier":"CVE-2021-23382","vulnerabilityDetails":"The package postcss before 8.2.13 are vulnerable to Regular Expression Denial of Service (ReDoS) via getAnnotationURL() and loadAnnotation() in lib/previous-map.js. The vulnerable regexes are caused mainly by the sub-pattern \\/\\*\\s* sourceMappingURL\u003d(.*).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23382","cvss3Severity":"medium","cvss3Score":"5.3","cvss3Metrics":{"A":"Low","AC":"Low","PR":"None","S":"Unchanged","C":"None","UI":"None","AV":"Network","I":"None"},"extraData":{}}</REMEDIATE> --> | non_main | cve medium detected in multiple libraries cve medium severity vulnerability vulnerable libraries postcss tgz postcss tgz postcss tgz postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file developers package json path to vulnerable library developers node modules perfectionist node modules postcss package json developers node modules postcss scss node modules postcss package json dependency hierarchy tailwindcss tgz root library perfectionist tgz postcss scss tgz x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file developers package json path to vulnerable library developers node modules postcss package json dependency hierarchy tailwindcss tgz root library x postcss tgz vulnerable library postcss tgz tool for transforming styles with js plugins library home page a href path to dependency file developers package json path to vulnerable library developers node modules postcss modules values node modules postcss package json developers node modules css loader node modules postcss package json developers node modules postcss functions node modules postcss package json developers node modules postcss easy import node modules postcss package json developers node modules postcss modules local by default node modules postcss package json developers node modules postcss modules extract imports node modules postcss package json developers node modules postcss modules scope node modules postcss package json developers node modules icss utils node modules postcss package json developers node modules postcss import node modules postcss package json dependency hierarchy postcss easy import tgz root library postcss import tgz x postcss tgz vulnerable library found in base branch development vulnerability details the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution postcss isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree tailwindcss perfectionist postcss scss postcss isminimumfixversionavailable true minimumfixversion postcss packagetype javascript node js packagename postcss packageversion packagefilepaths istransitivedependency true dependencytree tailwindcss postcss isminimumfixversionavailable true minimumfixversion postcss packagetype javascript node js packagename postcss packageversion packagefilepaths istransitivedependency true dependencytree postcss easy import postcss import postcss isminimumfixversionavailable true minimumfixversion postcss basebranches vulnerabilityidentifier cve vulnerabilitydetails the package postcss before are vulnerable to regular expression denial of service redos via getannotationurl and loadannotation in lib previous map js the vulnerable regexes are caused mainly by the sub pattern s sourcemappingurl vulnerabilityurl | 0 |
3,117 | 11,905,232,606 | IssuesEvent | 2020-03-30 18:12:26 | diofant/diofant | https://api.github.com/repos/diofant/diofant | opened | Cleanup polynomial initialization functions from polyutils, optimize like poly() | help wanted maintainability performance polys | E.g. _parallel_dict_from_expr() could avoid expand(). Also, probably there should be only "parallel" versions, i.e. parallel_dict_from_expr().
This address the following [TODO](https://github.com/diofant/diofant/blob/3d08f9ab8cd77359f97411382ad754b5dc09b96e/diofant/polys/rings.py#L115).
| True | Cleanup polynomial initialization functions from polyutils, optimize like poly() - E.g. _parallel_dict_from_expr() could avoid expand(). Also, probably there should be only "parallel" versions, i.e. parallel_dict_from_expr().
This address the following [TODO](https://github.com/diofant/diofant/blob/3d08f9ab8cd77359f97411382ad754b5dc09b96e/diofant/polys/rings.py#L115).
| main | cleanup polynomial initialization functions from polyutils optimize like poly e g parallel dict from expr could avoid expand also probably there should be only parallel versions i e parallel dict from expr this address the following | 1 |
295,797 | 25,506,638,792 | IssuesEvent | 2022-11-28 09:57:16 | keycloak/keycloak | https://api.github.com/repos/keycloak/keycloak | closed | Unable to run map-storage-jpa tests with custom Postgres image | area/testsuite kind/bug area/storage | ### Area
testsuite
### Describe the bug
According to the [documentation](https://github.com/keycloak/keycloak/blob/main/testsuite/integration-arquillian/HOW-TO-RUN.md#running-tests-with-jpa-map-storage), it is possible to run JPA Map storage tests using a custom Postgres image defined in system property `keycloak.map.storage.postgres.docker.image`.
### Version
19.0.0
### Expected behavior
The custom Postgres image is pulled and tests are executed.
### Actual behavior
The custom image is not pulled and produces the following error:
`Caused by: java.lang.IllegalStateException: Failed to verify that image 'MY_IMAGE' is a compatible substitute for 'postgres'. This generally means that you are trying to use an image that Testcontainers has not been designed to use. If this is deliberate, and if you are confident that the image is compatible, you should declare compatibility in code using the `asCompatibleSubstituteFor` method. For example:
2022-11-25 11:52:46 DockerImageName myImage = DockerImageName.parse("MY_IMAGE").asCompatibleSubstituteFor("postgres");
2022-11-25 11:52:46 and then use `myImage` instead.`
### How to Reproduce?
1. `mvn clean install -DskipTests`
2. `mvn -f testsuite/integration-arquillian/tests/base/pom.xml clean install -Pauth-server-undertow -Pmap-storage,map-storage-jpa,db-postgres -Dpageload.timeout=90000 -Dkeycloak.map.storage.postgres.docker.image=MY_IMAGE`
### Anything else?
_No response_ | 1.0 | Unable to run map-storage-jpa tests with custom Postgres image - ### Area
testsuite
### Describe the bug
According to the [documentation](https://github.com/keycloak/keycloak/blob/main/testsuite/integration-arquillian/HOW-TO-RUN.md#running-tests-with-jpa-map-storage), it is possible to run JPA Map storage tests using a custom Postgres image defined in system property `keycloak.map.storage.postgres.docker.image`.
### Version
19.0.0
### Expected behavior
The custom Postgres image is pulled and tests are executed.
### Actual behavior
The custom image is not pulled and produces the following error:
`Caused by: java.lang.IllegalStateException: Failed to verify that image 'MY_IMAGE' is a compatible substitute for 'postgres'. This generally means that you are trying to use an image that Testcontainers has not been designed to use. If this is deliberate, and if you are confident that the image is compatible, you should declare compatibility in code using the `asCompatibleSubstituteFor` method. For example:
2022-11-25 11:52:46 DockerImageName myImage = DockerImageName.parse("MY_IMAGE").asCompatibleSubstituteFor("postgres");
2022-11-25 11:52:46 and then use `myImage` instead.`
### How to Reproduce?
1. `mvn clean install -DskipTests`
2. `mvn -f testsuite/integration-arquillian/tests/base/pom.xml clean install -Pauth-server-undertow -Pmap-storage,map-storage-jpa,db-postgres -Dpageload.timeout=90000 -Dkeycloak.map.storage.postgres.docker.image=MY_IMAGE`
### Anything else?
_No response_ | non_main | unable to run map storage jpa tests with custom postgres image area testsuite describe the bug according to the it is possible to run jpa map storage tests using a custom postgres image defined in system property keycloak map storage postgres docker image version expected behavior the custom postgres image is pulled and tests are executed actual behavior the custom image is not pulled and produces the following error caused by java lang illegalstateexception failed to verify that image my image is a compatible substitute for postgres this generally means that you are trying to use an image that testcontainers has not been designed to use if this is deliberate and if you are confident that the image is compatible you should declare compatibility in code using the ascompatiblesubstitutefor method for example dockerimagename myimage dockerimagename parse my image ascompatiblesubstitutefor postgres and then use myimage instead how to reproduce mvn clean install dskiptests mvn f testsuite integration arquillian tests base pom xml clean install pauth server undertow pmap storage map storage jpa db postgres dpageload timeout dkeycloak map storage postgres docker image my image anything else no response | 0 |
2,045 | 6,894,652,324 | IssuesEvent | 2017-11-23 10:47:46 | dgets/DANT2a | https://api.github.com/repos/dgets/DANT2a | closed | Switch to console window debugging or learn debugger | enhancement maintainability | Things are getting a little too complicated for a `MessageBox` to take care of it easily. In lieu of not being able to handle unit testing or TDD just yet, we should really open up a console window for debugging stats here. Unless, of course, learning watch points & variable watches in the debugger doesn't prove to be too terrible. | True | Switch to console window debugging or learn debugger - Things are getting a little too complicated for a `MessageBox` to take care of it easily. In lieu of not being able to handle unit testing or TDD just yet, we should really open up a console window for debugging stats here. Unless, of course, learning watch points & variable watches in the debugger doesn't prove to be too terrible. | main | switch to console window debugging or learn debugger things are getting a little too complicated for a messagebox to take care of it easily in lieu of not being able to handle unit testing or tdd just yet we should really open up a console window for debugging stats here unless of course learning watch points variable watches in the debugger doesn t prove to be too terrible | 1 |
130,232 | 18,055,103,696 | IssuesEvent | 2021-09-20 07:05:41 | opensrp/web | https://api.github.com/repos/opensrp/web | closed | User management Design QA issues | Bug Report Design | - [x] Remove gear icon
- [x] Set search text 14px
- [x] in the admin menu, when a subsection is open but you are in a different section, the subsection is still white. Is this default Ant? Update the submenu to show selection even when in a different section but under the same submenu
- [x] Remove the `Required Actions` field | 1.0 | User management Design QA issues - - [x] Remove gear icon
- [x] Set search text 14px
- [x] in the admin menu, when a subsection is open but you are in a different section, the subsection is still white. Is this default Ant? Update the submenu to show selection even when in a different section but under the same submenu
- [x] Remove the `Required Actions` field | non_main | user management design qa issues remove gear icon set search text in the admin menu when a subsection is open but you are in a different section the subsection is still white is this default ant update the submenu to show selection even when in a different section but under the same submenu remove the required actions field | 0 |
994 | 4,758,682,855 | IssuesEvent | 2016-10-24 20:13:40 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | get_url prevents caching of data | affects_2.0 bug_report feature_idea waiting_on_maintainer | Issue Type:
Bug Report
Component Name:
get_url module
Ansible Version:
ansible 2.0.0.2
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
Ansible Configuration:
Stock.
Environment:
Ubuntu 14.04
Summary:
get_url seems to add headers that prevent efficient use of a caching proxy.
Steps To Reproduce:
I am trying to download some ISOs to multiple machines (via a proxy to
conserve bandwidth).
The ISO is being stored by the proxy, and the machine is using the proxy, but
it is downloading from the upstream source every time. Squid is showing in its
logs:
1454462392.008 532579 192.168.122.10 TCP_CLIENT_REFRESH_MISS/200 632291702 GET
http://mirrors.kernel.org/centos/7.2.1511/isos/x86_64/CentOS-7-x86_64-Minimal-1511.iso - HIER_DIRECT/198.145.20.143 application/octet-stream
According to the squid docs:
TCP_CLIENT_REFRESH_MISS
The client issued a "no-cache" pragma, or some analogous cache control command
along with the request. Thus, the cache has to refetch the object.
Using a standard wget (or, say, yum to retrieve packages) does not cause
CLIENT_REFRESH_MISSes.
Is there something in the get_url code that is causing the sending of a no
cache pragma? Or maybe it's not turning off some default option in the
underlying urllib (or whatever it uses under the hood)?
Expected Results:
File will be downloaded much faster, and from the cache.
Actual Results:
Squid is actually downloading the file again. | True | get_url prevents caching of data - Issue Type:
Bug Report
Component Name:
get_url module
Ansible Version:
ansible 2.0.0.2
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
Ansible Configuration:
Stock.
Environment:
Ubuntu 14.04
Summary:
get_url seems to add headers that prevent efficient use of a caching proxy.
Steps To Reproduce:
I am trying to download some ISOs to multiple machines (via a proxy to
conserve bandwidth).
The ISO is being stored by the proxy, and the machine is using the proxy, but
it is downloading from the upstream source every time. Squid is showing in its
logs:
1454462392.008 532579 192.168.122.10 TCP_CLIENT_REFRESH_MISS/200 632291702 GET
http://mirrors.kernel.org/centos/7.2.1511/isos/x86_64/CentOS-7-x86_64-Minimal-1511.iso - HIER_DIRECT/198.145.20.143 application/octet-stream
According to the squid docs:
TCP_CLIENT_REFRESH_MISS
The client issued a "no-cache" pragma, or some analogous cache control command
along with the request. Thus, the cache has to refetch the object.
Using a standard wget (or, say, yum to retrieve packages) does not cause
CLIENT_REFRESH_MISSes.
Is there something in the get_url code that is causing the sending of a no
cache pragma? Or maybe it's not turning off some default option in the
underlying urllib (or whatever it uses under the hood)?
Expected Results:
File will be downloaded much faster, and from the cache.
Actual Results:
Squid is actually downloading the file again. | main | get url prevents caching of data issue type bug report component name get url module ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides ansible configuration stock environment ubuntu summary get url seems to add headers that prevent efficient use of a caching proxy steps to reproduce i am trying to download some isos to multiple machines via a proxy to conserve bandwidth the iso is being stored by the proxy and the machine is using the proxy but it is downloading from the upstream source every time squid is showing in its logs tcp client refresh miss get hier direct application octet stream according to the squid docs tcp client refresh miss the client issued a no cache pragma or some analogous cache control command along with the request thus the cache has to refetch the object using a standard wget or say yum to retrieve packages does not cause client refresh misses is there something in the get url code that is causing the sending of a no cache pragma or maybe it s not turning off some default option in the underlying urllib or whatever it uses under the hood expected results file will be downloaded much faster and from the cache actual results squid is actually downloading the file again | 1 |
2,885 | 10,319,592,794 | IssuesEvent | 2019-08-30 17:57:59 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | closed | becoming group member and porting modules | Maintainer application | Hello, I got a request to port one of the modules I maintain on d.o to backdrop and I'd like to support that myself. I think I have to apply to join first correct?
It will be the d3 module, and I'm a maintainer and co-maintainer of a few other projects that I can port when the time comes as well.
| True | becoming group member and porting modules - Hello, I got a request to port one of the modules I maintain on d.o to backdrop and I'd like to support that myself. I think I have to apply to join first correct?
It will be the d3 module, and I'm a maintainer and co-maintainer of a few other projects that I can port when the time comes as well.
| main | becoming group member and porting modules hello i got a request to port one of the modules i maintain on d o to backdrop and i d like to support that myself i think i have to apply to join first correct it will be the module and i m a maintainer and co maintainer of a few other projects that i can port when the time comes as well | 1 |
1,883 | 6,577,516,314 | IssuesEvent | 2017-09-12 01:27:36 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Fact gathering randomly fails on a VM | affects_2.0 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
setup module
##### ANSIBLE VERSION
```
ansible 2.0.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
```
[defaults]
inventory=hosts
host_key_checking=false
nocows=1
fact_caching=jsonfile
fact_caching_connection=facts/
ansible_managed = Ansible managed: {file} modified on %Y-%m-%d by {uid} on {host}
scp_if_ssh=true
timeout=20
pipelining=true
[ssh_connection]
ssh_args = -o ControlMaster=auto -o ControlPersist=60s
```
##### OS / ENVIRONMENT
HOST: Ubuntu 15.10
Linux laptappy 4.2.0-34-generic #39-Ubuntu SMP Thu Mar 10 22:13:01 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
VM: Centos 7
Linux gl-docker 4.4.3-1.el7.elrepo.x86_64 #1 SMP Thu Feb 25 17:09:04 EST 2016 x86_64 x86_64 x86_64 GNU/Linux
##### SUMMARY
Trying to gather facts from a newly spun up vagrant box fails. I am not using the ansible vagrant provisioner, I am running ansible after the fact. I can connect all night and day with the connection string that ansible verbose output uses. After a successful fact gathering, I can run ansible commands all night and day against my node(s). Fact gathering is the only thing that chokes randomly.
##### STEPS TO REPRODUCE
The command I run
```
cd ansible ; ansible vagrant -m setup -a "fact_path=facts" -vvvv ; cd ..
```
Inventory. Regardless if I use the private key or password, the results are the same.
```
gl-docker ansible_connection=ssh ansible_host=127.0.0.1 ansible_port=2233 ansible_user=vagrant ansible_ssh_private_key_file=../.vagrant/machines/gl-docker/virtualbox/private_key
[vagrant]
gl-docker
```
I can run this command and everything appears to exit cleanly. I copy/pasted this string from the ansible verbose output.
```
for i in {1..200} ; do ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 'echo -e "#########\nHEY IM IN\n#########" ; exit' ; echo $?; echo $i ;do
```
I've also set the following variable to successfully and repeatedly run the setup module on the host itself.
```
export ANSIBLE_KEEP_REMOTE_FILES=1
```
##### EXPECTED RESULTS
I expect fact gathering to return successfully each and every time.
##### ACTUAL RESULTS
Fact gathering randomly fails. Here are two subsequent fact gathering runs.
```
Using /opt/work-repos/docker-microservice-devenv/ansible/ansible.cfg as config file
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776 `" )'"'"''
<127.0.0.1> PUT /tmp/tmpwzYwQD TO /home/vagrant/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776/setup
<127.0.0.1> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r '[127.0.0.1]'
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/vagrant/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776/setup; rm -rf "/home/vagrant/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776/" > /dev/null 2>&1'"'"''
gl-docker | FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "setup"
},
"module_stderr": "",
"module_stdout": "{\"invocation\": {\"module_args\": {\"filter\": \"*\", \"fact_path\": \"facts\"}}, \"changed\": false, \"_ansible_verbose_override\": true, \"ansible_facts\": {\"ansible_product_serial\": \"NA\", \"ansible_form_factor\": \"Other\", \"ansible_product_version\": \"1.2\", \"ansible_fips\": false, \"ansible_service_mgr\": \"systemd\", \"ansible_user_id\": \"vagrant\", \"module_setup\": true, \"ansible_memtotal_mb\": 2000, \"ansible_ssh_host_key_rsa_public\": \"AAAAB3NzaC1yc2EAAAADAQABAAABAQDJ8cfLKMpYB/mrhS3DuBTB6cTryHUzLHJ1gIk9Ro+QhbidvnAb9Br4182qytb2fmSB0kn+I4WbZBSDj2Kv9uicL8BVEyoMZtAsDZ0cHBsN5Su+Dk5mijG4PplAM8CjU6BxvhGZJ+lxuT4IMMCEMua7SmVVITaiebVbcTi7x61WsuOaKAAe+D9xaHcY0tlKvr0XaZxnyxtscbHLEc1vi6WUYlM1uLXdnMycE5DYlmsl5FfCkrtFeEG/xdbbcmBlrSl7WUwyVY0w1GPGBQFatEHf8PJhYTELNWVHR1rCMMP8Q/uiT9E/UkQHASS3XgRe+W6FqgD6pO6mOHkLXjoFVtsD\", \"ansible_ssh_host_key_ecdsa_public\": \"AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDB17PdGWfCxEJRNI1dj88lwt47Ah4eneGboTvG8V2Rctf7nGAOdLrB+//YrHPyb7mSiJp58/p+w8qkKy4sFQ38=\", \"ansible_distribution_version\": \"7.2.1511\", \"ansible_domain\": \"localdomain\", \"ansible_user_shell\": \"/bin/bash\", \"ansible_date_time\": {\"weekday_number\": \"5\", \"iso8601_basic_short\": \"20160325T132842\", \"tz\": \"UTC\", \"weeknumber\": \"12\", \"hour\": \"13\", \"year\": \"2016\", \"minute\": \"28\", \"tz_offset\": \"+0000\", \"month\": \"03\", \"epoch\": \"1458912522\", \"iso8601_micro\": \"2016-03-25T13:28:42.761267Z\", \"weekday\": \"Friday\", \"time\": \"13:28:42\", \"date\": \"2016-03-25\", \"iso8601\": \"2016-03-25T13:28:42Z\", \"day\": \"25\", \"iso8601_basic\": \"20160325T132842761179\", \"second\": \"42\"}, \"ansible_ssh_host_key_ed25519_public\": \"AAAAC3NzaC1lZDI1NTE5AAAAIHtYETIJC0EFGF1Mam7pYjiVxAVTNQp6yqADQoSS13bP\", \"ansible_processor_cores\": 1, \"ansible_virtualization_role\": \"guest\", \"ansible_env\": {\"LESSOPEN\": \"||/usr/bin/lesspipe.sh %s\", \"SSH_CLIENT\": \"10.0.2.2 57776 22\", \"SELINUX_USE_CURRENT_RANGE\": \"\", \"LOGNAME\": \"vagrant\", \"USER\": \"vagrant\", \"PATH\": \"/usr/local/bin:/usr/bin\", \"HOME\": \"/home/vagrant\", \"LANG\": \"en_US.UTF-8\", \"TERM\": \"xterm-256color\", \"SHELL\": \"/bin/bash\", \"SHLVL\": \"2\", \"LC_ALL\": \"en_US.UTF-8\", \"XDG_RUNTIME_DIR\": \"/run/user/1000\", \"SELINUX_ROLE_REQUESTED\": \"\", \"QT_GRAPHICSSYSTEM_CHECKED\": \"1\", \"XDG_SESSION_ID\": \"11\", \"_\": \"/usr/bin/python\", \"LC_MESSAGES\": \"en_US.UTF-8\", \"SSH_TTY\": \"/dev/pts/0\", \"SELINUX_LEVEL_REQUESTED\": \"\", \"PWD\": \"/home/vagrant\", \"MAIL\": \"/var/mail/vagrant\", \"SSH_CONNECTION\": \"10.0.2.2 57776 10.0.2.15 22\"}, \"ansible_processor_vcpus\": 1, \"ansible_docker0\": {\"macaddress\": \"02:42:d0:a0:e3:30\", \"interfaces\": [], \"mtu\": 1500, \"device\": \"docker0\", \"promisc\": false, \"stp\": false, \"ipv4\": {\"broadcast\": \"global\", \"netmask\": \"255.255.0.0\", \"network\": \"172.17.0.0\", \"address\": \"172.17.0.1\"}, \"active\": false, \"type\": \"bridge\", \"id\": \"8000.0242d0a0e330\"}, \"ansible_bios_version\": \"VirtualBox\", \"ansible_processor\": [\"GenuineIntel\", \"Intel(R) Core(TM) i7-4578U CPU @ 3.00GHz\"], \"ansible_virtualization_type\": \"virtualbox\", \"ansible_lo\": {\"mtu\": 65536, \"device\": \"lo\", \"promisc\": false, \"ipv4\": {\"broadcast\": \"host\", \"netmask\": \"255.0.0.0\", \"network\": \"127.0.0.0\", \"address\": \"127.0.0.1\"}, \"ipv6\": [{\"scope\": \"host\", \"prefix\": \"128\", \"address\": \"::1\"}], \"active\": true, \"type\": \"loopback\"}, \"ansible_userspace_bits\": \"64\", \"ansible_architecture\": \"x86_64\", \"ansible_default_ipv4\": {\"macaddress\": \"08:00:27:07:5e:92\", \"network\": \"10.0.2.0\", \"mtu\": 1500, \"broadcast\": \"10.0.2.255\", \"alias\": \"enp0s3\", \"netmask\": \"255.255.255.0\", \"address\": \"10.0.2.15\", \"interface\": \"enp0s3\", \"type\": \"ether\", \"gateway\": \"10.0.2.2\"}, \"ansible_swapfree_mb\": 1015, \"ansible_default_ipv6\": {}, \"ansible_distribution_release\": \"Core\", \"ansible_system_vendor\": \"innotek GmbH\", \"ansible_os_family\": \"RedHat\", \"ansible_cmdline\": {\"BOOT_IMAGE\": \"/vmlinuz-4.4.3-1.el7.elrepo.x86_64\", \"quiet\": true, \"rhgb\": true, \"rd.lvm.lv\": \"centos/swap\", \"crashkernel\": \"auto\", \"ro\": true, \"root\": \"/dev/mapper/centos-root\"}, \"ansible_mounts\": [{\"uuid\": \"319365c2-ad4a-4a6b-986a-0b050e96624c\", \"size_total\": 8986296320, \"mount\": \"/\", \"size_available\": 6826946560, \"fstype\": \"xfs\", \"device\": \"/dev/mapper/centos-root\", \"options\": \"rw,seclabel,relatime,attr2,inode64,noquota\"",
"msg": "MODULE FAILURE",
"parsed": false
}
###########################################################
###########################################################
Using /opt/work-repos/docker-microservice-devenv/ansible/ansible.cfg as config file
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005 `" )'"'"''
<127.0.0.1> PUT /tmp/tmpcA5aGq TO /home/vagrant/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005/setup
<127.0.0.1> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r '[127.0.0.1]'
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/vagrant/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005/setup; rm -rf "/home/vagrant/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005/" > /dev/null 2>&1'"'"''
gl-docker | SUCCESS => {
"ansible_facts": {
"ansible_all_ipv4_addresses": [
"172.17.0.1",
"10.0.2.15",
"10.100.101.111"
],
"ansible_all_ipv6_addresses": [
"fe80::a00:27ff:fe07:5e92",
"fe80::a00:27ff:fe63:1c7c"
],
"ansible_architecture": "x86_64",
"ansible_bios_date": "12/01/2006",
"ansible_bios_version": "VirtualBox",
"ansible_cmdline": {
"BOOT_IMAGE": "/vmlinuz-4.4.3-1.el7.elrepo.x86_64",
"crashkernel": "auto",
"quiet": true,
"rd.lvm.lv": "centos/swap",
"rhgb": true,
"ro": true,
"root": "/dev/mapper/centos-root"
},
"ansible_date_time": {
"date": "2016-03-25",
"day": "25",
"epoch": "1458912524",
"hour": "13",
"iso8601": "2016-03-25T13:28:44Z",
"iso8601_basic": "20160325T132844056072",
"iso8601_basic_short": "20160325T132844",
"iso8601_micro": "2016-03-25T13:28:44.056144Z",
"minute": "28",
"month": "03",
"second": "44",
"time": "13:28:44",
"tz": "UTC",
"tz_offset": "+0000",
"weekday": "Friday",
"weekday_number": "5",
"weeknumber": "12",
"year": "2016"
},
"ansible_default_ipv4": {
"address": "10.0.2.15",
"alias": "enp0s3",
"broadcast": "10.0.2.255",
"gateway": "10.0.2.2",
"interface": "enp0s3",
"macaddress": "08:00:27:07:5e:92",
"mtu": 1500,
"netmask": "255.255.255.0",
"network": "10.0.2.0",
"type": "ether"
},
"ansible_default_ipv6": {},
"ansible_devices": {
"sda": {
"holders": [],
"host": "IDE interface: Intel Corporation 82371AB/EB/MB PIIX4 IDE (rev 01)",
"model": "VBOX HARDDISK",
"partitions": {
"sda1": {
"sectors": "1024000",
"sectorsize": 512,
"size": "500.00 MB",
"start": "2048"
},
"sda2": {
"sectors": "19740672",
"sectorsize": 512,
"size": "9.41 GB",
"start": "1026048"
}
},
"removable": "0",
"rotational": "1",
"scheduler_mode": "deadline",
"sectors": "20766720",
"sectorsize": "512",
"size": "9.90 GB",
"support_discard": "0",
"vendor": "ATA"
}
},
"ansible_distribution": "CentOS",
"ansible_distribution_major_version": "7",
"ansible_distribution_release": "Core",
"ansible_distribution_version": "7.2.1511",
"ansible_dns": {
"nameservers": [
"10.0.2.3"
]
},
"ansible_docker0": {
"active": false,
"device": "docker0",
"id": "8000.0242d0a0e330",
"interfaces": [],
"ipv4": {
"address": "172.17.0.1",
"broadcast": "global",
"netmask": "255.255.0.0",
"network": "172.17.0.0"
},
"macaddress": "02:42:d0:a0:e3:30",
"mtu": 1500,
"promisc": false,
"stp": false,
"type": "bridge"
},
"ansible_domain": "localdomain",
"ansible_enp0s3": {
"active": true,
"device": "enp0s3",
"ipv4": {
"address": "10.0.2.15",
"broadcast": "10.0.2.255",
"netmask": "255.255.255.0",
"network": "10.0.2.0"
},
"ipv6": [
{
"address": "fe80::a00:27ff:fe07:5e92",
"prefix": "64",
"scope": "link"
}
],
"macaddress": "08:00:27:07:5e:92",
"module": "e1000",
"mtu": 1500,
"pciid": "0000:00:03.0",
"promisc": false,
"type": "ether"
},
"ansible_enp0s8": {
"active": true,
"device": "enp0s8",
"ipv4": {
"address": "10.100.101.111",
"broadcast": "10.100.101.255",
"netmask": "255.255.255.0",
"network": "10.100.101.0"
},
"ipv6": [
{
"address": "fe80::a00:27ff:fe63:1c7c",
"prefix": "64",
"scope": "link"
}
],
"macaddress": "08:00:27:63:1c:7c",
"module": "e1000",
"mtu": 1500,
"pciid": "0000:00:08.0",
"promisc": false,
"type": "ether"
},
"ansible_env": {
"HOME": "/home/vagrant",
"LANG": "en_US.UTF-8",
"LC_ALL": "en_US.UTF-8",
"LC_MESSAGES": "en_US.UTF-8",
"LESSOPEN": "||/usr/bin/lesspipe.sh %s",
"LOGNAME": "vagrant",
"MAIL": "/var/mail/vagrant",
"PATH": "/usr/local/bin:/usr/bin",
"PWD": "/home/vagrant",
"QT_GRAPHICSSYSTEM_CHECKED": "1",
"SELINUX_LEVEL_REQUESTED": "",
"SELINUX_ROLE_REQUESTED": "",
"SELINUX_USE_CURRENT_RANGE": "",
"SHELL": "/bin/bash",
"SHLVL": "2",
"SSH_CLIENT": "10.0.2.2 57776 22",
"SSH_CONNECTION": "10.0.2.2 57776 10.0.2.15 22",
"SSH_TTY": "/dev/pts/0",
"TERM": "xterm-256color",
"USER": "vagrant",
"XDG_RUNTIME_DIR": "/run/user/1000",
"XDG_SESSION_ID": "11",
"_": "/usr/bin/python"
},
"ansible_fips": false,
"ansible_form_factor": "Other",
"ansible_fqdn": "localhost.localdomain",
"ansible_hostname": "gl-docker",
"ansible_interfaces": [
"lo",
"docker0",
"enp0s3",
"enp0s8"
],
"ansible_kernel": "4.4.3-1.el7.elrepo.x86_64",
"ansible_lo": {
"active": true,
"device": "lo",
"ipv4": {
"address": "127.0.0.1",
"broadcast": "host",
"netmask": "255.0.0.0",
"network": "127.0.0.0"
},
"ipv6": [
{
"address": "::1",
"prefix": "128",
"scope": "host"
}
],
"mtu": 65536,
"promisc": false,
"type": "loopback"
},
"ansible_lsb": {
"codename": "Core",
"description": "CentOS Linux release 7.2.1511 (Core)",
"id": "CentOS",
"major_release": "7",
"release": "7.2.1511"
},
"ansible_machine": "x86_64",
"ansible_machine_id": "69c2e9ff6b3b4594b1a26db25287da79",
"ansible_memfree_mb": 1629,
"ansible_memory_mb": {
"nocache": {
"free": 1790,
"used": 210
},
"real": {
"free": 1629,
"total": 2000,
"used": 371
},
"swap": {
"cached": 0,
"free": 1015,
"total": 1015,
"used": 0
}
},
"ansible_memtotal_mb": 2000,
"ansible_mounts": [
{
"device": "/dev/mapper/centos-root",
"fstype": "xfs",
"mount": "/",
"options": "rw,seclabel,relatime,attr2,inode64,noquota",
"size_available": 6826946560,
"size_total": 8986296320,
"uuid": "319365c2-ad4a-4a6b-986a-0b050e96624c"
},
{
"device": "/dev/sda1",
"fstype": "xfs",
"mount": "/boot",
"options": "rw,seclabel,relatime,attr2,inode64,noquota",
"size_available": 320561152,
"size_total": 520794112,
"uuid": "9a2e2b29-a60f-484b-be03-30df74cc87a4"
}
],
"ansible_nodename": "gl-docker",
"ansible_os_family": "RedHat",
"ansible_pkg_mgr": "yum",
"ansible_processor": [
"GenuineIntel",
"Intel(R) Core(TM) i7-4578U CPU @ 3.00GHz"
],
"ansible_processor_cores": 1,
"ansible_processor_count": 1,
"ansible_processor_threads_per_core": 1,
"ansible_processor_vcpus": 1,
"ansible_product_name": "VirtualBox",
"ansible_product_serial": "NA",
"ansible_product_uuid": "NA",
"ansible_product_version": "1.2",
"ansible_python_version": "2.7.5",
"ansible_selinux": {
"config_mode": "permissive",
"mode": "permissive",
"policyvers": 30,
"status": "enabled",
"type": "targeted"
},
"ansible_service_mgr": "systemd",
"ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDB17PdGWfCxEJRNI1dj88lwt47Ah4eneGboTvG8V2Rctf7nGAOdLrB+//YrHPyb7mSiJp58/p+w8qkKy4sFQ38=",
"ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIHtYETIJC0EFGF1Mam7pYjiVxAVTNQp6yqADQoSS13bP",
"ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDJ8cfLKMpYB/mrhS3DuBTB6cTryHUzLHJ1gIk9Ro+QhbidvnAb9Br4182qytb2fmSB0kn+I4WbZBSDj2Kv9uicL8BVEyoMZtAsDZ0cHBsN5Su+Dk5mijG4PplAM8CjU6BxvhGZJ+lxuT4IMMCEMua7SmVVITaiebVbcTi7x61WsuOaKAAe+D9xaHcY0tlKvr0XaZxnyxtscbHLEc1vi6WUYlM1uLXdnMycE5DYlmsl5FfCkrtFeEG/xdbbcmBlrSl7WUwyVY0w1GPGBQFatEHf8PJhYTELNWVHR1rCMMP8Q/uiT9E/UkQHASS3XgRe+W6FqgD6pO6mOHkLXjoFVtsD",
"ansible_swapfree_mb": 1015,
"ansible_swaptotal_mb": 1015,
"ansible_system": "Linux",
"ansible_system_vendor": "innotek GmbH",
"ansible_uptime_seconds": 2173,
"ansible_user_dir": "/home/vagrant",
"ansible_user_gecos": "vagrant",
"ansible_user_gid": 1000,
"ansible_user_id": "vagrant",
"ansible_user_shell": "/bin/bash",
"ansible_user_uid": 1000,
"ansible_userspace_architecture": "x86_64",
"ansible_userspace_bits": "64",
"ansible_virtualization_role": "guest",
"ansible_virtualization_type": "virtualbox",
"module_setup": true
},
"changed": false,
"invocation": {
"module_args": {
"fact_path": "facts",
"filter": "*"
},
"module_name": "setup"
}
}
```
| True | Fact gathering randomly fails on a VM - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
setup module
##### ANSIBLE VERSION
```
ansible 2.0.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
```
[defaults]
inventory=hosts
host_key_checking=false
nocows=1
fact_caching=jsonfile
fact_caching_connection=facts/
ansible_managed = Ansible managed: {file} modified on %Y-%m-%d by {uid} on {host}
scp_if_ssh=true
timeout=20
pipelining=true
[ssh_connection]
ssh_args = -o ControlMaster=auto -o ControlPersist=60s
```
##### OS / ENVIRONMENT
HOST: Ubuntu 15.10
Linux laptappy 4.2.0-34-generic #39-Ubuntu SMP Thu Mar 10 22:13:01 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
VM: Centos 7
Linux gl-docker 4.4.3-1.el7.elrepo.x86_64 #1 SMP Thu Feb 25 17:09:04 EST 2016 x86_64 x86_64 x86_64 GNU/Linux
##### SUMMARY
Trying to gather facts from a newly spun up vagrant box fails. I am not using the ansible vagrant provisioner, I am running ansible after the fact. I can connect all night and day with the connection string that ansible verbose output uses. After a successful fact gathering, I can run ansible commands all night and day against my node(s). Fact gathering is the only thing that chokes randomly.
##### STEPS TO REPRODUCE
The command I run
```
cd ansible ; ansible vagrant -m setup -a "fact_path=facts" -vvvv ; cd ..
```
Inventory. Regardless if I use the private key or password, the results are the same.
```
gl-docker ansible_connection=ssh ansible_host=127.0.0.1 ansible_port=2233 ansible_user=vagrant ansible_ssh_private_key_file=../.vagrant/machines/gl-docker/virtualbox/private_key
[vagrant]
gl-docker
```
I can run this command and everything appears to exit cleanly. I copy/pasted this string from the ansible verbose output.
```
for i in {1..200} ; do ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 'echo -e "#########\nHEY IM IN\n#########" ; exit' ; echo $?; echo $i ;do
```
I've also set the following variable to successfully and repeatedly run the setup module on the host itself.
```
export ANSIBLE_KEEP_REMOTE_FILES=1
```
##### EXPECTED RESULTS
I expect fact gathering to return successfully each and every time.
##### ACTUAL RESULTS
Fact gathering randomly fails. Here are two subsequent fact gathering runs.
```
Using /opt/work-repos/docker-microservice-devenv/ansible/ansible.cfg as config file
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776 `" )'"'"''
<127.0.0.1> PUT /tmp/tmpwzYwQD TO /home/vagrant/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776/setup
<127.0.0.1> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r '[127.0.0.1]'
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/vagrant/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776/setup; rm -rf "/home/vagrant/.ansible/tmp/ansible-tmp-1458912522.44-162920354534776/" > /dev/null 2>&1'"'"''
gl-docker | FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "setup"
},
"module_stderr": "",
"module_stdout": "{\"invocation\": {\"module_args\": {\"filter\": \"*\", \"fact_path\": \"facts\"}}, \"changed\": false, \"_ansible_verbose_override\": true, \"ansible_facts\": {\"ansible_product_serial\": \"NA\", \"ansible_form_factor\": \"Other\", \"ansible_product_version\": \"1.2\", \"ansible_fips\": false, \"ansible_service_mgr\": \"systemd\", \"ansible_user_id\": \"vagrant\", \"module_setup\": true, \"ansible_memtotal_mb\": 2000, \"ansible_ssh_host_key_rsa_public\": \"AAAAB3NzaC1yc2EAAAADAQABAAABAQDJ8cfLKMpYB/mrhS3DuBTB6cTryHUzLHJ1gIk9Ro+QhbidvnAb9Br4182qytb2fmSB0kn+I4WbZBSDj2Kv9uicL8BVEyoMZtAsDZ0cHBsN5Su+Dk5mijG4PplAM8CjU6BxvhGZJ+lxuT4IMMCEMua7SmVVITaiebVbcTi7x61WsuOaKAAe+D9xaHcY0tlKvr0XaZxnyxtscbHLEc1vi6WUYlM1uLXdnMycE5DYlmsl5FfCkrtFeEG/xdbbcmBlrSl7WUwyVY0w1GPGBQFatEHf8PJhYTELNWVHR1rCMMP8Q/uiT9E/UkQHASS3XgRe+W6FqgD6pO6mOHkLXjoFVtsD\", \"ansible_ssh_host_key_ecdsa_public\": \"AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDB17PdGWfCxEJRNI1dj88lwt47Ah4eneGboTvG8V2Rctf7nGAOdLrB+//YrHPyb7mSiJp58/p+w8qkKy4sFQ38=\", \"ansible_distribution_version\": \"7.2.1511\", \"ansible_domain\": \"localdomain\", \"ansible_user_shell\": \"/bin/bash\", \"ansible_date_time\": {\"weekday_number\": \"5\", \"iso8601_basic_short\": \"20160325T132842\", \"tz\": \"UTC\", \"weeknumber\": \"12\", \"hour\": \"13\", \"year\": \"2016\", \"minute\": \"28\", \"tz_offset\": \"+0000\", \"month\": \"03\", \"epoch\": \"1458912522\", \"iso8601_micro\": \"2016-03-25T13:28:42.761267Z\", \"weekday\": \"Friday\", \"time\": \"13:28:42\", \"date\": \"2016-03-25\", \"iso8601\": \"2016-03-25T13:28:42Z\", \"day\": \"25\", \"iso8601_basic\": \"20160325T132842761179\", \"second\": \"42\"}, \"ansible_ssh_host_key_ed25519_public\": \"AAAAC3NzaC1lZDI1NTE5AAAAIHtYETIJC0EFGF1Mam7pYjiVxAVTNQp6yqADQoSS13bP\", \"ansible_processor_cores\": 1, \"ansible_virtualization_role\": \"guest\", \"ansible_env\": {\"LESSOPEN\": \"||/usr/bin/lesspipe.sh %s\", \"SSH_CLIENT\": \"10.0.2.2 57776 22\", \"SELINUX_USE_CURRENT_RANGE\": \"\", \"LOGNAME\": \"vagrant\", \"USER\": \"vagrant\", \"PATH\": \"/usr/local/bin:/usr/bin\", \"HOME\": \"/home/vagrant\", \"LANG\": \"en_US.UTF-8\", \"TERM\": \"xterm-256color\", \"SHELL\": \"/bin/bash\", \"SHLVL\": \"2\", \"LC_ALL\": \"en_US.UTF-8\", \"XDG_RUNTIME_DIR\": \"/run/user/1000\", \"SELINUX_ROLE_REQUESTED\": \"\", \"QT_GRAPHICSSYSTEM_CHECKED\": \"1\", \"XDG_SESSION_ID\": \"11\", \"_\": \"/usr/bin/python\", \"LC_MESSAGES\": \"en_US.UTF-8\", \"SSH_TTY\": \"/dev/pts/0\", \"SELINUX_LEVEL_REQUESTED\": \"\", \"PWD\": \"/home/vagrant\", \"MAIL\": \"/var/mail/vagrant\", \"SSH_CONNECTION\": \"10.0.2.2 57776 10.0.2.15 22\"}, \"ansible_processor_vcpus\": 1, \"ansible_docker0\": {\"macaddress\": \"02:42:d0:a0:e3:30\", \"interfaces\": [], \"mtu\": 1500, \"device\": \"docker0\", \"promisc\": false, \"stp\": false, \"ipv4\": {\"broadcast\": \"global\", \"netmask\": \"255.255.0.0\", \"network\": \"172.17.0.0\", \"address\": \"172.17.0.1\"}, \"active\": false, \"type\": \"bridge\", \"id\": \"8000.0242d0a0e330\"}, \"ansible_bios_version\": \"VirtualBox\", \"ansible_processor\": [\"GenuineIntel\", \"Intel(R) Core(TM) i7-4578U CPU @ 3.00GHz\"], \"ansible_virtualization_type\": \"virtualbox\", \"ansible_lo\": {\"mtu\": 65536, \"device\": \"lo\", \"promisc\": false, \"ipv4\": {\"broadcast\": \"host\", \"netmask\": \"255.0.0.0\", \"network\": \"127.0.0.0\", \"address\": \"127.0.0.1\"}, \"ipv6\": [{\"scope\": \"host\", \"prefix\": \"128\", \"address\": \"::1\"}], \"active\": true, \"type\": \"loopback\"}, \"ansible_userspace_bits\": \"64\", \"ansible_architecture\": \"x86_64\", \"ansible_default_ipv4\": {\"macaddress\": \"08:00:27:07:5e:92\", \"network\": \"10.0.2.0\", \"mtu\": 1500, \"broadcast\": \"10.0.2.255\", \"alias\": \"enp0s3\", \"netmask\": \"255.255.255.0\", \"address\": \"10.0.2.15\", \"interface\": \"enp0s3\", \"type\": \"ether\", \"gateway\": \"10.0.2.2\"}, \"ansible_swapfree_mb\": 1015, \"ansible_default_ipv6\": {}, \"ansible_distribution_release\": \"Core\", \"ansible_system_vendor\": \"innotek GmbH\", \"ansible_os_family\": \"RedHat\", \"ansible_cmdline\": {\"BOOT_IMAGE\": \"/vmlinuz-4.4.3-1.el7.elrepo.x86_64\", \"quiet\": true, \"rhgb\": true, \"rd.lvm.lv\": \"centos/swap\", \"crashkernel\": \"auto\", \"ro\": true, \"root\": \"/dev/mapper/centos-root\"}, \"ansible_mounts\": [{\"uuid\": \"319365c2-ad4a-4a6b-986a-0b050e96624c\", \"size_total\": 8986296320, \"mount\": \"/\", \"size_available\": 6826946560, \"fstype\": \"xfs\", \"device\": \"/dev/mapper/centos-root\", \"options\": \"rw,seclabel,relatime,attr2,inode64,noquota\"",
"msg": "MODULE FAILURE",
"parsed": false
}
###########################################################
###########################################################
Using /opt/work-repos/docker-microservice-devenv/ansible/ansible.cfg as config file
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'( umask 22 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005 `" && echo "` echo $HOME/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005 `" )'"'"''
<127.0.0.1> PUT /tmp/tmpcA5aGq TO /home/vagrant/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005/setup
<127.0.0.1> SSH: EXEC sftp -b - -C -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r '[127.0.0.1]'
<127.0.0.1> ESTABLISH SSH CONNECTION FOR USER: vagrant
<127.0.0.1> SSH: EXEC ssh -C -q -o ControlMaster=auto -o ControlPersist=60s -o StrictHostKeyChecking=no -o Port=2233 -o 'IdentityFile="../.vagrant/machines/gl-docker/virtualbox/private_key"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=vagrant -o ConnectTimeout=20 -o ControlPath=/home/phil/.ansible/cp/ansible-ssh-%h-%p-%r -tt 127.0.0.1 '/bin/sh -c '"'"'LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/vagrant/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005/setup; rm -rf "/home/vagrant/.ansible/tmp/ansible-tmp-1458912523.74-172174991201005/" > /dev/null 2>&1'"'"''
gl-docker | SUCCESS => {
"ansible_facts": {
"ansible_all_ipv4_addresses": [
"172.17.0.1",
"10.0.2.15",
"10.100.101.111"
],
"ansible_all_ipv6_addresses": [
"fe80::a00:27ff:fe07:5e92",
"fe80::a00:27ff:fe63:1c7c"
],
"ansible_architecture": "x86_64",
"ansible_bios_date": "12/01/2006",
"ansible_bios_version": "VirtualBox",
"ansible_cmdline": {
"BOOT_IMAGE": "/vmlinuz-4.4.3-1.el7.elrepo.x86_64",
"crashkernel": "auto",
"quiet": true,
"rd.lvm.lv": "centos/swap",
"rhgb": true,
"ro": true,
"root": "/dev/mapper/centos-root"
},
"ansible_date_time": {
"date": "2016-03-25",
"day": "25",
"epoch": "1458912524",
"hour": "13",
"iso8601": "2016-03-25T13:28:44Z",
"iso8601_basic": "20160325T132844056072",
"iso8601_basic_short": "20160325T132844",
"iso8601_micro": "2016-03-25T13:28:44.056144Z",
"minute": "28",
"month": "03",
"second": "44",
"time": "13:28:44",
"tz": "UTC",
"tz_offset": "+0000",
"weekday": "Friday",
"weekday_number": "5",
"weeknumber": "12",
"year": "2016"
},
"ansible_default_ipv4": {
"address": "10.0.2.15",
"alias": "enp0s3",
"broadcast": "10.0.2.255",
"gateway": "10.0.2.2",
"interface": "enp0s3",
"macaddress": "08:00:27:07:5e:92",
"mtu": 1500,
"netmask": "255.255.255.0",
"network": "10.0.2.0",
"type": "ether"
},
"ansible_default_ipv6": {},
"ansible_devices": {
"sda": {
"holders": [],
"host": "IDE interface: Intel Corporation 82371AB/EB/MB PIIX4 IDE (rev 01)",
"model": "VBOX HARDDISK",
"partitions": {
"sda1": {
"sectors": "1024000",
"sectorsize": 512,
"size": "500.00 MB",
"start": "2048"
},
"sda2": {
"sectors": "19740672",
"sectorsize": 512,
"size": "9.41 GB",
"start": "1026048"
}
},
"removable": "0",
"rotational": "1",
"scheduler_mode": "deadline",
"sectors": "20766720",
"sectorsize": "512",
"size": "9.90 GB",
"support_discard": "0",
"vendor": "ATA"
}
},
"ansible_distribution": "CentOS",
"ansible_distribution_major_version": "7",
"ansible_distribution_release": "Core",
"ansible_distribution_version": "7.2.1511",
"ansible_dns": {
"nameservers": [
"10.0.2.3"
]
},
"ansible_docker0": {
"active": false,
"device": "docker0",
"id": "8000.0242d0a0e330",
"interfaces": [],
"ipv4": {
"address": "172.17.0.1",
"broadcast": "global",
"netmask": "255.255.0.0",
"network": "172.17.0.0"
},
"macaddress": "02:42:d0:a0:e3:30",
"mtu": 1500,
"promisc": false,
"stp": false,
"type": "bridge"
},
"ansible_domain": "localdomain",
"ansible_enp0s3": {
"active": true,
"device": "enp0s3",
"ipv4": {
"address": "10.0.2.15",
"broadcast": "10.0.2.255",
"netmask": "255.255.255.0",
"network": "10.0.2.0"
},
"ipv6": [
{
"address": "fe80::a00:27ff:fe07:5e92",
"prefix": "64",
"scope": "link"
}
],
"macaddress": "08:00:27:07:5e:92",
"module": "e1000",
"mtu": 1500,
"pciid": "0000:00:03.0",
"promisc": false,
"type": "ether"
},
"ansible_enp0s8": {
"active": true,
"device": "enp0s8",
"ipv4": {
"address": "10.100.101.111",
"broadcast": "10.100.101.255",
"netmask": "255.255.255.0",
"network": "10.100.101.0"
},
"ipv6": [
{
"address": "fe80::a00:27ff:fe63:1c7c",
"prefix": "64",
"scope": "link"
}
],
"macaddress": "08:00:27:63:1c:7c",
"module": "e1000",
"mtu": 1500,
"pciid": "0000:00:08.0",
"promisc": false,
"type": "ether"
},
"ansible_env": {
"HOME": "/home/vagrant",
"LANG": "en_US.UTF-8",
"LC_ALL": "en_US.UTF-8",
"LC_MESSAGES": "en_US.UTF-8",
"LESSOPEN": "||/usr/bin/lesspipe.sh %s",
"LOGNAME": "vagrant",
"MAIL": "/var/mail/vagrant",
"PATH": "/usr/local/bin:/usr/bin",
"PWD": "/home/vagrant",
"QT_GRAPHICSSYSTEM_CHECKED": "1",
"SELINUX_LEVEL_REQUESTED": "",
"SELINUX_ROLE_REQUESTED": "",
"SELINUX_USE_CURRENT_RANGE": "",
"SHELL": "/bin/bash",
"SHLVL": "2",
"SSH_CLIENT": "10.0.2.2 57776 22",
"SSH_CONNECTION": "10.0.2.2 57776 10.0.2.15 22",
"SSH_TTY": "/dev/pts/0",
"TERM": "xterm-256color",
"USER": "vagrant",
"XDG_RUNTIME_DIR": "/run/user/1000",
"XDG_SESSION_ID": "11",
"_": "/usr/bin/python"
},
"ansible_fips": false,
"ansible_form_factor": "Other",
"ansible_fqdn": "localhost.localdomain",
"ansible_hostname": "gl-docker",
"ansible_interfaces": [
"lo",
"docker0",
"enp0s3",
"enp0s8"
],
"ansible_kernel": "4.4.3-1.el7.elrepo.x86_64",
"ansible_lo": {
"active": true,
"device": "lo",
"ipv4": {
"address": "127.0.0.1",
"broadcast": "host",
"netmask": "255.0.0.0",
"network": "127.0.0.0"
},
"ipv6": [
{
"address": "::1",
"prefix": "128",
"scope": "host"
}
],
"mtu": 65536,
"promisc": false,
"type": "loopback"
},
"ansible_lsb": {
"codename": "Core",
"description": "CentOS Linux release 7.2.1511 (Core)",
"id": "CentOS",
"major_release": "7",
"release": "7.2.1511"
},
"ansible_machine": "x86_64",
"ansible_machine_id": "69c2e9ff6b3b4594b1a26db25287da79",
"ansible_memfree_mb": 1629,
"ansible_memory_mb": {
"nocache": {
"free": 1790,
"used": 210
},
"real": {
"free": 1629,
"total": 2000,
"used": 371
},
"swap": {
"cached": 0,
"free": 1015,
"total": 1015,
"used": 0
}
},
"ansible_memtotal_mb": 2000,
"ansible_mounts": [
{
"device": "/dev/mapper/centos-root",
"fstype": "xfs",
"mount": "/",
"options": "rw,seclabel,relatime,attr2,inode64,noquota",
"size_available": 6826946560,
"size_total": 8986296320,
"uuid": "319365c2-ad4a-4a6b-986a-0b050e96624c"
},
{
"device": "/dev/sda1",
"fstype": "xfs",
"mount": "/boot",
"options": "rw,seclabel,relatime,attr2,inode64,noquota",
"size_available": 320561152,
"size_total": 520794112,
"uuid": "9a2e2b29-a60f-484b-be03-30df74cc87a4"
}
],
"ansible_nodename": "gl-docker",
"ansible_os_family": "RedHat",
"ansible_pkg_mgr": "yum",
"ansible_processor": [
"GenuineIntel",
"Intel(R) Core(TM) i7-4578U CPU @ 3.00GHz"
],
"ansible_processor_cores": 1,
"ansible_processor_count": 1,
"ansible_processor_threads_per_core": 1,
"ansible_processor_vcpus": 1,
"ansible_product_name": "VirtualBox",
"ansible_product_serial": "NA",
"ansible_product_uuid": "NA",
"ansible_product_version": "1.2",
"ansible_python_version": "2.7.5",
"ansible_selinux": {
"config_mode": "permissive",
"mode": "permissive",
"policyvers": 30,
"status": "enabled",
"type": "targeted"
},
"ansible_service_mgr": "systemd",
"ansible_ssh_host_key_ecdsa_public": "AAAAE2VjZHNhLXNoYTItbmlzdHAyNTYAAAAIbmlzdHAyNTYAAABBBDB17PdGWfCxEJRNI1dj88lwt47Ah4eneGboTvG8V2Rctf7nGAOdLrB+//YrHPyb7mSiJp58/p+w8qkKy4sFQ38=",
"ansible_ssh_host_key_ed25519_public": "AAAAC3NzaC1lZDI1NTE5AAAAIHtYETIJC0EFGF1Mam7pYjiVxAVTNQp6yqADQoSS13bP",
"ansible_ssh_host_key_rsa_public": "AAAAB3NzaC1yc2EAAAADAQABAAABAQDJ8cfLKMpYB/mrhS3DuBTB6cTryHUzLHJ1gIk9Ro+QhbidvnAb9Br4182qytb2fmSB0kn+I4WbZBSDj2Kv9uicL8BVEyoMZtAsDZ0cHBsN5Su+Dk5mijG4PplAM8CjU6BxvhGZJ+lxuT4IMMCEMua7SmVVITaiebVbcTi7x61WsuOaKAAe+D9xaHcY0tlKvr0XaZxnyxtscbHLEc1vi6WUYlM1uLXdnMycE5DYlmsl5FfCkrtFeEG/xdbbcmBlrSl7WUwyVY0w1GPGBQFatEHf8PJhYTELNWVHR1rCMMP8Q/uiT9E/UkQHASS3XgRe+W6FqgD6pO6mOHkLXjoFVtsD",
"ansible_swapfree_mb": 1015,
"ansible_swaptotal_mb": 1015,
"ansible_system": "Linux",
"ansible_system_vendor": "innotek GmbH",
"ansible_uptime_seconds": 2173,
"ansible_user_dir": "/home/vagrant",
"ansible_user_gecos": "vagrant",
"ansible_user_gid": 1000,
"ansible_user_id": "vagrant",
"ansible_user_shell": "/bin/bash",
"ansible_user_uid": 1000,
"ansible_userspace_architecture": "x86_64",
"ansible_userspace_bits": "64",
"ansible_virtualization_role": "guest",
"ansible_virtualization_type": "virtualbox",
"module_setup": true
},
"changed": false,
"invocation": {
"module_args": {
"fact_path": "facts",
"filter": "*"
},
"module_name": "setup"
}
}
```
| main | fact gathering randomly fails on a vm issue type bug report component name setup module ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration inventory hosts host key checking false nocows fact caching jsonfile fact caching connection facts ansible managed ansible managed file modified on y m d by uid on host scp if ssh true timeout pipelining true ssh args o controlmaster auto o controlpersist os environment host ubuntu linux laptappy generic ubuntu smp thu mar utc gnu linux vm centos linux gl docker elrepo smp thu feb est gnu linux summary trying to gather facts from a newly spun up vagrant box fails i am not using the ansible vagrant provisioner i am running ansible after the fact i can connect all night and day with the connection string that ansible verbose output uses after a successful fact gathering i can run ansible commands all night and day against my node s fact gathering is the only thing that chokes randomly steps to reproduce the command i run cd ansible ansible vagrant m setup a fact path facts vvvv cd inventory regardless if i use the private key or password the results are the same gl docker ansible connection ssh ansible host ansible port ansible user vagrant ansible ssh private key file vagrant machines gl docker virtualbox private key gl docker i can run this command and everything appears to exit cleanly i copy pasted this string from the ansible verbose output for i in do ssh c vvv o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile vagrant machines gl docker virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home phil ansible cp ansible ssh h p r tt echo e nhey im in n exit echo echo i do i ve also set the following variable to successfully and repeatedly run the setup module on the host itself export ansible keep remote files expected results i expect fact gathering to return successfully each and every time actual results fact gathering randomly fails here are two subsequent fact gathering runs using opt work repos docker microservice devenv ansible ansible cfg as config file establish ssh connection for user vagrant ssh exec ssh c q o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile vagrant machines gl docker virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home phil ansible cp ansible ssh h p r tt bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp put tmp tmpwzywqd to home vagrant ansible tmp ansible tmp setup ssh exec sftp b c o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile vagrant machines gl docker virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home phil ansible cp ansible ssh h p r establish ssh connection for user vagrant ssh exec ssh c q o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile vagrant machines gl docker virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home phil ansible cp ansible ssh h p r tt bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python home vagrant ansible tmp ansible tmp setup rm rf home vagrant ansible tmp ansible tmp dev null gl docker failed changed false failed true invocation module name setup module stderr module stdout invocation module args filter fact path facts changed false ansible verbose override true ansible facts ansible product serial na ansible form factor other ansible product version ansible fips false ansible service mgr systemd ansible user id vagrant module setup true ansible memtotal mb ansible ssh host key rsa public ansible ssh host key ecdsa public p ansible distribution version ansible domain localdomain ansible user shell bin bash ansible date time weekday number basic short tz utc weeknumber hour year minute tz offset month epoch micro weekday friday time date day basic second ansible ssh host key public ansible processor cores ansible virtualization role guest ansible env lessopen usr bin lesspipe sh s ssh client selinux use current range logname vagrant user vagrant path usr local bin usr bin home home vagrant lang en us utf term xterm shell bin bash shlvl lc all en us utf xdg runtime dir run user selinux role requested qt graphicssystem checked xdg session id usr bin python lc messages en us utf ssh tty dev pts selinux level requested pwd home vagrant mail var mail vagrant ssh connection ansible processor vcpus ansible macaddress interfaces mtu device promisc false stp false broadcast global netmask network address active false type bridge id ansible bios version virtualbox ansible processor ansible virtualization type virtualbox ansible lo mtu device lo promisc false broadcast host netmask network address active true type loopback ansible userspace bits ansible architecture ansible default macaddress network mtu broadcast alias netmask address interface type ether gateway ansible swapfree mb ansible default ansible distribution release core ansible system vendor innotek gmbh ansible os family redhat ansible cmdline boot image vmlinuz elrepo quiet true rhgb true rd lvm lv centos swap crashkernel auto ro true root dev mapper centos root ansible mounts uuid size total mount size available fstype xfs device dev mapper centos root options rw seclabel relatime noquota msg module failure parsed false using opt work repos docker microservice devenv ansible ansible cfg as config file establish ssh connection for user vagrant ssh exec ssh c q o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile vagrant machines gl docker virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home phil ansible cp ansible ssh h p r tt bin sh c umask mkdir p echo home ansible tmp ansible tmp echo echo home ansible tmp ansible tmp put tmp to home vagrant ansible tmp ansible tmp setup ssh exec sftp b c o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile vagrant machines gl docker virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home phil ansible cp ansible ssh h p r establish ssh connection for user vagrant ssh exec ssh c q o controlmaster auto o controlpersist o stricthostkeychecking no o port o identityfile vagrant machines gl docker virtualbox private key o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user vagrant o connecttimeout o controlpath home phil ansible cp ansible ssh h p r tt bin sh c lang en us utf lc all en us utf lc messages en us utf usr bin python home vagrant ansible tmp ansible tmp setup rm rf home vagrant ansible tmp ansible tmp dev null gl docker success ansible facts ansible all addresses ansible all addresses ansible architecture ansible bios date ansible bios version virtualbox ansible cmdline boot image vmlinuz elrepo crashkernel auto quiet true rd lvm lv centos swap rhgb true ro true root dev mapper centos root ansible date time date day epoch hour basic basic short micro minute month second time tz utc tz offset weekday friday weekday number weeknumber year ansible default address alias broadcast gateway interface macaddress mtu netmask network type ether ansible default ansible devices sda holders host ide interface intel corporation eb mb ide rev model vbox harddisk partitions sectors sectorsize size mb start sectors sectorsize size gb start removable rotational scheduler mode deadline sectors sectorsize size gb support discard vendor ata ansible distribution centos ansible distribution major version ansible distribution release core ansible distribution version ansible dns nameservers ansible active false device id interfaces address broadcast global netmask network macaddress mtu promisc false stp false type bridge ansible domain localdomain ansible active true device address broadcast netmask network address prefix scope link macaddress module mtu pciid promisc false type ether ansible active true device address broadcast netmask network address prefix scope link macaddress module mtu pciid promisc false type ether ansible env home home vagrant lang en us utf lc all en us utf lc messages en us utf lessopen usr bin lesspipe sh s logname vagrant mail var mail vagrant path usr local bin usr bin pwd home vagrant qt graphicssystem checked selinux level requested selinux role requested selinux use current range shell bin bash shlvl ssh client ssh connection ssh tty dev pts term xterm user vagrant xdg runtime dir run user xdg session id usr bin python ansible fips false ansible form factor other ansible fqdn localhost localdomain ansible hostname gl docker ansible interfaces lo ansible kernel elrepo ansible lo active true device lo address broadcast host netmask network address prefix scope host mtu promisc false type loopback ansible lsb codename core description centos linux release core id centos major release release ansible machine ansible machine id ansible memfree mb ansible memory mb nocache free used real free total used swap cached free total used ansible memtotal mb ansible mounts device dev mapper centos root fstype xfs mount options rw seclabel relatime noquota size available size total uuid device dev fstype xfs mount boot options rw seclabel relatime noquota size available size total uuid ansible nodename gl docker ansible os family redhat ansible pkg mgr yum ansible processor genuineintel intel r core tm cpu ansible processor cores ansible processor count ansible processor threads per core ansible processor vcpus ansible product name virtualbox ansible product serial na ansible product uuid na ansible product version ansible python version ansible selinux config mode permissive mode permissive policyvers status enabled type targeted ansible service mgr systemd ansible ssh host key ecdsa public p ansible ssh host key public ansible ssh host key rsa public ansible swapfree mb ansible swaptotal mb ansible system linux ansible system vendor innotek gmbh ansible uptime seconds ansible user dir home vagrant ansible user gecos vagrant ansible user gid ansible user id vagrant ansible user shell bin bash ansible user uid ansible userspace architecture ansible userspace bits ansible virtualization role guest ansible virtualization type virtualbox module setup true changed false invocation module args fact path facts filter module name setup | 1 |
277,327 | 24,060,102,625 | IssuesEvent | 2022-09-16 21:17:18 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | provideCellStatusBarItems called on metadata change | integration-test-failure | https://dev.azure.com/monacotools/Monaco/_build/results?buildId=185525&view=logs&j=9833cdaa-2e92-5f8d-56a8-239fd18fd628&t=9bc48eac-eb66-52c8-7250-d6490c5f3a15&l=587
```
1 failing
1) statusbar
provideCellStatusBarItems called on metadata change:
Error: asPromise TIMEOUT reached
at asPromise (extensions/vscode-api-tests/src/utils.ts:128:16)
at Context.<anonymous> (extensions/vscode-api-tests/src/singlefolder-tests/notebook.api.test.ts:280:34)
at processImmediate (node:internal/timers:466:21)
``` | 1.0 | provideCellStatusBarItems called on metadata change - https://dev.azure.com/monacotools/Monaco/_build/results?buildId=185525&view=logs&j=9833cdaa-2e92-5f8d-56a8-239fd18fd628&t=9bc48eac-eb66-52c8-7250-d6490c5f3a15&l=587
```
1 failing
1) statusbar
provideCellStatusBarItems called on metadata change:
Error: asPromise TIMEOUT reached
at asPromise (extensions/vscode-api-tests/src/utils.ts:128:16)
at Context.<anonymous> (extensions/vscode-api-tests/src/singlefolder-tests/notebook.api.test.ts:280:34)
at processImmediate (node:internal/timers:466:21)
``` | non_main | providecellstatusbaritems called on metadata change failing statusbar providecellstatusbaritems called on metadata change error aspromise timeout reached at aspromise extensions vscode api tests src utils ts at context extensions vscode api tests src singlefolder tests notebook api test ts at processimmediate node internal timers | 0 |
441,127 | 12,708,566,688 | IssuesEvent | 2020-06-23 10:46:32 | geosolutions-it/MapStore2 | https://api.github.com/repos/geosolutions-it/MapStore2 | opened | Measure tool has a bug when switching type of measure | Priority: High bug | ## Description
<!-- Add here a few sentences describing the bug. -->
There are some problems when you use measure tool and switch tool while drawing the measure
## How to reproduce
<!-- A list of steps to reproduce the bug -->
- open measure tool
- start drawing, do not complete the measure
- enable bearing
- as you can see some measure tooltips remains floating on the map
- check console
*Expected Result*
<!-- Describe here the expected result -->
to reset previous setup of measure (remove tooltips etc)
*Current Result*
<!-- Describe here the current behavior -->
this

- [x] Not browser related
<details><summary> <b>Browser info</b> </summary>
<!-- If browser related, please compile the following table -->
<!-- If your browser is not in the list please add a new row to the table with the version -->
(use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users)
| Browser Affected | Version |
|---|---|
|Internet Explorer| |
|Edge| |
|Chrome| |
|Firefox| |
|Safari| |
</details>
## Other useful information
<!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
| 1.0 | Measure tool has a bug when switching type of measure - ## Description
<!-- Add here a few sentences describing the bug. -->
There are some problems when you use measure tool and switch tool while drawing the measure
## How to reproduce
<!-- A list of steps to reproduce the bug -->
- open measure tool
- start drawing, do not complete the measure
- enable bearing
- as you can see some measure tooltips remains floating on the map
- check console
*Expected Result*
<!-- Describe here the expected result -->
to reset previous setup of measure (remove tooltips etc)
*Current Result*
<!-- Describe here the current behavior -->
this

- [x] Not browser related
<details><summary> <b>Browser info</b> </summary>
<!-- If browser related, please compile the following table -->
<!-- If your browser is not in the list please add a new row to the table with the version -->
(use this site: <a href="https://www.whatsmybrowser.org/">https://www.whatsmybrowser.org/</a> for non expert users)
| Browser Affected | Version |
|---|---|
|Internet Explorer| |
|Edge| |
|Chrome| |
|Firefox| |
|Safari| |
</details>
## Other useful information
<!-- error stack trace, screenshot, videos, or link to repository code are welcome -->
| non_main | measure tool has a bug when switching type of measure description there are some problems when you use measure tool and switch tool while drawing the measure how to reproduce open measure tool start drawing do not complete the measure enable bearing as you can see some measure tooltips remains floating on the map check console expected result to reset previous setup of measure remove tooltips etc current result this not browser related browser info use this site a href for non expert users browser affected version internet explorer edge chrome firefox safari other useful information | 0 |
528,254 | 15,362,931,339 | IssuesEvent | 2021-03-01 20:08:58 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | Replace GetService with GetServiceAsync everywhere in NuGet.Clients | Pipeline:Backlog Priority:2 Product:VS.Client Tenet:Performance | Follow up on #3775.
Recommended practice of acquiring VS services is always calling `GetServiceAsync` to avoid the high risk of deadlock. All the occurrences of `GetService` and `ServiceLocator` should be replaced.
| 1.0 | Replace GetService with GetServiceAsync everywhere in NuGet.Clients - Follow up on #3775.
Recommended practice of acquiring VS services is always calling `GetServiceAsync` to avoid the high risk of deadlock. All the occurrences of `GetService` and `ServiceLocator` should be replaced.
| non_main | replace getservice with getserviceasync everywhere in nuget clients follow up on recommended practice of acquiring vs services is always calling getserviceasync to avoid the high risk of deadlock all the occurrences of getservice and servicelocator should be replaced | 0 |
193,061 | 22,216,046,493 | IssuesEvent | 2022-06-08 01:50:21 | dpteam/RK3188_TABLET | https://api.github.com/repos/dpteam/RK3188_TABLET | reopened | CVE-2012-6548 (Low) detected in linux-yocto-4.12v3.1.10 | security vulnerability | ## CVE-2012-6548 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.12v3.1.10</b></p></summary>
<p>
<p>Linux 4.12 Embedded Kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.12>https://git.yoctoproject.org/git/linux-yocto-4.12</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The udf_encode_fh function in fs/udf/namei.c in the Linux kernel before 3.6 does not initialize a certain structure member, which allows local users to obtain sensitive information from kernel heap memory via a crafted application.
<p>Publish Date: 2013-03-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6548>CVE-2012-6548</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6548">https://nvd.nist.gov/vuln/detail/CVE-2012-6548</a></p>
<p>Release Date: 2013-03-15</p>
<p>Fix Resolution: 3.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2012-6548 (Low) detected in linux-yocto-4.12v3.1.10 - ## CVE-2012-6548 - Low Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-4.12v3.1.10</b></p></summary>
<p>
<p>Linux 4.12 Embedded Kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-4.12>https://git.yoctoproject.org/git/linux-yocto-4.12</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (0)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/low_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The udf_encode_fh function in fs/udf/namei.c in the Linux kernel before 3.6 does not initialize a certain structure member, which allows local users to obtain sensitive information from kernel heap memory via a crafted application.
<p>Publish Date: 2013-03-15
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2012-6548>CVE-2012-6548</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>2.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6548">https://nvd.nist.gov/vuln/detail/CVE-2012-6548</a></p>
<p>Release Date: 2013-03-15</p>
<p>Fix Resolution: 3.6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve low detected in linux yocto cve low severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details the udf encode fh function in fs udf namei c in the linux kernel before does not initialize a certain structure member which allows local users to obtain sensitive information from kernel heap memory via a crafted application publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
35,214 | 2,789,839,060 | IssuesEvent | 2015-05-08 21:49:47 | google/google-visualization-api-issues | https://api.github.com/repos/google/google-visualization-api-issues | opened | document.getElementById("google.visualiation.MotionChart-0") is null | Priority-Low Type-Enhancement | Original [issue 261](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=261) created by orwant on 2010-04-21T16:36:29.000Z:
<b>What would you like to see us add to this API?</b>
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| 1.0 | document.getElementById("google.visualiation.MotionChart-0") is null - Original [issue 261](https://code.google.com/p/google-visualization-api-issues/issues/detail?id=261) created by orwant on 2010-04-21T16:36:29.000Z:
<b>What would you like to see us add to this API?</b>
<b>What component is this issue related to (PieChart, LineChart, DataTable,</b>
<b>Query, etc)?</b>
<b>*********************************************************</b>
<b>For developers viewing this issue: please click the 'star' icon to be</b>
<b>notified of future changes, and to let us know how many of you are</b>
<b>interested in seeing it resolved.</b>
<b>*********************************************************</b>
| non_main | document getelementbyid google visualiation motionchart is null original created by orwant on what would you like to see us add to this api what component is this issue related to piechart linechart datatable query etc for developers viewing this issue please click the star icon to be notified of future changes and to let us know how many of you are interested in seeing it resolved | 0 |
94,421 | 27,188,750,470 | IssuesEvent | 2023-02-19 14:45:02 | MustardChef/WSABuilds | https://api.github.com/repos/MustardChef/WSABuilds | closed | WSA with BusyBox and Xposed Framework | Question Build Request | Hy, can you make a build with Busybox and Xposed framework already installed, I tried to install busybox normaly but the WSA crashes when i press the Install button, also tried to manualy create the xbin folder in the system folder using a root explorer and then copy the binaries but the WSA crashes again. It may be related to the way that the system handles the system storage because the root explorer shows 0.00mb free in the root subfolders and busybox says also that there is no space available for install | 1.0 | WSA with BusyBox and Xposed Framework - Hy, can you make a build with Busybox and Xposed framework already installed, I tried to install busybox normaly but the WSA crashes when i press the Install button, also tried to manualy create the xbin folder in the system folder using a root explorer and then copy the binaries but the WSA crashes again. It may be related to the way that the system handles the system storage because the root explorer shows 0.00mb free in the root subfolders and busybox says also that there is no space available for install | non_main | wsa with busybox and xposed framework hy can you make a build with busybox and xposed framework already installed i tried to install busybox normaly but the wsa crashes when i press the install button also tried to manualy create the xbin folder in the system folder using a root explorer and then copy the binaries but the wsa crashes again it may be related to the way that the system handles the system storage because the root explorer shows free in the root subfolders and busybox says also that there is no space available for install | 0 |
278,377 | 24,150,232,321 | IssuesEvent | 2022-09-21 23:21:29 | stratum-mining/stratum | https://api.github.com/repos/stratum-mining/stratum | opened | Fix broken interop-test in test.yaml GHA for macos | test important | `interop-test` GitHub actions job in `test.yaml` does not pass for macos os.
Logic to skip macos and just run ubuntu is added [here](hhttps://github.com/stratum-mining/stratum/commit/6c7b105f73554c091117d4d7b79af60146ce8f83) in #71. | 1.0 | Fix broken interop-test in test.yaml GHA for macos - `interop-test` GitHub actions job in `test.yaml` does not pass for macos os.
Logic to skip macos and just run ubuntu is added [here](hhttps://github.com/stratum-mining/stratum/commit/6c7b105f73554c091117d4d7b79af60146ce8f83) in #71. | non_main | fix broken interop test in test yaml gha for macos interop test github actions job in test yaml does not pass for macos os logic to skip macos and just run ubuntu is added h in | 0 |
2,989 | 10,820,516,389 | IssuesEvent | 2019-11-08 16:33:00 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | closed | Contrib Group Application - Views Selective Filters | Maintainer application | Hiya! Looking to become an official Backdrop contributor. I run [Mustardseed Media](http://mustardseedmedia.com) full time and I'm currently in the transition for my clients from D7 over to Backdrop. Apparently [I've been using Drupal (as Rob_Feature) for 13 years and 10 months](https://www.drupal.org/u/rob_feature). I've tried to contribute there as much as possible, but my contributions come more in the way of training and testing (although I've contributed a few things).
Anyway...I'm in the process of launching some new sites on Backdrop and finally had the opportunity of porting my first module: [Views Selective Filters](https://github.com/bobchristenson/views_selective_filters) (hopefully the first of many).
I'm not a hardcore developer, but certainly I know my way around the ecosystem (still learning differences between D7 and Backdrop). So, I'm always looking for help/advice for any projects I may port/develop in the future.
Let me know where we go from here! | True | Contrib Group Application - Views Selective Filters - Hiya! Looking to become an official Backdrop contributor. I run [Mustardseed Media](http://mustardseedmedia.com) full time and I'm currently in the transition for my clients from D7 over to Backdrop. Apparently [I've been using Drupal (as Rob_Feature) for 13 years and 10 months](https://www.drupal.org/u/rob_feature). I've tried to contribute there as much as possible, but my contributions come more in the way of training and testing (although I've contributed a few things).
Anyway...I'm in the process of launching some new sites on Backdrop and finally had the opportunity of porting my first module: [Views Selective Filters](https://github.com/bobchristenson/views_selective_filters) (hopefully the first of many).
I'm not a hardcore developer, but certainly I know my way around the ecosystem (still learning differences between D7 and Backdrop). So, I'm always looking for help/advice for any projects I may port/develop in the future.
Let me know where we go from here! | main | contrib group application views selective filters hiya looking to become an official backdrop contributor i run full time and i m currently in the transition for my clients from over to backdrop apparently i ve tried to contribute there as much as possible but my contributions come more in the way of training and testing although i ve contributed a few things anyway i m in the process of launching some new sites on backdrop and finally had the opportunity of porting my first module hopefully the first of many i m not a hardcore developer but certainly i know my way around the ecosystem still learning differences between and backdrop so i m always looking for help advice for any projects i may port develop in the future let me know where we go from here | 1 |
332,750 | 10,110,165,038 | IssuesEvent | 2019-07-30 09:39:02 | wso2/product-microgateway | https://api.github.com/repos/wso2/product-microgateway | opened | [toolkit] Error building when interceptor folder is missing | Priority/Normal Type/Bug | ### Description:
Following error occurs when a project without interceptor folder is build.
`org.wso2.apimgt.gateway.cli.exception.CLIInternalException: Error occurred while generating source code for the open API definitions.
at org.wso2.apimgt.gateway.cli.cmd.BuildCmd.execute(BuildCmd.java:120)
at java.util.Optional.ifPresent(Optional.java:159)
at org.wso2.apimgt.gateway.cli.cmd.Main.main(Main.java:48)
Caused by: java.nio.file.NoSuchFileException: /var/jenkins_home/workspace/MicroGatewayPipeline/MGW-Projects/bookstore/interceptors
at sun.nio.fs.UnixException.translateToIOException(UnixException.java:86)
at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102)
at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107)
at sun.nio.fs.UnixCopyFile.copy(UnixCopyFile.java:526)
at sun.nio.fs.UnixFileSystemProvider.copy(UnixFileSystemProvider.java:253)
at java.nio.file.Files.copy(Files.java:1274)
at org.wso2.apimgt.gateway.cli.utils.GatewayCmdUtils.copyFolder(GatewayCmdUtils.java:686)
at org.wso2.apimgt.gateway.cli.utils.GatewayCmdUtils.copyAndReplaceFolder(GatewayCmdUtils.java:635)
at org.wso2.apimgt.gateway.cli.cmd.BuildCmd.execute(BuildCmd.java:108)
... 2 more
`
### Steps to reproduce:
This happens when we check out project from git hub any other VCS.Because interceptor folder is empty, those folder will not be committed. Hence upon checkout and building it gives the error
### Affected Product Version:
3.0.1
### OS, DB, other environment details and versions:
- OS:
- Client:
- DB:
- Env (Docker/K8s):
---
### Optional Fields
#### Related Issues:
https://github.com/wso2/product-microgateway/issues/714
#### Suggested Labels:
<!--Only to be used by non-committers-->
#### Suggested Assignees:
<!--Only to be used by non-committers-->
| 1.0 | [toolkit] Error building when interceptor folder is missing - ### Description:
Following error occurs when a project without interceptor folder is build.
`org.wso2.apimgt.gateway.cli.exception.CLIInternalException: Error occurred while generating source code for the open API definitions.
at org.wso2.apimgt.gateway.cli.cmd.BuildCmd.execute(BuildCmd.java:120)
at java.util.Optional.ifPresent(Optional.java:159)
at org.wso2.apimgt.gateway.cli.cmd.Main.main(Main.java:48)
Caused by: java.nio.file.NoSuchFileException: /var/jenkins_home/workspace/MicroGatewayPipeline/MGW-Projects/bookstore/interceptors
at sun.nio.fs.UnixException.translateToIOException(UnixException.java:86)
at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:102)
at sun.nio.fs.UnixException.rethrowAsIOException(UnixException.java:107)
at sun.nio.fs.UnixCopyFile.copy(UnixCopyFile.java:526)
at sun.nio.fs.UnixFileSystemProvider.copy(UnixFileSystemProvider.java:253)
at java.nio.file.Files.copy(Files.java:1274)
at org.wso2.apimgt.gateway.cli.utils.GatewayCmdUtils.copyFolder(GatewayCmdUtils.java:686)
at org.wso2.apimgt.gateway.cli.utils.GatewayCmdUtils.copyAndReplaceFolder(GatewayCmdUtils.java:635)
at org.wso2.apimgt.gateway.cli.cmd.BuildCmd.execute(BuildCmd.java:108)
... 2 more
`
### Steps to reproduce:
This happens when we check out project from git hub any other VCS.Because interceptor folder is empty, those folder will not be committed. Hence upon checkout and building it gives the error
### Affected Product Version:
3.0.1
### OS, DB, other environment details and versions:
- OS:
- Client:
- DB:
- Env (Docker/K8s):
---
### Optional Fields
#### Related Issues:
https://github.com/wso2/product-microgateway/issues/714
#### Suggested Labels:
<!--Only to be used by non-committers-->
#### Suggested Assignees:
<!--Only to be used by non-committers-->
| non_main | error building when interceptor folder is missing description following error occurs when a project without interceptor folder is build org apimgt gateway cli exception cliinternalexception error occurred while generating source code for the open api definitions at org apimgt gateway cli cmd buildcmd execute buildcmd java at java util optional ifpresent optional java at org apimgt gateway cli cmd main main main java caused by java nio file nosuchfileexception var jenkins home workspace microgatewaypipeline mgw projects bookstore interceptors at sun nio fs unixexception translatetoioexception unixexception java at sun nio fs unixexception rethrowasioexception unixexception java at sun nio fs unixexception rethrowasioexception unixexception java at sun nio fs unixcopyfile copy unixcopyfile java at sun nio fs unixfilesystemprovider copy unixfilesystemprovider java at java nio file files copy files java at org apimgt gateway cli utils gatewaycmdutils copyfolder gatewaycmdutils java at org apimgt gateway cli utils gatewaycmdutils copyandreplacefolder gatewaycmdutils java at org apimgt gateway cli cmd buildcmd execute buildcmd java more steps to reproduce this happens when we check out project from git hub any other vcs because interceptor folder is empty those folder will not be committed hence upon checkout and building it gives the error affected product version os db other environment details and versions os client db env docker optional fields related issues suggested labels suggested assignees | 0 |
235,818 | 18,061,148,283 | IssuesEvent | 2021-09-20 14:06:46 | ita-social-projects/dokazovi-requirements | https://api.github.com/repos/ita-social-projects/dokazovi-requirements | opened | [Test for Story #189 ]Verify if the system shows the admin a warning message if the user don't put video in field 'Відео' and press button 'Опублікувати' | documentation test case | *Story link**
[#189 Story](https://github.com/ita-social-projects/dokazovi-requirements/issues/189)
### Status:
Not executed
### Title:
erify if the system shows the admin a warning message if the user don't put video in field 'Відео' and press button 'Опублікувати'
### Description:
### Pre-conditions:
1.Go to the web application as an admin.
2.Should be opened tab 'Відео'.
Step № | Test Steps | Test data | Expected result | Status (Pass/Fail/Not executed) | Notes
------------ | ------------ | ------------ | ------------ | ------------ | ------------
1 | Fill all field but in field 'Відео' don't put video| | User can fill fields | Not executed|Додайте, будь ласка, відео
2 | Press 'Опублікувати'| | Showing an error 'Додайте, будь ласка, відео'| Not Executed|
### Dependencies:
[#189 Story](https://github.com/ita-social-projects/dokazovi-requirements/issues/189)
### [Gantt Chart](https://docs.google.com/spreadsheets/d/1bgaEJDOf3OhfNRfP-WWPKmmZFW5C3blOUxamE3wSCbM/edit#gid=775577959)
| 1.0 | [Test for Story #189 ]Verify if the system shows the admin a warning message if the user don't put video in field 'Відео' and press button 'Опублікувати' - *Story link**
[#189 Story](https://github.com/ita-social-projects/dokazovi-requirements/issues/189)
### Status:
Not executed
### Title:
erify if the system shows the admin a warning message if the user don't put video in field 'Відео' and press button 'Опублікувати'
### Description:
### Pre-conditions:
1.Go to the web application as an admin.
2.Should be opened tab 'Відео'.
Step № | Test Steps | Test data | Expected result | Status (Pass/Fail/Not executed) | Notes
------------ | ------------ | ------------ | ------------ | ------------ | ------------
1 | Fill all field but in field 'Відео' don't put video| | User can fill fields | Not executed|Додайте, будь ласка, відео
2 | Press 'Опублікувати'| | Showing an error 'Додайте, будь ласка, відео'| Not Executed|
### Dependencies:
[#189 Story](https://github.com/ita-social-projects/dokazovi-requirements/issues/189)
### [Gantt Chart](https://docs.google.com/spreadsheets/d/1bgaEJDOf3OhfNRfP-WWPKmmZFW5C3blOUxamE3wSCbM/edit#gid=775577959)
| non_main | verify if the system shows the admin a warning message if the user don t put video in field відео and press button опублікувати story link status not executed title erify if the system shows the admin a warning message if the user don t put video in field відео and press button опублікувати description pre conditions go to the web application as an admin should be opened tab відео step № test steps test data expected result status pass fail not executed notes fill all field but in field відео don t put video user can fill fields not executed додайте будь ласка відео press опублікувати showing an error додайте будь ласка відео not executed dependencies | 0 |
1,127 | 4,997,919,320 | IssuesEvent | 2016-12-09 18:11:47 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Add state=query option for OS Packaging Modules | affects_2.1 feature_idea waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
- yum
- apt
- dnf (but this is in the ansible-modules-extras repo)
##### ANSIBLE VERSION
```
ansible 2.1.0.0
```
##### OS / ENVIRONMENT
<!----->
- Managing Linux
##### SUMMARY
<!--- Explain the problem briefly -->
I would like to query for the installation status of a particular package without actually changing the state of the package on the machine. This would allow one to register a variable to act upon in some other way.
It seems the way this is generally done is to run the shell module and execute commands to determine whether a package is installed. For example:
```
shell: rpm -q <package>
register: pkg_query
changed_when: false
```
Instead, I would like to use the OS Packaging Modules to query for the installation status of a particular package. This would be like adding a `state: query` option to some of the OS Packaging Modules which would then allow one to register a variable to capture the result. This would also help cleanup repetitive tasks just to handle different package managers: `apt`, `yum` and `dnf` are the ones that come to mind where this would be useful.
<!---##### STEPS TO REPRODUCE-->
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- You can also paste gist.github.com links for larger files -->
<!---##### EXPECTED RESULTS-->
<!--- What did you expect to happen when running the steps above? -->
<!---##### ACTUAL RESULTS-->
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
| True | Add state=query option for OS Packaging Modules - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
- yum
- apt
- dnf (but this is in the ansible-modules-extras repo)
##### ANSIBLE VERSION
```
ansible 2.1.0.0
```
##### OS / ENVIRONMENT
<!----->
- Managing Linux
##### SUMMARY
<!--- Explain the problem briefly -->
I would like to query for the installation status of a particular package without actually changing the state of the package on the machine. This would allow one to register a variable to act upon in some other way.
It seems the way this is generally done is to run the shell module and execute commands to determine whether a package is installed. For example:
```
shell: rpm -q <package>
register: pkg_query
changed_when: false
```
Instead, I would like to use the OS Packaging Modules to query for the installation status of a particular package. This would be like adding a `state: query` option to some of the OS Packaging Modules which would then allow one to register a variable to capture the result. This would also help cleanup repetitive tasks just to handle different package managers: `apt`, `yum` and `dnf` are the ones that come to mind where this would be useful.
<!---##### STEPS TO REPRODUCE-->
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- You can also paste gist.github.com links for larger files -->
<!---##### EXPECTED RESULTS-->
<!--- What did you expect to happen when running the steps above? -->
<!---##### ACTUAL RESULTS-->
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
| main | add state query option for os packaging modules issue type feature idea component name yum apt dnf but this is in the ansible modules extras repo ansible version ansible os environment managing linux summary i would like to query for the installation status of a particular package without actually changing the state of the package on the machine this would allow one to register a variable to act upon in some other way it seems the way this is generally done is to run the shell module and execute commands to determine whether a package is installed for example shell rpm q register pkg query changed when false instead i would like to use the os packaging modules to query for the installation status of a particular package this would be like adding a state query option to some of the os packaging modules which would then allow one to register a variable to capture the result this would also help cleanup repetitive tasks just to handle different package managers apt yum and dnf are the ones that come to mind where this would be useful for bugs show exactly how to reproduce the problem for new features show how the feature would be used | 1 |
442,128 | 30,818,403,532 | IssuesEvent | 2023-08-01 14:47:53 | iodepo/odis-arch | https://api.github.com/repos/iodepo/odis-arch | opened | Documentation focus | documentation | Creating a generic issue for holding some goals on the reworking of the documentation
- [ ] Need a simple single page that defines the requirements for partners to join; ODIS Cat entry, sitemap.xml, etc
- [ ] Perhaps look at the initial work (very basic) for partner on boarding and the decision tree in https://github.com/iodepo/odis-arch/tree/master/docs/decisiontree
- [ ] update the SHACL validation section to focus on the use for alignment to the search
We should also review all the current issues tagged with documentation: https://github.com/iodepo/odis-arch/labels/documentation | 1.0 | Documentation focus - Creating a generic issue for holding some goals on the reworking of the documentation
- [ ] Need a simple single page that defines the requirements for partners to join; ODIS Cat entry, sitemap.xml, etc
- [ ] Perhaps look at the initial work (very basic) for partner on boarding and the decision tree in https://github.com/iodepo/odis-arch/tree/master/docs/decisiontree
- [ ] update the SHACL validation section to focus on the use for alignment to the search
We should also review all the current issues tagged with documentation: https://github.com/iodepo/odis-arch/labels/documentation | non_main | documentation focus creating a generic issue for holding some goals on the reworking of the documentation need a simple single page that defines the requirements for partners to join odis cat entry sitemap xml etc perhaps look at the initial work very basic for partner on boarding and the decision tree in update the shacl validation section to focus on the use for alignment to the search we should also review all the current issues tagged with documentation | 0 |
4,996 | 25,709,013,064 | IssuesEvent | 2022-12-07 04:22:05 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | The event source and function provided mapping already exists. Please update or delete the existing mapping with UUID | area/cloudformation stage/needs-investigation maintainer/need-followup | **Description:**
I am trying to implement Dynamodb streams and a trigger for lambda
```
myFunction:
Type: AWS::Serverless::Function
Properties:
Policies:
- AWSLambdaBasicExecutionRole
- DynamoDBCrudPolicy:
TableName: !Ref TableName
- DynamoDBStreamReadPolicy:
TableName: !Ref TableName
StreamName:
!Select
- 2
- !Split
- "/"
- Fn::ImportValue:
!Sub ${TableName}-streams-arn
CodeUri: bin/
Handler: myFunctions
Events:
dynamodb:
Type: DynamoDB
Properties:
Stream:
Fn::ImportValue: !Sub ${TableName}-streams-arn
StartingPosition: LATEST
BatchSize: 10
```
The first deployment was successful, however after adding the `DynamoDBStreamReadPolicy` and changed `StartingPosition` from `TRIM_HORIZON` to `LATEST` I got the following error:
```
The event source arn (" arn:aws:dynamodb:.../stream/... ") and function (" myFunction-...") provided mapping already exists. Please update or delete the existing mapping with UUID xxxx-xxxx-xxx-xxx (Service: AWSLambda; Status Code: 409; Error Code: ResourceConflictException; Request ID: xxxx-xxxxxxx-xxxx-xxx)
```
**Observed result:**
Error
**Expected result:**
Probably error, but at this point I did not find any explanation
PS: **as my understanding once the event mapping is created must be replaced or deleted?** If yes why it says *"Please update or delete"*? If now I want to change the mapping configuration do I need to remove it, deploy, add it again and redeploy?
Thanks | True | The event source and function provided mapping already exists. Please update or delete the existing mapping with UUID - **Description:**
I am trying to implement Dynamodb streams and a trigger for lambda
```
myFunction:
Type: AWS::Serverless::Function
Properties:
Policies:
- AWSLambdaBasicExecutionRole
- DynamoDBCrudPolicy:
TableName: !Ref TableName
- DynamoDBStreamReadPolicy:
TableName: !Ref TableName
StreamName:
!Select
- 2
- !Split
- "/"
- Fn::ImportValue:
!Sub ${TableName}-streams-arn
CodeUri: bin/
Handler: myFunctions
Events:
dynamodb:
Type: DynamoDB
Properties:
Stream:
Fn::ImportValue: !Sub ${TableName}-streams-arn
StartingPosition: LATEST
BatchSize: 10
```
The first deployment was successful, however after adding the `DynamoDBStreamReadPolicy` and changed `StartingPosition` from `TRIM_HORIZON` to `LATEST` I got the following error:
```
The event source arn (" arn:aws:dynamodb:.../stream/... ") and function (" myFunction-...") provided mapping already exists. Please update or delete the existing mapping with UUID xxxx-xxxx-xxx-xxx (Service: AWSLambda; Status Code: 409; Error Code: ResourceConflictException; Request ID: xxxx-xxxxxxx-xxxx-xxx)
```
**Observed result:**
Error
**Expected result:**
Probably error, but at this point I did not find any explanation
PS: **as my understanding once the event mapping is created must be replaced or deleted?** If yes why it says *"Please update or delete"*? If now I want to change the mapping configuration do I need to remove it, deploy, add it again and redeploy?
Thanks | main | the event source and function provided mapping already exists please update or delete the existing mapping with uuid description i am trying to implement dynamodb streams and a trigger for lambda myfunction type aws serverless function properties policies awslambdabasicexecutionrole dynamodbcrudpolicy tablename ref tablename dynamodbstreamreadpolicy tablename ref tablename streamname select split fn importvalue sub tablename streams arn codeuri bin handler myfunctions events dynamodb type dynamodb properties stream fn importvalue sub tablename streams arn startingposition latest batchsize the first deployment was successful however after adding the dynamodbstreamreadpolicy and changed startingposition from trim horizon to latest i got the following error the event source arn arn aws dynamodb stream and function myfunction provided mapping already exists please update or delete the existing mapping with uuid xxxx xxxx xxx xxx service awslambda status code error code resourceconflictexception request id xxxx xxxxxxx xxxx xxx observed result error expected result probably error but at this point i did not find any explanation ps as my understanding once the event mapping is created must be replaced or deleted if yes why it says please update or delete if now i want to change the mapping configuration do i need to remove it deploy add it again and redeploy thanks | 1 |
1,223 | 5,217,706,301 | IssuesEvent | 2017-01-26 14:42:28 | duckduckgo/zeroclickinfo-goodies | https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies | closed | Conversions: Convert to inches | Maintainer Approved Status: PR Received Suggestion | Example:
**5'7" in inches**
[https://duckduckgo.com/?q=5%277%22+in+inches ](https://duckduckgo.com/?q=5%277%22+in+inches )
_"Google will display the conversion right at the top, but duckduckgo doesn’t."_
<img width="823" alt="screen shot 2017-01-21 at 10 09 21 am" src="https://cloud.githubusercontent.com/assets/126358/22175386/b29b3746-dfc1-11e6-9e16-ccf32f1eda84.png">
------
IA Page: http://duck.co/ia/view/conversions
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mintsoft | True | Conversions: Convert to inches - Example:
**5'7" in inches**
[https://duckduckgo.com/?q=5%277%22+in+inches ](https://duckduckgo.com/?q=5%277%22+in+inches )
_"Google will display the conversion right at the top, but duckduckgo doesn’t."_
<img width="823" alt="screen shot 2017-01-21 at 10 09 21 am" src="https://cloud.githubusercontent.com/assets/126358/22175386/b29b3746-dfc1-11e6-9e16-ccf32f1eda84.png">
------
IA Page: http://duck.co/ia/view/conversions
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @mintsoft | main | conversions convert to inches example in inches google will display the conversion right at the top but duckduckgo doesn’t img width alt screen shot at am src ia page mintsoft | 1 |
303,548 | 26,216,247,188 | IssuesEvent | 2023-01-04 11:12:09 | saleor/saleor-dashboard | https://api.github.com/repos/saleor/saleor-dashboard | closed | Create missing tests for staff users | automation tests |
As a user I should not be able to create staff member with not unique email - SALEOR_3508
As an staff user I should be able to change my password in account settings - SALEOR_3510
As an admin I want to update staff member - SALEOR_3507
As an admin I should not be able to update staff member with not unique email - SALEOR_3509
https://docs.google.com/spreadsheets/d/1XQ1-babjTVCsKpuDwU91kDt3C-80TJNJ/edit#gid=906353335 | 1.0 | Create missing tests for staff users -
As a user I should not be able to create staff member with not unique email - SALEOR_3508
As an staff user I should be able to change my password in account settings - SALEOR_3510
As an admin I want to update staff member - SALEOR_3507
As an admin I should not be able to update staff member with not unique email - SALEOR_3509
https://docs.google.com/spreadsheets/d/1XQ1-babjTVCsKpuDwU91kDt3C-80TJNJ/edit#gid=906353335 | non_main | create missing tests for staff users as a user i should not be able to create staff member with not unique email saleor as an staff user i should be able to change my password in account settings saleor as an admin i want to update staff member saleor as an admin i should not be able to update staff member with not unique email saleor | 0 |
444,411 | 12,811,771,768 | IssuesEvent | 2020-07-04 01:13:30 | afup/back-office | https://api.github.com/repos/afup/back-office | closed | Core/Design | Refactoring priority 1 | Alors pour le design, j'ai trois propositions :
- http://www.cliptheme.com/demo/rapido/
- http://hubancreative.com/projects/templates/presenter/?theme=coco3
- http://bigmedved.ru/orb/
J'ai abandonné braincrafted, pas intéressant je trouve
| 1.0 | Core/Design - Alors pour le design, j'ai trois propositions :
- http://www.cliptheme.com/demo/rapido/
- http://hubancreative.com/projects/templates/presenter/?theme=coco3
- http://bigmedved.ru/orb/
J'ai abandonné braincrafted, pas intéressant je trouve
| non_main | core design alors pour le design j ai trois propositions j ai abandonné braincrafted pas intéressant je trouve | 0 |
1,399 | 6,025,396,771 | IssuesEvent | 2017-06-08 08:35:48 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | win_acl module cannot modify ACLs for registry keys | affects_2.3 feature_idea waiting_on_maintainer windows | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_acl module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.3.0 (devel 05aed6e52e) last updated 2016/10/18 19:10:43 (GMT -700)
lib/ansible/modules/core: (detached HEAD b59b5d36e0) last updated 2016/10/18 19:10:47 (GMT -700)
lib/ansible/modules/extras: (detached HEAD 3f77bb6857) last updated 2016/10/18 19:10:51 (GMT -700)
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Control Node: Fedora 23
Host Machine: AWS Windows 2012 Member Server
##### SUMMARY
<!--- Explain the problem briefly -->
The win_acl module currently does not support modification of ACLs for registry keys - only for folders and files. There is not currently a module to modify ACLs for registry keys, which are used for customization/hardening of Windows OSes. This is needed for an ongoing Windows STIG Remediation engagement.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
run the following tasks below
<!--- Paste example playbooks or commands between quotes below -->
```
- name: "V-39331 dummy"
win_acl:
path: '\Windows\SYSVOL'
user: 'BUILTIN\Administrators'
rights: 'FullControl'
type: 'allow'
state: 'present'
inherit: 'ContainerInherit, ObjectInherit'
propogation: 'InheritOnly'
- name: "V-1152 dummy"
win_acl:
path: 'HKLM:\SYSTEM\CurrentControlSet\Control\SecurePipeServers\Winreg\'
user: 'BUILTIN\Administrators'
rights: 'FullControl'
type: 'allow'
state: 'present'
inherit: 'ContainerInherit, ObjectInherit'
propagation: 'None'
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The task list runs cleanly, successfully updating each ACE.
[ansiblePatchedModules.zip](https://github.com/ansible/ansible-modules-extras/files/537808/ansiblePatchedModules.zip)
[ansiblePatchedModules.zip](https://github.com/ansible/ansible-modules-extras/files/537809/ansiblePatchedModules.zip)
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
The first task runs fine because it is a folder, but the second task fails because it is a registry key. This is because under the hood permissions for registry keys are set with "New-Object System.Security.AccessControl.RegistryAccessRule" while folder and files are set with "New-Object System.Security.AccessControl.FileSystemAccessRule." Incorporating this functionality into the win_acl.ps1 and win_acl_inheritance.ps1 would involve patching both modules. I have attached _very_ basic patches so that the task list provided above works for both tasks.
<!--- Paste verbatim command output between quotes below -->
```
fatal: [54.71.248.237]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "win_acl"
},
"msg": "an error occured when attempting to present FullControl permission(s) on HKLM:\\SYSTEM\\CurrentControlSet\\Control\\SecurePipeServers\\Winreg\\ for BUILTIN\\Administrators"
}
```
| True | win_acl module cannot modify ACLs for registry keys - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_acl module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.3.0 (devel 05aed6e52e) last updated 2016/10/18 19:10:43 (GMT -700)
lib/ansible/modules/core: (detached HEAD b59b5d36e0) last updated 2016/10/18 19:10:47 (GMT -700)
lib/ansible/modules/extras: (detached HEAD 3f77bb6857) last updated 2016/10/18 19:10:51 (GMT -700)
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Control Node: Fedora 23
Host Machine: AWS Windows 2012 Member Server
##### SUMMARY
<!--- Explain the problem briefly -->
The win_acl module currently does not support modification of ACLs for registry keys - only for folders and files. There is not currently a module to modify ACLs for registry keys, which are used for customization/hardening of Windows OSes. This is needed for an ongoing Windows STIG Remediation engagement.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
run the following tasks below
<!--- Paste example playbooks or commands between quotes below -->
```
- name: "V-39331 dummy"
win_acl:
path: '\Windows\SYSVOL'
user: 'BUILTIN\Administrators'
rights: 'FullControl'
type: 'allow'
state: 'present'
inherit: 'ContainerInherit, ObjectInherit'
propogation: 'InheritOnly'
- name: "V-1152 dummy"
win_acl:
path: 'HKLM:\SYSTEM\CurrentControlSet\Control\SecurePipeServers\Winreg\'
user: 'BUILTIN\Administrators'
rights: 'FullControl'
type: 'allow'
state: 'present'
inherit: 'ContainerInherit, ObjectInherit'
propagation: 'None'
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The task list runs cleanly, successfully updating each ACE.
[ansiblePatchedModules.zip](https://github.com/ansible/ansible-modules-extras/files/537808/ansiblePatchedModules.zip)
[ansiblePatchedModules.zip](https://github.com/ansible/ansible-modules-extras/files/537809/ansiblePatchedModules.zip)
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
The first task runs fine because it is a folder, but the second task fails because it is a registry key. This is because under the hood permissions for registry keys are set with "New-Object System.Security.AccessControl.RegistryAccessRule" while folder and files are set with "New-Object System.Security.AccessControl.FileSystemAccessRule." Incorporating this functionality into the win_acl.ps1 and win_acl_inheritance.ps1 would involve patching both modules. I have attached _very_ basic patches so that the task list provided above works for both tasks.
<!--- Paste verbatim command output between quotes below -->
```
fatal: [54.71.248.237]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_name": "win_acl"
},
"msg": "an error occured when attempting to present FullControl permission(s) on HKLM:\\SYSTEM\\CurrentControlSet\\Control\\SecurePipeServers\\Winreg\\ for BUILTIN\\Administrators"
}
```
| main | win acl module cannot modify acls for registry keys issue type feature idea component name win acl module ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific control node fedora host machine aws windows member server summary the win acl module currently does not support modification of acls for registry keys only for folders and files there is not currently a module to modify acls for registry keys which are used for customization hardening of windows oses this is needed for an ongoing windows stig remediation engagement steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used run the following tasks below name v dummy win acl path windows sysvol user builtin administrators rights fullcontrol type allow state present inherit containerinherit objectinherit propogation inheritonly name v dummy win acl path hklm system currentcontrolset control securepipeservers winreg user builtin administrators rights fullcontrol type allow state present inherit containerinherit objectinherit propagation none expected results the task list runs cleanly successfully updating each ace actual results the first task runs fine because it is a folder but the second task fails because it is a registry key this is because under the hood permissions for registry keys are set with new object system security accesscontrol registryaccessrule while folder and files are set with new object system security accesscontrol filesystemaccessrule incorporating this functionality into the win acl and win acl inheritance would involve patching both modules i have attached very basic patches so that the task list provided above works for both tasks fatal failed changed false failed true invocation module name win acl msg an error occured when attempting to present fullcontrol permission s on hklm system currentcontrolset control securepipeservers winreg for builtin administrators | 1 |
89,264 | 17,850,633,395 | IssuesEvent | 2021-09-04 02:00:24 | BeccaLyria/discord-bot | https://api.github.com/repos/BeccaLyria/discord-bot | closed | [FEAT] - Slime Command | 🟩 priority: low 🏁 status: ready for dev ⭐ goal: addition 💻 aspect: code help wanted | ### What feature would you like to see?
Need to pull the slime command from Rosalia and add it to Becca. Command should be added as a `misc` slash command, and rather than setting the member's nickname should return an embed with the suggestion instead.
### Additional information
_No response_ | 1.0 | [FEAT] - Slime Command - ### What feature would you like to see?
Need to pull the slime command from Rosalia and add it to Becca. Command should be added as a `misc` slash command, and rather than setting the member's nickname should return an embed with the suggestion instead.
### Additional information
_No response_ | non_main | slime command what feature would you like to see need to pull the slime command from rosalia and add it to becca command should be added as a misc slash command and rather than setting the member s nickname should return an embed with the suggestion instead additional information no response | 0 |
254,595 | 21,797,347,682 | IssuesEvent | 2022-05-15 20:34:32 | mitmproxy/mitmproxy | https://api.github.com/repos/mitmproxy/mitmproxy | opened | PytestUnraisableExceptionWarning: RuntimeError: Event loop is closed | kind/bug area/tests | #### Problem Description
`test_proxyserver:test_dns` currently causes a `PytestUnraisableExceptionWarning` at a random other test when running the testsuite:
```python
test/mitmproxy/addons/test_readfile.py::TestReadFileStdin::test_stdin
/mnt/c/Users/user/git/mitmproxy/.tox/linux/py/lib/python3.10/site-packages/_pytest/unraisableexception.py:78: PytestUnraisableExceptionWarning: Exception ignored in: <coroutine object ConnectionHandler.handle_connection at 0x7fa0fbffbd80>
Traceback (most recent call last):
File "/usr/lib/python3.10/asyncio/queues.py", line 159, in get
await getter
GeneratorExit
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/mnt/c/Users/user/git/mitmproxy/mitmproxy/proxy/server.py", line 248, in handle_connection
data = await reader.read(65535)
File "/mnt/c/Users/user/git/mitmproxy/mitmproxy/net/udp.py", line 267, in read
return await self._packets.get()
File "/usr/lib/python3.10/asyncio/queues.py", line 161, in get
getter.cancel() # Just in case getter is not done yet.
File "/usr/lib/python3.10/asyncio/base_events.py", line 745, in call_soon
self._check_closed()
File "/usr/lib/python3.10/asyncio/base_events.py", line 510, in _check_closed
raise RuntimeError('Event loop is closed')
RuntimeError: Event loop is closed
warnings.warn(pytest.PytestUnraisableExceptionWarning(msg))
```
The problem is that pytest-asyncio closes the event loop while we still have an "active" UDP connection. One fix would be to just swallow these errors in `udp.py` (and return `b""`), alternatively we could manually cancel all active tasks in the respective test. I'm leaning towards the former. @Meitinger, any opinions? :)
#### Steps to reproduce the behavior:
1. `tox -e py`
#### System Information
```
Mitmproxy: 9.0.0.dev (+5, commit ef457a0)
Python: 3.10.2
OpenSSL: OpenSSL 1.1.1f 31 Mar 2020
Platform: Linux-4.4.0-19041-Microsoft-x86_64-with-glibc2.31
``` | 1.0 | PytestUnraisableExceptionWarning: RuntimeError: Event loop is closed - #### Problem Description
`test_proxyserver:test_dns` currently causes a `PytestUnraisableExceptionWarning` at a random other test when running the testsuite:
```python
test/mitmproxy/addons/test_readfile.py::TestReadFileStdin::test_stdin
/mnt/c/Users/user/git/mitmproxy/.tox/linux/py/lib/python3.10/site-packages/_pytest/unraisableexception.py:78: PytestUnraisableExceptionWarning: Exception ignored in: <coroutine object ConnectionHandler.handle_connection at 0x7fa0fbffbd80>
Traceback (most recent call last):
File "/usr/lib/python3.10/asyncio/queues.py", line 159, in get
await getter
GeneratorExit
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "/mnt/c/Users/user/git/mitmproxy/mitmproxy/proxy/server.py", line 248, in handle_connection
data = await reader.read(65535)
File "/mnt/c/Users/user/git/mitmproxy/mitmproxy/net/udp.py", line 267, in read
return await self._packets.get()
File "/usr/lib/python3.10/asyncio/queues.py", line 161, in get
getter.cancel() # Just in case getter is not done yet.
File "/usr/lib/python3.10/asyncio/base_events.py", line 745, in call_soon
self._check_closed()
File "/usr/lib/python3.10/asyncio/base_events.py", line 510, in _check_closed
raise RuntimeError('Event loop is closed')
RuntimeError: Event loop is closed
warnings.warn(pytest.PytestUnraisableExceptionWarning(msg))
```
The problem is that pytest-asyncio closes the event loop while we still have an "active" UDP connection. One fix would be to just swallow these errors in `udp.py` (and return `b""`), alternatively we could manually cancel all active tasks in the respective test. I'm leaning towards the former. @Meitinger, any opinions? :)
#### Steps to reproduce the behavior:
1. `tox -e py`
#### System Information
```
Mitmproxy: 9.0.0.dev (+5, commit ef457a0)
Python: 3.10.2
OpenSSL: OpenSSL 1.1.1f 31 Mar 2020
Platform: Linux-4.4.0-19041-Microsoft-x86_64-with-glibc2.31
``` | non_main | pytestunraisableexceptionwarning runtimeerror event loop is closed problem description test proxyserver test dns currently causes a pytestunraisableexceptionwarning at a random other test when running the testsuite python test mitmproxy addons test readfile py testreadfilestdin test stdin mnt c users user git mitmproxy tox linux py lib site packages pytest unraisableexception py pytestunraisableexceptionwarning exception ignored in traceback most recent call last file usr lib asyncio queues py line in get await getter generatorexit during handling of the above exception another exception occurred traceback most recent call last file mnt c users user git mitmproxy mitmproxy proxy server py line in handle connection data await reader read file mnt c users user git mitmproxy mitmproxy net udp py line in read return await self packets get file usr lib asyncio queues py line in get getter cancel just in case getter is not done yet file usr lib asyncio base events py line in call soon self check closed file usr lib asyncio base events py line in check closed raise runtimeerror event loop is closed runtimeerror event loop is closed warnings warn pytest pytestunraisableexceptionwarning msg the problem is that pytest asyncio closes the event loop while we still have an active udp connection one fix would be to just swallow these errors in udp py and return b alternatively we could manually cancel all active tasks in the respective test i m leaning towards the former meitinger any opinions steps to reproduce the behavior tox e py system information mitmproxy dev commit python openssl openssl mar platform linux microsoft with | 0 |
1,225 | 5,217,706,306 | IssuesEvent | 2017-01-26 14:42:28 | duckduckgo/zeroclickinfo-goodies | https://api.github.com/repos/duckduckgo/zeroclickinfo-goodies | closed | Conversions: Support “feet and inches” syntax | Improvement Maintainer Approved Status: PR Received Suggestion | As it stands, it only supports either [whole](https://duckduckgo.com/?q=5'+in+cm&ia=answer) or [decimal](https://duckduckgo.com/?q=5.5+feet+in+cm&ia=answer) feet. Queries such as [5'7" in cm](https://duckduckgo.com/?q=5'7"+in+cm) don’t produce results.
Likewise, it’d be useful to be able to convert in the other direction; [173cm in feet](https://duckduckgo.com/?q=173cm+in+feet&ia=answer) produces an answer in decimal feet. [173cm in feet and inches](https://duckduckgo.com/?q=173cm+in+feet+and+inches) produces no answer.
As someone who doesn’t use feet and inches regularly myself, but often interacts with people who do, it would be handy to be able to use DuckDuckGo to quickly convert.
| True | Conversions: Support “feet and inches” syntax - As it stands, it only supports either [whole](https://duckduckgo.com/?q=5'+in+cm&ia=answer) or [decimal](https://duckduckgo.com/?q=5.5+feet+in+cm&ia=answer) feet. Queries such as [5'7" in cm](https://duckduckgo.com/?q=5'7"+in+cm) don’t produce results.
Likewise, it’d be useful to be able to convert in the other direction; [173cm in feet](https://duckduckgo.com/?q=173cm+in+feet&ia=answer) produces an answer in decimal feet. [173cm in feet and inches](https://duckduckgo.com/?q=173cm+in+feet+and+inches) produces no answer.
As someone who doesn’t use feet and inches regularly myself, but often interacts with people who do, it would be handy to be able to use DuckDuckGo to quickly convert.
| main | conversions support “feet and inches” syntax as it stands it only supports either or feet queries such as don’t produce results likewise it’d be useful to be able to convert in the other direction produces an answer in decimal feet produces no answer as someone who doesn’t use feet and inches regularly myself but often interacts with people who do it would be handy to be able to use duckduckgo to quickly convert | 1 |
1,256 | 5,330,147,468 | IssuesEvent | 2017-02-15 16:25:44 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Error with module apt autoremove | affects_2.1 bug_report P2 waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
apt autoremove
##### ANSIBLE VERSION
``` bash
ansible 2.1.0.0
config file = /home/galindro/git_repos/devops/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
``` bash
$ cat ansible.cfg
[defaults]
host_key_checking = False
[ssh_connection]
pipelining=False
ssh_args = -o ForwardAgent=yes -o ControlMaster=auto -o ControlPersist=30m -o ControlPath="~/.ansible/tmp/ansible-ssh-%h-%p-%r"
```
I'm NOT using any ANSIBLE_\* environment variable
##### OS / ENVIRONMENT
```
No LSB modules are available.
Distributor ID: Debian
Description: Debian GNU/Linux 8.5 (jessie)
Release: 8.5
Codename: jessie
```
##### SUMMARY
I'm used the apt module with autoremove option and this error was showed when I run the task:
```
TASK [common : autoremove unused packages] *************************************
fatal: [172.30.2.214]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_RpjDPt/ansible_module_apt.py\", line 842, in <module>\r\n main()\r\n File \"/tmp/ansible_RpjDPt/ansible_module_apt.py\", line 802, in main\r\n for package in packages:\r\nTypeError: 'NoneType' object is not iterable\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
##### STEPS TO REPRODUCE
Create this task:
``` yaml
- name: autoremove unused packages
apt:
autoremove: yes
sudo: yes
tags:
- infra
```
Put that task inside a simple role
Create a playbook:
``` yaml
- hosts: all
role:
- myrolename
environment:
TMOUT: 600
```
Run the playbook
```
ansible-playbook -i ec2.py -v -b myplay.yml
```
P.S.: my host is an instance running on amazon ec2
##### EXPECTED RESULTS
This error:
``` bash
TASK [myrolename : autoremove unused packages] *************************************
fatal: [172.30.2.214]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_RpjDPt/ansible_module_apt.py\", line 842, in <module>\r\n main()\r\n File \"/tmp/ansible_RpjDPt/ansible_module_apt.py\", line 802, in main\r\n for package in packages:\r\nTypeError: 'NoneType' object is not iterable\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
##### ACTUAL RESULTS
Running with -vvvv:
```
TASK [myrolename : autoremove unused packages] *************************************
task path: /home/galindro/git_repos/devops/roles/myrolename/tasks/debian.yml:52
<172.30.2.214> ESTABLISH SSH CONNECTION FOR USER: ubuntu
<172.30.2.214> SSH: EXEC ssh -C -vvv -o ForwardAgent=yes -o ControlMaster=auto -o ControlPersist=30m -o 'ControlPath=~/.ansible/tmp/ansible-ssh-%h-%p-%r' -o StrictHostKeyChecking=no -o Port=40 -o 'IdentityFile="sbkey_20150204_app.pem"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ubuntu -o ConnectTimeout=10 172.30.2.214 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254 `" && echo ansible-tmp-1466722720.22-229804017504254="` echo $HOME/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254 `" ) && sleep 0'"'"''
<172.30.2.214> PUT /tmp/tmpBiapsW TO /home/ubuntu/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254/apt
<172.30.2.214> SSH: EXEC sftp -b - -C -vvv -o ForwardAgent=yes -o ControlMaster=auto -o ControlPersist=30m -o 'ControlPath=~/.ansible/tmp/ansible-ssh-%h-%p-%r' -o StrictHostKeyChecking=no -o Port=40 -o 'IdentityFile="sbkey_20150204_app.pem"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ubuntu -o ConnectTimeout=10 '[172.30.2.214]'
<172.30.2.214> ESTABLISH SSH CONNECTION FOR USER: ubuntu
<172.30.2.214> SSH: EXEC ssh -C -vvv -o ForwardAgent=yes -o ControlMaster=auto -o ControlPersist=30m -o 'ControlPath=~/.ansible/tmp/ansible-ssh-%h-%p-%r' -o StrictHostKeyChecking=no -o Port=40 -o 'IdentityFile="sbkey_20150204_app.pem"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ubuntu -o ConnectTimeout=10 -tt 172.30.2.214 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-yhgovfdopoujnzwegznjpbbaoqcpcnyd; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 TMOUT=600 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/ubuntu/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254/apt; rm -rf "/home/ubuntu/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
fatal: [172.30.2.214]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "apt"}, "module_stderr": "OpenSSH_6.7p1 Debian-5+deb8u2, OpenSSL 1.0.1t 3 May 2016\r\ndebug1: Reading configuration data /home/galindro/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 32546\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\nShared connection to 172.30.2.214 closed.\r\n", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_uvVsjU/ansible_module_apt.py\", line 842, in <module>\r\n main()\r\n File \"/tmp/ansible_uvVsjU/ansible_module_apt.py\", line 802, in main\r\n for package in packages:\r\nTypeError: 'NoneType' object is not iterable\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
| True | Error with module apt autoremove - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
apt autoremove
##### ANSIBLE VERSION
``` bash
ansible 2.1.0.0
config file = /home/galindro/git_repos/devops/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
``` bash
$ cat ansible.cfg
[defaults]
host_key_checking = False
[ssh_connection]
pipelining=False
ssh_args = -o ForwardAgent=yes -o ControlMaster=auto -o ControlPersist=30m -o ControlPath="~/.ansible/tmp/ansible-ssh-%h-%p-%r"
```
I'm NOT using any ANSIBLE_\* environment variable
##### OS / ENVIRONMENT
```
No LSB modules are available.
Distributor ID: Debian
Description: Debian GNU/Linux 8.5 (jessie)
Release: 8.5
Codename: jessie
```
##### SUMMARY
I'm used the apt module with autoremove option and this error was showed when I run the task:
```
TASK [common : autoremove unused packages] *************************************
fatal: [172.30.2.214]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_RpjDPt/ansible_module_apt.py\", line 842, in <module>\r\n main()\r\n File \"/tmp/ansible_RpjDPt/ansible_module_apt.py\", line 802, in main\r\n for package in packages:\r\nTypeError: 'NoneType' object is not iterable\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
##### STEPS TO REPRODUCE
Create this task:
``` yaml
- name: autoremove unused packages
apt:
autoremove: yes
sudo: yes
tags:
- infra
```
Put that task inside a simple role
Create a playbook:
``` yaml
- hosts: all
role:
- myrolename
environment:
TMOUT: 600
```
Run the playbook
```
ansible-playbook -i ec2.py -v -b myplay.yml
```
P.S.: my host is an instance running on amazon ec2
##### EXPECTED RESULTS
This error:
``` bash
TASK [myrolename : autoremove unused packages] *************************************
fatal: [172.30.2.214]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_RpjDPt/ansible_module_apt.py\", line 842, in <module>\r\n main()\r\n File \"/tmp/ansible_RpjDPt/ansible_module_apt.py\", line 802, in main\r\n for package in packages:\r\nTypeError: 'NoneType' object is not iterable\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
##### ACTUAL RESULTS
Running with -vvvv:
```
TASK [myrolename : autoremove unused packages] *************************************
task path: /home/galindro/git_repos/devops/roles/myrolename/tasks/debian.yml:52
<172.30.2.214> ESTABLISH SSH CONNECTION FOR USER: ubuntu
<172.30.2.214> SSH: EXEC ssh -C -vvv -o ForwardAgent=yes -o ControlMaster=auto -o ControlPersist=30m -o 'ControlPath=~/.ansible/tmp/ansible-ssh-%h-%p-%r' -o StrictHostKeyChecking=no -o Port=40 -o 'IdentityFile="sbkey_20150204_app.pem"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ubuntu -o ConnectTimeout=10 172.30.2.214 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254 `" && echo ansible-tmp-1466722720.22-229804017504254="` echo $HOME/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254 `" ) && sleep 0'"'"''
<172.30.2.214> PUT /tmp/tmpBiapsW TO /home/ubuntu/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254/apt
<172.30.2.214> SSH: EXEC sftp -b - -C -vvv -o ForwardAgent=yes -o ControlMaster=auto -o ControlPersist=30m -o 'ControlPath=~/.ansible/tmp/ansible-ssh-%h-%p-%r' -o StrictHostKeyChecking=no -o Port=40 -o 'IdentityFile="sbkey_20150204_app.pem"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ubuntu -o ConnectTimeout=10 '[172.30.2.214]'
<172.30.2.214> ESTABLISH SSH CONNECTION FOR USER: ubuntu
<172.30.2.214> SSH: EXEC ssh -C -vvv -o ForwardAgent=yes -o ControlMaster=auto -o ControlPersist=30m -o 'ControlPath=~/.ansible/tmp/ansible-ssh-%h-%p-%r' -o StrictHostKeyChecking=no -o Port=40 -o 'IdentityFile="sbkey_20150204_app.pem"' -o KbdInteractiveAuthentication=no -o PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o PasswordAuthentication=no -o User=ubuntu -o ConnectTimeout=10 -tt 172.30.2.214 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-yhgovfdopoujnzwegznjpbbaoqcpcnyd; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 TMOUT=600 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/ubuntu/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254/apt; rm -rf "/home/ubuntu/.ansible/tmp/ansible-tmp-1466722720.22-229804017504254/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
fatal: [172.30.2.214]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "apt"}, "module_stderr": "OpenSSH_6.7p1 Debian-5+deb8u2, OpenSSL 1.0.1t 3 May 2016\r\ndebug1: Reading configuration data /home/galindro/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 32546\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\nShared connection to 172.30.2.214 closed.\r\n", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_uvVsjU/ansible_module_apt.py\", line 842, in <module>\r\n main()\r\n File \"/tmp/ansible_uvVsjU/ansible_module_apt.py\", line 802, in main\r\n for package in packages:\r\nTypeError: 'NoneType' object is not iterable\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
| main | error with module apt autoremove issue type bug report component name apt autoremove ansible version bash ansible config file home galindro git repos devops ansible cfg configured module search path default w o overrides configuration bash cat ansible cfg host key checking false pipelining false ssh args o forwardagent yes o controlmaster auto o controlpersist o controlpath ansible tmp ansible ssh h p r i m not using any ansible environment variable os environment no lsb modules are available distributor id debian description debian gnu linux jessie release codename jessie summary i m used the apt module with autoremove option and this error was showed when i run the task task fatal failed changed false failed true module stderr module stdout traceback most recent call last r n file tmp ansible rpjdpt ansible module apt py line in r n main r n file tmp ansible rpjdpt ansible module apt py line in main r n for package in packages r ntypeerror nonetype object is not iterable r n msg module failure parsed false steps to reproduce create this task yaml name autoremove unused packages apt autoremove yes sudo yes tags infra put that task inside a simple role create a playbook yaml hosts all role myrolename environment tmout run the playbook ansible playbook i py v b myplay yml p s my host is an instance running on amazon expected results this error bash task fatal failed changed false failed true module stderr module stdout traceback most recent call last r n file tmp ansible rpjdpt ansible module apt py line in r n main r n file tmp ansible rpjdpt ansible module apt py line in main r n for package in packages r ntypeerror nonetype object is not iterable r n msg module failure parsed false actual results running with vvvv task task path home galindro git repos devops roles myrolename tasks debian yml establish ssh connection for user ubuntu ssh exec ssh c vvv o forwardagent yes o controlmaster auto o controlpersist o controlpath ansible tmp ansible ssh h p r o stricthostkeychecking no o port o identityfile sbkey app pem o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ubuntu o connecttimeout bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpbiapsw to home ubuntu ansible tmp ansible tmp apt ssh exec sftp b c vvv o forwardagent yes o controlmaster auto o controlpersist o controlpath ansible tmp ansible ssh h p r o stricthostkeychecking no o port o identityfile sbkey app pem o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ubuntu o connecttimeout establish ssh connection for user ubuntu ssh exec ssh c vvv o forwardagent yes o controlmaster auto o controlpersist o controlpath ansible tmp ansible ssh h p r o stricthostkeychecking no o port o identityfile sbkey app pem o kbdinteractiveauthentication no o preferredauthentications gssapi with mic gssapi keyex hostbased publickey o passwordauthentication no o user ubuntu o connecttimeout tt bin sh c sudo h s n u root bin sh c echo become success yhgovfdopoujnzwegznjpbbaoqcpcnyd lang en us utf lc all en us utf tmout lc messages en us utf usr bin python home ubuntu ansible tmp ansible tmp apt rm rf home ubuntu ansible tmp ansible tmp dev null sleep fatal failed changed false failed true invocation module name apt module stderr openssh debian openssl may r reading configuration data home galindro ssh config r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r fd setting o nonblock r mux client hello exchange master version r mux client forwards request forwardings local remote r mux client request session entering r mux client request alive entering r mux client request alive done pid r mux client request session session request sent r mux client request session master session id r mux client read packet read header failed broken pipe r received exit status from master r nshared connection to closed r n module stdout traceback most recent call last r n file tmp ansible uvvsju ansible module apt py line in r n main r n file tmp ansible uvvsju ansible module apt py line in main r n for package in packages r ntypeerror nonetype object is not iterable r n msg module failure parsed false | 1 |
5,024 | 25,783,591,698 | IssuesEvent | 2022-12-09 18:08:45 | Quansight-Labs/czi-scientific-python-mgmt | https://api.github.com/repos/Quansight-Labs/czi-scientific-python-mgmt | opened | Assist with release management to relieve maintainer burden | mission: maintainers 🚀 type: mission 🎯 | ## 📝 Summary
To put all of the above into the hands of the community, and to aid with adoption, we will help make releases for several ecosystem packages.
## 🚀 Tasks / Deliverables
See all issues labelled as [`mission: maintainers 🚀`](https://github.com/Quansight-Labs/czi-scientific-python-mgmt/issues?q=is%3Aopen+is%3Aissue+label%3A%22mission%3A+maintainers++%F0%9F%9A%80%22)
## 📋 Additional information
- | True | Assist with release management to relieve maintainer burden - ## 📝 Summary
To put all of the above into the hands of the community, and to aid with adoption, we will help make releases for several ecosystem packages.
## 🚀 Tasks / Deliverables
See all issues labelled as [`mission: maintainers 🚀`](https://github.com/Quansight-Labs/czi-scientific-python-mgmt/issues?q=is%3Aopen+is%3Aissue+label%3A%22mission%3A+maintainers++%F0%9F%9A%80%22)
## 📋 Additional information
- | main | assist with release management to relieve maintainer burden 📝 summary to put all of the above into the hands of the community and to aid with adoption we will help make releases for several ecosystem packages 🚀 tasks deliverables see all issues labelled as 📋 additional information | 1 |
3,293 | 12,681,063,179 | IssuesEvent | 2020-06-19 14:45:48 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | opened | Constructors shall not have ref or out parameters | Area: analyzer Area: maintainability feature | Constructors are responsible to create instances of the specific type. They should not be responsible to create other types that then are returned via ref or out parameters as that contradicts the idea of constructors. | True | Constructors shall not have ref or out parameters - Constructors are responsible to create instances of the specific type. They should not be responsible to create other types that then are returned via ref or out parameters as that contradicts the idea of constructors. | main | constructors shall not have ref or out parameters constructors are responsible to create instances of the specific type they should not be responsible to create other types that then are returned via ref or out parameters as that contradicts the idea of constructors | 1 |
1,762 | 2,518,902,212 | IssuesEvent | 2015-01-17 06:39:47 | rbsl-gaurav/rbframework | https://api.github.com/repos/rbsl-gaurav/rbframework | opened | Remove alert form Javascript code and introduce rb.console object | discuss Priority-High Priority-Normal task | <a href="https://github.com/mManishTrivedi"><img src="https://avatars.githubusercontent.com/u/2376789?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mManishTrivedi](https://github.com/mManishTrivedi)**
_Monday Jun 23, 2014 at 10:48 GMT_
_Originally opened as https://github.com/readybytes/rbframework/issues/40_
----
1. Dump data into console (Use console api)
2. Add specific condition like notification to end user or developer msgs (console enable when debug mode is enabled )
| 2.0 | Remove alert form Javascript code and introduce rb.console object - <a href="https://github.com/mManishTrivedi"><img src="https://avatars.githubusercontent.com/u/2376789?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mManishTrivedi](https://github.com/mManishTrivedi)**
_Monday Jun 23, 2014 at 10:48 GMT_
_Originally opened as https://github.com/readybytes/rbframework/issues/40_
----
1. Dump data into console (Use console api)
2. Add specific condition like notification to end user or developer msgs (console enable when debug mode is enabled )
| non_main | remove alert form javascript code and introduce rb console object issue by monday jun at gmt originally opened as dump data into console use console api add specific condition like notification to end user or developer msgs console enable when debug mode is enabled | 0 |
886 | 11,771,579,257 | IssuesEvent | 2020-03-16 00:37:30 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | closed | Portable Mode: Support auto update on Windows | feature-request portable-mode | Issue Type: <b>Bug</b>
Having the portable/zip version of vscode doesn't autoupdate, I have to manually update it by copying the zip.
I'd prefer if auto update worked because I'd imagine you could intelligently remove old files and persist my data - right now I don't know if I should delete everything and paste the new stuff (and lose my settings) or if it's OK to layer.
VS Code version: Code 1.40.0 (86405ea23e3937316009fc27c9361deee66ffbf5, 2019-11-06T17:02:13.381Z)
OS version: Windows_NT x64 10.0.18362
<!-- generated by issue reporter --> | True | Portable Mode: Support auto update on Windows - Issue Type: <b>Bug</b>
Having the portable/zip version of vscode doesn't autoupdate, I have to manually update it by copying the zip.
I'd prefer if auto update worked because I'd imagine you could intelligently remove old files and persist my data - right now I don't know if I should delete everything and paste the new stuff (and lose my settings) or if it's OK to layer.
VS Code version: Code 1.40.0 (86405ea23e3937316009fc27c9361deee66ffbf5, 2019-11-06T17:02:13.381Z)
OS version: Windows_NT x64 10.0.18362
<!-- generated by issue reporter --> | non_main | portable mode support auto update on windows issue type bug having the portable zip version of vscode doesn t autoupdate i have to manually update it by copying the zip i d prefer if auto update worked because i d imagine you could intelligently remove old files and persist my data right now i don t know if i should delete everything and paste the new stuff and lose my settings or if it s ok to layer vs code version code os version windows nt | 0 |
1,702 | 6,574,397,126 | IssuesEvent | 2017-09-11 12:44:35 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | apt module always returns CHANGED when using update_cache | affects_2.2 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
apt
##### ANSIBLE VERSION
```
ansible 2.2.0.0
```
##### CONFIGURATION
Default
##### OS / ENVIRONMENT
Debian Jessie 64 GNU/Linux
##### SUMMARY
When installing packages and using update_cache option apt module always returns state CHANGED even if packages are already installed.
In my opinion expected behavior would be:
1. when using update_cache AND installing package(s) return status CHANGED only when package has been installed or updated, no matter if cache has been updated or not.
2. when using update_cache without installing any package(s) return status CHANGED when cache has been updated.
##### STEPS TO REPRODUCE
```
---
- hosts: localhost
gather_facts: no
become: yes
tasks:
- apt:
name: htop
update_cache: yes
```
##### EXPECTED RESULTS
As in ansible 2.1:
```
PLAY [localhost] ***************************************************************
TASK [apt] *********************************************************************
ok: [localhost]
PLAY RECAP *********************************************************************
localhost : ok=1 changed=0 unreachable=0 failed=0
```
Part of debug output:
```
ok: [localhost] => {"cache_update_time": 1478621386, "cache_updated": true, "changed": false, "invocation": {"module_args": {"allow_unauthenticated": false, "autoremove": false, "cache_valid_time": null, "deb": null, "default_release": null, "dpkg_options": "force-confdef,force-confold", "force": false, "install_recommends": null, "name": "htop", "only_upgrade": false, "package": ["htop"], "purge": false, "state": "present", "update_cache": true, "upgrade": null}, "module_name": "apt"}}
```
##### ACTUAL RESULTS
```
PLAY [localhost] ***************************************************************
TASK [apt] *********************************************************************
changed: [localhost]
PLAY RECAP *********************************************************************
localhost : ok=1 changed=1 unreachable=0 failed=0
```
Part of debug output:
```
changed: [localhost] => {
"cache_update_time": 1478621193,
"cache_updated": true,
"changed": true,
"invocation": {
"module_args": {
"allow_unauthenticated": false,
"autoremove": false,
"cache_valid_time": 0,
"deb": null,
"default_release": null,
"dpkg_options": "force-confdef,force-confold",
"force": false,
"install_recommends": null,
"name": "htop",
"only_upgrade": false,
"package": [
"htop"
],
"purge": false,
"state": "present",
"update_cache": true,
"upgrade": null
},
"module_name": "apt"
}
}
``` | True | apt module always returns CHANGED when using update_cache - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
apt
##### ANSIBLE VERSION
```
ansible 2.2.0.0
```
##### CONFIGURATION
Default
##### OS / ENVIRONMENT
Debian Jessie 64 GNU/Linux
##### SUMMARY
When installing packages and using update_cache option apt module always returns state CHANGED even if packages are already installed.
In my opinion expected behavior would be:
1. when using update_cache AND installing package(s) return status CHANGED only when package has been installed or updated, no matter if cache has been updated or not.
2. when using update_cache without installing any package(s) return status CHANGED when cache has been updated.
##### STEPS TO REPRODUCE
```
---
- hosts: localhost
gather_facts: no
become: yes
tasks:
- apt:
name: htop
update_cache: yes
```
##### EXPECTED RESULTS
As in ansible 2.1:
```
PLAY [localhost] ***************************************************************
TASK [apt] *********************************************************************
ok: [localhost]
PLAY RECAP *********************************************************************
localhost : ok=1 changed=0 unreachable=0 failed=0
```
Part of debug output:
```
ok: [localhost] => {"cache_update_time": 1478621386, "cache_updated": true, "changed": false, "invocation": {"module_args": {"allow_unauthenticated": false, "autoremove": false, "cache_valid_time": null, "deb": null, "default_release": null, "dpkg_options": "force-confdef,force-confold", "force": false, "install_recommends": null, "name": "htop", "only_upgrade": false, "package": ["htop"], "purge": false, "state": "present", "update_cache": true, "upgrade": null}, "module_name": "apt"}}
```
##### ACTUAL RESULTS
```
PLAY [localhost] ***************************************************************
TASK [apt] *********************************************************************
changed: [localhost]
PLAY RECAP *********************************************************************
localhost : ok=1 changed=1 unreachable=0 failed=0
```
Part of debug output:
```
changed: [localhost] => {
"cache_update_time": 1478621193,
"cache_updated": true,
"changed": true,
"invocation": {
"module_args": {
"allow_unauthenticated": false,
"autoremove": false,
"cache_valid_time": 0,
"deb": null,
"default_release": null,
"dpkg_options": "force-confdef,force-confold",
"force": false,
"install_recommends": null,
"name": "htop",
"only_upgrade": false,
"package": [
"htop"
],
"purge": false,
"state": "present",
"update_cache": true,
"upgrade": null
},
"module_name": "apt"
}
}
``` | main | apt module always returns changed when using update cache issue type bug report component name apt ansible version ansible configuration default os environment debian jessie gnu linux summary when installing packages and using update cache option apt module always returns state changed even if packages are already installed in my opinion expected behavior would be when using update cache and installing package s return status changed only when package has been installed or updated no matter if cache has been updated or not when using update cache without installing any package s return status changed when cache has been updated steps to reproduce hosts localhost gather facts no become yes tasks apt name htop update cache yes expected results as in ansible play task ok play recap localhost ok changed unreachable failed part of debug output ok cache update time cache updated true changed false invocation module args allow unauthenticated false autoremove false cache valid time null deb null default release null dpkg options force confdef force confold force false install recommends null name htop only upgrade false package purge false state present update cache true upgrade null module name apt actual results play task changed play recap localhost ok changed unreachable failed part of debug output changed cache update time cache updated true changed true invocation module args allow unauthenticated false autoremove false cache valid time deb null default release null dpkg options force confdef force confold force false install recommends null name htop only upgrade false package htop purge false state present update cache true upgrade null module name apt | 1 |
1,692 | 6,574,191,942 | IssuesEvent | 2017-09-11 11:54:14 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | ios_facts: very limited functionality | affects_2.2 bug_report networking waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
ios_facts
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Distributor ID: Ubuntu
Description: Ubuntu 16.04.1 LTS
Release: 16.04
Codename: xenial
##### SUMMARY
<!--- Explain the problem briefly -->
looks like ios_facts module covering only cisco routers, but switches running different version of software failing with different commands failure
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Tried on three different devices:
WS-C4500X-16 - didn't worked bootflash:/cat4500e-universalk9.SPA.03.06.04.E.152-2.E4.bin
WS-C4507R - didn't worked slot0:cat4500-entservicesk9-mz.122-31.SGA7.bin
ISR4451-X/K9 - worked bootflash:/isr4400-universalk9.03.11.01.S.154-1.S1-std.SPA.bin
<!--- Paste example playbooks or commands between quotes below -->
```
- name: DEFINE
set_fact:
connection:
host: "{{ inventory_hostname }}"
username: "{{ creds['username'] }}"
password: "{{ creds['password'] }}"
- name: FACTS
ios_facts:
provider: "{{ connection }}"
gather_subset: all
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Need to improve functionality or add definition which type of cisco devices can be used
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
Below you can find example, with problem related to command which not existing in switch envrionment
<!--- Paste verbatim command output between quotes below -->
```
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: lgb3-wvrtcore-a#
fatal: [lgb3-wvrtcore-a]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_a6NdY9/ansible_module_ios_facts.py\", line 461, in <module>\n main()\n File \"/tmp/ansible_a6NdY9/ansible_module_ios_facts.py\", line 443, in main\n runner.run()\n File \"/tmp/ansible_a6NdY9/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 170, in run\n File \"/tmp/ansible_a6NdY9/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 98, in run_commands\n File \"/tmp/ansible_a6NdY9/ansible_modlib.zip/ansible/module_utils/shell.py\", line 253, in run_commands\n File \"/tmp/ansible_a6NdY9/ansible_modlib.zip/ansible/module_utils/shell.py\", line 250, in execute\nansible.module_utils.network.NetworkError: matched error in response: show memory statistics | include Processor\r\n ^\r\n% Invalid input detected at '^' marker.\r\n\r\nlgb3-wvrtcore-a#\n", "module_stdout": "", "msg": "MODULE FAILURE"}
```
| True | ios_facts: very limited functionality - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
ios_facts
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Distributor ID: Ubuntu
Description: Ubuntu 16.04.1 LTS
Release: 16.04
Codename: xenial
##### SUMMARY
<!--- Explain the problem briefly -->
looks like ios_facts module covering only cisco routers, but switches running different version of software failing with different commands failure
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Tried on three different devices:
WS-C4500X-16 - didn't worked bootflash:/cat4500e-universalk9.SPA.03.06.04.E.152-2.E4.bin
WS-C4507R - didn't worked slot0:cat4500-entservicesk9-mz.122-31.SGA7.bin
ISR4451-X/K9 - worked bootflash:/isr4400-universalk9.03.11.01.S.154-1.S1-std.SPA.bin
<!--- Paste example playbooks or commands between quotes below -->
```
- name: DEFINE
set_fact:
connection:
host: "{{ inventory_hostname }}"
username: "{{ creds['username'] }}"
password: "{{ creds['password'] }}"
- name: FACTS
ios_facts:
provider: "{{ connection }}"
gather_subset: all
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Need to improve functionality or add definition which type of cisco devices can be used
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
Below you can find example, with problem related to command which not existing in switch envrionment
<!--- Paste verbatim command output between quotes below -->
```
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: lgb3-wvrtcore-a#
fatal: [lgb3-wvrtcore-a]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_a6NdY9/ansible_module_ios_facts.py\", line 461, in <module>\n main()\n File \"/tmp/ansible_a6NdY9/ansible_module_ios_facts.py\", line 443, in main\n runner.run()\n File \"/tmp/ansible_a6NdY9/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 170, in run\n File \"/tmp/ansible_a6NdY9/ansible_modlib.zip/ansible/module_utils/netcli.py\", line 98, in run_commands\n File \"/tmp/ansible_a6NdY9/ansible_modlib.zip/ansible/module_utils/shell.py\", line 253, in run_commands\n File \"/tmp/ansible_a6NdY9/ansible_modlib.zip/ansible/module_utils/shell.py\", line 250, in execute\nansible.module_utils.network.NetworkError: matched error in response: show memory statistics | include Processor\r\n ^\r\n% Invalid input detected at '^' marker.\r\n\r\nlgb3-wvrtcore-a#\n", "module_stdout": "", "msg": "MODULE FAILURE"}
```
| main | ios facts very limited functionality issue type bug report component name ios facts ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific distributor id ubuntu description ubuntu lts release codename xenial summary looks like ios facts module covering only cisco routers but switches running different version of software failing with different commands failure steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used tried on three different devices ws didn t worked bootflash spa e bin ws didn t worked mz bin x worked bootflash s std spa bin name define set fact connection host inventory hostname username creds password creds name facts ios facts provider connection gather subset all expected results need to improve functionality or add definition which type of cisco devices can be used actual results below you can find example with problem related to command which not existing in switch envrionment an exception occurred during task execution to see the full traceback use vvv the error was wvrtcore a fatal failed changed false failed true module stderr traceback most recent call last n file tmp ansible ansible module ios facts py line in n main n file tmp ansible ansible module ios facts py line in main n runner run n file tmp ansible ansible modlib zip ansible module utils netcli py line in run n file tmp ansible ansible modlib zip ansible module utils netcli py line in run commands n file tmp ansible ansible modlib zip ansible module utils shell py line in run commands n file tmp ansible ansible modlib zip ansible module utils shell py line in execute nansible module utils network networkerror matched error in response show memory statistics include processor r n r n invalid input detected at marker r n r wvrtcore a n module stdout msg module failure | 1 |
100,726 | 12,554,788,155 | IssuesEvent | 2020-06-07 03:28:55 | EnigmaDragons/TheInterview | https://api.github.com/repos/EnigmaDragons/TheInterview | closed | Create Server Farm | design | A small security room with a card printer inside:
An administrative room with a card reader protecting it
A cooling system level with off and on inside server room
the server room should have columns with letters and then server racks with numbers in the columns, the servers that are hackable should have some form of open console to indicate they are hackable
the last column in the server room should have giant new signs | 1.0 | Create Server Farm - A small security room with a card printer inside:
An administrative room with a card reader protecting it
A cooling system level with off and on inside server room
the server room should have columns with letters and then server racks with numbers in the columns, the servers that are hackable should have some form of open console to indicate they are hackable
the last column in the server room should have giant new signs | non_main | create server farm a small security room with a card printer inside an administrative room with a card reader protecting it a cooling system level with off and on inside server room the server room should have columns with letters and then server racks with numbers in the columns the servers that are hackable should have some form of open console to indicate they are hackable the last column in the server room should have giant new signs | 0 |
365,069 | 25,519,225,123 | IssuesEvent | 2022-11-28 18:57:54 | avioconsulting/mule-opentelemetry-module | https://api.github.com/repos/avioconsulting/mule-opentelemetry-module | closed | Extracting httpStatus causes error for non-java values | bug documentation | HTTP Processor on source flows attempts to extracts the http response code from `vars.httpStatus` and set the span's status code with it -
https://github.com/avioconsulting/mule-opentelemetry-module/blob/main/src/main/java/com/avioconsulting/mule/opentelemetry/internal/processor/HttpProcessorComponent.java#L191-L198
This works when value `httpStatus` variable contains a java value. But when the value is set as some other media type, eg. `application/json`, the variable's value gets wrapped in an instance of stream provider with json value. In that case, `statusCode` gets the value of the classname instead of string representation of actual http status. This results in `Integer.parseInt(statusCode)` to fail with a NumberFormatException ( see #79 ).
One example of erroneous status variable declaration -
```
<set-variable variableName="httpStatus" value="#[output application/json --- 200]"/>
```
```
java.lang.NumberFormatException: For input string: "org.mule.runtime.core.internal.streaming.bytes.ManagedCursorStreamProvider@1ce7d975"
```
This leads to spans/traces not ending.
- [ ] Add a note in document for `httpStatus` variable to be a java value.
- [ ] Should we handle json value? Probably not needed. But maybe failure should be handled gracefully by logging the error/warning and letting span to end without status code.
| 1.0 | Extracting httpStatus causes error for non-java values - HTTP Processor on source flows attempts to extracts the http response code from `vars.httpStatus` and set the span's status code with it -
https://github.com/avioconsulting/mule-opentelemetry-module/blob/main/src/main/java/com/avioconsulting/mule/opentelemetry/internal/processor/HttpProcessorComponent.java#L191-L198
This works when value `httpStatus` variable contains a java value. But when the value is set as some other media type, eg. `application/json`, the variable's value gets wrapped in an instance of stream provider with json value. In that case, `statusCode` gets the value of the classname instead of string representation of actual http status. This results in `Integer.parseInt(statusCode)` to fail with a NumberFormatException ( see #79 ).
One example of erroneous status variable declaration -
```
<set-variable variableName="httpStatus" value="#[output application/json --- 200]"/>
```
```
java.lang.NumberFormatException: For input string: "org.mule.runtime.core.internal.streaming.bytes.ManagedCursorStreamProvider@1ce7d975"
```
This leads to spans/traces not ending.
- [ ] Add a note in document for `httpStatus` variable to be a java value.
- [ ] Should we handle json value? Probably not needed. But maybe failure should be handled gracefully by logging the error/warning and letting span to end without status code.
| non_main | extracting httpstatus causes error for non java values http processor on source flows attempts to extracts the http response code from vars httpstatus and set the span s status code with it this works when value httpstatus variable contains a java value but when the value is set as some other media type eg application json the variable s value gets wrapped in an instance of stream provider with json value in that case statuscode gets the value of the classname instead of string representation of actual http status this results in integer parseint statuscode to fail with a numberformatexception see one example of erroneous status variable declaration java lang numberformatexception for input string org mule runtime core internal streaming bytes managedcursorstreamprovider this leads to spans traces not ending add a note in document for httpstatus variable to be a java value should we handle json value probably not needed but maybe failure should be handled gracefully by logging the error warning and letting span to end without status code | 0 |
5,500 | 27,448,053,569 | IssuesEvent | 2023-03-02 15:31:11 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [a11y]: Disabled state on various components violates minimum contrast rules | type: a11y ♿ status: waiting for maintainer response 💬 | ### Package
carbon-components-react
### Browser
Chrome
### Operating System
MacOS
### Package version
"carbon-components": "^10.56.0" but also applicable to latest v11
### React version
v17
### Automated testing tool and ruleset
IBM equal access accessibility checker - latest
### Assistive technology
_No response_
### Description
Expect disabled components to have sufficient contrast to not violate accessibility rules
### WCAG 2.1 Violation
1.4.3 Contrast (Minimum): The visual presentation of text and images of text has a contrast ratio of at least 4.5:1, with a 3:1 ratio for large-scale text.
### Reproduction/example
https://react.carbondesignsystem.com/?path=/story/components-textinput--playground&args=disabled:true
### Steps to reproduce
Run IBM equal access accessibility checker on disabled carbon components and see the violation.
Additionally, you can go to the carbon design systems storybook and open the accessibility tab and see the violation.
<img width="1728" alt="Screenshot 2023-03-01 at 11 37 56 AM" src="https://user-images.githubusercontent.com/13371221/222204208-5eefc30b-44ed-4bb1-bb93-51454385af72.png">
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | True | [a11y]: Disabled state on various components violates minimum contrast rules - ### Package
carbon-components-react
### Browser
Chrome
### Operating System
MacOS
### Package version
"carbon-components": "^10.56.0" but also applicable to latest v11
### React version
v17
### Automated testing tool and ruleset
IBM equal access accessibility checker - latest
### Assistive technology
_No response_
### Description
Expect disabled components to have sufficient contrast to not violate accessibility rules
### WCAG 2.1 Violation
1.4.3 Contrast (Minimum): The visual presentation of text and images of text has a contrast ratio of at least 4.5:1, with a 3:1 ratio for large-scale text.
### Reproduction/example
https://react.carbondesignsystem.com/?path=/story/components-textinput--playground&args=disabled:true
### Steps to reproduce
Run IBM equal access accessibility checker on disabled carbon components and see the violation.
Additionally, you can go to the carbon design systems storybook and open the accessibility tab and see the violation.
<img width="1728" alt="Screenshot 2023-03-01 at 11 37 56 AM" src="https://user-images.githubusercontent.com/13371221/222204208-5eefc30b-44ed-4bb1-bb93-51454385af72.png">
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | main | disabled state on various components violates minimum contrast rules package carbon components react browser chrome operating system macos package version carbon components but also applicable to latest react version automated testing tool and ruleset ibm equal access accessibility checker latest assistive technology no response description expect disabled components to have sufficient contrast to not violate accessibility rules wcag violation contrast minimum the visual presentation of text and images of text has a contrast ratio of at least with a ratio for large scale text reproduction example steps to reproduce run ibm equal access accessibility checker on disabled carbon components and see the violation additionally you can go to the carbon design systems storybook and open the accessibility tab and see the violation img width alt screenshot at am src code of conduct i agree to follow this project s i checked the for duplicate problems | 1 |
4,575 | 23,772,750,740 | IssuesEvent | 2022-09-01 17:50:52 | pyOpenSci/software-review | https://api.github.com/repos/pyOpenSci/software-review | closed | Ocetrac: A Python package to track the spatiotemporal evolution of marine heatwaves | 1/editor-checks 2/seeking-reviewer(s) ⌛ pending-maintainer-response | Submitting Author: Hillary Scannell (@hscannell)
Package Name: Ocetrac
One-Line Description of Package: Ocetrac is a python package to detect and track the spatiotemporal evolution of marine heatwaves.
Repository Link: https://github.com/ocetrac/ocetrac
Version submitted: 0.1.4
Editor: TBD
Reviewer 1: TBD
Reviewer 2: TBD
Archive: TBD
Version accepted: TBD
---
## Description
Ocetrac is an analysis routine and data processing tool used to extract the spatial trajectories of marine heatwaves (MHW) over time. It utilizes both morphological image processing and multiple object tracking to provide a new set metrics including event size, location, intensity, and duration. We anticipate that these metrics will be incorporated into machine learning forecasts to predict when and where MHWs are likely to occur, with the intent for operational use in warning vulnerable coastal communities of physical risk. While the motivation behind developing Ocetrac was to study MHWs, the algorithm could be applied to track any geographically coherent spatiotemporal anomaly.
## Scope
- Please indicate which [category or categories][PackageCategories] this package falls under:
- [ ] Data retrieval
- [ ] Data extraction
- [ ] Data munging
- [ ] Data deposition
- [ ] Reproducibility
- [x] Geospatial
- [ ] Education
- [ ] Data visualization*
\* Please fill out a pre-submission inquiry before submitting a data visualization package. For more info, see [notes on categories][NotesOnCategories] of our guidebook.
- Explain how the and why the package falls under these categories (briefly, 1-2 sentences):
Ocetrac analyzes the spatiotemporal connectivity amongst geospatial anomalies. In doing so, it makes use of common morphological operations borrowed from multidimensional image processing.
- Who is the target audience and what are scientific applications of this package?
The target audiences are data analysts and physical scientists charged with understanding the spatiotemporal evolution of anomalous events.
- Are there other Python packages that accomplish the same thing? If so, how does yours differ?
To our knowledge, there is no other Python package that achieves the goals of Ocetrac.
- If you made a pre-submission enquiry, please paste the link to the corresponding issue, forum post, or other discussion, or `@tag` the editor you contacted:
## Technical checks
For details about the pyOpenSci packaging requirements, see our [packaging guide][PackagingGuide]. Confirm each of the following by checking the box. This package:
- [x] does not violate the Terms of Service of any service it interacts with.
- [x] has an [OSI approved license][OsiApprovedLicense].
- [x] contains a README with instructions for installing the development version.
- [x] includes documentation with examples for all functions.
- [x] contains a vignette with examples of its essential functions and uses.
- [x] has a test suite.
- [x] has continuous integration, such as Travis CI, AppVeyor, CircleCI, and/or others.
## Publication options
- [ ] Do you wish to automatically submit to the [Journal of Open Source Software][JournalOfOpenSourceSoftware]? If so:
<details>
<summary>JOSS Checks</summary>
- [ ] The package has an **obvious research application** according to JOSS's definition in their [submission requirements][JossSubmissionRequirements]. Be aware that completing the pyOpenSci review process **does not** guarantee acceptance to JOSS. Be sure to read their submission requirements (linked above) if you are interested in submitting to JOSS.
- [ ] The package is not a "minor utility" as defined by JOSS's [submission requirements][JossSubmissionRequirements]: "Minor ‘utility’ packages, including ‘thin’ API clients, are not acceptable." pyOpenSci welcomes these packages under "Data Retrieval", but JOSS has slightly different criteria.
- [ ] The package contains a `paper.md` matching [JOSS's requirements][JossPaperRequirements] with a high-level description in the package root or in `inst/`.
- [ ] The package is deposited in a long-term repository with the DOI:
*Note: Do not submit your package separately to JOSS*
</details>
## Are you OK with Reviewers Submitting Issues and/or pull requests to your Repo Directly?
This option will allow reviewers to open smaller issues that can then be linked to PR's rather than submitting a more dense text based review. It will also allow you to demonstrate addressing the issue via PR links.
- [x] Yes I am OK with reviewers submitting requested changes as issues to my repo. Reviewers will then link to the issues in their submitted review.
## Code of conduct
- [x] I agree to abide by [pyOpenSci's Code of Conduct][PyOpenSciCodeOfConduct] during the review process and in maintaining my package should it be accepted.
**P.S.** *Have feedback/comments about our review process? Leave a comment [here][Comments]
## Editor and Review Templates
[Editor and review templates can be found here][Templates]
[PackagingGuide]: https://www.pyopensci.org/contributing-guide/authoring/index.html#packaging-guide
[PackageCategories]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/aims-and-scope.html?highlight=data#package-categories
[NotesOnCategories]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/aims-and-scope.html?highlight=data#notes-on-categories
[JournalOfOpenSourceSoftware]: http://joss.theoj.org/
[JossSubmissionRequirements]: https://joss.readthedocs.io/en/latest/submitting.html#submission-requirements
[JossPaperRequirements]: https://joss.readthedocs.io/en/latest/submitting.html#what-should-my-paper-contain
[PyOpenSciCodeOfConduct]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/code-of-conduct.html?highlight=code%20conduct
[OsiApprovedLicense]: https://opensource.org/licenses
[Templates]: https://www.pyopensci.org/contributing-guide/appendices/templates.html
[Comments]: https://github.com/pyOpenSci/governance/issues/8 | True | Ocetrac: A Python package to track the spatiotemporal evolution of marine heatwaves - Submitting Author: Hillary Scannell (@hscannell)
Package Name: Ocetrac
One-Line Description of Package: Ocetrac is a python package to detect and track the spatiotemporal evolution of marine heatwaves.
Repository Link: https://github.com/ocetrac/ocetrac
Version submitted: 0.1.4
Editor: TBD
Reviewer 1: TBD
Reviewer 2: TBD
Archive: TBD
Version accepted: TBD
---
## Description
Ocetrac is an analysis routine and data processing tool used to extract the spatial trajectories of marine heatwaves (MHW) over time. It utilizes both morphological image processing and multiple object tracking to provide a new set metrics including event size, location, intensity, and duration. We anticipate that these metrics will be incorporated into machine learning forecasts to predict when and where MHWs are likely to occur, with the intent for operational use in warning vulnerable coastal communities of physical risk. While the motivation behind developing Ocetrac was to study MHWs, the algorithm could be applied to track any geographically coherent spatiotemporal anomaly.
## Scope
- Please indicate which [category or categories][PackageCategories] this package falls under:
- [ ] Data retrieval
- [ ] Data extraction
- [ ] Data munging
- [ ] Data deposition
- [ ] Reproducibility
- [x] Geospatial
- [ ] Education
- [ ] Data visualization*
\* Please fill out a pre-submission inquiry before submitting a data visualization package. For more info, see [notes on categories][NotesOnCategories] of our guidebook.
- Explain how the and why the package falls under these categories (briefly, 1-2 sentences):
Ocetrac analyzes the spatiotemporal connectivity amongst geospatial anomalies. In doing so, it makes use of common morphological operations borrowed from multidimensional image processing.
- Who is the target audience and what are scientific applications of this package?
The target audiences are data analysts and physical scientists charged with understanding the spatiotemporal evolution of anomalous events.
- Are there other Python packages that accomplish the same thing? If so, how does yours differ?
To our knowledge, there is no other Python package that achieves the goals of Ocetrac.
- If you made a pre-submission enquiry, please paste the link to the corresponding issue, forum post, or other discussion, or `@tag` the editor you contacted:
## Technical checks
For details about the pyOpenSci packaging requirements, see our [packaging guide][PackagingGuide]. Confirm each of the following by checking the box. This package:
- [x] does not violate the Terms of Service of any service it interacts with.
- [x] has an [OSI approved license][OsiApprovedLicense].
- [x] contains a README with instructions for installing the development version.
- [x] includes documentation with examples for all functions.
- [x] contains a vignette with examples of its essential functions and uses.
- [x] has a test suite.
- [x] has continuous integration, such as Travis CI, AppVeyor, CircleCI, and/or others.
## Publication options
- [ ] Do you wish to automatically submit to the [Journal of Open Source Software][JournalOfOpenSourceSoftware]? If so:
<details>
<summary>JOSS Checks</summary>
- [ ] The package has an **obvious research application** according to JOSS's definition in their [submission requirements][JossSubmissionRequirements]. Be aware that completing the pyOpenSci review process **does not** guarantee acceptance to JOSS. Be sure to read their submission requirements (linked above) if you are interested in submitting to JOSS.
- [ ] The package is not a "minor utility" as defined by JOSS's [submission requirements][JossSubmissionRequirements]: "Minor ‘utility’ packages, including ‘thin’ API clients, are not acceptable." pyOpenSci welcomes these packages under "Data Retrieval", but JOSS has slightly different criteria.
- [ ] The package contains a `paper.md` matching [JOSS's requirements][JossPaperRequirements] with a high-level description in the package root or in `inst/`.
- [ ] The package is deposited in a long-term repository with the DOI:
*Note: Do not submit your package separately to JOSS*
</details>
## Are you OK with Reviewers Submitting Issues and/or pull requests to your Repo Directly?
This option will allow reviewers to open smaller issues that can then be linked to PR's rather than submitting a more dense text based review. It will also allow you to demonstrate addressing the issue via PR links.
- [x] Yes I am OK with reviewers submitting requested changes as issues to my repo. Reviewers will then link to the issues in their submitted review.
## Code of conduct
- [x] I agree to abide by [pyOpenSci's Code of Conduct][PyOpenSciCodeOfConduct] during the review process and in maintaining my package should it be accepted.
**P.S.** *Have feedback/comments about our review process? Leave a comment [here][Comments]
## Editor and Review Templates
[Editor and review templates can be found here][Templates]
[PackagingGuide]: https://www.pyopensci.org/contributing-guide/authoring/index.html#packaging-guide
[PackageCategories]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/aims-and-scope.html?highlight=data#package-categories
[NotesOnCategories]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/aims-and-scope.html?highlight=data#notes-on-categories
[JournalOfOpenSourceSoftware]: http://joss.theoj.org/
[JossSubmissionRequirements]: https://joss.readthedocs.io/en/latest/submitting.html#submission-requirements
[JossPaperRequirements]: https://joss.readthedocs.io/en/latest/submitting.html#what-should-my-paper-contain
[PyOpenSciCodeOfConduct]: https://www.pyopensci.org/contributing-guide/open-source-software-peer-review/code-of-conduct.html?highlight=code%20conduct
[OsiApprovedLicense]: https://opensource.org/licenses
[Templates]: https://www.pyopensci.org/contributing-guide/appendices/templates.html
[Comments]: https://github.com/pyOpenSci/governance/issues/8 | main | ocetrac a python package to track the spatiotemporal evolution of marine heatwaves submitting author hillary scannell hscannell package name ocetrac one line description of package ocetrac is a python package to detect and track the spatiotemporal evolution of marine heatwaves repository link version submitted editor tbd reviewer tbd reviewer tbd archive tbd version accepted tbd description ocetrac is an analysis routine and data processing tool used to extract the spatial trajectories of marine heatwaves mhw over time it utilizes both morphological image processing and multiple object tracking to provide a new set metrics including event size location intensity and duration we anticipate that these metrics will be incorporated into machine learning forecasts to predict when and where mhws are likely to occur with the intent for operational use in warning vulnerable coastal communities of physical risk while the motivation behind developing ocetrac was to study mhws the algorithm could be applied to track any geographically coherent spatiotemporal anomaly scope please indicate which this package falls under data retrieval data extraction data munging data deposition reproducibility geospatial education data visualization please fill out a pre submission inquiry before submitting a data visualization package for more info see of our guidebook explain how the and why the package falls under these categories briefly sentences ocetrac analyzes the spatiotemporal connectivity amongst geospatial anomalies in doing so it makes use of common morphological operations borrowed from multidimensional image processing who is the target audience and what are scientific applications of this package the target audiences are data analysts and physical scientists charged with understanding the spatiotemporal evolution of anomalous events are there other python packages that accomplish the same thing if so how does yours differ to our knowledge there is no other python package that achieves the goals of ocetrac if you made a pre submission enquiry please paste the link to the corresponding issue forum post or other discussion or tag the editor you contacted technical checks for details about the pyopensci packaging requirements see our confirm each of the following by checking the box this package does not violate the terms of service of any service it interacts with has an contains a readme with instructions for installing the development version includes documentation with examples for all functions contains a vignette with examples of its essential functions and uses has a test suite has continuous integration such as travis ci appveyor circleci and or others publication options do you wish to automatically submit to the if so joss checks the package has an obvious research application according to joss s definition in their be aware that completing the pyopensci review process does not guarantee acceptance to joss be sure to read their submission requirements linked above if you are interested in submitting to joss the package is not a minor utility as defined by joss s minor ‘utility’ packages including ‘thin’ api clients are not acceptable pyopensci welcomes these packages under data retrieval but joss has slightly different criteria the package contains a paper md matching with a high level description in the package root or in inst the package is deposited in a long term repository with the doi note do not submit your package separately to joss are you ok with reviewers submitting issues and or pull requests to your repo directly this option will allow reviewers to open smaller issues that can then be linked to pr s rather than submitting a more dense text based review it will also allow you to demonstrate addressing the issue via pr links yes i am ok with reviewers submitting requested changes as issues to my repo reviewers will then link to the issues in their submitted review code of conduct i agree to abide by during the review process and in maintaining my package should it be accepted p s have feedback comments about our review process leave a comment editor and review templates | 1 |
5,086 | 25,998,604,646 | IssuesEvent | 2022-12-20 13:39:29 | cncf/glossary | https://api.github.com/repos/cncf/glossary | opened | Maintainer to-do list (English Glossary) | lang/en maintainers | - [ ] Rework microservices definition to microservices architecture and adjust services definition accordingly
- [ ] GitOps term
- [ ] Revisit observability | True | Maintainer to-do list (English Glossary) - - [ ] Rework microservices definition to microservices architecture and adjust services definition accordingly
- [ ] GitOps term
- [ ] Revisit observability | main | maintainer to do list english glossary rework microservices definition to microservices architecture and adjust services definition accordingly gitops term revisit observability | 1 |
36,934 | 15,097,609,080 | IssuesEvent | 2021-02-07 19:21:30 | KieronQuinn/TapTap | https://api.github.com/repos/KieronQuinn/TapTap | opened | Do not make issues about the service being killed by the system (they will be closed) | Service Killed | Due to OEMs clearly not caring and killing every background service in sight (including accessibility services - image requiring something like Tap, Tap to be able to _use_ your phone, only for it to be killed constantly) in the name of "battery optimisation", there has been an influx of issues related to the app being killed by the system.
**Please make sure you have followed the steps in the app to help prevent the service being killed by the system, including those on https://dontkillmyapp.com/, this is the only known way to prevent the app being killed**. It is also recommended you reboot your phone after running the steps for good measure, to reset any settings that may be cached.
Beyond this, there is no magic code that can make the system not kill the app, so I **cannot** do anything about this. Therefore, issues related to this will be tagged with "Service Killed" and be closed. | 1.0 | Do not make issues about the service being killed by the system (they will be closed) - Due to OEMs clearly not caring and killing every background service in sight (including accessibility services - image requiring something like Tap, Tap to be able to _use_ your phone, only for it to be killed constantly) in the name of "battery optimisation", there has been an influx of issues related to the app being killed by the system.
**Please make sure you have followed the steps in the app to help prevent the service being killed by the system, including those on https://dontkillmyapp.com/, this is the only known way to prevent the app being killed**. It is also recommended you reboot your phone after running the steps for good measure, to reset any settings that may be cached.
Beyond this, there is no magic code that can make the system not kill the app, so I **cannot** do anything about this. Therefore, issues related to this will be tagged with "Service Killed" and be closed. | non_main | do not make issues about the service being killed by the system they will be closed due to oems clearly not caring and killing every background service in sight including accessibility services image requiring something like tap tap to be able to use your phone only for it to be killed constantly in the name of battery optimisation there has been an influx of issues related to the app being killed by the system please make sure you have followed the steps in the app to help prevent the service being killed by the system including those on this is the only known way to prevent the app being killed it is also recommended you reboot your phone after running the steps for good measure to reset any settings that may be cached beyond this there is no magic code that can make the system not kill the app so i cannot do anything about this therefore issues related to this will be tagged with service killed and be closed | 0 |
434,957 | 12,529,753,576 | IssuesEvent | 2020-06-04 11:55:44 | weaveworks/eksctl | https://api.github.com/repos/weaveworks/eksctl | closed | asgMetricsCollection without a Metrics list fails to deploy | kind/bug priority/important-soon | **What happened?**
Attempting to create a NodeGroup that collects all metrics by default at 1Minute intervals failed:
```
[✖] creating CloudFormation stack "eksctl-clusterName-nodegroup-groupName": ValidationError: [/Resources/NodeGroup/Type/MetricsCollection/0/Metrics] 'null' values are not allowed in templates
```
Given the following in my NodeGroup:
```
asgMetricsCollection:
- granularity: 1Minute
```
Which is consistent with the [eksctl config schema](https://eksctl.io/usage/schema/) that lists only `granularity` as required, and the [CloudFormation schema documentation](https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-as-metricscollection.html#cfn-as-metricscollection-metrics) that if you omit Metrics, all metrics are enabled.
**What you expected to happen?**
The NodeGroup created successfully, and all ASG metrics are collected by CloudWatch.
**How to reproduce it?**
Create a nodegroup using the `asgMetricsCollection` as defined above.
**Anything else we need to know?**
Probably just a simple fix to [`metricsCollectionResource`](https://github.com/weaveworks/eksctl/blob/c84ce3d172c5e4ceef75dbb237d9b5717644e237/pkg/cfn/builder/nodegroup.go#L368) to not add the `"Metrics"` entry if `m.Metrics` was omitted.
I suspect an empty array will in fact disable all metrics, so the only workaround is to use the full list of metrics.
**Versions**
Please paste in the output of these commands:
```
PS C:\Projects\AWSPerfect> eksctl version
0.20.0
PS C:\Projects\AWSPerfect> kubectl version
Client Version: version.Info{Major:"1", Minor:"15", GitVersion:"v1.15.11", GitCommit:"d94a81c724ea8e1ccc9002d89b7fe81d58f89ede", GitTreeState:"clean", BuildDate:"2020-03-12T21:08:59Z", GoVersion:"go1.12.17", Compiler:"gc", Platform:"windows/amd64"}
Server Version: version.Info{Major:"1", Minor:"15+", GitVersion:"v1.15.11-eks-af3caf", GitCommit:"af3caf6136cd355f467083651cc1010a499f59b1", GitTreeState:"clean", BuildDate:"2020-03-27T21:51:36Z", GoVersion:"go1.12.17", Compiler:"gc", Platform:"linux/amd64"}
```
**Logs**
Error received is included above. | 1.0 | asgMetricsCollection without a Metrics list fails to deploy - **What happened?**
Attempting to create a NodeGroup that collects all metrics by default at 1Minute intervals failed:
```
[✖] creating CloudFormation stack "eksctl-clusterName-nodegroup-groupName": ValidationError: [/Resources/NodeGroup/Type/MetricsCollection/0/Metrics] 'null' values are not allowed in templates
```
Given the following in my NodeGroup:
```
asgMetricsCollection:
- granularity: 1Minute
```
Which is consistent with the [eksctl config schema](https://eksctl.io/usage/schema/) that lists only `granularity` as required, and the [CloudFormation schema documentation](https://docs.aws.amazon.com/AWSCloudFormation/latest/UserGuide/aws-properties-as-metricscollection.html#cfn-as-metricscollection-metrics) that if you omit Metrics, all metrics are enabled.
**What you expected to happen?**
The NodeGroup created successfully, and all ASG metrics are collected by CloudWatch.
**How to reproduce it?**
Create a nodegroup using the `asgMetricsCollection` as defined above.
**Anything else we need to know?**
Probably just a simple fix to [`metricsCollectionResource`](https://github.com/weaveworks/eksctl/blob/c84ce3d172c5e4ceef75dbb237d9b5717644e237/pkg/cfn/builder/nodegroup.go#L368) to not add the `"Metrics"` entry if `m.Metrics` was omitted.
I suspect an empty array will in fact disable all metrics, so the only workaround is to use the full list of metrics.
**Versions**
Please paste in the output of these commands:
```
PS C:\Projects\AWSPerfect> eksctl version
0.20.0
PS C:\Projects\AWSPerfect> kubectl version
Client Version: version.Info{Major:"1", Minor:"15", GitVersion:"v1.15.11", GitCommit:"d94a81c724ea8e1ccc9002d89b7fe81d58f89ede", GitTreeState:"clean", BuildDate:"2020-03-12T21:08:59Z", GoVersion:"go1.12.17", Compiler:"gc", Platform:"windows/amd64"}
Server Version: version.Info{Major:"1", Minor:"15+", GitVersion:"v1.15.11-eks-af3caf", GitCommit:"af3caf6136cd355f467083651cc1010a499f59b1", GitTreeState:"clean", BuildDate:"2020-03-27T21:51:36Z", GoVersion:"go1.12.17", Compiler:"gc", Platform:"linux/amd64"}
```
**Logs**
Error received is included above. | non_main | asgmetricscollection without a metrics list fails to deploy what happened attempting to create a nodegroup that collects all metrics by default at intervals failed creating cloudformation stack eksctl clustername nodegroup groupname validationerror null values are not allowed in templates given the following in my nodegroup asgmetricscollection granularity which is consistent with the that lists only granularity as required and the that if you omit metrics all metrics are enabled what you expected to happen the nodegroup created successfully and all asg metrics are collected by cloudwatch how to reproduce it create a nodegroup using the asgmetricscollection as defined above anything else we need to know probably just a simple fix to to not add the metrics entry if m metrics was omitted i suspect an empty array will in fact disable all metrics so the only workaround is to use the full list of metrics versions please paste in the output of these commands ps c projects awsperfect eksctl version ps c projects awsperfect kubectl version client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform windows server version version info major minor gitversion eks gitcommit gittreestate clean builddate goversion compiler gc platform linux logs error received is included above | 0 |
4,952 | 25,455,552,888 | IssuesEvent | 2022-11-24 13:55:27 | pace/bricks | https://api.github.com/repos/pace/bricks | closed | service generation: create docker-compose.yml | T::Maintainance | When generating a new service, also generate a docker-compose.yml for easily spinning up the service and its dependencies (database, ...). | True | service generation: create docker-compose.yml - When generating a new service, also generate a docker-compose.yml for easily spinning up the service and its dependencies (database, ...). | main | service generation create docker compose yml when generating a new service also generate a docker compose yml for easily spinning up the service and its dependencies database | 1 |
4,051 | 18,922,275,650 | IssuesEvent | 2021-11-17 04:10:16 | filecoin-project/filecoin-docs | https://api.github.com/repos/filecoin-project/filecoin-docs | closed | APIstruct import in Go JSON-RPC example no longer exists. | help wanted dif/hard effort/hours kind/bug kind/maintenance need/maintainers-input P1 status/ready topic/docs project/lotus | "github.com/filecoin-project/lotus/api/apistruct"
imported in the example seems to not exist anymore.
i will confirm if the example still works or not | True | APIstruct import in Go JSON-RPC example no longer exists. - "github.com/filecoin-project/lotus/api/apistruct"
imported in the example seems to not exist anymore.
i will confirm if the example still works or not | main | apistruct import in go json rpc example no longer exists github com filecoin project lotus api apistruct imported in the example seems to not exist anymore i will confirm if the example still works or not | 1 |
4,315 | 21,717,575,258 | IssuesEvent | 2022-05-10 19:34:18 | BioArchLinux/Packages | https://api.github.com/repos/BioArchLinux/Packages | closed | [MAINTAIN] r-bloodgen3module | maintain | <!--
Please report the error of one package in one issue! Use multi issues to report multi bugs.
Thanks!
-->
**Log of the bug**
<details>
```
[put the output here](https://log.bioarchlinux.org/2022-04-28T16%3A00%3A12/r-bloodgen3module.log#:~:text=**%20byte%2Dcompile%20and,dyn.load%0AExecution%20halted)
```
</details>
**Packages (please complete the following information):**
- Package Name: [e.g. iqtree]
**Description**
Add any other context about the problem here.
| True | [MAINTAIN] r-bloodgen3module - <!--
Please report the error of one package in one issue! Use multi issues to report multi bugs.
Thanks!
-->
**Log of the bug**
<details>
```
[put the output here](https://log.bioarchlinux.org/2022-04-28T16%3A00%3A12/r-bloodgen3module.log#:~:text=**%20byte%2Dcompile%20and,dyn.load%0AExecution%20halted)
```
</details>
**Packages (please complete the following information):**
- Package Name: [e.g. iqtree]
**Description**
Add any other context about the problem here.
| main | r please report the error of one package in one issue use multi issues to report multi bugs thanks log of the bug packages please complete the following information package name description add any other context about the problem here | 1 |
146,896 | 11,761,062,522 | IssuesEvent | 2020-03-13 21:01:22 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | Manual test run on Linux for 1.5.x - Release | OS/Linux QA/Yes release-notes/exclude tests | ## Per release specialty tests
- [x] Implement cosmetic blocking [tracking].([#5381](https://github.com/brave/brave-browser/issues/5381))
- [ ] Implement ad conversion attribution for Brave Ads.([#6536](https://github.com/brave/brave-browser/issues/6536))
- [ ] Improve compatibility with MetaMask and allow web3 provider selection.([#7503](https://github.com/brave/brave-browser/issues/7503))
- [x] Intermittent crash if Confirmations library is called before being instantiated.([#7728](https://github.com/brave/brave-browser/issues/7728))
### Installer
- [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window
### Data(Upgrade from previous release)
- [x] Make sure that data from the last version appears in the new version OK
- [x] With data from the last version, verify that
- [x] bookmarks on the bookmark toolbar and bookmark folders can be opened
- [x] cookies are preserved
- [x] installed extensions are retained and work correctly
- [x] opened tabs can be reloaded
- [x] stored passwords are preserved
- [x] unpinned tabs can be pinned
## Extensions/Plugins tests
- [x] Verify one item from Brave Update server is installable (Example: Ad-block DAT file on fresh extension)
- [x] Verify one item from Google Update server is installable (Example: Extensions from CWS)
- [x] Verify PDFJS, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled
- [x] Verify magnet links and .torrent files loads Torrent viewer page and able to download torrent
### CWS
- [x] Verify installing ABP from CWS shows warning message `NOT A RECOMMENDED BRAVE EXTENSION!` but still allows to install the extension
- [x] Verify installing LastPass from CWS doesn't show any warning message
### PDF
- [x] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf
- [x] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf
### Widevine
- [x] Verify `Widevine Notification` is shown when you visit Netflix for the first time
- [x] Test that you can stream on Netflix on a fresh profile after installing Widevine
### Bravery settings
- [x] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/
- [x] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/
- [x] Verify that toggling `Ads and trackers blocked` works as expected
- [ ] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items
- [x] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly
- [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked
### Fingerprint Tests
- [x] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00`
- [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on
- [x] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/
- [x] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on
### Rewards
- [x] Verify wallet is auto created after enabling rewards
- [x] Verify account balance shows correct BAT and USD value
- [x] Verify you are able to restore a wallet
- [x] Verify wallet address matches the QR code that is generated under `Add funds`
- [x] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel
- [ ] Verify adding funds via any of the currencies flows into wallet after specified amount of time
- [ ] Verify adding funds to an existing wallet with amount, adjusts the BAT value appropriately
- [x] Verify monthly budget shows correct BAT and USD value
- [x] Verify you are able to exclude a publisher from the auto-contribute table by clicking on the `x` in auto-contribute table and popup list of sites
- [x] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel
- [x] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %)
- [x] Verify when you click on `Send a tip`, the custom tip banner displays
- [x] Verify you are able to make one-time tip and they display in tips panel
- [x] Verify you are able to make recurring tip and they display in tips panel
- [x] Verify you can tip a verified publisher
- [x] Verify you can tip a verified YouTube creator
- [x] Verify tip panel shows a verified checkmark for a verified publisher/verified YouTube creator
- [x] Verify tip panel shows a message about unverified publisher
- [x] Verify BR panel shows message about an unverified publisher
- [x] Verify you are able to perform a contribution
- [x] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators
- [x] Verify that disabling Rewards and enabling it again does not lose state
- [x] Verify that disabling auto-contribute and enabling it again does not lose state
- [x] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings
- [x] Upgrade from older version
- [x] Verify the wallet balance is retained and wallet backup code isn't corrupted
- [x] Verify auto-contribute list is not lost after upgrade
- [x] Verify tips list is not lost after upgrade
- [x] Verify wallet panel transactions list is not lost after upgrade
### Ads Upgrade Tests:
- [x] Install 0.62.51 and enable Rewards (Ads are not available on this version). Update on `test` channel to the hotfix version. Verify Ads are off by default, should get a BAT logo notification to alert you that Ads are available.
- [x] Install 0.64.77 and enable Rewards. Ads are on by default. View an Ad. Update on `test` channel to the hotfix version. Verify Ads are still on after update, Ads panel information was not lost after upgrade, no BAT logo notification.
- [x] Install 0.64.77 and enable Rewards. Disable Ads. Update on `test` channel to the hotfix version. Verify Ads are still off after update, no BAT logo notification.
- [x] Install 1.4.95 and enable Rewards. Ads are on by default. View an ad. Update on `test` channel to the hotfix version. Verify Ads are still on after update, Ads panel information was not lost after upgrade, no BAT logo notification.
- [x] install 1.4.95 and enable Rewards. Disable Ads. Update on `test` channel to the hotfix version. Verify Ads are still off after update, no BAT logo notification.
### Tor Tabs
- [x] Visit https://check.torproject.org in a Tor window, ensure its shows success message for using a Tor exit node
- [x] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads
- [x] Visit https://protonirockerxow.onion/ in a Tor window, ensure login page is shown
- [x] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown
### Session storage
- [x] Temporarily move away your browser profile and test that a new profile is created when browser is launched
- macOS - `~/Library/Application\ Support/BraveSoftware/`
- Windows - `%userprofile%\appdata\Local\BraveSoftware\`
- Linux(Ubuntu) - `~/.config/BraveSoftware/`
- [x] Test that windows and tabs restore when closed, including active tab
- [x] Ensure that the tabs in the above session are being lazy loaded when the session is restored
## Update tests
- [x] Verify visiting `brave://settings/help` triggers update check
- [ ] Verify once update is downloaded, prompts to `Relaunch` to install update
## Chromium upgrade tests
- [x] Verify `brave://gpu` on Brave and `chrome://gpu` on Chrome are similar for the same Chromium version on both browsers
#### Adblock
- [x] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/`
#### Components
- [x] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components
## Crypto Wallets
- [x] ensure that you can create a new wallet without any issues
- [x] ensure that you can restore a previous CW wallet without any issues
- [x] ensure that you can restore a previous MM wallet without any issues
- [x] ensure that you can create a transaction (sending crypto) with a CW wallet
- [x] ensure that you can create a transaction (sending crypto) using a restored MM wallet | 1.0 | Manual test run on Linux for 1.5.x - Release - ## Per release specialty tests
- [x] Implement cosmetic blocking [tracking].([#5381](https://github.com/brave/brave-browser/issues/5381))
- [ ] Implement ad conversion attribution for Brave Ads.([#6536](https://github.com/brave/brave-browser/issues/6536))
- [ ] Improve compatibility with MetaMask and allow web3 provider selection.([#7503](https://github.com/brave/brave-browser/issues/7503))
- [x] Intermittent crash if Confirmations library is called before being instantiated.([#7728](https://github.com/brave/brave-browser/issues/7728))
### Installer
- [x] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window
### Data(Upgrade from previous release)
- [x] Make sure that data from the last version appears in the new version OK
- [x] With data from the last version, verify that
- [x] bookmarks on the bookmark toolbar and bookmark folders can be opened
- [x] cookies are preserved
- [x] installed extensions are retained and work correctly
- [x] opened tabs can be reloaded
- [x] stored passwords are preserved
- [x] unpinned tabs can be pinned
## Extensions/Plugins tests
- [x] Verify one item from Brave Update server is installable (Example: Ad-block DAT file on fresh extension)
- [x] Verify one item from Google Update server is installable (Example: Extensions from CWS)
- [x] Verify PDFJS, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled
- [x] Verify magnet links and .torrent files loads Torrent viewer page and able to download torrent
### CWS
- [x] Verify installing ABP from CWS shows warning message `NOT A RECOMMENDED BRAVE EXTENSION!` but still allows to install the extension
- [x] Verify installing LastPass from CWS doesn't show any warning message
### PDF
- [x] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf
- [x] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf
### Widevine
- [x] Verify `Widevine Notification` is shown when you visit Netflix for the first time
- [x] Test that you can stream on Netflix on a fresh profile after installing Widevine
### Bravery settings
- [x] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/
- [x] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/
- [x] Verify that toggling `Ads and trackers blocked` works as expected
- [ ] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items
- [x] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly
- [x] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked
### Fingerprint Tests
- [x] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00`
- [x] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on
- [x] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/
- [x] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on
### Rewards
- [x] Verify wallet is auto created after enabling rewards
- [x] Verify account balance shows correct BAT and USD value
- [x] Verify you are able to restore a wallet
- [x] Verify wallet address matches the QR code that is generated under `Add funds`
- [x] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel
- [ ] Verify adding funds via any of the currencies flows into wallet after specified amount of time
- [ ] Verify adding funds to an existing wallet with amount, adjusts the BAT value appropriately
- [x] Verify monthly budget shows correct BAT and USD value
- [x] Verify you are able to exclude a publisher from the auto-contribute table by clicking on the `x` in auto-contribute table and popup list of sites
- [x] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel
- [x] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %)
- [x] Verify when you click on `Send a tip`, the custom tip banner displays
- [x] Verify you are able to make one-time tip and they display in tips panel
- [x] Verify you are able to make recurring tip and they display in tips panel
- [x] Verify you can tip a verified publisher
- [x] Verify you can tip a verified YouTube creator
- [x] Verify tip panel shows a verified checkmark for a verified publisher/verified YouTube creator
- [x] Verify tip panel shows a message about unverified publisher
- [x] Verify BR panel shows message about an unverified publisher
- [x] Verify you are able to perform a contribution
- [x] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators
- [x] Verify that disabling Rewards and enabling it again does not lose state
- [x] Verify that disabling auto-contribute and enabling it again does not lose state
- [x] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings
- [x] Upgrade from older version
- [x] Verify the wallet balance is retained and wallet backup code isn't corrupted
- [x] Verify auto-contribute list is not lost after upgrade
- [x] Verify tips list is not lost after upgrade
- [x] Verify wallet panel transactions list is not lost after upgrade
### Ads Upgrade Tests:
- [x] Install 0.62.51 and enable Rewards (Ads are not available on this version). Update on `test` channel to the hotfix version. Verify Ads are off by default, should get a BAT logo notification to alert you that Ads are available.
- [x] Install 0.64.77 and enable Rewards. Ads are on by default. View an Ad. Update on `test` channel to the hotfix version. Verify Ads are still on after update, Ads panel information was not lost after upgrade, no BAT logo notification.
- [x] Install 0.64.77 and enable Rewards. Disable Ads. Update on `test` channel to the hotfix version. Verify Ads are still off after update, no BAT logo notification.
- [x] Install 1.4.95 and enable Rewards. Ads are on by default. View an ad. Update on `test` channel to the hotfix version. Verify Ads are still on after update, Ads panel information was not lost after upgrade, no BAT logo notification.
- [x] install 1.4.95 and enable Rewards. Disable Ads. Update on `test` channel to the hotfix version. Verify Ads are still off after update, no BAT logo notification.
### Tor Tabs
- [x] Visit https://check.torproject.org in a Tor window, ensure its shows success message for using a Tor exit node
- [x] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads
- [x] Visit https://protonirockerxow.onion/ in a Tor window, ensure login page is shown
- [x] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown
### Session storage
- [x] Temporarily move away your browser profile and test that a new profile is created when browser is launched
- macOS - `~/Library/Application\ Support/BraveSoftware/`
- Windows - `%userprofile%\appdata\Local\BraveSoftware\`
- Linux(Ubuntu) - `~/.config/BraveSoftware/`
- [x] Test that windows and tabs restore when closed, including active tab
- [x] Ensure that the tabs in the above session are being lazy loaded when the session is restored
## Update tests
- [x] Verify visiting `brave://settings/help` triggers update check
- [ ] Verify once update is downloaded, prompts to `Relaunch` to install update
## Chromium upgrade tests
- [x] Verify `brave://gpu` on Brave and `chrome://gpu` on Chrome are similar for the same Chromium version on both browsers
#### Adblock
- [x] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/`
#### Components
- [x] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components
## Crypto Wallets
- [x] ensure that you can create a new wallet without any issues
- [x] ensure that you can restore a previous CW wallet without any issues
- [x] ensure that you can restore a previous MM wallet without any issues
- [x] ensure that you can create a transaction (sending crypto) with a CW wallet
- [x] ensure that you can create a transaction (sending crypto) using a restored MM wallet | non_main | manual test run on linux for x release per release specialty tests implement cosmetic blocking implement ad conversion attribution for brave ads improve compatibility with metamask and allow provider selection intermittent crash if confirmations library is called before being instantiated installer check signature if os run spctl assess verbose applications brave browser beta app and make sure it returns accepted if windows right click on the brave installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window data upgrade from previous release make sure that data from the last version appears in the new version ok with data from the last version verify that bookmarks on the bookmark toolbar and bookmark folders can be opened cookies are preserved installed extensions are retained and work correctly opened tabs can be reloaded stored passwords are preserved unpinned tabs can be pinned extensions plugins tests verify one item from brave update server is installable example ad block dat file on fresh extension verify one item from google update server is installable example extensions from cws verify pdfjs torrent viewer extensions are installed automatically on fresh profile and cannot be disabled verify magnet links and torrent files loads torrent viewer page and able to download torrent cws verify installing abp from cws shows warning message not a recommended brave extension but still allows to install the extension verify installing lastpass from cws doesn t show any warning message pdf test that pdf is loaded over https at test that pdf is loaded over http at widevine verify widevine notification is shown when you visit netflix for the first time test that you can stream on netflix on a fresh profile after installing widevine bravery settings verify that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to verify that toggling ads and trackers blocked works as expected visit verify that safe browsing via our proxy works for all the listed items visit and then turn on script blocking page should not load allow it from the script blocking ui in the url bar and it should load the page correctly test that party storage results are blank at when party cookies are blocked and not blank when party cookies are unblocked fingerprint tests visit ensure blocked items are listed in shields result window should show got canvas fingerprint and got webgl fingerprint test that audio fingerprint is blocked at only when block all fingerprinting protection is on test that brave browser isn t detected on test that doesn t leak ip address when block all fingerprinting protection is on rewards verify wallet is auto created after enabling rewards verify account balance shows correct bat and usd value verify you are able to restore a wallet verify wallet address matches the qr code that is generated under add funds verify actions taken claiming grant tipping auto contribute display in wallet panel verify adding funds via any of the currencies flows into wallet after specified amount of time verify adding funds to an existing wallet with amount adjusts the bat value appropriately verify monthly budget shows correct bat and usd value verify you are able to exclude a publisher from the auto contribute table by clicking on the x in auto contribute table and popup list of sites verify you are able to exclude a publisher by using the toggle on the rewards panel verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify when you click on send a tip the custom tip banner displays verify you are able to make one time tip and they display in tips panel verify you are able to make recurring tip and they display in tips panel verify you can tip a verified publisher verify you can tip a verified youtube creator verify tip panel shows a verified checkmark for a verified publisher verified youtube creator verify tip panel shows a message about unverified publisher verify br panel shows message about an unverified publisher verify you are able to perform a contribution verify if you disable auto contribute you are still able to tip regular sites and youtube creators verify that disabling rewards and enabling it again does not lose state verify that disabling auto contribute and enabling it again does not lose state adjust min visit time in settings visit some sites and youtube channels to verify they are added to the table after the specified settings upgrade from older version verify the wallet balance is retained and wallet backup code isn t corrupted verify auto contribute list is not lost after upgrade verify tips list is not lost after upgrade verify wallet panel transactions list is not lost after upgrade ads upgrade tests install and enable rewards ads are not available on this version update on test channel to the hotfix version verify ads are off by default should get a bat logo notification to alert you that ads are available install and enable rewards ads are on by default view an ad update on test channel to the hotfix version verify ads are still on after update ads panel information was not lost after upgrade no bat logo notification install and enable rewards disable ads update on test channel to the hotfix version verify ads are still off after update no bat logo notification install and enable rewards ads are on by default view an ad update on test channel to the hotfix version verify ads are still on after update ads panel information was not lost after upgrade no bat logo notification install and enable rewards disable ads update on test channel to the hotfix version verify ads are still off after update no bat logo notification tor tabs visit in a tor window ensure its shows success message for using a tor exit node visit in a tor window note down exit node ip address do a hard refresh ctrl shift r cmd shift r ensure exit ip changes after page reloads visit in a tor window ensure login page is shown visit in a tor window ensure location isn t shown session storage temporarily move away your browser profile and test that a new profile is created when browser is launched macos library application support bravesoftware windows userprofile appdata local bravesoftware linux ubuntu config bravesoftware test that windows and tabs restore when closed including active tab ensure that the tabs in the above session are being lazy loaded when the session is restored update tests verify visiting brave settings help triggers update check verify once update is downloaded prompts to relaunch to install update chromium upgrade tests verify brave gpu on brave and chrome gpu on chrome are similar for the same chromium version on both browsers adblock verify referrer blocking works properly for tld visit and verify adblock works properly similar to components delete adblock folder from browser profile and restart browser visit brave components and verify brave ad block updater downloads and update the component repeat for all brave components crypto wallets ensure that you can create a new wallet without any issues ensure that you can restore a previous cw wallet without any issues ensure that you can restore a previous mm wallet without any issues ensure that you can create a transaction sending crypto with a cw wallet ensure that you can create a transaction sending crypto using a restored mm wallet | 0 |
32,895 | 6,133,223,622 | IssuesEvent | 2017-06-25 11:52:46 | TRIQS/triqs | https://api.github.com/repos/TRIQS/triqs | closed | Installation on Archlinux | Documentation | Install the following packages (TRIQS uses Python 2 but the default Python version on Archlinux is 3):
`yaourt -S boost boost-libs cmake git gcc gcc-fortran openmpi blas lapack fftw gmp hdf5 python2 python2-h5py python2-numpy python2-scipy python2-jinja python2-matplotlib python2-tornado python2-pyzmq python2-mako python2-mpi4py`
Then follow the usual installation, but add as argument to cmake :
`cmake -DCMAKE_INSTALL_PREFIX=path_to_install_directory ../triqs.src -DPYTHON_INTERPRETER=python2`
And after make and make install, do not use the deprecated ipytriqs_notebook, but first add the path of the TRIQS libraries :
`export PYTHONPATH=/your_path/INSTALL_DIR/lib/python2.7/site-packages:$PYTHONPATH
`
And then load notebook using
`jupyter notebook` and start a Python2 kernel
| 1.0 | Installation on Archlinux - Install the following packages (TRIQS uses Python 2 but the default Python version on Archlinux is 3):
`yaourt -S boost boost-libs cmake git gcc gcc-fortran openmpi blas lapack fftw gmp hdf5 python2 python2-h5py python2-numpy python2-scipy python2-jinja python2-matplotlib python2-tornado python2-pyzmq python2-mako python2-mpi4py`
Then follow the usual installation, but add as argument to cmake :
`cmake -DCMAKE_INSTALL_PREFIX=path_to_install_directory ../triqs.src -DPYTHON_INTERPRETER=python2`
And after make and make install, do not use the deprecated ipytriqs_notebook, but first add the path of the TRIQS libraries :
`export PYTHONPATH=/your_path/INSTALL_DIR/lib/python2.7/site-packages:$PYTHONPATH
`
And then load notebook using
`jupyter notebook` and start a Python2 kernel
| non_main | installation on archlinux install the following packages triqs uses python but the default python version on archlinux is yaourt s boost boost libs cmake git gcc gcc fortran openmpi blas lapack fftw gmp numpy scipy jinja matplotlib tornado pyzmq mako then follow the usual installation but add as argument to cmake cmake dcmake install prefix path to install directory triqs src dpython interpreter and after make and make install do not use the deprecated ipytriqs notebook but first add the path of the triqs libraries export pythonpath your path install dir lib site packages pythonpath and then load notebook using jupyter notebook and start a kernel | 0 |
6,306 | 4,216,730,275 | IssuesEvent | 2016-06-30 10:18:25 | ff36/halo-gui | https://api.github.com/repos/ff36/halo-gui | closed | Add Sector to network properties | change usability | Its not clear which sector the selected properties belong to. Lets add the sector to the properties display:

| True | Add Sector to network properties - Its not clear which sector the selected properties belong to. Lets add the sector to the properties display:

| non_main | add sector to network properties its not clear which sector the selected properties belong to lets add the sector to the properties display | 0 |
2,190 | 7,734,733,056 | IssuesEvent | 2018-05-27 06:09:25 | jenkinsci/ghprb-plugin | https://api.github.com/repos/jenkinsci/ghprb-plugin | opened | Release 1.41.0 | maintainer communication | This issue is to track the 1.41.0 release.
I have successfully tested upgrading from 1.40.0 stable to the master development branch.
Next up is to release the next version with CHANGELOG release notes. | True | Release 1.41.0 - This issue is to track the 1.41.0 release.
I have successfully tested upgrading from 1.40.0 stable to the master development branch.
Next up is to release the next version with CHANGELOG release notes. | main | release this issue is to track the release i have successfully tested upgrading from stable to the master development branch next up is to release the next version with changelog release notes | 1 |
328,209 | 28,105,617,798 | IssuesEvent | 2023-03-31 00:10:41 | Azure/azure-sdk-for-java | https://api.github.com/repos/Azure/azure-sdk-for-java | closed | Container Registry ReadMe Issue | Container Registry Client Docs needs-team-triage test-manual-pass | 1.
**Section** [link](https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/containerregistry/azure-containers-containerregistry#sync-examples)

**Suggestion:**
1. Link change to `https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/containerregistry/azure-containers-containerregistry#list-artifact-tags-with-anonymous-access`
2. Link change to `https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/containerregistry/azure-containers-containerregistry#upload`
@joshfree, @Mohit-Chakraborty, @mario-guerra, and @alzimmermsft for notification. | 1.0 | Container Registry ReadMe Issue - 1.
**Section** [link](https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/containerregistry/azure-containers-containerregistry#sync-examples)

**Suggestion:**
1. Link change to `https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/containerregistry/azure-containers-containerregistry#list-artifact-tags-with-anonymous-access`
2. Link change to `https://github.com/Azure/azure-sdk-for-java/tree/main/sdk/containerregistry/azure-containers-containerregistry#upload`
@joshfree, @Mohit-Chakraborty, @mario-guerra, and @alzimmermsft for notification. | non_main | container registry readme issue section suggestion link change to link change to joshfree mohit chakraborty mario guerra and alzimmermsft for notification | 0 |
218,156 | 16,960,086,212 | IssuesEvent | 2021-06-29 01:41:09 | anhdtqwerty/thpt | https://api.github.com/repos/anhdtqwerty/thpt | closed | Major | Quản lý Bộ môn | Thêm Bộ môn | Thêm và hiển thị thành công bộ môn mới bị trùng | dev-done test-verified | Thêm Bộ môn mới
Step:
1. Click "Thêm Bộ môn"
2. Nhập bộ môn có tên bị trùng hoặc có chứa dấu space vị trí đầu/ cuối
3. Bấm "Lưu"
Actual:
Thêm và hiển thị thành công bộ môn mới có tên bị trùng hoặc chứa space đầu/ cuối
Expect:
Thêm mới không thành công
Thông báo "Bộ môn đã tồn tại"
<img width="1320" alt="BM1" src="https://user-images.githubusercontent.com/82717419/121237176-f6832d00-c8c0-11eb-8811-c227bb2ee264.png">
<img width="1061" alt="BM2" src="https://user-images.githubusercontent.com/82717419/121237180-f7b45a00-c8c0-11eb-828a-540c785761e2.png">
<img width="1309" alt="BM3" src="https://user-images.githubusercontent.com/82717419/121237184-f8e58700-c8c0-11eb-81f7-a877ed0df9f3.png">
| 1.0 | Major | Quản lý Bộ môn | Thêm Bộ môn | Thêm và hiển thị thành công bộ môn mới bị trùng - Thêm Bộ môn mới
Step:
1. Click "Thêm Bộ môn"
2. Nhập bộ môn có tên bị trùng hoặc có chứa dấu space vị trí đầu/ cuối
3. Bấm "Lưu"
Actual:
Thêm và hiển thị thành công bộ môn mới có tên bị trùng hoặc chứa space đầu/ cuối
Expect:
Thêm mới không thành công
Thông báo "Bộ môn đã tồn tại"
<img width="1320" alt="BM1" src="https://user-images.githubusercontent.com/82717419/121237176-f6832d00-c8c0-11eb-8811-c227bb2ee264.png">
<img width="1061" alt="BM2" src="https://user-images.githubusercontent.com/82717419/121237180-f7b45a00-c8c0-11eb-828a-540c785761e2.png">
<img width="1309" alt="BM3" src="https://user-images.githubusercontent.com/82717419/121237184-f8e58700-c8c0-11eb-81f7-a877ed0df9f3.png">
| non_main | major quản lý bộ môn thêm bộ môn thêm và hiển thị thành công bộ môn mới bị trùng thêm bộ môn mới step click thêm bộ môn nhập bộ môn có tên bị trùng hoặc có chứa dấu space vị trí đầu cuối bấm lưu actual thêm và hiển thị thành công bộ môn mới có tên bị trùng hoặc chứa space đầu cuối expect thêm mới không thành công thông báo bộ môn đã tồn tại img width alt src img width alt src img width alt src | 0 |
830 | 4,467,637,507 | IssuesEvent | 2016-08-25 06:04:31 | duckduckgo/zeroclickinfo-fathead | https://api.github.com/repos/duckduckgo/zeroclickinfo-fathead | opened | Abbreviations: | Maintainer Input Requested |
On searching for `integrity meaning`, the url defaults to about and shows the abbreviation instead of the definition which seems like the intended query.
The default behavior in this case is more likely to show the definition or the dictionary meaning of the word.
------
IA Page: http://duck.co/ia/view/abbreviations
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @stands4 | True | Abbreviations: -
On searching for `integrity meaning`, the url defaults to about and shows the abbreviation instead of the definition which seems like the intended query.
The default behavior in this case is more likely to show the definition or the dictionary meaning of the word.
------
IA Page: http://duck.co/ia/view/abbreviations
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @stands4 | main | abbreviations on searching for integrity meaning the url defaults to about and shows the abbreviation instead of the definition which seems like the intended query the default behavior in this case is more likely to show the definition or the dictionary meaning of the word ia page | 1 |
706,540 | 24,276,375,106 | IssuesEvent | 2022-09-28 14:07:58 | nanoframework/Home | https://api.github.com/repos/nanoframework/Home | closed | Dead code in JsonToken class? | Status: Waiting triage Priority: Low Area: CL-nanoFramework.Json | ### Library/API/IoT binding
nanoFramework.Json
### Visual Studio version
_No response_
### .NET nanoFramework extension version
_No response_
### Target name(s)
_No response_
### Firmware version
_No response_
### Device capabilities
_No response_
### Description
ExitSerialization and EnterSerialization in JsonToken methods looks like not used. Investigate it and if possible remove.
### How to reproduce
_No response_
### Expected behaviour
_No response_
### Screenshots
_No response_
### Sample project or code
_No response_
### Aditional information
_No response_ | 1.0 | Dead code in JsonToken class? - ### Library/API/IoT binding
nanoFramework.Json
### Visual Studio version
_No response_
### .NET nanoFramework extension version
_No response_
### Target name(s)
_No response_
### Firmware version
_No response_
### Device capabilities
_No response_
### Description
ExitSerialization and EnterSerialization in JsonToken methods looks like not used. Investigate it and if possible remove.
### How to reproduce
_No response_
### Expected behaviour
_No response_
### Screenshots
_No response_
### Sample project or code
_No response_
### Aditional information
_No response_ | non_main | dead code in jsontoken class library api iot binding nanoframework json visual studio version no response net nanoframework extension version no response target name s no response firmware version no response device capabilities no response description exitserialization and enterserialization in jsontoken methods looks like not used investigate it and if possible remove how to reproduce no response expected behaviour no response screenshots no response sample project or code no response aditional information no response | 0 |
181,112 | 14,852,307,163 | IssuesEvent | 2021-01-18 08:24:21 | gbv/login-server | https://api.github.com/repos/gbv/login-server | opened | Add document/page with detailed explanation about Login Server | documentation | From #27:
> Add an end-user friendly help page that explains the purpose and functionality of login-server and single-sign-on with basic concepts such as
>
> Identity providers (e.g. ORCID)
> Identities
> Client applications (Cocoda)
> Sessions
> Service providers (e.g. jskos-server, Wikidata-write-access...)
> API
Either a document inside the repo (that could then be linked from the /help page), or a separate sub-page. | 1.0 | Add document/page with detailed explanation about Login Server - From #27:
> Add an end-user friendly help page that explains the purpose and functionality of login-server and single-sign-on with basic concepts such as
>
> Identity providers (e.g. ORCID)
> Identities
> Client applications (Cocoda)
> Sessions
> Service providers (e.g. jskos-server, Wikidata-write-access...)
> API
Either a document inside the repo (that could then be linked from the /help page), or a separate sub-page. | non_main | add document page with detailed explanation about login server from add an end user friendly help page that explains the purpose and functionality of login server and single sign on with basic concepts such as identity providers e g orcid identities client applications cocoda sessions service providers e g jskos server wikidata write access api either a document inside the repo that could then be linked from the help page or a separate sub page | 0 |
743,508 | 25,901,798,491 | IssuesEvent | 2022-12-15 06:39:24 | hashicorp/terraform-provider-google | https://api.github.com/repos/hashicorp/terraform-provider-google | closed | google_cloudbuild_trigger is missing some parameters | enhancement size/m priority/3 crosslinked | <!--- Please leave this line, it helps our automation: [issue-type:enhancement] --->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment. If the issue is assigned to the "modular-magician" user, it is either in the process of being autogenerated, or is planned to be autogenerated soon. If the issue is assigned to a user, that user is claiming responsibility for the issue. If the issue is assigned to "hashibot", a community member has claimed the issue already.
<!--- Thank you for keeping this note for the community --->
### Description
The resource is missing some parameters that described [here](https://cloud.google.com/cloud-build/docs/build-config#options) (ie: options. artifacts, etc)
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* google_cloudbuild_trigger
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```tf
resource "google_cloudbuild_trigger" "app" {
project = google_project.app.id
description = "Build app"
trigger_template {
project_id = google_project.app.id
repo_name = "app"
tag_name = var.tag_name
}
build {
step {
name = "gcr.io/cloud-builders/docker"
id = "Build"
args = [
"build",
"-f", "Dockerfile",
"-t", "eu.gcr.io/$PROJECT_ID/app:latest",
".",
]
}
images = [
"eu.gcr.io/$PROJECT_ID/app:latest",
]
options = {
machineType = "N1_HIGHCPU_8"
}
}
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
* #4844
| 1.0 | google_cloudbuild_trigger is missing some parameters - <!--- Please leave this line, it helps our automation: [issue-type:enhancement] --->
<!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment. If the issue is assigned to the "modular-magician" user, it is either in the process of being autogenerated, or is planned to be autogenerated soon. If the issue is assigned to a user, that user is claiming responsibility for the issue. If the issue is assigned to "hashibot", a community member has claimed the issue already.
<!--- Thank you for keeping this note for the community --->
### Description
The resource is missing some parameters that described [here](https://cloud.google.com/cloud-build/docs/build-config#options) (ie: options. artifacts, etc)
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* google_cloudbuild_trigger
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```tf
resource "google_cloudbuild_trigger" "app" {
project = google_project.app.id
description = "Build app"
trigger_template {
project_id = google_project.app.id
repo_name = "app"
tag_name = var.tag_name
}
build {
step {
name = "gcr.io/cloud-builders/docker"
id = "Build"
args = [
"build",
"-f", "Dockerfile",
"-t", "eu.gcr.io/$PROJECT_ID/app:latest",
".",
]
}
images = [
"eu.gcr.io/$PROJECT_ID/app:latest",
]
options = {
machineType = "N1_HIGHCPU_8"
}
}
}
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation?
--->
* #4844
| non_main | google cloudbuild trigger is missing some parameters community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment if the issue is assigned to the modular magician user it is either in the process of being autogenerated or is planned to be autogenerated soon if the issue is assigned to a user that user is claiming responsibility for the issue if the issue is assigned to hashibot a community member has claimed the issue already description the resource is missing some parameters that described ie options artifacts etc new or affected resource s google cloudbuild trigger potential terraform configuration tf resource google cloudbuild trigger app project google project app id description build app trigger template project id google project app id repo name app tag name var tag name build step name gcr io cloud builders docker id build args build f dockerfile t eu gcr io project id app latest images eu gcr io project id app latest options machinetype highcpu references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation | 0 |
3,537 | 13,922,851,966 | IssuesEvent | 2020-10-21 13:46:06 | hashicorp/terraform | https://api.github.com/repos/hashicorp/terraform | closed | Puppet provisioner fails to install Puppet on Windows Server 2016 with TLS/SSL Error | bug needs-maintainer provisioner/puppet v0.12 | ### Terraform Version
<!---
Run `terraform -v` to show the version, and paste the result between the ``` marks below.
If you are not running the latest version of Terraform, please try upgrading because your issue may have already been fixed.
-->
```
Terraform v0.12.18
+ provider.aws v2.42.0
```
### Terraform Configuration Files
<!--
Paste the relevant parts of your Terraform configuration between the ``` marks below.
For large Terraform configs, please use a service like Dropbox and share a link to the ZIP file. For security, you can also encrypt the files using our GPG public key.
-->
```hcl
resource "aws_instance" "instance" {
ami = data.aws_ami.windows-server.id
instance_type = var.instance_type
key_name = "provision_key"
availability_zone = "us-east-1b"
tags = {
Name = "example-vm"
}
get_password_data = true
vpc_security_group_ids = ["sg-1111111111"]
subnet_id = "subnet-11111111111"
user_data = <<-DATA
<script>
winrm quickconfig -q & winrm set winrm/config @{MaxTimeoutms="1800000"} & winrm set winrm/config/service @{AllowUnencrypted="true"} & winrm set winrm/config/service/auth @{Basic="true"}
</script>
<powershell>
Write-Host "setting up firewall"
netsh advfirewall firewall add rule name="WinRM in" protocol=TCP dir=in profile=any localport=5985 remoteip=any localip=any action=allow
Write-Host "restarting winrm"
Stop-Service winrm
Start-Service winrm
</powershell>
DATA
provisioner "puppet" {
server = "puppet.test.com"
connection {
host = coalesce(self.private_ip, self.public_ip)
type = "winrm"
user = "Administrator"
password = rsadecrypt(self.password_data, file("~/.ssh/provision_key"))
timeout = "10m"
}
open_source = false
certname = "example.test.com"
autosign = false
}
}
```
### Debug Output
https://gist.github.com/camara-tech/f407b8760def43f6c9da56fc45a88efe
### Crash Output
Not Applicable
### Expected Behavior
The terraform puppet provisioner successfully downloads the puppet agent installer for windows from the puppet enterprise server and successfully installs it.
### Actual Behavior
the terraform puppet provisioner failed with the following error from powershell:
```
The request was aborted: Could not create SSL/TLS secure channel.
```
### Steps to Reproduce
terraform init
terraform apply
### Additional Context
In our environment, we have completely disabled TLS protocols below TLS 1.2. Also, our puppet enterprise console uses the same Powershell that is present in the puppet provisioner for windows with the addition of the following line:
```
[System.Net.ServicePointManager]::SecurityProtocol = [Net.SecurityProtocolType]::Tls12;
```
Perhaps that needs to be added to the existing provisioner?
### References
Not Applicable
| True | Puppet provisioner fails to install Puppet on Windows Server 2016 with TLS/SSL Error - ### Terraform Version
<!---
Run `terraform -v` to show the version, and paste the result between the ``` marks below.
If you are not running the latest version of Terraform, please try upgrading because your issue may have already been fixed.
-->
```
Terraform v0.12.18
+ provider.aws v2.42.0
```
### Terraform Configuration Files
<!--
Paste the relevant parts of your Terraform configuration between the ``` marks below.
For large Terraform configs, please use a service like Dropbox and share a link to the ZIP file. For security, you can also encrypt the files using our GPG public key.
-->
```hcl
resource "aws_instance" "instance" {
ami = data.aws_ami.windows-server.id
instance_type = var.instance_type
key_name = "provision_key"
availability_zone = "us-east-1b"
tags = {
Name = "example-vm"
}
get_password_data = true
vpc_security_group_ids = ["sg-1111111111"]
subnet_id = "subnet-11111111111"
user_data = <<-DATA
<script>
winrm quickconfig -q & winrm set winrm/config @{MaxTimeoutms="1800000"} & winrm set winrm/config/service @{AllowUnencrypted="true"} & winrm set winrm/config/service/auth @{Basic="true"}
</script>
<powershell>
Write-Host "setting up firewall"
netsh advfirewall firewall add rule name="WinRM in" protocol=TCP dir=in profile=any localport=5985 remoteip=any localip=any action=allow
Write-Host "restarting winrm"
Stop-Service winrm
Start-Service winrm
</powershell>
DATA
provisioner "puppet" {
server = "puppet.test.com"
connection {
host = coalesce(self.private_ip, self.public_ip)
type = "winrm"
user = "Administrator"
password = rsadecrypt(self.password_data, file("~/.ssh/provision_key"))
timeout = "10m"
}
open_source = false
certname = "example.test.com"
autosign = false
}
}
```
### Debug Output
https://gist.github.com/camara-tech/f407b8760def43f6c9da56fc45a88efe
### Crash Output
Not Applicable
### Expected Behavior
The terraform puppet provisioner successfully downloads the puppet agent installer for windows from the puppet enterprise server and successfully installs it.
### Actual Behavior
the terraform puppet provisioner failed with the following error from powershell:
```
The request was aborted: Could not create SSL/TLS secure channel.
```
### Steps to Reproduce
terraform init
terraform apply
### Additional Context
In our environment, we have completely disabled TLS protocols below TLS 1.2. Also, our puppet enterprise console uses the same Powershell that is present in the puppet provisioner for windows with the addition of the following line:
```
[System.Net.ServicePointManager]::SecurityProtocol = [Net.SecurityProtocolType]::Tls12;
```
Perhaps that needs to be added to the existing provisioner?
### References
Not Applicable
| main | puppet provisioner fails to install puppet on windows server with tls ssl error terraform version run terraform v to show the version and paste the result between the marks below if you are not running the latest version of terraform please try upgrading because your issue may have already been fixed terraform provider aws terraform configuration files paste the relevant parts of your terraform configuration between the marks below for large terraform configs please use a service like dropbox and share a link to the zip file for security you can also encrypt the files using our gpg public key hcl resource aws instance instance ami data aws ami windows server id instance type var instance type key name provision key availability zone us east tags name example vm get password data true vpc security group ids subnet id subnet user data data winrm quickconfig q winrm set winrm config maxtimeoutms winrm set winrm config service allowunencrypted true winrm set winrm config service auth basic true write host setting up firewall netsh advfirewall firewall add rule name winrm in protocol tcp dir in profile any localport remoteip any localip any action allow write host restarting winrm stop service winrm start service winrm data provisioner puppet server puppet test com connection host coalesce self private ip self public ip type winrm user administrator password rsadecrypt self password data file ssh provision key timeout open source false certname example test com autosign false debug output crash output not applicable expected behavior the terraform puppet provisioner successfully downloads the puppet agent installer for windows from the puppet enterprise server and successfully installs it actual behavior the terraform puppet provisioner failed with the following error from powershell the request was aborted could not create ssl tls secure channel steps to reproduce terraform init terraform apply additional context in our environment we have completely disabled tls protocols below tls also our puppet enterprise console uses the same powershell that is present in the puppet provisioner for windows with the addition of the following line securityprotocol perhaps that needs to be added to the existing provisioner references not applicable | 1 |
796 | 4,402,739,259 | IssuesEvent | 2016-08-11 03:09:16 | gogits/gogs | https://api.github.com/repos/gogits/gogs | closed | [FR] git-daemon-export-ok | kind/enhancement status/assigned to maintainer | Would it be possible to add an option to ensure presence of `git-daemon-export-ok` file for public repositories and its absence for private ones? The presence of this file guides the `git-daemon` used to provide `git://` protocol as to which repositories to actually expose.
AFAICT, using `git-daemon` to expose repos read-only in parallel with gogs works, but there is no distinction for private / public ones, so it exposes either none of them (cause the file is not present), or all of them (if run with `--export-all`). | True | [FR] git-daemon-export-ok - Would it be possible to add an option to ensure presence of `git-daemon-export-ok` file for public repositories and its absence for private ones? The presence of this file guides the `git-daemon` used to provide `git://` protocol as to which repositories to actually expose.
AFAICT, using `git-daemon` to expose repos read-only in parallel with gogs works, but there is no distinction for private / public ones, so it exposes either none of them (cause the file is not present), or all of them (if run with `--export-all`). | main | git daemon export ok would it be possible to add an option to ensure presence of git daemon export ok file for public repositories and its absence for private ones the presence of this file guides the git daemon used to provide git protocol as to which repositories to actually expose afaict using git daemon to expose repos read only in parallel with gogs works but there is no distinction for private public ones so it exposes either none of them cause the file is not present or all of them if run with export all | 1 |
575,759 | 17,049,203,244 | IssuesEvent | 2021-07-06 06:41:46 | gnosis/ido-ux | https://api.github.com/repos/gnosis/ido-ux | closed | Mobile: app menu overlaps app's content when scroll up and down the screen | QA QA passed bug low priority | Mobile app menu overlaps app's content when scroll up and down the screen.

**Expected**: the menu is fixed | 1.0 | Mobile: app menu overlaps app's content when scroll up and down the screen - Mobile app menu overlaps app's content when scroll up and down the screen.

**Expected**: the menu is fixed | non_main | mobile app menu overlaps app s content when scroll up and down the screen mobile app menu overlaps app s content when scroll up and down the screen expected the menu is fixed | 0 |
1,687 | 6,574,166,501 | IssuesEvent | 2017-09-11 11:47:24 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Synchronize module ignores remote_user parameter | affects_2.2 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
Module `synchronize`
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
##### OS / ENVIRONMENT
macOS Sierra
##### SUMMARY
Synchronize module doesn't use user specified in `remote_user` to connect to another system.
Works correctly in `2.0.2.0`.
##### STEPS TO REPRODUCE
You need a machine that runs Ansible.
A host, let's call it `staging_machine`, where the `ansible_user` is `staging_user`.
Another host called `my_server`, where the `ansible_user` is `deploy_user`.
Both the ansible machine and `staging_machine` have keys on `my_server` for `deploy_user`.
Then run
```
- name: Deploy java versions and scripts
hosts: my_server
any_errors_fatal: true
tasks:
- name: "Install JDK"
synchronize: src="jdk_folder" dest="deploy_area/jdkfolder/"
delegate_to: staging_machine
remote_user: "deploy_user"
```
##### EXPECTED RESULTS
Files being copied.
##### ACTUAL RESULTS
Server `staging_machine` tries to connect with it's own ansible_user `staging_user` instead of with `deploy_user`.
```
TASK [Install JDK] *************************************************************
staging_user@my_server's password:
```
This works in `2.0.2.0`. | True | Synchronize module ignores remote_user parameter - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
Module `synchronize`
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
##### OS / ENVIRONMENT
macOS Sierra
##### SUMMARY
Synchronize module doesn't use user specified in `remote_user` to connect to another system.
Works correctly in `2.0.2.0`.
##### STEPS TO REPRODUCE
You need a machine that runs Ansible.
A host, let's call it `staging_machine`, where the `ansible_user` is `staging_user`.
Another host called `my_server`, where the `ansible_user` is `deploy_user`.
Both the ansible machine and `staging_machine` have keys on `my_server` for `deploy_user`.
Then run
```
- name: Deploy java versions and scripts
hosts: my_server
any_errors_fatal: true
tasks:
- name: "Install JDK"
synchronize: src="jdk_folder" dest="deploy_area/jdkfolder/"
delegate_to: staging_machine
remote_user: "deploy_user"
```
##### EXPECTED RESULTS
Files being copied.
##### ACTUAL RESULTS
Server `staging_machine` tries to connect with it's own ansible_user `staging_user` instead of with `deploy_user`.
```
TASK [Install JDK] *************************************************************
staging_user@my_server's password:
```
This works in `2.0.2.0`. | main | synchronize module ignores remote user parameter issue type bug report component name module synchronize ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration os environment macos sierra summary synchronize module doesn t use user specified in remote user to connect to another system works correctly in steps to reproduce you need a machine that runs ansible a host let s call it staging machine where the ansible user is staging user another host called my server where the ansible user is deploy user both the ansible machine and staging machine have keys on my server for deploy user then run name deploy java versions and scripts hosts my server any errors fatal true tasks name install jdk synchronize src jdk folder dest deploy area jdkfolder delegate to staging machine remote user deploy user expected results files being copied actual results server staging machine tries to connect with it s own ansible user staging user instead of with deploy user task staging user my server s password this works in | 1 |
3,990 | 18,445,186,131 | IssuesEvent | 2021-10-15 00:19:59 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | "Please install sqlite3 package manually" but I have installed it. | stage/needs-investigation area/local/invoke maintainer/need-response | <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- Briefly describe the bug you are facing.-->
"Please install sqlite3 package manually" but I have installed it.
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
https://github.com/tylerlong/sam-app-demo
```
sam build && sam local invoke NewFunction -e events/event.json
```
### Observed result:
<!-- Please provide command output with `--debug` flag set. -->
```
sam build && sam local invoke NewFunction -e events/event.json --debug
Building codeuri: /Users/tyler.liu/src/ts/sam-app-demo/hello-world runtime: nodejs14.x metadata: {} functions: ['HelloWorldFunction', 'NewFunction']
Running NodejsNpmBuilder:NpmPack
Running NodejsNpmBuilder:CopyNpmrc
Running NodejsNpmBuilder:CopySource
Running NodejsNpmBuilder:NpmInstall
Running NodejsNpmBuilder:CleanUpNpmrc
Build Succeeded
Built Artifacts : .aws-sam/build
Built Template : .aws-sam/build/template.yaml
Commands you can use next
=========================
[*] Invoke Function: sam local invoke
[*] Deploy: sam deploy --guided
2021-07-01 12:08:51,366 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2021-07-01 12:08:51,366 | Using config file: samconfig.toml, config environment: default
2021-07-01 12:08:51,366 | Expand command line arguments to:
2021-07-01 12:08:51,366 | --template_file=/Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/template.yaml --event=events/event.json --function_logical_id=NewFunction --no_event --layer_cache_basedir=/Users/tyler.liu/.aws-sam/layers-pkg --container_host=localhost --container_host_interface=127.0.0.1
2021-07-01 12:08:51,367 | local invoke command is called
2021-07-01 12:08:51,375 | No Parameters detected in the template
2021-07-01 12:08:51,416 | 3 stacks found in the template
2021-07-01 12:08:51,417 | No Parameters detected in the template
2021-07-01 12:08:51,440 | 3 resources found in the stack
2021-07-01 12:08:51,440 | No Parameters detected in the template
2021-07-01 12:08:51,463 | Found Serverless function with name='HelloWorldFunction' and CodeUri='HelloWorldFunction'
2021-07-01 12:08:51,463 | --base-dir is not presented, adjusting uri HelloWorldFunction relative to /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/template.yaml
2021-07-01 12:08:51,464 | Found Serverless function with name='NewFunction' and CodeUri='NewFunction'
2021-07-01 12:08:51,464 | --base-dir is not presented, adjusting uri NewFunction relative to /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/template.yaml
2021-07-01 12:08:51,474 | Found one Lambda function with name 'NewFunction'
2021-07-01 12:08:51,474 | Invoking app.newHandler (nodejs14.x)
2021-07-01 12:08:51,474 | Environment variables overrides data is standard format
2021-07-01 12:08:51,474 | Loading AWS credentials from session with profile 'None'
2021-07-01 12:08:51,486 | Resolving code path. Cwd=/Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build, CodeUri=/Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/NewFunction
2021-07-01 12:08:51,486 | Resolved absolute path to code is /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/NewFunction
2021-07-01 12:08:51,486 | Code /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/NewFunction is not a zip/jar file
2021-07-01 12:08:51,503 | Skip pulling image and use local one: amazon/aws-sam-cli-emulation-image-nodejs14.x:rapid-1.24.1.
2021-07-01 12:08:51,503 | Mounting /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/NewFunction as /var/task:ro,delegated inside runtime container
2021-07-01 12:08:52,003 | Starting a timer for 3 seconds for function 'NewFunction'
START RequestId: 532ba6d5-3ed1-4674-8b18-e9d14a041740 Version: $LATEST
2021-07-01T19:08:53.741Z undefined ERROR Uncaught Exception {"errorType":"Error","errorMessage":"Please install sqlite3 package manually","stack":["Error: Please install sqlite3 package manually"," at ConnectionManager._loadDialectModule (/var/task/node_modules/sequelize/lib/dialects/abstract/connection-manager.js:88:15)"," at new ConnectionManager (/var/task/node_modules/sequelize/lib/dialects/sqlite/connection-manager.js:22:21)"," at new SqliteDialect (/var/task/node_modules/sequelize/lib/dialects/sqlite/index.js:14:30)"," at new Sequelize (/var/task/node_modules/sequelize/lib/sequelize.js:334:20)"," at Object.<anonymous> (/var/task/node_modules/ringcentral-chatbot/dist/models/sequelize.js:12:19)"," at Module._compile (internal/modules/cjs/loader.js:1085:14)"," at Object.Module._extensions..js (internal/modules/cjs/loader.js:1114:10)"," at Module.load (internal/modules/cjs/loader.js:950:32)"," at Function.Module._load (internal/modules/cjs/loader.js:790:14)"," at Module.require (internal/modules/cjs/loader.js:974:19)"," at require (internal/modules/cjs/helpers.js:92:18)"," at Object.<anonymous> (/var/task/node_modules/ringcentral-chatbot/dist/models/Bot.js:16:42)"," at Module._compile (internal/modules/cjs/loader.js:1085:14)"," at Object.Module._extensions..js (internal/modules/cjs/loader.js:1114:10)"," at Module.load (internal/modules/cjs/loader.js:950:32)"," at Function.Module._load (internal/modules/cjs/loader.js:790:14)"," at Module.require (internal/modules/cjs/loader.js:974:19)"," at require (internal/modules/cjs/helpers.js:92:18)"," at Object.<anonymous> (/var/task/node_modules/ringcentral-chatbot/dist/models/index.js:19:12)"," at Module._compile (internal/modules/cjs/loader.js:1085:14)"," at Object.Module._extensions..js (internal/modules/cjs/loader.js:1114:10)"," at Module.load (internal/modules/cjs/loader.js:950:32)"]}
time="2021-07-01T19:08:53.747" level=panic msg="ReplyStream not available"
2021/07/01 19:08:53 http: panic serving 127.0.0.1:50584: &{0xc00013e000 map[] 2021-07-01 19:08:53.747549909 +0000 UTC m=+1.748889415 panic <nil> ReplyStream not available <nil> <nil> }
goroutine 41 [running]:
net/http.(*conn).serve.func1(0xc0000be1e0)
/usr/local/go/src/net/http/server.go:1800 +0x139
panic(0x866640, 0xc0000ee0e0)
/usr/local/go/src/runtime/panic.go:975 +0x3e3
github.com/sirupsen/logrus.Entry.log(0xc00013e000, 0xc0000a2330, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/entry.go:259 +0x335
github.com/sirupsen/logrus.(*Entry).Log(0xc0000ee000, 0xc000000000, 0xc0000f7588, 0x1, 0x1)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/entry.go:287 +0xeb
github.com/sirupsen/logrus.(*Logger).Log(0xc00013e000, 0xc000000000, 0xc0000f7588, 0x1, 0x1)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/logger.go:193 +0x7d
github.com/sirupsen/logrus.(*Logger).Panic(...)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/logger.go:234
github.com/sirupsen/logrus.Panic(...)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/exported.go:129
go.amzn.com/lambda/rapi/rendering.RenderInteropError(0x9097c0, 0xc0000dc1c0, 0xc0000d4500, 0x902b60, 0xc0000961a0)
/LambdaRuntimeLocal/lambda/rapi/rendering/rendering.go:292 +0x9a
go.amzn.com/lambda/rapi/handler.(*initErrorHandler).ServeHTTP(0xc00018c0a0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/LambdaRuntimeLocal/lambda/rapi/handler/initerror.go:52 +0x519
net/http.HandlerFunc.ServeHTTP(0xc000182380, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/usr/local/go/src/net/http/server.go:2041 +0x44
github.com/go-chi/chi.(*Mux).routeHTTP(0xc000180120, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:431 +0x278
net/http.HandlerFunc.ServeHTTP(0xc00018c030, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/usr/local/go/src/net/http/server.go:2041 +0x44
go.amzn.com/lambda/rapi/middleware.RuntimeReleaseMiddleware.func1.1(0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/LambdaRuntimeLocal/lambda/rapi/middleware/middleware.go:100 +0xea
net/http.HandlerFunc.ServeHTTP(0xc000182180, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/usr/local/go/src/net/http/server.go:2041 +0x44
go.amzn.com/lambda/rapi/middleware.AccessLogMiddleware.func1.1(0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/LambdaRuntimeLocal/lambda/rapi/middleware/middleware.go:77 +0x170
net/http.HandlerFunc.ServeHTTP(0xc0001821a0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/usr/local/go/src/net/http/server.go:2041 +0x44
go.amzn.com/lambda/rapi/middleware.AppCtxMiddleware.func1.1(0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/LambdaRuntimeLocal/lambda/rapi/middleware/middleware.go:66 +0x77
net/http.HandlerFunc.ServeHTTP(0xc00018a2d0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/usr/local/go/src/net/http/server.go:2041 +0x44
github.com/go-chi/chi.(*Mux).ServeHTTP(0xc000180120, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:70 +0x513
github.com/go-chi/chi.(*Mux).Mount.func1(0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:298 +0x118
net/http.HandlerFunc.ServeHTTP(0xc0001823e0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/usr/local/go/src/net/http/server.go:2041 +0x44
github.com/go-chi/chi.(*Mux).routeHTTP(0xc0001800c0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:431 +0x278
net/http.HandlerFunc.ServeHTTP(0xc00018c0c0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/usr/local/go/src/net/http/server.go:2041 +0x44
github.com/go-chi/chi.(*Mux).ServeHTTP(0xc0001800c0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4300)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:86 +0x2b2
net/http.serverHandler.ServeHTTP(0xc00019a000, 0x9097c0, 0xc0000dc1c0, 0xc0000d4300)
/usr/local/go/src/net/http/server.go:2836 +0xa3
net/http.(*conn).serve(0xc0000be1e0, 0x90a800, 0xc0000c22c0)
/usr/local/go/src/net/http/server.go:1924 +0x86c
created by net/http.(*Server).Serve
/usr/local/go/src/net/http/server.go:2962 +0x35c
2021-07-01T19:08:53.749Z undefined ERROR Uncaught Exception {"errorType":"Error","errorMessage":"socket hang up","code":"ECONNRESET","stack":["Error: socket hang up"," at connResetException (internal/errors.js:628:14)"," at Socket.socketOnEnd (_http_client.js:499:23)"," at Socket.emit (events.js:387:35)"," at endReadableNT (internal/streams/readable.js:1317:12)"," at processTicksAndRejections (internal/process/task_queues.js:82:21)"]}
time="2021-07-01T19:08:53.759" level=error msg="Init failed" InvokeID= error="Runtime exited with error: exit status 129"
time="2021-07-01T19:08:53.76" level=error msg="INIT DONE failed: Runtime.ExitError"
2021-07-01 12:08:53,956 | Cleaning all decompressed code dirs
2021-07-01 12:08:53,958 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': 'a21fc39b-d5d8-4975-8933-08bbbc776389', 'installationId': 'c6e31722-acda-4282-a011-7dcd88ad68ea', 'sessionId': '5f450b72-fb5e-4c21-8d48-6a5be40d5f6b', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.11', 'samcliVersion': '1.24.1', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam local invoke', 'duration': 2591, 'exitReason': 'success', 'exitCode': 0}}]}
2021-07-01 12:08:55,143 | Telemetry response: 200
2021-07-01 12:08:55,144 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2021-07-01 12:08:55,144 | Sending Telemetry: {'metrics': [{'runtimeMetric': {'requestId': '5d22ca34-18ce-4443-b9f8-8df7337434e2', 'installationId': 'c6e31722-acda-4282-a011-7dcd88ad68ea', 'sessionId': '5f450b72-fb5e-4c21-8d48-6a5be40d5f6b', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.11', 'samcliVersion': '1.24.1', 'runtimes': ['nodejs14.x']}}]}
2021-07-01 12:08:55,375 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
```
### Expected result:
<!-- Describe what you expected. -->
Not sure but at least is should NOT complain "Please install sqlite3 package manually".
I have installed it already: https://github.com/tylerlong/sam-app-demo/blob/main/hello-world/package.json#L14
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: macOS 10.14.6
2. `sam --version`: SAM CLI, version 1.24.1
3. AWS region: local test only
`Add --debug flag to command you are running`
| True | "Please install sqlite3 package manually" but I have installed it. - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
<!-- Briefly describe the bug you are facing.-->
"Please install sqlite3 package manually" but I have installed it.
### Steps to reproduce:
<!-- Provide detailed steps to replicate the bug, including steps from third party tools (CDK, etc.) -->
https://github.com/tylerlong/sam-app-demo
```
sam build && sam local invoke NewFunction -e events/event.json
```
### Observed result:
<!-- Please provide command output with `--debug` flag set. -->
```
sam build && sam local invoke NewFunction -e events/event.json --debug
Building codeuri: /Users/tyler.liu/src/ts/sam-app-demo/hello-world runtime: nodejs14.x metadata: {} functions: ['HelloWorldFunction', 'NewFunction']
Running NodejsNpmBuilder:NpmPack
Running NodejsNpmBuilder:CopyNpmrc
Running NodejsNpmBuilder:CopySource
Running NodejsNpmBuilder:NpmInstall
Running NodejsNpmBuilder:CleanUpNpmrc
Build Succeeded
Built Artifacts : .aws-sam/build
Built Template : .aws-sam/build/template.yaml
Commands you can use next
=========================
[*] Invoke Function: sam local invoke
[*] Deploy: sam deploy --guided
2021-07-01 12:08:51,366 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2021-07-01 12:08:51,366 | Using config file: samconfig.toml, config environment: default
2021-07-01 12:08:51,366 | Expand command line arguments to:
2021-07-01 12:08:51,366 | --template_file=/Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/template.yaml --event=events/event.json --function_logical_id=NewFunction --no_event --layer_cache_basedir=/Users/tyler.liu/.aws-sam/layers-pkg --container_host=localhost --container_host_interface=127.0.0.1
2021-07-01 12:08:51,367 | local invoke command is called
2021-07-01 12:08:51,375 | No Parameters detected in the template
2021-07-01 12:08:51,416 | 3 stacks found in the template
2021-07-01 12:08:51,417 | No Parameters detected in the template
2021-07-01 12:08:51,440 | 3 resources found in the stack
2021-07-01 12:08:51,440 | No Parameters detected in the template
2021-07-01 12:08:51,463 | Found Serverless function with name='HelloWorldFunction' and CodeUri='HelloWorldFunction'
2021-07-01 12:08:51,463 | --base-dir is not presented, adjusting uri HelloWorldFunction relative to /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/template.yaml
2021-07-01 12:08:51,464 | Found Serverless function with name='NewFunction' and CodeUri='NewFunction'
2021-07-01 12:08:51,464 | --base-dir is not presented, adjusting uri NewFunction relative to /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/template.yaml
2021-07-01 12:08:51,474 | Found one Lambda function with name 'NewFunction'
2021-07-01 12:08:51,474 | Invoking app.newHandler (nodejs14.x)
2021-07-01 12:08:51,474 | Environment variables overrides data is standard format
2021-07-01 12:08:51,474 | Loading AWS credentials from session with profile 'None'
2021-07-01 12:08:51,486 | Resolving code path. Cwd=/Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build, CodeUri=/Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/NewFunction
2021-07-01 12:08:51,486 | Resolved absolute path to code is /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/NewFunction
2021-07-01 12:08:51,486 | Code /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/NewFunction is not a zip/jar file
2021-07-01 12:08:51,503 | Skip pulling image and use local one: amazon/aws-sam-cli-emulation-image-nodejs14.x:rapid-1.24.1.
2021-07-01 12:08:51,503 | Mounting /Users/tyler.liu/src/ts/sam-app-demo/.aws-sam/build/NewFunction as /var/task:ro,delegated inside runtime container
2021-07-01 12:08:52,003 | Starting a timer for 3 seconds for function 'NewFunction'
START RequestId: 532ba6d5-3ed1-4674-8b18-e9d14a041740 Version: $LATEST
2021-07-01T19:08:53.741Z undefined ERROR Uncaught Exception {"errorType":"Error","errorMessage":"Please install sqlite3 package manually","stack":["Error: Please install sqlite3 package manually"," at ConnectionManager._loadDialectModule (/var/task/node_modules/sequelize/lib/dialects/abstract/connection-manager.js:88:15)"," at new ConnectionManager (/var/task/node_modules/sequelize/lib/dialects/sqlite/connection-manager.js:22:21)"," at new SqliteDialect (/var/task/node_modules/sequelize/lib/dialects/sqlite/index.js:14:30)"," at new Sequelize (/var/task/node_modules/sequelize/lib/sequelize.js:334:20)"," at Object.<anonymous> (/var/task/node_modules/ringcentral-chatbot/dist/models/sequelize.js:12:19)"," at Module._compile (internal/modules/cjs/loader.js:1085:14)"," at Object.Module._extensions..js (internal/modules/cjs/loader.js:1114:10)"," at Module.load (internal/modules/cjs/loader.js:950:32)"," at Function.Module._load (internal/modules/cjs/loader.js:790:14)"," at Module.require (internal/modules/cjs/loader.js:974:19)"," at require (internal/modules/cjs/helpers.js:92:18)"," at Object.<anonymous> (/var/task/node_modules/ringcentral-chatbot/dist/models/Bot.js:16:42)"," at Module._compile (internal/modules/cjs/loader.js:1085:14)"," at Object.Module._extensions..js (internal/modules/cjs/loader.js:1114:10)"," at Module.load (internal/modules/cjs/loader.js:950:32)"," at Function.Module._load (internal/modules/cjs/loader.js:790:14)"," at Module.require (internal/modules/cjs/loader.js:974:19)"," at require (internal/modules/cjs/helpers.js:92:18)"," at Object.<anonymous> (/var/task/node_modules/ringcentral-chatbot/dist/models/index.js:19:12)"," at Module._compile (internal/modules/cjs/loader.js:1085:14)"," at Object.Module._extensions..js (internal/modules/cjs/loader.js:1114:10)"," at Module.load (internal/modules/cjs/loader.js:950:32)"]}
time="2021-07-01T19:08:53.747" level=panic msg="ReplyStream not available"
2021/07/01 19:08:53 http: panic serving 127.0.0.1:50584: &{0xc00013e000 map[] 2021-07-01 19:08:53.747549909 +0000 UTC m=+1.748889415 panic <nil> ReplyStream not available <nil> <nil> }
goroutine 41 [running]:
net/http.(*conn).serve.func1(0xc0000be1e0)
/usr/local/go/src/net/http/server.go:1800 +0x139
panic(0x866640, 0xc0000ee0e0)
/usr/local/go/src/runtime/panic.go:975 +0x3e3
github.com/sirupsen/logrus.Entry.log(0xc00013e000, 0xc0000a2330, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, 0x0, ...)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/entry.go:259 +0x335
github.com/sirupsen/logrus.(*Entry).Log(0xc0000ee000, 0xc000000000, 0xc0000f7588, 0x1, 0x1)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/entry.go:287 +0xeb
github.com/sirupsen/logrus.(*Logger).Log(0xc00013e000, 0xc000000000, 0xc0000f7588, 0x1, 0x1)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/logger.go:193 +0x7d
github.com/sirupsen/logrus.(*Logger).Panic(...)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/logger.go:234
github.com/sirupsen/logrus.Panic(...)
/go/pkg/mod/github.com/sirupsen/logrus@v1.6.0/exported.go:129
go.amzn.com/lambda/rapi/rendering.RenderInteropError(0x9097c0, 0xc0000dc1c0, 0xc0000d4500, 0x902b60, 0xc0000961a0)
/LambdaRuntimeLocal/lambda/rapi/rendering/rendering.go:292 +0x9a
go.amzn.com/lambda/rapi/handler.(*initErrorHandler).ServeHTTP(0xc00018c0a0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/LambdaRuntimeLocal/lambda/rapi/handler/initerror.go:52 +0x519
net/http.HandlerFunc.ServeHTTP(0xc000182380, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/usr/local/go/src/net/http/server.go:2041 +0x44
github.com/go-chi/chi.(*Mux).routeHTTP(0xc000180120, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:431 +0x278
net/http.HandlerFunc.ServeHTTP(0xc00018c030, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/usr/local/go/src/net/http/server.go:2041 +0x44
go.amzn.com/lambda/rapi/middleware.RuntimeReleaseMiddleware.func1.1(0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/LambdaRuntimeLocal/lambda/rapi/middleware/middleware.go:100 +0xea
net/http.HandlerFunc.ServeHTTP(0xc000182180, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/usr/local/go/src/net/http/server.go:2041 +0x44
go.amzn.com/lambda/rapi/middleware.AccessLogMiddleware.func1.1(0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/LambdaRuntimeLocal/lambda/rapi/middleware/middleware.go:77 +0x170
net/http.HandlerFunc.ServeHTTP(0xc0001821a0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4500)
/usr/local/go/src/net/http/server.go:2041 +0x44
go.amzn.com/lambda/rapi/middleware.AppCtxMiddleware.func1.1(0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/LambdaRuntimeLocal/lambda/rapi/middleware/middleware.go:66 +0x77
net/http.HandlerFunc.ServeHTTP(0xc00018a2d0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/usr/local/go/src/net/http/server.go:2041 +0x44
github.com/go-chi/chi.(*Mux).ServeHTTP(0xc000180120, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:70 +0x513
github.com/go-chi/chi.(*Mux).Mount.func1(0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:298 +0x118
net/http.HandlerFunc.ServeHTTP(0xc0001823e0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/usr/local/go/src/net/http/server.go:2041 +0x44
github.com/go-chi/chi.(*Mux).routeHTTP(0xc0001800c0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:431 +0x278
net/http.HandlerFunc.ServeHTTP(0xc00018c0c0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4400)
/usr/local/go/src/net/http/server.go:2041 +0x44
github.com/go-chi/chi.(*Mux).ServeHTTP(0xc0001800c0, 0x9097c0, 0xc0000dc1c0, 0xc0000d4300)
/go/pkg/mod/github.com/go-chi/chi@v4.1.2+incompatible/mux.go:86 +0x2b2
net/http.serverHandler.ServeHTTP(0xc00019a000, 0x9097c0, 0xc0000dc1c0, 0xc0000d4300)
/usr/local/go/src/net/http/server.go:2836 +0xa3
net/http.(*conn).serve(0xc0000be1e0, 0x90a800, 0xc0000c22c0)
/usr/local/go/src/net/http/server.go:1924 +0x86c
created by net/http.(*Server).Serve
/usr/local/go/src/net/http/server.go:2962 +0x35c
2021-07-01T19:08:53.749Z undefined ERROR Uncaught Exception {"errorType":"Error","errorMessage":"socket hang up","code":"ECONNRESET","stack":["Error: socket hang up"," at connResetException (internal/errors.js:628:14)"," at Socket.socketOnEnd (_http_client.js:499:23)"," at Socket.emit (events.js:387:35)"," at endReadableNT (internal/streams/readable.js:1317:12)"," at processTicksAndRejections (internal/process/task_queues.js:82:21)"]}
time="2021-07-01T19:08:53.759" level=error msg="Init failed" InvokeID= error="Runtime exited with error: exit status 129"
time="2021-07-01T19:08:53.76" level=error msg="INIT DONE failed: Runtime.ExitError"
2021-07-01 12:08:53,956 | Cleaning all decompressed code dirs
2021-07-01 12:08:53,958 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': 'a21fc39b-d5d8-4975-8933-08bbbc776389', 'installationId': 'c6e31722-acda-4282-a011-7dcd88ad68ea', 'sessionId': '5f450b72-fb5e-4c21-8d48-6a5be40d5f6b', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.11', 'samcliVersion': '1.24.1', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam local invoke', 'duration': 2591, 'exitReason': 'success', 'exitCode': 0}}]}
2021-07-01 12:08:55,143 | Telemetry response: 200
2021-07-01 12:08:55,144 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2021-07-01 12:08:55,144 | Sending Telemetry: {'metrics': [{'runtimeMetric': {'requestId': '5d22ca34-18ce-4443-b9f8-8df7337434e2', 'installationId': 'c6e31722-acda-4282-a011-7dcd88ad68ea', 'sessionId': '5f450b72-fb5e-4c21-8d48-6a5be40d5f6b', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.8.11', 'samcliVersion': '1.24.1', 'runtimes': ['nodejs14.x']}}]}
2021-07-01 12:08:55,375 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
```
### Expected result:
<!-- Describe what you expected. -->
Not sure but at least is should NOT complain "Please install sqlite3 package manually".
I have installed it already: https://github.com/tylerlong/sam-app-demo/blob/main/hello-world/package.json#L14
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: macOS 10.14.6
2. `sam --version`: SAM CLI, version 1.24.1
3. AWS region: local test only
`Add --debug flag to command you are running`
| main | please install package manually but i have installed it make sure we don t have an existing issue that reports the bug you are seeing both open and closed if you do find an existing issue re open or add a comment to that issue instead of creating a new one description please install package manually but i have installed it steps to reproduce sam build sam local invoke newfunction e events event json observed result sam build sam local invoke newfunction e events event json debug building codeuri users tyler liu src ts sam app demo hello world runtime x metadata functions running nodejsnpmbuilder npmpack running nodejsnpmbuilder copynpmrc running nodejsnpmbuilder copysource running nodejsnpmbuilder npminstall running nodejsnpmbuilder cleanupnpmrc build succeeded built artifacts aws sam build built template aws sam build template yaml commands you can use next invoke function sam local invoke deploy sam deploy guided telemetry endpoint configured to be using config file samconfig toml config environment default expand command line arguments to template file users tyler liu src ts sam app demo aws sam build template yaml event events event json function logical id newfunction no event layer cache basedir users tyler liu aws sam layers pkg container host localhost container host interface local invoke command is called no parameters detected in the template stacks found in the template no parameters detected in the template resources found in the stack no parameters detected in the template found serverless function with name helloworldfunction and codeuri helloworldfunction base dir is not presented adjusting uri helloworldfunction relative to users tyler liu src ts sam app demo aws sam build template yaml found serverless function with name newfunction and codeuri newfunction base dir is not presented adjusting uri newfunction relative to users tyler liu src ts sam app demo aws sam build template yaml found one lambda function with name newfunction invoking app newhandler x environment variables overrides data is standard format loading aws credentials from session with profile none resolving code path cwd users tyler liu src ts sam app demo aws sam build codeuri users tyler liu src ts sam app demo aws sam build newfunction resolved absolute path to code is users tyler liu src ts sam app demo aws sam build newfunction code users tyler liu src ts sam app demo aws sam build newfunction is not a zip jar file skip pulling image and use local one amazon aws sam cli emulation image x rapid mounting users tyler liu src ts sam app demo aws sam build newfunction as var task ro delegated inside runtime container starting a timer for seconds for function newfunction start requestid version latest undefined error uncaught exception errortype error errormessage please install package manually stack time level panic msg replystream not available http panic serving map utc m panic replystream not available goroutine net http conn serve usr local go src net http server go panic usr local go src runtime panic go github com sirupsen logrus entry log go pkg mod github com sirupsen logrus entry go github com sirupsen logrus entry log go pkg mod github com sirupsen logrus entry go github com sirupsen logrus logger log go pkg mod github com sirupsen logrus logger go github com sirupsen logrus logger panic go pkg mod github com sirupsen logrus logger go github com sirupsen logrus panic go pkg mod github com sirupsen logrus exported go go amzn com lambda rapi rendering renderinteroperror lambdaruntimelocal lambda rapi rendering rendering go go amzn com lambda rapi handler initerrorhandler servehttp lambdaruntimelocal lambda rapi handler initerror go net http handlerfunc servehttp usr local go src net http server go github com go chi chi mux routehttp go pkg mod github com go chi chi incompatible mux go net http handlerfunc servehttp usr local go src net http server go go amzn com lambda rapi middleware runtimereleasemiddleware lambdaruntimelocal lambda rapi middleware middleware go net http handlerfunc servehttp usr local go src net http server go go amzn com lambda rapi middleware accesslogmiddleware lambdaruntimelocal lambda rapi middleware middleware go net http handlerfunc servehttp usr local go src net http server go go amzn com lambda rapi middleware appctxmiddleware lambdaruntimelocal lambda rapi middleware middleware go net http handlerfunc servehttp usr local go src net http server go github com go chi chi mux servehttp go pkg mod github com go chi chi incompatible mux go github com go chi chi mux mount go pkg mod github com go chi chi incompatible mux go net http handlerfunc servehttp usr local go src net http server go github com go chi chi mux routehttp go pkg mod github com go chi chi incompatible mux go net http handlerfunc servehttp usr local go src net http server go github com go chi chi mux servehttp go pkg mod github com go chi chi incompatible mux go net http serverhandler servehttp usr local go src net http server go net http conn serve usr local go src net http server go created by net http server serve usr local go src net http server go undefined error uncaught exception errortype error errormessage socket hang up code econnreset stack time level error msg init failed invokeid error runtime exited with error exit status time level error msg init done failed runtime exiterror cleaning all decompressed code dirs sending telemetry metrics telemetry response telemetry endpoint configured to be sending telemetry metrics httpsconnectionpool host aws serverless tools telemetry us west amazonaws com port read timed out read timeout expected result not sure but at least is should not complain please install package manually i have installed it already additional environment details ex windows mac amazon linux etc os macos sam version sam cli version aws region local test only add debug flag to command you are running | 1 |
1,056 | 4,865,275,973 | IssuesEvent | 2016-11-14 20:18:48 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | args alias unsupported by service module | affects_2.2 bug_report in progress waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
service module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file =
configured module search path = Default w/o overrides
```
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ubuntu 16.04
##### SUMMARY
<!--- Explain the problem briefly -->
The 'args' alias for the 'arguments' parameter is not recognized as valid by the service module.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- hosts: localhost
connection: local
gather_facts: no
tasks:
- service:
name: ssh
state: started
args: ""
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The playbook to complete successfully.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
fatal: [localhost]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_args": {
"args": "",
"name": "ssh",
"state": "started"
}
},
"msg": "unsupported parameter for module: args"
}
```
| True | args alias unsupported by service module - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
service module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file =
configured module search path = Default w/o overrides
```
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ubuntu 16.04
##### SUMMARY
<!--- Explain the problem briefly -->
The 'args' alias for the 'arguments' parameter is not recognized as valid by the service module.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- hosts: localhost
connection: local
gather_facts: no
tasks:
- service:
name: ssh
state: started
args: ""
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The playbook to complete successfully.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
fatal: [localhost]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_args": {
"args": "",
"name": "ssh",
"state": "started"
}
},
"msg": "unsupported parameter for module: args"
}
```
| main | args alias unsupported by service module issue type bug report component name service module ansible version ansible config file configured module search path default w o overrides os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific ubuntu summary the args alias for the arguments parameter is not recognized as valid by the service module steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used hosts localhost connection local gather facts no tasks service name ssh state started args expected results the playbook to complete successfully actual results fatal failed changed false failed true invocation module args args name ssh state started msg unsupported parameter for module args | 1 |
20,240 | 10,479,699,006 | IssuesEvent | 2019-09-24 05:18:35 | scriptex/webpack-mpa | https://api.github.com/repos/scriptex/webpack-mpa | closed | CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js | security vulnerability | ## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/webpack-mpa/node_modules/vm-browserify/example/run/index.html</p>
<p>Path to vulnerable library: /webpack-mpa/node_modules/vm-browserify/example/run/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/scriptex/webpack-mpa/commit/8b39c112f82dbc91fff837dd14751007b9e4839d">8b39c112f82dbc91fff837dd14751007b9e4839d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2012-6708 (Medium) detected in jquery-1.7.1.min.js - ## CVE-2012-6708 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.7.1.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.7.1/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/webpack-mpa/node_modules/vm-browserify/example/run/index.html</p>
<p>Path to vulnerable library: /webpack-mpa/node_modules/vm-browserify/example/run/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.7.1.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/scriptex/webpack-mpa/commit/8b39c112f82dbc91fff837dd14751007b9e4839d">8b39c112f82dbc91fff837dd14751007b9e4839d</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 1.9.0 is vulnerable to Cross-site Scripting (XSS) attacks. The jQuery(strInput) function does not differentiate selectors from HTML in a reliable fashion. In vulnerable versions, jQuery determined whether the input was HTML by looking for the '<' character anywhere in the string, giving attackers more flexibility when attempting to construct a malicious payload. In fixed versions, jQuery only deems the input to be HTML if it explicitly starts with the '<' character, limiting exploitability only to attackers who can control the beginning of a string, which is far less common.
<p>Publish Date: 2018-01-18
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2012-6708>CVE-2012-6708</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2012-6708">https://nvd.nist.gov/vuln/detail/CVE-2012-6708</a></p>
<p>Release Date: 2018-01-18</p>
<p>Fix Resolution: jQuery - v1.9.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm webpack mpa node modules vm browserify example run index html path to vulnerable library webpack mpa node modules vm browserify example run index html dependency hierarchy x jquery min js vulnerable library found in head commit a href vulnerability details jquery before is vulnerable to cross site scripting xss attacks the jquery strinput function does not differentiate selectors from html in a reliable fashion in vulnerable versions jquery determined whether the input was html by looking for the character anywhere in the string giving attackers more flexibility when attempting to construct a malicious payload in fixed versions jquery only deems the input to be html if it explicitly starts with the character limiting exploitability only to attackers who can control the beginning of a string which is far less common publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery step up your open source security game with whitesource | 0 |
3,842 | 16,760,806,841 | IssuesEvent | 2021-06-13 18:48:05 | code-gambit/VT-WebClient | https://api.github.com/repos/code-gambit/VT-WebClient | closed | [BUG] password field doesn't have visibility toggle button | Type: Maintainance UI | **Describe the bug**
While signing in the password has no password visibility toggle switch, hence the user has no option to see the entered password.
Area Where Bug Occurred
- [x] UI/UX
- [ ] Business Logic
**Screenshot**

**How to reproduce the error/bug**
1. Visit `base_url/authenticate`.
2. Observe the password filed, it has no option to make the password visible.
**Additional context**
Password field must have an eye icon, clicking which makes password visibility toggle.
**Before submitting an issue check the below**
- [x] I have synced all my node versions as mentioned in the project"
- [x] I am using the same version of npm as is the project
- [x] My current branch is in sync with the development branch
| True | [BUG] password field doesn't have visibility toggle button - **Describe the bug**
While signing in the password has no password visibility toggle switch, hence the user has no option to see the entered password.
Area Where Bug Occurred
- [x] UI/UX
- [ ] Business Logic
**Screenshot**

**How to reproduce the error/bug**
1. Visit `base_url/authenticate`.
2. Observe the password filed, it has no option to make the password visible.
**Additional context**
Password field must have an eye icon, clicking which makes password visibility toggle.
**Before submitting an issue check the below**
- [x] I have synced all my node versions as mentioned in the project"
- [x] I am using the same version of npm as is the project
- [x] My current branch is in sync with the development branch
| main | password field doesn t have visibility toggle button describe the bug while signing in the password has no password visibility toggle switch hence the user has no option to see the entered password area where bug occurred ui ux business logic screenshot how to reproduce the error bug visit base url authenticate observe the password filed it has no option to make the password visible additional context password field must have an eye icon clicking which makes password visibility toggle before submitting an issue check the below i have synced all my node versions as mentioned in the project i am using the same version of npm as is the project my current branch is in sync with the development branch | 1 |
4,132 | 19,601,822,405 | IssuesEvent | 2022-01-06 02:49:00 | DynamoRIO/dynamorio | https://api.github.com/repos/DynamoRIO/dynamorio | closed | update to clang-format-9 | Maintainability | We are currently using clang-format-6 but it is getting difficult to maintain that on new systems. This is a proposal to update to clang-format-9. Some files do have drastic differences with version 9, such as optionsx.h, so we'd do a single reformat across the code base. | True | update to clang-format-9 - We are currently using clang-format-6 but it is getting difficult to maintain that on new systems. This is a proposal to update to clang-format-9. Some files do have drastic differences with version 9, such as optionsx.h, so we'd do a single reformat across the code base. | main | update to clang format we are currently using clang format but it is getting difficult to maintain that on new systems this is a proposal to update to clang format some files do have drastic differences with version such as optionsx h so we d do a single reformat across the code base | 1 |
284,180 | 24,581,921,736 | IssuesEvent | 2022-10-13 16:15:29 | magma/magma | https://api.github.com/repos/magma/magma | opened | Add make magma.deb workflow to CI dashboard | component: ci LTE-Integration-Test | Add [this workflow](https://github.com/magma/magma/blob/master/.github/workflows/lte-integ-test-bazel-magma-deb.yml) to the [CI dashboard](https://magma-ci.web.app/). As discussed in the DevOps meeting, it would be good to have this in the dashboard so that we monitor our production debian artifacts more closely. | 1.0 | Add make magma.deb workflow to CI dashboard - Add [this workflow](https://github.com/magma/magma/blob/master/.github/workflows/lte-integ-test-bazel-magma-deb.yml) to the [CI dashboard](https://magma-ci.web.app/). As discussed in the DevOps meeting, it would be good to have this in the dashboard so that we monitor our production debian artifacts more closely. | non_main | add make magma deb workflow to ci dashboard add to the as discussed in the devops meeting it would be good to have this in the dashboard so that we monitor our production debian artifacts more closely | 0 |
1,458 | 6,306,739,670 | IssuesEvent | 2017-07-21 21:55:18 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Verify that cron_file argument is valid. | affects_2.0 feature_idea waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
cron
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.0.0.2
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ubuntu
##### SUMMARY
<!--- Explain the problem briefly -->
Cron will only read cron-files without . (dots) in them, particularly
`[The filenames] must consist solely of upper- and lower-case letters, digits, underscores, and hyphens. This means that they cannot contain any dots.`
It would be nice if
1) The cron-module gave a warning if the user ran cron_file with an illegal filename, and
2) this were mentioned in the documentation of the cron-module, and
3) the cron-module had a "sanitise" option which would replace dots with some optional character.
Point 1 is the most important, it will save others from the going through the same debugging as we had to ;-)
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
```
| True | Verify that cron_file argument is valid. - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
cron
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.0.0.2
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ubuntu
##### SUMMARY
<!--- Explain the problem briefly -->
Cron will only read cron-files without . (dots) in them, particularly
`[The filenames] must consist solely of upper- and lower-case letters, digits, underscores, and hyphens. This means that they cannot contain any dots.`
It would be nice if
1) The cron-module gave a warning if the user ran cron_file with an illegal filename, and
2) this were mentioned in the documentation of the cron-module, and
3) the cron-module had a "sanitise" option which would replace dots with some optional character.
Point 1 is the most important, it will save others from the going through the same debugging as we had to ;-)
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
```
| main | verify that cron file argument is valid issue type feature idea component name cron ansible version ansible config file configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific ubuntu summary cron will only read cron files without dots in them particularly must consist solely of upper and lower case letters digits underscores and hyphens this means that they cannot contain any dots it would be nice if the cron module gave a warning if the user ran cron file with an illegal filename and this were mentioned in the documentation of the cron module and the cron module had a sanitise option which would replace dots with some optional character point is the most important it will save others from the going through the same debugging as we had to steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used expected results actual results | 1 |
334 | 3,135,496,669 | IssuesEvent | 2015-09-10 15:29:46 | tgstation/-tg-station | https://api.github.com/repos/tgstation/-tg-station | closed | Drone dispenser cooldown cannot be modified | Consistency Issue Easy Fix Feature Request Maintainability - Hinders improvements Not a bug | The dispenser cooldown is hardcoded rather than a var. This inconsistent with almost every other thing that has a cooldown, and makes it so the cooldown time between drones is impossible to modify. | True | Drone dispenser cooldown cannot be modified - The dispenser cooldown is hardcoded rather than a var. This inconsistent with almost every other thing that has a cooldown, and makes it so the cooldown time between drones is impossible to modify. | main | drone dispenser cooldown cannot be modified the dispenser cooldown is hardcoded rather than a var this inconsistent with almost every other thing that has a cooldown and makes it so the cooldown time between drones is impossible to modify | 1 |
3,288 | 12,561,328,133 | IssuesEvent | 2020-06-08 01:08:31 | short-d/short | https://api.github.com/repos/short-d/short | reopened | [Refactor] Check error types in tests. | Go maintainability | **What is frustrating you?**
Test cases that expect errors usually pass if it has any error, which is not correct if the error is different from what is expected.
For example:
Test case struct has the field:
https://github.com/short-d/short/blob/e506d994fb6b8225e7b9f714931acb3d045b5547/backend/app/usecase/url/urlcreator_test.go#L38 Then checking the expected behavior is done in this way:
https://github.com/short-d/short/blob/e506d994fb6b8225e7b9f714931acb3d045b5547/backend/app/usecase/url/urlcreator_test.go#L192-L193
**Your solution**
Tests should pass only if the error type is precisely the one expected.
| True | [Refactor] Check error types in tests. - **What is frustrating you?**
Test cases that expect errors usually pass if it has any error, which is not correct if the error is different from what is expected.
For example:
Test case struct has the field:
https://github.com/short-d/short/blob/e506d994fb6b8225e7b9f714931acb3d045b5547/backend/app/usecase/url/urlcreator_test.go#L38 Then checking the expected behavior is done in this way:
https://github.com/short-d/short/blob/e506d994fb6b8225e7b9f714931acb3d045b5547/backend/app/usecase/url/urlcreator_test.go#L192-L193
**Your solution**
Tests should pass only if the error type is precisely the one expected.
| main | check error types in tests what is frustrating you test cases that expect errors usually pass if it has any error which is not correct if the error is different from what is expected for example test case struct has the field then checking the expected behavior is done in this way your solution tests should pass only if the error type is precisely the one expected | 1 |
635,407 | 20,387,022,895 | IssuesEvent | 2022-02-22 08:12:26 | enso-org/enso | https://api.github.com/repos/enso-org/enso | closed | up_to doesnt work correctly when range start is 0 | Type: Bug Status: Help Wanted Difficulty: Beginner Change: Non-Breaking Priority: High Size: Small Status: Good First Issue Category: Stdlib | <!--
Please ensure that you are running the latest version of Enso before reporting
the bug! It may have been fixed since.
-->
### General Summary
the up_to operator doesnt work correctly when range starts at 0.
### Steps to Reproduce
1. open the enso IDE
2. enter the graph below (this is the error case - range 0..6 returns a list [0..4] with a length of 5)

3. change the range start to 1 -( correct use - range 1..6 returns a list [1..5] with a length of 5)

| 1.0 | up_to doesnt work correctly when range start is 0 - <!--
Please ensure that you are running the latest version of Enso before reporting
the bug! It may have been fixed since.
-->
### General Summary
the up_to operator doesnt work correctly when range starts at 0.
### Steps to Reproduce
1. open the enso IDE
2. enter the graph below (this is the error case - range 0..6 returns a list [0..4] with a length of 5)

3. change the range start to 1 -( correct use - range 1..6 returns a list [1..5] with a length of 5)

| non_main | up to doesnt work correctly when range start is please ensure that you are running the latest version of enso before reporting the bug it may have been fixed since general summary the up to operator doesnt work correctly when range starts at steps to reproduce open the enso ide enter the graph below this is the error case range returns a list with a length of change the range start to correct use range returns a list with a length of | 0 |
1,493 | 6,122,924,471 | IssuesEvent | 2017-06-23 02:02:42 | Endogix/WebFormWeaver | https://api.github.com/repos/Endogix/WebFormWeaver | opened | Export form structure | architecture business logic feature presentation logic | Allow the user to export the form structure when they are finished editing the form, through a button near the bottom.
Exporting will show a modal with the code - as optionally either HTML code, JSON, or XML.
There should also be the option to specify an endpoint that this script can connect to through AJAX to save to a server, with another option to specify a function to call when the connection is complete (most likely to be a redirect function).
## Acceptance criteria
User should be able to:
- [ ] Export the form as HTML code
- [ ] Export the form as a JSON string
- [ ] Export the form as an XML string
- [ ] Export the form to an endpoint through AJAX (either JSON or XML)
- [ ] Configure a function in the options to call when then connection to the endpoint is complete | 1.0 | Export form structure - Allow the user to export the form structure when they are finished editing the form, through a button near the bottom.
Exporting will show a modal with the code - as optionally either HTML code, JSON, or XML.
There should also be the option to specify an endpoint that this script can connect to through AJAX to save to a server, with another option to specify a function to call when the connection is complete (most likely to be a redirect function).
## Acceptance criteria
User should be able to:
- [ ] Export the form as HTML code
- [ ] Export the form as a JSON string
- [ ] Export the form as an XML string
- [ ] Export the form to an endpoint through AJAX (either JSON or XML)
- [ ] Configure a function in the options to call when then connection to the endpoint is complete | non_main | export form structure allow the user to export the form structure when they are finished editing the form through a button near the bottom exporting will show a modal with the code as optionally either html code json or xml there should also be the option to specify an endpoint that this script can connect to through ajax to save to a server with another option to specify a function to call when the connection is complete most likely to be a redirect function acceptance criteria user should be able to export the form as html code export the form as a json string export the form as an xml string export the form to an endpoint through ajax either json or xml configure a function in the options to call when then connection to the endpoint is complete | 0 |
204,433 | 15,442,276,956 | IssuesEvent | 2021-03-08 07:27:35 | trevorNgo/Measure2.0 | https://api.github.com/repos/trevorNgo/Measure2.0 | opened | CS4ZP6 Tester Feedback: Archive Year Term have wrong notification time | tester | Environment
Google Chrome
Windows 10

The job finished time should be current time
Please list the reproduction steps
Login as Admin
Navigate to Archive Year Term to the end | 1.0 | CS4ZP6 Tester Feedback: Archive Year Term have wrong notification time - Environment
Google Chrome
Windows 10

The job finished time should be current time
Please list the reproduction steps
Login as Admin
Navigate to Archive Year Term to the end | non_main | tester feedback archive year term have wrong notification time environment google chrome windows the job finished time should be current time please list the reproduction steps login as admin navigate to archive year term to the end | 0 |
2,087 | 7,097,823,626 | IssuesEvent | 2018-01-14 23:29:18 | AndrewJGregory/Procify | https://api.github.com/repos/AndrewJGregory/Procify | closed | Playlist jbuilder views | maintainability | The jbuilder views for playlists (index, new, and show) use the same internal code. This logic should be extracted into partial and be reused by all three views rather than having duplicate code. | True | Playlist jbuilder views - The jbuilder views for playlists (index, new, and show) use the same internal code. This logic should be extracted into partial and be reused by all three views rather than having duplicate code. | main | playlist jbuilder views the jbuilder views for playlists index new and show use the same internal code this logic should be extracted into partial and be reused by all three views rather than having duplicate code | 1 |
409,183 | 11,958,115,321 | IssuesEvent | 2020-04-04 16:52:32 | Thiago-Viotto/Meus-Filmes-App | https://api.github.com/repos/Thiago-Viotto/Meus-Filmes-App | opened | [Filmes] Botões estão mal diagramados | bug priority_low | ### Expected result
Botões alinhados
### Actual result

| 1.0 | [Filmes] Botões estão mal diagramados - ### Expected result
Botões alinhados
### Actual result

| non_main | botões estão mal diagramados expected result botões alinhados actual result | 0 |
2,999 | 10,974,552,114 | IssuesEvent | 2019-11-29 09:24:44 | test-kitchen/kitchen-azurerm | https://api.github.com/repos/test-kitchen/kitchen-azurerm | closed | public_ip and vnet_id are currently mutually exclusive. | maintainer unable to repro | # Version:
0.14.9
# Environment:
N/A
# Scenario:
Azure supports a public IP when a machine is joined to a vnet. We need the public IP for winrm orchestration -- so test kitchen can continue to run chef -- and the VM to be on the subnet so it can access IP addresses on the subnet. We'd like to avoid having to be connected to the same vnet using a vpn client.
# Steps to Reproduce:
Just use kitchen.yml example 4 - deploy VM to existing virtual network/subnet (use for ExpressRoute/VPN scenarios) and set public_ip flag to true. Check the portal, and the machine is not added to the vnet.
# Expected Result:
I'm expecting the machine to be added to the vnet but also have a public ip.
# Actual Result:
The machine is NOT added to the vnet but does have a public ip.
| True | public_ip and vnet_id are currently mutually exclusive. - # Version:
0.14.9
# Environment:
N/A
# Scenario:
Azure supports a public IP when a machine is joined to a vnet. We need the public IP for winrm orchestration -- so test kitchen can continue to run chef -- and the VM to be on the subnet so it can access IP addresses on the subnet. We'd like to avoid having to be connected to the same vnet using a vpn client.
# Steps to Reproduce:
Just use kitchen.yml example 4 - deploy VM to existing virtual network/subnet (use for ExpressRoute/VPN scenarios) and set public_ip flag to true. Check the portal, and the machine is not added to the vnet.
# Expected Result:
I'm expecting the machine to be added to the vnet but also have a public ip.
# Actual Result:
The machine is NOT added to the vnet but does have a public ip.
| main | public ip and vnet id are currently mutually exclusive version environment n a scenario azure supports a public ip when a machine is joined to a vnet we need the public ip for winrm orchestration so test kitchen can continue to run chef and the vm to be on the subnet so it can access ip addresses on the subnet we d like to avoid having to be connected to the same vnet using a vpn client steps to reproduce just use kitchen yml example deploy vm to existing virtual network subnet use for expressroute vpn scenarios and set public ip flag to true check the portal and the machine is not added to the vnet expected result i m expecting the machine to be added to the vnet but also have a public ip actual result the machine is not added to the vnet but does have a public ip | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.