Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
895 | 4,554,275,277 | IssuesEvent | 2016-09-13 08:53:05 | Particular/NServiceBus.SqlServer | https://api.github.com/repos/Particular/NServiceBus.SqlServer | closed | Multi instance connection provider callback | State: In Progress - Maintainer Prio Tag: Maintainer Prio Type: Bug | ### Problem
The callback is called for satellites / subqueues too. This has the result that a satellite queue can use a different connection than the endpoint queue. This even works in V3 as the current version of the multi-instance sample demonstrates but in V2 accessing a satellite queue seem to result in using the connection that is used for the endpoint.
The string value passed to the callback contains the satellite name and the schema. This value cannot really be parsed. Using a **StartsWith** is also not possible as this would not work for overlapping endpoint names
* Satelite queues must be accessed using the same connection as the endpoint queue
* The callback should not be called specifying the subqueue / schema / etc as this prevents from using a value comparison and even if it was called it should not require to parse the string value these values.
Following PR fixes the multi instance sample:
* https://github.com/Particular/docs.particular.net/pull/1607
### Solution
Currently there is no (easy) way to check at the SqlServer transport level if the stringified address that is passed from the Core represents a local endpoints queue or remote destination.
The problems are caused mainly due to the fact the SqlServer addresses do not have a structure apart from parts before and after `@`. For e.g. `Sales.Something` might be a satellite of logical endpoint named `Sale` or it might be a separate logical endpoint altogether. As a result it is not possible to provide helper methods of type `x.BelongsToEndpoint` and `x.IsQueue`. For the same reason it is not possible not to call the factory method only for main queue and not satellites or to pass only endpoint name and not full transport address. This is the case in V2 as well where the pull type of configuration methods are called also for satellites.
As a result the decision was made to remove schema part when passing address to the factory and secondly to provide a better sample that would provide more idiomatic approach to returning the connections.
### PoA
- [x] Change extension method and pass only queue name to the factory method
- [x] Provide a better sample for configuring multi-instance
| True | Multi instance connection provider callback - ### Problem
The callback is called for satellites / subqueues too. This has the result that a satellite queue can use a different connection than the endpoint queue. This even works in V3 as the current version of the multi-instance sample demonstrates but in V2 accessing a satellite queue seem to result in using the connection that is used for the endpoint.
The string value passed to the callback contains the satellite name and the schema. This value cannot really be parsed. Using a **StartsWith** is also not possible as this would not work for overlapping endpoint names
* Satelite queues must be accessed using the same connection as the endpoint queue
* The callback should not be called specifying the subqueue / schema / etc as this prevents from using a value comparison and even if it was called it should not require to parse the string value these values.
Following PR fixes the multi instance sample:
* https://github.com/Particular/docs.particular.net/pull/1607
### Solution
Currently there is no (easy) way to check at the SqlServer transport level if the stringified address that is passed from the Core represents a local endpoints queue or remote destination.
The problems are caused mainly due to the fact the SqlServer addresses do not have a structure apart from parts before and after `@`. For e.g. `Sales.Something` might be a satellite of logical endpoint named `Sale` or it might be a separate logical endpoint altogether. As a result it is not possible to provide helper methods of type `x.BelongsToEndpoint` and `x.IsQueue`. For the same reason it is not possible not to call the factory method only for main queue and not satellites or to pass only endpoint name and not full transport address. This is the case in V2 as well where the pull type of configuration methods are called also for satellites.
As a result the decision was made to remove schema part when passing address to the factory and secondly to provide a better sample that would provide more idiomatic approach to returning the connections.
### PoA
- [x] Change extension method and pass only queue name to the factory method
- [x] Provide a better sample for configuring multi-instance
| main | multi instance connection provider callback problem the callback is called for satellites subqueues too this has the result that a satellite queue can use a different connection than the endpoint queue this even works in as the current version of the multi instance sample demonstrates but in accessing a satellite queue seem to result in using the connection that is used for the endpoint the string value passed to the callback contains the satellite name and the schema this value cannot really be parsed using a startswith is also not possible as this would not work for overlapping endpoint names satelite queues must be accessed using the same connection as the endpoint queue the callback should not be called specifying the subqueue schema etc as this prevents from using a value comparison and even if it was called it should not require to parse the string value these values following pr fixes the multi instance sample solution currently there is no easy way to check at the sqlserver transport level if the stringified address that is passed from the core represents a local endpoints queue or remote destination the problems are caused mainly due to the fact the sqlserver addresses do not have a structure apart from parts before and after for e g sales something might be a satellite of logical endpoint named sale or it might be a separate logical endpoint altogether as a result it is not possible to provide helper methods of type x belongstoendpoint and x isqueue for the same reason it is not possible not to call the factory method only for main queue and not satellites or to pass only endpoint name and not full transport address this is the case in as well where the pull type of configuration methods are called also for satellites as a result the decision was made to remove schema part when passing address to the factory and secondly to provide a better sample that would provide more idiomatic approach to returning the connections poa change extension method and pass only queue name to the factory method provide a better sample for configuring multi instance | 1 |
1,528 | 6,572,217,542 | IssuesEvent | 2017-09-11 00:10:52 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | lvol does not work with size suffix 'b' or 'B' | affects_1.9 bug_report easyfix waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### ANSIBLE VERSION
```
ansible 1.9.3
configured module search path = None
```
##### CONFIGURATION
<!---
-->
pipelining=True
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Control host: openSuSE 13.1
Managed host: Scientific Linux 7.2
##### SUMMARY
<!--- Explain the problem briefly -->
lvol produces an internal parse error when trying to create a volume with a size given in bytes.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
```
ansible <remote host> -i hosts -u root -m lvol -a "lv='scratch' vg='vg_system' size='10000000B'"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
That the lv is created.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
```
node520 | FAILED >> {
"failed": true,
"msg": "OpenSSH_6.2p2, OpenSSL 1.0.1k 8 Jan 2015\ndebug1: Reading configuration data /homes/mathphys/sfs/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 18: Applying options for *\r\ndebug1: /etc/ssh/ssh_config line 47: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug1: mux_client_request_session: master session id: 2\r\nX11 forwarding request failed\r\nTraceback (most recent call last):\n File \"<stdin>\", line 1852, in <module>\n File \"<stdin>\", line 167, in main\n File \"<stdin>\", line 93, in parse_lvs\nValueError: invalid literal for int() with base 10: '209715200B'\n",
"parsed": false
}
```
| True | lvol does not work with size suffix 'b' or 'B' - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### ANSIBLE VERSION
```
ansible 1.9.3
configured module search path = None
```
##### CONFIGURATION
<!---
-->
pipelining=True
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Control host: openSuSE 13.1
Managed host: Scientific Linux 7.2
##### SUMMARY
<!--- Explain the problem briefly -->
lvol produces an internal parse error when trying to create a volume with a size given in bytes.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
```
ansible <remote host> -i hosts -u root -m lvol -a "lv='scratch' vg='vg_system' size='10000000B'"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
That the lv is created.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
```
node520 | FAILED >> {
"failed": true,
"msg": "OpenSSH_6.2p2, OpenSSL 1.0.1k 8 Jan 2015\ndebug1: Reading configuration data /homes/mathphys/sfs/.ssh/config\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 18: Applying options for *\r\ndebug1: /etc/ssh/ssh_config line 47: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug1: mux_client_request_session: master session id: 2\r\nX11 forwarding request failed\r\nTraceback (most recent call last):\n File \"<stdin>\", line 1852, in <module>\n File \"<stdin>\", line 167, in main\n File \"<stdin>\", line 93, in parse_lvs\nValueError: invalid literal for int() with base 10: '209715200B'\n",
"parsed": false
}
```
| main | lvol does not work with size suffix b or b issue type bug report ansible version ansible configured module search path none configuration pipelining true os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific control host opensuse managed host scientific linux summary lvol produces an internal parse error when trying to create a volume with a size given in bytes steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used ansible i hosts u root m lvol a lv scratch vg vg system size expected results that the lv is created actual results failed failed true msg openssh openssl jan reading configuration data homes mathphys sfs ssh config r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r etc ssh ssh config line applying options for r auto mux trying existing master r mux client request session master session id r forwarding request failed r ntraceback most recent call last n file line in n file line in main n file line in parse lvs nvalueerror invalid literal for int with base n parsed false | 1 |
4,079 | 19,268,995,936 | IssuesEvent | 2021-12-10 01:31:30 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | No way to pass additional arguments to delve for Go debugging | area/debugging maintainer/need-response | ### Description:
Delve does not support `DEBUGGER_ARGS` or any other way to pass arguments to the debugger. This is important to get around `Version of Delve is too old for this version of Go (maximum supported version 1.14, suppress this error with --check-go-version=false)` errors when trying to debug go 1.15 lambdas with goland 2020.1/2020.2
### Steps to reproduce:
Try passing any DEBUGGER_ARGS to delve. `delveArgs` in `runHandler` does not take any additional arguments from the environment or otherwise.
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: mac
2. `sam --version`: 1.18.1
3. AWS region: Didn't get that far
| True | No way to pass additional arguments to delve for Go debugging - ### Description:
Delve does not support `DEBUGGER_ARGS` or any other way to pass arguments to the debugger. This is important to get around `Version of Delve is too old for this version of Go (maximum supported version 1.14, suppress this error with --check-go-version=false)` errors when trying to debug go 1.15 lambdas with goland 2020.1/2020.2
### Steps to reproduce:
Try passing any DEBUGGER_ARGS to delve. `delveArgs` in `runHandler` does not take any additional arguments from the environment or otherwise.
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: mac
2. `sam --version`: 1.18.1
3. AWS region: Didn't get that far
| main | no way to pass additional arguments to delve for go debugging description delve does not support debugger args or any other way to pass arguments to the debugger this is important to get around version of delve is too old for this version of go maximum supported version suppress this error with check go version false errors when trying to debug go lambdas with goland steps to reproduce try passing any debugger args to delve delveargs in runhandler does not take any additional arguments from the environment or otherwise additional environment details ex windows mac amazon linux etc os mac sam version aws region didn t get that far | 1 |
112,965 | 17,114,077,831 | IssuesEvent | 2021-07-11 00:31:46 | AlexRogalskiy/ws-documents | https://api.github.com/repos/AlexRogalskiy/ws-documents | closed | CVE-2021-21342 (High) detected in xstream-1.4.11.1.jar | no-issue-activity security vulnerability | ## CVE-2021-21342 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.11.1.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p>
<p>Path to dependency file: ws-documents/modules/document-generator/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-netflix-eureka-client-2.2.5.RELEASE.jar (Root Library)
- :x: **xstream-1.4.11.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/ws-documents/commit/f0ca974033183c586fb41762fb7d7f4730a99acc">f0ca974033183c586fb41762fb7d7f4730a99acc</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects. XStream creates therefore new instances based on these type information. An attacker can manipulate the processed input stream and replace or inject objects, that result in a server-side forgery request. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21342>CVE-2021-21342</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-hvv8-336g-rx3m">https://github.com/x-stream/xstream/security/advisories/GHSA-hvv8-336g-rx3m</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-21342 (High) detected in xstream-1.4.11.1.jar - ## CVE-2021-21342 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xstream-1.4.11.1.jar</b></p></summary>
<p>XStream is a serialization library from Java objects to XML and back.</p>
<p>Library home page: <a href="http://x-stream.github.io">http://x-stream.github.io</a></p>
<p>Path to dependency file: ws-documents/modules/document-generator/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar,/home/wss-scanner/.m2/repository/com/thoughtworks/xstream/xstream/1.4.11.1/xstream-1.4.11.1.jar</p>
<p>
Dependency Hierarchy:
- spring-cloud-starter-netflix-eureka-client-2.2.5.RELEASE.jar (Root Library)
- :x: **xstream-1.4.11.1.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/AlexRogalskiy/ws-documents/commit/f0ca974033183c586fb41762fb7d7f4730a99acc">f0ca974033183c586fb41762fb7d7f4730a99acc</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
XStream is a Java library to serialize objects to XML and back again. In XStream before version 1.4.16, there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects. XStream creates therefore new instances based on these type information. An attacker can manipulate the processed input stream and replace or inject objects, that result in a server-side forgery request. No user is affected, who followed the recommendation to setup XStream's security framework with a whitelist limited to the minimal required types. If you rely on XStream's default blacklist of the Security Framework, you will have to use at least version 1.4.16.
<p>Publish Date: 2021-03-23
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-21342>CVE-2021-21342</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/x-stream/xstream/security/advisories/GHSA-hvv8-336g-rx3m">https://github.com/x-stream/xstream/security/advisories/GHSA-hvv8-336g-rx3m</a></p>
<p>Release Date: 2021-03-23</p>
<p>Fix Resolution: com.thoughtworks.xstream:xstream:1.4.16</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in xstream jar cve high severity vulnerability vulnerable library xstream jar xstream is a serialization library from java objects to xml and back library home page a href path to dependency file ws documents modules document generator pom xml path to vulnerable library home wss scanner repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar home wss scanner repository com thoughtworks xstream xstream xstream jar dependency hierarchy spring cloud starter netflix eureka client release jar root library x xstream jar vulnerable library found in head commit a href vulnerability details xstream is a java library to serialize objects to xml and back again in xstream before version there is a vulnerability where the processed stream at unmarshalling time contains type information to recreate the formerly written objects xstream creates therefore new instances based on these type information an attacker can manipulate the processed input stream and replace or inject objects that result in a server side forgery request no user is affected who followed the recommendation to setup xstream s security framework with a whitelist limited to the minimal required types if you rely on xstream s default blacklist of the security framework you will have to use at least version publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com thoughtworks xstream xstream step up your open source security game with whitesource | 0 |
559,116 | 16,550,389,307 | IssuesEvent | 2021-05-28 07:54:36 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | accounts.google.com - site is not usable | browser-firefox-mobile bugbug-probability-high engine-gecko priority-critical | <!-- @browser: Firefox Mobile 88.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:88.0) Gecko/88.0 Firefox/88.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/74685 -->
**URL**: https://accounts.google.com/signin/oauth/consent?authuser=0
**Browser / Version**: Firefox Mobile 88.0
**Operating System**: Android 11
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Doesn't load. Struck at same page where loader keeps on loading.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | accounts.google.com - site is not usable - <!-- @browser: Firefox Mobile 88.0 -->
<!-- @ua_header: Mozilla/5.0 (Android 11; Mobile; rv:88.0) Gecko/88.0 Firefox/88.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/74685 -->
**URL**: https://accounts.google.com/signin/oauth/consent?authuser=0
**Browser / Version**: Firefox Mobile 88.0
**Operating System**: Android 11
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
Doesn't load. Struck at same page where loader keeps on loading.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | accounts google com site is not usable url browser version firefox mobile operating system android tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce doesn t load struck at same page where loader keeps on loading browser configuration none from with ❤️ | 0 |
644,170 | 20,968,931,158 | IssuesEvent | 2022-03-28 09:30:48 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | soundcloud.com - site is not usable | browser-firefox priority-critical engine-gecko | <!-- @browser: Firefox 56.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:56.0) Gecko/20100101 Firefox/56.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/101511 -->
**URL**: https://soundcloud.com
**Browser / Version**: Firefox 56.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Unable to login
**Steps to Reproduce**:
Soundcloud, I am already acct holder. Using current Firefox w/usual addons. Have disabled umatrix, allowed all on NoScript, etc.
Soundcloud displays almost nothing. Mozilla recommends "forget about this site". So, I cleared cache, history, passwords, everything from FF. Site loads properly. Try login. Login says "we think you are robot. Try again". Endless loop. User agent switcher doesn't affect anything, still loads almost blank page.
Open site using Chrome. Everything normal, can log in. Chrome also has umatrix and ublock Origin, but disabled for Soundcloud, same as Firefox.
Works on Chrome, not on Firefox.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | soundcloud.com - site is not usable - <!-- @browser: Firefox 56.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; WOW64; rv:56.0) Gecko/20100101 Firefox/56.0 -->
<!-- @reported_with: unknown -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/101511 -->
**URL**: https://soundcloud.com
**Browser / Version**: Firefox 56.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Unable to login
**Steps to Reproduce**:
Soundcloud, I am already acct holder. Using current Firefox w/usual addons. Have disabled umatrix, allowed all on NoScript, etc.
Soundcloud displays almost nothing. Mozilla recommends "forget about this site". So, I cleared cache, history, passwords, everything from FF. Site loads properly. Try login. Login says "we think you are robot. Try again". Endless loop. User agent switcher doesn't affect anything, still loads almost blank page.
Open site using Chrome. Everything normal, can log in. Chrome also has umatrix and ublock Origin, but disabled for Soundcloud, same as Firefox.
Works on Chrome, not on Firefox.
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | soundcloud com site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description unable to login steps to reproduce soundcloud i am already acct holder using current firefox w usual addons have disabled umatrix allowed all on noscript etc soundcloud displays almost nothing mozilla recommends forget about this site so i cleared cache history passwords everything from ff site loads properly try login login says we think you are robot try again endless loop user agent switcher doesn t affect anything still loads almost blank page open site using chrome everything normal can log in chrome also has umatrix and ublock origin but disabled for soundcloud same as firefox works on chrome not on firefox browser configuration none from with ❤️ | 0 |
5,091 | 26,006,744,234 | IssuesEvent | 2022-12-20 20:10:25 | camunda/zeebe | https://api.github.com/repos/camunda/zeebe | opened | Refactor internal communication layer | kind/toil area/performance area/reliability area/observability area/maintainability | <!--
In case you have questions about our software we encourage everyone to participate in our community via the
- Camunda Platform community forum https://forum.camunda.io/ or
- Slack https://camunda-cloud.slack.com/ (For invite: https://camunda-slack-invite.herokuapp.com/)
There you can exchange ideas with other Zeebe and Camunda Platform 8 users, as well as the product developers, and use the search to find answer to similar questions.
This issue template is used by the Zeebe engineers to create general tasks.
-->
_Sorry if there exist already an issue, but I was not able to find it._
**Description**
We have seen recently that we have many issues with the Atomix based internal networking/communication layer, which we use between gateway and broker and between broker - broker.
For example https://github.com/zeebe-io/zeebe-chaos/issues/294 were we send request over several minutes without detecting that the node was already gone. When trying to fix this issue via https://github.com/camunda/zeebe/pull/11307 it turned out to be rather hard to test and reason about (about the code in general).
This means right now the networking part is hard to maintain, hard to test and somehow a blackhole, since there are no metrics and no good logging.
Ideally we should spent some time to refactor this part of our system to be more confident in our networking, this would include introducing better visibility (logging + metrics ), improve maintability and readability and very important reduce the complexity and improve the testability.
We have already thought several times about (brought up initially by @npepinpe ) it to replace it also with grpc which would be a good opportunity, which also comes with lot of costs and risks of course. We need to discuss this further within the team.
| True | Refactor internal communication layer - <!--
In case you have questions about our software we encourage everyone to participate in our community via the
- Camunda Platform community forum https://forum.camunda.io/ or
- Slack https://camunda-cloud.slack.com/ (For invite: https://camunda-slack-invite.herokuapp.com/)
There you can exchange ideas with other Zeebe and Camunda Platform 8 users, as well as the product developers, and use the search to find answer to similar questions.
This issue template is used by the Zeebe engineers to create general tasks.
-->
_Sorry if there exist already an issue, but I was not able to find it._
**Description**
We have seen recently that we have many issues with the Atomix based internal networking/communication layer, which we use between gateway and broker and between broker - broker.
For example https://github.com/zeebe-io/zeebe-chaos/issues/294 were we send request over several minutes without detecting that the node was already gone. When trying to fix this issue via https://github.com/camunda/zeebe/pull/11307 it turned out to be rather hard to test and reason about (about the code in general).
This means right now the networking part is hard to maintain, hard to test and somehow a blackhole, since there are no metrics and no good logging.
Ideally we should spent some time to refactor this part of our system to be more confident in our networking, this would include introducing better visibility (logging + metrics ), improve maintability and readability and very important reduce the complexity and improve the testability.
We have already thought several times about (brought up initially by @npepinpe ) it to replace it also with grpc which would be a good opportunity, which also comes with lot of costs and risks of course. We need to discuss this further within the team.
| main | refactor internal communication layer in case you have questions about our software we encourage everyone to participate in our community via the camunda platform community forum or slack for invite there you can exchange ideas with other zeebe and camunda platform users as well as the product developers and use the search to find answer to similar questions this issue template is used by the zeebe engineers to create general tasks sorry if there exist already an issue but i was not able to find it description we have seen recently that we have many issues with the atomix based internal networking communication layer which we use between gateway and broker and between broker broker for example were we send request over several minutes without detecting that the node was already gone when trying to fix this issue via it turned out to be rather hard to test and reason about about the code in general this means right now the networking part is hard to maintain hard to test and somehow a blackhole since there are no metrics and no good logging ideally we should spent some time to refactor this part of our system to be more confident in our networking this would include introducing better visibility logging metrics improve maintability and readability and very important reduce the complexity and improve the testability we have already thought several times about brought up initially by npepinpe it to replace it also with grpc which would be a good opportunity which also comes with lot of costs and risks of course we need to discuss this further within the team | 1 |
2,050 | 6,926,474,489 | IssuesEvent | 2017-11-30 19:17:57 | sinatra/sinatra | https://api.github.com/repos/sinatra/sinatra | closed | Help Wanted && Good First Issue && Wishlist | Maintainer | * ["Help Wanted"](https://github.com/sinatra/sinatra/labels/help%20wanted): Anyone willing to pitch in is open to contribute to this ticket as they see fit (will try to add context / summarize or ask for requirements)
* ["Good First Issue"](https://github.com/sinatra/sinatra/labels/good%20first%20issue): Potential first time contributors should start here
* ["Wishlist"](https://github.com/sinatra/sinatra/labels/Wishlist): All the things I wish we had but have no time for | True | Help Wanted && Good First Issue && Wishlist - * ["Help Wanted"](https://github.com/sinatra/sinatra/labels/help%20wanted): Anyone willing to pitch in is open to contribute to this ticket as they see fit (will try to add context / summarize or ask for requirements)
* ["Good First Issue"](https://github.com/sinatra/sinatra/labels/good%20first%20issue): Potential first time contributors should start here
* ["Wishlist"](https://github.com/sinatra/sinatra/labels/Wishlist): All the things I wish we had but have no time for | main | help wanted good first issue wishlist anyone willing to pitch in is open to contribute to this ticket as they see fit will try to add context summarize or ask for requirements potential first time contributors should start here all the things i wish we had but have no time for | 1 |
4,627 | 23,980,726,906 | IssuesEvent | 2022-09-13 14:52:26 | exercism/python | https://api.github.com/repos/exercism/python | closed | [New Concept Exercise] : list-comprehensions | x:status/claimed x:size/large claimed 🐾 maintainer action required❕ new exercise ✨ | This issue describes how to implement the `list-comprehensions` concept exercise for the Python track.
## Getting started
**Please please please read the docs before starting.** Posting PRs without reading these docs will be a lot more frustrating for you during the review cycle, and exhaust Exercism's maintainers' time. So, before diving into the implementation, please read up on the following documents:
- [Contributing to Exercism](https://exercism.org/docs/building) | [Exercism and GitHub](https://exercism.org/docs/building/github) | [Contributor Pull Request Guide](https://exercism.org/docs/building/github/contributors-pull-request-guide)
- [What are those Weird Task Tags about?](https://exercism.org/docs/building/product/tasks)
- [Building Language Tracks: An Overview](https://exercism.org/docs/building/tracks)
- [What are Concepts?](https://exercism.org/docs/building/tracks/concepts)
- [Concept Exercise Specifications](https://exercism.org/docs/building/tracks/concept-exercises)
- [Concept Specifications](https://exercism.org/docs/building/tracks/concepts)
- [Exercism Formatting and Style Guide](https://exercism.org/docs/building/markdown/style-guide)
- [Exercism Markdown Specification](https://exercism.org/docs/building/markdown/markdown)
- [Reputation](https://exercism.org/docs/using/product/reputation)
## Goal
The goal of this exercise is to teach the syntax and variants of `list comprehensions` in Python.
## Learning objectives
- Understand how a `list comprehension` relates to the `loop` + `append` method of creating/computing a `List`.
- Create a list comprehension from a `loop` + `append`
- Create a list comprehension from a `string`, `List`, or other `iterable` (_such as `zip()` or `dict.items()`_)
- Use one or more conditions/operators/methods to filter comprehension inputs
- Use methods or logic to format the elements (output members) of the comprehension
- Create a _nested comprehension_
- Create a _nested comprehension_ with one or more formatting or filtering conditions
## Out of scope
- Memory and performance characteristics and optimizations
- `generators` and `generator expressions` in `list comprehensions`
- How `map()` and `filter()` relate to or differ from `list comprehensions`
- Using the `assignment expression` (_walrus operator_) with a `list comprehension`
## Concepts
- `list-comprehensions`
## Prerequisites
- `basics`
- `bools`
- `comparisons`
- `conditionals`
- `lists`
- `list-methods`
- `loops`
- `iteration`
- `sequences`
## Resources to refer to
- [List Comprehensions (Python official docs)](https://docs.python.org/3/tutorial/datastructures.html#list-comprehensions)
- [Nested List Comprehensions (Python official docs)](https://docs.python.org/3/tutorial/datastructures.html#nested-list-comprehensions)
- [When to Use a List Comprehension in Python (Real Python)](https://realpython.com/list-comprehension-python/)
- [Python List Comprehensions: Explained Visually (Trey Hunner)](https://treyhunner.com/2015/12/python-list-comprehensions-now-in-color/)
### Hints
- `List Comprehensions` section of the Python docs tutorial: [List Comprehensions](https://docs.python.org/3/tutorial/datastructures.html#list-comprehensions)
- This animated GIF from Trey Hunner: [List Comprehensions: The Movie](https://treyhunner.com/images/list-comprehension-condition.gif)
### After
- `comprehension syntax` for other data structures such as `sets` and `dictionaries`
- `generators` and `generator expressions`
- `generators` and `generator expressions` in `list comprehensions`
## Representer
No changes required.
## Analyzer
No changes rquired.
## Implementing
- Tests should be written using unittest.TestCase, and the test file named list_comprehensions_test.py.
- [How to Implement a Concept Exercise in Python](https://github.com/exercism/v3/blob/master/languages/python/reference/implementing-a-concept-exercise.md)
- [make-concept-exercise Utility](https://github.com/exercism/v3/tree/master/languages/python/bin)
## Help
If you have any questions while implementing the exercise, please post the questions as comments in this issue.
| True | [New Concept Exercise] : list-comprehensions - This issue describes how to implement the `list-comprehensions` concept exercise for the Python track.
## Getting started
**Please please please read the docs before starting.** Posting PRs without reading these docs will be a lot more frustrating for you during the review cycle, and exhaust Exercism's maintainers' time. So, before diving into the implementation, please read up on the following documents:
- [Contributing to Exercism](https://exercism.org/docs/building) | [Exercism and GitHub](https://exercism.org/docs/building/github) | [Contributor Pull Request Guide](https://exercism.org/docs/building/github/contributors-pull-request-guide)
- [What are those Weird Task Tags about?](https://exercism.org/docs/building/product/tasks)
- [Building Language Tracks: An Overview](https://exercism.org/docs/building/tracks)
- [What are Concepts?](https://exercism.org/docs/building/tracks/concepts)
- [Concept Exercise Specifications](https://exercism.org/docs/building/tracks/concept-exercises)
- [Concept Specifications](https://exercism.org/docs/building/tracks/concepts)
- [Exercism Formatting and Style Guide](https://exercism.org/docs/building/markdown/style-guide)
- [Exercism Markdown Specification](https://exercism.org/docs/building/markdown/markdown)
- [Reputation](https://exercism.org/docs/using/product/reputation)
## Goal
The goal of this exercise is to teach the syntax and variants of `list comprehensions` in Python.
## Learning objectives
- Understand how a `list comprehension` relates to the `loop` + `append` method of creating/computing a `List`.
- Create a list comprehension from a `loop` + `append`
- Create a list comprehension from a `string`, `List`, or other `iterable` (_such as `zip()` or `dict.items()`_)
- Use one or more conditions/operators/methods to filter comprehension inputs
- Use methods or logic to format the elements (output members) of the comprehension
- Create a _nested comprehension_
- Create a _nested comprehension_ with one or more formatting or filtering conditions
## Out of scope
- Memory and performance characteristics and optimizations
- `generators` and `generator expressions` in `list comprehensions`
- How `map()` and `filter()` relate to or differ from `list comprehensions`
- Using the `assignment expression` (_walrus operator_) with a `list comprehension`
## Concepts
- `list-comprehensions`
## Prerequisites
- `basics`
- `bools`
- `comparisons`
- `conditionals`
- `lists`
- `list-methods`
- `loops`
- `iteration`
- `sequences`
## Resources to refer to
- [List Comprehensions (Python official docs)](https://docs.python.org/3/tutorial/datastructures.html#list-comprehensions)
- [Nested List Comprehensions (Python official docs)](https://docs.python.org/3/tutorial/datastructures.html#nested-list-comprehensions)
- [When to Use a List Comprehension in Python (Real Python)](https://realpython.com/list-comprehension-python/)
- [Python List Comprehensions: Explained Visually (Trey Hunner)](https://treyhunner.com/2015/12/python-list-comprehensions-now-in-color/)
### Hints
- `List Comprehensions` section of the Python docs tutorial: [List Comprehensions](https://docs.python.org/3/tutorial/datastructures.html#list-comprehensions)
- This animated GIF from Trey Hunner: [List Comprehensions: The Movie](https://treyhunner.com/images/list-comprehension-condition.gif)
### After
- `comprehension syntax` for other data structures such as `sets` and `dictionaries`
- `generators` and `generator expressions`
- `generators` and `generator expressions` in `list comprehensions`
## Representer
No changes required.
## Analyzer
No changes rquired.
## Implementing
- Tests should be written using unittest.TestCase, and the test file named list_comprehensions_test.py.
- [How to Implement a Concept Exercise in Python](https://github.com/exercism/v3/blob/master/languages/python/reference/implementing-a-concept-exercise.md)
- [make-concept-exercise Utility](https://github.com/exercism/v3/tree/master/languages/python/bin)
## Help
If you have any questions while implementing the exercise, please post the questions as comments in this issue.
| main | list comprehensions this issue describes how to implement the list comprehensions concept exercise for the python track getting started please please please read the docs before starting posting prs without reading these docs will be a lot more frustrating for you during the review cycle and exhaust exercism s maintainers time so before diving into the implementation please read up on the following documents goal the goal of this exercise is to teach the syntax and variants of list comprehensions in python learning objectives understand how a list comprehension relates to the loop append method of creating computing a list create a list comprehension from a loop append create a list comprehension from a string list or other iterable such as zip or dict items use one or more conditions operators methods to filter comprehension inputs use methods or logic to format the elements output members of the comprehension create a nested comprehension create a nested comprehension with one or more formatting or filtering conditions out of scope memory and performance characteristics and optimizations generators and generator expressions in list comprehensions how map and filter relate to or differ from list comprehensions using the assignment expression walrus operator with a list comprehension concepts list comprehensions prerequisites basics bools comparisons conditionals lists list methods loops iteration sequences resources to refer to hints list comprehensions section of the python docs tutorial this animated gif from trey hunner after comprehension syntax for other data structures such as sets and dictionaries generators and generator expressions generators and generator expressions in list comprehensions representer no changes required analyzer no changes rquired implementing tests should be written using unittest testcase and the test file named list comprehensions test py help if you have any questions while implementing the exercise please post the questions as comments in this issue | 1 |
36,166 | 8,056,848,647 | IssuesEvent | 2018-08-02 13:54:02 | usnistgov/ACVP | https://api.github.com/repos/usnistgov/ACVP | closed | CTR DRBG: test vectors without PR seem wrong | Numerical code bug Spec error | I get the following vectors for predictionresistance == false:
```
"testType" : "AFT",
"mode" : "AES-128",
"derFunc" : true,
"predResistance" : false,
"entropyInputLen" : 128,
"reSeed" : true,
"nonceLen" : 64,
"persoStringLen" : 0,
"additionalInputLen" : 0,
"returnedBitsLen" : 512,
"tests" : [ {
"tcId" : 1,
"entropyInput" : "CAED3A22E6CEF0801CB84C393CAD1B10",
"nonce" : "214AC39FACDA5E84",
"persoString" : "",
"otherInput" : [ {
"additionalInput" : "",
"entropyInput" : "5DBF858FCF24267F3C1CBFD2B3CAF805"
}, {
"additionalInput" : "",
"entropyInput" : ""
}, {
"additionalInput" : "",
"entropyInput" : ""
} ]
}, {
```
The providing of otherInput seems wrong as we do not have prediction resistance. Besides, there are 3 entries in otherInput. | 1.0 | CTR DRBG: test vectors without PR seem wrong - I get the following vectors for predictionresistance == false:
```
"testType" : "AFT",
"mode" : "AES-128",
"derFunc" : true,
"predResistance" : false,
"entropyInputLen" : 128,
"reSeed" : true,
"nonceLen" : 64,
"persoStringLen" : 0,
"additionalInputLen" : 0,
"returnedBitsLen" : 512,
"tests" : [ {
"tcId" : 1,
"entropyInput" : "CAED3A22E6CEF0801CB84C393CAD1B10",
"nonce" : "214AC39FACDA5E84",
"persoString" : "",
"otherInput" : [ {
"additionalInput" : "",
"entropyInput" : "5DBF858FCF24267F3C1CBFD2B3CAF805"
}, {
"additionalInput" : "",
"entropyInput" : ""
}, {
"additionalInput" : "",
"entropyInput" : ""
} ]
}, {
```
The providing of otherInput seems wrong as we do not have prediction resistance. Besides, there are 3 entries in otherInput. | non_main | ctr drbg test vectors without pr seem wrong i get the following vectors for predictionresistance false testtype aft mode aes derfunc true predresistance false entropyinputlen reseed true noncelen persostringlen additionalinputlen returnedbitslen tests tcid entropyinput nonce persostring otherinput additionalinput entropyinput additionalinput entropyinput additionalinput entropyinput the providing of otherinput seems wrong as we do not have prediction resistance besides there are entries in otherinput | 0 |
2,909 | 10,342,857,233 | IssuesEvent | 2019-09-04 07:37:25 | sapcc/maia | https://api.github.com/repos/sapcc/maia | opened | Update prometheus dependencies from 1.8 to latest stable | maintainability | Update the used prometheus packages to latest stable to avoid future incompatibilities.
Most of the packages are still available in later versions of Prometheus, except for `storage/metric`. From that package we use only `LabelMatcher` which has been replaced by a new type of the same name in package `prompb`. | True | Update prometheus dependencies from 1.8 to latest stable - Update the used prometheus packages to latest stable to avoid future incompatibilities.
Most of the packages are still available in later versions of Prometheus, except for `storage/metric`. From that package we use only `LabelMatcher` which has been replaced by a new type of the same name in package `prompb`. | main | update prometheus dependencies from to latest stable update the used prometheus packages to latest stable to avoid future incompatibilities most of the packages are still available in later versions of prometheus except for storage metric from that package we use only labelmatcher which has been replaced by a new type of the same name in package prompb | 1 |
640,815 | 20,799,993,636 | IssuesEvent | 2022-03-17 13:05:07 | gofiber/fiber | https://api.github.com/repos/gofiber/fiber | reopened | C.JSON not work on go 1.18 | 🚨 High Priority ☢️ Bug | in go 1.17.8 c.json work fine,but after i upgrade go version , c.json not work report "runtime error empty pointer" | 1.0 | C.JSON not work on go 1.18 - in go 1.17.8 c.json work fine,but after i upgrade go version , c.json not work report "runtime error empty pointer" | non_main | c json not work on go in go c json work fine but after i upgrade go version c json not work report runtime error empty pointer | 0 |
56,261 | 6,971,583,801 | IssuesEvent | 2017-12-11 14:30:20 | GetTerminus/terminus-ui | https://api.github.com/repos/GetTerminus/terminus-ui | opened | Collapsable sidebar / gutter | Focus: component Needs: design Needs: exploration Type: feature | We need a sidebar for actions etc.
- This sidebar will originally be only for admin functionality.
- The sidebar will receive all the actions that it can include.
- Selecting an action will emit that action.
- Sidebar should include a fuzzy finder search bar for quick command access
- Autocomplete by typing, enter selects the highlighted option. | 1.0 | Collapsable sidebar / gutter - We need a sidebar for actions etc.
- This sidebar will originally be only for admin functionality.
- The sidebar will receive all the actions that it can include.
- Selecting an action will emit that action.
- Sidebar should include a fuzzy finder search bar for quick command access
- Autocomplete by typing, enter selects the highlighted option. | non_main | collapsable sidebar gutter we need a sidebar for actions etc this sidebar will originally be only for admin functionality the sidebar will receive all the actions that it can include selecting an action will emit that action sidebar should include a fuzzy finder search bar for quick command access autocomplete by typing enter selects the highlighted option | 0 |
45,759 | 5,731,673,871 | IssuesEvent | 2017-04-21 13:04:31 | pruje/libbash.sh | https://api.github.com/repos/pruje/libbash.sh | closed | lb_result() does not work on macOS without specifying an exit code | bug macOS NEED TEST | When using `lb_result` without argument (exit code), it is supposed to pick up the last result (`$?`).
It works on Linux systems, but apparently not on macOS which returns always 0 as last result.
Need to test this further and specify it in documentation. | 1.0 | lb_result() does not work on macOS without specifying an exit code - When using `lb_result` without argument (exit code), it is supposed to pick up the last result (`$?`).
It works on Linux systems, but apparently not on macOS which returns always 0 as last result.
Need to test this further and specify it in documentation. | non_main | lb result does not work on macos without specifying an exit code when using lb result without argument exit code it is supposed to pick up the last result it works on linux systems but apparently not on macos which returns always as last result need to test this further and specify it in documentation | 0 |
2,978 | 10,723,468,957 | IssuesEvent | 2019-10-27 19:00:34 | vostpt/mobile-app | https://api.github.com/repos/vostpt/mobile-app | closed | Contributors Screen | Needs Maintainers Help hacktoberfest | **Description**
Create screen with all the contributors for the app
**File Location**
```
- presentation
|__ ui
```
**Requirements**
- From the Github API, get the current contributors information and display it in form of list with:
- User image
- User Name
- User description
**UI**
<img width="364" alt="imagem" src="https://user-images.githubusercontent.com/10728633/63040338-29bf2580-bebd-11e9-86d2-0741927e0392.png">
**Notes**
Expect the following object:
```
{
"username": "String",
"name": "String",
"profilePicture": "String"
"category": CategoryEnum
}
```
With Category enum having the values:
```
copy,
design,
development
```
| True | Contributors Screen - **Description**
Create screen with all the contributors for the app
**File Location**
```
- presentation
|__ ui
```
**Requirements**
- From the Github API, get the current contributors information and display it in form of list with:
- User image
- User Name
- User description
**UI**
<img width="364" alt="imagem" src="https://user-images.githubusercontent.com/10728633/63040338-29bf2580-bebd-11e9-86d2-0741927e0392.png">
**Notes**
Expect the following object:
```
{
"username": "String",
"name": "String",
"profilePicture": "String"
"category": CategoryEnum
}
```
With Category enum having the values:
```
copy,
design,
development
```
| main | contributors screen description create screen with all the contributors for the app file location presentation ui requirements from the github api get the current contributors information and display it in form of list with user image user name user description ui img width alt imagem src notes expect the following object username string name string profilepicture string category categoryenum with category enum having the values copy design development | 1 |
31,039 | 2,731,255,441 | IssuesEvent | 2015-04-16 19:16:39 | NuGet/Home | https://api.github.com/repos/NuGet/Home | closed | Functional test on satellite package is failing | 2 - Working Area:SDK Priority:1 Type:Bug | Following Functional tests on satellite package are failing
InstallSatellitePackageCopiesFilesToRuntimeFolderWhenInstalledAsDependency
InstallSatellitePackageCopiesFilesToExistingRuntimePackageFolder
InstallingSatellitePackageOnlyCopiesCultureSpecificLibFolderContents
<!---
@huboard:{"order":2347.79296875,"milestone_order":271,"custom_state":""}
-->
| 1.0 | Functional test on satellite package is failing - Following Functional tests on satellite package are failing
InstallSatellitePackageCopiesFilesToRuntimeFolderWhenInstalledAsDependency
InstallSatellitePackageCopiesFilesToExistingRuntimePackageFolder
InstallingSatellitePackageOnlyCopiesCultureSpecificLibFolderContents
<!---
@huboard:{"order":2347.79296875,"milestone_order":271,"custom_state":""}
-->
| non_main | functional test on satellite package is failing following functional tests on satellite package are failing installsatellitepackagecopiesfilestoruntimefolderwheninstalledasdependency installsatellitepackagecopiesfilestoexistingruntimepackagefolder installingsatellitepackageonlycopiesculturespecificlibfoldercontents huboard order milestone order custom state | 0 |
418,743 | 28,125,609,227 | IssuesEvent | 2023-03-31 17:26:03 | ekumenlabs/gauges2 | https://api.github.com/repos/ekumenlabs/gauges2 | opened | Document the repository | documentation | ### Definition of done
Provide information about:
- [ ] Requirements & Install
- [ ] Platform being targeted
- [ ] Offered widgets
- [ ] Getting started: Explain how to use them.
| 1.0 | Document the repository - ### Definition of done
Provide information about:
- [ ] Requirements & Install
- [ ] Platform being targeted
- [ ] Offered widgets
- [ ] Getting started: Explain how to use them.
| non_main | document the repository definition of done provide information about requirements install platform being targeted offered widgets getting started explain how to use them | 0 |
175,605 | 6,552,506,098 | IssuesEvent | 2017-09-05 18:37:37 | strongloop/loopback-connector-swagger | https://api.github.com/repos/strongloop/loopback-connector-swagger | closed | Expose a subset of operations depending on the attached Model | feature needs-priority stale | A single Swagger document typically describes multiple resources (`tags` in swagger spec, I think they map to`api`s in swagger-client terminology).
Right now, our swagger connector always exposes all operations on every model that's attached to it. For example, if we have a LoopBack service exposing "Product" and "Category" models, then the client model created via our swagger connector will have methods like `Product_findById` and `Category_findById`.
What I would like to see instead, is to use the model name as a filter to decide which operations should be mixed into the model attached to a swagger-connector backed datasource. Using my example above, when I attach `Product` and `Category` models to the same datasource, each of these models will have a single `findById` method invoking the correct operation under the hood.
IMO, we can think about tags/apis as table names in database world. When a model is attached to a MySQL datasource, it does not receive methods for manipulating all tables available in the database, only methods to work with data in the specific table configured by the model.
Coupling the (client) model name and the swagger tag may not work in all cases. I think we should allow developers to customise the tag used to filter methods to inject to a given model, using the model name as the default.
Examples:
```js
// only methods for "Product" tag are mixed in
{
"name": "Product",
}
// only methods for "Product" tag are mixed in
{
"name": "ProductService",
"swagger": {
"tag": "Product" // perhaps use "api" as the config key?
}
}
// all methods are mixed in - backwards compatibility mode
{
"name": "ShoppingService",
"swagger": {
// I am not sure how to express this configuration,
// below are few examples to consider
"tag": "*",
// or
"includeAllTags": true
}
}
```
Note that @raymondfeng has already implemented a similar mechanism in https://github.com/strongloop/loopback-swagger/pull/81, it applies when generating a LoopBack server stub implementation.
cc @ritch | 1.0 | Expose a subset of operations depending on the attached Model - A single Swagger document typically describes multiple resources (`tags` in swagger spec, I think they map to`api`s in swagger-client terminology).
Right now, our swagger connector always exposes all operations on every model that's attached to it. For example, if we have a LoopBack service exposing "Product" and "Category" models, then the client model created via our swagger connector will have methods like `Product_findById` and `Category_findById`.
What I would like to see instead, is to use the model name as a filter to decide which operations should be mixed into the model attached to a swagger-connector backed datasource. Using my example above, when I attach `Product` and `Category` models to the same datasource, each of these models will have a single `findById` method invoking the correct operation under the hood.
IMO, we can think about tags/apis as table names in database world. When a model is attached to a MySQL datasource, it does not receive methods for manipulating all tables available in the database, only methods to work with data in the specific table configured by the model.
Coupling the (client) model name and the swagger tag may not work in all cases. I think we should allow developers to customise the tag used to filter methods to inject to a given model, using the model name as the default.
Examples:
```js
// only methods for "Product" tag are mixed in
{
"name": "Product",
}
// only methods for "Product" tag are mixed in
{
"name": "ProductService",
"swagger": {
"tag": "Product" // perhaps use "api" as the config key?
}
}
// all methods are mixed in - backwards compatibility mode
{
"name": "ShoppingService",
"swagger": {
// I am not sure how to express this configuration,
// below are few examples to consider
"tag": "*",
// or
"includeAllTags": true
}
}
```
Note that @raymondfeng has already implemented a similar mechanism in https://github.com/strongloop/loopback-swagger/pull/81, it applies when generating a LoopBack server stub implementation.
cc @ritch | non_main | expose a subset of operations depending on the attached model a single swagger document typically describes multiple resources tags in swagger spec i think they map to api s in swagger client terminology right now our swagger connector always exposes all operations on every model that s attached to it for example if we have a loopback service exposing product and category models then the client model created via our swagger connector will have methods like product findbyid and category findbyid what i would like to see instead is to use the model name as a filter to decide which operations should be mixed into the model attached to a swagger connector backed datasource using my example above when i attach product and category models to the same datasource each of these models will have a single findbyid method invoking the correct operation under the hood imo we can think about tags apis as table names in database world when a model is attached to a mysql datasource it does not receive methods for manipulating all tables available in the database only methods to work with data in the specific table configured by the model coupling the client model name and the swagger tag may not work in all cases i think we should allow developers to customise the tag used to filter methods to inject to a given model using the model name as the default examples js only methods for product tag are mixed in name product only methods for product tag are mixed in name productservice swagger tag product perhaps use api as the config key all methods are mixed in backwards compatibility mode name shoppingservice swagger i am not sure how to express this configuration below are few examples to consider tag or includealltags true note that raymondfeng has already implemented a similar mechanism in it applies when generating a loopback server stub implementation cc ritch | 0 |
2,085 | 7,094,110,236 | IssuesEvent | 2018-01-12 23:53:52 | AndrewJGregory/Procify | https://api.github.com/repos/AndrewJGregory/Procify | closed | Coupled props with inner collection | maintainability | In `InnerCollection`, there are props being passed down to whatever component is being rendered. This led to a nasty bug for the artist show page that is now fixed: props from the container were overrode with props from `InnerCollection`. Renaming the props in the container is a temporary fix, but ultimately the props need to be removed and not passed down to the component rendered inside of `InnerCollection` because the containers should do that job of filtering out the necessary state. | True | Coupled props with inner collection - In `InnerCollection`, there are props being passed down to whatever component is being rendered. This led to a nasty bug for the artist show page that is now fixed: props from the container were overrode with props from `InnerCollection`. Renaming the props in the container is a temporary fix, but ultimately the props need to be removed and not passed down to the component rendered inside of `InnerCollection` because the containers should do that job of filtering out the necessary state. | main | coupled props with inner collection in innercollection there are props being passed down to whatever component is being rendered this led to a nasty bug for the artist show page that is now fixed props from the container were overrode with props from innercollection renaming the props in the container is a temporary fix but ultimately the props need to be removed and not passed down to the component rendered inside of innercollection because the containers should do that job of filtering out the necessary state | 1 |
4,931 | 25,342,306,901 | IssuesEvent | 2022-11-18 23:12:06 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | Reusing logical ID of autogenerated HttpApi clobbers DefinitionBody | type/bug stage/bug-repro maintainer/need-response | **Description:**
Howdy, I've noticed there appears to be some undefined behavior when using re-using the auto-generated ID of the 'AWS::Serverless::HttpApi' resource - specifically the logical ID 'ServerlessHttpApi'. I have not checked whether the REST API is affected as well.
**Steps to reproduce the issue:**
1. Create a simple template that does not specify an 'AWS::Serverless::HttpApi', e.g. rely on auto generation by creating a serverless function type with a HttpApi event.
result: Everything works as expected
2. Add the HTTP API resource and specify an OpenAPI document - e.g.;
```
ServerlessHttpApi:
Type: AWS::Serverless::HttpApi
Properties:
DefinitionBody:
'Fn::Transform':
Name: AWS::Include
Parameters:
Location: api.yaml
```
**Observed Result**: The lambda integration fields are present, but the body of the OpenAPI document has not been carried over to the Body property of 'AWS::ApiGatewayV2::Api`
**expected**: the transformed final template will include the OpenAPI definition in the 'AWS::ApiGatewayV2::Api' body field, with the lambda integration fields merged into the template.
If a logical ID is specified other than 'ServerlessHttpApi', everything works as expected; DefinitionBody and Integration definitions are merged as expected. This only seems to occurs when using the same logical ID that the auto-generated API would use.
Happy to dig in and contribute an appropriate fix if this an issue - whether that be documentation or a code change.
| True | Reusing logical ID of autogenerated HttpApi clobbers DefinitionBody - **Description:**
Howdy, I've noticed there appears to be some undefined behavior when using re-using the auto-generated ID of the 'AWS::Serverless::HttpApi' resource - specifically the logical ID 'ServerlessHttpApi'. I have not checked whether the REST API is affected as well.
**Steps to reproduce the issue:**
1. Create a simple template that does not specify an 'AWS::Serverless::HttpApi', e.g. rely on auto generation by creating a serverless function type with a HttpApi event.
result: Everything works as expected
2. Add the HTTP API resource and specify an OpenAPI document - e.g.;
```
ServerlessHttpApi:
Type: AWS::Serverless::HttpApi
Properties:
DefinitionBody:
'Fn::Transform':
Name: AWS::Include
Parameters:
Location: api.yaml
```
**Observed Result**: The lambda integration fields are present, but the body of the OpenAPI document has not been carried over to the Body property of 'AWS::ApiGatewayV2::Api`
**expected**: the transformed final template will include the OpenAPI definition in the 'AWS::ApiGatewayV2::Api' body field, with the lambda integration fields merged into the template.
If a logical ID is specified other than 'ServerlessHttpApi', everything works as expected; DefinitionBody and Integration definitions are merged as expected. This only seems to occurs when using the same logical ID that the auto-generated API would use.
Happy to dig in and contribute an appropriate fix if this an issue - whether that be documentation or a code change.
| main | reusing logical id of autogenerated httpapi clobbers definitionbody description howdy i ve noticed there appears to be some undefined behavior when using re using the auto generated id of the aws serverless httpapi resource specifically the logical id serverlesshttpapi i have not checked whether the rest api is affected as well steps to reproduce the issue create a simple template that does not specify an aws serverless httpapi e g rely on auto generation by creating a serverless function type with a httpapi event result everything works as expected add the http api resource and specify an openapi document e g serverlesshttpapi type aws serverless httpapi properties definitionbody fn transform name aws include parameters location api yaml observed result the lambda integration fields are present but the body of the openapi document has not been carried over to the body property of aws api expected the transformed final template will include the openapi definition in the aws api body field with the lambda integration fields merged into the template if a logical id is specified other than serverlesshttpapi everything works as expected definitionbody and integration definitions are merged as expected this only seems to occurs when using the same logical id that the auto generated api would use happy to dig in and contribute an appropriate fix if this an issue whether that be documentation or a code change | 1 |
407,441 | 11,913,503,632 | IssuesEvent | 2020-03-31 12:07:25 | WebAhead/wajjbat-social | https://api.github.com/repos/WebAhead/wajjbat-social | closed | Add ingerdients and steps | T4h priority-2 | -Add ingredients when clicking the add button, in a list.
-Add how to prepare steps when clicking on the add button, in a list. | 1.0 | Add ingerdients and steps - -Add ingredients when clicking the add button, in a list.
-Add how to prepare steps when clicking on the add button, in a list. | non_main | add ingerdients and steps add ingredients when clicking the add button in a list add how to prepare steps when clicking on the add button in a list | 0 |
1,204 | 5,140,117,905 | IssuesEvent | 2017-01-12 03:06:31 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | closed | Bug report: Uninstalling gpgtools fails on trying to remove CLI binaries | awaiting maintainer feedback | #### Description of issue
I'm attempting to uninstall GPGTools, but the uninstallation process fails as it tries to cleanup CLI links that don't appear to exist. From what I can tell, it's attempting to first check that they exist, but something's going wrong that's hard for me to diagnose with my limited knowledge of how cask works.
<details><summary>Output of `brew cask uninstall gpgtools --verbose`</summary>
```
==> Running uninstall process for gpgtools; your password may be necessary
==> Removing launchctl service org.gpgtools.macgpg2.shutdown-gpg-agent
==> Removing launchctl service org.gpgtools.Libmacgpg.xpc
==> Removing launchctl service org.gpgtools.gpgmail.enable-bundles
==> Removing launchctl service org.gpgtools.gpgmail.user-uuid-patcher
==> Removing launchctl service org.gpgtools.gpgmail.uuid-patcher
==> Removing launchctl service org.gpgtools.macgpg2.fix
==> Removing launchctl service org.gpgtools.macgpg2.updater
==> Quitting application ID com.apple.mail
==> Quitting application ID org.gpgtools.gpgkeychainaccess
==> Quitting application ID org.gpgtools.gpgservices
==> Removing files from pkgutil Bill-of-Materials
==> Removing files: ["/Applications/GPG Keychain Access.app", "/Applications/GPG Keychain.app", "/usr/local/MacGPG2", "/Library/Services/GPGServices.service", "/Library/Mail/Bundles/GPGMail.mailbundle", "/Library/PreferencePanes/GPGPreferences.prefPane"]
Error: Command failed to execute!
==> Failed command:
/bin/bash -c [[ \"$(/usr/bin/readlink /usr/local/bin/gpg2)\" =~ MacGPG2 ]] && /bin/rm -- /usr/local/bin/gpg2
==> Standard Output of failed command:
==> Standard Error of failed command:
==> Exit status of failed command:
#<Process::Status: pid 78542 exit 1>
Error: Kernel.exit
```
</details>
<details><summary>Output of `brew doctor`</summary>
```
Your system is ready to brew.
```
</details>
<details><summary>Output of `brew cask doctor`</summary>
```
==> macOS Release:
10.12.2
==> Hardware Architecture:
intel-64
==> Ruby Version:
2.0.0-p648
==> Ruby Path:
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/bin/ruby
==> Homebrew Version:
Homebrew 1.1.7-11-gebf3d93
Homebrew/homebrew-core (git revision dfb6; last commit 2017-01-11)
==> Homebrew Executable Path:
/usr/local/bin/brew
==> Homebrew Cellar Path:
/usr/local/Cellar
==> Homebrew Repository Path:
/usr/local/Homebrew
==> Homebrew Origin:
https://github.com/Homebrew/brew
==> Homebrew-Cask Version:
Homebrew-Cask 1.1.7-11-gebf3d93
caskroom/homebrew-cask (git revision b1b9; last commit 2017-01-11)
==> Homebrew-Cask Install Location:
<NONE>
==> Homebrew-Cask Staging Location:
/usr/local/Caskroom
==> Homebrew-Cask Cached Downloads:
/Users/[USER_NAME]/Library/Caches/Homebrew/Cask
0 files, 0B
==> Homebrew-Cask Default Tap Path:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask
==> Homebrew-Cask Alternate Cask Taps:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts
==> Homebrew-Cask Default Tap Cask Count:
3511
==> Contents of $LOAD_PATH:
/usr/local/Homebrew/Library/Homebrew/cask/lib
/usr/local/Homebrew/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin16
/Library/Ruby/Site/2.0.0/universal-darwin16
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16
==> Contents of $RUBYLIB Environment Variable:
<NONE>
==> Contents of $RUBYOPT Environment Variable:
<NONE>
==> Contents of $RUBYPATH Environment Variable:
<NONE>
==> Contents of $RBENV_VERSION Environment Variable:
<NONE>
==> Contents of $CHRUBY_VERSION Environment Variable:
<NONE>
==> Contents of $GEM_HOME Environment Variable:
<NONE>
==> Contents of $GEM_PATH Environment Variable:
<NONE>
==> Contents of $BUNDLE_PATH Environment Variable:
<NONE>
==> Contents of $PATH Environment Variable:
PATH="/Users/[USER_NAME]/bin:/Users/[USER_NAME]/Library/Python/2.7/bin:/Users/[USER_NAME]/.gem/ruby/2.0.0/bin:/Users/[USER_NAME]/go/bin:/usr/local/bin:/usr/local/sbin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/MacGPG2/bin:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-bundle/cmd:/usr/local/Homebrew/Library/Homebrew/cask/cmd:/usr/local/Homebrew/Library/Homebrew/shims/scm"
==> Contents of $SHELL Environment Variable:
SHELL="/usr/local/bin/zsh"
==> Contents of Locale Environment Variables:
LANG="en_US.UTF-8"
```
</details>
| True | Bug report: Uninstalling gpgtools fails on trying to remove CLI binaries - #### Description of issue
I'm attempting to uninstall GPGTools, but the uninstallation process fails as it tries to cleanup CLI links that don't appear to exist. From what I can tell, it's attempting to first check that they exist, but something's going wrong that's hard for me to diagnose with my limited knowledge of how cask works.
<details><summary>Output of `brew cask uninstall gpgtools --verbose`</summary>
```
==> Running uninstall process for gpgtools; your password may be necessary
==> Removing launchctl service org.gpgtools.macgpg2.shutdown-gpg-agent
==> Removing launchctl service org.gpgtools.Libmacgpg.xpc
==> Removing launchctl service org.gpgtools.gpgmail.enable-bundles
==> Removing launchctl service org.gpgtools.gpgmail.user-uuid-patcher
==> Removing launchctl service org.gpgtools.gpgmail.uuid-patcher
==> Removing launchctl service org.gpgtools.macgpg2.fix
==> Removing launchctl service org.gpgtools.macgpg2.updater
==> Quitting application ID com.apple.mail
==> Quitting application ID org.gpgtools.gpgkeychainaccess
==> Quitting application ID org.gpgtools.gpgservices
==> Removing files from pkgutil Bill-of-Materials
==> Removing files: ["/Applications/GPG Keychain Access.app", "/Applications/GPG Keychain.app", "/usr/local/MacGPG2", "/Library/Services/GPGServices.service", "/Library/Mail/Bundles/GPGMail.mailbundle", "/Library/PreferencePanes/GPGPreferences.prefPane"]
Error: Command failed to execute!
==> Failed command:
/bin/bash -c [[ \"$(/usr/bin/readlink /usr/local/bin/gpg2)\" =~ MacGPG2 ]] && /bin/rm -- /usr/local/bin/gpg2
==> Standard Output of failed command:
==> Standard Error of failed command:
==> Exit status of failed command:
#<Process::Status: pid 78542 exit 1>
Error: Kernel.exit
```
</details>
<details><summary>Output of `brew doctor`</summary>
```
Your system is ready to brew.
```
</details>
<details><summary>Output of `brew cask doctor`</summary>
```
==> macOS Release:
10.12.2
==> Hardware Architecture:
intel-64
==> Ruby Version:
2.0.0-p648
==> Ruby Path:
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/bin/ruby
==> Homebrew Version:
Homebrew 1.1.7-11-gebf3d93
Homebrew/homebrew-core (git revision dfb6; last commit 2017-01-11)
==> Homebrew Executable Path:
/usr/local/bin/brew
==> Homebrew Cellar Path:
/usr/local/Cellar
==> Homebrew Repository Path:
/usr/local/Homebrew
==> Homebrew Origin:
https://github.com/Homebrew/brew
==> Homebrew-Cask Version:
Homebrew-Cask 1.1.7-11-gebf3d93
caskroom/homebrew-cask (git revision b1b9; last commit 2017-01-11)
==> Homebrew-Cask Install Location:
<NONE>
==> Homebrew-Cask Staging Location:
/usr/local/Caskroom
==> Homebrew-Cask Cached Downloads:
/Users/[USER_NAME]/Library/Caches/Homebrew/Cask
0 files, 0B
==> Homebrew-Cask Default Tap Path:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-cask
==> Homebrew-Cask Alternate Cask Taps:
/usr/local/Homebrew/Library/Taps/caskroom/homebrew-fonts
==> Homebrew-Cask Default Tap Cask Count:
3511
==> Contents of $LOAD_PATH:
/usr/local/Homebrew/Library/Homebrew/cask/lib
/usr/local/Homebrew/Library/Homebrew
/Library/Ruby/Site/2.0.0
/Library/Ruby/Site/2.0.0/x86_64-darwin16
/Library/Ruby/Site/2.0.0/universal-darwin16
/Library/Ruby/Site
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby/2.0.0/universal-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/vendor_ruby
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/x86_64-darwin16
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/universal-darwin16
==> Contents of $RUBYLIB Environment Variable:
<NONE>
==> Contents of $RUBYOPT Environment Variable:
<NONE>
==> Contents of $RUBYPATH Environment Variable:
<NONE>
==> Contents of $RBENV_VERSION Environment Variable:
<NONE>
==> Contents of $CHRUBY_VERSION Environment Variable:
<NONE>
==> Contents of $GEM_HOME Environment Variable:
<NONE>
==> Contents of $GEM_PATH Environment Variable:
<NONE>
==> Contents of $BUNDLE_PATH Environment Variable:
<NONE>
==> Contents of $PATH Environment Variable:
PATH="/Users/[USER_NAME]/bin:/Users/[USER_NAME]/Library/Python/2.7/bin:/Users/[USER_NAME]/.gem/ruby/2.0.0/bin:/Users/[USER_NAME]/go/bin:/usr/local/bin:/usr/local/sbin:/usr/local/bin:/usr/bin:/bin:/usr/sbin:/sbin:/opt/X11/bin:/usr/local/MacGPG2/bin:/usr/local/Homebrew/Library/Taps/homebrew/homebrew-bundle/cmd:/usr/local/Homebrew/Library/Homebrew/cask/cmd:/usr/local/Homebrew/Library/Homebrew/shims/scm"
==> Contents of $SHELL Environment Variable:
SHELL="/usr/local/bin/zsh"
==> Contents of Locale Environment Variables:
LANG="en_US.UTF-8"
```
</details>
| main | bug report uninstalling gpgtools fails on trying to remove cli binaries description of issue i m attempting to uninstall gpgtools but the uninstallation process fails as it tries to cleanup cli links that don t appear to exist from what i can tell it s attempting to first check that they exist but something s going wrong that s hard for me to diagnose with my limited knowledge of how cask works output of brew cask uninstall gpgtools verbose running uninstall process for gpgtools your password may be necessary removing launchctl service org gpgtools shutdown gpg agent removing launchctl service org gpgtools libmacgpg xpc removing launchctl service org gpgtools gpgmail enable bundles removing launchctl service org gpgtools gpgmail user uuid patcher removing launchctl service org gpgtools gpgmail uuid patcher removing launchctl service org gpgtools fix removing launchctl service org gpgtools updater quitting application id com apple mail quitting application id org gpgtools gpgkeychainaccess quitting application id org gpgtools gpgservices removing files from pkgutil bill of materials removing files error command failed to execute failed command bin bash c bin rm usr local bin standard output of failed command standard error of failed command exit status of failed command error kernel exit output of brew doctor your system is ready to brew output of brew cask doctor macos release hardware architecture intel ruby version ruby path system library frameworks ruby framework versions usr bin ruby homebrew version homebrew homebrew homebrew core git revision last commit homebrew executable path usr local bin brew homebrew cellar path usr local cellar homebrew repository path usr local homebrew homebrew origin homebrew cask version homebrew cask caskroom homebrew cask git revision last commit homebrew cask install location homebrew cask staging location usr local caskroom homebrew cask cached downloads users library caches homebrew cask files homebrew cask default tap path usr local homebrew library taps caskroom homebrew cask homebrew cask alternate cask taps usr local homebrew library taps caskroom homebrew fonts homebrew cask default tap cask count contents of load path usr local homebrew library homebrew cask lib usr local homebrew library homebrew library ruby site library ruby site library ruby site universal library ruby site system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby vendor ruby universal system library frameworks ruby framework versions usr lib ruby vendor ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby system library frameworks ruby framework versions usr lib ruby universal contents of rubylib environment variable contents of rubyopt environment variable contents of rubypath environment variable contents of rbenv version environment variable contents of chruby version environment variable contents of gem home environment variable contents of gem path environment variable contents of bundle path environment variable contents of path environment variable path users bin users library python bin users gem ruby bin users go bin usr local bin usr local sbin usr local bin usr bin bin usr sbin sbin opt bin usr local bin usr local homebrew library taps homebrew homebrew bundle cmd usr local homebrew library homebrew cask cmd usr local homebrew library homebrew shims scm contents of shell environment variable shell usr local bin zsh contents of locale environment variables lang en us utf | 1 |
5,416 | 27,192,573,611 | IssuesEvent | 2023-02-19 23:53:43 | Lissy93/dashy | https://api.github.com/repos/Lissy93/dashy | closed | [BUG] Opening first page to non-root URL causes 404 | 🐛 Bug 👤 Awaiting Maintainer Response | ### Environment
Self-Hosted (Docker)
### System
Docker 20.10.22
### Version
2.1.1
### Describe the problem
Reproducible directly from the docker image - without nginx or anything else in-between.
`docker run --rm -it -p 4322:80 lissy93/dashy`
With the default configuration ('history' routing), navigating directly to a non-root route will cause an error to be returned by the `node` server.
## Reproduction steps
1. Open `localhost:4322/workspace/`
2. **`Cannot GET /workspace/` is served** (see screenshot below)
I've seen issues that may be similar - but none that outline the issue clearly like this.
Generally the [advice is to statically host dashy](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#404-on-static-hosting) and/or [just use `hash` routing instead](https://github.com/Lissy93/dashy/issues/628#issuecomment-1116098016). I believe this can be fixed in `server.js` and am working on opening a PR.
### Additional info

### Please tick the boxes
- [X] You have explained the issue clearly, and included all relevant info
- [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy
- [X] You've checked that this [issue hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue)
- [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide 
- [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct) | True | [BUG] Opening first page to non-root URL causes 404 - ### Environment
Self-Hosted (Docker)
### System
Docker 20.10.22
### Version
2.1.1
### Describe the problem
Reproducible directly from the docker image - without nginx or anything else in-between.
`docker run --rm -it -p 4322:80 lissy93/dashy`
With the default configuration ('history' routing), navigating directly to a non-root route will cause an error to be returned by the `node` server.
## Reproduction steps
1. Open `localhost:4322/workspace/`
2. **`Cannot GET /workspace/` is served** (see screenshot below)
I've seen issues that may be similar - but none that outline the issue clearly like this.
Generally the [advice is to statically host dashy](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#404-on-static-hosting) and/or [just use `hash` routing instead](https://github.com/Lissy93/dashy/issues/628#issuecomment-1116098016). I believe this can be fixed in `server.js` and am working on opening a PR.
### Additional info

### Please tick the boxes
- [X] You have explained the issue clearly, and included all relevant info
- [X] You are using a [supported](https://github.com/Lissy93/dashy/blob/master/.github/SECURITY.md#supported-versions) version of Dashy
- [X] You've checked that this [issue hasn't already been raised](https://github.com/Lissy93/dashy/issues?q=is%3Aissue)
- [X] You've checked the [docs](https://github.com/Lissy93/dashy/tree/master/docs#readme) and [troubleshooting](https://github.com/Lissy93/dashy/blob/master/docs/troubleshooting.md#troubleshooting) guide 
- [X] You agree to the [code of conduct](https://github.com/Lissy93/dashy/blob/master/.github/CODE_OF_CONDUCT.md#contributor-covenant-code-of-conduct) | main | opening first page to non root url causes environment self hosted docker system docker version describe the problem reproducible directly from the docker image without nginx or anything else in between docker run rm it p dashy with the default configuration history routing navigating directly to a non root route will cause an error to be returned by the node server reproduction steps open localhost workspace cannot get workspace is served see screenshot below i ve seen issues that may be similar but none that outline the issue clearly like this generally the and or i believe this can be fixed in server js and am working on opening a pr additional info please tick the boxes you have explained the issue clearly and included all relevant info you are using a version of dashy you ve checked that this you ve checked the and guide you agree to the | 1 |
5,041 | 5,399,368,097 | IssuesEvent | 2017-02-27 19:16:51 | dotnet/cli | https://api.github.com/repos/dotnet/cli | closed | Ensure CLI version is SemVer compatible | blocking-release infrastructure RC3 known issue | Make sure CLI's release versions are "1.0.0" and not "1.0.0-001571", i.e., without the prerelease or the build tags.
As per the shared FX spec for SDK, the host is supposed to choose the latest `production` CLI when a CLI version cannot be inferred from a `global.json` probe. The keyword here is `production`. The host needs a way to know to pick a `production` release in a directory.
Example:
```
- dotnet\
- SDK\
+ 1.0.0\
+ 1.0.1\
+ 1.0.2-001573\
```
The host will choose `1.0.1` unless a `global.json` specified otherwise. The host chose `1.0.1` because it did not have `prerelease` or `build` tags and is the latest version.
| 1.0 | Ensure CLI version is SemVer compatible - Make sure CLI's release versions are "1.0.0" and not "1.0.0-001571", i.e., without the prerelease or the build tags.
As per the shared FX spec for SDK, the host is supposed to choose the latest `production` CLI when a CLI version cannot be inferred from a `global.json` probe. The keyword here is `production`. The host needs a way to know to pick a `production` release in a directory.
Example:
```
- dotnet\
- SDK\
+ 1.0.0\
+ 1.0.1\
+ 1.0.2-001573\
```
The host will choose `1.0.1` unless a `global.json` specified otherwise. The host chose `1.0.1` because it did not have `prerelease` or `build` tags and is the latest version.
| non_main | ensure cli version is semver compatible make sure cli s release versions are and not i e without the prerelease or the build tags as per the shared fx spec for sdk the host is supposed to choose the latest production cli when a cli version cannot be inferred from a global json probe the keyword here is production the host needs a way to know to pick a production release in a directory example dotnet sdk the host will choose unless a global json specified otherwise the host chose because it did not have prerelease or build tags and is the latest version | 0 |
32,570 | 6,827,507,453 | IssuesEvent | 2017-11-08 17:14:21 | primefaces/primefaces | https://api.github.com/repos/primefaces/primefaces | closed | p:textEditor in p:inplace sets property null if form submitted when editor not visible | defect | I suspect the value is still submitted despite it is not visible, and since it is not visible, the value is null.
## 1) Environment
- PrimeFaces version: 6.0.18, 6.1
- Does it work on the newest released PrimeFaces version? Version? No
- Does it work on the newest sources in GitHub? Probably not
- Application server + version: Wildfly 10.1
- Affected browsers: All
## 2) Expected behavior
If the form is submitted while p:textEditor not visible, the field should retain its value.
## 3) Actual behavior
If the form is submitted while p:textEditor not visible, the field is set to null.
## 4) Steps to reproduce
Press button in below sample
## 5) Sample XHTML
```
<h:form>
<p:inplace editor="true">
<f:facet name="output">
<h:outputText value="Some text to be able to toggle (other bug probably)" />
</f:facet>
<f:facet name="input">
<p:textEditor value="#{backingBean.property}"/>
</f:facet>
<p:ajax event="save" listener="#{backingBean.save}"/>
</p:inplace>
<p:commandButton value="Press this to lose property value" action="#{backingBean.someaction}" />
</h:form>
```
Workaround: add process="@this" to the p:commandButton.
## 6) Sample bean
```java
@Named
@ViewScoped
public class BackingBean implements Serializable {
private String property;
// getter, setter
public void save() {
}
public void someAction() {
// here property is null if p:textEdit was not visible when button pressed
}
}
```
| 1.0 | p:textEditor in p:inplace sets property null if form submitted when editor not visible - I suspect the value is still submitted despite it is not visible, and since it is not visible, the value is null.
## 1) Environment
- PrimeFaces version: 6.0.18, 6.1
- Does it work on the newest released PrimeFaces version? Version? No
- Does it work on the newest sources in GitHub? Probably not
- Application server + version: Wildfly 10.1
- Affected browsers: All
## 2) Expected behavior
If the form is submitted while p:textEditor not visible, the field should retain its value.
## 3) Actual behavior
If the form is submitted while p:textEditor not visible, the field is set to null.
## 4) Steps to reproduce
Press button in below sample
## 5) Sample XHTML
```
<h:form>
<p:inplace editor="true">
<f:facet name="output">
<h:outputText value="Some text to be able to toggle (other bug probably)" />
</f:facet>
<f:facet name="input">
<p:textEditor value="#{backingBean.property}"/>
</f:facet>
<p:ajax event="save" listener="#{backingBean.save}"/>
</p:inplace>
<p:commandButton value="Press this to lose property value" action="#{backingBean.someaction}" />
</h:form>
```
Workaround: add process="@this" to the p:commandButton.
## 6) Sample bean
```java
@Named
@ViewScoped
public class BackingBean implements Serializable {
private String property;
// getter, setter
public void save() {
}
public void someAction() {
// here property is null if p:textEdit was not visible when button pressed
}
}
```
| non_main | p texteditor in p inplace sets property null if form submitted when editor not visible i suspect the value is still submitted despite it is not visible and since it is not visible the value is null environment primefaces version does it work on the newest released primefaces version version no does it work on the newest sources in github probably not application server version wildfly affected browsers all expected behavior if the form is submitted while p texteditor not visible the field should retain its value actual behavior if the form is submitted while p texteditor not visible the field is set to null steps to reproduce press button in below sample sample xhtml workaround add process this to the p commandbutton sample bean java named viewscoped public class backingbean implements serializable private string property getter setter public void save public void someaction here property is null if p textedit was not visible when button pressed | 0 |
591 | 4,082,801,739 | IssuesEvent | 2016-05-31 14:04:45 | Particular/ServicePulse | https://api.github.com/repos/Particular/ServicePulse | closed | Expanded mobile nav menu won't auto-close after clicking to switch screens | Impact: S Size: S State: In Progress - Maintainer Prio Tag: Maintainer Prio Tag: Triaged Type: Bug | ### Symptom
When using navigation menu on mobile device it doesn't collapse after clicking outside of menu.
### Who's affected
Users who use Service Pulse on mobile devices.
### Steps to reproduce:
- switch to an SM breakpoint
- open nav menu
- click to switch another screen | True | Expanded mobile nav menu won't auto-close after clicking to switch screens - ### Symptom
When using navigation menu on mobile device it doesn't collapse after clicking outside of menu.
### Who's affected
Users who use Service Pulse on mobile devices.
### Steps to reproduce:
- switch to an SM breakpoint
- open nav menu
- click to switch another screen | main | expanded mobile nav menu won t auto close after clicking to switch screens symptom when using navigation menu on mobile device it doesn t collapse after clicking outside of menu who s affected users who use service pulse on mobile devices steps to reproduce switch to an sm breakpoint open nav menu click to switch another screen | 1 |
3,743 | 15,730,106,446 | IssuesEvent | 2021-03-29 15:33:09 | ipfs/go-ipfs | https://api.github.com/repos/ipfs/go-ipfs | closed | Use go-multiaddr-dns DefaultResolver instead of std lib Resolver | exp/beginner kind/enhancement need/maintainers-input | Would be nice to swap out the `net.LookupTXT` in the dns resolver for the `madns.DefaultResolver.LookupTXT` function.
This would allow one to just override the `madns.DefaultResolver` and it would work for both libp2p and IPFS. As it stands now there is no way to configure this resolver.
https://github.com/ipfs/go-ipfs/blob/f006fc164771d6ab2e2b2608992bc6e5d1f17a8f/namesys/dns.go#L28 | True | Use go-multiaddr-dns DefaultResolver instead of std lib Resolver - Would be nice to swap out the `net.LookupTXT` in the dns resolver for the `madns.DefaultResolver.LookupTXT` function.
This would allow one to just override the `madns.DefaultResolver` and it would work for both libp2p and IPFS. As it stands now there is no way to configure this resolver.
https://github.com/ipfs/go-ipfs/blob/f006fc164771d6ab2e2b2608992bc6e5d1f17a8f/namesys/dns.go#L28 | main | use go multiaddr dns defaultresolver instead of std lib resolver would be nice to swap out the net lookuptxt in the dns resolver for the madns defaultresolver lookuptxt function this would allow one to just override the madns defaultresolver and it would work for both and ipfs as it stands now there is no way to configure this resolver | 1 |
20,103 | 2,622,185,362 | IssuesEvent | 2015-03-04 00:20:36 | byzhang/signal-collect | https://api.github.com/repos/byzhang/signal-collect | closed | Rework aggregation operations | 1.0 auto-migrated Priority-Medium Type-Enhancement | ```
An aggregation operation should be expressed by an instance of an aggregation
class instead of 3 parameters to an aggregation function.
```
Original issue reported on code.google.com by `philip.stutz` on 12 Oct 2011 at 12:29 | 1.0 | Rework aggregation operations - ```
An aggregation operation should be expressed by an instance of an aggregation
class instead of 3 parameters to an aggregation function.
```
Original issue reported on code.google.com by `philip.stutz` on 12 Oct 2011 at 12:29 | non_main | rework aggregation operations an aggregation operation should be expressed by an instance of an aggregation class instead of parameters to an aggregation function original issue reported on code google com by philip stutz on oct at | 0 |
54,411 | 30,113,866,694 | IssuesEvent | 2023-06-30 09:52:10 | modin-project/modin | https://api.github.com/repos/modin-project/modin | closed | remove `__make_init_labels_args` explicit calls that materialize axes | Performance 🚀 | By doing this, we will improve asynchronous code execution. | True | remove `__make_init_labels_args` explicit calls that materialize axes - By doing this, we will improve asynchronous code execution. | non_main | remove make init labels args explicit calls that materialize axes by doing this we will improve asynchronous code execution | 0 |
3,607 | 14,570,158,039 | IssuesEvent | 2020-12-17 14:02:38 | jupyterlab/jupyterlab-hdf5 | https://api.github.com/repos/jupyterlab/jupyterlab-hdf5 | closed | Modernize signature of `get` method in backend | tag:backend type:maintainence | I keep forgetting about this, but
https://github.com/jupyterlab/jupyterlab-hdf5/blob/8ea9fc03360a322cbac54db82222d21f0604f0b2/jupyterlab_hdf/baseHandler.py#L108-L109
needs to become
```python
from tornado import web
@web.authenticated
async def get(self):
```
This will:
- update the (deprecated) tornado coroutine to the native python async support
- add https authentication to the endpoint, which is just a Good Idea
@loichuder If you're currently writing unittests for the enpoints, you'll probably want to make sure this change has been made before you get in too deep (it may subtlely change some behavior). You can just add this change to unittest PR, or I can take care of it if you prefer | True | Modernize signature of `get` method in backend - I keep forgetting about this, but
https://github.com/jupyterlab/jupyterlab-hdf5/blob/8ea9fc03360a322cbac54db82222d21f0604f0b2/jupyterlab_hdf/baseHandler.py#L108-L109
needs to become
```python
from tornado import web
@web.authenticated
async def get(self):
```
This will:
- update the (deprecated) tornado coroutine to the native python async support
- add https authentication to the endpoint, which is just a Good Idea
@loichuder If you're currently writing unittests for the enpoints, you'll probably want to make sure this change has been made before you get in too deep (it may subtlely change some behavior). You can just add this change to unittest PR, or I can take care of it if you prefer | main | modernize signature of get method in backend i keep forgetting about this but needs to become python from tornado import web web authenticated async def get self this will update the deprecated tornado coroutine to the native python async support add https authentication to the endpoint which is just a good idea loichuder if you re currently writing unittests for the enpoints you ll probably want to make sure this change has been made before you get in too deep it may subtlely change some behavior you can just add this change to unittest pr or i can take care of it if you prefer | 1 |
12,514 | 9,669,616,364 | IssuesEvent | 2019-05-21 17:49:20 | Azure/azure-powershell | https://api.github.com/repos/Azure/azure-powershell | closed | Add support for configuring SSL settings of Api Management instance using Powershell cmdlets | API Management Service Attention | We're trying to automate our deployment of Api Management using the Api Management git repositories. However, Api Management currently doesn't support configuration for SSL settings of an instance using Azure PowerShell cmdlets. We would like to have that support so that we can manage APIM completely through Azure Powershell instead of having to change things in the Portal. | 1.0 | Add support for configuring SSL settings of Api Management instance using Powershell cmdlets - We're trying to automate our deployment of Api Management using the Api Management git repositories. However, Api Management currently doesn't support configuration for SSL settings of an instance using Azure PowerShell cmdlets. We would like to have that support so that we can manage APIM completely through Azure Powershell instead of having to change things in the Portal. | non_main | add support for configuring ssl settings of api management instance using powershell cmdlets we re trying to automate our deployment of api management using the api management git repositories however api management currently doesn t support configuration for ssl settings of an instance using azure powershell cmdlets we would like to have that support so that we can manage apim completely through azure powershell instead of having to change things in the portal | 0 |
2,243 | 7,893,828,248 | IssuesEvent | 2018-06-28 19:22:27 | arcticicestudio/nord-visual-studio-code | https://api.github.com/repos/arcticicestudio/nord-visual-studio-code | closed | Replace Gulp with npm scripts | context-workflow scope-maintainability type-improvement | [gulp.js][gulp] is a awesome tool to create simple or complex build workflows with streams, but there are also some drawbacks:
* Gulp plugins often get out of date and don’t support new features from the underlying library
* The project depends on the Gulp plugin author and not the library that is being wrapped
* Documentation for Gulp plugins are typically not very robust or comprehensive
* Debugging a Gulp plugin can be frustrating (is the problem in the plugin or in the underlying library?)
* If a plugin doesn't exist already for Gulp you need to write it yourself
* Errors that occur inside a Gulp wrapper aren't always handled very well
All currently implemented Gulp tasks can be migrated to npm scripts by directly using the CLI API of the wrapped library.
[gulp]: https://gulpjs.com | True | Replace Gulp with npm scripts - [gulp.js][gulp] is a awesome tool to create simple or complex build workflows with streams, but there are also some drawbacks:
* Gulp plugins often get out of date and don’t support new features from the underlying library
* The project depends on the Gulp plugin author and not the library that is being wrapped
* Documentation for Gulp plugins are typically not very robust or comprehensive
* Debugging a Gulp plugin can be frustrating (is the problem in the plugin or in the underlying library?)
* If a plugin doesn't exist already for Gulp you need to write it yourself
* Errors that occur inside a Gulp wrapper aren't always handled very well
All currently implemented Gulp tasks can be migrated to npm scripts by directly using the CLI API of the wrapped library.
[gulp]: https://gulpjs.com | main | replace gulp with npm scripts is a awesome tool to create simple or complex build workflows with streams but there are also some drawbacks gulp plugins often get out of date and don’t support new features from the underlying library the project depends on the gulp plugin author and not the library that is being wrapped documentation for gulp plugins are typically not very robust or comprehensive debugging a gulp plugin can be frustrating is the problem in the plugin or in the underlying library if a plugin doesn t exist already for gulp you need to write it yourself errors that occur inside a gulp wrapper aren t always handled very well all currently implemented gulp tasks can be migrated to npm scripts by directly using the cli api of the wrapped library | 1 |
3,140 | 12,055,389,746 | IssuesEvent | 2020-04-15 12:55:04 | arcticicestudio/igloo | https://api.github.com/repos/arcticicestudio/igloo | opened | “Atom“ snowblock decommission | scope-maintainability scope-quality snowblock-atom type-task | Related to #248
---
[Atom][] is not used anymore since at least February 23 2019 and was replaced by [Visual Studio Code][vscode] in #179. Therefore the `atom` _snowblock_ will be removed while the data is still available through the _Git_ repository history/logs.
[atom]: https://atom.io
[vscode]: https://code.visualstudio.com
| True | “Atom“ snowblock decommission - Related to #248
---
[Atom][] is not used anymore since at least February 23 2019 and was replaced by [Visual Studio Code][vscode] in #179. Therefore the `atom` _snowblock_ will be removed while the data is still available through the _Git_ repository history/logs.
[atom]: https://atom.io
[vscode]: https://code.visualstudio.com
| main | “atom“ snowblock decommission related to is not used anymore since at least february and was replaced by in therefore the atom snowblock will be removed while the data is still available through the git repository history logs | 1 |
4,693 | 24,211,820,065 | IssuesEvent | 2022-09-26 00:18:52 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Irregular test failure: /mob/living/simple_animal/mouse was unable to be GC'd | Maintainability/Hinders improvements | ## Reproduction:
```
## REF SEARCH Beginning search for references to a /mob/living/simple_animal/mouse.
## REF SEARCH Finished searching globals
## REF SEARCH Finished searching native globals
## REF SEARCH Finished searching atoms
## REF SEARCH Found /mob/living/simple_animal/mouse [0x300011d] in list Datums -> /datum/spatial_grid_cell [0x21007b6d] -> hearing_contents (list).
## REF SEARCH Finished searching datums
## REF SEARCH Finished searching clients
## REF SEARCH Completed search for references to a /mob/living/simple_animal/mouse.
## TESTING: GC: -- [0x300011d] | /mob/living/simple_animal/mouse was unable to be GC'd --
Error: /mob/living/simple_animal/mouse hard deleted 1 times out of a total del count of 26
FAIL: /datum/unit_test/create_and_destroy 518.4s
REASON #1: /mob/living/simple_animal/mouse hard deleted 1 times out of a total del count of 26 at code/modules/unit_tests/create_and_destroy.dm:170
```
https://github.com/tgstation/tgstation/actions/runs/3120786625/jobs/5061636026 | True | Irregular test failure: /mob/living/simple_animal/mouse was unable to be GC'd - ## Reproduction:
```
## REF SEARCH Beginning search for references to a /mob/living/simple_animal/mouse.
## REF SEARCH Finished searching globals
## REF SEARCH Finished searching native globals
## REF SEARCH Finished searching atoms
## REF SEARCH Found /mob/living/simple_animal/mouse [0x300011d] in list Datums -> /datum/spatial_grid_cell [0x21007b6d] -> hearing_contents (list).
## REF SEARCH Finished searching datums
## REF SEARCH Finished searching clients
## REF SEARCH Completed search for references to a /mob/living/simple_animal/mouse.
## TESTING: GC: -- [0x300011d] | /mob/living/simple_animal/mouse was unable to be GC'd --
Error: /mob/living/simple_animal/mouse hard deleted 1 times out of a total del count of 26
FAIL: /datum/unit_test/create_and_destroy 518.4s
REASON #1: /mob/living/simple_animal/mouse hard deleted 1 times out of a total del count of 26 at code/modules/unit_tests/create_and_destroy.dm:170
```
https://github.com/tgstation/tgstation/actions/runs/3120786625/jobs/5061636026 | main | irregular test failure mob living simple animal mouse was unable to be gc d reproduction ref search beginning search for references to a mob living simple animal mouse ref search finished searching globals ref search finished searching native globals ref search finished searching atoms ref search found mob living simple animal mouse in list datums datum spatial grid cell hearing contents list ref search finished searching datums ref search finished searching clients ref search completed search for references to a mob living simple animal mouse testing gc mob living simple animal mouse was unable to be gc d error mob living simple animal mouse hard deleted times out of a total del count of fail datum unit test create and destroy reason mob living simple animal mouse hard deleted times out of a total del count of at code modules unit tests create and destroy dm | 1 |
293,459 | 25,294,439,509 | IssuesEvent | 2022-11-17 04:52:57 | Hekili/hekili | https://api.github.com/repos/Hekili/hekili | closed | Arms: Bladestorm recommended while on CD with Hurricane talented | accepted recommendation in testing retail | ### Before You Begin
- [X] I confirm that I have downloaded the latest version of the addon.
- [X] I am not playing on a private server.
- [X] I checked for an [existing, open ticket](https://github.com/Hekili/hekili/labels/recommendation) for this issue and was not able to find one.
- [X] I edited the title of this issue (above) so that it describes the issue I am reporting.
- [X] I am reporting an issue with the default priority included with the specialization (imported or edited priorities are not supported).
### WoW Version
Retail (Dragonflight)
### Describe the Issue
Hekili addon seems to be recommending Bladestorm (icon) while its on cooldown. It is also not picking up my keybind for bladestorm
### How to Reproduce
1. Go into aoe situation
2. Use bladestorm
3. Hekili still recommends bladestorm
### Talent Loadout
BcEAAAAAAAAAAAAAAAAAAAAAAAQKRikkEkDEalEJBAAAABEJKkQUEJQIwBoIRIAQIAAAAAAAAAtEAIE
### Snapshot
https://pastebin.com/BEk6tt0j
### Raidbots Sim Report (Link)
_No response_
### Additional Information
Hope the snapshot caught it but i think it did
### Contact Information
Twizz#2915 | 1.0 | Arms: Bladestorm recommended while on CD with Hurricane talented - ### Before You Begin
- [X] I confirm that I have downloaded the latest version of the addon.
- [X] I am not playing on a private server.
- [X] I checked for an [existing, open ticket](https://github.com/Hekili/hekili/labels/recommendation) for this issue and was not able to find one.
- [X] I edited the title of this issue (above) so that it describes the issue I am reporting.
- [X] I am reporting an issue with the default priority included with the specialization (imported or edited priorities are not supported).
### WoW Version
Retail (Dragonflight)
### Describe the Issue
Hekili addon seems to be recommending Bladestorm (icon) while its on cooldown. It is also not picking up my keybind for bladestorm
### How to Reproduce
1. Go into aoe situation
2. Use bladestorm
3. Hekili still recommends bladestorm
### Talent Loadout
BcEAAAAAAAAAAAAAAAAAAAAAAAQKRikkEkDEalEJBAAAABEJKkQUEJQIwBoIRIAQIAAAAAAAAAtEAIE
### Snapshot
https://pastebin.com/BEk6tt0j
### Raidbots Sim Report (Link)
_No response_
### Additional Information
Hope the snapshot caught it but i think it did
### Contact Information
Twizz#2915 | non_main | arms bladestorm recommended while on cd with hurricane talented before you begin i confirm that i have downloaded the latest version of the addon i am not playing on a private server i checked for an for this issue and was not able to find one i edited the title of this issue above so that it describes the issue i am reporting i am reporting an issue with the default priority included with the specialization imported or edited priorities are not supported wow version retail dragonflight describe the issue hekili addon seems to be recommending bladestorm icon while its on cooldown it is also not picking up my keybind for bladestorm how to reproduce go into aoe situation use bladestorm hekili still recommends bladestorm talent loadout bceaaaaaaaaaaaaaaaaaaaaaaaqkrikkekdealejbaaaabejkkquejqiwboiriaqiaaaaaaaaateaie snapshot raidbots sim report link no response additional information hope the snapshot caught it but i think it did contact information twizz | 0 |
2,073 | 7,024,832,173 | IssuesEvent | 2017-12-23 00:17:34 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Floor code needs to be refactored | Maintainability/Hinders improvements Not a bug | The recent refactoring of map files to use different floor types is causing predictable bugs and a new refactoring of floor code is necessary to fix things once and for all.
| True | Floor code needs to be refactored - The recent refactoring of map files to use different floor types is causing predictable bugs and a new refactoring of floor code is necessary to fix things once and for all.
| main | floor code needs to be refactored the recent refactoring of map files to use different floor types is causing predictable bugs and a new refactoring of floor code is necessary to fix things once and for all | 1 |
1,606 | 6,572,392,383 | IssuesEvent | 2017-09-11 01:58:10 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | consul_kv: support KV transactions | affects_2.1 feature_idea waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
consul_kv
##### ANSIBLE VERSION
```
ansible 2.1.2.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
The `consul_kv` module currently only supports modifying individual KV values. However, starting with consul 0.7, there is an endpoint that allows packaging a set of KV updates (up to 64, apparently) into a single atomic transaction: https://www.consul.io/docs/agent/http/kv.html#txn
It would be ideal if this module supported this transactional behavior where it is available, perhaps by overriding the meaning of `with_items` similar to how package manager modules like yum/apt do.
##### STEPS TO REPRODUCE
N/A
##### EXPECTED RESULTS
N/A
##### ACTUAL RESULTS
N/A
| True | consul_kv: support KV transactions - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
consul_kv
##### ANSIBLE VERSION
```
ansible 2.1.2.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
The `consul_kv` module currently only supports modifying individual KV values. However, starting with consul 0.7, there is an endpoint that allows packaging a set of KV updates (up to 64, apparently) into a single atomic transaction: https://www.consul.io/docs/agent/http/kv.html#txn
It would be ideal if this module supported this transactional behavior where it is available, perhaps by overriding the meaning of `with_items` similar to how package manager modules like yum/apt do.
##### STEPS TO REPRODUCE
N/A
##### EXPECTED RESULTS
N/A
##### ACTUAL RESULTS
N/A
| main | consul kv support kv transactions issue type feature idea component name consul kv ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration n a os environment n a summary the consul kv module currently only supports modifying individual kv values however starting with consul there is an endpoint that allows packaging a set of kv updates up to apparently into a single atomic transaction it would be ideal if this module supported this transactional behavior where it is available perhaps by overriding the meaning of with items similar to how package manager modules like yum apt do steps to reproduce n a expected results n a actual results n a | 1 |
525,543 | 15,255,934,346 | IssuesEvent | 2021-02-20 18:05:00 | kubernetes/minikube | https://api.github.com/repos/kubernetes/minikube | closed | [go.mod] build error: package go.opentelemetry.io/otel | kind/bug priority/important-soon | /kind bug
When I built minikube binary, the errors of `go mod` happend.
```
$ make
go build -tags "" -ldflags="-X k8s.io/minikube/pkg/version.version=v1.17.1 -X k8s.io/minikube/pkg/version.isoVersion=v1.17.0 -X k8s.io/minikube/pkg/version.gitCommitID="051caa1d68b2812095efadc536edb76ec4c6e010-dirty" -X k8s.io/minikube/pkg/version.storageProvisionerVersion=v4" -o out/minikube k8s.io/minikube/cmd/minikube
go: finding module for package go.opentelemetry.io/otel/api/trace
go: finding module for package go.opentelemetry.io/otel/api/global
../../../pkg/mod/github.com/!google!cloud!platform/opentelemetry-operations-go/exporter/trace@v0.13.0/cloudtrace.go:28:2: module go.opentelemetry.io/otel@latest found (v0.17.0), but does not contain package go.opentelemetry.io/otel/api/global
../../../pkg/mod/github.com/!google!cloud!platform/opentelemetry-operations-go/exporter/trace@v0.13.0/cloudtrace.go:29:2: module go.opentelemetry.io/otel@latest found (v0.17.0), but does not contain package go.opentelemetry.io/otel/api/trace
make: *** [out/minikube] Error 1
```
This is caused by the missmatch of module version.
https://github.com/kubernetes/minikube/blob/051caa1d68b2812095efadc536edb76ec4c6e010/pkg/trace/gcp.go#L24-L30
`go.opentelemetry.io/otel/api/trace` and `go.opentelemetry.io/otel/api/global` are old packages at v0.13.0.
In v0.17.0, these package don't exist. And these absent causes the errors of `go mod`.
We need to update other modules to take consistency. | 1.0 | [go.mod] build error: package go.opentelemetry.io/otel - /kind bug
When I built minikube binary, the errors of `go mod` happend.
```
$ make
go build -tags "" -ldflags="-X k8s.io/minikube/pkg/version.version=v1.17.1 -X k8s.io/minikube/pkg/version.isoVersion=v1.17.0 -X k8s.io/minikube/pkg/version.gitCommitID="051caa1d68b2812095efadc536edb76ec4c6e010-dirty" -X k8s.io/minikube/pkg/version.storageProvisionerVersion=v4" -o out/minikube k8s.io/minikube/cmd/minikube
go: finding module for package go.opentelemetry.io/otel/api/trace
go: finding module for package go.opentelemetry.io/otel/api/global
../../../pkg/mod/github.com/!google!cloud!platform/opentelemetry-operations-go/exporter/trace@v0.13.0/cloudtrace.go:28:2: module go.opentelemetry.io/otel@latest found (v0.17.0), but does not contain package go.opentelemetry.io/otel/api/global
../../../pkg/mod/github.com/!google!cloud!platform/opentelemetry-operations-go/exporter/trace@v0.13.0/cloudtrace.go:29:2: module go.opentelemetry.io/otel@latest found (v0.17.0), but does not contain package go.opentelemetry.io/otel/api/trace
make: *** [out/minikube] Error 1
```
This is caused by the missmatch of module version.
https://github.com/kubernetes/minikube/blob/051caa1d68b2812095efadc536edb76ec4c6e010/pkg/trace/gcp.go#L24-L30
`go.opentelemetry.io/otel/api/trace` and `go.opentelemetry.io/otel/api/global` are old packages at v0.13.0.
In v0.17.0, these package don't exist. And these absent causes the errors of `go mod`.
We need to update other modules to take consistency. | non_main | build error package go opentelemetry io otel kind bug when i built minikube binary the errors of go mod happend make go build tags ldflags x io minikube pkg version version x io minikube pkg version isoversion x io minikube pkg version gitcommitid dirty x io minikube pkg version storageprovisionerversion o out minikube io minikube cmd minikube go finding module for package go opentelemetry io otel api trace go finding module for package go opentelemetry io otel api global pkg mod github com google cloud platform opentelemetry operations go exporter trace cloudtrace go module go opentelemetry io otel latest found but does not contain package go opentelemetry io otel api global pkg mod github com google cloud platform opentelemetry operations go exporter trace cloudtrace go module go opentelemetry io otel latest found but does not contain package go opentelemetry io otel api trace make error this is caused by the missmatch of module version go opentelemetry io otel api trace and go opentelemetry io otel api global are old packages at in these package don t exist and these absent causes the errors of go mod we need to update other modules to take consistency | 0 |
2,451 | 8,639,870,594 | IssuesEvent | 2018-11-23 22:14:45 | F5OEO/rpitx | https://api.github.com/repos/F5OEO/rpitx | closed | Cant boot with dt-blob.bin | V1 related (not maintained) | I have a Raspberry pi 3 v1.2 and as part of the install steps it copies dt-blob.bin to the /boot folder. (This is for the gpio assignments I believe) On the next reboot, the raspberry pi just stops booting and blinks 7 times (kernel.img not found) upon removing dt-blob.bin the pi immediately begins to boot successfully. What's wrong? I tried compiling my own from the dt-blob.dts in the src/ folder but to no avail. What's happening? Raspbian jessie, usb boot. All updates installed. | True | Cant boot with dt-blob.bin - I have a Raspberry pi 3 v1.2 and as part of the install steps it copies dt-blob.bin to the /boot folder. (This is for the gpio assignments I believe) On the next reboot, the raspberry pi just stops booting and blinks 7 times (kernel.img not found) upon removing dt-blob.bin the pi immediately begins to boot successfully. What's wrong? I tried compiling my own from the dt-blob.dts in the src/ folder but to no avail. What's happening? Raspbian jessie, usb boot. All updates installed. | main | cant boot with dt blob bin i have a raspberry pi and as part of the install steps it copies dt blob bin to the boot folder this is for the gpio assignments i believe on the next reboot the raspberry pi just stops booting and blinks times kernel img not found upon removing dt blob bin the pi immediately begins to boot successfully what s wrong i tried compiling my own from the dt blob dts in the src folder but to no avail what s happening raspbian jessie usb boot all updates installed | 1 |
5,786 | 30,649,232,364 | IssuesEvent | 2023-07-25 07:45:15 | jupyter-naas/awesome-notebooks | https://api.github.com/repos/jupyter-naas/awesome-notebooks | closed | Python - Convert string boolean to boolean | templates maintainer | This notebook will show how to convert a string boolean to a boolean type in Python. It is usefull for data cleaning and data manipulation.
| True | Python - Convert string boolean to boolean - This notebook will show how to convert a string boolean to a boolean type in Python. It is usefull for data cleaning and data manipulation.
| main | python convert string boolean to boolean this notebook will show how to convert a string boolean to a boolean type in python it is usefull for data cleaning and data manipulation | 1 |
293,558 | 22,062,309,178 | IssuesEvent | 2022-05-30 19:49:58 | mapstruct/mapstruct | https://api.github.com/repos/mapstruct/mapstruct | closed | Error in the readme ( what is mapstract paragraph ) | documentation | To create a mapping between two types, declare a mapper class like this:
```
@Mapper
public interface CarMapper {
CarMapper INSTANCE = Mappers.getMapper( CarMapper.class );
@Mapping(target = "seatCount", source = "numberOfSeats")
CarDto carToCarDto(Car car);
}
```
I guess it should be "declare a mapper interface like this: " | 1.0 | Error in the readme ( what is mapstract paragraph ) - To create a mapping between two types, declare a mapper class like this:
```
@Mapper
public interface CarMapper {
CarMapper INSTANCE = Mappers.getMapper( CarMapper.class );
@Mapping(target = "seatCount", source = "numberOfSeats")
CarDto carToCarDto(Car car);
}
```
I guess it should be "declare a mapper interface like this: " | non_main | error in the readme what is mapstract paragraph to create a mapping between two types declare a mapper class like this mapper public interface carmapper carmapper instance mappers getmapper carmapper class mapping target seatcount source numberofseats cardto cartocardto car car i guess it should be declare a mapper interface like this | 0 |
207,925 | 15,859,941,525 | IssuesEvent | 2021-04-08 08:35:59 | rovergulf/ngx-slice-kit | https://api.github.com/repos/rovergulf/ngx-slice-kit | opened | update coverage results to handle codecov reports | testing | add reporter:
https://istanbul.js.org/docs/advanced/alternative-reporters/#lcovonly
```js
{ type: 'lcovonly' },
``` | 1.0 | update coverage results to handle codecov reports - add reporter:
https://istanbul.js.org/docs/advanced/alternative-reporters/#lcovonly
```js
{ type: 'lcovonly' },
``` | non_main | update coverage results to handle codecov reports add reporter js type lcovonly | 0 |
4,777 | 24,606,987,955 | IssuesEvent | 2022-10-14 17:12:03 | duckduckgo/zeroclickinfo-longtail | https://api.github.com/repos/duckduckgo/zeroclickinfo-longtail | closed | MetroLyrics: Wrong result | Bug Maintainer Input Requested Type: Relevancy | Searching for "On the plane blackfield" brings up lyrics for a Ed Sheeran song.
---
IA Page: http://duck.co/ia/view/lyrics
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @jdorweiler
| True | MetroLyrics: Wrong result - Searching for "On the plane blackfield" brings up lyrics for a Ed Sheeran song.
---
IA Page: http://duck.co/ia/view/lyrics
[Maintainer](http://docs.duckduckhack.com/maintaining/guidelines.html): @jdorweiler
| main | metrolyrics wrong result searching for on the plane blackfield brings up lyrics for a ed sheeran song ia page jdorweiler | 1 |
5,339 | 26,939,605,987 | IssuesEvent | 2023-02-08 00:24:39 | mozilla/foundation.mozilla.org | https://api.github.com/repos/mozilla/foundation.mozilla.org | closed | Update Spacing Units on Tailwind Config with updated design tokens | engineering frontend maintain | Should be done sooner rather then later!
## Current implementation
| Tailwind Unit | Bootstrap Unit | Size (rem) |
| --------------- | ---------------- | ----------- |
| 1 | 1 | 0.25rem |
| 2 | 2 | 0.50rem |
| 3 | | 0.75rem |
| 4 | 3 | 1rem |
| 5 | 4 | 1.5rem |
| 6 | | 2rem |
| 7 | 5 | 3rem |
## Current Design Tokens
https://www.figma.com/file/JT9Dvdyc96AqQ3HJlnA9Qq/WIP-Mozilla-Foundation-Design-Style-Library
## Converting Guide
| Tailwind Unit | Old Size | New Size |
| --------------- | ---------------- | ----------- |
| 0 | 0 | 0 |
| px | 1px | 1px |
| rem | | 1rem |
| 1 | 0.25rem | 2px |
| 2 | 0.5rem | 4px |
| 3 | 0.75rem | 6px |
| 4 | 1rem | 8px |
| 5 | 1.5rem | 10px |
| 6 | 2rem | 12px |
| 7 | 3rem | 14px |
| 8 | | 16px |
| 9 | | 18px |
| 10 | | 20px |
| 12 | | 24px |
| 14 | | 28px |
| 16 | | 32px |
| 18 | | 36px |
| 20 | | 40px |
| 22 | | 44px |
| 24 | | 48px |
| 28 | | 56px |
| 32 | | 64px |
| 40 | | 80px |
| 48 | | 96px |
| 56 | | 112px |
| 64 | | 128px |
| 72 | | 144px |
| 80 | | 160px |
| 88 | | 176px |
| 96 | | 192px |
| 104 | | 208px |
| 112 | | 224px |
| 120 | | 240px |
| 128 | | 256px |
| 144 | | 288px |
| 160 | | 320px |
| 192 | | 384px |
Dev Tasks:
- [ ] Replace width/height tailwind classes
- [ ] Replace margin/padding tailwind classes
- [ ] Replace arbitrary tailwind classes that include new spacing units Ex. `tw-m-[20px]`
- [ ] Inspect both `.html` and `.scss` files | True | Update Spacing Units on Tailwind Config with updated design tokens - Should be done sooner rather then later!
## Current implementation
| Tailwind Unit | Bootstrap Unit | Size (rem) |
| --------------- | ---------------- | ----------- |
| 1 | 1 | 0.25rem |
| 2 | 2 | 0.50rem |
| 3 | | 0.75rem |
| 4 | 3 | 1rem |
| 5 | 4 | 1.5rem |
| 6 | | 2rem |
| 7 | 5 | 3rem |
## Current Design Tokens
https://www.figma.com/file/JT9Dvdyc96AqQ3HJlnA9Qq/WIP-Mozilla-Foundation-Design-Style-Library
## Converting Guide
| Tailwind Unit | Old Size | New Size |
| --------------- | ---------------- | ----------- |
| 0 | 0 | 0 |
| px | 1px | 1px |
| rem | | 1rem |
| 1 | 0.25rem | 2px |
| 2 | 0.5rem | 4px |
| 3 | 0.75rem | 6px |
| 4 | 1rem | 8px |
| 5 | 1.5rem | 10px |
| 6 | 2rem | 12px |
| 7 | 3rem | 14px |
| 8 | | 16px |
| 9 | | 18px |
| 10 | | 20px |
| 12 | | 24px |
| 14 | | 28px |
| 16 | | 32px |
| 18 | | 36px |
| 20 | | 40px |
| 22 | | 44px |
| 24 | | 48px |
| 28 | | 56px |
| 32 | | 64px |
| 40 | | 80px |
| 48 | | 96px |
| 56 | | 112px |
| 64 | | 128px |
| 72 | | 144px |
| 80 | | 160px |
| 88 | | 176px |
| 96 | | 192px |
| 104 | | 208px |
| 112 | | 224px |
| 120 | | 240px |
| 128 | | 256px |
| 144 | | 288px |
| 160 | | 320px |
| 192 | | 384px |
Dev Tasks:
- [ ] Replace width/height tailwind classes
- [ ] Replace margin/padding tailwind classes
- [ ] Replace arbitrary tailwind classes that include new spacing units Ex. `tw-m-[20px]`
- [ ] Inspect both `.html` and `.scss` files | main | update spacing units on tailwind config with updated design tokens should be done sooner rather then later current implementation tailwind unit bootstrap unit size rem current design tokens converting guide tailwind unit old size new size px rem dev tasks replace width height tailwind classes replace margin padding tailwind classes replace arbitrary tailwind classes that include new spacing units ex tw m inspect both html and scss files | 1 |
5,154 | 26,256,643,610 | IssuesEvent | 2023-01-06 01:44:23 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | Bug: sam build fails on unmodified Hello World Typescript example | stage/needs-investigation maintainer/need-followup | <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
sam build fails on unmodified example
### Steps to reproduce:
`$ sam init` with the following choices:
Template: `1 - AWS Quick Start Templates`
Quick Start template: `1 - Hello World Example`
Runtime: `12 - nodejs18.x`
Package type: `1 - Zip`
Template `2 - Hello World Example TypeScript`
X-Ray tracing, CloudWatch: `no`
```
$ cd sam-app
$ sam build
```
Note: `npm install; npm run test` succeeds without error
### Observed result:
```
$ sam build --debug
2022-12-29 15:52:47,119 | Config file location: /tmp/sam-app/samconfig.toml
2022-12-29 15:52:47,120 | Config file '/tmp/sam-app/samconfig.toml' does not exist
2022-12-29 15:52:47,125 | Using SAM Template at /tmp/sam-app/template.yaml
2022-12-29 15:52:47,174 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2022-12-29 15:52:47,175 | Using config file: samconfig.toml, config environment: default
2022-12-29 15:52:47,175 | Expand command line arguments to:
2022-12-29 15:52:47,175 | --template_file=/tmp/sam-app/template.yaml --build_dir=.aws-sam/build --cache_dir=.aws-sam/cache
2022-12-29 15:52:47,319 | 'build' command is called
2022-12-29 15:52:47,325 | No Parameters detected in the template
2022-12-29 15:52:47,338 | There is no customer defined id or cdk path defined for resource HelloWorldFunction, so we will use the resource logical id as the resource id
2022-12-29 15:52:47,338 | There is no customer defined id or cdk path defined for resource ServerlessRestApi, so we will use the resource logical id as the resource id
2022-12-29 15:52:47,339 | 0 stacks found in the template
2022-12-29 15:52:47,339 | No Parameters detected in the template
2022-12-29 15:52:47,352 | There is no customer defined id or cdk path defined for resource HelloWorldFunction, so we will use the resource logical id as the resource id
2022-12-29 15:52:47,352 | There is no customer defined id or cdk path defined for resource ServerlessRestApi, so we will use the resource logical id as the resource id
2022-12-29 15:52:47,353 | 2 resources found in the stack
2022-12-29 15:52:47,353 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello-world/'
2022-12-29 15:52:47,353 | --base-dir is not presented, adjusting uri hello-world/ relative to /tmp/sam-app/template.yaml
2022-12-29 15:52:47,359 | 2 resources found in the stack
2022-12-29 15:52:47,359 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello-world/'
2022-12-29 15:52:47,359 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello-world/'
2022-12-29 15:52:47,359 | --base-dir is not presented, adjusting uri hello-world/ relative to /tmp/sam-app/template.yaml
2022-12-29 15:52:47,360 | Instantiating build definitions
2022-12-29 15:52:47,360 | No previous build graph found, generating new one
2022-12-29 15:52:47,360 | Unique function build definition found, adding as new (Function Build Definition: BuildDefinition(nodejs18.x, /tmp/sam-app/hello-world, Zip, , 1cf07abd-4df9-47d3-a970-711f7864556a, {'BuildMethod': 'esbuild', 'BuildProperties': {'Minify': True, 'Target': 'es2020', 'Sourcemap': True, 'EntryPoints': ['app.ts']}}, {}, x86_64, []), Function: Function(function_id='HelloWorldFunction', name='HelloWorldFunction', functionname='HelloWorldFunction', runtime='nodejs18.x', memory=None, timeout=3, handler='app.lambdaHandler', imageuri=None, packagetype='Zip', imageconfig=None, codeuri='/tmp/sam-app/hello-world', environment=None, rolearn=None, layers=[], events={'HelloWorld': {'Type': 'Api', 'Properties': {'Path': '/hello', 'Method': 'get', 'RestApiId': 'ServerlessRestApi'}}}, metadata={'BuildMethod': 'esbuild', 'BuildProperties': {'Minify': True, 'Target': 'es2020', 'Sourcemap': True, 'EntryPoints': ['app.ts']}, 'SamResourceId': 'HelloWorldFunction'}, inlinecode=None, codesign_config_arn=None, architectures=['x86_64'], function_url_config=None, stack_path=''))
2022-12-29 15:52:47,361 | Building codeuri: /tmp/sam-app/hello-world runtime: nodejs18.x metadata: {'BuildMethod': 'esbuild', 'BuildProperties': {'Minify': True, 'Target': 'es2020', 'Sourcemap': True, 'EntryPoints': ['app.ts']}} architecture: x86_64 functions: HelloWorldFunction
2022-12-29 15:52:47,361 | Building to following folder /tmp/sam-app/.aws-sam/build/HelloWorldFunction
2022-12-29 15:52:47,362 | Loading workflow module 'aws_lambda_builders.workflows'
2022-12-29 15:52:47,364 | Registering workflow 'PythonPipBuilder' with capability 'Capability(language='python', dependency_manager='pip', application_framework=None)'
2022-12-29 15:52:47,364 | Registering workflow 'NodejsNpmBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)'
2022-12-29 15:52:47,365 | Registering workflow 'RubyBundlerBuilder' with capability 'Capability(language='ruby', dependency_manager='bundler', application_framework=None)'
2022-12-29 15:52:47,366 | Registering workflow 'GoModulesBuilder' with capability 'Capability(language='go', dependency_manager='modules', application_framework=None)'
2022-12-29 15:52:47,367 | Registering workflow 'JavaGradleWorkflow' with capability 'Capability(language='java', dependency_manager='gradle', application_framework=None)'
2022-12-29 15:52:47,368 | Registering workflow 'JavaMavenWorkflow' with capability 'Capability(language='java', dependency_manager='maven', application_framework=None)'
2022-12-29 15:52:47,369 | Registering workflow 'DotnetCliPackageBuilder' with capability 'Capability(language='dotnet', dependency_manager='cli-package', application_framework=None)'
2022-12-29 15:52:47,370 | Registering workflow 'CustomMakeBuilder' with capability 'Capability(language='provided', dependency_manager=None, application_framework=None)'
2022-12-29 15:52:47,371 | Registering workflow 'NodejsNpmEsbuildBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm-esbuild', application_framework=None)'
2022-12-29 15:52:47,371 | Found workflow 'NodejsNpmEsbuildBuilder' to support capabilities 'Capability(language='nodejs', dependency_manager='npm-esbuild', application_framework=None)'
2022-12-29 15:52:47,372 | executing NPM: ['npm', 'bin']
2022-12-29 15:52:47,698 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2022-12-29 15:52:47,699 | Unable to find Click Context for getting session_id.
2022-12-29 15:52:47,699 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': 'dfd84683-08ab-4079-8242-6cc50f9d5044', 'installationId': 'd0401db6-4ad8-461c-aeb4-07cce42e8d89', 'sessionId': 'f8cbf88d-6bbf-4f58-a3bd-2b258a7fe252', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.7.10', 'samcliVersion': '1.67.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam build', 'metricSpecificAttributes': {'projectType': 'CFN', 'gitOrigin': None, 'projectName': '593ab2ca51e925b9f6c2f258bc55ed5926cf6d2c78239a685d65907e4ec7edd3', 'initialCommit': None}, 'duration': 523, 'exitReason': 'NpmExecutionError', 'exitCode': 255}}]}
2022-12-29 15:52:47,700 | Sending Telemetry: {'metrics': [{'events': {'requestId': 'b159d1fd-f009-4e56-bb4a-724ba258ac90', 'installationId': 'd0401db6-4ad8-461c-aeb4-07cce42e8d89', 'sessionId': 'f8cbf88d-6bbf-4f58-a3bd-2b258a7fe252', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.7.10', 'samcliVersion': '1.67.0', 'metricSpecificAttributes': {'events': [{'event_name': 'BuildWorkflowUsed', 'event_value': 'nodejs-npm-esbuild', 'thread_id': 139922124578816, 'time_stamp': '2022-12-29 20:52:47.361'}]}}}]}
2022-12-29 15:52:48,485 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
Traceback (most recent call last):
File "samcli/__main__.py", line 12, in <module>
2022-12-29 15:52:48,486 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
File "click/core.py", line 1130, in __call__
File "click/core.py", line 1055, in main
File "click/core.py", line 1657, in invoke
File "click/core.py", line 1404, in invoke
File "click/core.py", line 760, in invoke
File "click/decorators.py", line 84, in new_func
File "click/core.py", line 760, in invoke
File "samcli/lib/telemetry/metric.py", line 194, in wrapped
File "samcli/lib/telemetry/metric.py", line 140, in wrapped
File "samcli/lib/utils/version_checker.py", line 41, in wrapped
File "samcli/cli/main.py", line 86, in wrapper
File "samcli/commands/build/command.py", line 206, in cli
File "samcli/commands/build/command.py", line 276, in do_cli
File "samcli/commands/build/build_context.py", line 256, in run
File "samcli/lib/build/app_builder.py", line 210, in build
File "samcli/lib/build/build_strategy.py", line 80, in build
File "samcli/lib/build/build_strategy.py", line 90, in _build_functions
File "samcli/lib/build/build_strategy.py", line 174, in build_single_function_definition
File "samcli/lib/build/app_builder.py", line 705, in _build_function
File "samcli/lib/build/app_builder.py", line 859, in _build_function_in_process
File "aws_lambda_builders/builder.py", line 161, in build
File "aws_lambda_builders/workflows/nodejs_npm_esbuild/workflow.py", line 56, in __init__
File "aws_lambda_builders/workflows/nodejs_npm_esbuild/workflow.py", line 218, in _get_esbuild_subprocess
File "aws_lambda_builders/workflows/nodejs_npm/npm.py", line 85, in run
aws_lambda_builders.workflows.nodejs_npm.npm.NpmExecutionError: NPM Failed:
[444677] Failed to execute script '__main__' due to unhandled exception!
```
### Expected result:
I would expect "Build Succeeded" much like what happens when I use the Javascript `Hello World Example` like so:
```$ sam build
Building codeuri: /tmp/sam-app-js/hello-world runtime: nodejs18.x metadata: {} architecture: x86_64 functions: HelloWorldFunction
Running NodejsNpmBuilder:NpmPack
Running NodejsNpmBuilder:CopyNpmrcAndLockfile
Running NodejsNpmBuilder:CopySource
Running NodejsNpmBuilder:NpmInstall
Running NodejsNpmBuilder:CleanUpNpmrc
Running NodejsNpmBuilder:LockfileCleanUp
Build Succeeded
Built Artifacts : .aws-sam/build
Built Template : .aws-sam/build/template.yaml
```
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
```
$ cat /etc/issue
Ubuntu 22.04.1 LTS
$ aws --version
aws-cli/2.9.11 Python/3.9.11 Linux/5.15.0-56-generic exe/x86_64.ubuntu.22 prompt/off
$ sam --version
SAM CLI, version 1.67.0
region = us-east-1
$ node --version
v18.12.1
$ npm --version
9.1.3
``` | True | Bug: sam build fails on unmodified Hello World Typescript example - <!-- Make sure we don't have an existing Issue that reports the bug you are seeing (both open and closed).
If you do find an existing Issue, re-open or add a comment to that Issue instead of creating a new one. -->
### Description:
sam build fails on unmodified example
### Steps to reproduce:
`$ sam init` with the following choices:
Template: `1 - AWS Quick Start Templates`
Quick Start template: `1 - Hello World Example`
Runtime: `12 - nodejs18.x`
Package type: `1 - Zip`
Template `2 - Hello World Example TypeScript`
X-Ray tracing, CloudWatch: `no`
```
$ cd sam-app
$ sam build
```
Note: `npm install; npm run test` succeeds without error
### Observed result:
```
$ sam build --debug
2022-12-29 15:52:47,119 | Config file location: /tmp/sam-app/samconfig.toml
2022-12-29 15:52:47,120 | Config file '/tmp/sam-app/samconfig.toml' does not exist
2022-12-29 15:52:47,125 | Using SAM Template at /tmp/sam-app/template.yaml
2022-12-29 15:52:47,174 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2022-12-29 15:52:47,175 | Using config file: samconfig.toml, config environment: default
2022-12-29 15:52:47,175 | Expand command line arguments to:
2022-12-29 15:52:47,175 | --template_file=/tmp/sam-app/template.yaml --build_dir=.aws-sam/build --cache_dir=.aws-sam/cache
2022-12-29 15:52:47,319 | 'build' command is called
2022-12-29 15:52:47,325 | No Parameters detected in the template
2022-12-29 15:52:47,338 | There is no customer defined id or cdk path defined for resource HelloWorldFunction, so we will use the resource logical id as the resource id
2022-12-29 15:52:47,338 | There is no customer defined id or cdk path defined for resource ServerlessRestApi, so we will use the resource logical id as the resource id
2022-12-29 15:52:47,339 | 0 stacks found in the template
2022-12-29 15:52:47,339 | No Parameters detected in the template
2022-12-29 15:52:47,352 | There is no customer defined id or cdk path defined for resource HelloWorldFunction, so we will use the resource logical id as the resource id
2022-12-29 15:52:47,352 | There is no customer defined id or cdk path defined for resource ServerlessRestApi, so we will use the resource logical id as the resource id
2022-12-29 15:52:47,353 | 2 resources found in the stack
2022-12-29 15:52:47,353 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello-world/'
2022-12-29 15:52:47,353 | --base-dir is not presented, adjusting uri hello-world/ relative to /tmp/sam-app/template.yaml
2022-12-29 15:52:47,359 | 2 resources found in the stack
2022-12-29 15:52:47,359 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello-world/'
2022-12-29 15:52:47,359 | Found Serverless function with name='HelloWorldFunction' and CodeUri='hello-world/'
2022-12-29 15:52:47,359 | --base-dir is not presented, adjusting uri hello-world/ relative to /tmp/sam-app/template.yaml
2022-12-29 15:52:47,360 | Instantiating build definitions
2022-12-29 15:52:47,360 | No previous build graph found, generating new one
2022-12-29 15:52:47,360 | Unique function build definition found, adding as new (Function Build Definition: BuildDefinition(nodejs18.x, /tmp/sam-app/hello-world, Zip, , 1cf07abd-4df9-47d3-a970-711f7864556a, {'BuildMethod': 'esbuild', 'BuildProperties': {'Minify': True, 'Target': 'es2020', 'Sourcemap': True, 'EntryPoints': ['app.ts']}}, {}, x86_64, []), Function: Function(function_id='HelloWorldFunction', name='HelloWorldFunction', functionname='HelloWorldFunction', runtime='nodejs18.x', memory=None, timeout=3, handler='app.lambdaHandler', imageuri=None, packagetype='Zip', imageconfig=None, codeuri='/tmp/sam-app/hello-world', environment=None, rolearn=None, layers=[], events={'HelloWorld': {'Type': 'Api', 'Properties': {'Path': '/hello', 'Method': 'get', 'RestApiId': 'ServerlessRestApi'}}}, metadata={'BuildMethod': 'esbuild', 'BuildProperties': {'Minify': True, 'Target': 'es2020', 'Sourcemap': True, 'EntryPoints': ['app.ts']}, 'SamResourceId': 'HelloWorldFunction'}, inlinecode=None, codesign_config_arn=None, architectures=['x86_64'], function_url_config=None, stack_path=''))
2022-12-29 15:52:47,361 | Building codeuri: /tmp/sam-app/hello-world runtime: nodejs18.x metadata: {'BuildMethod': 'esbuild', 'BuildProperties': {'Minify': True, 'Target': 'es2020', 'Sourcemap': True, 'EntryPoints': ['app.ts']}} architecture: x86_64 functions: HelloWorldFunction
2022-12-29 15:52:47,361 | Building to following folder /tmp/sam-app/.aws-sam/build/HelloWorldFunction
2022-12-29 15:52:47,362 | Loading workflow module 'aws_lambda_builders.workflows'
2022-12-29 15:52:47,364 | Registering workflow 'PythonPipBuilder' with capability 'Capability(language='python', dependency_manager='pip', application_framework=None)'
2022-12-29 15:52:47,364 | Registering workflow 'NodejsNpmBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm', application_framework=None)'
2022-12-29 15:52:47,365 | Registering workflow 'RubyBundlerBuilder' with capability 'Capability(language='ruby', dependency_manager='bundler', application_framework=None)'
2022-12-29 15:52:47,366 | Registering workflow 'GoModulesBuilder' with capability 'Capability(language='go', dependency_manager='modules', application_framework=None)'
2022-12-29 15:52:47,367 | Registering workflow 'JavaGradleWorkflow' with capability 'Capability(language='java', dependency_manager='gradle', application_framework=None)'
2022-12-29 15:52:47,368 | Registering workflow 'JavaMavenWorkflow' with capability 'Capability(language='java', dependency_manager='maven', application_framework=None)'
2022-12-29 15:52:47,369 | Registering workflow 'DotnetCliPackageBuilder' with capability 'Capability(language='dotnet', dependency_manager='cli-package', application_framework=None)'
2022-12-29 15:52:47,370 | Registering workflow 'CustomMakeBuilder' with capability 'Capability(language='provided', dependency_manager=None, application_framework=None)'
2022-12-29 15:52:47,371 | Registering workflow 'NodejsNpmEsbuildBuilder' with capability 'Capability(language='nodejs', dependency_manager='npm-esbuild', application_framework=None)'
2022-12-29 15:52:47,371 | Found workflow 'NodejsNpmEsbuildBuilder' to support capabilities 'Capability(language='nodejs', dependency_manager='npm-esbuild', application_framework=None)'
2022-12-29 15:52:47,372 | executing NPM: ['npm', 'bin']
2022-12-29 15:52:47,698 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2022-12-29 15:52:47,699 | Unable to find Click Context for getting session_id.
2022-12-29 15:52:47,699 | Sending Telemetry: {'metrics': [{'commandRun': {'requestId': 'dfd84683-08ab-4079-8242-6cc50f9d5044', 'installationId': 'd0401db6-4ad8-461c-aeb4-07cce42e8d89', 'sessionId': 'f8cbf88d-6bbf-4f58-a3bd-2b258a7fe252', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.7.10', 'samcliVersion': '1.67.0', 'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam build', 'metricSpecificAttributes': {'projectType': 'CFN', 'gitOrigin': None, 'projectName': '593ab2ca51e925b9f6c2f258bc55ed5926cf6d2c78239a685d65907e4ec7edd3', 'initialCommit': None}, 'duration': 523, 'exitReason': 'NpmExecutionError', 'exitCode': 255}}]}
2022-12-29 15:52:47,700 | Sending Telemetry: {'metrics': [{'events': {'requestId': 'b159d1fd-f009-4e56-bb4a-724ba258ac90', 'installationId': 'd0401db6-4ad8-461c-aeb4-07cce42e8d89', 'sessionId': 'f8cbf88d-6bbf-4f58-a3bd-2b258a7fe252', 'executionEnvironment': 'CLI', 'ci': False, 'pyversion': '3.7.10', 'samcliVersion': '1.67.0', 'metricSpecificAttributes': {'events': [{'event_name': 'BuildWorkflowUsed', 'event_value': 'nodejs-npm-esbuild', 'thread_id': 139922124578816, 'time_stamp': '2022-12-29 20:52:47.361'}]}}}]}
2022-12-29 15:52:48,485 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
Traceback (most recent call last):
File "samcli/__main__.py", line 12, in <module>
2022-12-29 15:52:48,486 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
File "click/core.py", line 1130, in __call__
File "click/core.py", line 1055, in main
File "click/core.py", line 1657, in invoke
File "click/core.py", line 1404, in invoke
File "click/core.py", line 760, in invoke
File "click/decorators.py", line 84, in new_func
File "click/core.py", line 760, in invoke
File "samcli/lib/telemetry/metric.py", line 194, in wrapped
File "samcli/lib/telemetry/metric.py", line 140, in wrapped
File "samcli/lib/utils/version_checker.py", line 41, in wrapped
File "samcli/cli/main.py", line 86, in wrapper
File "samcli/commands/build/command.py", line 206, in cli
File "samcli/commands/build/command.py", line 276, in do_cli
File "samcli/commands/build/build_context.py", line 256, in run
File "samcli/lib/build/app_builder.py", line 210, in build
File "samcli/lib/build/build_strategy.py", line 80, in build
File "samcli/lib/build/build_strategy.py", line 90, in _build_functions
File "samcli/lib/build/build_strategy.py", line 174, in build_single_function_definition
File "samcli/lib/build/app_builder.py", line 705, in _build_function
File "samcli/lib/build/app_builder.py", line 859, in _build_function_in_process
File "aws_lambda_builders/builder.py", line 161, in build
File "aws_lambda_builders/workflows/nodejs_npm_esbuild/workflow.py", line 56, in __init__
File "aws_lambda_builders/workflows/nodejs_npm_esbuild/workflow.py", line 218, in _get_esbuild_subprocess
File "aws_lambda_builders/workflows/nodejs_npm/npm.py", line 85, in run
aws_lambda_builders.workflows.nodejs_npm.npm.NpmExecutionError: NPM Failed:
[444677] Failed to execute script '__main__' due to unhandled exception!
```
### Expected result:
I would expect "Build Succeeded" much like what happens when I use the Javascript `Hello World Example` like so:
```$ sam build
Building codeuri: /tmp/sam-app-js/hello-world runtime: nodejs18.x metadata: {} architecture: x86_64 functions: HelloWorldFunction
Running NodejsNpmBuilder:NpmPack
Running NodejsNpmBuilder:CopyNpmrcAndLockfile
Running NodejsNpmBuilder:CopySource
Running NodejsNpmBuilder:NpmInstall
Running NodejsNpmBuilder:CleanUpNpmrc
Running NodejsNpmBuilder:LockfileCleanUp
Build Succeeded
Built Artifacts : .aws-sam/build
Built Template : .aws-sam/build/template.yaml
```
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
```
$ cat /etc/issue
Ubuntu 22.04.1 LTS
$ aws --version
aws-cli/2.9.11 Python/3.9.11 Linux/5.15.0-56-generic exe/x86_64.ubuntu.22 prompt/off
$ sam --version
SAM CLI, version 1.67.0
region = us-east-1
$ node --version
v18.12.1
$ npm --version
9.1.3
``` | main | bug sam build fails on unmodified hello world typescript example make sure we don t have an existing issue that reports the bug you are seeing both open and closed if you do find an existing issue re open or add a comment to that issue instead of creating a new one description sam build fails on unmodified example steps to reproduce sam init with the following choices template aws quick start templates quick start template hello world example runtime x package type zip template hello world example typescript x ray tracing cloudwatch no cd sam app sam build note npm install npm run test succeeds without error observed result sam build debug config file location tmp sam app samconfig toml config file tmp sam app samconfig toml does not exist using sam template at tmp sam app template yaml telemetry endpoint configured to be using config file samconfig toml config environment default expand command line arguments to template file tmp sam app template yaml build dir aws sam build cache dir aws sam cache build command is called no parameters detected in the template there is no customer defined id or cdk path defined for resource helloworldfunction so we will use the resource logical id as the resource id there is no customer defined id or cdk path defined for resource serverlessrestapi so we will use the resource logical id as the resource id stacks found in the template no parameters detected in the template there is no customer defined id or cdk path defined for resource helloworldfunction so we will use the resource logical id as the resource id there is no customer defined id or cdk path defined for resource serverlessrestapi so we will use the resource logical id as the resource id resources found in the stack found serverless function with name helloworldfunction and codeuri hello world base dir is not presented adjusting uri hello world relative to tmp sam app template yaml resources found in the stack found serverless function with name helloworldfunction and codeuri hello world found serverless function with name helloworldfunction and codeuri hello world base dir is not presented adjusting uri hello world relative to tmp sam app template yaml instantiating build definitions no previous build graph found generating new one unique function build definition found adding as new function build definition builddefinition x tmp sam app hello world zip buildmethod esbuild buildproperties minify true target sourcemap true entrypoints function function function id helloworldfunction name helloworldfunction functionname helloworldfunction runtime x memory none timeout handler app lambdahandler imageuri none packagetype zip imageconfig none codeuri tmp sam app hello world environment none rolearn none layers events helloworld type api properties path hello method get restapiid serverlessrestapi metadata buildmethod esbuild buildproperties minify true target sourcemap true entrypoints samresourceid helloworldfunction inlinecode none codesign config arn none architectures function url config none stack path building codeuri tmp sam app hello world runtime x metadata buildmethod esbuild buildproperties minify true target sourcemap true entrypoints architecture functions helloworldfunction building to following folder tmp sam app aws sam build helloworldfunction loading workflow module aws lambda builders workflows registering workflow pythonpipbuilder with capability capability language python dependency manager pip application framework none registering workflow nodejsnpmbuilder with capability capability language nodejs dependency manager npm application framework none registering workflow rubybundlerbuilder with capability capability language ruby dependency manager bundler application framework none registering workflow gomodulesbuilder with capability capability language go dependency manager modules application framework none registering workflow javagradleworkflow with capability capability language java dependency manager gradle application framework none registering workflow javamavenworkflow with capability capability language java dependency manager maven application framework none registering workflow dotnetclipackagebuilder with capability capability language dotnet dependency manager cli package application framework none registering workflow custommakebuilder with capability capability language provided dependency manager none application framework none registering workflow nodejsnpmesbuildbuilder with capability capability language nodejs dependency manager npm esbuild application framework none found workflow nodejsnpmesbuildbuilder to support capabilities capability language nodejs dependency manager npm esbuild application framework none executing npm telemetry endpoint configured to be unable to find click context for getting session id sending telemetry metrics sending telemetry metrics httpsconnectionpool host aws serverless tools telemetry us west amazonaws com port read timed out read timeout traceback most recent call last file samcli main py line in httpsconnectionpool host aws serverless tools telemetry us west amazonaws com port read timed out read timeout file click core py line in call file click core py line in main file click core py line in invoke file click core py line in invoke file click core py line in invoke file click decorators py line in new func file click core py line in invoke file samcli lib telemetry metric py line in wrapped file samcli lib telemetry metric py line in wrapped file samcli lib utils version checker py line in wrapped file samcli cli main py line in wrapper file samcli commands build command py line in cli file samcli commands build command py line in do cli file samcli commands build build context py line in run file samcli lib build app builder py line in build file samcli lib build build strategy py line in build file samcli lib build build strategy py line in build functions file samcli lib build build strategy py line in build single function definition file samcli lib build app builder py line in build function file samcli lib build app builder py line in build function in process file aws lambda builders builder py line in build file aws lambda builders workflows nodejs npm esbuild workflow py line in init file aws lambda builders workflows nodejs npm esbuild workflow py line in get esbuild subprocess file aws lambda builders workflows nodejs npm npm py line in run aws lambda builders workflows nodejs npm npm npmexecutionerror npm failed failed to execute script main due to unhandled exception expected result i would expect build succeeded much like what happens when i use the javascript hello world example like so sam build building codeuri tmp sam app js hello world runtime x metadata architecture functions helloworldfunction running nodejsnpmbuilder npmpack running nodejsnpmbuilder copynpmrcandlockfile running nodejsnpmbuilder copysource running nodejsnpmbuilder npminstall running nodejsnpmbuilder cleanupnpmrc running nodejsnpmbuilder lockfilecleanup build succeeded built artifacts aws sam build built template aws sam build template yaml additional environment details ex windows mac amazon linux etc cat etc issue ubuntu lts aws version aws cli python linux generic exe ubuntu prompt off sam version sam cli version region us east node version npm version | 1 |
140,998 | 5,427,664,206 | IssuesEvent | 2017-03-03 14:00:35 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | Strided indexing behavior | bug high priority | Currently it seems `x[i:j:k]` with k>1 doesn't work when:
(1) j>len(x) or
(2) (j-i)%k != 0
```
a = list(range(10))
x = torch.Tensor(a)
x[::2] # works
x[::3] # fails
x[::4] # fails
x[::5] # works
# In fact, right now there's no pytorch equivalent to
a[::4] # returns [0,4,8]
x[:8:4] # works, but only returns [0,4]
```
This likely has to do with narrow not allowing condition 1:
https://github.com/pytorch/pytorch/blob/67f94557ff26428ac911d8c08c7f9b619a41950e/torch/csrc/generic/Tensor.cpp#L477 | 1.0 | Strided indexing behavior - Currently it seems `x[i:j:k]` with k>1 doesn't work when:
(1) j>len(x) or
(2) (j-i)%k != 0
```
a = list(range(10))
x = torch.Tensor(a)
x[::2] # works
x[::3] # fails
x[::4] # fails
x[::5] # works
# In fact, right now there's no pytorch equivalent to
a[::4] # returns [0,4,8]
x[:8:4] # works, but only returns [0,4]
```
This likely has to do with narrow not allowing condition 1:
https://github.com/pytorch/pytorch/blob/67f94557ff26428ac911d8c08c7f9b619a41950e/torch/csrc/generic/Tensor.cpp#L477 | non_main | strided indexing behavior currently it seems x with k doesn t work when j len x or j i k a list range x torch tensor a x works x fails x fails x works in fact right now there s no pytorch equivalent to a returns x works but only returns this likely has to do with narrow not allowing condition | 0 |
4,336 | 21,786,655,769 | IssuesEvent | 2022-05-14 08:30:01 | Numble-challenge-Team/client | https://api.github.com/repos/Numble-challenge-Team/client | closed | 프론트엔드 개발 환경 설정 | maintain | ### ISSUE
- Type: chore
- Page: -
### 변경 사항
- lint, prettier fix 명령어 추가
- tsconfig include 값에 next.config.js 추가
- 프로젝트 컬러 시스템 수정
- material ui sheet 적용 | True | 프론트엔드 개발 환경 설정 - ### ISSUE
- Type: chore
- Page: -
### 변경 사항
- lint, prettier fix 명령어 추가
- tsconfig include 값에 next.config.js 추가
- 프로젝트 컬러 시스템 수정
- material ui sheet 적용 | main | 프론트엔드 개발 환경 설정 issue type chore page 변경 사항 lint prettier fix 명령어 추가 tsconfig include 값에 next config js 추가 프로젝트 컬러 시스템 수정 material ui sheet 적용 | 1 |
35,324 | 9,561,926,362 | IssuesEvent | 2019-05-04 03:34:12 | openshiftio/openshift.io | https://api.github.com/repos/openshiftio/openshift.io | closed | Build: Add contract testing | team/build-cd type/user-story | Fabric8-auth [has started](https://github.com/fabric8-services/fabric8-auth/pull/711) to integrate contract testing from :
https://github.com/fabric8io/fabric8-test/tree/master/contract_tests/go/auth-api
Figure out how we can integrate that for fabric8-build
Fabric8 Build issue: https://github.com/fabric8-services/fabric8-build/issues/103 | 1.0 | Build: Add contract testing - Fabric8-auth [has started](https://github.com/fabric8-services/fabric8-auth/pull/711) to integrate contract testing from :
https://github.com/fabric8io/fabric8-test/tree/master/contract_tests/go/auth-api
Figure out how we can integrate that for fabric8-build
Fabric8 Build issue: https://github.com/fabric8-services/fabric8-build/issues/103 | non_main | build add contract testing auth to integrate contract testing from figure out how we can integrate that for build build issue | 0 |
2,179 | 7,641,243,110 | IssuesEvent | 2018-05-08 03:35:40 | tgstation/tgstation | https://api.github.com/repos/tgstation/tgstation | closed | Space torus code doesn't work right and isn't needed, but is too nice to destroy | Maintainability/Hinders improvements | Ever since randomized space, having convoluted space torus z transitions hasn't really been needed. Since everything out there is random, their relative order to each other is irrelevant. That alone wouldn't warrant its removal.
The problem is the space torus is absolutely rife with self looping z levels, by my count EVERY z level self loops either north/south or east/west. Consequently the geometry of space is always either 9x1 or 1x9. Either you're self looping or your looping through everything. There is absolutely no reason we couldn't just use the 9 cross linked z leves in a 3x3 grid formation and be done with it and remove all self loops.
It's just that the current code [space_transition.dm] just cannot properly do that. Attempting to expand the grid to a 3x3 square leads to a formation where the z levels form a zigzag diagonally jointed mess. There doesn't seem to be any proper way to do this and it'd be too much of a shame to just scrap it entirely.
A note on ministation z5: Ministation is the only station not using lavaland, and that's important because with lavaland there's a proper 9 cross linked z levels, but without it there's 10. HOWEVER there's no mining shuttle at all on ministation, and in fact their z5 is self linking, meaning it's a dead zone where nothing can ever enter or leave. This means you don't need to give special consideration to it.
Tried to do this myself but not great at reading the code so whatever :cow2:
| True | Space torus code doesn't work right and isn't needed, but is too nice to destroy - Ever since randomized space, having convoluted space torus z transitions hasn't really been needed. Since everything out there is random, their relative order to each other is irrelevant. That alone wouldn't warrant its removal.
The problem is the space torus is absolutely rife with self looping z levels, by my count EVERY z level self loops either north/south or east/west. Consequently the geometry of space is always either 9x1 or 1x9. Either you're self looping or your looping through everything. There is absolutely no reason we couldn't just use the 9 cross linked z leves in a 3x3 grid formation and be done with it and remove all self loops.
It's just that the current code [space_transition.dm] just cannot properly do that. Attempting to expand the grid to a 3x3 square leads to a formation where the z levels form a zigzag diagonally jointed mess. There doesn't seem to be any proper way to do this and it'd be too much of a shame to just scrap it entirely.
A note on ministation z5: Ministation is the only station not using lavaland, and that's important because with lavaland there's a proper 9 cross linked z levels, but without it there's 10. HOWEVER there's no mining shuttle at all on ministation, and in fact their z5 is self linking, meaning it's a dead zone where nothing can ever enter or leave. This means you don't need to give special consideration to it.
Tried to do this myself but not great at reading the code so whatever :cow2:
| main | space torus code doesn t work right and isn t needed but is too nice to destroy ever since randomized space having convoluted space torus z transitions hasn t really been needed since everything out there is random their relative order to each other is irrelevant that alone wouldn t warrant its removal the problem is the space torus is absolutely rife with self looping z levels by my count every z level self loops either north south or east west consequently the geometry of space is always either or either you re self looping or your looping through everything there is absolutely no reason we couldn t just use the cross linked z leves in a grid formation and be done with it and remove all self loops it s just that the current code just cannot properly do that attempting to expand the grid to a square leads to a formation where the z levels form a zigzag diagonally jointed mess there doesn t seem to be any proper way to do this and it d be too much of a shame to just scrap it entirely a note on ministation ministation is the only station not using lavaland and that s important because with lavaland there s a proper cross linked z levels but without it there s however there s no mining shuttle at all on ministation and in fact their is self linking meaning it s a dead zone where nothing can ever enter or leave this means you don t need to give special consideration to it tried to do this myself but not great at reading the code so whatever | 1 |
1,007 | 4,781,925,474 | IssuesEvent | 2016-10-28 11:16:20 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | win_nssm idempodence problems | affects_2.2 bug_report waiting_on_maintainer windows | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
win_nssm
##### ANSIBLE VERSION
2.2.0 (devel as of August 1st 2016)
```
```
##### CONFIGURATION
##### OS / ENVIRONMENT
Ansible node: Ubuntu 14.04
Target node: Windows Server 2012R2
##### SUMMARY
Might be an edge case, but it showed up as we switched directories for our Logstash clusters.
If win_nssm finds an existing service and reconfigures it (for example, if the old service was C:\test\myfile.exe and a new config points to D:\test\myfile.exe), the service will keep its old app startup (AppDirectory) path - in this example AppDirectory will still point to C:\test. Not sure if this is a bug in win_nssm or in the ansible module. Maybe it would be a good idea to expose AppDirectory as an option in the Ansible module so that this could be specified
##### STEPS TO REPRODUCE
see above
##### EXPECTED RESULTS
win_nssm reconfigures the service with the correct AppDirectory
##### ACTUAL RESULTS
win_nssm configures the service, keeping the old value for AppDirectory
| True | win_nssm idempodence problems - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
win_nssm
##### ANSIBLE VERSION
2.2.0 (devel as of August 1st 2016)
```
```
##### CONFIGURATION
##### OS / ENVIRONMENT
Ansible node: Ubuntu 14.04
Target node: Windows Server 2012R2
##### SUMMARY
Might be an edge case, but it showed up as we switched directories for our Logstash clusters.
If win_nssm finds an existing service and reconfigures it (for example, if the old service was C:\test\myfile.exe and a new config points to D:\test\myfile.exe), the service will keep its old app startup (AppDirectory) path - in this example AppDirectory will still point to C:\test. Not sure if this is a bug in win_nssm or in the ansible module. Maybe it would be a good idea to expose AppDirectory as an option in the Ansible module so that this could be specified
##### STEPS TO REPRODUCE
see above
##### EXPECTED RESULTS
win_nssm reconfigures the service with the correct AppDirectory
##### ACTUAL RESULTS
win_nssm configures the service, keeping the old value for AppDirectory
| main | win nssm idempodence problems issue type bug report component name win nssm ansible version devel as of august configuration os environment ansible node ubuntu target node windows server summary might be an edge case but it showed up as we switched directories for our logstash clusters if win nssm finds an existing service and reconfigures it for example if the old service was c test myfile exe and a new config points to d test myfile exe the service will keep its old app startup appdirectory path in this example appdirectory will still point to c test not sure if this is a bug in win nssm or in the ansible module maybe it would be a good idea to expose appdirectory as an option in the ansible module so that this could be specified steps to reproduce see above expected results win nssm reconfigures the service with the correct appdirectory actual results win nssm configures the service keeping the old value for appdirectory | 1 |
758,535 | 26,559,170,301 | IssuesEvent | 2023-01-20 14:32:42 | wp-media/wp-rocket | https://api.github.com/repos/wp-media/wp-rocket | closed | Potential redirect loop when using 3.11.4, trailing slash redirection | type: bug priority: 🔥critical module: cache severity: critical | **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version - 3.11.4
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
When permalinks are set to be without `/` and website is in the subdirectory we might experience redirect loop.
**To Reproduce**
Steps to reproduce the behavior:
1. Use website placed in subdirectory (could be multisite)
2. Change the permalinks to be without `/`
3. Visit frontend in incognito
**Expected behavior**
No redirect loop
**Additional context**
Still not sure if there are any other factors needed to reproduce.
**Temporary solution**
Change the permalinks settings to be with `/`
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| 1.0 | Potential redirect loop when using 3.11.4, trailing slash redirection - **Before submitting an issue please check that you’ve completed the following steps:**
- Made sure you’re on the latest version - 3.11.4
- Used the search feature to ensure that the bug hasn’t been reported before
**Describe the bug**
When permalinks are set to be without `/` and website is in the subdirectory we might experience redirect loop.
**To Reproduce**
Steps to reproduce the behavior:
1. Use website placed in subdirectory (could be multisite)
2. Change the permalinks to be without `/`
3. Visit frontend in incognito
**Expected behavior**
No redirect loop
**Additional context**
Still not sure if there are any other factors needed to reproduce.
**Temporary solution**
Change the permalinks settings to be with `/`
**Backlog Grooming (for WP Media dev team use only)**
- [ ] Reproduce the problem
- [ ] Identify the root cause
- [ ] Scope a solution
- [ ] Estimate the effort
| non_main | potential redirect loop when using trailing slash redirection before submitting an issue please check that you’ve completed the following steps made sure you’re on the latest version used the search feature to ensure that the bug hasn’t been reported before describe the bug when permalinks are set to be without and website is in the subdirectory we might experience redirect loop to reproduce steps to reproduce the behavior use website placed in subdirectory could be multisite change the permalinks to be without visit frontend in incognito expected behavior no redirect loop additional context still not sure if there are any other factors needed to reproduce temporary solution change the permalinks settings to be with backlog grooming for wp media dev team use only reproduce the problem identify the root cause scope a solution estimate the effort | 0 |
4,164 | 19,981,501,422 | IssuesEvent | 2022-01-30 00:39:29 | thumbor/thumbor-bootcamp | https://api.github.com/repos/thumbor/thumbor-bootcamp | opened | [Bootcamp Task] Change thumbor-url command to click | task L2 python maintainability | ## Areas of Expertise
Python
## Summary
Change the `thumbor-url` command to use [click](https://click.palletsprojects.com/en/8.0.x/).
## Involved Modules
* [thumbor](https://github.com/thumbor/thumbor/) - [url_composer](https://github.com/thumbor/thumbor/blob/master/thumbor/url_composer.py)
## Task Relevance
By upgrading the handling of the cli parts of thumbor to a more established library like click we ensure maintainability for the future.
## Further Details
Completing this task means the `thumbor-url` command in thumbor is handled by click and not by optparse (deprecated module in Python 3).
## How to complete this task?
To complete this task, follow this workflow:
1. [Fork the involved repositories](http://help.github.com/fork-a-repo/)
2. In each repository there's a documentation on how to install dependencies and initialize your environment
3. Hack, in no particular order:
- Write code & tests
- Write new tests
- Write docs
- Improve design
- Check that all tests pass
- Repeat until you're satisfied
4. [Submit a pull request](https://docs.github.com/en/github/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request).
| True | [Bootcamp Task] Change thumbor-url command to click - ## Areas of Expertise
Python
## Summary
Change the `thumbor-url` command to use [click](https://click.palletsprojects.com/en/8.0.x/).
## Involved Modules
* [thumbor](https://github.com/thumbor/thumbor/) - [url_composer](https://github.com/thumbor/thumbor/blob/master/thumbor/url_composer.py)
## Task Relevance
By upgrading the handling of the cli parts of thumbor to a more established library like click we ensure maintainability for the future.
## Further Details
Completing this task means the `thumbor-url` command in thumbor is handled by click and not by optparse (deprecated module in Python 3).
## How to complete this task?
To complete this task, follow this workflow:
1. [Fork the involved repositories](http://help.github.com/fork-a-repo/)
2. In each repository there's a documentation on how to install dependencies and initialize your environment
3. Hack, in no particular order:
- Write code & tests
- Write new tests
- Write docs
- Improve design
- Check that all tests pass
- Repeat until you're satisfied
4. [Submit a pull request](https://docs.github.com/en/github/collaborating-with-pull-requests/proposing-changes-to-your-work-with-pull-requests/creating-a-pull-request).
| main | change thumbor url command to click areas of expertise python summary change the thumbor url command to use involved modules task relevance by upgrading the handling of the cli parts of thumbor to a more established library like click we ensure maintainability for the future further details completing this task means the thumbor url command in thumbor is handled by click and not by optparse deprecated module in python how to complete this task to complete this task follow this workflow in each repository there s a documentation on how to install dependencies and initialize your environment hack in no particular order write code tests write new tests write docs improve design check that all tests pass repeat until you re satisfied | 1 |
5,556 | 27,804,499,447 | IssuesEvent | 2023-03-17 18:33:00 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | Defining CORS when ApiKeyRequired is true results in an OPTIONS method that requires an API key | type/bug area/swagger area/resource/api type/bug-triaged maintainer/need-followup | <!--
Before reporting a new issue, make sure we don't have any duplicates already open or closed by
searching the issues list. If there is a duplicate, re-open or add a comment to the
existing issue instead of creating a new one. If you are reporting a bug,
make sure to include relevant information asked below to help with debugging.
## GENERAL HELP QUESTIONS ##
Github Issues is for bug reports and feature requests. If you have general support
questions, the following locations are a good place:
- Post a question in StackOverflow with "aws-sam" tag
-->
**Description:**
<!-- Briefly describe the problem you are facing -->
When setting up an API Gateway with {proxy+} integration through SAM, having a required API key prevents the CORS definition from properly responding to OPTIONS requests, since they require an API key as well.
**Steps to reproduce the issue:**
1. Define an Api that requires an API key:
```
Resources:
BasicAWSApiGateway:
Type: AWS::Serverless::Api
Properties:
StageName: !Ref Environment
Auth:
ApiKeyRequired: true
Cors: "'*'"
DefinitionBody:
swagger: "2.0.0"
info:
version: 1.0.0
paths:
/{proxy+}:
x-amazon-apigateway-any-method:
produces:
- application/json
parameters:
- in: path
name: proxy
required: true
type: string
x-amazon-apigateway-integration:
uri: !Sub "arn:aws:apigateway:${AWS::Region}:lambda:path/2015-03-31/functions/arn:aws:lambda:${AWS::Region}:${AWS::AccountId}:function:${LambdaName}-${Environment}/invocations"
httpMethod: POST
type: aws_proxy
passthroughBehavior: when_no_match
credentials: !GetAtt ApiGatewayExecutionRole.Arn
```
**Observed result:**
The `OPTIONS` method in API Gateway requires an API Key, blocking CORS since browsers don't add a `x-api-key` header.
**Expected result:**
The `OPTIONS` method should not require an API Key.
| True | Defining CORS when ApiKeyRequired is true results in an OPTIONS method that requires an API key - <!--
Before reporting a new issue, make sure we don't have any duplicates already open or closed by
searching the issues list. If there is a duplicate, re-open or add a comment to the
existing issue instead of creating a new one. If you are reporting a bug,
make sure to include relevant information asked below to help with debugging.
## GENERAL HELP QUESTIONS ##
Github Issues is for bug reports and feature requests. If you have general support
questions, the following locations are a good place:
- Post a question in StackOverflow with "aws-sam" tag
-->
**Description:**
<!-- Briefly describe the problem you are facing -->
When setting up an API Gateway with {proxy+} integration through SAM, having a required API key prevents the CORS definition from properly responding to OPTIONS requests, since they require an API key as well.
**Steps to reproduce the issue:**
1. Define an Api that requires an API key:
```
Resources:
BasicAWSApiGateway:
Type: AWS::Serverless::Api
Properties:
StageName: !Ref Environment
Auth:
ApiKeyRequired: true
Cors: "'*'"
DefinitionBody:
swagger: "2.0.0"
info:
version: 1.0.0
paths:
/{proxy+}:
x-amazon-apigateway-any-method:
produces:
- application/json
parameters:
- in: path
name: proxy
required: true
type: string
x-amazon-apigateway-integration:
uri: !Sub "arn:aws:apigateway:${AWS::Region}:lambda:path/2015-03-31/functions/arn:aws:lambda:${AWS::Region}:${AWS::AccountId}:function:${LambdaName}-${Environment}/invocations"
httpMethod: POST
type: aws_proxy
passthroughBehavior: when_no_match
credentials: !GetAtt ApiGatewayExecutionRole.Arn
```
**Observed result:**
The `OPTIONS` method in API Gateway requires an API Key, blocking CORS since browsers don't add a `x-api-key` header.
**Expected result:**
The `OPTIONS` method should not require an API Key.
| main | defining cors when apikeyrequired is true results in an options method that requires an api key before reporting a new issue make sure we don t have any duplicates already open or closed by searching the issues list if there is a duplicate re open or add a comment to the existing issue instead of creating a new one if you are reporting a bug make sure to include relevant information asked below to help with debugging general help questions github issues is for bug reports and feature requests if you have general support questions the following locations are a good place post a question in stackoverflow with aws sam tag description when setting up an api gateway with proxy integration through sam having a required api key prevents the cors definition from properly responding to options requests since they require an api key as well steps to reproduce the issue define an api that requires an api key resources basicawsapigateway type aws serverless api properties stagename ref environment auth apikeyrequired true cors definitionbody swagger info version paths proxy x amazon apigateway any method produces application json parameters in path name proxy required true type string x amazon apigateway integration uri sub arn aws apigateway aws region lambda path functions arn aws lambda aws region aws accountid function lambdaname environment invocations httpmethod post type aws proxy passthroughbehavior when no match credentials getatt apigatewayexecutionrole arn observed result the options method in api gateway requires an api key blocking cors since browsers don t add a x api key header expected result the options method should not require an api key | 1 |
4,542 | 23,661,594,266 | IssuesEvent | 2022-08-26 16:05:59 | deislabs/spiderlightning | https://api.github.com/repos/deislabs/spiderlightning | closed | sanity: variances in wit-bindgen-* references | 🐛 bug 🚧 maintainer issue | **Description of the bug**
Not really a bug, just an observation: during `make build` I noticed that 2 different references to `wit-bindgen-*` modules are used:
```
...
Compiling wit-bindgen-gen-rust v0.1.0 (https://github.com/Mossaka/wit-bindgen?rev=8252b0e39c7495f647ec0b0898721a7c641fc6c8#8252b0e3)
Compiling wit-bindgen-gen-rust v0.1.0 (https://github.com/bytecodealliance/wit-bindgen?rev=a79a4be33d76ddf62839ba71602c26a96610ef7c#a79a4be3)
...
Compiling wit-bindgen-gen-rust v0.1.0 (https://github.com/Mossaka/wit-bindgen?rev=8252b0e39c7495f647ec0b0898721a7c641fc6c8#8252b0e3)
Compiling wit-bindgen-gen-rust v0.1.0 (https://github.com/bytecodealliance/wit-bindgen?rev=a79a4be33d76ddf62839ba71602c26a96610ef7c#a79a4be3)
...
Compiling wit-bindgen-wasmtime-impl v0.1.0 (https://github.com/Mossaka/wit-bindgen?rev=8252b0e39c7495f647ec0b0898721a7c641fc6c8#8252b0e3)
Compiling wit-bindgen-wasmtime-impl v0.1.0 (https://github.com/bytecodealliance/wit-bindgen?rev=a79a4be33d76ddf62839ba71602c26a96610ef7c#a79a4be3)
...
Compiling wit-bindgen-wasmtime v0.1.0 (https://github.com/bytecodealliance/wit-bindgen?rev=a79a4be33d76ddf62839ba71602c26a96610ef7c#a79a4be3)
Compiling wit-bindgen-wasmtime v0.1.0 (https://github.com/Mossaka/wit-bindgen?rev=8252b0e39c7495f647ec0b0898721a7c641fc6c8#8252b0e3)
...
```
Is this intentional? If not, what should be the primary reference used?
**To Reproduce**
Either `make build` and follow carefully or check for dependency references with
```
$ grep -r "wit-bindgen-wasmtime" --include=Cargo.toml
crates/lockd/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/pubsub/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/runtime-configs/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/http-api/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/Mossaka/wit-bindgen", rev = "8252b0e39c7495f647ec0b0898721a7c641fc6c8" }
crates/events/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/mq/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/http/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/Mossaka/wit-bindgen", rev = "8252b0e39c7495f647ec0b0898721a7c641fc6c8" }
crates/events-api/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/kv/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/runtime/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
slight/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
```
**Additional context**
Just forked from `main` and wanted to make a clean build & install. | True | sanity: variances in wit-bindgen-* references - **Description of the bug**
Not really a bug, just an observation: during `make build` I noticed that 2 different references to `wit-bindgen-*` modules are used:
```
...
Compiling wit-bindgen-gen-rust v0.1.0 (https://github.com/Mossaka/wit-bindgen?rev=8252b0e39c7495f647ec0b0898721a7c641fc6c8#8252b0e3)
Compiling wit-bindgen-gen-rust v0.1.0 (https://github.com/bytecodealliance/wit-bindgen?rev=a79a4be33d76ddf62839ba71602c26a96610ef7c#a79a4be3)
...
Compiling wit-bindgen-gen-rust v0.1.0 (https://github.com/Mossaka/wit-bindgen?rev=8252b0e39c7495f647ec0b0898721a7c641fc6c8#8252b0e3)
Compiling wit-bindgen-gen-rust v0.1.0 (https://github.com/bytecodealliance/wit-bindgen?rev=a79a4be33d76ddf62839ba71602c26a96610ef7c#a79a4be3)
...
Compiling wit-bindgen-wasmtime-impl v0.1.0 (https://github.com/Mossaka/wit-bindgen?rev=8252b0e39c7495f647ec0b0898721a7c641fc6c8#8252b0e3)
Compiling wit-bindgen-wasmtime-impl v0.1.0 (https://github.com/bytecodealliance/wit-bindgen?rev=a79a4be33d76ddf62839ba71602c26a96610ef7c#a79a4be3)
...
Compiling wit-bindgen-wasmtime v0.1.0 (https://github.com/bytecodealliance/wit-bindgen?rev=a79a4be33d76ddf62839ba71602c26a96610ef7c#a79a4be3)
Compiling wit-bindgen-wasmtime v0.1.0 (https://github.com/Mossaka/wit-bindgen?rev=8252b0e39c7495f647ec0b0898721a7c641fc6c8#8252b0e3)
...
```
Is this intentional? If not, what should be the primary reference used?
**To Reproduce**
Either `make build` and follow carefully or check for dependency references with
```
$ grep -r "wit-bindgen-wasmtime" --include=Cargo.toml
crates/lockd/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/pubsub/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/runtime-configs/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/http-api/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/Mossaka/wit-bindgen", rev = "8252b0e39c7495f647ec0b0898721a7c641fc6c8" }
crates/events/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/mq/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/http/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/Mossaka/wit-bindgen", rev = "8252b0e39c7495f647ec0b0898721a7c641fc6c8" }
crates/events-api/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/kv/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
crates/runtime/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
slight/Cargo.toml:wit-bindgen-wasmtime = { git = "https://github.com/bytecodealliance/wit-bindgen", rev = "a79a4be33d76ddf62839ba71602c26a96610ef7c" }
```
**Additional context**
Just forked from `main` and wanted to make a clean build & install. | main | sanity variances in wit bindgen references description of the bug not really a bug just an observation during make build i noticed that different references to wit bindgen modules are used compiling wit bindgen gen rust compiling wit bindgen gen rust compiling wit bindgen gen rust compiling wit bindgen gen rust compiling wit bindgen wasmtime impl compiling wit bindgen wasmtime impl compiling wit bindgen wasmtime compiling wit bindgen wasmtime is this intentional if not what should be the primary reference used to reproduce either make build and follow carefully or check for dependency references with grep r wit bindgen wasmtime include cargo toml crates lockd cargo toml wit bindgen wasmtime git rev crates pubsub cargo toml wit bindgen wasmtime git rev crates runtime configs cargo toml wit bindgen wasmtime git rev crates http api cargo toml wit bindgen wasmtime git rev crates events cargo toml wit bindgen wasmtime git rev crates mq cargo toml wit bindgen wasmtime git rev crates http cargo toml wit bindgen wasmtime git rev crates events api cargo toml wit bindgen wasmtime git rev crates kv cargo toml wit bindgen wasmtime git rev crates runtime cargo toml wit bindgen wasmtime git rev slight cargo toml wit bindgen wasmtime git rev additional context just forked from main and wanted to make a clean build install | 1 |
1,995 | 6,705,470,043 | IssuesEvent | 2017-10-12 00:28:04 | cannawen/metric_units_reddit_bot | https://api.github.com/repos/cannawen/metric_units_reddit_bot | closed | Handling ranges better | feature request in progress maintainer approved | Currently, the bot handles ranges in a non-intuitive way by displaying each value separately, the current ranges handling can be made much better instead if we process them separately in `findPotentialConversions` and `formatConversion` to do the following:
- Display ranges in comment as ranges in response, for example, `1-2 miles ≈ 1.6-3.2 km`
- Handle many more joining keywords, such as `1 to 2 miles ≈ 1.6 to 3.2 km`, `1 or 2 miles ≈ 1.6 or 3.2 km` and similar.
I think I can work on this(if the changes make sense?) | True | Handling ranges better - Currently, the bot handles ranges in a non-intuitive way by displaying each value separately, the current ranges handling can be made much better instead if we process them separately in `findPotentialConversions` and `formatConversion` to do the following:
- Display ranges in comment as ranges in response, for example, `1-2 miles ≈ 1.6-3.2 km`
- Handle many more joining keywords, such as `1 to 2 miles ≈ 1.6 to 3.2 km`, `1 or 2 miles ≈ 1.6 or 3.2 km` and similar.
I think I can work on this(if the changes make sense?) | main | handling ranges better currently the bot handles ranges in a non intuitive way by displaying each value separately the current ranges handling can be made much better instead if we process them separately in findpotentialconversions and formatconversion to do the following display ranges in comment as ranges in response for example miles ≈ km handle many more joining keywords such as to miles ≈ to km or miles ≈ or km and similar i think i can work on this if the changes make sense | 1 |
138,197 | 11,194,311,075 | IssuesEvent | 2020-01-03 00:22:09 | GMLC-TDC/HELICS | https://api.github.com/repos/GMLC-TDC/HELICS | closed | MSYS2 test build | testing | related to #751.
Some recent changes with the network detection broke the MSYS2 build. The fix was pretty easy and is included in #777, so that isn't the issue. But as a lesson learned we probably should have a cron build or something to test a build in MSYS2 to make sure that build type continues to work in the future. | 1.0 | MSYS2 test build - related to #751.
Some recent changes with the network detection broke the MSYS2 build. The fix was pretty easy and is included in #777, so that isn't the issue. But as a lesson learned we probably should have a cron build or something to test a build in MSYS2 to make sure that build type continues to work in the future. | non_main | test build related to some recent changes with the network detection broke the build the fix was pretty easy and is included in so that isn t the issue but as a lesson learned we probably should have a cron build or something to test a build in to make sure that build type continues to work in the future | 0 |
152,256 | 13,449,799,229 | IssuesEvent | 2020-09-08 17:28:41 | mozilla/uniffi-rs | https://api.github.com/repos/mozilla/uniffi-rs | closed | Document uniffi-bindgen | documentation | Our README.md doesn't include any information about the CLI, we should probably add it. In #212 We add a couple of mentions of it, but we might benefit from having a section there explaining it.
- This should explain how one can download the tool (currently using `cargo install --path`, once it's on crates.io we'll add that too)
- Maybe an explanation of how the tool fits in the grand scheme of things.
- A mention of how to find more information (telling users to run the `--help` may be enough here)
┆Issue is synchronized with this [Jira Task](https://jira.mozilla.com/browse/SYNC-1638)
| 1.0 | Document uniffi-bindgen - Our README.md doesn't include any information about the CLI, we should probably add it. In #212 We add a couple of mentions of it, but we might benefit from having a section there explaining it.
- This should explain how one can download the tool (currently using `cargo install --path`, once it's on crates.io we'll add that too)
- Maybe an explanation of how the tool fits in the grand scheme of things.
- A mention of how to find more information (telling users to run the `--help` may be enough here)
┆Issue is synchronized with this [Jira Task](https://jira.mozilla.com/browse/SYNC-1638)
| non_main | document uniffi bindgen our readme md doesn t include any information about the cli we should probably add it in we add a couple of mentions of it but we might benefit from having a section there explaining it this should explain how one can download the tool currently using cargo install path once it s on crates io we ll add that too maybe an explanation of how the tool fits in the grand scheme of things a mention of how to find more information telling users to run the help may be enough here ┆issue is synchronized with this | 0 |
210,730 | 7,194,133,103 | IssuesEvent | 2018-02-04 00:17:40 | ozelevrim/EvrimNews | https://api.github.com/repos/ozelevrim/EvrimNews | closed | Remove unnecessary files from project to decrease size | enhancement priority: high | Remove unnecessary files from project to decrease size.
https://stackoverflow.com/questions/6373482/remove-all-unused-resources-from-an-android-project | 1.0 | Remove unnecessary files from project to decrease size - Remove unnecessary files from project to decrease size.
https://stackoverflow.com/questions/6373482/remove-all-unused-resources-from-an-android-project | non_main | remove unnecessary files from project to decrease size remove unnecessary files from project to decrease size | 0 |
47,605 | 12,061,500,200 | IssuesEvent | 2020-04-16 00:00:09 | osquery/osquery | https://api.github.com/repos/osquery/osquery | opened | Not enough space to build on the CI or how to reduce binary bloat | CI/CD build | # Description
Lately Azure seems to have reduced the free space available on their VMs and specifically on Linux.
They've always guaranteed 10GB of space, but for debug builds we use more than that.
Currently the Linux VMs have actually 12GB free (taken from the running Docker image), which is not always enough.
To reduce the space usage we have implemented a flag to disable debug symbols when build in Debug, since they are not currently used anyway; all the third party libraries are always built in Release mode
I also very recently reduced the ccache cache size from it's 5G max, to 2G; this gives us some breathing space but I don't know how long it will lasts.
Following some numbers to understand how big the build folder is:
```
Built on Linux, tests included
Release 3.7G
Debug 13G
RelWithDebInfo 12G
Debug NoDebugSymbols 5.3G
```
Looking at the last mode, Debug + No Symbols, we have:
```
~492M for *.o
~429M for *.a
~4018M for *-test (all the tests)
~66M osqueryd
98 test binaries
```
So with ccache originally at ~5G, we had 10G total; then add the source size, which is ~2G and you can see how we were hitting the limit.
# What we can try to do
I think we could try to act on the following points:
- [ ] Try to reduce the number of separate test binaries
- [ ] Verify that OSQUERY_NO_DEBUG_SYMBOLS works correctly and that `-g0` is passed to all targets as the last `-g` option
- [ ] Review test and library dependencies, correct and clean them up
- [ ] Try to better isolate libraries that require to be linked with whole-archive, so that they are smaller.
- [ ] Try to identify templated code bloat and reduce it
# Notes
- For the first point, I think that to be really effective, either we have to merge a lot of small tests together or few big ones.
There's a downside with merging though, which is that having less but bigger binaries will slow down linking further (on Windows is especially slow), and therefore will slow down a bit the development cycle of changing something and recompiling.
I believe we could add some machinery in CMake to be able to merge and split targets via options, so that in a "development mode" one could keep things separated and have faster linking times, while in a "CI" mode it merges them.
- For the third point, this is partially tracked by https://github.com/osquery/osquery/issues/5916.
The idea here is that by correcting all the dependencies, a lot of unneeded includes are removed and so .o get smaller.
I don't expect this to have a huge impact, but it's good nonetheless for correctness, not having to specify in CMake a ton of deps that should be already carried over by other dependencies, and should slightly decrease build time.
- For the fourth point, this I think is one of the major hitters.
We use whole-archive because we have static libraries that make use of global static initialization of variables to trigger class constructors, to auto register tables.
While this is something that simplify managing registration of tables, causes some other issues (beyond having to use whole-archive); but talking about whole-archive, if the linker tries to link a library that does such use of initialization and the variable/symbol is not used actually referenced in the executable, then it will drop the entire logic, which means no initialization will happen.
That's because as a quick workaround we force the linker to include ALL symbols from that library.
The side effect, as highlighted, is that we get no unused symbol discarding, at least of the ones we really don't care about, so we bloat the executable.
Last time I tried to drop whole-archive from the extension library, the build size was reduced by several GBs (although this was tested with a debug build WITH symbols, as far as I recall).
Some ways to solve this:
1) The initialization/registration code is put into a .h which is then included by the executable (the osqueryd target); this means that that code actually is part of the executable and cannot be dropped by the linker, so no need to have whole-archive on the static library.
2) Try to refactor each library that contains such static initialization, so that it contains less unrelated code, so to reduce bloating
3) Do not use static init, which means manually listing init functions to be called in the code.
This has the disadvantage that the list has to be manually created, but at the same time the advantages are that the initialization time has less issues (initialization order fiasco, not possible to skip initialization of some table based on runtime logic) and we skip whole-archive entirely
- The last point is potentially the less rewarding of all three, because the major issue I think comes from third party libraries more than our code.
Moreover to be able to do a major impact we would need to be able to use C++17.
Here the bloat mostly comes from function signatures. In many cases, templated functions simply have long type names as parameters, and there's not much one can do.
Though what one can do is to shorten the return type signature by using `auto` (this might not be possible without C++17).
Another thing is not using `std::function` but a lambda, which results in a shorter string.
There are also cases where it's possible that too many combinations of concrete template parameters and functions with those parameters are created, starting from a single combination of those parameters. This though often happens in code that uses metaprogramming heavily.
A tool that's useful to check what is bloating some binaries is https://github.com/google/bloaty, specifically `bloaty <binary> -d compileunits -s file -n 0`; you would need to compile osquery in RelWithDebInfo, to see the source file names. | 1.0 | Not enough space to build on the CI or how to reduce binary bloat - # Description
Lately Azure seems to have reduced the free space available on their VMs and specifically on Linux.
They've always guaranteed 10GB of space, but for debug builds we use more than that.
Currently the Linux VMs have actually 12GB free (taken from the running Docker image), which is not always enough.
To reduce the space usage we have implemented a flag to disable debug symbols when build in Debug, since they are not currently used anyway; all the third party libraries are always built in Release mode
I also very recently reduced the ccache cache size from it's 5G max, to 2G; this gives us some breathing space but I don't know how long it will lasts.
Following some numbers to understand how big the build folder is:
```
Built on Linux, tests included
Release 3.7G
Debug 13G
RelWithDebInfo 12G
Debug NoDebugSymbols 5.3G
```
Looking at the last mode, Debug + No Symbols, we have:
```
~492M for *.o
~429M for *.a
~4018M for *-test (all the tests)
~66M osqueryd
98 test binaries
```
So with ccache originally at ~5G, we had 10G total; then add the source size, which is ~2G and you can see how we were hitting the limit.
# What we can try to do
I think we could try to act on the following points:
- [ ] Try to reduce the number of separate test binaries
- [ ] Verify that OSQUERY_NO_DEBUG_SYMBOLS works correctly and that `-g0` is passed to all targets as the last `-g` option
- [ ] Review test and library dependencies, correct and clean them up
- [ ] Try to better isolate libraries that require to be linked with whole-archive, so that they are smaller.
- [ ] Try to identify templated code bloat and reduce it
# Notes
- For the first point, I think that to be really effective, either we have to merge a lot of small tests together or few big ones.
There's a downside with merging though, which is that having less but bigger binaries will slow down linking further (on Windows is especially slow), and therefore will slow down a bit the development cycle of changing something and recompiling.
I believe we could add some machinery in CMake to be able to merge and split targets via options, so that in a "development mode" one could keep things separated and have faster linking times, while in a "CI" mode it merges them.
- For the third point, this is partially tracked by https://github.com/osquery/osquery/issues/5916.
The idea here is that by correcting all the dependencies, a lot of unneeded includes are removed and so .o get smaller.
I don't expect this to have a huge impact, but it's good nonetheless for correctness, not having to specify in CMake a ton of deps that should be already carried over by other dependencies, and should slightly decrease build time.
- For the fourth point, this I think is one of the major hitters.
We use whole-archive because we have static libraries that make use of global static initialization of variables to trigger class constructors, to auto register tables.
While this is something that simplify managing registration of tables, causes some other issues (beyond having to use whole-archive); but talking about whole-archive, if the linker tries to link a library that does such use of initialization and the variable/symbol is not used actually referenced in the executable, then it will drop the entire logic, which means no initialization will happen.
That's because as a quick workaround we force the linker to include ALL symbols from that library.
The side effect, as highlighted, is that we get no unused symbol discarding, at least of the ones we really don't care about, so we bloat the executable.
Last time I tried to drop whole-archive from the extension library, the build size was reduced by several GBs (although this was tested with a debug build WITH symbols, as far as I recall).
Some ways to solve this:
1) The initialization/registration code is put into a .h which is then included by the executable (the osqueryd target); this means that that code actually is part of the executable and cannot be dropped by the linker, so no need to have whole-archive on the static library.
2) Try to refactor each library that contains such static initialization, so that it contains less unrelated code, so to reduce bloating
3) Do not use static init, which means manually listing init functions to be called in the code.
This has the disadvantage that the list has to be manually created, but at the same time the advantages are that the initialization time has less issues (initialization order fiasco, not possible to skip initialization of some table based on runtime logic) and we skip whole-archive entirely
- The last point is potentially the less rewarding of all three, because the major issue I think comes from third party libraries more than our code.
Moreover to be able to do a major impact we would need to be able to use C++17.
Here the bloat mostly comes from function signatures. In many cases, templated functions simply have long type names as parameters, and there's not much one can do.
Though what one can do is to shorten the return type signature by using `auto` (this might not be possible without C++17).
Another thing is not using `std::function` but a lambda, which results in a shorter string.
There are also cases where it's possible that too many combinations of concrete template parameters and functions with those parameters are created, starting from a single combination of those parameters. This though often happens in code that uses metaprogramming heavily.
A tool that's useful to check what is bloating some binaries is https://github.com/google/bloaty, specifically `bloaty <binary> -d compileunits -s file -n 0`; you would need to compile osquery in RelWithDebInfo, to see the source file names. | non_main | not enough space to build on the ci or how to reduce binary bloat description lately azure seems to have reduced the free space available on their vms and specifically on linux they ve always guaranteed of space but for debug builds we use more than that currently the linux vms have actually free taken from the running docker image which is not always enough to reduce the space usage we have implemented a flag to disable debug symbols when build in debug since they are not currently used anyway all the third party libraries are always built in release mode i also very recently reduced the ccache cache size from it s max to this gives us some breathing space but i don t know how long it will lasts following some numbers to understand how big the build folder is built on linux tests included release debug relwithdebinfo debug nodebugsymbols looking at the last mode debug no symbols we have for o for a for test all the tests osqueryd test binaries so with ccache originally at we had total then add the source size which is and you can see how we were hitting the limit what we can try to do i think we could try to act on the following points try to reduce the number of separate test binaries verify that osquery no debug symbols works correctly and that is passed to all targets as the last g option review test and library dependencies correct and clean them up try to better isolate libraries that require to be linked with whole archive so that they are smaller try to identify templated code bloat and reduce it notes for the first point i think that to be really effective either we have to merge a lot of small tests together or few big ones there s a downside with merging though which is that having less but bigger binaries will slow down linking further on windows is especially slow and therefore will slow down a bit the development cycle of changing something and recompiling i believe we could add some machinery in cmake to be able to merge and split targets via options so that in a development mode one could keep things separated and have faster linking times while in a ci mode it merges them for the third point this is partially tracked by the idea here is that by correcting all the dependencies a lot of unneeded includes are removed and so o get smaller i don t expect this to have a huge impact but it s good nonetheless for correctness not having to specify in cmake a ton of deps that should be already carried over by other dependencies and should slightly decrease build time for the fourth point this i think is one of the major hitters we use whole archive because we have static libraries that make use of global static initialization of variables to trigger class constructors to auto register tables while this is something that simplify managing registration of tables causes some other issues beyond having to use whole archive but talking about whole archive if the linker tries to link a library that does such use of initialization and the variable symbol is not used actually referenced in the executable then it will drop the entire logic which means no initialization will happen that s because as a quick workaround we force the linker to include all symbols from that library the side effect as highlighted is that we get no unused symbol discarding at least of the ones we really don t care about so we bloat the executable last time i tried to drop whole archive from the extension library the build size was reduced by several gbs although this was tested with a debug build with symbols as far as i recall some ways to solve this the initialization registration code is put into a h which is then included by the executable the osqueryd target this means that that code actually is part of the executable and cannot be dropped by the linker so no need to have whole archive on the static library try to refactor each library that contains such static initialization so that it contains less unrelated code so to reduce bloating do not use static init which means manually listing init functions to be called in the code this has the disadvantage that the list has to be manually created but at the same time the advantages are that the initialization time has less issues initialization order fiasco not possible to skip initialization of some table based on runtime logic and we skip whole archive entirely the last point is potentially the less rewarding of all three because the major issue i think comes from third party libraries more than our code moreover to be able to do a major impact we would need to be able to use c here the bloat mostly comes from function signatures in many cases templated functions simply have long type names as parameters and there s not much one can do though what one can do is to shorten the return type signature by using auto this might not be possible without c another thing is not using std function but a lambda which results in a shorter string there are also cases where it s possible that too many combinations of concrete template parameters and functions with those parameters are created starting from a single combination of those parameters this though often happens in code that uses metaprogramming heavily a tool that s useful to check what is bloating some binaries is specifically bloaty d compileunits s file n you would need to compile osquery in relwithdebinfo to see the source file names | 0 |
255,074 | 27,484,728,738 | IssuesEvent | 2023-03-04 01:12:47 | panasalap/linux-4.1.15 | https://api.github.com/repos/panasalap/linux-4.1.15 | opened | CVE-2018-12232 (Medium) detected in linuxlinux-4.1.17 | security vulnerability | ## CVE-2018-12232 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.17</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In net/socket.c in the Linux kernel through 4.17.1, there is a race condition between fchownat and close in cases where they target the same socket file descriptor, related to the sock_close and sockfs_setattr functions. fchownat does not increment the file descriptor reference count, which allows close to set the socket to NULL during fchownat's execution, leading to a NULL pointer dereference and system crash.
<p>Publish Date: 2018-06-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-12232>CVE-2018-12232</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-12232">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-12232</a></p>
<p>Release Date: 2018-06-12</p>
<p>Fix Resolution: v4.18-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2018-12232 (Medium) detected in linuxlinux-4.1.17 - ## CVE-2018-12232 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.1.17</b></p></summary>
<p>
<p>Apache Software Foundation (ASF)</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In net/socket.c in the Linux kernel through 4.17.1, there is a race condition between fchownat and close in cases where they target the same socket file descriptor, related to the sock_close and sockfs_setattr functions. fchownat does not increment the file descriptor reference count, which allows close to set the socket to NULL during fchownat's execution, leading to a NULL pointer dereference and system crash.
<p>Publish Date: 2018-06-12
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2018-12232>CVE-2018-12232</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-12232">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2018-12232</a></p>
<p>Release Date: 2018-06-12</p>
<p>Fix Resolution: v4.18-rc1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in base branch master vulnerable source files vulnerability details in net socket c in the linux kernel through there is a race condition between fchownat and close in cases where they target the same socket file descriptor related to the sock close and sockfs setattr functions fchownat does not increment the file descriptor reference count which allows close to set the socket to null during fchownat s execution leading to a null pointer dereference and system crash publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
178,882 | 6,619,766,402 | IssuesEvent | 2017-09-21 13:28:36 | NREL/EnergyPlus | https://api.github.com/repos/NREL/EnergyPlus | closed | OA control type crash | Priority1 | Issue overview
--------------
Testing new features from #6223 and #6252. Running the defect file using Chicago weather with debug version, the program was crashed on line 2908 in MixedAir.cc.
The index of cOAFlowMethodTypes is 9, while the cOAFlowMethodTypes is sized to 6.
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus 8.8.0 (as of 2017-09-12)
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| 1.0 | OA control type crash - Issue overview
--------------
Testing new features from #6223 and #6252. Running the defect file using Chicago weather with debug version, the program was crashed on line 2908 in MixedAir.cc.
The index of cOAFlowMethodTypes is 9, while the cOAFlowMethodTypes is sized to 6.
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus 8.8.0 (as of 2017-09-12)
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [ ] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
| non_main | oa control type crash issue overview testing new features from and running the defect file using chicago weather with debug version the program was crashed on line in mixedair cc the index of coaflowmethodtypes is while the coaflowmethodtypes is sized to details some additional details for this issue if relevant platform operating system version version of energyplus as of checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect | 0 |
2,318 | 8,303,638,383 | IssuesEvent | 2018-09-21 18:15:52 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | remove deprecated `quiet` keyword | maintainability | We deprecated `quiet=True` in favor of `verbose=False` a long time ago and we announced that this feature would be removed in 0.17.0. However, `quiet` is still accepted in various parts (analysis module).
The `quiet` kwarg and the related code and tests should be finally removed. | True | remove deprecated `quiet` keyword - We deprecated `quiet=True` in favor of `verbose=False` a long time ago and we announced that this feature would be removed in 0.17.0. However, `quiet` is still accepted in various parts (analysis module).
The `quiet` kwarg and the related code and tests should be finally removed. | main | remove deprecated quiet keyword we deprecated quiet true in favor of verbose false a long time ago and we announced that this feature would be removed in however quiet is still accepted in various parts analysis module the quiet kwarg and the related code and tests should be finally removed | 1 |
888 | 4,553,054,288 | IssuesEvent | 2016-09-13 02:19:12 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | gc_storage has misnamed arguments in documentation (or code) | affects_2.0 bug_report cloud gce P3 waiting_on_maintainer | **Issue Type:**
Bug Report
**Ansible Version:**
master
**Summary:**
https://github.com/ansible/ansible-modules-core/blob/f4625a3dd104e245a80ff547deb75f0de880d24f/cloud/google/gc_storage.py
The documentation lists "gcs_access_key" (line 73), but the code which loads this uses "gs_access_key" (line 346, etc). Same for "gcs_secret_key". | True | gc_storage has misnamed arguments in documentation (or code) - **Issue Type:**
Bug Report
**Ansible Version:**
master
**Summary:**
https://github.com/ansible/ansible-modules-core/blob/f4625a3dd104e245a80ff547deb75f0de880d24f/cloud/google/gc_storage.py
The documentation lists "gcs_access_key" (line 73), but the code which loads this uses "gs_access_key" (line 346, etc). Same for "gcs_secret_key". | main | gc storage has misnamed arguments in documentation or code issue type bug report ansible version master summary the documentation lists gcs access key line but the code which loads this uses gs access key line etc same for gcs secret key | 1 |
2,617 | 8,876,650,916 | IssuesEvent | 2019-01-12 17:01:57 | Homebrew/homebrew-cask | https://api.github.com/repos/Homebrew/homebrew-cask | opened | What to do regarding java development casks | awaiting maintainer feedback discussion | Refs https://github.com/Homebrew/homebrew-cask/pull/57374#issuecomment-453760711. Ping @Homebrew/cask.
I feel like the multitude of Java/JVM/JDK casks is too much. Even which software to call `java` is a contentious subject that pops up once in a while. Those casks need non-trivial amounts of spaghetti code (i.e. non-standard-cask-code), there are a bunch of them, all have slightly different reasons for existing, and more than once we’ve had to rethink how they’re organised. We’ve tried to make sense of them yet again recently, but the person doing most of the work in that regard has since left.
Our options, as I see them:
1. Continue adding these as they’re submitted, increasing the mess, eventually having to remove a bunch of them because nobody can make heads or tails of it anymore, or just leave them hanging and broken, as has happened in the past.
2. Have someone (preferably a core maintainer, but not mandatory) commit to understanding all of these different softwares, why they exist, decide which should be included or not, and keep them in good shape. This person will be made a [code owner](https://blog.github.com/2017-07-06-introducing-code-owners/) to deal with all of these. To improve the [lottery factor](https://en.wikipedia.org/wiki/Bus_factor), this person should also keep a document that keeps track of all these different casks, why they exist, and why they need the extra code to be made functional (particularly useful in cases where they differ, if any).
3. Keep the `java` cask and move everything else to yet another tap.
4. Keep the `java` cask and delete everything else. If everyone else needs them, let them make their own tap.
My takes on these suggestions: **1** looks like a recipe for failure and burnout. Both **3** and **4** might restore order, *but* I’d prefer if that decision would be made by someone with the knowledge described in **2**. **4** will piss off more people, and **3** won’t really ease our burden, just shift it. I’ll definitely not be the person to tackle **2**. I dislike java and the stress java casks have given HBC repos in the past and that would make impartiality in my handling of it difficult. However, I would likely help that person tightening the documentation.
Naturally, any other solutions are welcome, or even opinions if you think this ins’t an issue. Right now, I feel this is a problem, even if an unconscious one. Casks with spaghetti code tend to be left longer without being fixed, java being no exception. | True | What to do regarding java development casks - Refs https://github.com/Homebrew/homebrew-cask/pull/57374#issuecomment-453760711. Ping @Homebrew/cask.
I feel like the multitude of Java/JVM/JDK casks is too much. Even which software to call `java` is a contentious subject that pops up once in a while. Those casks need non-trivial amounts of spaghetti code (i.e. non-standard-cask-code), there are a bunch of them, all have slightly different reasons for existing, and more than once we’ve had to rethink how they’re organised. We’ve tried to make sense of them yet again recently, but the person doing most of the work in that regard has since left.
Our options, as I see them:
1. Continue adding these as they’re submitted, increasing the mess, eventually having to remove a bunch of them because nobody can make heads or tails of it anymore, or just leave them hanging and broken, as has happened in the past.
2. Have someone (preferably a core maintainer, but not mandatory) commit to understanding all of these different softwares, why they exist, decide which should be included or not, and keep them in good shape. This person will be made a [code owner](https://blog.github.com/2017-07-06-introducing-code-owners/) to deal with all of these. To improve the [lottery factor](https://en.wikipedia.org/wiki/Bus_factor), this person should also keep a document that keeps track of all these different casks, why they exist, and why they need the extra code to be made functional (particularly useful in cases where they differ, if any).
3. Keep the `java` cask and move everything else to yet another tap.
4. Keep the `java` cask and delete everything else. If everyone else needs them, let them make their own tap.
My takes on these suggestions: **1** looks like a recipe for failure and burnout. Both **3** and **4** might restore order, *but* I’d prefer if that decision would be made by someone with the knowledge described in **2**. **4** will piss off more people, and **3** won’t really ease our burden, just shift it. I’ll definitely not be the person to tackle **2**. I dislike java and the stress java casks have given HBC repos in the past and that would make impartiality in my handling of it difficult. However, I would likely help that person tightening the documentation.
Naturally, any other solutions are welcome, or even opinions if you think this ins’t an issue. Right now, I feel this is a problem, even if an unconscious one. Casks with spaghetti code tend to be left longer without being fixed, java being no exception. | main | what to do regarding java development casks refs ping homebrew cask i feel like the multitude of java jvm jdk casks is too much even which software to call java is a contentious subject that pops up once in a while those casks need non trivial amounts of spaghetti code i e non standard cask code there are a bunch of them all have slightly different reasons for existing and more than once we’ve had to rethink how they’re organised we’ve tried to make sense of them yet again recently but the person doing most of the work in that regard has since left our options as i see them continue adding these as they’re submitted increasing the mess eventually having to remove a bunch of them because nobody can make heads or tails of it anymore or just leave them hanging and broken as has happened in the past have someone preferably a core maintainer but not mandatory commit to understanding all of these different softwares why they exist decide which should be included or not and keep them in good shape this person will be made a to deal with all of these to improve the this person should also keep a document that keeps track of all these different casks why they exist and why they need the extra code to be made functional particularly useful in cases where they differ if any keep the java cask and move everything else to yet another tap keep the java cask and delete everything else if everyone else needs them let them make their own tap my takes on these suggestions looks like a recipe for failure and burnout both and might restore order but i’d prefer if that decision would be made by someone with the knowledge described in will piss off more people and won’t really ease our burden just shift it i’ll definitely not be the person to tackle i dislike java and the stress java casks have given hbc repos in the past and that would make impartiality in my handling of it difficult however i would likely help that person tightening the documentation naturally any other solutions are welcome or even opinions if you think this ins’t an issue right now i feel this is a problem even if an unconscious one casks with spaghetti code tend to be left longer without being fixed java being no exception | 1 |
625 | 4,146,912,046 | IssuesEvent | 2016-06-15 03:12:35 | Microsoft/DirectXTK | https://api.github.com/repos/Microsoft/DirectXTK | opened | Remove VS 2012 adapter code | maintainence | As part of dropping VS 2012 projects and Windows phone 8.0 support, can clean up the following code
* Remove C4005 disable for ``stdint.h`` (workaround for bug with VS 2010 + Windows 7 SDK)
* Remove C4481 disable for "override is an extension" (workaround for VS 2010 bug)
* Remove ``DIRECTX_STD_CALLCONV`` std::function workaround for VS 2012
* Remove ``DIRECTX_CTOR_DEFAULT`` / ``DIRECTX_CTOR_DELETE`` macros and just use =default, =delete directly (VS 2013 or later supports this)
* Remove DirectXMath 3.03 adapters for 3.06 constructs (workaround for Windows 8.0 SDK)
* Make use of ``std::make_unique<>`` (C++14 draft feature supported in VS 2013)
* Remove some guarded code patterns for Windows XP (i.e. functions that were added to Windows Vista)
* Make consistent use of ``= {}`` to initialize memory to zero (C++11 brace init behavior fixed in VS 2013)
* Remove legacy ``WCHAR`` Win32 type and use ``wchar_t`` | True | Remove VS 2012 adapter code - As part of dropping VS 2012 projects and Windows phone 8.0 support, can clean up the following code
* Remove C4005 disable for ``stdint.h`` (workaround for bug with VS 2010 + Windows 7 SDK)
* Remove C4481 disable for "override is an extension" (workaround for VS 2010 bug)
* Remove ``DIRECTX_STD_CALLCONV`` std::function workaround for VS 2012
* Remove ``DIRECTX_CTOR_DEFAULT`` / ``DIRECTX_CTOR_DELETE`` macros and just use =default, =delete directly (VS 2013 or later supports this)
* Remove DirectXMath 3.03 adapters for 3.06 constructs (workaround for Windows 8.0 SDK)
* Make use of ``std::make_unique<>`` (C++14 draft feature supported in VS 2013)
* Remove some guarded code patterns for Windows XP (i.e. functions that were added to Windows Vista)
* Make consistent use of ``= {}`` to initialize memory to zero (C++11 brace init behavior fixed in VS 2013)
* Remove legacy ``WCHAR`` Win32 type and use ``wchar_t`` | main | remove vs adapter code as part of dropping vs projects and windows phone support can clean up the following code remove disable for stdint h workaround for bug with vs windows sdk remove disable for override is an extension workaround for vs bug remove directx std callconv std function workaround for vs remove directx ctor default directx ctor delete macros and just use default delete directly vs or later supports this remove directxmath adapters for constructs workaround for windows sdk make use of std make unique c draft feature supported in vs remove some guarded code patterns for windows xp i e functions that were added to windows vista make consistent use of to initialize memory to zero c brace init behavior fixed in vs remove legacy wchar type and use wchar t | 1 |
52,548 | 10,880,950,570 | IssuesEvent | 2019-11-17 14:40:51 | square/moshi | https://api.github.com/repos/square/moshi | closed | Suppress deprecation warnings if any properties or target classes are deprecated | Kotlin - code gen enhancement | This prevents using `warningsAsError` or `-Werror` in compilations otherwise
Thanks to @erichkleung for the report | 1.0 | Suppress deprecation warnings if any properties or target classes are deprecated - This prevents using `warningsAsError` or `-Werror` in compilations otherwise
Thanks to @erichkleung for the report | non_main | suppress deprecation warnings if any properties or target classes are deprecated this prevents using warningsaserror or werror in compilations otherwise thanks to erichkleung for the report | 0 |
4,488 | 23,375,624,253 | IssuesEvent | 2022-08-11 02:28:06 | restqa/restqa | https://api.github.com/repos/restqa/restqa | closed | [Dashboard] Create new file on the editor section | enhancement wontfix pair with maintainer | Hello 👋,
### 👀 Background
While using RestQA from command line, we can create file on our own since we have access to the file systems.
However if we are using RestQA on dashboard mode we have limited access to the file system.
### ✌️ What is the actual behavior?
Currently there is no way to create a new file from the dashboard.
### 🕵️♀️ How to reproduce the current behavior?
1. Install RestQA `npm i -g @restqa/restqa`
2. Initiate a RestQA project `restqa init`
3. Run the dashboard `restqa dashboard`
4. Access to the dashboard from your browser
5. Select Editor from the left menu.
### 🤞 What is the expected behavior?
In the editor section from the dashboard we should have a way to interact with the file system in order to create a new feature file.
### 😎 Proposed solution.
- [ ] Add a new button on the editor section of the dashboard: `Create a new feature file`
- [ ] Only allow the creation of `.feature` file extension
- [ ] Create an new endpoint on the API layer `POST /api/project/features/{path}` accepting `plain/text` (check swagger on https://localhost:8081/api while dashboard is running)
- [ ] Do not allow the creation of file with empty text
- [ ] Do not allow the creation of file if the dashboard is running on read only mode
- [ ] Find the best UI approach for the button and the file name field.
Cheers. | True | [Dashboard] Create new file on the editor section - Hello 👋,
### 👀 Background
While using RestQA from command line, we can create file on our own since we have access to the file systems.
However if we are using RestQA on dashboard mode we have limited access to the file system.
### ✌️ What is the actual behavior?
Currently there is no way to create a new file from the dashboard.
### 🕵️♀️ How to reproduce the current behavior?
1. Install RestQA `npm i -g @restqa/restqa`
2. Initiate a RestQA project `restqa init`
3. Run the dashboard `restqa dashboard`
4. Access to the dashboard from your browser
5. Select Editor from the left menu.
### 🤞 What is the expected behavior?
In the editor section from the dashboard we should have a way to interact with the file system in order to create a new feature file.
### 😎 Proposed solution.
- [ ] Add a new button on the editor section of the dashboard: `Create a new feature file`
- [ ] Only allow the creation of `.feature` file extension
- [ ] Create an new endpoint on the API layer `POST /api/project/features/{path}` accepting `plain/text` (check swagger on https://localhost:8081/api while dashboard is running)
- [ ] Do not allow the creation of file with empty text
- [ ] Do not allow the creation of file if the dashboard is running on read only mode
- [ ] Find the best UI approach for the button and the file name field.
Cheers. | main | create new file on the editor section hello 👋 👀 background while using restqa from command line we can create file on our own since we have access to the file systems however if we are using restqa on dashboard mode we have limited access to the file system ✌️ what is the actual behavior currently there is no way to create a new file from the dashboard 🕵️♀️ how to reproduce the current behavior install restqa npm i g restqa restqa initiate a restqa project restqa init run the dashboard restqa dashboard access to the dashboard from your browser select editor from the left menu 🤞 what is the expected behavior in the editor section from the dashboard we should have a way to interact with the file system in order to create a new feature file 😎 proposed solution add a new button on the editor section of the dashboard create a new feature file only allow the creation of feature file extension create an new endpoint on the api layer post api project features path accepting plain text check swagger on while dashboard is running do not allow the creation of file with empty text do not allow the creation of file if the dashboard is running on read only mode find the best ui approach for the button and the file name field cheers | 1 |
3,898 | 17,358,152,505 | IssuesEvent | 2021-07-29 16:44:15 | ipfs/go-ipfs | https://api.github.com/repos/ipfs/go-ipfs | opened | Automate copying of signed binaries from dist.ipfs.io/go-ipfs | kind/maintenance need/maintainer-input | ### Checklist
- [X] My issue is specific & actionable.
- [X] I am not suggesting a protocol enhancement.
- [X] I have searched on the [issue tracker](https://github.com/ipfs/go-ipfs/issues?q=is%3Aissue) for my issue.
### Description
## Current state
- macOS signing is being added in https://github.com/ipfs/distributions/pull/367 which means `dist.ipfs.io` is the source of the truth when it comes to official binaries
- We want binaries attached to releases at https://github.com/ipfs/go-ipfs/releases to be signed as well, so they need to come from `/ipns/dist.ipfs.io/go-ipfs/`
## Proposed change
(iiuc) binaries attached to GitHub release are copied from `/ipns/dist.ipfs.io/go-ipfs/` manually by @aschmahmann or @Stebalien – we should automate this.
## Solution design
Create a GitHub Action workflow (triggered manually via `workflow_dispatch` and maybe daily via `cron` to avoid human error) that iterates over every `vA.B.C` release and attaches any missing files from `/ipns/dist.ipfs.io/go-ipfs/VA.B.C` (including manifest and `.cid` and `.sha512` files) if not present yet.
This ensures https://github.com/ipfs/go-ipfs/releases is in sync with dist.ipfs.io. | True | Automate copying of signed binaries from dist.ipfs.io/go-ipfs - ### Checklist
- [X] My issue is specific & actionable.
- [X] I am not suggesting a protocol enhancement.
- [X] I have searched on the [issue tracker](https://github.com/ipfs/go-ipfs/issues?q=is%3Aissue) for my issue.
### Description
## Current state
- macOS signing is being added in https://github.com/ipfs/distributions/pull/367 which means `dist.ipfs.io` is the source of the truth when it comes to official binaries
- We want binaries attached to releases at https://github.com/ipfs/go-ipfs/releases to be signed as well, so they need to come from `/ipns/dist.ipfs.io/go-ipfs/`
## Proposed change
(iiuc) binaries attached to GitHub release are copied from `/ipns/dist.ipfs.io/go-ipfs/` manually by @aschmahmann or @Stebalien – we should automate this.
## Solution design
Create a GitHub Action workflow (triggered manually via `workflow_dispatch` and maybe daily via `cron` to avoid human error) that iterates over every `vA.B.C` release and attaches any missing files from `/ipns/dist.ipfs.io/go-ipfs/VA.B.C` (including manifest and `.cid` and `.sha512` files) if not present yet.
This ensures https://github.com/ipfs/go-ipfs/releases is in sync with dist.ipfs.io. | main | automate copying of signed binaries from dist ipfs io go ipfs checklist my issue is specific actionable i am not suggesting a protocol enhancement i have searched on the for my issue description current state macos signing is being added in which means dist ipfs io is the source of the truth when it comes to official binaries we want binaries attached to releases at to be signed as well so they need to come from ipns dist ipfs io go ipfs proposed change iiuc binaries attached to github release are copied from ipns dist ipfs io go ipfs manually by aschmahmann or stebalien – we should automate this solution design create a github action workflow triggered manually via workflow dispatch and maybe daily via cron to avoid human error that iterates over every va b c release and attaches any missing files from ipns dist ipfs io go ipfs va b c including manifest and cid and files if not present yet this ensures is in sync with dist ipfs io | 1 |
2,924 | 10,424,483,147 | IssuesEvent | 2019-09-16 13:43:26 | chocolatey-community/chocolatey-package-requests | https://api.github.com/repos/chocolatey-community/chocolatey-package-requests | reopened | AMD Graphics Driver | Status: Available For Maintainer(s) | <!--
* Please ensure the package does not already exist in the Chocolatey Community Repository - https://chocolatey.org/packages - by using a relevant search.
* Please ensure there is no existing open package request.
* Please ensure the issue title starts with 'RFP - ' - for example 'RFP - Adobe Reader'
* Please ensure you have both the Software Project URL and the Software Download URL before continuing.
NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication
PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM.
-->
## Checklist
- [x] The package I am requesting does not already exist on https://chocolatey.org/packages;
- [x] There is no open issue for this package;
- [x] The issue title starts 'RFP - ';
- [x] The download URL is public and not locked behind a paywall / login;
## Package Details
Software project URL : https://www.amd.com/en/support
Direct download URL for the software / installer : https://drivers.amd.com/drivers/installer/19.30/beta/radeon-software-adrenalin-2019-19.9.1-minimalsetup-190904_web.exe
Software summary / short description : This software automatically scans for the gpu in the system and downloads and install required drivers.
<!-- ## Package Expectations
Here you can make suggestions on what you would expect the package to do outside of 'installing' - eg. adding icons to the desktop
--> | True | AMD Graphics Driver - <!--
* Please ensure the package does not already exist in the Chocolatey Community Repository - https://chocolatey.org/packages - by using a relevant search.
* Please ensure there is no existing open package request.
* Please ensure the issue title starts with 'RFP - ' - for example 'RFP - Adobe Reader'
* Please ensure you have both the Software Project URL and the Software Download URL before continuing.
NOTE: Keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the Chocolatey community - https://github.com/chocolatey/chocolatey-package-requests/blob/master/README.md#etiquette-regarding-communication
PLEASE REMOVE ALL COMMENTS ONCE YOU HAVE READ THEM.
-->
## Checklist
- [x] The package I am requesting does not already exist on https://chocolatey.org/packages;
- [x] There is no open issue for this package;
- [x] The issue title starts 'RFP - ';
- [x] The download URL is public and not locked behind a paywall / login;
## Package Details
Software project URL : https://www.amd.com/en/support
Direct download URL for the software / installer : https://drivers.amd.com/drivers/installer/19.30/beta/radeon-software-adrenalin-2019-19.9.1-minimalsetup-190904_web.exe
Software summary / short description : This software automatically scans for the gpu in the system and downloads and install required drivers.
<!-- ## Package Expectations
Here you can make suggestions on what you would expect the package to do outside of 'installing' - eg. adding icons to the desktop
--> | main | amd graphics driver please ensure the package does not already exist in the chocolatey community repository by using a relevant search please ensure there is no existing open package request please ensure the issue title starts with rfp for example rfp adobe reader please ensure you have both the software project url and the software download url before continuing note keep in mind we have an etiquette regarding communication that we expect folks to observe when they are looking for support in the chocolatey community please remove all comments once you have read them checklist the package i am requesting does not already exist on there is no open issue for this package the issue title starts rfp the download url is public and not locked behind a paywall login package details software project url direct download url for the software installer software summary short description this software automatically scans for the gpu in the system and downloads and install required drivers package expectations here you can make suggestions on what you would expect the package to do outside of installing eg adding icons to the desktop | 1 |
138,143 | 30,823,150,837 | IssuesEvent | 2023-08-01 17:52:21 | kitsudaiki/Hanami-AI | https://api.github.com/repos/kitsudaiki/Hanami-AI | opened | new core data-structure | Code cleanup / QA Documentation new Feature / improvement Test Usability Breaking: Snapshots Breaking: Templates | ### Description
The actual data-structure changed multiple times over the last years. Some parts are still leftovers from the first PoC's and actual changed were build around this, for example the segments. Especially when in tried to port the core to GPU I ran into multiple problems A new simpler and better structure have to be used to make things faster, better scalable and more compatible for the GPU.
| 1.0 | new core data-structure - ### Description
The actual data-structure changed multiple times over the last years. Some parts are still leftovers from the first PoC's and actual changed were build around this, for example the segments. Especially when in tried to port the core to GPU I ran into multiple problems A new simpler and better structure have to be used to make things faster, better scalable and more compatible for the GPU.
| non_main | new core data structure description the actual data structure changed multiple times over the last years some parts are still leftovers from the first poc s and actual changed were build around this for example the segments especially when in tried to port the core to gpu i ran into multiple problems a new simpler and better structure have to be used to make things faster better scalable and more compatible for the gpu | 0 |
105,278 | 16,637,027,222 | IssuesEvent | 2021-06-04 01:09:12 | veshitala/flask-blogger | https://api.github.com/repos/veshitala/flask-blogger | opened | CVE-2021-28677 (High) detected in Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl | security vulnerability | ## CVE-2021-28677 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to vulnerable library: flask-blogger/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in Pillow before 8.2.0. For EPS data, the readline implementation used in EPSImageFile has to deal with any combination of \r and \n as line endings. It used an accidentally quadratic method of accumulating lines while looking for a line ending. A malicious EPS file could use this to perform a DoS of Pillow in the open phase, before an image was accepted for opening.
<p>Publish Date: 2021-06-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28677>CVE-2021-28677</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28677</a></p>
<p>Release Date: 2021-06-02</p>
<p>Fix Resolution: Pillow - 8.2.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2021-28677 (High) detected in Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl - ## CVE-2021-28677 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</b></p></summary>
<p>Python Imaging Library (Fork)</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl">https://files.pythonhosted.org/packages/0d/f3/421598450cb9503f4565d936860763b5af413a61009d87a5ab1e34139672/Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl</a></p>
<p>Path to vulnerable library: flask-blogger/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Pillow-5.4.1-cp27-cp27mu-manylinux1_x86_64.whl** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in Pillow before 8.2.0. For EPS data, the readline implementation used in EPSImageFile has to deal with any combination of \r and \n as line endings. It used an accidentally quadratic method of accumulating lines while looking for a line ending. A malicious EPS file could use this to perform a DoS of Pillow in the open phase, before an image was accepted for opening.
<p>Publish Date: 2021-06-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-28677>CVE-2021-28677</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28677">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2021-28677</a></p>
<p>Release Date: 2021-06-02</p>
<p>Fix Resolution: Pillow - 8.2.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in pillow whl cve high severity vulnerability vulnerable library pillow whl python imaging library fork library home page a href path to vulnerable library flask blogger requirements txt dependency hierarchy x pillow whl vulnerable library vulnerability details an issue was discovered in pillow before for eps data the readline implementation used in epsimagefile has to deal with any combination of r and n as line endings it used an accidentally quadratic method of accumulating lines while looking for a line ending a malicious eps file could use this to perform a dos of pillow in the open phase before an image was accepted for opening publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution pillow step up your open source security game with whitesource | 0 |
1,084 | 4,932,105,553 | IssuesEvent | 2016-11-28 12:33:03 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Exception when using nxos_snmp_community | affects_2.3 bug_report networking waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
nxos_snmp_community
##### ANSIBLE VERSION
```
ansible 2.3.0~git20161010.03765ba
or
ansible 2.2.0.0-0.1.rc1
(ansible 2.1.2.0-1 does not include the module)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
inventory = ./hosts
gathering = explicit
roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/Roles/roles
private_role_vars = yes
log_path = /var/log/ansible.log
fact_caching = redis
fact_caching_timeout = 86400
retry_files_enabled = False
##### OS / ENVIRONMENT
- host: Ubuntu 16.04 4.4.0
- target: **NX-OSv 7.3(0)D1(1)**
##### SUMMARY
cf. title
##### STEPS TO REPRODUCE
**Inventory ./hosts**:
```
[all:vars]
nms_mgt_ip_address=172.21.100.1
[spines]
NX_OSv_Spine_11 ansible_host=172.21.100.11
NX_OSv_Spine_12 ansible_host=172.21.100.12
```
Structure passed as "**provider**": connections.ssh
Defined in group_vars/nx_osv/connections.yml and a symbolic link in roles/nxos_snmp/defaults points to nx_osv
```
connections
...
nxapi:
transport: nxapi
host: "{{ ansible_host }}"
# ansible_port
port: "{{ http.port }}"
# ansible_user
username: admin
password: xxxxxxxx
# enable_secret_password
auth_pass: xxxxxxxx
# http or https
use_ssl: no
validate_certs: "{{ https.validate_certs }}"
```
**Role**: nxos_snmp:
```
- include_vars: "../defaults/{{ os_family }}/connections.yml"
...
- name: Configuring the ACL for authorized NMS in new SNMPv2c
nxos_config:
provider: "{{ connections.nxapi }}"
parents:
- "ip access-list authorized-snmp-nms"
lines:
- "permit ip host {{ nms_mgt_ip_address }} host {{ ansible_host }} log"
register: result
- name: Configuring RO/RW community string in new IPv4/SNMPv2c
nxos_snmp_community:
provider: "{{ connections.nxapi }}"
community: whatever
access: rw
acl: authorized-snmp-nms
state: present
register: result
```
**Playbook**:
```
- name: Configuring SNMP on NX-OS/NX-OSv
hosts:
- nx_osv
roles:
- nxos_snmp
```
##### EXPECTED RESULTS
The community should be configured without an exception.
##### ACTUAL RESULTS
```
TASK [nxos_snmp : Configuring the ACL for authorized NMS in new SNMPv2c] *******
changed: [NX_OSv_Spine_12] => {"changed": true, "updates": ["ip access-list authorized-snmp-nms", "permit ip host 172.21.100.1 host 172.21.100.12 log"], "warnings": []}
changed: [NX_OSv_Spine_11] => {"changed": true, "updates": ["ip access-list authorized-snmp-nms", "permit ip host 172.21.100.1 host 172.21.100.11 log"], "warnings": []}
...
TASK [nxos_snmp : Configuring RO/RW community string in new IPv4/SNMPv2c] ******
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: TypeError: string indices must be integers
fatal: [NX_OSv_Spine_11]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_55m6Ao/ansible_module_nxos_snmp_community.py\", line 499, in <module>\n main()\n File \"/tmp/ansible_55m6Ao/ansible_module_nxos_snmp_community.py\", line 458, in main\n existing = get_snmp_community(module, community)\n File \"/tmp/ansible_55m6Ao/ansible_module_nxos_snmp_community.py\", line 389, in get_snmp_community\n community_table = data['TABLE_snmp_community']['ROW_snmp_community']\nTypeError: string indices must be integers\n", "module_stdout": "", "msg": "MODULE FAILURE"}
```
No issue when configuring through the CLI:
```
NX_OSv_Spine_11(config)# snmp-server community whatever rw
NX_OSv_Spine_11(config)# snmp-server community whatever use-ipv4acl authorized-snmp-nms
NX_OSv_Spine_11(config)#
```
| True | Exception when using nxos_snmp_community - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
nxos_snmp_community
##### ANSIBLE VERSION
```
ansible 2.3.0~git20161010.03765ba
or
ansible 2.2.0.0-0.1.rc1
(ansible 2.1.2.0-1 does not include the module)
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
inventory = ./hosts
gathering = explicit
roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/Roles/roles
private_role_vars = yes
log_path = /var/log/ansible.log
fact_caching = redis
fact_caching_timeout = 86400
retry_files_enabled = False
##### OS / ENVIRONMENT
- host: Ubuntu 16.04 4.4.0
- target: **NX-OSv 7.3(0)D1(1)**
##### SUMMARY
cf. title
##### STEPS TO REPRODUCE
**Inventory ./hosts**:
```
[all:vars]
nms_mgt_ip_address=172.21.100.1
[spines]
NX_OSv_Spine_11 ansible_host=172.21.100.11
NX_OSv_Spine_12 ansible_host=172.21.100.12
```
Structure passed as "**provider**": connections.ssh
Defined in group_vars/nx_osv/connections.yml and a symbolic link in roles/nxos_snmp/defaults points to nx_osv
```
connections
...
nxapi:
transport: nxapi
host: "{{ ansible_host }}"
# ansible_port
port: "{{ http.port }}"
# ansible_user
username: admin
password: xxxxxxxx
# enable_secret_password
auth_pass: xxxxxxxx
# http or https
use_ssl: no
validate_certs: "{{ https.validate_certs }}"
```
**Role**: nxos_snmp:
```
- include_vars: "../defaults/{{ os_family }}/connections.yml"
...
- name: Configuring the ACL for authorized NMS in new SNMPv2c
nxos_config:
provider: "{{ connections.nxapi }}"
parents:
- "ip access-list authorized-snmp-nms"
lines:
- "permit ip host {{ nms_mgt_ip_address }} host {{ ansible_host }} log"
register: result
- name: Configuring RO/RW community string in new IPv4/SNMPv2c
nxos_snmp_community:
provider: "{{ connections.nxapi }}"
community: whatever
access: rw
acl: authorized-snmp-nms
state: present
register: result
```
**Playbook**:
```
- name: Configuring SNMP on NX-OS/NX-OSv
hosts:
- nx_osv
roles:
- nxos_snmp
```
##### EXPECTED RESULTS
The community should be configured without an exception.
##### ACTUAL RESULTS
```
TASK [nxos_snmp : Configuring the ACL for authorized NMS in new SNMPv2c] *******
changed: [NX_OSv_Spine_12] => {"changed": true, "updates": ["ip access-list authorized-snmp-nms", "permit ip host 172.21.100.1 host 172.21.100.12 log"], "warnings": []}
changed: [NX_OSv_Spine_11] => {"changed": true, "updates": ["ip access-list authorized-snmp-nms", "permit ip host 172.21.100.1 host 172.21.100.11 log"], "warnings": []}
...
TASK [nxos_snmp : Configuring RO/RW community string in new IPv4/SNMPv2c] ******
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: TypeError: string indices must be integers
fatal: [NX_OSv_Spine_11]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/tmp/ansible_55m6Ao/ansible_module_nxos_snmp_community.py\", line 499, in <module>\n main()\n File \"/tmp/ansible_55m6Ao/ansible_module_nxos_snmp_community.py\", line 458, in main\n existing = get_snmp_community(module, community)\n File \"/tmp/ansible_55m6Ao/ansible_module_nxos_snmp_community.py\", line 389, in get_snmp_community\n community_table = data['TABLE_snmp_community']['ROW_snmp_community']\nTypeError: string indices must be integers\n", "module_stdout": "", "msg": "MODULE FAILURE"}
```
No issue when configuring through the CLI:
```
NX_OSv_Spine_11(config)# snmp-server community whatever rw
NX_OSv_Spine_11(config)# snmp-server community whatever use-ipv4acl authorized-snmp-nms
NX_OSv_Spine_11(config)#
```
| main | exception when using nxos snmp community issue type bug report component name nxos snmp community ansible version ansible or ansible ansible does not include the module config file etc ansible ansible cfg configured module search path default w o overrides configuration inventory hosts gathering explicit roles path home actionmystique program files ubuntu ansible roles roles private role vars yes log path var log ansible log fact caching redis fact caching timeout retry files enabled false os environment host ubuntu target nx osv summary cf title steps to reproduce inventory hosts nms mgt ip address nx osv spine ansible host nx osv spine ansible host structure passed as provider connections ssh defined in group vars nx osv connections yml and a symbolic link in roles nxos snmp defaults points to nx osv connections nxapi transport nxapi host ansible host ansible port port http port ansible user username admin password xxxxxxxx enable secret password auth pass xxxxxxxx http or https use ssl no validate certs https validate certs role nxos snmp include vars defaults os family connections yml name configuring the acl for authorized nms in new nxos config provider connections nxapi parents ip access list authorized snmp nms lines permit ip host nms mgt ip address host ansible host log register result name configuring ro rw community string in new nxos snmp community provider connections nxapi community whatever access rw acl authorized snmp nms state present register result playbook name configuring snmp on nx os nx osv hosts nx osv roles nxos snmp expected results the community should be configured without an exception actual results task changed changed true updates warnings changed changed true updates warnings task an exception occurred during task execution to see the full traceback use vvv the error was typeerror string indices must be integers fatal failed changed false failed true module stderr traceback most recent call last n file tmp ansible ansible module nxos snmp community py line in n main n file tmp ansible ansible module nxos snmp community py line in main n existing get snmp community module community n file tmp ansible ansible module nxos snmp community py line in get snmp community n community table data ntypeerror string indices must be integers n module stdout msg module failure no issue when configuring through the cli nx osv spine config snmp server community whatever rw nx osv spine config snmp server community whatever use authorized snmp nms nx osv spine config | 1 |
1,693 | 6,574,204,082 | IssuesEvent | 2017-09-11 11:57:30 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Broken pipe; Sharing connection closed. | affects_2.2 bug_report waiting_on_maintainer | Sorry if there are similar issues but I think I have a specific case.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
command module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = ../my_path/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
```
host_key_checking = False
```
##### OS / ENVIRONMENT
Mac OS 10.11.6
OpenSSH_6.9p1, LibreSSL 2.1.8
##### SUMMARY
This happens sporadically. It appears to happen once and then on subsequent runs the error disappears. After waiting for some time running the play again will trigger the same error and re-running it the error no longer appears.
##### STEPS TO REPRODUCE
Running a command task that runs for 4 hosts and the task is delegated to a single host (adding peers to a cluster master).
##### EXPECTED RESULTS
Expected SSH connection and all well.
##### ACTUAL RESULTS
Fails to connect to delegated host.
```
"msg": "Failed to connect to the host via ssh: OpenSSH_6.9p1, LibreSSL 2.1.8\r\ndebug1: Reading configuration data /Users/jonnymcc/.ssh/config\r\ndebug1: /Users/jonnymcc/.ssh/config line 30: Applying options for *\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 21: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 17898\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 8\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Control master terminated unexpectedly\r\n",
"unreachable": true
```
| True | Broken pipe; Sharing connection closed. - Sorry if there are similar issues but I think I have a specific case.
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
command module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = ../my_path/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
```
host_key_checking = False
```
##### OS / ENVIRONMENT
Mac OS 10.11.6
OpenSSH_6.9p1, LibreSSL 2.1.8
##### SUMMARY
This happens sporadically. It appears to happen once and then on subsequent runs the error disappears. After waiting for some time running the play again will trigger the same error and re-running it the error no longer appears.
##### STEPS TO REPRODUCE
Running a command task that runs for 4 hosts and the task is delegated to a single host (adding peers to a cluster master).
##### EXPECTED RESULTS
Expected SSH connection and all well.
##### ACTUAL RESULTS
Fails to connect to delegated host.
```
"msg": "Failed to connect to the host via ssh: OpenSSH_6.9p1, LibreSSL 2.1.8\r\ndebug1: Reading configuration data /Users/jonnymcc/.ssh/config\r\ndebug1: /Users/jonnymcc/.ssh/config line 30: Applying options for *\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 21: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 17898\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 8\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Control master terminated unexpectedly\r\n",
"unreachable": true
```
| main | broken pipe sharing connection closed sorry if there are similar issues but i think i have a specific case issue type bug report component name command module ansible version ansible config file my path ansible cfg configured module search path default w o overrides configuration host key checking false os environment mac os openssh libressl summary this happens sporadically it appears to happen once and then on subsequent runs the error disappears after waiting for some time running the play again will trigger the same error and re running it the error no longer appears steps to reproduce running a command task that runs for hosts and the task is delegated to a single host adding peers to a cluster master expected results expected ssh connection and all well actual results fails to connect to delegated host msg failed to connect to the host via ssh openssh libressl r reading configuration data users jonnymcc ssh config r users jonnymcc ssh config line applying options for r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r fd setting o nonblock r mux client hello exchange master version r mux client forwards request forwardings local remote r mux client request session entering r mux client request alive entering r mux client request alive done pid r mux client request session session request sent r mux client request session master session id r mux client read packet read header failed broken pipe r control master terminated unexpectedly r n unreachable true | 1 |
1,449 | 6,287,707,199 | IssuesEvent | 2017-07-19 15:27:50 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | Variable expansion has changed when doing 'with_items' with the 'package' module | affects_2.3 bug_report module needs_maintainer support:core | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
package
##### ANSIBLE VERSION
```
ansible 2.3.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
python version = 2.7.13 (default, May 10 2017, 20:04:28) [GCC 6.3.1 20161221 (Red Hat 6.3.1-1)]
```
##### CONFIGURATION
No special configs in the scope of this bug.
##### OS / ENVIRONMENT
Ansible is ran from Fedora 25/26, managing hosts on Fedora 25/26 and Debian.
##### SUMMARY
When I used Ansible 2.2 I used group_vars to have common software package lists which I then used with the module 'package' to install them on RPM and DEB based systems. That way I didn't have to duplicate any package lists for the modules 'dnf' and 'apt' and also didn't have to administrate separate lists.
Now since 2.3 this has changed. It now checks these packages one-by-one when using these vars with the 'with_items' iteration function. I think [this](https://github.com/ansible/ansible/commit/c005ae3cd95bca0267f57e26abcbe0b8e128c8fd) commit is responsible for this new behavior. This kind of makes the 'package' module unfit now to be used as a universal module for all supported package managers... Please revert this change and fix whatever problems the expansion of variables creates, because disabling a feature isn't a fix in my opinion (nor a feature).
##### STEPS TO REPRODUCE
Create a group_var for a role and include e.g. the following.
```
ansible_dependencies_fedora:
- python-dnf
- libselinux-python
- libsemanage-python
- policycoreutils-python
```
Then use this variable in the appropriate role and expand it with the 'with_items' function within the 'package' module.
```
- name: install ansible dependencies
package: name={{ item }} state=present
with_items:
- "{{ ansible_dependencies_fedora }}"
when: ansible_distribution == "Fedora"
tags:
- setup_ansible
- install_packages
```
##### EXPECTED RESULTS
In Ansible 2.2 this worked fine and this is desirable because the advantage of using variable expansion with a universal module for package management is to use a single set of packages for multiple systems with different package dependency solvers.
##### ACTUAL RESULTS
It iterates the items one-by-one instead of treating the expansion as one entire set of items. It's working fine when e.g. using the module 'dnf', but of course stating the explicit module for every package dependency solver results in duplication of this list. Which defeats the purpose of having a universal package module.
I've reverted the change of commit c005ae3cd95bca0267f57e26abcbe0b8e128c8fd manually on my system. Now everything works again as expected, please make this work again as it did in Ansible 2.2!
| True | Variable expansion has changed when doing 'with_items' with the 'package' module - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
package
##### ANSIBLE VERSION
```
ansible 2.3.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
python version = 2.7.13 (default, May 10 2017, 20:04:28) [GCC 6.3.1 20161221 (Red Hat 6.3.1-1)]
```
##### CONFIGURATION
No special configs in the scope of this bug.
##### OS / ENVIRONMENT
Ansible is ran from Fedora 25/26, managing hosts on Fedora 25/26 and Debian.
##### SUMMARY
When I used Ansible 2.2 I used group_vars to have common software package lists which I then used with the module 'package' to install them on RPM and DEB based systems. That way I didn't have to duplicate any package lists for the modules 'dnf' and 'apt' and also didn't have to administrate separate lists.
Now since 2.3 this has changed. It now checks these packages one-by-one when using these vars with the 'with_items' iteration function. I think [this](https://github.com/ansible/ansible/commit/c005ae3cd95bca0267f57e26abcbe0b8e128c8fd) commit is responsible for this new behavior. This kind of makes the 'package' module unfit now to be used as a universal module for all supported package managers... Please revert this change and fix whatever problems the expansion of variables creates, because disabling a feature isn't a fix in my opinion (nor a feature).
##### STEPS TO REPRODUCE
Create a group_var for a role and include e.g. the following.
```
ansible_dependencies_fedora:
- python-dnf
- libselinux-python
- libsemanage-python
- policycoreutils-python
```
Then use this variable in the appropriate role and expand it with the 'with_items' function within the 'package' module.
```
- name: install ansible dependencies
package: name={{ item }} state=present
with_items:
- "{{ ansible_dependencies_fedora }}"
when: ansible_distribution == "Fedora"
tags:
- setup_ansible
- install_packages
```
##### EXPECTED RESULTS
In Ansible 2.2 this worked fine and this is desirable because the advantage of using variable expansion with a universal module for package management is to use a single set of packages for multiple systems with different package dependency solvers.
##### ACTUAL RESULTS
It iterates the items one-by-one instead of treating the expansion as one entire set of items. It's working fine when e.g. using the module 'dnf', but of course stating the explicit module for every package dependency solver results in duplication of this list. Which defeats the purpose of having a universal package module.
I've reverted the change of commit c005ae3cd95bca0267f57e26abcbe0b8e128c8fd manually on my system. Now everything works again as expected, please make this work again as it did in Ansible 2.2!
| main | variable expansion has changed when doing with items with the package module issue type bug report component name package ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides python version default may configuration no special configs in the scope of this bug os environment ansible is ran from fedora managing hosts on fedora and debian summary when i used ansible i used group vars to have common software package lists which i then used with the module package to install them on rpm and deb based systems that way i didn t have to duplicate any package lists for the modules dnf and apt and also didn t have to administrate separate lists now since this has changed it now checks these packages one by one when using these vars with the with items iteration function i think commit is responsible for this new behavior this kind of makes the package module unfit now to be used as a universal module for all supported package managers please revert this change and fix whatever problems the expansion of variables creates because disabling a feature isn t a fix in my opinion nor a feature steps to reproduce create a group var for a role and include e g the following ansible dependencies fedora python dnf libselinux python libsemanage python policycoreutils python then use this variable in the appropriate role and expand it with the with items function within the package module name install ansible dependencies package name item state present with items ansible dependencies fedora when ansible distribution fedora tags setup ansible install packages expected results in ansible this worked fine and this is desirable because the advantage of using variable expansion with a universal module for package management is to use a single set of packages for multiple systems with different package dependency solvers actual results it iterates the items one by one instead of treating the expansion as one entire set of items it s working fine when e g using the module dnf but of course stating the explicit module for every package dependency solver results in duplication of this list which defeats the purpose of having a universal package module i ve reverted the change of commit manually on my system now everything works again as expected please make this work again as it did in ansible | 1 |
2,312 | 8,286,445,647 | IssuesEvent | 2018-09-19 04:52:43 | M4thG33k/TombManyGraves2 | https://api.github.com/repos/M4thG33k/TombManyGraves2 | closed | Mod no longer maintained | No longer maintained | As of July 2018, I (M4thG33k) have decided to stop updating this mod.
The code will, of course, remain available. If someone else wishes to continue the work on the mod, feel free to do so. I'd hate to see the mod vanish completely (even if I don't have a hand in it), but I, sadly, do not have the time to keep up with the game anymore. | True | Mod no longer maintained - As of July 2018, I (M4thG33k) have decided to stop updating this mod.
The code will, of course, remain available. If someone else wishes to continue the work on the mod, feel free to do so. I'd hate to see the mod vanish completely (even if I don't have a hand in it), but I, sadly, do not have the time to keep up with the game anymore. | main | mod no longer maintained as of july i have decided to stop updating this mod the code will of course remain available if someone else wishes to continue the work on the mod feel free to do so i d hate to see the mod vanish completely even if i don t have a hand in it but i sadly do not have the time to keep up with the game anymore | 1 |
1,160 | 2,599,418,135 | IssuesEvent | 2015-02-23 08:19:08 | mlively/Phake | https://api.github.com/repos/mlively/Phake | closed | Problem invoking a method named "eval" on a mock | Defect | First things first, thanks for the hard work you put into `Phake`, it's great!
As for the issue, I'm trying to mock [`phpredis`](https://github.com/nicolasff/phpredis) and then use `verify`/`when` on it's [`eval`](https://github.com/nicolasff/phpredis#eval) method, which leads to some odd behaviour.
I expected it to allow me to `verify`/`when` like I can with it's other methods, but instead it just throws the following exception any time `eval` is called on the mock:
```
RedisException: Redis server went away
```
Which is very strange as that's coming from the actual `phpredis` object, which I haven't instantiated.
I'm creating the mock like this:
```php
$this->redis = Phake::mock('redis');
```
I did dig around the `Phake` codebase and found `eval` in `reservedWords` inside the [`MockClass`](https://github.com/mlively/Phake/blob/master/src/Phake/ClassGenerator/MockClass.php#L109) which is no doubt the issue.
As a quick test, I removed `eval` from the `reservedWords` array, but then invoking `eval` on the mock resulted in this:
```
Phake_ClassGenerator_MockClass::instantiate(): The script tried to execute a method or access a property of an incomplete object. Please ensure that the class definition "redis_PHAKE547e744bc8771" of the object you are trying to operate on was loaded _before_ unserialize() gets called or provide a __autoload() function to load the class definition
```
Is there a work around for this situation or could the code be updated to handle this? | 1.0 | Problem invoking a method named "eval" on a mock - First things first, thanks for the hard work you put into `Phake`, it's great!
As for the issue, I'm trying to mock [`phpredis`](https://github.com/nicolasff/phpredis) and then use `verify`/`when` on it's [`eval`](https://github.com/nicolasff/phpredis#eval) method, which leads to some odd behaviour.
I expected it to allow me to `verify`/`when` like I can with it's other methods, but instead it just throws the following exception any time `eval` is called on the mock:
```
RedisException: Redis server went away
```
Which is very strange as that's coming from the actual `phpredis` object, which I haven't instantiated.
I'm creating the mock like this:
```php
$this->redis = Phake::mock('redis');
```
I did dig around the `Phake` codebase and found `eval` in `reservedWords` inside the [`MockClass`](https://github.com/mlively/Phake/blob/master/src/Phake/ClassGenerator/MockClass.php#L109) which is no doubt the issue.
As a quick test, I removed `eval` from the `reservedWords` array, but then invoking `eval` on the mock resulted in this:
```
Phake_ClassGenerator_MockClass::instantiate(): The script tried to execute a method or access a property of an incomplete object. Please ensure that the class definition "redis_PHAKE547e744bc8771" of the object you are trying to operate on was loaded _before_ unserialize() gets called or provide a __autoload() function to load the class definition
```
Is there a work around for this situation or could the code be updated to handle this? | non_main | problem invoking a method named eval on a mock first things first thanks for the hard work you put into phake it s great as for the issue i m trying to mock and then use verify when on it s method which leads to some odd behaviour i expected it to allow me to verify when like i can with it s other methods but instead it just throws the following exception any time eval is called on the mock redisexception redis server went away which is very strange as that s coming from the actual phpredis object which i haven t instantiated i m creating the mock like this php this redis phake mock redis i did dig around the phake codebase and found eval in reservedwords inside the which is no doubt the issue as a quick test i removed eval from the reservedwords array but then invoking eval on the mock resulted in this phake classgenerator mockclass instantiate the script tried to execute a method or access a property of an incomplete object please ensure that the class definition redis of the object you are trying to operate on was loaded before unserialize gets called or provide a autoload function to load the class definition is there a work around for this situation or could the code be updated to handle this | 0 |
1,643 | 6,572,667,453 | IssuesEvent | 2017-09-11 04:14:24 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | IPA: modules report failed state when no changes are made | affects_2.3 bug_report waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
ipa
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.3.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Linux Mint 18
##### SUMMARY
<!--- Explain the problem briefly -->
Instead of doing nothing ansible modules report failed state when no changes are made.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Run ipa modules imputing cyrilic data multiple times
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Create needed groups
ipa_group:
state: present
name: "{{ item }}"
description: "{{ users_to_add | selectattr('group','equalto',item) | map(attribute='group_desc') | list | first }}"
ipa_host: "{{ global_host }}"
ipa_user: "{{ global_user }}"
ipa_pass: "{{ global_pass }}"
validate_certs: no
user: "{{ users_to_add | selectattr('group','equalto',item) | map(attribute='login') | list }}"
with_items: "{{ users_to_add | map(attribute='group') | list | unique }}"
ignore_errors: true
users_to_add:
- username: Harley Quinn
login: 90987264
cluster_zone: Default
group: mininform
group_desc: "Some random data for description"
personal_data:
- first_name: Harley
last_name: Quinn
mail: harley@gmail.com
telnum: +79788880132
jobtitle: Министр
- username: Vasya Pupkin
login: 77777777
cluster_zone: Default
group: mininform
group_desc: "Some random data for description"
personal_data:
- first_name: Vasya
last_name: Pupkin
mail: vasya@gmail.com
telnum: +7970000805
jobtitle: Зам. Министра
- username: Vladimir Putin
login: 33333333
cluster_zone: Default
group: el_presidente
group_desc: "Русский президент"
personal_data:
- first_name: Владимир
last_name: Путин
mail: putin@gmail.com
telnum: +7970000000
jobtitle: Эль президенте
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Ansible won't report failed state
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
ignoring
failed: [ipa.kr3tec2h.loc] (item=({u'username': u'Medvedev', u'login': 88899823, u'cluster_zone': u'Default', u'group': u'el_presidente', u'group_d
esc': u'\u0410\u0432\u0430\u0442\u0430\u0440 \u043f\u0440\u0435\u0437\u0438\u0434\u0435\u043d\u0442\u0430'}, {u'mail': u'aang@gmail.com', u'first
_name': u'\u0410\u0432\u0430\u0442\u0430\u0440', u'last_name': u'\u0410\u0430\u043d\u0433', u'jobtitle': u'\u0410\u0432\u0430\u0442\u0430\u0440',
u'telnum': 7970000002})) => {"failed": true, "item": [{"cluster_zone": "Default", "group": "el_presidente", "group_desc": "Аватар президента", "
login": 88899823, "username": "Medvedev"}, {"first_name": "Аватар", "jobtitle": "Аватар", "last_name": "Аанг", "mail": "aang@gmail.com", "telnum"
: 7970000002}], "msg": "repsonse user_mod: no modifications to be performed"}
```
| True | IPA: modules report failed state when no changes are made - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
ipa
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.3.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Linux Mint 18
##### SUMMARY
<!--- Explain the problem briefly -->
Instead of doing nothing ansible modules report failed state when no changes are made.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Run ipa modules imputing cyrilic data multiple times
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Create needed groups
ipa_group:
state: present
name: "{{ item }}"
description: "{{ users_to_add | selectattr('group','equalto',item) | map(attribute='group_desc') | list | first }}"
ipa_host: "{{ global_host }}"
ipa_user: "{{ global_user }}"
ipa_pass: "{{ global_pass }}"
validate_certs: no
user: "{{ users_to_add | selectattr('group','equalto',item) | map(attribute='login') | list }}"
with_items: "{{ users_to_add | map(attribute='group') | list | unique }}"
ignore_errors: true
users_to_add:
- username: Harley Quinn
login: 90987264
cluster_zone: Default
group: mininform
group_desc: "Some random data for description"
personal_data:
- first_name: Harley
last_name: Quinn
mail: harley@gmail.com
telnum: +79788880132
jobtitle: Министр
- username: Vasya Pupkin
login: 77777777
cluster_zone: Default
group: mininform
group_desc: "Some random data for description"
personal_data:
- first_name: Vasya
last_name: Pupkin
mail: vasya@gmail.com
telnum: +7970000805
jobtitle: Зам. Министра
- username: Vladimir Putin
login: 33333333
cluster_zone: Default
group: el_presidente
group_desc: "Русский президент"
personal_data:
- first_name: Владимир
last_name: Путин
mail: putin@gmail.com
telnum: +7970000000
jobtitle: Эль президенте
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Ansible won't report failed state
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
ignoring
failed: [ipa.kr3tec2h.loc] (item=({u'username': u'Medvedev', u'login': 88899823, u'cluster_zone': u'Default', u'group': u'el_presidente', u'group_d
esc': u'\u0410\u0432\u0430\u0442\u0430\u0440 \u043f\u0440\u0435\u0437\u0438\u0434\u0435\u043d\u0442\u0430'}, {u'mail': u'aang@gmail.com', u'first
_name': u'\u0410\u0432\u0430\u0442\u0430\u0440', u'last_name': u'\u0410\u0430\u043d\u0433', u'jobtitle': u'\u0410\u0432\u0430\u0442\u0430\u0440',
u'telnum': 7970000002})) => {"failed": true, "item": [{"cluster_zone": "Default", "group": "el_presidente", "group_desc": "Аватар президента", "
login": 88899823, "username": "Medvedev"}, {"first_name": "Аватар", "jobtitle": "Аватар", "last_name": "Аанг", "mail": "aang@gmail.com", "telnum"
: 7970000002}], "msg": "repsonse user_mod: no modifications to be performed"}
```
| main | ipa modules report failed state when no changes are made issue type bug report component name ipa ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific linux mint summary instead of doing nothing ansible modules report failed state when no changes are made steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used run ipa modules imputing cyrilic data multiple times name create needed groups ipa group state present name item description users to add selectattr group equalto item map attribute group desc list first ipa host global host ipa user global user ipa pass global pass validate certs no user users to add selectattr group equalto item map attribute login list with items users to add map attribute group list unique ignore errors true users to add username harley quinn login cluster zone default group mininform group desc some random data for description personal data first name harley last name quinn mail harley gmail com telnum jobtitle министр username vasya pupkin login cluster zone default group mininform group desc some random data for description personal data first name vasya last name pupkin mail vasya gmail com telnum jobtitle зам министра username vladimir putin login cluster zone default group el presidente group desc русский президент personal data first name владимир last name путин mail putin gmail com telnum jobtitle эль президенте expected results ansible won t report failed state actual results ignoring failed item u username u medvedev u login u cluster zone u default u group u el presidente u group d esc u u mail u aang gmail com u first name u u last name u u jobtitle u u telnum failed true item cluster zone default group el presidente group desc аватар президента login username medvedev first name аватар jobtitle аватар last name аанг mail aang gmail com telnum msg repsonse user mod no modifications to be performed | 1 |
730,514 | 25,176,669,726 | IssuesEvent | 2022-11-11 09:52:37 | googleapis/google-api-nodejs-client | https://api.github.com/repos/googleapis/google-api-nodejs-client | opened | Domain-wide Delegation only works when providing credentials via `keyFile` | type: bug priority: p2 | It seems that making API calls (e.g., to the Google Drive API) impersonating another user via Domain-Wide Delegation only works when creating a client using the `keyFile` option (and _not_ when relying on automatically provided auth credentials).
Specifically, I am trying to call the Google Drive API from a Google Cloud Function. I rely on the "Default App Engine Service Account" for authorizing my client:
```javascript
// here, credentials are automatically obtained by getClient:
const auth = await google.auth.getClient({
clientOptions: {
subject: 'email@to.impersonate',
},
scopes: [ /* list of required scopes */ ],
});
const driveAPI = google.drive({ version: 'v3', auth });
// make some requests - however, they are NOT made by 'email@to.impersonate',
// but by the Default App Engine Service Account :(
```
The above code works, but requests are _not_ made as `email@to.impersonate`, but instead as the Default App Engine Service Account. To be clear, I have configured Domain-Wide Delegation for the Default App Engine Service Account in the Google Workspaces Admin UI (https://admin.google.com).
Now, the above code _does work_ once I slightly extend it by providing the Default App Engine Service Account's credentials via the `keyFile` option:
```javascript
// here, credentials are explicitly provided via keyFile:
const auth = await google.auth.getClient({
clientOptions: {
subject: 'email@to.impersonate',
},
keyFile: './path/to/keyFile.json',
scopes: [ /* list of required scopes */ ],
});
const driveAPI = google.drive({ version: 'v3', auth });
// make some requests - they are now made as 'email@to.impersonate' :)
```
Ideally, I'd like to avoid having to manually provide the Default App Engine Service Account credentials using a file.
It seems this issue was mentioned as part of discussions around https://github.com/googleapis/google-auth-library-nodejs/issues/916, specifically in this comment https://github.com/googleapis/google-auth-library-nodejs/issues/916#issuecomment-1080595410.
---
#### Environment details
- OS: Google Cloud functions
- Node.js version: 16
- npm version: 7 (?)
- `googleapis` version: 105.0.0
| 1.0 | Domain-wide Delegation only works when providing credentials via `keyFile` - It seems that making API calls (e.g., to the Google Drive API) impersonating another user via Domain-Wide Delegation only works when creating a client using the `keyFile` option (and _not_ when relying on automatically provided auth credentials).
Specifically, I am trying to call the Google Drive API from a Google Cloud Function. I rely on the "Default App Engine Service Account" for authorizing my client:
```javascript
// here, credentials are automatically obtained by getClient:
const auth = await google.auth.getClient({
clientOptions: {
subject: 'email@to.impersonate',
},
scopes: [ /* list of required scopes */ ],
});
const driveAPI = google.drive({ version: 'v3', auth });
// make some requests - however, they are NOT made by 'email@to.impersonate',
// but by the Default App Engine Service Account :(
```
The above code works, but requests are _not_ made as `email@to.impersonate`, but instead as the Default App Engine Service Account. To be clear, I have configured Domain-Wide Delegation for the Default App Engine Service Account in the Google Workspaces Admin UI (https://admin.google.com).
Now, the above code _does work_ once I slightly extend it by providing the Default App Engine Service Account's credentials via the `keyFile` option:
```javascript
// here, credentials are explicitly provided via keyFile:
const auth = await google.auth.getClient({
clientOptions: {
subject: 'email@to.impersonate',
},
keyFile: './path/to/keyFile.json',
scopes: [ /* list of required scopes */ ],
});
const driveAPI = google.drive({ version: 'v3', auth });
// make some requests - they are now made as 'email@to.impersonate' :)
```
Ideally, I'd like to avoid having to manually provide the Default App Engine Service Account credentials using a file.
It seems this issue was mentioned as part of discussions around https://github.com/googleapis/google-auth-library-nodejs/issues/916, specifically in this comment https://github.com/googleapis/google-auth-library-nodejs/issues/916#issuecomment-1080595410.
---
#### Environment details
- OS: Google Cloud functions
- Node.js version: 16
- npm version: 7 (?)
- `googleapis` version: 105.0.0
| non_main | domain wide delegation only works when providing credentials via keyfile it seems that making api calls e g to the google drive api impersonating another user via domain wide delegation only works when creating a client using the keyfile option and not when relying on automatically provided auth credentials specifically i am trying to call the google drive api from a google cloud function i rely on the default app engine service account for authorizing my client javascript here credentials are automatically obtained by getclient const auth await google auth getclient clientoptions subject email to impersonate scopes const driveapi google drive version auth make some requests however they are not made by email to impersonate but by the default app engine service account the above code works but requests are not made as email to impersonate but instead as the default app engine service account to be clear i have configured domain wide delegation for the default app engine service account in the google workspaces admin ui now the above code does work once i slightly extend it by providing the default app engine service account s credentials via the keyfile option javascript here credentials are explicitly provided via keyfile const auth await google auth getclient clientoptions subject email to impersonate keyfile path to keyfile json scopes const driveapi google drive version auth make some requests they are now made as email to impersonate ideally i d like to avoid having to manually provide the default app engine service account credentials using a file it seems this issue was mentioned as part of discussions around specifically in this comment environment details os google cloud functions node js version npm version googleapis version | 0 |
1,146 | 5,004,924,100 | IssuesEvent | 2016-12-12 08:53:23 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | Ad-hoc shell module freeze and never return to linux prompt | affects_2.1 bug_report waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
shell
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible-2.1.1.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
Nothing configured
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Red Hat Enterprise Linux Server release 6.8 (Santiago)
##### SUMMARY
<!--- Explain the problem briefly -->
When running a remote command with the module shell in ansible, for a list of servers and with sudo,
the command run fine in the remote servers but the ansible command never return to the linux prompt. Crtl+C doesn´t work for cancel. strace of the process shows an infinite loop for the ansible process. The only way to finish the run is to kill the ansible process.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
In a Red Had 6.8 with python 2.6.6 and the ansible version 2.1.1.0 run this with a file filled with several servers:
#> ansible -u <user> -i /tmp/list all -k -s -m shell -a "uptime"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The command never ends. Cannot be stoped/canceled with Crtl+c and have to be killed.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
Using /etc/ansible/ansible.cfg as config file
SSH password:
Loaded callback minimal of type stdout, v2.0
<XXXXX1> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX2> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX3> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX4> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX2> SSH: EXEC sshpass -d20 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r XXXXX2 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208 `" && echo ansible-tmp-1480423293.83-231739456228208="` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208 `" ) && sleep 0'"'"''
<XXXXX3> SSH: EXEC sshpass -d21 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r XXXXX3 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207 `" && echo ansible-tmp-1480423293.84-86346344087207="` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207 `" ) && sleep 0'"'"''
<XXXXX1> SSH: EXEC sshpass -d19 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r XXXXX1 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458 `" && echo ansible-tmp-1480423293.83-223315962231458="` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458 `" ) && sleep 0'"'"''
<XXXXX4> SSH: EXEC sshpass -d18 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r XXXXX4 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320 `" && echo ansible-tmp-1480423293.84-137653613389320="` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320 `" ) && sleep 0'"'"''
<XXXXX2> PUT /tmp/tmpXJBHQT TO /home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208/command
<XXXXX2> SSH: EXEC sshpass -d20 sftp -o BatchMode=no -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r '[XXXXX2]'
<XXXXX3> PUT /tmp/tmpnRVLB6 TO /home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207/command
<XXXXX3> SSH: EXEC sshpass -d21 sftp -o BatchMode=no -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r '[XXXXX3]'
<XXXXX2> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX2> SSH: EXEC sshpass -d20 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r -tt XXXXX2 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-ruzkfjtlzgbojjrcqkwsmvyrqbjlqkvq; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208/command; rm -rf "/home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
<XXXXX3> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX3> SSH: EXEC sshpass -d21 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r -tt XXXXX3 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-imlcvpsrsyiqbutgbwqjncqupwglioyl; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207/command; rm -rf "/home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
XXXXX3 | SUCCESS | rc=0 >>
13:41:33 up 38 days, 9:14, 1 user, load average: 0.00, 0.00, 0.00
<XXXXX4> PUT /tmp/tmpXJBHQT TO /home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320/command
<XXXXX4> SSH: EXEC sshpass -d18 sftp -o BatchMode=no -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r '[XXXXX4]'
XXXXX2 | SUCCESS | rc=0 >>
13:41:34 up 38 days, 8:11, 10 users, load average: 0.00, 0.00, 0.00
<XXXXX4> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX4> SSH: EXEC sshpass -d18 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r -tt XXXXX4 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-hafdetflhotpilnjdzjaagamprxblawg; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320/command; rm -rf "/home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
XXXXX4 | SUCCESS | rc=0 >>
13:41:33 up 26 days, 1:16, 1 user, load average: 0.00, 0.00, 0.00
<XXXXX1> PUT /tmp/tmpXJBHQT TO /home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458/command
<XXXXX1> SSH: EXEC sshpass -d19 sftp -o BatchMode=no -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r '[XXXXX1]'
<XXXXX1> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX1> SSH: EXEC sshpass -d19 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r -tt XXXXX1 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-qodqnawzjywbdlugkfynvmkpczcswokg; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458/command; rm -rf "/home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
XXXXX1 | SUCCESS | rc=0 >>
13:41:45 up 26 days, 1:13, 1 user, load average: 0.00, 0.00, 0.00
```
[Debug_ansible_shell.txt](https://github.com/ansible/ansible-modules-core/files/619617/Debug_ansible_shell.txt) | True | Ad-hoc shell module freeze and never return to linux prompt - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
shell
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible-2.1.1.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
Nothing configured
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Red Hat Enterprise Linux Server release 6.8 (Santiago)
##### SUMMARY
<!--- Explain the problem briefly -->
When running a remote command with the module shell in ansible, for a list of servers and with sudo,
the command run fine in the remote servers but the ansible command never return to the linux prompt. Crtl+C doesn´t work for cancel. strace of the process shows an infinite loop for the ansible process. The only way to finish the run is to kill the ansible process.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
In a Red Had 6.8 with python 2.6.6 and the ansible version 2.1.1.0 run this with a file filled with several servers:
#> ansible -u <user> -i /tmp/list all -k -s -m shell -a "uptime"
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The command never ends. Cannot be stoped/canceled with Crtl+c and have to be killed.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
Using /etc/ansible/ansible.cfg as config file
SSH password:
Loaded callback minimal of type stdout, v2.0
<XXXXX1> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX2> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX3> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX4> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX2> SSH: EXEC sshpass -d20 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r XXXXX2 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208 `" && echo ansible-tmp-1480423293.83-231739456228208="` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208 `" ) && sleep 0'"'"''
<XXXXX3> SSH: EXEC sshpass -d21 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r XXXXX3 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207 `" && echo ansible-tmp-1480423293.84-86346344087207="` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207 `" ) && sleep 0'"'"''
<XXXXX1> SSH: EXEC sshpass -d19 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r XXXXX1 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458 `" && echo ansible-tmp-1480423293.83-223315962231458="` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458 `" ) && sleep 0'"'"''
<XXXXX4> SSH: EXEC sshpass -d18 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r XXXXX4 '/bin/sh -c '"'"'( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320 `" && echo ansible-tmp-1480423293.84-137653613389320="` echo $HOME/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320 `" ) && sleep 0'"'"''
<XXXXX2> PUT /tmp/tmpXJBHQT TO /home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208/command
<XXXXX2> SSH: EXEC sshpass -d20 sftp -o BatchMode=no -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r '[XXXXX2]'
<XXXXX3> PUT /tmp/tmpnRVLB6 TO /home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207/command
<XXXXX3> SSH: EXEC sshpass -d21 sftp -o BatchMode=no -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r '[XXXXX3]'
<XXXXX2> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX2> SSH: EXEC sshpass -d20 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r -tt XXXXX2 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-ruzkfjtlzgbojjrcqkwsmvyrqbjlqkvq; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208/command; rm -rf "/home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-231739456228208/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
<XXXXX3> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX3> SSH: EXEC sshpass -d21 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r -tt XXXXX3 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-imlcvpsrsyiqbutgbwqjncqupwglioyl; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207/command; rm -rf "/home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-86346344087207/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
XXXXX3 | SUCCESS | rc=0 >>
13:41:33 up 38 days, 9:14, 1 user, load average: 0.00, 0.00, 0.00
<XXXXX4> PUT /tmp/tmpXJBHQT TO /home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320/command
<XXXXX4> SSH: EXEC sshpass -d18 sftp -o BatchMode=no -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r '[XXXXX4]'
XXXXX2 | SUCCESS | rc=0 >>
13:41:34 up 38 days, 8:11, 10 users, load average: 0.00, 0.00, 0.00
<XXXXX4> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX4> SSH: EXEC sshpass -d18 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r -tt XXXXX4 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-hafdetflhotpilnjdzjaagamprxblawg; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320/command; rm -rf "/home/<user>/.ansible/tmp/ansible-tmp-1480423293.84-137653613389320/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
XXXXX4 | SUCCESS | rc=0 >>
13:41:33 up 26 days, 1:16, 1 user, load average: 0.00, 0.00, 0.00
<XXXXX1> PUT /tmp/tmpXJBHQT TO /home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458/command
<XXXXX1> SSH: EXEC sshpass -d19 sftp -o BatchMode=no -b - -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r '[XXXXX1]'
<XXXXX1> ESTABLISH SSH CONNECTION FOR USER: <user>
<XXXXX1> SSH: EXEC sshpass -d19 ssh -C -vvv -o ControlMaster=auto -o ControlPersist=60s -o User=<user> -o ConnectTimeout=10 -o ControlPath=/root/.ansible/cp/ansible-ssh-%h-%p-%r -tt XXXXX1 '/bin/sh -c '"'"'sudo -H -S -n -u root /bin/sh -c '"'"'"'"'"'"'"'"'echo BECOME-SUCCESS-qodqnawzjywbdlugkfynvmkpczcswokg; LANG=en_US.UTF-8 LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python /home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458/command; rm -rf "/home/<user>/.ansible/tmp/ansible-tmp-1480423293.83-223315962231458/" > /dev/null 2>&1'"'"'"'"'"'"'"'"' && sleep 0'"'"''
XXXXX1 | SUCCESS | rc=0 >>
13:41:45 up 26 days, 1:13, 1 user, load average: 0.00, 0.00, 0.00
```
[Debug_ansible_shell.txt](https://github.com/ansible/ansible-modules-core/files/619617/Debug_ansible_shell.txt) | main | ad hoc shell module freeze and never return to linux prompt issue type bug report component name shell ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables nothing configured os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific red hat enterprise linux server release santiago summary when running a remote command with the module shell in ansible for a list of servers and with sudo the command run fine in the remote servers but the ansible command never return to the linux prompt crtl c doesn´t work for cancel strace of the process shows an infinite loop for the ansible process the only way to finish the run is to kill the ansible process steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used in a red had with python and the ansible version run this with a file filled with several servers ansible u i tmp list all k s m shell a uptime expected results the command never ends cannot be stoped canceled with crtl c and have to be killed actual results using etc ansible ansible cfg as config file ssh password loaded callback minimal of type stdout establish ssh connection for user establish ssh connection for user establish ssh connection for user establish ssh connection for user ssh exec sshpass ssh c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep ssh exec sshpass ssh c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep ssh exec sshpass ssh c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep ssh exec sshpass ssh c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpxjbhqt to home ansible tmp ansible tmp command ssh exec sshpass sftp o batchmode no b c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r put tmp to home ansible tmp ansible tmp command ssh exec sshpass sftp o batchmode no b c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r establish ssh connection for user ssh exec sshpass ssh c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r tt bin sh c sudo h s n u root bin sh c echo become success ruzkfjtlzgbojjrcqkwsmvyrqbjlqkvq lang en us utf lc all en us utf lc messages en us utf usr bin python home ansible tmp ansible tmp command rm rf home ansible tmp ansible tmp dev null sleep establish ssh connection for user ssh exec sshpass ssh c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r tt bin sh c sudo h s n u root bin sh c echo become success imlcvpsrsyiqbutgbwqjncqupwglioyl lang en us utf lc all en us utf lc messages en us utf usr bin python home ansible tmp ansible tmp command rm rf home ansible tmp ansible tmp dev null sleep success rc up days user load average put tmp tmpxjbhqt to home ansible tmp ansible tmp command ssh exec sshpass sftp o batchmode no b c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r success rc up days users load average establish ssh connection for user ssh exec sshpass ssh c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r tt bin sh c sudo h s n u root bin sh c echo become success hafdetflhotpilnjdzjaagamprxblawg lang en us utf lc all en us utf lc messages en us utf usr bin python home ansible tmp ansible tmp command rm rf home ansible tmp ansible tmp dev null sleep success rc up days user load average put tmp tmpxjbhqt to home ansible tmp ansible tmp command ssh exec sshpass sftp o batchmode no b c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r establish ssh connection for user ssh exec sshpass ssh c vvv o controlmaster auto o controlpersist o user o connecttimeout o controlpath root ansible cp ansible ssh h p r tt bin sh c sudo h s n u root bin sh c echo become success qodqnawzjywbdlugkfynvmkpczcswokg lang en us utf lc all en us utf lc messages en us utf usr bin python home ansible tmp ansible tmp command rm rf home ansible tmp ansible tmp dev null sleep success rc up days user load average | 1 |
5,778 | 30,624,025,967 | IssuesEvent | 2023-07-24 10:11:23 | jupyter-naas/awesome-notebooks | https://api.github.com/repos/jupyter-naas/awesome-notebooks | closed | GitHub - Get DataFrame from project view | templates maintainer | This notebook will show how to return a dataframe from project view using BeautifulSoup. It is usefull for organizations to quickly get data from GitHub project view.
| True | GitHub - Get DataFrame from project view - This notebook will show how to return a dataframe from project view using BeautifulSoup. It is usefull for organizations to quickly get data from GitHub project view.
| main | github get dataframe from project view this notebook will show how to return a dataframe from project view using beautifulsoup it is usefull for organizations to quickly get data from github project view | 1 |
295,742 | 9,100,781,183 | IssuesEvent | 2019-02-20 09:28:02 | jenkins-x/jx | https://api.github.com/repos/jenkins-x/jx | closed | any options dont inicialiaze web pages for home of jenkins x | area/install kind/question priority/awaiting-more-evidence | any options dont inicialiaze web pages for home of jenkins x
I try whit all options like:
jenkins-x chartmuseum
jenkins-x-docker-registry
jenkins-x-mongodb
jenkins-x-monocular-api
jenkins-x-monocular-prender
jenkins-x-monocular-ui
what is te home page that i shoudl view?
thanks for the response ;)
### Summary
### Steps to reproduce the behavior
### Expected behavior
### Actual behavior
### Jx version
The output of `jx version` is:
```
COPY OUTPUT HERE
```
### Jenkins type
<!--
Select which Jenkins installation type are you using.
-->
- [ ] Classic Jenkins
- [ ] Serverless Jenkins
### Kubernetes cluster
<!--
What kind of Kubernetes cluster are you using & how did you create it?
-->
### Operating system / Environment
<!--
In which environment are you running the jx CLI?
-->
| 1.0 | any options dont inicialiaze web pages for home of jenkins x - any options dont inicialiaze web pages for home of jenkins x
I try whit all options like:
jenkins-x chartmuseum
jenkins-x-docker-registry
jenkins-x-mongodb
jenkins-x-monocular-api
jenkins-x-monocular-prender
jenkins-x-monocular-ui
what is te home page that i shoudl view?
thanks for the response ;)
### Summary
### Steps to reproduce the behavior
### Expected behavior
### Actual behavior
### Jx version
The output of `jx version` is:
```
COPY OUTPUT HERE
```
### Jenkins type
<!--
Select which Jenkins installation type are you using.
-->
- [ ] Classic Jenkins
- [ ] Serverless Jenkins
### Kubernetes cluster
<!--
What kind of Kubernetes cluster are you using & how did you create it?
-->
### Operating system / Environment
<!--
In which environment are you running the jx CLI?
-->
| non_main | any options dont inicialiaze web pages for home of jenkins x any options dont inicialiaze web pages for home of jenkins x i try whit all options like jenkins x chartmuseum jenkins x docker registry jenkins x mongodb jenkins x monocular api jenkins x monocular prender jenkins x monocular ui what is te home page that i shoudl view thanks for the response summary steps to reproduce the behavior expected behavior actual behavior jx version the output of jx version is copy output here jenkins type select which jenkins installation type are you using classic jenkins serverless jenkins kubernetes cluster what kind of kubernetes cluster are you using how did you create it operating system environment in which environment are you running the jx cli | 0 |
611,888 | 18,983,824,468 | IssuesEvent | 2021-11-21 11:19:25 | obsproject/obs-websocket | https://api.github.com/repos/obsproject/obs-websocket | opened | Feature Request: Record pausing/resuming events | Low Priority Requires OBS Changes Issue: Feature Request | ### Feature Request Type
RPC Event
### Feature Request Type (Other)
_No response_
### Requested Feature
Add:
- `OBS_WEBSOCKET_OUTPUT_PAUSING`
- `OBS_WEBSOCKET_OUTPUT_RESUMING`
### Requested Feature Usage Scenario
Simply for feature-completeness | 1.0 | Feature Request: Record pausing/resuming events - ### Feature Request Type
RPC Event
### Feature Request Type (Other)
_No response_
### Requested Feature
Add:
- `OBS_WEBSOCKET_OUTPUT_PAUSING`
- `OBS_WEBSOCKET_OUTPUT_RESUMING`
### Requested Feature Usage Scenario
Simply for feature-completeness | non_main | feature request record pausing resuming events feature request type rpc event feature request type other no response requested feature add obs websocket output pausing obs websocket output resuming requested feature usage scenario simply for feature completeness | 0 |
84,948 | 24,473,311,440 | IssuesEvent | 2022-10-07 23:13:40 | micrometer-metrics/micrometer | https://api.github.com/repos/micrometer-metrics/micrometer | closed | Upgrade Spotless plugin to 6.11.0 | build dependency-upgrade | ```bash
./gradlew resolveAndLockAll --no-build-cache --update-locks 'com.diffplug.spotless:spotless-plugin-gradle'
``` | 1.0 | Upgrade Spotless plugin to 6.11.0 - ```bash
./gradlew resolveAndLockAll --no-build-cache --update-locks 'com.diffplug.spotless:spotless-plugin-gradle'
``` | non_main | upgrade spotless plugin to bash gradlew resolveandlockall no build cache update locks com diffplug spotless spotless plugin gradle | 0 |
248,157 | 18,858,041,038 | IssuesEvent | 2021-11-12 09:19:07 | botr99/pe | https://api.github.com/repos/botr99/pe | opened | Main and MainApp hyperlinks redirect to wrong place | type.DocumentationBug severity.Low | 
Click on the hyperlinks.
Expected: Team repo's Main and MainApp
Actual: AB3's Main and MainApp.
<!--session: 1636703842339-a70816ef-1274-413f-99eb-247a66e0f067-->
<!--Version: Web v3.4.1--> | 1.0 | Main and MainApp hyperlinks redirect to wrong place - 
Click on the hyperlinks.
Expected: Team repo's Main and MainApp
Actual: AB3's Main and MainApp.
<!--session: 1636703842339-a70816ef-1274-413f-99eb-247a66e0f067-->
<!--Version: Web v3.4.1--> | non_main | main and mainapp hyperlinks redirect to wrong place click on the hyperlinks expected team repo s main and mainapp actual s main and mainapp | 0 |
5,639 | 28,369,171,422 | IssuesEvent | 2023-04-12 15:43:52 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | Improve TSV serialization for tab and newline characters in cell value when copying cells to the clipboard | type: enhancement work: frontend status: ready restricted: maintainers | ## Current behavior
- You can copy the selected cells in a sheet via Ctrl+C (Cmd+C on Mac OS).
- Mathesar writes one string to the clipboard, a TSV representation of the cells.
- The TSV serialization logic is very simplistic. Columns are separated by tab characters. Rows are separated by newline characters.
- If a cell value contains a tab or newline character, the TSV serialization will not accurately represent the cell values.
You can test this out by setting a cell value to some text which contains a newline. Either use Ctrl+Enter to manually enter a newline, or paste some text into the cell which contains a newline. When rendering the cell in select mode, the newline won't display. It will show again when you enter edit mode on the cell though.
Here I have four cells, and I have newlines entered between "apple" and "banana" and also between "broccoli" and "carrot"

Copying the cells gives this text:
```
apple
banana cherry
asparagus broccoli
carrot
```
If you paste that text into another spreadsheet application, you'll get six cells instead of four. That's not good
## Desired behavior
- Our [Clipboard Interaction specs](https://wiki.mathesar.org/en/design/specs/clipboard-interactions) has a [TSV section](https://wiki.mathesar.org/en/design/specs/clipboard-interactions#tsv-data-representation) with more details about how the serialization logic should work.
- In the above example, copying those four cells should give this text:
```
"apple
banana" cherry
asparagus "broccoli
carrot"
```
The double quotes ensure that the newlines are parsed as cell content rather than row delimiters.
## Implementation
- We also need to deserialize the TSV string during paste, and that logic is more complex. As such, we should see if we can find a 3rd party library to handle the serialization and deserialization for us.
| True | Improve TSV serialization for tab and newline characters in cell value when copying cells to the clipboard - ## Current behavior
- You can copy the selected cells in a sheet via Ctrl+C (Cmd+C on Mac OS).
- Mathesar writes one string to the clipboard, a TSV representation of the cells.
- The TSV serialization logic is very simplistic. Columns are separated by tab characters. Rows are separated by newline characters.
- If a cell value contains a tab or newline character, the TSV serialization will not accurately represent the cell values.
You can test this out by setting a cell value to some text which contains a newline. Either use Ctrl+Enter to manually enter a newline, or paste some text into the cell which contains a newline. When rendering the cell in select mode, the newline won't display. It will show again when you enter edit mode on the cell though.
Here I have four cells, and I have newlines entered between "apple" and "banana" and also between "broccoli" and "carrot"

Copying the cells gives this text:
```
apple
banana cherry
asparagus broccoli
carrot
```
If you paste that text into another spreadsheet application, you'll get six cells instead of four. That's not good
## Desired behavior
- Our [Clipboard Interaction specs](https://wiki.mathesar.org/en/design/specs/clipboard-interactions) has a [TSV section](https://wiki.mathesar.org/en/design/specs/clipboard-interactions#tsv-data-representation) with more details about how the serialization logic should work.
- In the above example, copying those four cells should give this text:
```
"apple
banana" cherry
asparagus "broccoli
carrot"
```
The double quotes ensure that the newlines are parsed as cell content rather than row delimiters.
## Implementation
- We also need to deserialize the TSV string during paste, and that logic is more complex. As such, we should see if we can find a 3rd party library to handle the serialization and deserialization for us.
| main | improve tsv serialization for tab and newline characters in cell value when copying cells to the clipboard current behavior you can copy the selected cells in a sheet via ctrl c cmd c on mac os mathesar writes one string to the clipboard a tsv representation of the cells the tsv serialization logic is very simplistic columns are separated by tab characters rows are separated by newline characters if a cell value contains a tab or newline character the tsv serialization will not accurately represent the cell values you can test this out by setting a cell value to some text which contains a newline either use ctrl enter to manually enter a newline or paste some text into the cell which contains a newline when rendering the cell in select mode the newline won t display it will show again when you enter edit mode on the cell though here i have four cells and i have newlines entered between apple and banana and also between broccoli and carrot copying the cells gives this text apple banana cherry asparagus broccoli carrot if you paste that text into another spreadsheet application you ll get six cells instead of four that s not good desired behavior our has a with more details about how the serialization logic should work in the above example copying those four cells should give this text apple banana cherry asparagus broccoli carrot the double quotes ensure that the newlines are parsed as cell content rather than row delimiters implementation we also need to deserialize the tsv string during paste and that logic is more complex as such we should see if we can find a party library to handle the serialization and deserialization for us | 1 |
345,756 | 24,873,661,401 | IssuesEvent | 2022-10-27 17:09:58 | WFCD/.github | https://api.github.com/repos/WFCD/.github | opened | Update contributing guidelines | documentation | Some things that are missing:
- [ ] general local tests and verifications
- [ ] add commit hooks to all projects for lint, test
- [ ] add commit hooks for conventional commit message
- [ ] add specifications of commit message format
- [ ] add details for which projects use semantic-release
- [ ] add badges for semantic-release and which format they use | 1.0 | Update contributing guidelines - Some things that are missing:
- [ ] general local tests and verifications
- [ ] add commit hooks to all projects for lint, test
- [ ] add commit hooks for conventional commit message
- [ ] add specifications of commit message format
- [ ] add details for which projects use semantic-release
- [ ] add badges for semantic-release and which format they use | non_main | update contributing guidelines some things that are missing general local tests and verifications add commit hooks to all projects for lint test add commit hooks for conventional commit message add specifications of commit message format add details for which projects use semantic release add badges for semantic release and which format they use | 0 |
1,094 | 4,955,656,077 | IssuesEvent | 2016-12-01 21:03:50 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | jenkins_plugin giving Python type error | affects_2.3 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`jenkins_plugin`
##### ANSIBLE VERSION
```
2.3.0
```
##### CONFIGURATION
No settings have been changed
##### OS / ENVIRONMENT
* Execution node : Ubuntu 16.10
* Managed node : CentOS7
##### SUMMARY
I am using the `jenkins_plugin` module to install plugins to a jenkins server I am deploying. The specific task is :
```
- name: install Jenkins plugins
jenkins_plugin:
owner: jenkins
group: jenkins
name: "{{ item }}"
state: latest
with_dependencies: yes
with_items: "{{ jenkins_plugins }}"
tags:
- plugins
```
where `plugins` is a list of plugins which we need (see https://github.com/AAROC/CODE-RADE-jenkins-role/blob/master/vars/main.yml)
##### STEPS TO REPRODUCE
The [CODE-RADE](https://galaxy.ansible.com/AAROC/CODE-RADE-jenkins-role/) role is used in [the playbook](https://github.com/AAROC/DevOps/blob/master/Ansible/CODE-RADE-jenkins-server.yml) . The role could be used to reproduce the error, but there are some specific things about our setup which we can't share freely.
Running the playbook results in the error shown in "ACUTAL RESULTS" **on subsequent runs**. Note that the plugins were installed correctly the first time, and the error arose _after a removal of the plugins_ (`rm -rf /var/lib/jenkins/plugins`) and a re-run of the playbook to check consistency and idempotency.
##### EXPECTED RESULTS
The plugin is installed correctly.
##### ACTUAL RESULTS
Each plugin task in the list failed :
```
failed: [196.21.182.7] (item=project-build-times) => {
"failed": true,
"invocation": {
"module_name": "jenkins_plugin"
},
"item": "project-build-times",
"module_stderr": "Shared connection to 196.21.182.7 closed.\r\n",
"module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_vYIwwC/ansible_module_jenkins_plugin.py\", line 829, in <module>\r\n main()\r\n File \"/tmp/ansible_vYIwwC/ansible_module_jenkins_plugin.py\", line 812, in main\r\n changed = jp.install()\r\n File \"/tmp/ansible_vYIwwC/ansible_module_jenkins_plugin.py\", line 481, in install\r\n self._write_file(plugin_file, r)\r\n File \"/tmp/ansible_vYIwwC/ansible_module_jenkins_plugin.py\", line 646, in _write_file\r\n fd = open(tmp_f, 'wb')\r\nTypeError: coercing to Unicode: need string or buffer, tuple found\r\n",
"msg": "MODULE FAILURE"
}
```
The main issue it seems is : **TypeError: coercing to Unicode: need string or buffer, tuple found** | True | jenkins_plugin giving Python type error - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
`jenkins_plugin`
##### ANSIBLE VERSION
```
2.3.0
```
##### CONFIGURATION
No settings have been changed
##### OS / ENVIRONMENT
* Execution node : Ubuntu 16.10
* Managed node : CentOS7
##### SUMMARY
I am using the `jenkins_plugin` module to install plugins to a jenkins server I am deploying. The specific task is :
```
- name: install Jenkins plugins
jenkins_plugin:
owner: jenkins
group: jenkins
name: "{{ item }}"
state: latest
with_dependencies: yes
with_items: "{{ jenkins_plugins }}"
tags:
- plugins
```
where `plugins` is a list of plugins which we need (see https://github.com/AAROC/CODE-RADE-jenkins-role/blob/master/vars/main.yml)
##### STEPS TO REPRODUCE
The [CODE-RADE](https://galaxy.ansible.com/AAROC/CODE-RADE-jenkins-role/) role is used in [the playbook](https://github.com/AAROC/DevOps/blob/master/Ansible/CODE-RADE-jenkins-server.yml) . The role could be used to reproduce the error, but there are some specific things about our setup which we can't share freely.
Running the playbook results in the error shown in "ACUTAL RESULTS" **on subsequent runs**. Note that the plugins were installed correctly the first time, and the error arose _after a removal of the plugins_ (`rm -rf /var/lib/jenkins/plugins`) and a re-run of the playbook to check consistency and idempotency.
##### EXPECTED RESULTS
The plugin is installed correctly.
##### ACTUAL RESULTS
Each plugin task in the list failed :
```
failed: [196.21.182.7] (item=project-build-times) => {
"failed": true,
"invocation": {
"module_name": "jenkins_plugin"
},
"item": "project-build-times",
"module_stderr": "Shared connection to 196.21.182.7 closed.\r\n",
"module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_vYIwwC/ansible_module_jenkins_plugin.py\", line 829, in <module>\r\n main()\r\n File \"/tmp/ansible_vYIwwC/ansible_module_jenkins_plugin.py\", line 812, in main\r\n changed = jp.install()\r\n File \"/tmp/ansible_vYIwwC/ansible_module_jenkins_plugin.py\", line 481, in install\r\n self._write_file(plugin_file, r)\r\n File \"/tmp/ansible_vYIwwC/ansible_module_jenkins_plugin.py\", line 646, in _write_file\r\n fd = open(tmp_f, 'wb')\r\nTypeError: coercing to Unicode: need string or buffer, tuple found\r\n",
"msg": "MODULE FAILURE"
}
```
The main issue it seems is : **TypeError: coercing to Unicode: need string or buffer, tuple found** | main | jenkins plugin giving python type error issue type bug report component name jenkins plugin ansible version configuration no settings have been changed os environment execution node ubuntu managed node summary i am using the jenkins plugin module to install plugins to a jenkins server i am deploying the specific task is name install jenkins plugins jenkins plugin owner jenkins group jenkins name item state latest with dependencies yes with items jenkins plugins tags plugins where plugins is a list of plugins which we need see steps to reproduce the role is used in the role could be used to reproduce the error but there are some specific things about our setup which we can t share freely running the playbook results in the error shown in acutal results on subsequent runs note that the plugins were installed correctly the first time and the error arose after a removal of the plugins rm rf var lib jenkins plugins and a re run of the playbook to check consistency and idempotency expected results the plugin is installed correctly actual results each plugin task in the list failed failed item project build times failed true invocation module name jenkins plugin item project build times module stderr shared connection to closed r n module stdout traceback most recent call last r n file tmp ansible vyiwwc ansible module jenkins plugin py line in r n main r n file tmp ansible vyiwwc ansible module jenkins plugin py line in main r n changed jp install r n file tmp ansible vyiwwc ansible module jenkins plugin py line in install r n self write file plugin file r r n file tmp ansible vyiwwc ansible module jenkins plugin py line in write file r n fd open tmp f wb r ntypeerror coercing to unicode need string or buffer tuple found r n msg module failure the main issue it seems is typeerror coercing to unicode need string or buffer tuple found | 1 |
5,089 | 26,005,683,296 | IssuesEvent | 2022-12-20 19:07:23 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | opened | AttributeError attname from the records endpoint | type: enhancement work: frontend status: ready restricted: maintainers | I have been getting this API error intermittently when making a GET request to the records endpoint. It always happens after making some rather significant DDL request, like importing a table or extracting columns -- though I can't reproduce it 100% of the time.
<details>
<summary>Traceback</summary>
```
Environment:
Request Method: GET
Request URL: http://localhost:8000/api/db/v0/tables/1824/records/?limit=500&offset=0
Django Version: 3.1.14
Python Version: 3.9.16
Installed Applications:
['django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'django_filters',
'django_property_filter',
'mathesar']
Installed Middleware:
['django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
'mathesar.middleware.CursorClosedHandlerMiddleware',
'mathesar.middleware.PasswordChangeNeededMiddleware',
'django_userforeignkey.middleware.UserForeignKeyMiddleware',
'django_request_cache.middleware.RequestCacheMiddleware']
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__
return self._index[key]
The above exception ('attname') was the direct cause of the following exception:
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view
return view_func(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view
return self.dispatch(request, *args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch
response = self.handle_exception(exc)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception
response = exception_handler(exc, context)
File "/code/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler
raise exc
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch
response = handler(request, *args, **kwargs)
File "/code/mathesar/api/db/viewsets/records.py", line 55, in list
column_names_to_ids = table.get_column_name_id_bidirectional_map()
File "/code/mathesar/models/base.py", line 554, in get_column_name_id_bidirectional_map
columns_map = bidict({column.name: column.id for column in columns})
File "/usr/local/lib/python3.9/site-packages/django/db/models/query.py", line 287, in __iter__
self._fetch_all()
File "/code/mathesar/utils/prefetch.py", line 158, in _fetch_all
prefetcher.fetch(obj_list, name, self.model, forwarders)
File "/code/mathesar/utils/prefetch.py", line 270, in fetch
related_data = self.filter(data_mapping.keys(), data_mapping.values())
File "/code/mathesar/models/base.py", line 225, in filter
return get_map_of_attnum_to_column_name(
File "/code/db/columns/operations/select.py", line 91, in get_map_of_attnum_to_column_name
triples_of_col_info = _get_triples_of_column_name_and_attnum_and_table_oid(
File "/code/db/columns/operations/select.py", line 104, in _get_triples_of_column_name_and_attnum_and_table_oid
statement = _statement_for_triples_of_column_name_and_attnum_and_table_oid(
File "/code/db/columns/operations/select.py", line 180, in _statement_for_triples_of_column_name_and_attnum_and_table_oid
sel = select(pg_attribute.c.attname, pg_attribute.c.attnum, pg_attribute.c.attrelid)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__
util.raise_(AttributeError(key), replace_context=err)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
Exception Type: AttributeError at /api/db/v0/tables/1824/records/
Exception Value: attname
```
</details>
https://user-images.githubusercontent.com/42411/208746906-86d06e41-18d0-45da-8d39-32aa8fddecba.mp4
| True | AttributeError attname from the records endpoint - I have been getting this API error intermittently when making a GET request to the records endpoint. It always happens after making some rather significant DDL request, like importing a table or extracting columns -- though I can't reproduce it 100% of the time.
<details>
<summary>Traceback</summary>
```
Environment:
Request Method: GET
Request URL: http://localhost:8000/api/db/v0/tables/1824/records/?limit=500&offset=0
Django Version: 3.1.14
Python Version: 3.9.16
Installed Applications:
['django.contrib.admin',
'django.contrib.auth',
'django.contrib.contenttypes',
'django.contrib.sessions',
'django.contrib.messages',
'django.contrib.staticfiles',
'rest_framework',
'django_filters',
'django_property_filter',
'mathesar']
Installed Middleware:
['django.middleware.security.SecurityMiddleware',
'django.contrib.sessions.middleware.SessionMiddleware',
'django.middleware.common.CommonMiddleware',
'django.middleware.csrf.CsrfViewMiddleware',
'django.contrib.auth.middleware.AuthenticationMiddleware',
'django.contrib.messages.middleware.MessageMiddleware',
'django.middleware.clickjacking.XFrameOptionsMiddleware',
'mathesar.middleware.CursorClosedHandlerMiddleware',
'mathesar.middleware.PasswordChangeNeededMiddleware',
'django_userforeignkey.middleware.UserForeignKeyMiddleware',
'django_request_cache.middleware.RequestCacheMiddleware']
Traceback (most recent call last):
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1167, in __getattr__
return self._index[key]
The above exception ('attname') was the direct cause of the following exception:
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/exception.py", line 47, in inner
response = get_response(request)
File "/usr/local/lib/python3.9/site-packages/django/core/handlers/base.py", line 181, in _get_response
response = wrapped_callback(request, *callback_args, **callback_kwargs)
File "/usr/local/lib/python3.9/site-packages/django/views/decorators/csrf.py", line 54, in wrapped_view
return view_func(*args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/viewsets.py", line 125, in view
return self.dispatch(request, *args, **kwargs)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 509, in dispatch
response = self.handle_exception(exc)
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 466, in handle_exception
response = exception_handler(exc, context)
File "/code/mathesar/exception_handlers.py", line 59, in mathesar_exception_handler
raise exc
File "/usr/local/lib/python3.9/site-packages/rest_framework/views.py", line 506, in dispatch
response = handler(request, *args, **kwargs)
File "/code/mathesar/api/db/viewsets/records.py", line 55, in list
column_names_to_ids = table.get_column_name_id_bidirectional_map()
File "/code/mathesar/models/base.py", line 554, in get_column_name_id_bidirectional_map
columns_map = bidict({column.name: column.id for column in columns})
File "/usr/local/lib/python3.9/site-packages/django/db/models/query.py", line 287, in __iter__
self._fetch_all()
File "/code/mathesar/utils/prefetch.py", line 158, in _fetch_all
prefetcher.fetch(obj_list, name, self.model, forwarders)
File "/code/mathesar/utils/prefetch.py", line 270, in fetch
related_data = self.filter(data_mapping.keys(), data_mapping.values())
File "/code/mathesar/models/base.py", line 225, in filter
return get_map_of_attnum_to_column_name(
File "/code/db/columns/operations/select.py", line 91, in get_map_of_attnum_to_column_name
triples_of_col_info = _get_triples_of_column_name_and_attnum_and_table_oid(
File "/code/db/columns/operations/select.py", line 104, in _get_triples_of_column_name_and_attnum_and_table_oid
statement = _statement_for_triples_of_column_name_and_attnum_and_table_oid(
File "/code/db/columns/operations/select.py", line 180, in _statement_for_triples_of_column_name_and_attnum_and_table_oid
sel = select(pg_attribute.c.attname, pg_attribute.c.attnum, pg_attribute.c.attrelid)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/sql/base.py", line 1169, in __getattr__
util.raise_(AttributeError(key), replace_context=err)
File "/usr/local/lib/python3.9/site-packages/sqlalchemy/util/compat.py", line 207, in raise_
raise exception
Exception Type: AttributeError at /api/db/v0/tables/1824/records/
Exception Value: attname
```
</details>
https://user-images.githubusercontent.com/42411/208746906-86d06e41-18d0-45da-8d39-32aa8fddecba.mp4
| main | attributeerror attname from the records endpoint i have been getting this api error intermittently when making a get request to the records endpoint it always happens after making some rather significant ddl request like importing a table or extracting columns though i can t reproduce it of the time traceback environment request method get request url django version python version installed applications django contrib admin django contrib auth django contrib contenttypes django contrib sessions django contrib messages django contrib staticfiles rest framework django filters django property filter mathesar installed middleware django middleware security securitymiddleware django contrib sessions middleware sessionmiddleware django middleware common commonmiddleware django middleware csrf csrfviewmiddleware django contrib auth middleware authenticationmiddleware django contrib messages middleware messagemiddleware django middleware clickjacking xframeoptionsmiddleware mathesar middleware cursorclosedhandlermiddleware mathesar middleware passwordchangeneededmiddleware django userforeignkey middleware userforeignkeymiddleware django request cache middleware requestcachemiddleware traceback most recent call last file usr local lib site packages sqlalchemy sql base py line in getattr return self index the above exception attname was the direct cause of the following exception file usr local lib site packages django core handlers exception py line in inner response get response request file usr local lib site packages django core handlers base py line in get response response wrapped callback request callback args callback kwargs file usr local lib site packages django views decorators csrf py line in wrapped view return view func args kwargs file usr local lib site packages rest framework viewsets py line in view return self dispatch request args kwargs file usr local lib site packages rest framework views py line in dispatch response self handle exception exc file usr local lib site packages rest framework views py line in handle exception response exception handler exc context file code mathesar exception handlers py line in mathesar exception handler raise exc file usr local lib site packages rest framework views py line in dispatch response handler request args kwargs file code mathesar api db viewsets records py line in list column names to ids table get column name id bidirectional map file code mathesar models base py line in get column name id bidirectional map columns map bidict column name column id for column in columns file usr local lib site packages django db models query py line in iter self fetch all file code mathesar utils prefetch py line in fetch all prefetcher fetch obj list name self model forwarders file code mathesar utils prefetch py line in fetch related data self filter data mapping keys data mapping values file code mathesar models base py line in filter return get map of attnum to column name file code db columns operations select py line in get map of attnum to column name triples of col info get triples of column name and attnum and table oid file code db columns operations select py line in get triples of column name and attnum and table oid statement statement for triples of column name and attnum and table oid file code db columns operations select py line in statement for triples of column name and attnum and table oid sel select pg attribute c attname pg attribute c attnum pg attribute c attrelid file usr local lib site packages sqlalchemy sql base py line in getattr util raise attributeerror key replace context err file usr local lib site packages sqlalchemy util compat py line in raise raise exception exception type attributeerror at api db tables records exception value attname | 1 |
747,508 | 26,088,839,162 | IssuesEvent | 2022-12-26 08:14:37 | mypfinance/mypfinance | https://api.github.com/repos/mypfinance/mypfinance | opened | Change accessToken storage on frontend | high priority research & development frontend backend | Currently, we store the access token in local storage and the frontend extracts it from there.
This is considered bad practice and easily hackable by a malicious user.
The goal of this issue is to refactor the application to store the accessToken in a cookie as opposed to in local storage! And to implement it in the auth-header so that all API calls work! | 1.0 | Change accessToken storage on frontend - Currently, we store the access token in local storage and the frontend extracts it from there.
This is considered bad practice and easily hackable by a malicious user.
The goal of this issue is to refactor the application to store the accessToken in a cookie as opposed to in local storage! And to implement it in the auth-header so that all API calls work! | non_main | change accesstoken storage on frontend currently we store the access token in local storage and the frontend extracts it from there this is considered bad practice and easily hackable by a malicious user the goal of this issue is to refactor the application to store the accesstoken in a cookie as opposed to in local storage and to implement it in the auth header so that all api calls work | 0 |
1,432 | 6,220,962,974 | IssuesEvent | 2017-07-10 02:49:42 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | define a MDAnalysis.analysis user interface | API Component-Analysis maintainability parallelization policy usability | **tl,dr**: _The `MDAnalysis.analysis` modules do not have a unified user interface, which is bad for users and bad for developers. We need to come up with a set of rules describing the analysis modules' user interface._
# Divergent user interface in `MDAnalysis.analysis`
The `MDAnalysis.analysis` (and `MDAnalysis.visualization`) modules collect various kinds of "tools" to analyze simulations; in some sense, they are responsible for the "analysis" in MDAnalysis. However, while we have been pretty stringent about what our API inside the "core" should look like, we have been [much less prescriptive with analysis](/MDAnalysis/mdanalysis/wiki/Style-Guide). To a good degree this reflects the reality that code is mainly contributed by researchers that wrote something to get a particularly job done and then realized that it might be usable for the rest of the community – of course, that's exactly what we want for a user-driven open source project! On the other hand, there seems to be a growing feeling among developers that we should have a more uniform interface to the analysis tools as well.
Ideally, all our analysis tools should have a common philosophy and share a common set of options. Being able to use different analysis tools "out of the box" once you have a basic understanding of how it works makes for a good overall user experience.
From the developer side, it promotes code re-use and modularization with subsequent improvements in testing coverage and code reliability.
# Using `AnalysisBase`
@richardjgowers wrote a prototype `MDAnalysis.analysis.base.AnalysisBase` class and in recent code reviews on contributions to analysis we have been pushing for basing analysis code on this class. But in discussions such as on PR #708 it is becoming clear that we should settle on what we expect of the analysis code to do, not the least so that developers, who spend a significant amount of time just cleaning up old mess when they implement code fixes and add new features, know where to set priorities and what is expected of them.
`AnalysisBase` outlines how to structure typical frame-based analysis but it does not really say (yet) what a user should be able to expect from analysis tools.
# Different models for the user interface
Some of the current analysis tools come with additional methods to immediately plot data, many are able to write intermediate and final data to a file for reuse (and perhaps are even able to re-read the file, and perform plotting without needing to reanalyze a trajectory), most of the store results as numpy arrays in an attribute `results` (often a `dict` for multiple results).
A more purist approach is to just return final data structures, throw away intermediates and do not even store final results, and let the user do all downstream processing and plotting.
I can see four broadly defined models how we could handle the user interface:
1. ~~**Anarchy**~~ : Do not prescribe any user interface and let each analysis tool writer decide what's best and most appropriate.
2. ~~**Minimalist**~~ (or developer-friendly?):
- class-based: prescribe use of `AnalysisBase` and stipulate that `run()` returns all computed data.
- function-based: only provide a function that performs the data reduction and returns all computed data
3. ~~**Baroque**~~ (or user-friendly?): prescribe `AnalysisBase` with additional features, for example (discussion needed!)
- `plot()` for a simple visualization of the data (remember that sometimes data plotting is pretty involved, see for instance, `PSAnalysis.plot()`!)
- `save()` to store data as a file on disk
- `to_df()` to return as a `pandas.DataFrame`
For any of these features you need to store the data inside the class somewhere.
4. ~~**Eclecticism**~~ : Somewhere between Minimalist and Baroque with some features mandatory and other optional (but which ones?).
5. **Bauhaus** (the emerging consensus from the discussion below: a cohesive reduction to a common set of functional elements together with minimalist inspirations.)
- Prescribe `AnalysisBase` with a common feature set (like **Baroque**) with the goal to have a unified and utilitarian interface.
- Provide the core numerical analysis (especially for frame-based analysis) as a function in the same module. This function is used in the `_single_frame()` method.
_Feel free to edit/add to the list._
# What do we need?
I am asking @MDAnalysis/coredevs (and anyone else interested) to chime in with opinions on what to do. The final outcome of this issue should be a consensus on set of rules (or a statement of the absence of rules for option 1) on how code in analysis ought to interface with the user. These rules will then become part of the [Developer Guide](https://github.com/MDAnalysis/mdanalysis/wiki/Guide-for-Developers).
# History
- 2016-02-15: added to list of options the **Bauhaus** model (best-of-both-worlds) as emerging from discussions below and added note to **minimalist** along what @jandom originally proposed.
- 2016-02-22: consensus appears to be to go for the **Bauhaus** design model
| True | define a MDAnalysis.analysis user interface - **tl,dr**: _The `MDAnalysis.analysis` modules do not have a unified user interface, which is bad for users and bad for developers. We need to come up with a set of rules describing the analysis modules' user interface._
# Divergent user interface in `MDAnalysis.analysis`
The `MDAnalysis.analysis` (and `MDAnalysis.visualization`) modules collect various kinds of "tools" to analyze simulations; in some sense, they are responsible for the "analysis" in MDAnalysis. However, while we have been pretty stringent about what our API inside the "core" should look like, we have been [much less prescriptive with analysis](/MDAnalysis/mdanalysis/wiki/Style-Guide). To a good degree this reflects the reality that code is mainly contributed by researchers that wrote something to get a particularly job done and then realized that it might be usable for the rest of the community – of course, that's exactly what we want for a user-driven open source project! On the other hand, there seems to be a growing feeling among developers that we should have a more uniform interface to the analysis tools as well.
Ideally, all our analysis tools should have a common philosophy and share a common set of options. Being able to use different analysis tools "out of the box" once you have a basic understanding of how it works makes for a good overall user experience.
From the developer side, it promotes code re-use and modularization with subsequent improvements in testing coverage and code reliability.
# Using `AnalysisBase`
@richardjgowers wrote a prototype `MDAnalysis.analysis.base.AnalysisBase` class and in recent code reviews on contributions to analysis we have been pushing for basing analysis code on this class. But in discussions such as on PR #708 it is becoming clear that we should settle on what we expect of the analysis code to do, not the least so that developers, who spend a significant amount of time just cleaning up old mess when they implement code fixes and add new features, know where to set priorities and what is expected of them.
`AnalysisBase` outlines how to structure typical frame-based analysis but it does not really say (yet) what a user should be able to expect from analysis tools.
# Different models for the user interface
Some of the current analysis tools come with additional methods to immediately plot data, many are able to write intermediate and final data to a file for reuse (and perhaps are even able to re-read the file, and perform plotting without needing to reanalyze a trajectory), most of the store results as numpy arrays in an attribute `results` (often a `dict` for multiple results).
A more purist approach is to just return final data structures, throw away intermediates and do not even store final results, and let the user do all downstream processing and plotting.
I can see four broadly defined models how we could handle the user interface:
1. ~~**Anarchy**~~ : Do not prescribe any user interface and let each analysis tool writer decide what's best and most appropriate.
2. ~~**Minimalist**~~ (or developer-friendly?):
- class-based: prescribe use of `AnalysisBase` and stipulate that `run()` returns all computed data.
- function-based: only provide a function that performs the data reduction and returns all computed data
3. ~~**Baroque**~~ (or user-friendly?): prescribe `AnalysisBase` with additional features, for example (discussion needed!)
- `plot()` for a simple visualization of the data (remember that sometimes data plotting is pretty involved, see for instance, `PSAnalysis.plot()`!)
- `save()` to store data as a file on disk
- `to_df()` to return as a `pandas.DataFrame`
For any of these features you need to store the data inside the class somewhere.
4. ~~**Eclecticism**~~ : Somewhere between Minimalist and Baroque with some features mandatory and other optional (but which ones?).
5. **Bauhaus** (the emerging consensus from the discussion below: a cohesive reduction to a common set of functional elements together with minimalist inspirations.)
- Prescribe `AnalysisBase` with a common feature set (like **Baroque**) with the goal to have a unified and utilitarian interface.
- Provide the core numerical analysis (especially for frame-based analysis) as a function in the same module. This function is used in the `_single_frame()` method.
_Feel free to edit/add to the list._
# What do we need?
I am asking @MDAnalysis/coredevs (and anyone else interested) to chime in with opinions on what to do. The final outcome of this issue should be a consensus on set of rules (or a statement of the absence of rules for option 1) on how code in analysis ought to interface with the user. These rules will then become part of the [Developer Guide](https://github.com/MDAnalysis/mdanalysis/wiki/Guide-for-Developers).
# History
- 2016-02-15: added to list of options the **Bauhaus** model (best-of-both-worlds) as emerging from discussions below and added note to **minimalist** along what @jandom originally proposed.
- 2016-02-22: consensus appears to be to go for the **Bauhaus** design model
| main | define a mdanalysis analysis user interface tl dr the mdanalysis analysis modules do not have a unified user interface which is bad for users and bad for developers we need to come up with a set of rules describing the analysis modules user interface divergent user interface in mdanalysis analysis the mdanalysis analysis and mdanalysis visualization modules collect various kinds of tools to analyze simulations in some sense they are responsible for the analysis in mdanalysis however while we have been pretty stringent about what our api inside the core should look like we have been mdanalysis mdanalysis wiki style guide to a good degree this reflects the reality that code is mainly contributed by researchers that wrote something to get a particularly job done and then realized that it might be usable for the rest of the community – of course that s exactly what we want for a user driven open source project on the other hand there seems to be a growing feeling among developers that we should have a more uniform interface to the analysis tools as well ideally all our analysis tools should have a common philosophy and share a common set of options being able to use different analysis tools out of the box once you have a basic understanding of how it works makes for a good overall user experience from the developer side it promotes code re use and modularization with subsequent improvements in testing coverage and code reliability using analysisbase richardjgowers wrote a prototype mdanalysis analysis base analysisbase class and in recent code reviews on contributions to analysis we have been pushing for basing analysis code on this class but in discussions such as on pr it is becoming clear that we should settle on what we expect of the analysis code to do not the least so that developers who spend a significant amount of time just cleaning up old mess when they implement code fixes and add new features know where to set priorities and what is expected of them analysisbase outlines how to structure typical frame based analysis but it does not really say yet what a user should be able to expect from analysis tools different models for the user interface some of the current analysis tools come with additional methods to immediately plot data many are able to write intermediate and final data to a file for reuse and perhaps are even able to re read the file and perform plotting without needing to reanalyze a trajectory most of the store results as numpy arrays in an attribute results often a dict for multiple results a more purist approach is to just return final data structures throw away intermediates and do not even store final results and let the user do all downstream processing and plotting i can see four broadly defined models how we could handle the user interface anarchy do not prescribe any user interface and let each analysis tool writer decide what s best and most appropriate minimalist or developer friendly class based prescribe use of analysisbase and stipulate that run returns all computed data function based only provide a function that performs the data reduction and returns all computed data baroque or user friendly prescribe analysisbase with additional features for example discussion needed plot for a simple visualization of the data remember that sometimes data plotting is pretty involved see for instance psanalysis plot save to store data as a file on disk to df to return as a pandas dataframe for any of these features you need to store the data inside the class somewhere eclecticism somewhere between minimalist and baroque with some features mandatory and other optional but which ones bauhaus the emerging consensus from the discussion below a cohesive reduction to a common set of functional elements together with minimalist inspirations prescribe analysisbase with a common feature set like baroque with the goal to have a unified and utilitarian interface provide the core numerical analysis especially for frame based analysis as a function in the same module this function is used in the single frame method feel free to edit add to the list what do we need i am asking mdanalysis coredevs and anyone else interested to chime in with opinions on what to do the final outcome of this issue should be a consensus on set of rules or a statement of the absence of rules for option on how code in analysis ought to interface with the user these rules will then become part of the history added to list of options the bauhaus model best of both worlds as emerging from discussions below and added note to minimalist along what jandom originally proposed consensus appears to be to go for the bauhaus design model | 1 |
252 | 3,005,398,567 | IssuesEvent | 2015-07-26 21:31:07 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | opened | Allow relative paths in `uninstall :delete` | core enhancement waiting-maintainer-feedback | Currently, relative paths are only allowed in `zap :delete`. I don't see a need for this restriction, and the consensus in #11530 seems to be that we should get rid of it.
Before I dive into the code though, I'd like to confirm with @caskroom/maintainers that this is the right thing to do. | True | Allow relative paths in `uninstall :delete` - Currently, relative paths are only allowed in `zap :delete`. I don't see a need for this restriction, and the consensus in #11530 seems to be that we should get rid of it.
Before I dive into the code though, I'd like to confirm with @caskroom/maintainers that this is the right thing to do. | main | allow relative paths in uninstall delete currently relative paths are only allowed in zap delete i don t see a need for this restriction and the consensus in seems to be that we should get rid of it before i dive into the code though i d like to confirm with caskroom maintainers that this is the right thing to do | 1 |
538,078 | 15,762,065,456 | IssuesEvent | 2021-03-31 10:39:46 | wso2/product-microgateway | https://api.github.com/repos/wso2/product-microgateway | closed | Enforcer environment variables can not resolve part of the string | Priority/Normal Type/Bug envoy-gw | ### Description:
If the enforcer related config has a config like below, it will not resolve from environment variable properly.
`foo=test:$env{FOO_VAL}`
It will resolve foo value with env variable value from FOO_VAL and removes the `test` part from the config, which is incorrect.
### Steps to reproduce:
### Affected Product Version:
4.0.0-m9
### Environment details (with versions):
- OS:
- Client:
- Env (Docker/K8s):
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
| 1.0 | Enforcer environment variables can not resolve part of the string - ### Description:
If the enforcer related config has a config like below, it will not resolve from environment variable properly.
`foo=test:$env{FOO_VAL}`
It will resolve foo value with env variable value from FOO_VAL and removes the `test` part from the config, which is incorrect.
### Steps to reproduce:
### Affected Product Version:
4.0.0-m9
### Environment details (with versions):
- OS:
- Client:
- Env (Docker/K8s):
---
### Optional Fields
#### Related Issues:
<!-- Any related issues from this/other repositories-->
#### Suggested Labels:
<!--Only to be used by non-members-->
#### Suggested Assignees:
<!--Only to be used by non-members-->
| non_main | enforcer environment variables can not resolve part of the string description if the enforcer related config has a config like below it will not resolve from environment variable properly foo test env foo val it will resolve foo value with env variable value from foo val and removes the test part from the config which is incorrect steps to reproduce affected product version environment details with versions os client env docker optional fields related issues suggested labels suggested assignees | 0 |
3,451 | 13,215,591,691 | IssuesEvent | 2020-08-17 00:12:37 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | Timeout configuration for rax_dns_record | affects_2.1 bot_closed cloud collection collection:community.general feature module needs_collection_redirect needs_maintainer needs_triage support:community | From @enekofb on 2016-08-26T13:57:58Z
##### ISSUE TYPE
- Feature
##### COMPONENT NAME
rax_dns_record
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
nothing changed
##### OS / ENVIRONMENT
mac osx 10.9.5
##### SUMMARY
Trying to add a dns record through ansible, rax_dns_record hits pyrax default dns timeout resulting in
"msg": "The API call to '/domains/4789666/records' did not complete after 5 seconds."
So even if 5 seconds should be enough in the majority of the cases, it would be great to be able to set a new timeout.
Copied from original issue: ansible/ansible-modules-core#4558
| True | Timeout configuration for rax_dns_record - From @enekofb on 2016-08-26T13:57:58Z
##### ISSUE TYPE
- Feature
##### COMPONENT NAME
rax_dns_record
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file =
configured module search path = Default w/o overrides
```
##### CONFIGURATION
nothing changed
##### OS / ENVIRONMENT
mac osx 10.9.5
##### SUMMARY
Trying to add a dns record through ansible, rax_dns_record hits pyrax default dns timeout resulting in
"msg": "The API call to '/domains/4789666/records' did not complete after 5 seconds."
So even if 5 seconds should be enough in the majority of the cases, it would be great to be able to set a new timeout.
Copied from original issue: ansible/ansible-modules-core#4558
| main | timeout configuration for rax dns record from enekofb on issue type feature component name rax dns record ansible version ansible config file configured module search path default w o overrides configuration nothing changed os environment mac osx summary trying to add a dns record through ansible rax dns record hits pyrax default dns timeout resulting in msg the api call to domains records did not complete after seconds so even if seconds should be enough in the majority of the cases it would be great to be able to set a new timeout copied from original issue ansible ansible modules core | 1 |
3,207 | 12,236,610,589 | IssuesEvent | 2020-05-04 16:37:40 | RockefellerArchiveCenter/aurora | https://api.github.com/repos/RockefellerArchiveCenter/aurora | closed | Refactor custom Javascript on appraisal queue page | maintainability | ## Describe the bug
The custom JS on the appraisal queue page is really messy and difficult to decipher. There is a lot of repetition which could be factored out to function.
## To reproduce
Look at the code!
## Expected behavior
I should be able to quickly understand what the JS is and does without having to read through a lot of similar but slightly different functions.
## Impact on your work
This makes updating and maintaining this code over the long haul quite difficult. | True | Refactor custom Javascript on appraisal queue page - ## Describe the bug
The custom JS on the appraisal queue page is really messy and difficult to decipher. There is a lot of repetition which could be factored out to function.
## To reproduce
Look at the code!
## Expected behavior
I should be able to quickly understand what the JS is and does without having to read through a lot of similar but slightly different functions.
## Impact on your work
This makes updating and maintaining this code over the long haul quite difficult. | main | refactor custom javascript on appraisal queue page describe the bug the custom js on the appraisal queue page is really messy and difficult to decipher there is a lot of repetition which could be factored out to function to reproduce look at the code expected behavior i should be able to quickly understand what the js is and does without having to read through a lot of similar but slightly different functions impact on your work this makes updating and maintaining this code over the long haul quite difficult | 1 |
360,613 | 25,299,514,458 | IssuesEvent | 2022-11-17 09:39:39 | score-spec/docs | https://api.github.com/repos/score-spec/docs | opened | 404 link | documentation | **Link to document**
URL: https://score.dev/manifesto
**Describe issue**
Hi!
The README.md links to the manifesto but it's a 404
**Suggested fix**
I don't know the correct link | 1.0 | 404 link - **Link to document**
URL: https://score.dev/manifesto
**Describe issue**
Hi!
The README.md links to the manifesto but it's a 404
**Suggested fix**
I don't know the correct link | non_main | link link to document url describe issue hi the readme md links to the manifesto but it s a suggested fix i don t know the correct link | 0 |
69,664 | 7,157,072,710 | IssuesEvent | 2018-01-26 18:33:17 | rampantpixels/rpmalloc | https://api.github.com/repos/rampantpixels/rpmalloc | closed | out of bounds read in _memory_allocate_large_from_heap | bug need test case | https://github.com/rampantpixels/rpmalloc/blob/82fbcc0af8f46d640024dbb5f612d7b7f0552ec5/rpmalloc/rpmalloc.c#L915
...there is wrong testing order. (idx < LARGE_CLASS_COUNT) have to be before !heap->large_cache[idx] because if idx == LARGE_CLASS_COUNT (after iteration) then you read from undefined memory (out-of-bounds). | 1.0 | out of bounds read in _memory_allocate_large_from_heap - https://github.com/rampantpixels/rpmalloc/blob/82fbcc0af8f46d640024dbb5f612d7b7f0552ec5/rpmalloc/rpmalloc.c#L915
...there is wrong testing order. (idx < LARGE_CLASS_COUNT) have to be before !heap->large_cache[idx] because if idx == LARGE_CLASS_COUNT (after iteration) then you read from undefined memory (out-of-bounds). | non_main | out of bounds read in memory allocate large from heap there is wrong testing order idx large cache because if idx large class count after iteration then you read from undefined memory out of bounds | 0 |
363,643 | 25,460,055,765 | IssuesEvent | 2022-11-24 17:44:19 | joshuacc/ahkpm | https://api.github.com/repos/joshuacc/ahkpm | opened | Rework the Quick Start on ahkpm.dev | documentation | It's a bit out of date, and not as reader friendly as it could be. | 1.0 | Rework the Quick Start on ahkpm.dev - It's a bit out of date, and not as reader friendly as it could be. | non_main | rework the quick start on ahkpm dev it s a bit out of date and not as reader friendly as it could be | 0 |
182,821 | 30,988,612,587 | IssuesEvent | 2023-08-09 01:21:49 | Lvght/e-andifes | https://api.github.com/repos/Lvght/e-andifes | closed | Prototipar tela de Gestão de Usuários | design usuário 1 | - [x] Tela de visualizar todos os usuários
- [x] Tela de detalhes para cada usuário
- [x] Deleção ou desativação
- [x] Edição
- [x] Tela de cadastro de usuário | 1.0 | Prototipar tela de Gestão de Usuários - - [x] Tela de visualizar todos os usuários
- [x] Tela de detalhes para cada usuário
- [x] Deleção ou desativação
- [x] Edição
- [x] Tela de cadastro de usuário | non_main | prototipar tela de gestão de usuários tela de visualizar todos os usuários tela de detalhes para cada usuário deleção ou desativação edição tela de cadastro de usuário | 0 |
1,801 | 6,575,923,403 | IssuesEvent | 2017-09-11 17:51:04 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | mysql_user host_all arguments conversion string formatting error | affects_2.1 bug_report waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
mysql_user
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ubuntu Trusty/Xenial
##### SUMMARY
Attempting to use the mysql_user module to change the password for the root user while using the 'host_all: yes' instead of iterating all hosts with 'host: "{{ item }}"' + 'with_items' results in error "not all arguments converted during string formatting"
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Change password
mysql_user:
login_user: "root"
login_password: "abc"
name: "root"
password: "xyz"
host_all: yes
```
or
```
- name: Change password
mysql_user:
name: "root"
password: "xyz"
host_all: yes
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The mysql_user should be able to iterate through mysql hosts and change the password of the root user against hosts in mysql similar to when using the 'host: "{{ item }}"' + 'with_items'
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
fatal: [192.168.1.127]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "mysql_user"}, "module_stderr": "OpenSSH_6.6.1, OpenSSL 1.0.1f 6 Jan 2014\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 27404\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\nShared connection to 192.168.1.127 closed.\r\n", "module_stdout": "\r\nTraceback (most recent call last):\r\n File \"/tmp/ansible_BJCDub/ansible_module_mysql_user.py\", line 582, in <module>\r\n main()\r\n File \"/tmp/ansible_BJCDub/ansible_module_mysql_user.py\", line 554, in main\r\n if user_exists(cursor, user, host, host_all):\r\n File \"/tmp/ansible_BJCDub/ansible_module_mysql_user.py\", line 216, in user_exists\r\n cursor.execute(\"SELECT count(*) FROM user WHERE user = %s\", user)\r\n File \"/usr/local/lib/python2.7/dist-packages/MySQLdb/cursors.py\", line 187, in execute\r\n query = query % tuple([db.literal(item) for item in args])\r\nTypeError: not all arguments converted during string formatting\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
| True | mysql_user host_all arguments conversion string formatting error - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
mysql_user
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Ubuntu Trusty/Xenial
##### SUMMARY
Attempting to use the mysql_user module to change the password for the root user while using the 'host_all: yes' instead of iterating all hosts with 'host: "{{ item }}"' + 'with_items' results in error "not all arguments converted during string formatting"
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Change password
mysql_user:
login_user: "root"
login_password: "abc"
name: "root"
password: "xyz"
host_all: yes
```
or
```
- name: Change password
mysql_user:
name: "root"
password: "xyz"
host_all: yes
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The mysql_user should be able to iterate through mysql hosts and change the password of the root user against hosts in mysql similar to when using the 'host: "{{ item }}"' + 'with_items'
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
fatal: [192.168.1.127]: FAILED! => {"changed": false, "failed": true, "invocation": {"module_name": "mysql_user"}, "module_stderr": "OpenSSH_6.6.1, OpenSSL 1.0.1f 6 Jan 2014\r\ndebug1: Reading configuration data /etc/ssh/ssh_config\r\ndebug1: /etc/ssh/ssh_config line 19: Applying options for *\r\ndebug1: auto-mux: Trying existing master\r\ndebug2: fd 3 setting O_NONBLOCK\r\ndebug2: mux_client_hello_exchange: master version 4\r\ndebug3: mux_client_forwards: request forwardings: 0 local, 0 remote\r\ndebug3: mux_client_request_session: entering\r\ndebug3: mux_client_request_alive: entering\r\ndebug3: mux_client_request_alive: done pid = 27404\r\ndebug3: mux_client_request_session: session request sent\r\ndebug1: mux_client_request_session: master session id: 2\r\ndebug3: mux_client_read_packet: read header failed: Broken pipe\r\ndebug2: Received exit status from master 0\r\nShared connection to 192.168.1.127 closed.\r\n", "module_stdout": "\r\nTraceback (most recent call last):\r\n File \"/tmp/ansible_BJCDub/ansible_module_mysql_user.py\", line 582, in <module>\r\n main()\r\n File \"/tmp/ansible_BJCDub/ansible_module_mysql_user.py\", line 554, in main\r\n if user_exists(cursor, user, host, host_all):\r\n File \"/tmp/ansible_BJCDub/ansible_module_mysql_user.py\", line 216, in user_exists\r\n cursor.execute(\"SELECT count(*) FROM user WHERE user = %s\", user)\r\n File \"/usr/local/lib/python2.7/dist-packages/MySQLdb/cursors.py\", line 187, in execute\r\n query = query % tuple([db.literal(item) for item in args])\r\nTypeError: not all arguments converted during string formatting\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
| main | mysql user host all arguments conversion string formatting error issue type bug report component name mysql user ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific ubuntu trusty xenial summary attempting to use the mysql user module to change the password for the root user while using the host all yes instead of iterating all hosts with host item with items results in error not all arguments converted during string formatting steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name change password mysql user login user root login password abc name root password xyz host all yes or name change password mysql user name root password xyz host all yes expected results the mysql user should be able to iterate through mysql hosts and change the password of the root user against hosts in mysql similar to when using the host item with items actual results fatal failed changed false failed true invocation module name mysql user module stderr openssh openssl jan r reading configuration data etc ssh ssh config r etc ssh ssh config line applying options for r auto mux trying existing master r fd setting o nonblock r mux client hello exchange master version r mux client forwards request forwardings local remote r mux client request session entering r mux client request alive entering r mux client request alive done pid r mux client request session session request sent r mux client request session master session id r mux client read packet read header failed broken pipe r received exit status from master r nshared connection to closed r n module stdout r ntraceback most recent call last r n file tmp ansible bjcdub ansible module mysql user py line in r n main r n file tmp ansible bjcdub ansible module mysql user py line in main r n if user exists cursor user host host all r n file tmp ansible bjcdub ansible module mysql user py line in user exists r n cursor execute select count from user where user s user r n file usr local lib dist packages mysqldb cursors py line in execute r n query query tuple r ntypeerror not all arguments converted during string formatting r n msg module failure parsed false | 1 |
823 | 4,445,331,355 | IssuesEvent | 2016-08-20 01:07:03 | OpenLightingProject/ola | https://api.github.com/repos/OpenLightingProject/ola | closed | gcc6 build issues | bug Maintainability OpSys-Linux | Hi,
`std::auto_ptr` is deprecated in the latest C++ standard. As of version 6, GCC emits a warning for this. Due to ola's default of enabling `-Werror` for builds, this causes builds with gcc6 to fail.
By adding `-Wno-error=deprecated-declarations`, this can be worked around without having to disable *all* warning->error conversions, but perhaps ola should consider migrating away from deprecated classes. | True | gcc6 build issues - Hi,
`std::auto_ptr` is deprecated in the latest C++ standard. As of version 6, GCC emits a warning for this. Due to ola's default of enabling `-Werror` for builds, this causes builds with gcc6 to fail.
By adding `-Wno-error=deprecated-declarations`, this can be worked around without having to disable *all* warning->error conversions, but perhaps ola should consider migrating away from deprecated classes. | main | build issues hi std auto ptr is deprecated in the latest c standard as of version gcc emits a warning for this due to ola s default of enabling werror for builds this causes builds with to fail by adding wno error deprecated declarations this can be worked around without having to disable all warning error conversions but perhaps ola should consider migrating away from deprecated classes | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.