Unnamed: 0 int64 1 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 3 438 | labels stringlengths 4 308 | body stringlengths 7 254k | index stringclasses 7 values | text_combine stringlengths 96 254k | label stringclasses 2 values | text stringlengths 96 246k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
3,964 | 18,106,937,091 | IssuesEvent | 2021-09-22 20:13:56 | bcurran3/ChocolateyPackages | https://api.github.com/repos/bcurran3/ChocolateyPackages | closed | Update TeamSpeak Server install to version 3.13.6 | Done !LOOKING_FOR_NEW_MAINTAINER! | I used your package to install a teamspeak server the other day. While it did install as expected, the server fails to launch properly since it is an outdated version.
If you could update to 3.13.6, i (and the 2 other people on earth that still use teamspeak) would greatly appreciate it!
Here are the links and checksums that you should need to update the install package:
`$url = 'https://files.teamspeak-services.com/releases/server/3.13.6/teamspeak3-server_win32-3.13.6.zip'
$checksum = 'f52cddfd2f1297756afa12d758718b3ee34b88c0e7962fb75985877f480e0d71'
$url64 = 'https://files.teamspeak-services.com/releases/server/3.13.6/teamspeak3-server_win64-3.13.6.zip'
$checksum64 = '6ffe2fad26cd68cbd8c23f7361469f7eaba9b267f320daf9cce90ed28e4b3bca'`
Thank you! | True | Update TeamSpeak Server install to version 3.13.6 - I used your package to install a teamspeak server the other day. While it did install as expected, the server fails to launch properly since it is an outdated version.
If you could update to 3.13.6, i (and the 2 other people on earth that still use teamspeak) would greatly appreciate it!
Here are the links and checksums that you should need to update the install package:
`$url = 'https://files.teamspeak-services.com/releases/server/3.13.6/teamspeak3-server_win32-3.13.6.zip'
$checksum = 'f52cddfd2f1297756afa12d758718b3ee34b88c0e7962fb75985877f480e0d71'
$url64 = 'https://files.teamspeak-services.com/releases/server/3.13.6/teamspeak3-server_win64-3.13.6.zip'
$checksum64 = '6ffe2fad26cd68cbd8c23f7361469f7eaba9b267f320daf9cce90ed28e4b3bca'`
Thank you! | main | update teamspeak server install to version i used your package to install a teamspeak server the other day while it did install as expected the server fails to launch properly since it is an outdated version if you could update to i and the other people on earth that still use teamspeak would greatly appreciate it here are the links and checksums that you should need to update the install package url checksum thank you | 1 |
2,491 | 8,650,706,893 | IssuesEvent | 2018-11-26 23:37:04 | react-navigation/react-navigation | https://api.github.com/repos/react-navigation/react-navigation | closed | createBottomTabNavigator gets route-name instead of tabBarLabel in some case | needs action from maintainer | Hi, having some troubles translating tabs using react-i18next I tested with your example at
https://github.com/i18next/react-i18next/tree/master/example/react-native-expo
I changed the stack used in wrappedStack (App.js) from StackNavigator to BottomTabNavigator using createBottomTabNavigator.
As long as all routes are defined directly with screens all works as expected. When for example defining first route/tab using a stack (to navigate lets say from a settings tab to a screen changing language) and second route with screen then the "stack-route" is labeled with the route-name instead of using tabBarLabel or title at least.
What I am doing wrong? Is there a way to change (translate) the route-name? I don't think so because of stack handling.
react-i18next (issue #484) saying: "overall i guess it's more a problem / limitation of react-navigation"
so I opened this issue.
I made changes to app.js, i18n.js, Home.js, Page2.js and added Language.js.
### Current Behavior
tabbar is labeled with route-name when route is a stack
### Expected Behavior
tabbar labeled with tabBarLabel as it is when route is a screen
### How to reproduce
https://github.com/HaJo10/react-i18next
### Your Environment
| software | version
| ---------------- | -------
| react-navigation | ^2.11.2
| react-native | 0.55.4
| node | 10.4.0
| npm or yarn | npm 6.2.0
| True | createBottomTabNavigator gets route-name instead of tabBarLabel in some case - Hi, having some troubles translating tabs using react-i18next I tested with your example at
https://github.com/i18next/react-i18next/tree/master/example/react-native-expo
I changed the stack used in wrappedStack (App.js) from StackNavigator to BottomTabNavigator using createBottomTabNavigator.
As long as all routes are defined directly with screens all works as expected. When for example defining first route/tab using a stack (to navigate lets say from a settings tab to a screen changing language) and second route with screen then the "stack-route" is labeled with the route-name instead of using tabBarLabel or title at least.
What I am doing wrong? Is there a way to change (translate) the route-name? I don't think so because of stack handling.
react-i18next (issue #484) saying: "overall i guess it's more a problem / limitation of react-navigation"
so I opened this issue.
I made changes to app.js, i18n.js, Home.js, Page2.js and added Language.js.
### Current Behavior
tabbar is labeled with route-name when route is a stack
### Expected Behavior
tabbar labeled with tabBarLabel as it is when route is a screen
### How to reproduce
https://github.com/HaJo10/react-i18next
### Your Environment
| software | version
| ---------------- | -------
| react-navigation | ^2.11.2
| react-native | 0.55.4
| node | 10.4.0
| npm or yarn | npm 6.2.0
| main | createbottomtabnavigator gets route name instead of tabbarlabel in some case hi having some troubles translating tabs using react i tested with your example at i changed the stack used in wrappedstack app js from stacknavigator to bottomtabnavigator using createbottomtabnavigator as long as all routes are defined directly with screens all works as expected when for example defining first route tab using a stack to navigate lets say from a settings tab to a screen changing language and second route with screen then the stack route is labeled with the route name instead of using tabbarlabel or title at least what i am doing wrong is there a way to change translate the route name i don t think so because of stack handling react issue saying overall i guess it s more a problem limitation of react navigation so i opened this issue i made changes to app js js home js js and added language js current behavior tabbar is labeled with route name when route is a stack expected behavior tabbar labeled with tabbarlabel as it is when route is a screen how to reproduce your environment software version react navigation react native node npm or yarn npm | 1 |
232,827 | 25,706,363,113 | IssuesEvent | 2022-12-07 01:06:42 | temporalio/sdk-dotnet | https://api.github.com/repos/temporalio/sdk-dotnet | closed | google.protobuf.3.21.1.nupkg: 1 vulnerabilities (highest severity is: 5.5) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>google.protobuf.3.21.1.nupkg</b></p></summary>
<p>C# runtime library for Protocol Buffers - Google's data interchange format.</p>
<p>Library home page: <a href="https://api.nuget.org/packages/google.protobuf.3.21.1.nupkg">https://api.nuget.org/packages/google.protobuf.3.21.1.nupkg</a></p>
<p>Path to dependency file: /Src/SDK/Api.Client.ManagedGrpc/Temporal.Api.Client.ManagedGrpc.csproj</p>
<p>Path to vulnerable library: /et/packages/google.protobuf/3.21.1/google.protobuf.3.21.1.nupkg</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-1941](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-1941) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | google.protobuf.3.21.1.nupkg | Direct | Google.Protobuf - 3.18.3,3.19.5,3.20.2,3.21.6;protobuf-python - 3.18.3,3.19.5,3.20.2,4.21.6 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-1941</summary>
### Vulnerable Library - <b>google.protobuf.3.21.1.nupkg</b></p>
<p>C# runtime library for Protocol Buffers - Google's data interchange format.</p>
<p>Library home page: <a href="https://api.nuget.org/packages/google.protobuf.3.21.1.nupkg">https://api.nuget.org/packages/google.protobuf.3.21.1.nupkg</a></p>
<p>Path to dependency file: /Src/SDK/Api.Client.ManagedGrpc/Temporal.Api.Client.ManagedGrpc.csproj</p>
<p>Path to vulnerable library: /et/packages/google.protobuf/3.21.1/google.protobuf.3.21.1.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **google.protobuf.3.21.1.nupkg** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A parsing vulnerability for the MessageSet type in the ProtocolBuffers versions prior to and including 3.16.1, 3.17.3, 3.18.2, 3.19.4, 3.20.1 and 3.21.5 for protobuf-cpp, and versions prior to and including 3.16.1, 3.17.3, 3.18.2, 3.19.4, 3.20.1 and 4.21.5 for protobuf-python can lead to out of memory failures. A specially crafted message with multiple key-value per elements creates parsing issues, and can lead to a Denial of Service against services receiving unsanitized input. We recommend upgrading to versions 3.18.3, 3.19.5, 3.20.2, 3.21.6 for protobuf-cpp and 3.18.3, 3.19.5, 3.20.2, 4.21.6 for protobuf-python. Versions for 3.16 and 3.17 are no longer updated.
<p>Publish Date: 2022-09-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-1941>CVE-2022-1941</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cloud.google.com/support/bulletins#GCP-2022-019">https://cloud.google.com/support/bulletins#GCP-2022-019</a></p>
<p>Release Date: 2022-09-22</p>
<p>Fix Resolution: Google.Protobuf - 3.18.3,3.19.5,3.20.2,3.21.6;protobuf-python - 3.18.3,3.19.5,3.20.2,4.21.6</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | True | google.protobuf.3.21.1.nupkg: 1 vulnerabilities (highest severity is: 5.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>google.protobuf.3.21.1.nupkg</b></p></summary>
<p>C# runtime library for Protocol Buffers - Google's data interchange format.</p>
<p>Library home page: <a href="https://api.nuget.org/packages/google.protobuf.3.21.1.nupkg">https://api.nuget.org/packages/google.protobuf.3.21.1.nupkg</a></p>
<p>Path to dependency file: /Src/SDK/Api.Client.ManagedGrpc/Temporal.Api.Client.ManagedGrpc.csproj</p>
<p>Path to vulnerable library: /et/packages/google.protobuf/3.21.1/google.protobuf.3.21.1.nupkg</p>
<p>
</details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-1941](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-1941) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | google.protobuf.3.21.1.nupkg | Direct | Google.Protobuf - 3.18.3,3.19.5,3.20.2,3.21.6;protobuf-python - 3.18.3,3.19.5,3.20.2,4.21.6 | ✅ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-1941</summary>
### Vulnerable Library - <b>google.protobuf.3.21.1.nupkg</b></p>
<p>C# runtime library for Protocol Buffers - Google's data interchange format.</p>
<p>Library home page: <a href="https://api.nuget.org/packages/google.protobuf.3.21.1.nupkg">https://api.nuget.org/packages/google.protobuf.3.21.1.nupkg</a></p>
<p>Path to dependency file: /Src/SDK/Api.Client.ManagedGrpc/Temporal.Api.Client.ManagedGrpc.csproj</p>
<p>Path to vulnerable library: /et/packages/google.protobuf/3.21.1/google.protobuf.3.21.1.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **google.protobuf.3.21.1.nupkg** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
A parsing vulnerability for the MessageSet type in the ProtocolBuffers versions prior to and including 3.16.1, 3.17.3, 3.18.2, 3.19.4, 3.20.1 and 3.21.5 for protobuf-cpp, and versions prior to and including 3.16.1, 3.17.3, 3.18.2, 3.19.4, 3.20.1 and 4.21.5 for protobuf-python can lead to out of memory failures. A specially crafted message with multiple key-value per elements creates parsing issues, and can lead to a Denial of Service against services receiving unsanitized input. We recommend upgrading to versions 3.18.3, 3.19.5, 3.20.2, 3.21.6 for protobuf-cpp and 3.18.3, 3.19.5, 3.20.2, 4.21.6 for protobuf-python. Versions for 3.16 and 3.17 are no longer updated.
<p>Publish Date: 2022-09-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-1941>CVE-2022-1941</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cloud.google.com/support/bulletins#GCP-2022-019">https://cloud.google.com/support/bulletins#GCP-2022-019</a></p>
<p>Release Date: 2022-09-22</p>
<p>Fix Resolution: Google.Protobuf - 3.18.3,3.19.5,3.20.2,3.21.6;protobuf-python - 3.18.3,3.19.5,3.20.2,4.21.6</p>
</p>
<p></p>
:rescue_worker_helmet: Automatic Remediation is available for this issue
</details>
***
<p>:rescue_worker_helmet: Automatic Remediation is available for this issue.</p> | non_main | google protobuf nupkg vulnerabilities highest severity is autoclosed vulnerable library google protobuf nupkg c runtime library for protocol buffers google s data interchange format library home page a href path to dependency file src sdk api client managedgrpc temporal api client managedgrpc csproj path to vulnerable library et packages google protobuf google protobuf nupkg vulnerabilities cve severity cvss dependency type fixed in remediation available medium google protobuf nupkg direct google protobuf protobuf python details cve vulnerable library google protobuf nupkg c runtime library for protocol buffers google s data interchange format library home page a href path to dependency file src sdk api client managedgrpc temporal api client managedgrpc csproj path to vulnerable library et packages google protobuf google protobuf nupkg dependency hierarchy x google protobuf nupkg vulnerable library found in base branch master vulnerability details a parsing vulnerability for the messageset type in the protocolbuffers versions prior to and including and for protobuf cpp and versions prior to and including and for protobuf python can lead to out of memory failures a specially crafted message with multiple key value per elements creates parsing issues and can lead to a denial of service against services receiving unsanitized input we recommend upgrading to versions for protobuf cpp and for protobuf python versions for and are no longer updated publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution google protobuf protobuf python rescue worker helmet automatic remediation is available for this issue rescue worker helmet automatic remediation is available for this issue | 0 |
216,108 | 24,229,858,563 | IssuesEvent | 2022-09-26 17:15:44 | Path-Check/gaen-mobile | https://api.github.com/repos/Path-Check/gaen-mobile | closed | CVE-2022-0144 (High) detected in shelljs-0.8.4.tgz - autoclosed | security vulnerability | ## CVE-2022-0144 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shelljs-0.8.4.tgz</b></p></summary>
<p>Portable Unix shell commands for Node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/shelljs/-/shelljs-0.8.4.tgz">https://registry.npmjs.org/shelljs/-/shelljs-0.8.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/shelljs/package.json</p>
<p>
Dependency Hierarchy:
- react-native-0.64.2.tgz (Root Library)
- :x: **shelljs-0.8.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Path-Check/gaen-mobile/commit/b58c238936a70cca77ab2ffa266d4ffa8d1a45d3">b58c238936a70cca77ab2ffa266d4ffa8d1a45d3</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
shelljs is vulnerable to Improper Privilege Management
<p>Publish Date: 2022-01-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0144>CVE-2022-0144</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-01-11</p>
<p>Fix Resolution (shelljs): 0.8.5</p>
<p>Direct dependency fix Resolution (react-native): 0.65.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-0144 (High) detected in shelljs-0.8.4.tgz - autoclosed - ## CVE-2022-0144 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>shelljs-0.8.4.tgz</b></p></summary>
<p>Portable Unix shell commands for Node.js</p>
<p>Library home page: <a href="https://registry.npmjs.org/shelljs/-/shelljs-0.8.4.tgz">https://registry.npmjs.org/shelljs/-/shelljs-0.8.4.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/shelljs/package.json</p>
<p>
Dependency Hierarchy:
- react-native-0.64.2.tgz (Root Library)
- :x: **shelljs-0.8.4.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/Path-Check/gaen-mobile/commit/b58c238936a70cca77ab2ffa266d4ffa8d1a45d3">b58c238936a70cca77ab2ffa266d4ffa8d1a45d3</a></p>
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
shelljs is vulnerable to Improper Privilege Management
<p>Publish Date: 2022-01-11
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0144>CVE-2022-0144</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2022-01-11</p>
<p>Fix Resolution (shelljs): 0.8.5</p>
<p>Direct dependency fix Resolution (react-native): 0.65.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve high detected in shelljs tgz autoclosed cve high severity vulnerability vulnerable library shelljs tgz portable unix shell commands for node js library home page a href path to dependency file package json path to vulnerable library node modules shelljs package json dependency hierarchy react native tgz root library x shelljs tgz vulnerable library found in head commit a href found in base branch develop vulnerability details shelljs is vulnerable to improper privilege management publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version release date fix resolution shelljs direct dependency fix resolution react native step up your open source security game with mend | 0 |
271,130 | 29,299,164,564 | IssuesEvent | 2023-05-25 01:06:39 | panasalap/linux-4.19.72_test1 | https://api.github.com/repos/panasalap/linux-4.19.72_test1 | opened | CVE-2023-33203 (Medium) detected in linux-yoctov5.4.51 | Mend: dependency security vulnerability | ## CVE-2023-33203 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/f1b7c617b9b8f4135ab2f75a0c407cc44d43683f">f1b7c617b9b8f4135ab2f75a0c407cc44d43683f</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel before 6.2.9 has a race condition and resultant use-after-free in drivers/net/ethernet/qualcomm/emac/emac.c if a physically proximate attacker unplugs an emac based device.
<p>Publish Date: 2023-05-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-33203>CVE-2023-33203</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-33203">https://www.linuxkernelcves.com/cves/CVE-2023-33203</a></p>
<p>Release Date: 2023-05-18</p>
<p>Fix Resolution: v4.14.312,v4.19.280,v5.4.240,v5.10.177,v5.15.105,v6.1.22,v6.2.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2023-33203 (Medium) detected in linux-yoctov5.4.51 - ## CVE-2023-33203 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yoctov5.4.51</b></p></summary>
<p>
<p>Yocto Linux Embedded kernel</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto>https://git.yoctoproject.org/git/linux-yocto</a></p>
<p>Found in HEAD commit: <a href="https://github.com/panasalap/linux-4.19.72/commit/f1b7c617b9b8f4135ab2f75a0c407cc44d43683f">f1b7c617b9b8f4135ab2f75a0c407cc44d43683f</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/net/ethernet/qualcomm/emac/emac.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The Linux kernel before 6.2.9 has a race condition and resultant use-after-free in drivers/net/ethernet/qualcomm/emac/emac.c if a physically proximate attacker unplugs an emac based device.
<p>Publish Date: 2023-05-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-33203>CVE-2023-33203</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2023-33203">https://www.linuxkernelcves.com/cves/CVE-2023-33203</a></p>
<p>Release Date: 2023-05-18</p>
<p>Fix Resolution: v4.14.312,v4.19.280,v5.4.240,v5.10.177,v5.15.105,v6.1.22,v6.2.9</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in linux cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel library home page a href found in head commit a href found in base branch master vulnerable source files drivers net ethernet qualcomm emac emac c drivers net ethernet qualcomm emac emac c vulnerability details the linux kernel before has a race condition and resultant use after free in drivers net ethernet qualcomm emac emac c if a physically proximate attacker unplugs an emac based device publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
2,476 | 8,639,909,264 | IssuesEvent | 2018-11-23 22:36:45 | F5OEO/rpitx | https://api.github.com/repos/F5OEO/rpitx | closed | Pi transmitting unwanted signals while transmitting by SSB | V1 related (not maintained) | Thanks for rpitx by the way.
I"m doing something like this to transmit a jt65 cq of a pre-recorded file I recorded via ssb on an original raspberry pi.
sudo /home/pi/radio/rpitx/rpitx -m IQ -i /home/pi/radio/rpitx/jt65/cq.wav.ssb -f 14077.6
I can pick myself up on another receiver and decode the jt65 file 20m away. The peculiar thing is my transmit signal is also transmitting a local am or fm radio station alongside the wav file which makes for a right dirty signal. I have a low pass filter for 20m connected between the pi and the antenna.
Has anyone any thoughts on cleaning up this signal? I don't see this behaviour when I use the VFO option as a test. The pi is currently inside a biscuit tin.
Many thanks,
Chris
| True | Pi transmitting unwanted signals while transmitting by SSB - Thanks for rpitx by the way.
I"m doing something like this to transmit a jt65 cq of a pre-recorded file I recorded via ssb on an original raspberry pi.
sudo /home/pi/radio/rpitx/rpitx -m IQ -i /home/pi/radio/rpitx/jt65/cq.wav.ssb -f 14077.6
I can pick myself up on another receiver and decode the jt65 file 20m away. The peculiar thing is my transmit signal is also transmitting a local am or fm radio station alongside the wav file which makes for a right dirty signal. I have a low pass filter for 20m connected between the pi and the antenna.
Has anyone any thoughts on cleaning up this signal? I don't see this behaviour when I use the VFO option as a test. The pi is currently inside a biscuit tin.
Many thanks,
Chris
| main | pi transmitting unwanted signals while transmitting by ssb thanks for rpitx by the way i m doing something like this to transmit a cq of a pre recorded file i recorded via ssb on an original raspberry pi sudo home pi radio rpitx rpitx m iq i home pi radio rpitx cq wav ssb f i can pick myself up on another receiver and decode the file away the peculiar thing is my transmit signal is also transmitting a local am or fm radio station alongside the wav file which makes for a right dirty signal i have a low pass filter for connected between the pi and the antenna has anyone any thoughts on cleaning up this signal i don t see this behaviour when i use the vfo option as a test the pi is currently inside a biscuit tin many thanks chris | 1 |
3,584 | 14,444,784,067 | IssuesEvent | 2020-12-07 21:48:38 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [Search a11y] aria-hidden property is set to true on focusable search elements | status: needs triage 🕵️♀️ status: waiting for maintainer response 💬 type: a11y ♿ | ## Environment
> Operating system
> Browser
> Automated testing tool and ruleset
> Assistive technology used to verify
## Detailed description
> What version of the Carbon Design System are you using?
Carbon-components-react
> What did you expect to happen?
aria-hidden should be false for focusable elements.
> What happened instead?
aria-hidden is set to true.
> What WCAG 2.1 checkpoint does the issue violate?
Rule ID: aria-hidden-focus
Ruleset: axe-core 3.4
User Impact: Serious
WCAG: 4.1.2
## Steps to reproduce the issue
1. Find a persistent data table search
2. Inspect the search input element.
3. Check value for aria-hidden.
[https://codesandbox.io/s/datatable-with-batch-actions-uolbz](https://codesandbox.io/s/datatable-with-batch-actions-uolbz
)
## Additional information
Using the aria-hidden="true" attribute on an element removes the element and ALL of its child nodes from the accessibility API making it completely inaccessible to screen readers and other assistive technologies. Aria-hidden may be used with extreme caution to hide visibly rendered content from assistive technologies only if the act of hiding this content is intended to improve the experience for users of assistive technologies by removing redundant or extraneous content. If aria-hidden is used to hide visible content from screen readers, the identical or equivalent meaning and functionality must be exposed to assistive technologies.
aria-hidden should not be true for items that allow input or focus. | True | [Search a11y] aria-hidden property is set to true on focusable search elements - ## Environment
> Operating system
> Browser
> Automated testing tool and ruleset
> Assistive technology used to verify
## Detailed description
> What version of the Carbon Design System are you using?
Carbon-components-react
> What did you expect to happen?
aria-hidden should be false for focusable elements.
> What happened instead?
aria-hidden is set to true.
> What WCAG 2.1 checkpoint does the issue violate?
Rule ID: aria-hidden-focus
Ruleset: axe-core 3.4
User Impact: Serious
WCAG: 4.1.2
## Steps to reproduce the issue
1. Find a persistent data table search
2. Inspect the search input element.
3. Check value for aria-hidden.
[https://codesandbox.io/s/datatable-with-batch-actions-uolbz](https://codesandbox.io/s/datatable-with-batch-actions-uolbz
)
## Additional information
Using the aria-hidden="true" attribute on an element removes the element and ALL of its child nodes from the accessibility API making it completely inaccessible to screen readers and other assistive technologies. Aria-hidden may be used with extreme caution to hide visibly rendered content from assistive technologies only if the act of hiding this content is intended to improve the experience for users of assistive technologies by removing redundant or extraneous content. If aria-hidden is used to hide visible content from screen readers, the identical or equivalent meaning and functionality must be exposed to assistive technologies.
aria-hidden should not be true for items that allow input or focus. | main | aria hidden property is set to true on focusable search elements environment operating system browser automated testing tool and ruleset assistive technology used to verify detailed description what version of the carbon design system are you using carbon components react what did you expect to happen aria hidden should be false for focusable elements what happened instead aria hidden is set to true what wcag checkpoint does the issue violate rule id aria hidden focus ruleset axe core user impact serious wcag steps to reproduce the issue find a persistent data table search inspect the search input element check value for aria hidden additional information using the aria hidden true attribute on an element removes the element and all of its child nodes from the accessibility api making it completely inaccessible to screen readers and other assistive technologies aria hidden may be used with extreme caution to hide visibly rendered content from assistive technologies only if the act of hiding this content is intended to improve the experience for users of assistive technologies by removing redundant or extraneous content if aria hidden is used to hide visible content from screen readers the identical or equivalent meaning and functionality must be exposed to assistive technologies aria hidden should not be true for items that allow input or focus | 1 |
3,052 | 11,420,242,440 | IssuesEvent | 2020-02-03 09:44:04 | OpenRefine/OpenRefine | https://api.github.com/repos/OpenRefine/OpenRefine | closed | Migrate from com.google.refine package to org.openrefine | enhancement maintainability | We have inherited from the `com.google.refine` package when Google Refine transitioned to OpenRefine. It made sense to keep using this package to maintain compatibility with extensions.
We aim to maintain the 3.x release series with the same `com.google.refine` package, so as not to spuriously break extensions. Version 4.0 will bring a lot of breaking changes, so it would make sense to do the switch to `org.openrefine` between the two.
This will ensure that the package we use is consistent with the Maven groupId of the artifacts we publish, for #2254. | True | Migrate from com.google.refine package to org.openrefine - We have inherited from the `com.google.refine` package when Google Refine transitioned to OpenRefine. It made sense to keep using this package to maintain compatibility with extensions.
We aim to maintain the 3.x release series with the same `com.google.refine` package, so as not to spuriously break extensions. Version 4.0 will bring a lot of breaking changes, so it would make sense to do the switch to `org.openrefine` between the two.
This will ensure that the package we use is consistent with the Maven groupId of the artifacts we publish, for #2254. | main | migrate from com google refine package to org openrefine we have inherited from the com google refine package when google refine transitioned to openrefine it made sense to keep using this package to maintain compatibility with extensions we aim to maintain the x release series with the same com google refine package so as not to spuriously break extensions version will bring a lot of breaking changes so it would make sense to do the switch to org openrefine between the two this will ensure that the package we use is consistent with the maven groupid of the artifacts we publish for | 1 |
252,900 | 27,271,379,909 | IssuesEvent | 2023-02-22 22:43:13 | snowflakedb/libsnowflakeclient | https://api.github.com/repos/snowflakedb/libsnowflakeclient | closed | util-linuxv2.36.2: 2 vulnerabilities (highest severity is: 5.5) | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>util-linuxv2.36.2</b></p></summary>
<p>
<p>The util-linux code repository.</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p>
</p>
</p></p>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chsh.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chfn.c</b>
</p>
<p></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-0563](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0563) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | util-linuxv2.36.2 | Direct | v2.37.4 | ❌ |
| [CVE-2021-3995](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3995) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.7 | util-linuxv2.36.2 | Direct | v2.37.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0563</summary>
### Vulnerable Library - <b>util-linuxv2.36.2</b></p>
<p>
<p>The util-linux code repository.</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (2)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chsh.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chfn.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
A flaw was found in the util-linux chfn and chsh utilities when compiled with Readline support. The Readline library uses an "INPUTRC" environment variable to get a path to the library config file. When the library cannot parse the specified file, it prints an error message containing data from the file. This flaw allows an unprivileged user to read root-owned files, potentially leading to privilege escalation. This flaw affects util-linux versions prior to 2.37.4.
<p>Publish Date: 2022-02-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0563>CVE-2022-0563</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=2053151">https://bugzilla.redhat.com/show_bug.cgi?id=2053151</a></p>
<p>Release Date: 2022-02-21</p>
<p>Fix Resolution: v2.37.4</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-3995</summary>
### Vulnerable Library - <b>util-linuxv2.36.2</b></p>
<p>
<p>The util-linux code repository.</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/libmount/src/mountP.h</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
A logic error was found in the libmount library of util-linux in the function that allows an unprivileged user to unmount a FUSE filesystem. This flaw allows an unprivileged local attacker to unmount FUSE filesystems that belong to certain other users who have a UID that is a prefix of the UID of the attacker in its string form. An attacker may use this flaw to cause a denial of service to applications that use the affected filesystems.
<p>Publish Date: 2021-11-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3995>CVE-2021-3995</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2021-3995">https://security-tracker.debian.org/tracker/CVE-2021-3995</a></p>
<p>Release Date: 2021-11-22</p>
<p>Fix Resolution: v2.37.3</p>
</p>
<p></p>
</details> | True | util-linuxv2.36.2: 2 vulnerabilities (highest severity is: 5.5) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>util-linuxv2.36.2</b></p></summary>
<p>
<p>The util-linux code repository.</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p>
</p>
</p></p>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chsh.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chfn.c</b>
</p>
<p></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | --- | --- |
| [CVE-2022-0563](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0563) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.5 | util-linuxv2.36.2 | Direct | v2.37.4 | ❌ |
| [CVE-2021-3995](https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3995) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 4.7 | util-linuxv2.36.2 | Direct | v2.37.3 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-0563</summary>
### Vulnerable Library - <b>util-linuxv2.36.2</b></p>
<p>
<p>The util-linux code repository.</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (2)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chsh.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/login-utils/chfn.c</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
A flaw was found in the util-linux chfn and chsh utilities when compiled with Readline support. The Readline library uses an "INPUTRC" environment variable to get a path to the library config file. When the library cannot parse the specified file, it prints an error message containing data from the file. This flaw allows an unprivileged user to read root-owned files, potentially leading to privilege escalation. This flaw affects util-linux versions prior to 2.37.4.
<p>Publish Date: 2022-02-21
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0563>CVE-2022-0563</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://bugzilla.redhat.com/show_bug.cgi?id=2053151">https://bugzilla.redhat.com/show_bug.cgi?id=2053151</a></p>
<p>Release Date: 2022-02-21</p>
<p>Fix Resolution: v2.37.4</p>
</p>
<p></p>
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2021-3995</summary>
### Vulnerable Library - <b>util-linuxv2.36.2</b></p>
<p>
<p>The util-linux code repository.</p>
<p>Library home page: <a href=https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git>https://git.kernel.org/pub/scm/utils/util-linux/util-linux.git</a></p>
<p>Found in base branch: <b>master</b></p></p>
</p></p>
### Vulnerable Source Files (1)
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/deps/util-linux.tar/util-linux/libmount/src/mountP.h</b>
</p>
<p></p>
</p>
<p></p>
### Vulnerability Details
<p>
A logic error was found in the libmount library of util-linux in the function that allows an unprivileged user to unmount a FUSE filesystem. This flaw allows an unprivileged local attacker to unmount FUSE filesystems that belong to certain other users who have a UID that is a prefix of the UID of the attacker in its string form. An attacker may use this flaw to cause a denial of service to applications that use the affected filesystems.
<p>Publish Date: 2021-11-22
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3995>CVE-2021-3995</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>4.7</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://security-tracker.debian.org/tracker/CVE-2021-3995">https://security-tracker.debian.org/tracker/CVE-2021-3995</a></p>
<p>Release Date: 2021-11-22</p>
<p>Fix Resolution: v2.37.3</p>
</p>
<p></p>
</details> | non_main | util vulnerabilities highest severity is vulnerable library util the util linux code repository library home page a href vulnerable source files deps util linux tar util linux login utils chsh c deps util linux tar util linux login utils chfn c vulnerabilities cve severity cvss dependency type fixed in remediation available medium util direct medium util direct details cve vulnerable library util the util linux code repository library home page a href found in base branch master vulnerable source files deps util linux tar util linux login utils chsh c deps util linux tar util linux login utils chfn c vulnerability details a flaw was found in the util linux chfn and chsh utilities when compiled with readline support the readline library uses an inputrc environment variable to get a path to the library config file when the library cannot parse the specified file it prints an error message containing data from the file this flaw allows an unprivileged user to read root owned files potentially leading to privilege escalation this flaw affects util linux versions prior to publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution cve vulnerable library util the util linux code repository library home page a href found in base branch master vulnerable source files deps util linux tar util linux libmount src mountp h vulnerability details a logic error was found in the libmount library of util linux in the function that allows an unprivileged user to unmount a fuse filesystem this flaw allows an unprivileged local attacker to unmount fuse filesystems that belong to certain other users who have a uid that is a prefix of the uid of the attacker in its string form an attacker may use this flaw to cause a denial of service to applications that use the affected filesystems publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
3,163 | 12,226,508,740 | IssuesEvent | 2020-05-03 11:15:49 | gfleetwood/asteres | https://api.github.com/repos/gfleetwood/asteres | opened | daroczig/logger (157297209) | R maintain | https://github.com/daroczig/logger
A lightweight, modern and flexible, log4j and futile.logger inspired logging utility for R | True | daroczig/logger (157297209) - https://github.com/daroczig/logger
A lightweight, modern and flexible, log4j and futile.logger inspired logging utility for R | main | daroczig logger a lightweight modern and flexible and futile logger inspired logging utility for r | 1 |
1,391 | 6,025,275,444 | IssuesEvent | 2017-06-08 08:15:58 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | win_iis_webapppool - not able to set all the properties | affects_2.1 bug_report waiting_on_maintainer windows | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_iis_webapppool
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
Default config
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
ansible is running on Ubuntu Trusty and the systems being managed are linux and Windows.
##### SUMMARY
<!--- Explain the problem briefly -->
Tried to change the value of Identity under Process Model in advanced settings of an appPool and the change did not take effect. Tried to change the value of other properties under Process Model and none gets changed, even though the output of ansible says changed. IIS being managed is 7.5 version. Can only few properties be modified of IIS from the ansible module as of now? If yes, please specify that in the documentation.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Set Identity Type to Network Service.
win_iis_webapppool:
name: UIAPI
attributes: identityType:2
state: restarted
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The Identity Type of the app pool should have changed to Network Service, but it did not.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
The value of identityType did not change, even though ansible reported that it has been changed.
<!--- Paste verbatim command output between quotes below -->
```
<ServerName> ESTABLISH WINRM CONNECTION FOR USER: UserName on PORT 5985 TO ServerName
<ServerName> EXEC Set-StrictMode -Version Latest
(New-Item -Type Directory -Path $env:temp -Name "ansible-tmp-1471345999.53-80318070680631").FullName | Write-Host -Separator '';
<ServerName> PUT "/tmp/tmpm8UpFI" TO "C:\Users\userID\AppData\Local\Temp\ansible-tmp-1471345999.53-80318070680631\win_iis_webapppool.ps1"
<ServerName> EXEC Set-StrictMode -Version Latest
Try
{
& 'C:\Users\userID\AppData\Local\Temp\ansible-tmp-1471345999.53-80318070680631\win_iis_webapppool.ps1'
}
Catch
{
$_obj = @{ failed = $true }
If ($_.Exception.GetType)
{
$_obj.Add('msg', $_.Exception.Message)
}
Else
{
$_obj.Add('msg', $_.ToString())
}
If ($_.InvocationInfo.PositionMessage)
{
$_obj.Add('exception', $_.InvocationInfo.PositionMessage)
}
ElseIf ($_.ScriptStackTrace)
{
$_obj.Add('exception', $_.ScriptStackTrace)
}
Try
{
$_obj.Add('error_record', ($_ | ConvertTo-Json | ConvertFrom-Json))
}
Catch
{
}
Echo $_obj | ConvertTo-Json -Compress -Depth 99
Exit 1
}
Finally { Remove-Item "C:\Users\userID\AppData\Local\Temp\ansible-tmp-1471345999.53-80318070680631" -Force -Recurse -ErrorAction SilentlyContinue }
changed: [ServerName] => {"attributes": {"identityType": "2"}, "changed": true, "info": {"attributes": {"CLRConfigFile": "", "applicationPoolSid": "S-1-5-82-
3266571588-702921383-1992572147-801121604-1441276117", "autoStart": true, "enable32BitAppOnWin64": false, "enableConfigurationOverride": true, "managedPipelineMode": 0, "
managedRuntimeLoader": "webengine4.dll", "managedRuntimeVersion": "v4.0", "name": "UIAPI", "passAnonymousToken": true, "queueLength": 1000, "startMode": 1, "state": 1},
"name": "UIAPI", "state": "Started"}, "invocation": {"module_name": "win_iis_webapppool"}}
PLAY RECAP *********************************************************************
ServerName : ok=1 changed=1 unreachable=0 failed=0
```
| True | win_iis_webapppool - not able to set all the properties - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
- Feature Idea
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
win_iis_webapppool
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
Default config
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
ansible is running on Ubuntu Trusty and the systems being managed are linux and Windows.
##### SUMMARY
<!--- Explain the problem briefly -->
Tried to change the value of Identity under Process Model in advanced settings of an appPool and the change did not take effect. Tried to change the value of other properties under Process Model and none gets changed, even though the output of ansible says changed. IIS being managed is 7.5 version. Can only few properties be modified of IIS from the ansible module as of now? If yes, please specify that in the documentation.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Set Identity Type to Network Service.
win_iis_webapppool:
name: UIAPI
attributes: identityType:2
state: restarted
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
The Identity Type of the app pool should have changed to Network Service, but it did not.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
The value of identityType did not change, even though ansible reported that it has been changed.
<!--- Paste verbatim command output between quotes below -->
```
<ServerName> ESTABLISH WINRM CONNECTION FOR USER: UserName on PORT 5985 TO ServerName
<ServerName> EXEC Set-StrictMode -Version Latest
(New-Item -Type Directory -Path $env:temp -Name "ansible-tmp-1471345999.53-80318070680631").FullName | Write-Host -Separator '';
<ServerName> PUT "/tmp/tmpm8UpFI" TO "C:\Users\userID\AppData\Local\Temp\ansible-tmp-1471345999.53-80318070680631\win_iis_webapppool.ps1"
<ServerName> EXEC Set-StrictMode -Version Latest
Try
{
& 'C:\Users\userID\AppData\Local\Temp\ansible-tmp-1471345999.53-80318070680631\win_iis_webapppool.ps1'
}
Catch
{
$_obj = @{ failed = $true }
If ($_.Exception.GetType)
{
$_obj.Add('msg', $_.Exception.Message)
}
Else
{
$_obj.Add('msg', $_.ToString())
}
If ($_.InvocationInfo.PositionMessage)
{
$_obj.Add('exception', $_.InvocationInfo.PositionMessage)
}
ElseIf ($_.ScriptStackTrace)
{
$_obj.Add('exception', $_.ScriptStackTrace)
}
Try
{
$_obj.Add('error_record', ($_ | ConvertTo-Json | ConvertFrom-Json))
}
Catch
{
}
Echo $_obj | ConvertTo-Json -Compress -Depth 99
Exit 1
}
Finally { Remove-Item "C:\Users\userID\AppData\Local\Temp\ansible-tmp-1471345999.53-80318070680631" -Force -Recurse -ErrorAction SilentlyContinue }
changed: [ServerName] => {"attributes": {"identityType": "2"}, "changed": true, "info": {"attributes": {"CLRConfigFile": "", "applicationPoolSid": "S-1-5-82-
3266571588-702921383-1992572147-801121604-1441276117", "autoStart": true, "enable32BitAppOnWin64": false, "enableConfigurationOverride": true, "managedPipelineMode": 0, "
managedRuntimeLoader": "webengine4.dll", "managedRuntimeVersion": "v4.0", "name": "UIAPI", "passAnonymousToken": true, "queueLength": 1000, "startMode": 1, "state": 1},
"name": "UIAPI", "state": "Started"}, "invocation": {"module_name": "win_iis_webapppool"}}
PLAY RECAP *********************************************************************
ServerName : ok=1 changed=1 unreachable=0 failed=0
```
| main | win iis webapppool not able to set all the properties issue type bug report feature idea component name win iis webapppool ansible version ansible configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables default config os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific ansible is running on ubuntu trusty and the systems being managed are linux and windows summary tried to change the value of identity under process model in advanced settings of an apppool and the change did not take effect tried to change the value of other properties under process model and none gets changed even though the output of ansible says changed iis being managed is version can only few properties be modified of iis from the ansible module as of now if yes please specify that in the documentation steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name set identity type to network service win iis webapppool name uiapi attributes identitytype state restarted expected results the identity type of the app pool should have changed to network service but it did not actual results the value of identitytype did not change even though ansible reported that it has been changed establish winrm connection for user username on port to servername exec set strictmode version latest new item type directory path env temp name ansible tmp fullname write host separator put tmp to c users userid appdata local temp ansible tmp win iis webapppool exec set strictmode version latest try c users userid appdata local temp ansible tmp win iis webapppool catch obj failed true if exception gettype obj add msg exception message else obj add msg tostring if invocationinfo positionmessage obj add exception invocationinfo positionmessage elseif scriptstacktrace obj add exception scriptstacktrace try obj add error record convertto json convertfrom json catch echo obj convertto json compress depth exit finally remove item c users userid appdata local temp ansible tmp force recurse erroraction silentlycontinue changed attributes identitytype changed true info attributes clrconfigfile applicationpoolsid s autostart true false enableconfigurationoverride true managedpipelinemode managedruntimeloader dll managedruntimeversion name uiapi passanonymoustoken true queuelength startmode state name uiapi state started invocation module name win iis webapppool play recap servername ok changed unreachable failed | 1 |
260,635 | 8,212,682,546 | IssuesEvent | 2018-09-04 17:05:56 | phetsims/axon | https://api.github.com/repos/phetsims/axon | closed | detect Property loops | priority:2-high | A Property loop occurs when a Property's `set` method is entered before a previous call to `set` exits. PhET-iO makes it necessary to deal with these loops because it results in intermediate/redundant data in the message stream. See for example https://github.com/phetsims/hookes-law/issues/52.
PhetioObject previously was responsible for detecting Property loops, but that was removed in https://github.com/phetsims/tandem/issues/57. And I believe it was @samreid who suggested that this responsibility does not belong in PhET-iO; it belongs in axon.
In https://github.com/phetsims/tandem/issues/57#issuecomment-396712827, I suggested a way to add responsibility (conditionally) to Property, reproduced below. And I found this to be invaluable in troubleshooting https://github.com/phetsims/hookes-law/issues/52. I'd like to see this added to Property.
```js
// @private
_notifyListeners: function( oldValue ) {
// We must short circuit based on tandem here as a guard against the toStateObject calls
this.tandem.isSuppliedAndEnabled() && this.startEvent( 'model', 'changed', {
oldValue: this.phetioType.elementType.toStateObject( oldValue ),
newValue: this.phetioType.elementType.toStateObject( this.get() ),
units: this.phetioType && this.phetioType.units
}, this.changeEventOptions );
// notify listeners, optionally detect loops where this Property is set again before this completes.
assert && assert( !this.notifying || !phet.chipper.queryParameters.detectPropertyLoops,
'Property loop detected, value=' + this.get() + ', oldValue=' + oldValue );
this.notifying = true;
this.changedEmitter.emit2( this.get(), oldValue );
this.notifying = false;
this.tandem.isSuppliedAndEnabled() && this.endEvent();
},
````
| 1.0 | detect Property loops - A Property loop occurs when a Property's `set` method is entered before a previous call to `set` exits. PhET-iO makes it necessary to deal with these loops because it results in intermediate/redundant data in the message stream. See for example https://github.com/phetsims/hookes-law/issues/52.
PhetioObject previously was responsible for detecting Property loops, but that was removed in https://github.com/phetsims/tandem/issues/57. And I believe it was @samreid who suggested that this responsibility does not belong in PhET-iO; it belongs in axon.
In https://github.com/phetsims/tandem/issues/57#issuecomment-396712827, I suggested a way to add responsibility (conditionally) to Property, reproduced below. And I found this to be invaluable in troubleshooting https://github.com/phetsims/hookes-law/issues/52. I'd like to see this added to Property.
```js
// @private
_notifyListeners: function( oldValue ) {
// We must short circuit based on tandem here as a guard against the toStateObject calls
this.tandem.isSuppliedAndEnabled() && this.startEvent( 'model', 'changed', {
oldValue: this.phetioType.elementType.toStateObject( oldValue ),
newValue: this.phetioType.elementType.toStateObject( this.get() ),
units: this.phetioType && this.phetioType.units
}, this.changeEventOptions );
// notify listeners, optionally detect loops where this Property is set again before this completes.
assert && assert( !this.notifying || !phet.chipper.queryParameters.detectPropertyLoops,
'Property loop detected, value=' + this.get() + ', oldValue=' + oldValue );
this.notifying = true;
this.changedEmitter.emit2( this.get(), oldValue );
this.notifying = false;
this.tandem.isSuppliedAndEnabled() && this.endEvent();
},
````
| non_main | detect property loops a property loop occurs when a property s set method is entered before a previous call to set exits phet io makes it necessary to deal with these loops because it results in intermediate redundant data in the message stream see for example phetioobject previously was responsible for detecting property loops but that was removed in and i believe it was samreid who suggested that this responsibility does not belong in phet io it belongs in axon in i suggested a way to add responsibility conditionally to property reproduced below and i found this to be invaluable in troubleshooting i d like to see this added to property js private notifylisteners function oldvalue we must short circuit based on tandem here as a guard against the tostateobject calls this tandem issuppliedandenabled this startevent model changed oldvalue this phetiotype elementtype tostateobject oldvalue newvalue this phetiotype elementtype tostateobject this get units this phetiotype this phetiotype units this changeeventoptions notify listeners optionally detect loops where this property is set again before this completes assert assert this notifying phet chipper queryparameters detectpropertyloops property loop detected value this get oldvalue oldvalue this notifying true this changedemitter this get oldvalue this notifying false this tandem issuppliedandenabled this endevent | 0 |
610,969 | 18,941,325,846 | IssuesEvent | 2021-11-18 03:27:06 | boomerang-io/roadmap | https://api.github.com/repos/boomerang-io/roadmap | closed | Task View Result Parameters not showing | bug priority: high project: flow | **Describe the bug**
Result Properties are not shown for a task in activity. UI Is trying to access `outputs` service is returning `results`
| 1.0 | Task View Result Parameters not showing - **Describe the bug**
Result Properties are not shown for a task in activity. UI Is trying to access `outputs` service is returning `results`
| non_main | task view result parameters not showing describe the bug result properties are not shown for a task in activity ui is trying to access outputs service is returning results | 0 |
4,188 | 20,376,866,690 | IssuesEvent | 2022-02-21 16:27:37 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | Request: Add indicator line - data table row component | type: enhancement 💡 component: data-table proposal: not pursuing status: waiting for maintainer response 💬 | ### Summary
In Watson Knowledge catalog we allow users to click a row to display meta data in context of the UI in a side panel. Users need an easy way to get information during their workflows, without losing context or navigating away from their current screens. Therefore we have to indicate the row in a table that was selected. The suer needs to have some visual hint for the currently displayed item.
Here examples:


This would require an indicator line as addition to the current data table row component (dev +design):
> blue indicator line 2 px on the far right of the row (#0F62FE, Blue 60)
See also: https://ibm.box.com/s/eyfek094u0r1tihcr42qt8xvzfmqd37b
### Justification
https://ibm.box.com/s/j4n4kru1wzuuw1g0jj0zcdj9hnn51yq4
This relatively small feature differentiates us from our competitors, and is a clear need coming out directly from user research. It is a pain point that was found and solved in our tools in the past. I the past this was solved by large tooltips. The difference now is that times change, tooltips do not provide the accessibility we want or the real estate for all use cases so we reworked it into an already familiar solution.To make sure people know which item is selected on the page an indicator is required for the specific row item in a table.
### Specific timeline issues / requests
Release May Cloud Pak for Data, Watson Knowledge Catalog
| True | Request: Add indicator line - data table row component - ### Summary
In Watson Knowledge catalog we allow users to click a row to display meta data in context of the UI in a side panel. Users need an easy way to get information during their workflows, without losing context or navigating away from their current screens. Therefore we have to indicate the row in a table that was selected. The suer needs to have some visual hint for the currently displayed item.
Here examples:


This would require an indicator line as addition to the current data table row component (dev +design):
> blue indicator line 2 px on the far right of the row (#0F62FE, Blue 60)
See also: https://ibm.box.com/s/eyfek094u0r1tihcr42qt8xvzfmqd37b
### Justification
https://ibm.box.com/s/j4n4kru1wzuuw1g0jj0zcdj9hnn51yq4
This relatively small feature differentiates us from our competitors, and is a clear need coming out directly from user research. It is a pain point that was found and solved in our tools in the past. I the past this was solved by large tooltips. The difference now is that times change, tooltips do not provide the accessibility we want or the real estate for all use cases so we reworked it into an already familiar solution.To make sure people know which item is selected on the page an indicator is required for the specific row item in a table.
### Specific timeline issues / requests
Release May Cloud Pak for Data, Watson Knowledge Catalog
| main | request add indicator line data table row component summary in watson knowledge catalog we allow users to click a row to display meta data in context of the ui in a side panel users need an easy way to get information during their workflows without losing context or navigating away from their current screens therefore we have to indicate the row in a table that was selected the suer needs to have some visual hint for the currently displayed item here examples this would require an indicator line as addition to the current data table row component dev design blue indicator line px on the far right of the row blue see also justification this relatively small feature differentiates us from our competitors and is a clear need coming out directly from user research it is a pain point that was found and solved in our tools in the past i the past this was solved by large tooltips the difference now is that times change tooltips do not provide the accessibility we want or the real estate for all use cases so we reworked it into an already familiar solution to make sure people know which item is selected on the page an indicator is required for the specific row item in a table specific timeline issues requests release may cloud pak for data watson knowledge catalog | 1 |
1,587 | 6,572,365,298 | IssuesEvent | 2017-09-11 01:45:02 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | nmcli: Missing deps, and a missing verification for that deps | affects_2.1 bug_report networking waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
nmclie module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
N/A
##### SUMMARY
from IRC
misc: that's a missing deps, and a missing verification for that deps
<!--- Explain the problem briefly -->
```
- name: Make {{ freeipa_ip }} the only nameserver
nmcli: conn_name=eth0 dns4=["{{ freeipa_ip }}"] state=present
fatal: [test]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_xHM2AZ/ansible_module_nmcli.py\", line 385, in <module>\r\n from gi.repository import NetworkManager, NMClient\r\nImportError: cannot import name NetworkManager\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
```
| True | nmcli: Missing deps, and a missing verification for that deps - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
nmclie module
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
N/A
##### SUMMARY
from IRC
misc: that's a missing deps, and a missing verification for that deps
<!--- Explain the problem briefly -->
```
- name: Make {{ freeipa_ip }} the only nameserver
nmcli: conn_name=eth0 dns4=["{{ freeipa_ip }}"] state=present
fatal: [test]: FAILED! => {"changed": false, "failed": true, "module_stderr": "", "module_stdout": "Traceback (most recent call last):\r\n File \"/tmp/ansible_xHM2AZ/ansible_module_nmcli.py\", line 385, in <module>\r\n from gi.repository import NetworkManager, NMClient\r\nImportError: cannot import name NetworkManager\r\n", "msg": "MODULE FAILURE", "parsed": false}
```
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
```
| main | nmcli missing deps and a missing verification for that deps issue type bug report component name nmclie module ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific n a summary from irc misc that s a missing deps and a missing verification for that deps name make freeipa ip the only nameserver nmcli conn name state present fatal failed changed false failed true module stderr module stdout traceback most recent call last r n file tmp ansible ansible module nmcli py line in r n from gi repository import networkmanager nmclient r nimporterror cannot import name networkmanager r n msg module failure parsed false steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used expected results actual results | 1 |
122,385 | 26,123,932,128 | IssuesEvent | 2022-12-28 15:54:47 | Clueless-Community/seamless-ui | https://api.github.com/repos/Clueless-Community/seamless-ui | closed | Improve Store Buttons/src/store-button-google-play-darkmode.html | codepeak 22 issue:1 | Need to improve this component on `Store Buttons/src/store-button-google-play-darkmode.html`
add more sizes
## Improvements required :

## Any reference image?

## Any external reference?
You can checkout [this]() one for better understanding.
| 1.0 | Improve Store Buttons/src/store-button-google-play-darkmode.html - Need to improve this component on `Store Buttons/src/store-button-google-play-darkmode.html`
add more sizes
## Improvements required :

## Any reference image?

## Any external reference?
You can checkout [this]() one for better understanding.
| non_main | improve store buttons src store button google play darkmode html need to improve this component on store buttons src store button google play darkmode html add more sizes improvements required any reference image any external reference you can checkout one for better understanding | 0 |
3,512 | 13,725,260,411 | IssuesEvent | 2020-10-03 17:43:39 | amyjko/faculty | https://api.github.com/repos/amyjko/faculty | closed | Merge reviewing service into a single reviewing list | content maintainability | Just show different levels. And add some optional descriptions of changes I made in leadership positions, especially program chairing. | True | Merge reviewing service into a single reviewing list - Just show different levels. And add some optional descriptions of changes I made in leadership positions, especially program chairing. | main | merge reviewing service into a single reviewing list just show different levels and add some optional descriptions of changes i made in leadership positions especially program chairing | 1 |
2,997 | 10,924,807,259 | IssuesEvent | 2019-11-22 10:57:22 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | opened | Do not assign null to listitems when iterating over a list | Area: analyzer Area: maintainability feature | It may happen that developers want to clear a list and its contents. In such cases you can see code like following:
```C#
var list = new List<object>();
foreach (var item in list)
{
item = null;
}
list.ForEach(item => item = null);
list.ForEach(item => { item = null });
```
That code does not do what the developer tries to acchieve.
The developer tries to set `null` inside the list, so that the list will only contain a number of `null` values.
However, as `item` is always a copy, only the copy (reference) will be set to `null` and not the original value.
So we should warn about it. | True | Do not assign null to listitems when iterating over a list - It may happen that developers want to clear a list and its contents. In such cases you can see code like following:
```C#
var list = new List<object>();
foreach (var item in list)
{
item = null;
}
list.ForEach(item => item = null);
list.ForEach(item => { item = null });
```
That code does not do what the developer tries to acchieve.
The developer tries to set `null` inside the list, so that the list will only contain a number of `null` values.
However, as `item` is always a copy, only the copy (reference) will be set to `null` and not the original value.
So we should warn about it. | main | do not assign null to listitems when iterating over a list it may happen that developers want to clear a list and its contents in such cases you can see code like following c var list new list foreach var item in list item null list foreach item item null list foreach item item null that code does not do what the developer tries to acchieve the developer tries to set null inside the list so that the list will only contain a number of null values however as item is always a copy only the copy reference will be set to null and not the original value so we should warn about it | 1 |
793 | 4,390,064,153 | IssuesEvent | 2016-08-09 01:05:40 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | reopened | File mode 0755 fails for a tar in `unarchive` | bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
Module: `unarchive`
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.0.0
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
`unarchive` fails with a `mode` of `0755`.
##### STEPS TO REPRODUCE
Create a `tar.gz` file (or any format that will be processed by [`TgzArchive`](https://github.com/ansible/ansible-modules-core/blob/devel/files/unarchive.py#L537)):
```
$ mkdir mydir
$ tar -czf mydir.tar.gz mydir
```
Use a playbook similar to the following (the important bit is the file mode):
```
- hosts: localhost
tasks:
- name: extract dir
unarchive:
src: "./mydir.tar.gz"
dest: "./"
mode: 0755
```
##### EXPECTED RESULTS
I would expect the playbook to run without error.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
$ ansible-playbook myplaybook.yml
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [extract dir] *************************************************************
fatal: [localhost]: FAILED! => {"changed": false, "dest": "./", "failed": true, "gid": 1000, "group": "jesse", "handler": "TgzArchive", "mode": "0775", "msg": "Unexpected error when accessing exploded file: [Errno 2] No such file or directory: './mydir'", "owner": "jesse", "size": 4096, "src": "/home/jesse/.ansible/tmp/ansible-tmp-1467149922.64-77975041845860/source", "state": "directory", "uid": 1000}
NO MORE HOSTS LEFT *************************************************************
to retry, use: --limit @myplaybook.retry
PLAY RECAP *********************************************************************
localhost : ok=1 changed=0 unreachable=0 failed=1
```
Running the same playbook, but replacing the file mode `0755` with `0777` runs without error:
```
$ ansible-playbook myplaybook.yml
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [extract dir] *************************************************************
changed: [localhost]
PLAY RECAP *********************************************************************
localhost : ok=2 changed=1 unreachable=0 failed=0
```
The error message in itself does not tell us a lot, but if we add some debug traces ([`check_results`](https://github.com/ansible/ansible-modules-core/blob/devel/files/unarchive.py#L778)), we see that the following `tar` command is being executed:
```
/bin/tar -C "./" -dz --mode="493" -f "/home/jesse/.ansible/tmp/ansible-tmp-1467150531.52-232348981801230/source"
```
and that `tar` is not happy about it:
```
/bin/tar: Invalid mode given on option
/bin/tar: Error is not recoverable: exiting now
```
The integer `493` passed to tar as a mode happens to be the decimal representation of `0755`. Tar complains because this is an invalid octal number for the `mode`. The file mode is being passed as decimal because it is passed to `tar` after casting it directly to a string (i.e. results in a decimal representation) in [`is_unarchived`](https://github.com/ansible/ansible-modules-core/blob/devel/files/unarchive.py#L585) and [`unarchive`](https://github.com/ansible/ansible-modules-core/blob/devel/files/unarchive.py#L628).
However, it looks like `tar` does not even care about `--mode` on `--diff` or `--extract` (emphasis is mine):
> `--mode=permissions'
**When adding files to an archive**, tar will use permissions for the archive members, rather than the permissions from the files.
My understanding is that passing this parameter should not have an impact on the files handled by `-d` or `-x`. This would explain why the issue was hard to encounter; only a file mode that would yield an invalid octal number *after* a conversion to a decimal number would make it fail. It does not have an effect on the final mode set.
Is there a particular reason for passing `--mode` when calling `tar -x` or `tar -d`?
If this is indeed a bug and I am not misconfiguring my playbook in some way, the workaround is rather simple -- using a symbolic representation for the file mode works like a charm.
If possible, I would be pleased to work on a PR to fix the issue in order to get involved in the project and its contribution process. | True | File mode 0755 fails for a tar in `unarchive` - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
Module: `unarchive`
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.1.0.0
```
##### CONFIGURATION
N/A
##### OS / ENVIRONMENT
N/A
##### SUMMARY
`unarchive` fails with a `mode` of `0755`.
##### STEPS TO REPRODUCE
Create a `tar.gz` file (or any format that will be processed by [`TgzArchive`](https://github.com/ansible/ansible-modules-core/blob/devel/files/unarchive.py#L537)):
```
$ mkdir mydir
$ tar -czf mydir.tar.gz mydir
```
Use a playbook similar to the following (the important bit is the file mode):
```
- hosts: localhost
tasks:
- name: extract dir
unarchive:
src: "./mydir.tar.gz"
dest: "./"
mode: 0755
```
##### EXPECTED RESULTS
I would expect the playbook to run without error.
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
$ ansible-playbook myplaybook.yml
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [extract dir] *************************************************************
fatal: [localhost]: FAILED! => {"changed": false, "dest": "./", "failed": true, "gid": 1000, "group": "jesse", "handler": "TgzArchive", "mode": "0775", "msg": "Unexpected error when accessing exploded file: [Errno 2] No such file or directory: './mydir'", "owner": "jesse", "size": 4096, "src": "/home/jesse/.ansible/tmp/ansible-tmp-1467149922.64-77975041845860/source", "state": "directory", "uid": 1000}
NO MORE HOSTS LEFT *************************************************************
to retry, use: --limit @myplaybook.retry
PLAY RECAP *********************************************************************
localhost : ok=1 changed=0 unreachable=0 failed=1
```
Running the same playbook, but replacing the file mode `0755` with `0777` runs without error:
```
$ ansible-playbook myplaybook.yml
PLAY [localhost] ***************************************************************
TASK [setup] *******************************************************************
ok: [localhost]
TASK [extract dir] *************************************************************
changed: [localhost]
PLAY RECAP *********************************************************************
localhost : ok=2 changed=1 unreachable=0 failed=0
```
The error message in itself does not tell us a lot, but if we add some debug traces ([`check_results`](https://github.com/ansible/ansible-modules-core/blob/devel/files/unarchive.py#L778)), we see that the following `tar` command is being executed:
```
/bin/tar -C "./" -dz --mode="493" -f "/home/jesse/.ansible/tmp/ansible-tmp-1467150531.52-232348981801230/source"
```
and that `tar` is not happy about it:
```
/bin/tar: Invalid mode given on option
/bin/tar: Error is not recoverable: exiting now
```
The integer `493` passed to tar as a mode happens to be the decimal representation of `0755`. Tar complains because this is an invalid octal number for the `mode`. The file mode is being passed as decimal because it is passed to `tar` after casting it directly to a string (i.e. results in a decimal representation) in [`is_unarchived`](https://github.com/ansible/ansible-modules-core/blob/devel/files/unarchive.py#L585) and [`unarchive`](https://github.com/ansible/ansible-modules-core/blob/devel/files/unarchive.py#L628).
However, it looks like `tar` does not even care about `--mode` on `--diff` or `--extract` (emphasis is mine):
> `--mode=permissions'
**When adding files to an archive**, tar will use permissions for the archive members, rather than the permissions from the files.
My understanding is that passing this parameter should not have an impact on the files handled by `-d` or `-x`. This would explain why the issue was hard to encounter; only a file mode that would yield an invalid octal number *after* a conversion to a decimal number would make it fail. It does not have an effect on the final mode set.
Is there a particular reason for passing `--mode` when calling `tar -x` or `tar -d`?
If this is indeed a bug and I am not misconfiguring my playbook in some way, the workaround is rather simple -- using a symbolic representation for the file mode works like a charm.
If possible, I would be pleased to work on a PR to fix the issue in order to get involved in the project and its contribution process. | main | file mode fails for a tar in unarchive issue type bug report component name module unarchive ansible version ansible configuration n a os environment n a summary unarchive fails with a mode of steps to reproduce create a tar gz file or any format that will be processed by mkdir mydir tar czf mydir tar gz mydir use a playbook similar to the following the important bit is the file mode hosts localhost tasks name extract dir unarchive src mydir tar gz dest mode expected results i would expect the playbook to run without error actual results ansible playbook myplaybook yml play task ok task fatal failed changed false dest failed true gid group jesse handler tgzarchive mode msg unexpected error when accessing exploded file no such file or directory mydir owner jesse size src home jesse ansible tmp ansible tmp source state directory uid no more hosts left to retry use limit myplaybook retry play recap localhost ok changed unreachable failed running the same playbook but replacing the file mode with runs without error ansible playbook myplaybook yml play task ok task changed play recap localhost ok changed unreachable failed the error message in itself does not tell us a lot but if we add some debug traces we see that the following tar command is being executed bin tar c dz mode f home jesse ansible tmp ansible tmp source and that tar is not happy about it bin tar invalid mode given on option bin tar error is not recoverable exiting now the integer passed to tar as a mode happens to be the decimal representation of tar complains because this is an invalid octal number for the mode the file mode is being passed as decimal because it is passed to tar after casting it directly to a string i e results in a decimal representation in and however it looks like tar does not even care about mode on diff or extract emphasis is mine mode permissions when adding files to an archive tar will use permissions for the archive members rather than the permissions from the files my understanding is that passing this parameter should not have an impact on the files handled by d or x this would explain why the issue was hard to encounter only a file mode that would yield an invalid octal number after a conversion to a decimal number would make it fail it does not have an effect on the final mode set is there a particular reason for passing mode when calling tar x or tar d if this is indeed a bug and i am not misconfiguring my playbook in some way the workaround is rather simple using a symbolic representation for the file mode works like a charm if possible i would be pleased to work on a pr to fix the issue in order to get involved in the project and its contribution process | 1 |
1,358 | 5,865,047,597 | IssuesEvent | 2017-05-13 00:19:53 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | Install Chocolatey from Internal Source | affects_2.3 feature_idea waiting_on_maintainer windows | Was asked if this was available and it doesn't appear it is from looking at the source. @nitzmahone
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
win_chocolatey
##### ANSIBLE VERSION
##### CONFIGURATION
##### OS / ENVIRONMENT
Windows
##### SUMMARY
Folks would like to be able to have Chocolatey installed from internal sources when using the Ansible module, particularly when they are completely offline.
This would be a good capability added to the module. It is already supported by other config mgmt tools so those could provide good references on how to add it to the chocolatey ansible module.
| True | Install Chocolatey from Internal Source - Was asked if this was available and it doesn't appear it is from looking at the source. @nitzmahone
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
win_chocolatey
##### ANSIBLE VERSION
##### CONFIGURATION
##### OS / ENVIRONMENT
Windows
##### SUMMARY
Folks would like to be able to have Chocolatey installed from internal sources when using the Ansible module, particularly when they are completely offline.
This would be a good capability added to the module. It is already supported by other config mgmt tools so those could provide good references on how to add it to the chocolatey ansible module.
| main | install chocolatey from internal source was asked if this was available and it doesn t appear it is from looking at the source nitzmahone issue type feature idea component name win chocolatey ansible version configuration os environment windows summary folks would like to be able to have chocolatey installed from internal sources when using the ansible module particularly when they are completely offline this would be a good capability added to the module it is already supported by other config mgmt tools so those could provide good references on how to add it to the chocolatey ansible module | 1 |
1,721 | 6,574,493,498 | IssuesEvent | 2017-09-11 13:05:52 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | dellos9_command ansible hangs after reload command issued to remote device. | affects_2.2 bug_report networking waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
dellos9_command
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /home/emarq/Solutions.Network.Automation/MAS/Ansible/dell/force10/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
```
---
- name: AGG base config
hosts: baseagg
vars:
node: agg
cli:
host: "{{ ansible_host }}"
transport: cli
username: admin
ssh_keyfile: /srv/tftpboot/my-rsa.pub
roles:
- deployconfig
roles/deployconfig/
├── tasks
│ └── main.yml
- name: reload
dellos9_command:
provider: "{{ cli }}"
commands: "reload no-confirm"
```
##### OS / ENVIRONMENT
```
Linux ansible 4.4.0-45-generic #66-Ubuntu SMP Wed Oct 19 14:12:37 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
```
##### SUMMARY
Create a task where the dellos9_command can issue a remote reload of the switch. Ansible will never timeout after the switch has rebooted.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
ansible-playbook masdbaseconfig.yml --limit s6000 -vvvvv
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
After waiting 60+ minutes for the job to fail, I had to issue a ctrl+c to break the process.
TASK [deployconfig : reload] ***************************************************
task path: /home/emarq/Solutions.Network.Automation/MAS/Ansible/dell/force10/roles/deployconfig/tasks/main.yml:31
Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/network/dellos9/dellos9_command.py
<10.10.234.96> ESTABLISH LOCAL CONNECTION FOR USER: emarq
<10.10.234.96> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675 `" && echo ansible-tmp-1478116357.69-28845854088675="` echo $HOME/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675 `" ) && sleep 0'
<10.10.234.96> PUT /tmp/tmpWKMxtR TO /home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/dellos9_command.py
<10.10.234.96> EXEC /bin/sh -c 'chmod u+x /home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/ /home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/dellos9_command.py && sleep 0'
<10.10.234.96> EXEC /bin/sh -c '/usr/bin/python /home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/dellos9_command.py; rm -rf "/home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/" > /dev/null 2>&1 && sleep 0'
CProcess WorkerProcess-19:
Traceback (most recent call last):
File "/usr/lib/python2.7/multiprocessing/process.py", line 258, in _bootstrap
self.run()
File "/usr/lib/python2.7/dist-packages/ansible/executor/process/worker.py", line 112, in run
self._rslt_q
File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 119, in run
res = self._execute()
File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 490, in _execute
result = self._handler.run(task_vars=variables)
File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/normal.py", line 33, in run
results = merge_hash(results, self._execute_module(tmp=tmp, task_vars=task_vars))
File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/__init__.py", line 643, in _execute_module
res = self._low_level_execute_command(cmd, sudoable=sudoable, in_data=in_data)
File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/__init__.py", line 754, in _low_level_execute_command
rc, stdout, stderr = self._connection.exec_command(cmd, in_data=in_data, sudoable=sudoable)
File "/usr/lib/python2.7/dist-packages/ansible/plugins/connection/local.py", line 114, in exec_command
stdout, stderr = p.communicate(in_data)
File "/usr/lib/python2.7/subprocess.py", line 800, in communicate
return self._communicate(input)
File "/usr/lib/python2.7/subprocess.py", line 1417, in _communicate
stdout, stderr = self._communicate_with_poll(input)
File "/usr/lib/python2.7/subprocess.py", line 1471, in _communicate_with_poll
ready = poller.poll()
KeyboardInterrupt
[ERROR]: User interrupted execution
```
| True | dellos9_command ansible hangs after reload command issued to remote device. - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
dellos9_command
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0.0
config file = /home/emarq/Solutions.Network.Automation/MAS/Ansible/dell/force10/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
```
---
- name: AGG base config
hosts: baseagg
vars:
node: agg
cli:
host: "{{ ansible_host }}"
transport: cli
username: admin
ssh_keyfile: /srv/tftpboot/my-rsa.pub
roles:
- deployconfig
roles/deployconfig/
├── tasks
│ └── main.yml
- name: reload
dellos9_command:
provider: "{{ cli }}"
commands: "reload no-confirm"
```
##### OS / ENVIRONMENT
```
Linux ansible 4.4.0-45-generic #66-Ubuntu SMP Wed Oct 19 14:12:37 UTC 2016 x86_64 x86_64 x86_64 GNU/Linux
```
##### SUMMARY
Create a task where the dellos9_command can issue a remote reload of the switch. Ansible will never timeout after the switch has rebooted.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
ansible-playbook masdbaseconfig.yml --limit s6000 -vvvvv
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
After waiting 60+ minutes for the job to fail, I had to issue a ctrl+c to break the process.
TASK [deployconfig : reload] ***************************************************
task path: /home/emarq/Solutions.Network.Automation/MAS/Ansible/dell/force10/roles/deployconfig/tasks/main.yml:31
Using module file /usr/lib/python2.7/dist-packages/ansible/modules/core/network/dellos9/dellos9_command.py
<10.10.234.96> ESTABLISH LOCAL CONNECTION FOR USER: emarq
<10.10.234.96> EXEC /bin/sh -c '( umask 77 && mkdir -p "` echo $HOME/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675 `" && echo ansible-tmp-1478116357.69-28845854088675="` echo $HOME/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675 `" ) && sleep 0'
<10.10.234.96> PUT /tmp/tmpWKMxtR TO /home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/dellos9_command.py
<10.10.234.96> EXEC /bin/sh -c 'chmod u+x /home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/ /home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/dellos9_command.py && sleep 0'
<10.10.234.96> EXEC /bin/sh -c '/usr/bin/python /home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/dellos9_command.py; rm -rf "/home/emarq/.ansible/tmp/ansible-tmp-1478116357.69-28845854088675/" > /dev/null 2>&1 && sleep 0'
CProcess WorkerProcess-19:
Traceback (most recent call last):
File "/usr/lib/python2.7/multiprocessing/process.py", line 258, in _bootstrap
self.run()
File "/usr/lib/python2.7/dist-packages/ansible/executor/process/worker.py", line 112, in run
self._rslt_q
File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 119, in run
res = self._execute()
File "/usr/lib/python2.7/dist-packages/ansible/executor/task_executor.py", line 490, in _execute
result = self._handler.run(task_vars=variables)
File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/normal.py", line 33, in run
results = merge_hash(results, self._execute_module(tmp=tmp, task_vars=task_vars))
File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/__init__.py", line 643, in _execute_module
res = self._low_level_execute_command(cmd, sudoable=sudoable, in_data=in_data)
File "/usr/lib/python2.7/dist-packages/ansible/plugins/action/__init__.py", line 754, in _low_level_execute_command
rc, stdout, stderr = self._connection.exec_command(cmd, in_data=in_data, sudoable=sudoable)
File "/usr/lib/python2.7/dist-packages/ansible/plugins/connection/local.py", line 114, in exec_command
stdout, stderr = p.communicate(in_data)
File "/usr/lib/python2.7/subprocess.py", line 800, in communicate
return self._communicate(input)
File "/usr/lib/python2.7/subprocess.py", line 1417, in _communicate
stdout, stderr = self._communicate_with_poll(input)
File "/usr/lib/python2.7/subprocess.py", line 1471, in _communicate_with_poll
ready = poller.poll()
KeyboardInterrupt
[ERROR]: User interrupted execution
```
| main | command ansible hangs after reload command issued to remote device issue type bug report component name command ansible version ansible config file home emarq solutions network automation mas ansible dell ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables name agg base config hosts baseagg vars node agg cli host ansible host transport cli username admin ssh keyfile srv tftpboot my rsa pub roles deployconfig roles deployconfig ├── tasks │ └── main yml name reload command provider cli commands reload no confirm os environment linux ansible generic ubuntu smp wed oct utc gnu linux summary create a task where the command can issue a remote reload of the switch ansible will never timeout after the switch has rebooted steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used ansible playbook masdbaseconfig yml limit vvvvv expected results actual results after waiting minutes for the job to fail i had to issue a ctrl c to break the process task task path home emarq solutions network automation mas ansible dell roles deployconfig tasks main yml using module file usr lib dist packages ansible modules core network command py establish local connection for user emarq exec bin sh c umask mkdir p echo home ansible tmp ansible tmp echo ansible tmp echo home ansible tmp ansible tmp sleep put tmp tmpwkmxtr to home emarq ansible tmp ansible tmp command py exec bin sh c chmod u x home emarq ansible tmp ansible tmp home emarq ansible tmp ansible tmp command py sleep exec bin sh c usr bin python home emarq ansible tmp ansible tmp command py rm rf home emarq ansible tmp ansible tmp dev null sleep cprocess workerprocess traceback most recent call last file usr lib multiprocessing process py line in bootstrap self run file usr lib dist packages ansible executor process worker py line in run self rslt q file usr lib dist packages ansible executor task executor py line in run res self execute file usr lib dist packages ansible executor task executor py line in execute result self handler run task vars variables file usr lib dist packages ansible plugins action normal py line in run results merge hash results self execute module tmp tmp task vars task vars file usr lib dist packages ansible plugins action init py line in execute module res self low level execute command cmd sudoable sudoable in data in data file usr lib dist packages ansible plugins action init py line in low level execute command rc stdout stderr self connection exec command cmd in data in data sudoable sudoable file usr lib dist packages ansible plugins connection local py line in exec command stdout stderr p communicate in data file usr lib subprocess py line in communicate return self communicate input file usr lib subprocess py line in communicate stdout stderr self communicate with poll input file usr lib subprocess py line in communicate with poll ready poller poll keyboardinterrupt user interrupted execution | 1 |
858 | 4,525,197,099 | IssuesEvent | 2016-09-07 03:18:34 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | ec2_asg_facts: Find by name using wrong key in search | aws bug_report cloud waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ec2_asg_facts
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0 (devel 4457985062) last updated 2016/09/02 09:55:19 (GMT -400)
lib/ansible/modules/core: (detached HEAD 7e79c59d38) last updated 2016/09/02 09:58:46 (GMT -400)
lib/ansible/modules/extras: (detached HEAD e8a5442345) last updated 2016/09/02 09:58:47 (GMT -400)
config file = /Users/jeffreyearl/Projects/mps-platform/ansible-infrastructure/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
N/A
##### SUMMARY
<!--- Explain the problem briefly -->
When using the 'name' parameter for this module, the results from boto3 are being searched with the snake_case key before the dictionary has been converted for the results.
To fix, line 305 of the module:
```
matched_name = name_prog.search(asg['auto_scaling_group_name'])
```
should be
```
matched_name = name_prog.search(asg['AutoScalingGroupName'])
```
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Add this task to any playbook setup to use AWS:
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Get existing ASG info
ec2_asg_facts:
name: "foo"
register: asgs
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Module should return asg's that have a name starting with "foo"
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
Exception
<!--- Paste verbatim command output between quotes below -->
```
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: KeyError: 'auto_scaling_group_name'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/var/folders/72/8wk3w0sn035cnsw147fzwzfh0000gn/T/ansible_qeZWU2/ansible_module_ec2_asg_facts.py\", line 351, in <module>\n main()\n File \"/var/folders/72/8wk3w0sn035cnsw147fzwzfh0000gn/T/ansible_qeZWU2/ansible_module_ec2_asg_facts.py\", line 343, in main\n results = find_asgs(autoscaling, module, name=asg_name, tags=asg_tags)\n File \"/var/folders/72/8wk3w0sn035cnsw147fzwzfh0000gn/T/ansible_qeZWU2/ansible_module_ec2_asg_facts.py\", line 305, in find_asgs\n matched_name = name_prog.search(asg['auto_scaling_group_name'])\nKeyError: 'auto_scaling_group_name'\n", "module_stdout": "", "msg": "MODULE FAILURE"}
```
| True | ec2_asg_facts: Find by name using wrong key in search - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
ec2_asg_facts
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
ansible 2.2.0 (devel 4457985062) last updated 2016/09/02 09:55:19 (GMT -400)
lib/ansible/modules/core: (detached HEAD 7e79c59d38) last updated 2016/09/02 09:58:46 (GMT -400)
lib/ansible/modules/extras: (detached HEAD e8a5442345) last updated 2016/09/02 09:58:47 (GMT -400)
config file = /Users/jeffreyearl/Projects/mps-platform/ansible-infrastructure/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
<!---
Mention any settings you have changed/added/removed in ansible.cfg
(or using the ANSIBLE_* environment variables).
-->
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
N/A
##### SUMMARY
<!--- Explain the problem briefly -->
When using the 'name' parameter for this module, the results from boto3 are being searched with the snake_case key before the dictionary has been converted for the results.
To fix, line 305 of the module:
```
matched_name = name_prog.search(asg['auto_scaling_group_name'])
```
should be
```
matched_name = name_prog.search(asg['AutoScalingGroupName'])
```
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
Add this task to any playbook setup to use AWS:
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Get existing ASG info
ec2_asg_facts:
name: "foo"
register: asgs
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
Module should return asg's that have a name starting with "foo"
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with high verbosity (-vvvv) -->
Exception
<!--- Paste verbatim command output between quotes below -->
```
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: KeyError: 'auto_scaling_group_name'
fatal: [localhost]: FAILED! => {"changed": false, "failed": true, "module_stderr": "Traceback (most recent call last):\n File \"/var/folders/72/8wk3w0sn035cnsw147fzwzfh0000gn/T/ansible_qeZWU2/ansible_module_ec2_asg_facts.py\", line 351, in <module>\n main()\n File \"/var/folders/72/8wk3w0sn035cnsw147fzwzfh0000gn/T/ansible_qeZWU2/ansible_module_ec2_asg_facts.py\", line 343, in main\n results = find_asgs(autoscaling, module, name=asg_name, tags=asg_tags)\n File \"/var/folders/72/8wk3w0sn035cnsw147fzwzfh0000gn/T/ansible_qeZWU2/ansible_module_ec2_asg_facts.py\", line 305, in find_asgs\n matched_name = name_prog.search(asg['auto_scaling_group_name'])\nKeyError: 'auto_scaling_group_name'\n", "module_stdout": "", "msg": "MODULE FAILURE"}
```
| main | asg facts find by name using wrong key in search issue type bug report component name asg facts ansible version ansible devel last updated gmt lib ansible modules core detached head last updated gmt lib ansible modules extras detached head last updated gmt config file users jeffreyearl projects mps platform ansible infrastructure ansible cfg configured module search path default w o overrides configuration mention any settings you have changed added removed in ansible cfg or using the ansible environment variables os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific n a summary when using the name parameter for this module the results from are being searched with the snake case key before the dictionary has been converted for the results to fix line of the module matched name name prog search asg should be matched name name prog search asg steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used add this task to any playbook setup to use aws name get existing asg info asg facts name foo register asgs expected results module should return asg s that have a name starting with foo actual results exception an exception occurred during task execution to see the full traceback use vvv the error was keyerror auto scaling group name fatal failed changed false failed true module stderr traceback most recent call last n file var folders t ansible ansible module asg facts py line in n main n file var folders t ansible ansible module asg facts py line in main n results find asgs autoscaling module name asg name tags asg tags n file var folders t ansible ansible module asg facts py line in find asgs n matched name name prog search asg nkeyerror auto scaling group name n module stdout msg module failure | 1 |
2,745 | 9,784,431,948 | IssuesEvent | 2019-06-08 19:09:15 | arcticicestudio/igloo | https://api.github.com/repos/arcticicestudio/igloo | closed | Link complete Taskwarrior "hooks" and "scripts" folders | scope-maintainability snowblock-taskwarrior type-improvement | Instead of linking single files from the `hooks` and `scripts` Taskwarrior folders both folders will be linked directly. | True | Link complete Taskwarrior "hooks" and "scripts" folders - Instead of linking single files from the `hooks` and `scripts` Taskwarrior folders both folders will be linked directly. | main | link complete taskwarrior hooks and scripts folders instead of linking single files from the hooks and scripts taskwarrior folders both folders will be linked directly | 1 |
420,331 | 28,244,975,217 | IssuesEvent | 2023-04-06 10:00:45 | juaml/julearn | https://api.github.com/repos/juaml/julearn | closed | [DOC] Example of transform_until | documentation | Create an example in which we can see all the transformation steps the data is going through before the classification algorithms. | 1.0 | [DOC] Example of transform_until - Create an example in which we can see all the transformation steps the data is going through before the classification algorithms. | non_main | example of transform until create an example in which we can see all the transformation steps the data is going through before the classification algorithms | 0 |
311,568 | 9,535,444,277 | IssuesEvent | 2019-04-30 06:53:24 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.slideshare.net - A Download window is opened when accessing the site | browser-firefox priority-critical severity-minor | <!-- @browser: Firefox Nightly 68.0a1 (2019-03-22) -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.86 Safari/537.36 -->
<!-- @reported_with: -->
**URL**: https://www.slideshare.net/
**Browser / Version**: Firefox Nightly 68.0a1 (2019-03-22)
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: A Download window is opened when accessing the site
**Steps to Reproduce**:
1. Navigate to: https://www.slideshare.net/
**Expected Behavior:**
The site is opened.
**Actual Behavior:**
The site is opened and a Download window is opened.
**Notes:**
1. The issue is not reproducible on Chrome.
2. Screenshot attached.
**Watchers:**
@softvision-oana-arbuzov
@softvision-sergiulogigan
@cipriansv
sv;
[](https://webcompat.com/uploads/2019/3/0dcf4a1f-9664-48f9-9fb8-285871f1f716.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.slideshare.net - A Download window is opened when accessing the site - <!-- @browser: Firefox Nightly 68.0a1 (2019-03-22) -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.86 Safari/537.36 -->
<!-- @reported_with: -->
**URL**: https://www.slideshare.net/
**Browser / Version**: Firefox Nightly 68.0a1 (2019-03-22)
**Operating System**: Windows 10
**Tested Another Browser**: Yes
**Problem type**: Something else
**Description**: A Download window is opened when accessing the site
**Steps to Reproduce**:
1. Navigate to: https://www.slideshare.net/
**Expected Behavior:**
The site is opened.
**Actual Behavior:**
The site is opened and a Download window is opened.
**Notes:**
1. The issue is not reproducible on Chrome.
2. Screenshot attached.
**Watchers:**
@softvision-oana-arbuzov
@softvision-sergiulogigan
@cipriansv
sv;
[](https://webcompat.com/uploads/2019/3/0dcf4a1f-9664-48f9-9fb8-285871f1f716.jpg)
<details>
<summary>Browser Configuration</summary>
<ul>
<li>None</li>
</ul>
</details>
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | a download window is opened when accessing the site url browser version firefox nightly operating system windows tested another browser yes problem type something else description a download window is opened when accessing the site steps to reproduce navigate to expected behavior the site is opened actual behavior the site is opened and a download window is opened notes the issue is not reproducible on chrome screenshot attached watchers softvision oana arbuzov softvision sergiulogigan cipriansv sv browser configuration none from with ❤️ | 0 |
534,025 | 15,607,670,083 | IssuesEvent | 2021-03-19 09:39:27 | OnTopicCMS/OnTopic-Library | https://api.github.com/repos/OnTopicCMS/OnTopic-Library | closed | Bug: CachedTopicRepository.Load() should return null if topicKey is empty | Area: Entity Priority: 0 Severity 2: Major Status 5: Complete Type: Bug | Currently, `CachedTopicRepository.Load()` will return the root topic if the `topicKey` is empty. An empty `topicKey` should be treated the same as a null `topicKey`, and result in a null `Topic` being returned. The current behavior could have a lot of unexpected consequences. | 1.0 | Bug: CachedTopicRepository.Load() should return null if topicKey is empty - Currently, `CachedTopicRepository.Load()` will return the root topic if the `topicKey` is empty. An empty `topicKey` should be treated the same as a null `topicKey`, and result in a null `Topic` being returned. The current behavior could have a lot of unexpected consequences. | non_main | bug cachedtopicrepository load should return null if topickey is empty currently cachedtopicrepository load will return the root topic if the topickey is empty an empty topickey should be treated the same as a null topickey and result in a null topic being returned the current behavior could have a lot of unexpected consequences | 0 |
799,060 | 28,300,565,241 | IssuesEvent | 2023-04-10 05:27:54 | googleapis/google-cloud-ruby | https://api.github.com/repos/googleapis/google-cloud-ruby | closed | [Nightly CI Failures] Failures detected for google-cloud-workflows | type: bug priority: p1 nightly failure | At 2023-04-09 08:51:49 UTC, detected failures in google-cloud-workflows for: yard
report_key_48d0c8774055c49008a9fc13a959feaf | 1.0 | [Nightly CI Failures] Failures detected for google-cloud-workflows - At 2023-04-09 08:51:49 UTC, detected failures in google-cloud-workflows for: yard
report_key_48d0c8774055c49008a9fc13a959feaf | non_main | failures detected for google cloud workflows at utc detected failures in google cloud workflows for yard report key | 0 |
106,969 | 9,198,944,524 | IssuesEvent | 2019-03-07 13:54:42 | NativeScript/nativescript-cli | https://api.github.com/repos/NativeScript/nativescript-cli | closed | Cannot run qunit tests | bug in testing | - CLI: 5.2.1
1. tns create TestApp --js
2. tns test init qunit
3. tns test android (or iOS)
Actual result:
Successfully synced application org.nativescript.TestApp on device emulator-5562.
?[32m01 03 2019 21:03:31.910:INFO [NativeScript / 23 (6.0; Android SDK built for x86)]: ?[39mConnected on socket pWvC_YFcNDTviXfRAAAA with id NativeScriptUnit-6647
?[33m01 03 2019 21:03:32.921:WARN [NativeScript / 23 (6.0; Android SDK built for x86)]: ?[39mAdapter did not report total number of specs.
NativeScript / 23 (6.0; Android SDK built for x86): Executed 0 of 0 SUCCESS (0 secs / 0 secs)
?[1A?[2KNativeScript / 23 (6.0; Android SDK built for x86) /base/node_modules/qunit/qunit/qunit.css?dd28ae0cce0104eca6ff99f54387788378943c3e at line 0 FAILED
SyntaxError: Invalid or unexpected token
NativeScript / 23 (6.0; Android SDK built for x86): Executed 1 of 0 (1 FAILED) (0 secs / 0 secs)
?[1A?[2KNativeScript / 23 (6.0; Android SDK built for x86): Executed 1 of 0 (1 FAILED) ERROR (0.121 secs / 0 secs)
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Test run failed. | 1.0 | Cannot run qunit tests - - CLI: 5.2.1
1. tns create TestApp --js
2. tns test init qunit
3. tns test android (or iOS)
Actual result:
Successfully synced application org.nativescript.TestApp on device emulator-5562.
?[32m01 03 2019 21:03:31.910:INFO [NativeScript / 23 (6.0; Android SDK built for x86)]: ?[39mConnected on socket pWvC_YFcNDTviXfRAAAA with id NativeScriptUnit-6647
?[33m01 03 2019 21:03:32.921:WARN [NativeScript / 23 (6.0; Android SDK built for x86)]: ?[39mAdapter did not report total number of specs.
NativeScript / 23 (6.0; Android SDK built for x86): Executed 0 of 0 SUCCESS (0 secs / 0 secs)
?[1A?[2KNativeScript / 23 (6.0; Android SDK built for x86) /base/node_modules/qunit/qunit/qunit.css?dd28ae0cce0104eca6ff99f54387788378943c3e at line 0 FAILED
SyntaxError: Invalid or unexpected token
NativeScript / 23 (6.0; Android SDK built for x86): Executed 1 of 0 (1 FAILED) (0 secs / 0 secs)
?[1A?[2KNativeScript / 23 (6.0; Android SDK built for x86): Executed 1 of 0 (1 FAILED) ERROR (0.121 secs / 0 secs)
Note: Some input files use or override a deprecated API.
Note: Recompile with -Xlint:deprecation for details.
Test run failed. | non_main | cannot run qunit tests cli tns create testapp js tns test init qunit tns test android or ios actual result successfully synced application org nativescript testapp on device emulator on socket pwvc yfcndtvixfraaaa with id nativescriptunit did not report total number of specs nativescript android sdk built for executed of success secs secs android sdk built for base node modules qunit qunit qunit css at line failed syntaxerror invalid or unexpected token nativescript android sdk built for executed of failed secs secs android sdk built for executed of failed error secs secs note some input files use or override a deprecated api note recompile with xlint deprecation for details test run failed | 0 |
473,920 | 13,649,312,027 | IssuesEvent | 2020-09-26 13:51:51 | ahmedkaludi/accelerated-mobile-pages | https://api.github.com/repos/ahmedkaludi/accelerated-mobile-pages | closed | Add custom classes in Body tag in AMP in all Designs | NEXT UPDATE [Priority: HIGH] enhancement | Add custom classes in Body tag in AMP in all Designs
1. for singular, add post-id
2. current version number of AMPforWP | 1.0 | Add custom classes in Body tag in AMP in all Designs - Add custom classes in Body tag in AMP in all Designs
1. for singular, add post-id
2. current version number of AMPforWP | non_main | add custom classes in body tag in amp in all designs add custom classes in body tag in amp in all designs for singular add post id current version number of ampforwp | 0 |
196,550 | 22,442,139,490 | IssuesEvent | 2022-06-21 02:34:21 | ConnectionMaster/create-probot-app | https://api.github.com/repos/ConnectionMaster/create-probot-app | opened | CVE-2022-33987 (Medium) detected in got-6.7.1.tgz, got-7.1.0.tgz | security vulnerability | ## CVE-2022-33987 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>got-6.7.1.tgz</b>, <b>got-7.1.0.tgz</b></p></summary>
<p>
<details><summary><b>got-6.7.1.tgz</b></p></summary>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-6.7.1.tgz">https://registry.npmjs.org/got/-/got-6.7.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/got/package.json,/node_modules/got/package.json</p>
<p>
Dependency Hierarchy:
- egad-0.2.0.tgz (Root Library)
- download-git-repo-1.1.0.tgz
- download-5.0.3.tgz
- :x: **got-6.7.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>got-7.1.0.tgz</b></p></summary>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-7.1.0.tgz">https://registry.npmjs.org/got/-/got-7.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/gh-got/node_modules/got/package.json</p>
<p>
Dependency Hierarchy:
- conjecture-0.1.2.tgz (Root Library)
- github-username-4.1.0.tgz
- gh-got-6.0.0.tgz
- :x: **got-7.1.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The got package before 12.1.0 for Node.js allows a redirect to a UNIX socket.
<p>Publish Date: 2022-06-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p>
<p>Release Date: 2022-06-18</p>
<p>Fix Resolution: got - 11.8.5,12.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-33987 (Medium) detected in got-6.7.1.tgz, got-7.1.0.tgz - ## CVE-2022-33987 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>got-6.7.1.tgz</b>, <b>got-7.1.0.tgz</b></p></summary>
<p>
<details><summary><b>got-6.7.1.tgz</b></p></summary>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-6.7.1.tgz">https://registry.npmjs.org/got/-/got-6.7.1.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/npm/node_modules/got/package.json,/node_modules/got/package.json</p>
<p>
Dependency Hierarchy:
- egad-0.2.0.tgz (Root Library)
- download-git-repo-1.1.0.tgz
- download-5.0.3.tgz
- :x: **got-6.7.1.tgz** (Vulnerable Library)
</details>
<details><summary><b>got-7.1.0.tgz</b></p></summary>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-7.1.0.tgz">https://registry.npmjs.org/got/-/got-7.1.0.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/gh-got/node_modules/got/package.json</p>
<p>
Dependency Hierarchy:
- conjecture-0.1.2.tgz (Root Library)
- github-username-4.1.0.tgz
- gh-got-6.0.0.tgz
- :x: **got-7.1.0.tgz** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The got package before 12.1.0 for Node.js allows a redirect to a UNIX socket.
<p>Publish Date: 2022-06-18
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-33987>CVE-2022-33987</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p>
<p>Release Date: 2022-06-18</p>
<p>Fix Resolution: got - 11.8.5,12.1.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_main | cve medium detected in got tgz got tgz cve medium severity vulnerability vulnerable libraries got tgz got tgz got tgz simplified http requests library home page a href path to dependency file package json path to vulnerable library node modules npm node modules got package json node modules got package json dependency hierarchy egad tgz root library download git repo tgz download tgz x got tgz vulnerable library got tgz simplified http requests library home page a href path to dependency file package json path to vulnerable library node modules gh got node modules got package json dependency hierarchy conjecture tgz root library github username tgz gh got tgz x got tgz vulnerable library found in base branch master vulnerability details the got package before for node js allows a redirect to a unix socket publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution got step up your open source security game with mend | 0 |
1,669 | 6,574,071,177 | IssuesEvent | 2017-09-11 11:21:14 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | EC2 instance_initiated_shutdown_behavior default should be more intelligent | affects_2.2 aws bug_report cloud waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
EC2
##### ANSIBLE VERSION
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = ['/usr/share/ansible']
```
##### OS / ENVIRONMENT
N/A
##### SUMMARY
When creating an instance-store backed instance, the default of "stop" on instance_initiated_shutdown_behavior is not a valid default. Explicitly setting to "terminate" resolves the issue.
##### STEPS TO REPRODUCE
Running a playbook with the following task
<!--- Paste example playbooks or commands between quotes below -->
```
- name: create web servers instance 1
ec2:
image: "{{ image }}"
instance_type: "{{ instance_type }}"
keypair: "{{ keypair }}"
instance_tags:
Name: "{{ role }}-v2-01"
service: "tn"
region: "{{ region }}"
zone: "{{ region }}a"
group: "{{ aws_security_group }}"
wait: true
monitoring: no
exact_count: 1
count_tag:
Name: "{{ role }}-v2-01"
register: ec2_info
```
##### EXPECTED RESULTS
Instance to be created
##### ACTUAL RESULTS
Playbook bombs out
```
fatal: [localhost]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_args": {
"assign_public_ip": false,
"aws_access_key": null,
"aws_secret_key": null,
"count": 1,
"count_tag": "{'Name': 'talis-com-server-v2-01'}",
"ebs_optimized": false,
"ec2_url": null,
"exact_count": 1,
"group": [
"staging",
"talis.com-auto-v2"
],
"group_id": null,
"id": null,
"image": "ami-bd36b0ce",
"instance_ids": null,
"instance_initiated_shutdown_behavior": null,
"instance_profile_name": null,
"instance_tags": {
"Name": "talis-com-server-v2-01",
"service": "tn"
},
"instance_type": "m1.small",
"kernel": null,
"key_name": "keypair",
"keypair": "keypair",
"monitoring": false,
"network_interfaces": null,
"placement_group": null,
"private_ip": null,
"profile": null,
"ramdisk": null,
"region": "eu-west-1",
"security_token": null,
"source_dest_check": true,
"spot_launch_group": null,
"spot_price": null,
"spot_type": "one-time",
"spot_wait_timeout": "600",
"state": "present",
"tenancy": "default",
"termination_protection": null,
"user_data": null,
"validate_certs": true,
"volumes": null,
"vpc_subnet_id": null,
"wait": true,
"wait_timeout": "300",
"zone": "eu-west-1a"
},
"module_name": "ec2"
},
"msg": "Instance creation failed => InvalidParameterCombination: The attribute instanceInitiatedShutdownBehavior can only be used for EBS-backed images."
}
```
| True | EC2 instance_initiated_shutdown_behavior default should be more intelligent - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
EC2
##### ANSIBLE VERSION
```
ansible 2.2.0.0
config file = /etc/ansible/ansible.cfg
configured module search path = ['/usr/share/ansible']
```
##### OS / ENVIRONMENT
N/A
##### SUMMARY
When creating an instance-store backed instance, the default of "stop" on instance_initiated_shutdown_behavior is not a valid default. Explicitly setting to "terminate" resolves the issue.
##### STEPS TO REPRODUCE
Running a playbook with the following task
<!--- Paste example playbooks or commands between quotes below -->
```
- name: create web servers instance 1
ec2:
image: "{{ image }}"
instance_type: "{{ instance_type }}"
keypair: "{{ keypair }}"
instance_tags:
Name: "{{ role }}-v2-01"
service: "tn"
region: "{{ region }}"
zone: "{{ region }}a"
group: "{{ aws_security_group }}"
wait: true
monitoring: no
exact_count: 1
count_tag:
Name: "{{ role }}-v2-01"
register: ec2_info
```
##### EXPECTED RESULTS
Instance to be created
##### ACTUAL RESULTS
Playbook bombs out
```
fatal: [localhost]: FAILED! => {
"changed": false,
"failed": true,
"invocation": {
"module_args": {
"assign_public_ip": false,
"aws_access_key": null,
"aws_secret_key": null,
"count": 1,
"count_tag": "{'Name': 'talis-com-server-v2-01'}",
"ebs_optimized": false,
"ec2_url": null,
"exact_count": 1,
"group": [
"staging",
"talis.com-auto-v2"
],
"group_id": null,
"id": null,
"image": "ami-bd36b0ce",
"instance_ids": null,
"instance_initiated_shutdown_behavior": null,
"instance_profile_name": null,
"instance_tags": {
"Name": "talis-com-server-v2-01",
"service": "tn"
},
"instance_type": "m1.small",
"kernel": null,
"key_name": "keypair",
"keypair": "keypair",
"monitoring": false,
"network_interfaces": null,
"placement_group": null,
"private_ip": null,
"profile": null,
"ramdisk": null,
"region": "eu-west-1",
"security_token": null,
"source_dest_check": true,
"spot_launch_group": null,
"spot_price": null,
"spot_type": "one-time",
"spot_wait_timeout": "600",
"state": "present",
"tenancy": "default",
"termination_protection": null,
"user_data": null,
"validate_certs": true,
"volumes": null,
"vpc_subnet_id": null,
"wait": true,
"wait_timeout": "300",
"zone": "eu-west-1a"
},
"module_name": "ec2"
},
"msg": "Instance creation failed => InvalidParameterCombination: The attribute instanceInitiatedShutdownBehavior can only be used for EBS-backed images."
}
```
| main | instance initiated shutdown behavior default should be more intelligent issue type bug report component name ansible version ansible config file etc ansible ansible cfg configured module search path os environment n a summary when creating an instance store backed instance the default of stop on instance initiated shutdown behavior is not a valid default explicitly setting to terminate resolves the issue steps to reproduce running a playbook with the following task name create web servers instance image image instance type instance type keypair keypair instance tags name role service tn region region zone region a group aws security group wait true monitoring no exact count count tag name role register info expected results instance to be created actual results playbook bombs out fatal failed changed false failed true invocation module args assign public ip false aws access key null aws secret key null count count tag name talis com server ebs optimized false url null exact count group staging talis com auto group id null id null image ami instance ids null instance initiated shutdown behavior null instance profile name null instance tags name talis com server service tn instance type small kernel null key name keypair keypair keypair monitoring false network interfaces null placement group null private ip null profile null ramdisk null region eu west security token null source dest check true spot launch group null spot price null spot type one time spot wait timeout state present tenancy default termination protection null user data null validate certs true volumes null vpc subnet id null wait true wait timeout zone eu west module name msg instance creation failed invalidparametercombination the attribute instanceinitiatedshutdownbehavior can only be used for ebs backed images | 1 |
155,145 | 13,612,719,023 | IssuesEvent | 2020-09-23 10:44:08 | parallaxsecond/parsec-book | https://api.github.com/repos/parallaxsecond/parsec-book | closed | Make sure the right info is in the right place | documentation enhancement | Information regarding the same topic should be in the same location in the book, otherwise it is hard to maintain.
Have a look at the book and try to regroup information where it should be. | 1.0 | Make sure the right info is in the right place - Information regarding the same topic should be in the same location in the book, otherwise it is hard to maintain.
Have a look at the book and try to regroup information where it should be. | non_main | make sure the right info is in the right place information regarding the same topic should be in the same location in the book otherwise it is hard to maintain have a look at the book and try to regroup information where it should be | 0 |
3,546 | 14,016,889,151 | IssuesEvent | 2020-10-29 15:01:50 | luksan47/mars | https://api.github.com/repos/luksan47/mars | opened | Release Urán 3.5 | maintainer needed | Collecting the upgrade steps here:
- Create a backup.
- Follow steps in #495 to create API token.
- Test print job cancellation with long documents. | True | Release Urán 3.5 - Collecting the upgrade steps here:
- Create a backup.
- Follow steps in #495 to create API token.
- Test print job cancellation with long documents. | main | release urán collecting the upgrade steps here create a backup follow steps in to create api token test print job cancellation with long documents | 1 |
96,872 | 12,168,637,628 | IssuesEvent | 2020-04-27 13:00:39 | EscolaDeSaudePublica/iSaude | https://api.github.com/repos/EscolaDeSaudePublica/iSaude | opened | Criar tela de Boas Vindas | Design | Fazer pelo menos dois slides apresentando o aplicativo e suas áreas, para ser apresentado na primeira vez que for aberto. | 1.0 | Criar tela de Boas Vindas - Fazer pelo menos dois slides apresentando o aplicativo e suas áreas, para ser apresentado na primeira vez que for aberto. | non_main | criar tela de boas vindas fazer pelo menos dois slides apresentando o aplicativo e suas áreas para ser apresentado na primeira vez que for aberto | 0 |
30,189 | 8,490,009,208 | IssuesEvent | 2018-10-26 22:06:59 | orbeon/orbeon-forms | https://api.github.com/repos/orbeon/orbeon-forms | closed | Form Builder language selector no longer floats to the right | Form Builder Regression | <img width="988" alt="screen shot 2018-10-23 at 9 40 01 am" src="https://user-images.githubusercontent.com/105769/47376587-dff41d00-d6a7-11e8-80b7-9c838cee2a40.png">
| 1.0 | Form Builder language selector no longer floats to the right - <img width="988" alt="screen shot 2018-10-23 at 9 40 01 am" src="https://user-images.githubusercontent.com/105769/47376587-dff41d00-d6a7-11e8-80b7-9c838cee2a40.png">
| non_main | form builder language selector no longer floats to the right img width alt screen shot at am src | 0 |
2,650 | 9,083,347,076 | IssuesEvent | 2019-02-17 19:45:39 | lrozenblyum/chess | https://api.github.com/repos/lrozenblyum/chess | opened | Enforce Maven3 usage | CI maintainability | Caused by #47
Discovered during #277
We should enforce maven3 usage in the project. Now we don't have forced validation of Maven.
Command execution
versions:display-plugin-updates
shows
[ERROR] Project does not define required minimum version of Maven.
[ERROR] Update the pom.xml to contain maven-enforcer-plugin to
[ERROR] force the Maven version which is needed to build this project.
[ERROR] See https://maven.apache.org/enforcer/enforcer-rules/requireMavenVersion.html
[ERROR] Using the minimum version of Maven: 3.0.5
| True | Enforce Maven3 usage - Caused by #47
Discovered during #277
We should enforce maven3 usage in the project. Now we don't have forced validation of Maven.
Command execution
versions:display-plugin-updates
shows
[ERROR] Project does not define required minimum version of Maven.
[ERROR] Update the pom.xml to contain maven-enforcer-plugin to
[ERROR] force the Maven version which is needed to build this project.
[ERROR] See https://maven.apache.org/enforcer/enforcer-rules/requireMavenVersion.html
[ERROR] Using the minimum version of Maven: 3.0.5
| main | enforce usage caused by discovered during we should enforce usage in the project now we don t have forced validation of maven command execution versions display plugin updates shows project does not define required minimum version of maven update the pom xml to contain maven enforcer plugin to force the maven version which is needed to build this project see using the minimum version of maven | 1 |
4,246 | 21,045,370,874 | IssuesEvent | 2022-03-31 15:34:00 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | StateMachine Events should support ScheduleExpression | area/event-source type/feature stage/needs-feedback stage/pm-review maintainer/need-response area/step-function/event-source | Currently, the only way to execute a Step Function based on an EventBridge schedule is by using `AWS::Events::Rule`. It would be great to simplify this by specifying a `ScheduleExpression` directly under `Type: EventBridgeRule` like this:
```
MyStateMachine:
Type: AWS::Serverless::StateMachine
Properties:
DefinitionUri: statemachine/mystatemachine.asl.json
DefinitionSubstitutions:
MyFunctionArn: !GetAtt MyFunction.Arn
Events:
EventBridgeRule:
Type: EventBridgeRule
Properties:
Description: Run my Step Function every 10 mins
ScheduleExpression: "rate(10 minutes)"
```
This should also spare me from having to create and wire up corresponding roles to grant perms like `states:StartExecution` to `events.amazonaws.com` and so forth. | True | StateMachine Events should support ScheduleExpression - Currently, the only way to execute a Step Function based on an EventBridge schedule is by using `AWS::Events::Rule`. It would be great to simplify this by specifying a `ScheduleExpression` directly under `Type: EventBridgeRule` like this:
```
MyStateMachine:
Type: AWS::Serverless::StateMachine
Properties:
DefinitionUri: statemachine/mystatemachine.asl.json
DefinitionSubstitutions:
MyFunctionArn: !GetAtt MyFunction.Arn
Events:
EventBridgeRule:
Type: EventBridgeRule
Properties:
Description: Run my Step Function every 10 mins
ScheduleExpression: "rate(10 minutes)"
```
This should also spare me from having to create and wire up corresponding roles to grant perms like `states:StartExecution` to `events.amazonaws.com` and so forth. | main | statemachine events should support scheduleexpression currently the only way to execute a step function based on an eventbridge schedule is by using aws events rule it would be great to simplify this by specifying a scheduleexpression directly under type eventbridgerule like this mystatemachine type aws serverless statemachine properties definitionuri statemachine mystatemachine asl json definitionsubstitutions myfunctionarn getatt myfunction arn events eventbridgerule type eventbridgerule properties description run my step function every mins scheduleexpression rate minutes this should also spare me from having to create and wire up corresponding roles to grant perms like states startexecution to events amazonaws com and so forth | 1 |
5,379 | 27,037,525,241 | IssuesEvent | 2023-02-12 23:28:07 | EMS-TU-Ilmenau/fastmat | https://api.github.com/repos/EMS-TU-Ilmenau/fastmat | opened | Migrate CI from travis-ci.org to travis-ci.com | maintainance | Seems that travis-ci.org is deprecated and should be migrated to the service under TLD .com | True | Migrate CI from travis-ci.org to travis-ci.com - Seems that travis-ci.org is deprecated and should be migrated to the service under TLD .com | main | migrate ci from travis ci org to travis ci com seems that travis ci org is deprecated and should be migrated to the service under tld com | 1 |
2,926 | 10,437,056,264 | IssuesEvent | 2019-09-17 21:01:32 | backdrop-ops/contrib | https://api.github.com/repos/backdrop-ops/contrib | closed | Port Request: role_delegation module | Maintainer application | I would like to get the [Role Delegation](https://www.drupal.org/project/role_delegation) module ported. I have created [this issue](https://www.drupal.org/project/role_delegation/issues/3077306) on Drupal.org.
I have done the[ initial port](https://github.com/kelizoliva/role_delegation) which went mostly fine, but I am running into issues with how users are handled in Backdrop vs. Drupal 7. Would love to work with someone seeing this through, if anyone would like to help. I and Palante are happy to continue to maintain the module going forward, as we hope to be using it long-term in a large deployment project here at Palante.
Would also like this ticket to serve as my application to join the Backdrop Contrib group! I am generally interested in helping to move modules along where I can! | True | Port Request: role_delegation module - I would like to get the [Role Delegation](https://www.drupal.org/project/role_delegation) module ported. I have created [this issue](https://www.drupal.org/project/role_delegation/issues/3077306) on Drupal.org.
I have done the[ initial port](https://github.com/kelizoliva/role_delegation) which went mostly fine, but I am running into issues with how users are handled in Backdrop vs. Drupal 7. Would love to work with someone seeing this through, if anyone would like to help. I and Palante are happy to continue to maintain the module going forward, as we hope to be using it long-term in a large deployment project here at Palante.
Would also like this ticket to serve as my application to join the Backdrop Contrib group! I am generally interested in helping to move modules along where I can! | main | port request role delegation module i would like to get the module ported i have created on drupal org i have done the which went mostly fine but i am running into issues with how users are handled in backdrop vs drupal would love to work with someone seeing this through if anyone would like to help i and palante are happy to continue to maintain the module going forward as we hope to be using it long term in a large deployment project here at palante would also like this ticket to serve as my application to join the backdrop contrib group i am generally interested in helping to move modules along where i can | 1 |
4,976 | 25,547,925,885 | IssuesEvent | 2022-11-29 20:34:19 | bazelbuild/intellij | https://api.github.com/repos/bazelbuild/intellij | closed | Can't run this project with 2022.3 | type: bug product: IntelliJ lang: java awaiting-maintainer | ### Description of the bug:
_No response_
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
1. Import this project using IntelliJ 2022.3
2. Follow the guide described in [DEV_IDE_SETUP.md](https://github.com/bazelbuild/intellij/blob/0edc361d5906f5edbb1b9002cd7ad425a317c89c/DEV_IDE_SETUP.md)
3. Click the Debug button (spawning the "Run Plugin" run configuration)
4. Open any Bazel project
5. Open "Project structure" dialog
### Which Intellij IDE are you using? Please provide the specific version.
_No response_
### What programming languages and tools are you using? Please provide specific versions.
_No response_
### What Bazel plugin version are you using?
_No response_
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_ | True | Can't run this project with 2022.3 - ### Description of the bug:
_No response_
### What's the simplest, easiest way to reproduce this bug? Please provide a minimal example if possible.
1. Import this project using IntelliJ 2022.3
2. Follow the guide described in [DEV_IDE_SETUP.md](https://github.com/bazelbuild/intellij/blob/0edc361d5906f5edbb1b9002cd7ad425a317c89c/DEV_IDE_SETUP.md)
3. Click the Debug button (spawning the "Run Plugin" run configuration)
4. Open any Bazel project
5. Open "Project structure" dialog
### Which Intellij IDE are you using? Please provide the specific version.
_No response_
### What programming languages and tools are you using? Please provide specific versions.
_No response_
### What Bazel plugin version are you using?
_No response_
### Have you found anything relevant by searching the web?
_No response_
### Any other information, logs, or outputs that you want to share?
_No response_ | main | can t run this project with description of the bug no response what s the simplest easiest way to reproduce this bug please provide a minimal example if possible import this project using intellij follow the guide described in click the debug button spawning the run plugin run configuration open any bazel project open project structure dialog which intellij ide are you using please provide the specific version no response what programming languages and tools are you using please provide specific versions no response what bazel plugin version are you using no response have you found anything relevant by searching the web no response any other information logs or outputs that you want to share no response | 1 |
41,210 | 5,315,679,560 | IssuesEvent | 2017-02-13 17:55:34 | nextcloud/ios | https://api.github.com/repos/nextcloud/ios | opened | Simplify »Upload / New« menu | 1. to develop design enhancement | Currently clicking the **+** for upload / new folder opens a huge list of things:

Instead it should only be a small popover with »Upload photo« and »Create folder«. »New note« is also useful. And last an entry for »More …«. That would streamline the experience here. | 1.0 | Simplify »Upload / New« menu - Currently clicking the **+** for upload / new folder opens a huge list of things:

Instead it should only be a small popover with »Upload photo« and »Create folder«. »New note« is also useful. And last an entry for »More …«. That would streamline the experience here. | non_main | simplify »upload new« menu currently clicking the for upload new folder opens a huge list of things instead it should only be a small popover with »upload photo« and »create folder« »new note« is also useful and last an entry for »more …« that would streamline the experience here | 0 |
2,692 | 9,396,179,714 | IssuesEvent | 2019-04-08 06:20:06 | RalfKoban/MiKo-Analyzers | https://api.github.com/repos/RalfKoban/MiKo-Analyzers | closed | Methods that return 'Task' of 'Task<T>' should never return null | Area: analyzer Area: maintainability feature review | If methods return a task, then they should never return null. They should return a completed task instead.
Note: Be aware of `async` methods. They are allowed to return null because the generated statemachine will return a task. | True | Methods that return 'Task' of 'Task<T>' should never return null - If methods return a task, then they should never return null. They should return a completed task instead.
Note: Be aware of `async` methods. They are allowed to return null because the generated statemachine will return a task. | main | methods that return task of task should never return null if methods return a task then they should never return null they should return a completed task instead note be aware of async methods they are allowed to return null because the generated statemachine will return a task | 1 |
5,397 | 27,115,667,349 | IssuesEvent | 2023-02-15 18:21:53 | VA-Explorer/va_explorer | https://api.github.com/repos/VA-Explorer/va_explorer | closed | Require confirmation when changing VAs | Type: Maintainance good first issue Domain: Frontend Status: Inactive | **What is the expected state?**
As a user with permissions to change VAs, I expect the UI to prompt me if I'm sure I'd like to change the VA by either reverting an edit or by saving a new edit.
**What is the actual state?**
When reverting an edit, the process is immediately applied without confirmation
**Relevant context**
- `va_explorer/templates/va_data_management/edit.html`
- `va_explorer/va_data_management/views.py::Edit`
- `va_explorer/va_data_management/views.py::Reset`
| True | Require confirmation when changing VAs - **What is the expected state?**
As a user with permissions to change VAs, I expect the UI to prompt me if I'm sure I'd like to change the VA by either reverting an edit or by saving a new edit.
**What is the actual state?**
When reverting an edit, the process is immediately applied without confirmation
**Relevant context**
- `va_explorer/templates/va_data_management/edit.html`
- `va_explorer/va_data_management/views.py::Edit`
- `va_explorer/va_data_management/views.py::Reset`
| main | require confirmation when changing vas what is the expected state as a user with permissions to change vas i expect the ui to prompt me if i m sure i d like to change the va by either reverting an edit or by saving a new edit what is the actual state when reverting an edit the process is immediately applied without confirmation relevant context va explorer templates va data management edit html va explorer va data management views py edit va explorer va data management views py reset | 1 |
191,381 | 6,828,543,733 | IssuesEvent | 2017-11-08 20:48:00 | Aubron/scoreshots-templates | https://api.github.com/repos/Aubron/scoreshots-templates | opened | Towson, Literally A Calendar | Priority: Low Status: Review Needed | ### Requested by:
Towson University
Probably Public
## Template Description:
This is a scheduling template. It's actually in the shape of a calendar.
We do technically have the ability to make templates of these dimensions now, so the last real barrier to making these is gone. There are some unique UI concerns that we'd have to address in order to tackle it, though.
In defense of the requester, a lot of teams do have this sort of structure built into their website. This seems less like an image you'd want to view within a limited post, and more like one you might want to distribute and reshare.
## Dynamic Considerations:
This will need to be scalable. I wasn't given a specific minimum/maximum value (though I did ask), so for now we'll assume the maximum value is 12, with 6 per row, as this is what is displayed in the preview.
## Additional Materials

| 1.0 | Towson, Literally A Calendar - ### Requested by:
Towson University
Probably Public
## Template Description:
This is a scheduling template. It's actually in the shape of a calendar.
We do technically have the ability to make templates of these dimensions now, so the last real barrier to making these is gone. There are some unique UI concerns that we'd have to address in order to tackle it, though.
In defense of the requester, a lot of teams do have this sort of structure built into their website. This seems less like an image you'd want to view within a limited post, and more like one you might want to distribute and reshare.
## Dynamic Considerations:
This will need to be scalable. I wasn't given a specific minimum/maximum value (though I did ask), so for now we'll assume the maximum value is 12, with 6 per row, as this is what is displayed in the preview.
## Additional Materials

| non_main | towson literally a calendar requested by towson university probably public template description this is a scheduling template it s actually in the shape of a calendar we do technically have the ability to make templates of these dimensions now so the last real barrier to making these is gone there are some unique ui concerns that we d have to address in order to tackle it though in defense of the requester a lot of teams do have this sort of structure built into their website this seems less like an image you d want to view within a limited post and more like one you might want to distribute and reshare dynamic considerations this will need to be scalable i wasn t given a specific minimum maximum value though i did ask so for now we ll assume the maximum value is with per row as this is what is displayed in the preview additional materials | 0 |
4,171 | 19,985,487,367 | IssuesEvent | 2022-01-30 15:47:41 | BioArchLinux/Packages | https://api.github.com/repos/BioArchLinux/Packages | opened | [MAINTAIN] r-lpsymphony | maintain | <!--
Please report the error of one package in one issue! Use multi issues to report multi bugs.
Thanks!
-->
**Log of the bug**
<details>
```
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinSnapshot.cpp -fPIC -DPIC -o CoinSnapshot.o
/bin/sh ../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c -o CoinWarmStartBasis.lo CoinWarmStartBasis.cpp
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinWarmStartBasis.cpp -fPIC -DPIC -o CoinWarmStartBasis.o
/bin/sh ../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c -o CoinWarmStartVector.lo CoinWarmStartVector.cpp
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinWarmStartVector.cpp -fPIC -DPIC -o CoinWarmStartVector.o
/bin/sh ../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c -o CoinWarmStartDual.lo CoinWarmStartDual.cpp
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinWarmStartDual.cpp -fPIC -DPIC -o CoinWarmStartDual.o
/bin/sh ../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c -o CoinWarmStartPrimalDual.lo CoinWarmStartPrimalDual.cpp
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinWarmStartPrimalDual.cpp -fPIC -DPIC -o CoinWarmStartPrimalDual.o
/bin/sh ../../libtool --tag=CXX --mode=link g++ -w -g -O2 -DCOINUTILS_BUILD -Wl,-O1,--sort-common,--as-needed,-z,relro,-z,now -o libCoinUtils.la -rpath /tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/lib -no-undefined -version-info 13:6:10 CoinAlloc.lo CoinBuild.lo CoinDenseVector.lo CoinError.lo CoinFactorization1.lo CoinFactorization2.lo CoinFactorization3.lo CoinFactorization4.lo CoinSimpFactorization.lo CoinDenseFactorization.lo CoinOslFactorization.lo CoinOslFactorization2.lo CoinOslFactorization3.lo CoinFileIO.lo CoinFinite.lo CoinIndexedVector.lo CoinLpIO.lo CoinMessage.lo CoinMessageHandler.lo CoinModel.lo CoinStructuredModel.lo CoinModelUseful.lo CoinModelUseful2.lo CoinMpsIO.lo CoinPackedMatrix.lo CoinPackedVector.lo CoinPackedVectorBase.lo CoinParam.lo CoinParamUtils.lo CoinPostsolveMatrix.lo CoinPrePostsolveMatrix.lo CoinPresolveDoubleton.lo CoinPresolveDual.lo CoinPresolveDupcol.lo CoinPresolveEmpty.lo CoinPresolveFixed.lo CoinPresolveForcing.lo CoinPresolveHelperFunctions.lo CoinPresolveImpliedFree.lo CoinPresolveIsolated.lo CoinPresolveMatrix.lo CoinPresolvePsdebug.lo CoinPresolveMonitor.lo CoinPresolveSingleton.lo CoinPresolveSubst.lo CoinPresolveTighten.lo CoinPresolveTripleton.lo CoinPresolveUseless.lo CoinPresolveZeros.lo CoinRational.lo CoinSearchTree.lo CoinShallowPackedVector.lo CoinSnapshot.lo CoinWarmStartBasis.lo CoinWarmStartVector.lo CoinWarmStartDual.lo CoinWarmStartPrimalDual.lo -lm
mkdir .libs
ar cru .libs/libCoinUtils.a CoinAlloc.o CoinBuild.o CoinDenseVector.o CoinError.o CoinFactorization1.o CoinFactorization2.o CoinFactorization3.o CoinFactorization4.o CoinSimpFactorization.o CoinDenseFactorization.o CoinOslFactorization.o CoinOslFactorization2.o CoinOslFactorization3.o CoinFileIO.o CoinFinite.o CoinIndexedVector.o CoinLpIO.o CoinMessage.o CoinMessageHandler.o CoinModel.o CoinStructuredModel.o CoinModelUseful.o CoinModelUseful2.o CoinMpsIO.o CoinPackedMatrix.o CoinPackedVector.o CoinPackedVectorBase.o CoinParam.o CoinParamUtils.o CoinPostsolveMatrix.o CoinPrePostsolveMatrix.o CoinPresolveDoubleton.o CoinPresolveDual.o CoinPresolveDupcol.o CoinPresolveEmpty.o CoinPresolveFixed.o CoinPresolveForcing.o CoinPresolveHelperFunctions.o CoinPresolveImpliedFree.o CoinPresolveIsolated.o CoinPresolveMatrix.o CoinPresolvePsdebug.o CoinPresolveMonitor.o CoinPresolveSingleton.o CoinPresolveSubst.o CoinPresolveTighten.o CoinPresolveTripleton.o CoinPresolveUseless.o CoinPresolveZeros.o CoinRational.o CoinSearchTree.o CoinShallowPackedVector.o CoinSnapshot.o CoinWarmStartBasis.o CoinWarmStartVector.o CoinWarmStartDual.o CoinWarmStartPrimalDual.o
ar: `u' modifier ignored since `D' is the default (see `U')
ranlib .libs/libCoinUtils.a
creating libCoinUtils.la
(cd .libs && rm -f libCoinUtils.la && ln -s ../libCoinUtils.la libCoinUtils.la)
make[4]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils/src'
make[3]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils/src'
make[3]: Entering directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils'
make[3]: Nothing to be done for 'all-am'.
make[3]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils'
make[2]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils'
Making all in Osi
make[2]: Entering directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi'
Making all in src/Osi
make[3]: Entering directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi/src/Osi'
make all-am
make[4]: Entering directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi/src/Osi'
/bin/sh ../../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -I/tmp/RtmprwBDNn/Rbuild12daa92b38f71c/lpsymphony/src/SYMPHONY/include/coin -D_FORTIFY_SOURCE=2 -w -g -O2 -DOSI_BUILD -c -o OsiAuxInfo.lo OsiAuxInfo.cpp
g++ -DHAVE_CONFIG_H -I. -I. -I/tmp/RtmprwBDNn/Rbuild12daa92b38f71c/lpsymphony/src/SYMPHONY/include/coin -D_FORTIFY_SOURCE=2 -w -g -O2 -DOSI_BUILD -c OsiAuxInfo.cpp -fPIC -DPIC -o OsiAuxInfo.o
OsiAuxInfo.cpp:10:10: fatal error: CoinPragma.hpp: No such file or directory
10 | #include "CoinPragma.hpp"
| ^~~~~~~~~~~~~~~~
compilation terminated.
make[4]: *** [Makefile:535: OsiAuxInfo.lo] Error 1
make[4]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi/src/Osi'
make[3]: *** [Makefile:407: all] Error 2
make[3]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi/src/Osi'
make[2]: *** [Makefile:532: all-recursive] Error 1
make[2]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi'
make[1]: *** [Makefile:324: all-recursive] Error 1
make[1]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY'
make: *** [Makevars:13: SYMPHONY.ts] Error 2
ERROR: compilation failed for package ‘lpsymphony’
```
</details>
**Packages (please complete the following information):**
- Package Name: r-lpsymphony
**Description**
https://log.bioarchlinux.org/2022-01-29T15%3A51%3A16/r-lpsymphony.log
| True | [MAINTAIN] r-lpsymphony - <!--
Please report the error of one package in one issue! Use multi issues to report multi bugs.
Thanks!
-->
**Log of the bug**
<details>
```
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinSnapshot.cpp -fPIC -DPIC -o CoinSnapshot.o
/bin/sh ../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c -o CoinWarmStartBasis.lo CoinWarmStartBasis.cpp
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinWarmStartBasis.cpp -fPIC -DPIC -o CoinWarmStartBasis.o
/bin/sh ../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c -o CoinWarmStartVector.lo CoinWarmStartVector.cpp
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinWarmStartVector.cpp -fPIC -DPIC -o CoinWarmStartVector.o
/bin/sh ../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c -o CoinWarmStartDual.lo CoinWarmStartDual.cpp
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinWarmStartDual.cpp -fPIC -DPIC -o CoinWarmStartDual.o
/bin/sh ../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c -o CoinWarmStartPrimalDual.lo CoinWarmStartPrimalDual.cpp
g++ -DHAVE_CONFIG_H -I. -I. -D_FORTIFY_SOURCE=2 -w -g -O2 -DCOINUTILS_BUILD -c CoinWarmStartPrimalDual.cpp -fPIC -DPIC -o CoinWarmStartPrimalDual.o
/bin/sh ../../libtool --tag=CXX --mode=link g++ -w -g -O2 -DCOINUTILS_BUILD -Wl,-O1,--sort-common,--as-needed,-z,relro,-z,now -o libCoinUtils.la -rpath /tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/lib -no-undefined -version-info 13:6:10 CoinAlloc.lo CoinBuild.lo CoinDenseVector.lo CoinError.lo CoinFactorization1.lo CoinFactorization2.lo CoinFactorization3.lo CoinFactorization4.lo CoinSimpFactorization.lo CoinDenseFactorization.lo CoinOslFactorization.lo CoinOslFactorization2.lo CoinOslFactorization3.lo CoinFileIO.lo CoinFinite.lo CoinIndexedVector.lo CoinLpIO.lo CoinMessage.lo CoinMessageHandler.lo CoinModel.lo CoinStructuredModel.lo CoinModelUseful.lo CoinModelUseful2.lo CoinMpsIO.lo CoinPackedMatrix.lo CoinPackedVector.lo CoinPackedVectorBase.lo CoinParam.lo CoinParamUtils.lo CoinPostsolveMatrix.lo CoinPrePostsolveMatrix.lo CoinPresolveDoubleton.lo CoinPresolveDual.lo CoinPresolveDupcol.lo CoinPresolveEmpty.lo CoinPresolveFixed.lo CoinPresolveForcing.lo CoinPresolveHelperFunctions.lo CoinPresolveImpliedFree.lo CoinPresolveIsolated.lo CoinPresolveMatrix.lo CoinPresolvePsdebug.lo CoinPresolveMonitor.lo CoinPresolveSingleton.lo CoinPresolveSubst.lo CoinPresolveTighten.lo CoinPresolveTripleton.lo CoinPresolveUseless.lo CoinPresolveZeros.lo CoinRational.lo CoinSearchTree.lo CoinShallowPackedVector.lo CoinSnapshot.lo CoinWarmStartBasis.lo CoinWarmStartVector.lo CoinWarmStartDual.lo CoinWarmStartPrimalDual.lo -lm
mkdir .libs
ar cru .libs/libCoinUtils.a CoinAlloc.o CoinBuild.o CoinDenseVector.o CoinError.o CoinFactorization1.o CoinFactorization2.o CoinFactorization3.o CoinFactorization4.o CoinSimpFactorization.o CoinDenseFactorization.o CoinOslFactorization.o CoinOslFactorization2.o CoinOslFactorization3.o CoinFileIO.o CoinFinite.o CoinIndexedVector.o CoinLpIO.o CoinMessage.o CoinMessageHandler.o CoinModel.o CoinStructuredModel.o CoinModelUseful.o CoinModelUseful2.o CoinMpsIO.o CoinPackedMatrix.o CoinPackedVector.o CoinPackedVectorBase.o CoinParam.o CoinParamUtils.o CoinPostsolveMatrix.o CoinPrePostsolveMatrix.o CoinPresolveDoubleton.o CoinPresolveDual.o CoinPresolveDupcol.o CoinPresolveEmpty.o CoinPresolveFixed.o CoinPresolveForcing.o CoinPresolveHelperFunctions.o CoinPresolveImpliedFree.o CoinPresolveIsolated.o CoinPresolveMatrix.o CoinPresolvePsdebug.o CoinPresolveMonitor.o CoinPresolveSingleton.o CoinPresolveSubst.o CoinPresolveTighten.o CoinPresolveTripleton.o CoinPresolveUseless.o CoinPresolveZeros.o CoinRational.o CoinSearchTree.o CoinShallowPackedVector.o CoinSnapshot.o CoinWarmStartBasis.o CoinWarmStartVector.o CoinWarmStartDual.o CoinWarmStartPrimalDual.o
ar: `u' modifier ignored since `D' is the default (see `U')
ranlib .libs/libCoinUtils.a
creating libCoinUtils.la
(cd .libs && rm -f libCoinUtils.la && ln -s ../libCoinUtils.la libCoinUtils.la)
make[4]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils/src'
make[3]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils/src'
make[3]: Entering directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils'
make[3]: Nothing to be done for 'all-am'.
make[3]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils'
make[2]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/CoinUtils'
Making all in Osi
make[2]: Entering directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi'
Making all in src/Osi
make[3]: Entering directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi/src/Osi'
make all-am
make[4]: Entering directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi/src/Osi'
/bin/sh ../../../libtool --tag=CXX --mode=compile g++ -DHAVE_CONFIG_H -I. -I`echo .` -I/tmp/RtmprwBDNn/Rbuild12daa92b38f71c/lpsymphony/src/SYMPHONY/include/coin -D_FORTIFY_SOURCE=2 -w -g -O2 -DOSI_BUILD -c -o OsiAuxInfo.lo OsiAuxInfo.cpp
g++ -DHAVE_CONFIG_H -I. -I. -I/tmp/RtmprwBDNn/Rbuild12daa92b38f71c/lpsymphony/src/SYMPHONY/include/coin -D_FORTIFY_SOURCE=2 -w -g -O2 -DOSI_BUILD -c OsiAuxInfo.cpp -fPIC -DPIC -o OsiAuxInfo.o
OsiAuxInfo.cpp:10:10: fatal error: CoinPragma.hpp: No such file or directory
10 | #include "CoinPragma.hpp"
| ^~~~~~~~~~~~~~~~
compilation terminated.
make[4]: *** [Makefile:535: OsiAuxInfo.lo] Error 1
make[4]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi/src/Osi'
make[3]: *** [Makefile:407: all] Error 2
make[3]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi/src/Osi'
make[2]: *** [Makefile:532: all-recursive] Error 1
make[2]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY/Osi'
make[1]: *** [Makefile:324: all-recursive] Error 1
make[1]: Leaving directory '/tmp/Rtmprhkc3h/R.INSTALL409469cff59/lpsymphony/src/SYMPHONY'
make: *** [Makevars:13: SYMPHONY.ts] Error 2
ERROR: compilation failed for package ‘lpsymphony’
```
</details>
**Packages (please complete the following information):**
- Package Name: r-lpsymphony
**Description**
https://log.bioarchlinux.org/2022-01-29T15%3A51%3A16/r-lpsymphony.log
| main | r lpsymphony please report the error of one package in one issue use multi issues to report multi bugs thanks log of the bug g dhave config h i i d fortify source w g dcoinutils build c coinsnapshot cpp fpic dpic o coinsnapshot o bin sh libtool tag cxx mode compile g dhave config h i i echo d fortify source w g dcoinutils build c o coinwarmstartbasis lo coinwarmstartbasis cpp g dhave config h i i d fortify source w g dcoinutils build c coinwarmstartbasis cpp fpic dpic o coinwarmstartbasis o bin sh libtool tag cxx mode compile g dhave config h i i echo d fortify source w g dcoinutils build c o coinwarmstartvector lo coinwarmstartvector cpp g dhave config h i i d fortify source w g dcoinutils build c coinwarmstartvector cpp fpic dpic o coinwarmstartvector o bin sh libtool tag cxx mode compile g dhave config h i i echo d fortify source w g dcoinutils build c o coinwarmstartdual lo coinwarmstartdual cpp g dhave config h i i d fortify source w g dcoinutils build c coinwarmstartdual cpp fpic dpic o coinwarmstartdual o bin sh libtool tag cxx mode compile g dhave config h i i echo d fortify source w g dcoinutils build c o coinwarmstartprimaldual lo coinwarmstartprimaldual cpp g dhave config h i i d fortify source w g dcoinutils build c coinwarmstartprimaldual cpp fpic dpic o coinwarmstartprimaldual o bin sh libtool tag cxx mode link g w g dcoinutils build wl sort common as needed z relro z now o libcoinutils la rpath tmp r lpsymphony src symphony lib no undefined version info coinalloc lo coinbuild lo coindensevector lo coinerror lo lo lo lo lo coinsimpfactorization lo coindensefactorization lo coinoslfactorization lo lo lo coinfileio lo coinfinite lo coinindexedvector lo coinlpio lo coinmessage lo coinmessagehandler lo coinmodel lo coinstructuredmodel lo coinmodeluseful lo lo coinmpsio lo coinpackedmatrix lo coinpackedvector lo coinpackedvectorbase lo coinparam lo coinparamutils lo coinpostsolvematrix lo coinprepostsolvematrix lo coinpresolvedoubleton lo coinpresolvedual lo coinpresolvedupcol lo coinpresolveempty lo coinpresolvefixed lo coinpresolveforcing lo coinpresolvehelperfunctions lo coinpresolveimpliedfree lo coinpresolveisolated lo coinpresolvematrix lo coinpresolvepsdebug lo coinpresolvemonitor lo coinpresolvesingleton lo coinpresolvesubst lo coinpresolvetighten lo coinpresolvetripleton lo coinpresolveuseless lo coinpresolvezeros lo coinrational lo coinsearchtree lo coinshallowpackedvector lo coinsnapshot lo coinwarmstartbasis lo coinwarmstartvector lo coinwarmstartdual lo coinwarmstartprimaldual lo lm mkdir libs ar cru libs libcoinutils a coinalloc o coinbuild o coindensevector o coinerror o o o o o coinsimpfactorization o coindensefactorization o coinoslfactorization o o o coinfileio o coinfinite o coinindexedvector o coinlpio o coinmessage o coinmessagehandler o coinmodel o coinstructuredmodel o coinmodeluseful o o coinmpsio o coinpackedmatrix o coinpackedvector o coinpackedvectorbase o coinparam o coinparamutils o coinpostsolvematrix o coinprepostsolvematrix o coinpresolvedoubleton o coinpresolvedual o coinpresolvedupcol o coinpresolveempty o coinpresolvefixed o coinpresolveforcing o coinpresolvehelperfunctions o coinpresolveimpliedfree o coinpresolveisolated o coinpresolvematrix o coinpresolvepsdebug o coinpresolvemonitor o coinpresolvesingleton o coinpresolvesubst o coinpresolvetighten o coinpresolvetripleton o coinpresolveuseless o coinpresolvezeros o coinrational o coinsearchtree o coinshallowpackedvector o coinsnapshot o coinwarmstartbasis o coinwarmstartvector o coinwarmstartdual o coinwarmstartprimaldual o ar u modifier ignored since d is the default see u ranlib libs libcoinutils a creating libcoinutils la cd libs rm f libcoinutils la ln s libcoinutils la libcoinutils la make leaving directory tmp r lpsymphony src symphony coinutils src make leaving directory tmp r lpsymphony src symphony coinutils src make entering directory tmp r lpsymphony src symphony coinutils make nothing to be done for all am make leaving directory tmp r lpsymphony src symphony coinutils make leaving directory tmp r lpsymphony src symphony coinutils making all in osi make entering directory tmp r lpsymphony src symphony osi making all in src osi make entering directory tmp r lpsymphony src symphony osi src osi make all am make entering directory tmp r lpsymphony src symphony osi src osi bin sh libtool tag cxx mode compile g dhave config h i i echo i tmp rtmprwbdnn lpsymphony src symphony include coin d fortify source w g dosi build c o osiauxinfo lo osiauxinfo cpp g dhave config h i i i tmp rtmprwbdnn lpsymphony src symphony include coin d fortify source w g dosi build c osiauxinfo cpp fpic dpic o osiauxinfo o osiauxinfo cpp fatal error coinpragma hpp no such file or directory include coinpragma hpp compilation terminated make error make leaving directory tmp r lpsymphony src symphony osi src osi make error make leaving directory tmp r lpsymphony src symphony osi src osi make error make leaving directory tmp r lpsymphony src symphony osi make error make leaving directory tmp r lpsymphony src symphony make error error compilation failed for package ‘lpsymphony’ packages please complete the following information package name r lpsymphony description | 1 |
1,715 | 6,574,461,391 | IssuesEvent | 2017-09-11 12:58:59 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | junos_config broken configurations and idempotency | affects_2.2 bug_report networking waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
junos_config
##### ANSIBLE VERSION
```
ansible 2.2.0.0
config file = /home/centos/provisioning-metal/ansible.cfg
configured module search path = ['ephemeral_roles/plugins/library']
```
##### CONFIGURATION
Roles, plugins, inventory paths.
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
CentOS 7 managing a Juniper EX3300 stack running 12.3R8.7
##### SUMMARY
The junos_config deploys broken/inconsistent configurations every time I run it.
##### STEPS TO REPRODUCE
Starting with blank port configuration:
```
logan@stack01.rack92> show configuration interfaces ae11
{master:0}
logan@stack01.rack92> show configuration interfaces ge-0/0/6
{master:0}
logan@stack01.rack92> show configuration interfaces ge-1/0/6
{master:0}
```
The playbook used is the following:
```
- debug:
msg: "{{ switch_commands | default([]) }}"
when: "{{ not switch_commands_fact | skipped }}"
- name: Apply JunOS configuration
local_action:
module: junos_config
lines: "{{ switch_commands | default([]) }}"
provider: "{{ hostvars[switch_port_hostname][hostvars[switch_port_hostname]['network_provider']] }}"
when: "{{ not switch_commands_fact | skipped }}"
```
First run of playbook:
```
TASK [debug] *******************************************************************
ok: [lsn-mc1002] => {
"msg": [
"set interfaces vlan.2003 family inet address 10.3.8.1/24",
"set vlans ANSIBLE-MANAGEMENT vlan-id 2003",
"set vlans ANSIBLE-MANAGEMENT l3-interface vlan.2003",
"delete interfaces ae11",
"set interfaces ae11 description \"lsn-mc1002\"",
"set interfaces ae11 aggregated-ether-options lacp active",
"set interfaces ae11.0 family ethernet-switching port-mode access",
"set interfaces ae11.0 family ethernet-switching vlan members ANSIBLE-MANAGEMENT"
]
}
TASK [Apply JunOS configuration] ***********************************************
changed: [lsn-mc1002 -> localhost]
TASK [debug] *******************************************************************
ok: [lsn-mc1002] => {
"msg": [
"set interfaces vlan.2003 family inet address 10.3.8.1/24",
"set vlans ANSIBLE-MANAGEMENT vlan-id 2003",
"set vlans ANSIBLE-MANAGEMENT l3-interface vlan.2003",
"delete interfaces ge-0/0/6",
"set interfaces ge-0/0/6 description \"lsn-mc1002\"",
"set interfaces ge-0/0/6 ether-options 802.3ad ae11"
]
}
TASK [Apply JunOS configuration] ***********************************************
changed: [lsn-mc1002 -> localhost]
TASK [debug] *******************************************************************
ok: [lsn-mc1002] => {
"msg": [
"set interfaces vlan.2003 family inet address 10.3.8.1/24",
"set vlans ANSIBLE-MANAGEMENT vlan-id 2003",
"set vlans ANSIBLE-MANAGEMENT l3-interface vlan.2003",
"delete interfaces ge-1/0/6",
"set interfaces ge-1/0/6 description \"lsn-mc1002\"",
"set interfaces ge-1/0/6 ether-options 802.3ad ae11"
]
}
TASK [Apply JunOS configuration] ***********************************************
changed: [lsn-mc1002 -> localhost]
```
Resulting configuration from first run looks good:
```
logan@stack01.rack92> show configuration interfaces ae11
description lsn-mc1002;
aggregated-ether-options {
lacp {
active;
}
}
unit 0 {
family ethernet-switching {
port-mode access;
vlan {
members ANSIBLE-MANAGEMENT;
}
}
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-0/0/6
description lsn-mc1002;
ether-options {
802.3ad ae11;
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-1/0/6
description lsn-mc1002;
ether-options {
802.3ad ae11;
}
{master:0}
```
Second run. Ansible output is identical so I'll refrain from pasting it again.
```
logan@stack01.rack92> show configuration interfaces ae11
description lsn-mc1002;
unit 0 {
family ethernet-switching {
port-mode access;
vlan {
members ANSIBLE-MANAGEMENT;
}
}
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-0/0/6
description lsn-mc1002;
{master:0}
logan@stack01.rack92> show configuration interfaces ge-1/0/6
description lsn-mc1002;
{master:0}
```
3rd run:
```
logan@stack01.rack92> show configuration interfaces ae11
description lsn-mc1002;
aggregated-ether-options {
lacp {
active;
}
}
unit 0 {
family ethernet-switching {
port-mode access;
vlan {
members ANSIBLE-MANAGEMENT;
}
}
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-0/0/6
description lsn-mc1002;
ether-options {
802.3ad ae11;
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-1/0/6
description lsn-mc1002;
ether-options {
802.3ad ae11;
}
{master:0}
```
Subsequent runs oscillate back and forth between the broken and proper configurations. | True | junos_config broken configurations and idempotency - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
junos_config
##### ANSIBLE VERSION
```
ansible 2.2.0.0
config file = /home/centos/provisioning-metal/ansible.cfg
configured module search path = ['ephemeral_roles/plugins/library']
```
##### CONFIGURATION
Roles, plugins, inventory paths.
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
CentOS 7 managing a Juniper EX3300 stack running 12.3R8.7
##### SUMMARY
The junos_config deploys broken/inconsistent configurations every time I run it.
##### STEPS TO REPRODUCE
Starting with blank port configuration:
```
logan@stack01.rack92> show configuration interfaces ae11
{master:0}
logan@stack01.rack92> show configuration interfaces ge-0/0/6
{master:0}
logan@stack01.rack92> show configuration interfaces ge-1/0/6
{master:0}
```
The playbook used is the following:
```
- debug:
msg: "{{ switch_commands | default([]) }}"
when: "{{ not switch_commands_fact | skipped }}"
- name: Apply JunOS configuration
local_action:
module: junos_config
lines: "{{ switch_commands | default([]) }}"
provider: "{{ hostvars[switch_port_hostname][hostvars[switch_port_hostname]['network_provider']] }}"
when: "{{ not switch_commands_fact | skipped }}"
```
First run of playbook:
```
TASK [debug] *******************************************************************
ok: [lsn-mc1002] => {
"msg": [
"set interfaces vlan.2003 family inet address 10.3.8.1/24",
"set vlans ANSIBLE-MANAGEMENT vlan-id 2003",
"set vlans ANSIBLE-MANAGEMENT l3-interface vlan.2003",
"delete interfaces ae11",
"set interfaces ae11 description \"lsn-mc1002\"",
"set interfaces ae11 aggregated-ether-options lacp active",
"set interfaces ae11.0 family ethernet-switching port-mode access",
"set interfaces ae11.0 family ethernet-switching vlan members ANSIBLE-MANAGEMENT"
]
}
TASK [Apply JunOS configuration] ***********************************************
changed: [lsn-mc1002 -> localhost]
TASK [debug] *******************************************************************
ok: [lsn-mc1002] => {
"msg": [
"set interfaces vlan.2003 family inet address 10.3.8.1/24",
"set vlans ANSIBLE-MANAGEMENT vlan-id 2003",
"set vlans ANSIBLE-MANAGEMENT l3-interface vlan.2003",
"delete interfaces ge-0/0/6",
"set interfaces ge-0/0/6 description \"lsn-mc1002\"",
"set interfaces ge-0/0/6 ether-options 802.3ad ae11"
]
}
TASK [Apply JunOS configuration] ***********************************************
changed: [lsn-mc1002 -> localhost]
TASK [debug] *******************************************************************
ok: [lsn-mc1002] => {
"msg": [
"set interfaces vlan.2003 family inet address 10.3.8.1/24",
"set vlans ANSIBLE-MANAGEMENT vlan-id 2003",
"set vlans ANSIBLE-MANAGEMENT l3-interface vlan.2003",
"delete interfaces ge-1/0/6",
"set interfaces ge-1/0/6 description \"lsn-mc1002\"",
"set interfaces ge-1/0/6 ether-options 802.3ad ae11"
]
}
TASK [Apply JunOS configuration] ***********************************************
changed: [lsn-mc1002 -> localhost]
```
Resulting configuration from first run looks good:
```
logan@stack01.rack92> show configuration interfaces ae11
description lsn-mc1002;
aggregated-ether-options {
lacp {
active;
}
}
unit 0 {
family ethernet-switching {
port-mode access;
vlan {
members ANSIBLE-MANAGEMENT;
}
}
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-0/0/6
description lsn-mc1002;
ether-options {
802.3ad ae11;
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-1/0/6
description lsn-mc1002;
ether-options {
802.3ad ae11;
}
{master:0}
```
Second run. Ansible output is identical so I'll refrain from pasting it again.
```
logan@stack01.rack92> show configuration interfaces ae11
description lsn-mc1002;
unit 0 {
family ethernet-switching {
port-mode access;
vlan {
members ANSIBLE-MANAGEMENT;
}
}
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-0/0/6
description lsn-mc1002;
{master:0}
logan@stack01.rack92> show configuration interfaces ge-1/0/6
description lsn-mc1002;
{master:0}
```
3rd run:
```
logan@stack01.rack92> show configuration interfaces ae11
description lsn-mc1002;
aggregated-ether-options {
lacp {
active;
}
}
unit 0 {
family ethernet-switching {
port-mode access;
vlan {
members ANSIBLE-MANAGEMENT;
}
}
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-0/0/6
description lsn-mc1002;
ether-options {
802.3ad ae11;
}
{master:0}
logan@stack01.rack92> show configuration interfaces ge-1/0/6
description lsn-mc1002;
ether-options {
802.3ad ae11;
}
{master:0}
```
Subsequent runs oscillate back and forth between the broken and proper configurations. | main | junos config broken configurations and idempotency issue type bug report component name junos config ansible version ansible config file home centos provisioning metal ansible cfg configured module search path configuration roles plugins inventory paths os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific centos managing a juniper stack running summary the junos config deploys broken inconsistent configurations every time i run it steps to reproduce starting with blank port configuration logan show configuration interfaces master logan show configuration interfaces ge master logan show configuration interfaces ge master the playbook used is the following debug msg switch commands default when not switch commands fact skipped name apply junos configuration local action module junos config lines switch commands default provider hostvars when not switch commands fact skipped first run of playbook task ok msg set interfaces vlan family inet address set vlans ansible management vlan id set vlans ansible management interface vlan delete interfaces set interfaces description lsn set interfaces aggregated ether options lacp active set interfaces family ethernet switching port mode access set interfaces family ethernet switching vlan members ansible management task changed task ok msg set interfaces vlan family inet address set vlans ansible management vlan id set vlans ansible management interface vlan delete interfaces ge set interfaces ge description lsn set interfaces ge ether options task changed task ok msg set interfaces vlan family inet address set vlans ansible management vlan id set vlans ansible management interface vlan delete interfaces ge set interfaces ge description lsn set interfaces ge ether options task changed resulting configuration from first run looks good logan show configuration interfaces description lsn aggregated ether options lacp active unit family ethernet switching port mode access vlan members ansible management master logan show configuration interfaces ge description lsn ether options master logan show configuration interfaces ge description lsn ether options master second run ansible output is identical so i ll refrain from pasting it again logan show configuration interfaces description lsn unit family ethernet switching port mode access vlan members ansible management master logan show configuration interfaces ge description lsn master logan show configuration interfaces ge description lsn master run logan show configuration interfaces description lsn aggregated ether options lacp active unit family ethernet switching port mode access vlan members ansible management master logan show configuration interfaces ge description lsn ether options master logan show configuration interfaces ge description lsn ether options master subsequent runs oscillate back and forth between the broken and proper configurations | 1 |
2,982 | 10,735,672,440 | IssuesEvent | 2019-10-29 09:17:21 | diofant/diofant | https://api.github.com/repos/diofant/diofant | opened | Remove _randint/randrange helpers | maintainability utilities | Probably, random.choice(s) may be used, when seed is a sequence. Everything is in CPython 3.6. | True | Remove _randint/randrange helpers - Probably, random.choice(s) may be used, when seed is a sequence. Everything is in CPython 3.6. | main | remove randint randrange helpers probably random choice s may be used when seed is a sequence everything is in cpython | 1 |
3,668 | 14,996,232,169 | IssuesEvent | 2021-01-29 15:19:51 | ansible/ansible | https://api.github.com/repos/ansible/ansible | closed | Syntax improvements for complex "when" conditionals | affects_2.8 feature needs_maintainer support:core | ##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
Please could the `when:` directive be extended with additional syntactic sugar for both AND and OR?
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the new feature, module, plugin or task below -->
Not sure what is meant here. However, I would like the sugar to ideally use directives named `any` and `all`. This would provide a degree of similarity to Python's [`all()`](https://docs.python.org/3/library/functions.html#all) and [`any()`](https://docs.python.org/3/library/functions.html#any) functions.
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
This would allow over-complex and bracket-y conditions to be rewritten in a more understandable way. The syntax would also be better when looking at diffs. A common use case is when targetting `tasks` or `import_*`s at multiple (distro, version) combinations. Currently one might write
```yaml
- import_tasks: stuff.yml
when:
- ((ansible_distribution == "Debian" and
ansible_distribution_major_version in ["8", "9"]) or
(ansible_distribution == "CentOS" and
ansible_distribution_major_version in ["6", "7"]))
```
In future, perhaps one could write the following instead?
```yaml
# HYPOTHETICAL ONLY
- import_tasks: stuff.yml
when:
- any:
- all:
- ansible_distribution == "Debian"
- ansible_distribution_major_version in ["8", "9"]
- all:
- ansible_distribution == "CentOS"
- ansible_distribution_major_version in ["6", "7"]
``` | True | Syntax improvements for complex "when" conditionals - ##### SUMMARY
<!--- Describe the new feature/improvement briefly below -->
Please could the `when:` directive be extended with additional syntactic sugar for both AND and OR?
##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
<!--- Write the short name of the new feature, module, plugin or task below -->
Not sure what is meant here. However, I would like the sugar to ideally use directives named `any` and `all`. This would provide a degree of similarity to Python's [`all()`](https://docs.python.org/3/library/functions.html#all) and [`any()`](https://docs.python.org/3/library/functions.html#any) functions.
##### ADDITIONAL INFORMATION
<!--- Describe how the feature would be used, why it is needed and what it would solve -->
This would allow over-complex and bracket-y conditions to be rewritten in a more understandable way. The syntax would also be better when looking at diffs. A common use case is when targetting `tasks` or `import_*`s at multiple (distro, version) combinations. Currently one might write
```yaml
- import_tasks: stuff.yml
when:
- ((ansible_distribution == "Debian" and
ansible_distribution_major_version in ["8", "9"]) or
(ansible_distribution == "CentOS" and
ansible_distribution_major_version in ["6", "7"]))
```
In future, perhaps one could write the following instead?
```yaml
# HYPOTHETICAL ONLY
- import_tasks: stuff.yml
when:
- any:
- all:
- ansible_distribution == "Debian"
- ansible_distribution_major_version in ["8", "9"]
- all:
- ansible_distribution == "CentOS"
- ansible_distribution_major_version in ["6", "7"]
``` | main | syntax improvements for complex when conditionals summary please could the when directive be extended with additional syntactic sugar for both and and or issue type feature idea component name not sure what is meant here however i would like the sugar to ideally use directives named any and all this would provide a degree of similarity to python s and functions additional information this would allow over complex and bracket y conditions to be rewritten in a more understandable way the syntax would also be better when looking at diffs a common use case is when targetting tasks or import s at multiple distro version combinations currently one might write yaml import tasks stuff yml when ansible distribution debian and ansible distribution major version in or ansible distribution centos and ansible distribution major version in in future perhaps one could write the following instead yaml hypothetical only import tasks stuff yml when any all ansible distribution debian ansible distribution major version in all ansible distribution centos ansible distribution major version in | 1 |
1,442 | 6,265,079,807 | IssuesEvent | 2017-07-16 14:27:04 | ocaml/opam-repository | https://api.github.com/repos/ocaml/opam-repository | closed | ocamlnet 4.x requires nettle-dev when gnutls is used on Debian | depext needs maintainer action | When `gnutls` is enabled and installed but `nettle-dev` is not installed I get this error (note that GnuTLS _is_ installed, and if I install `nettle-dev` then the build succeeds):
```
# command ./configure -bindir /home/edwin/.opam/system/bin -disable-gssapi -enable-gnutls -enable-pcre -enable-gtk2 -disable-zip -with-nethttpd
[...]
# Checking whether lablgtk2 has GMain.Io.add_watch with list support... yes
# Checking for GnuTLS... not found
# Required library GnuTLS not found!
### stderr ###
# Welcome to Ocamlnet version 4.0.2
# Package nettle was not found in the pkg-config search path.
# Perhaps you should add the directory containing `nettle.pc'
# to the PKG_CONFIG_PATH environment variable
# No package 'nettle' found
# Package nettle was not found in the pkg-config search path.
# Perhaps you should add the directory containing `nettle.pc'
# to the PKG_CONFIG_PATH environment variable
# No package 'nettle' found
```
| True | ocamlnet 4.x requires nettle-dev when gnutls is used on Debian - When `gnutls` is enabled and installed but `nettle-dev` is not installed I get this error (note that GnuTLS _is_ installed, and if I install `nettle-dev` then the build succeeds):
```
# command ./configure -bindir /home/edwin/.opam/system/bin -disable-gssapi -enable-gnutls -enable-pcre -enable-gtk2 -disable-zip -with-nethttpd
[...]
# Checking whether lablgtk2 has GMain.Io.add_watch with list support... yes
# Checking for GnuTLS... not found
# Required library GnuTLS not found!
### stderr ###
# Welcome to Ocamlnet version 4.0.2
# Package nettle was not found in the pkg-config search path.
# Perhaps you should add the directory containing `nettle.pc'
# to the PKG_CONFIG_PATH environment variable
# No package 'nettle' found
# Package nettle was not found in the pkg-config search path.
# Perhaps you should add the directory containing `nettle.pc'
# to the PKG_CONFIG_PATH environment variable
# No package 'nettle' found
```
| main | ocamlnet x requires nettle dev when gnutls is used on debian when gnutls is enabled and installed but nettle dev is not installed i get this error note that gnutls is installed and if i install nettle dev then the build succeeds command configure bindir home edwin opam system bin disable gssapi enable gnutls enable pcre enable disable zip with nethttpd checking whether has gmain io add watch with list support yes checking for gnutls not found required library gnutls not found stderr welcome to ocamlnet version package nettle was not found in the pkg config search path perhaps you should add the directory containing nettle pc to the pkg config path environment variable no package nettle found package nettle was not found in the pkg config search path perhaps you should add the directory containing nettle pc to the pkg config path environment variable no package nettle found | 1 |
2,980 | 10,731,445,180 | IssuesEvent | 2019-10-28 19:34:13 | andrewdetorres/Real-Estate-System | https://api.github.com/repos/andrewdetorres/Real-Estate-System | opened | Return type declarations | Type: Maintainence | Ensure return type decelerations are being used across code.
`/**
* Example
*
* @param integer $id
* @return array
*/
public static function example($abc) : array {
return [];
}` | True | Return type declarations - Ensure return type decelerations are being used across code.
`/**
* Example
*
* @param integer $id
* @return array
*/
public static function example($abc) : array {
return [];
}` | main | return type declarations ensure return type decelerations are being used across code example param integer id return array public static function example abc array return | 1 |
243,020 | 7,852,387,850 | IssuesEvent | 2018-06-20 14:29:29 | SoMa-Project/ec_grasp_planner | https://api.github.com/repos/SoMa-Project/ec_grasp_planner | opened | Limit maximum velocity during PostGraspRotate | enhancement priority | The trajectory generated by the HA for PostGraspRotate (for a grasped real cucumber) includes a part in which the WAM is moving at least one of its joints quite fast. This results in the max velocity safety being triggered and in case its not, it simply looks scary.
Adding a parameter to handarm_parameters.py to control this behavior would be great.
Something worth trying out is, if setting a lower vmax in the HA is already solving this issue (would still need to adjustable from the parameter file) | 1.0 | Limit maximum velocity during PostGraspRotate - The trajectory generated by the HA for PostGraspRotate (for a grasped real cucumber) includes a part in which the WAM is moving at least one of its joints quite fast. This results in the max velocity safety being triggered and in case its not, it simply looks scary.
Adding a parameter to handarm_parameters.py to control this behavior would be great.
Something worth trying out is, if setting a lower vmax in the HA is already solving this issue (would still need to adjustable from the parameter file) | non_main | limit maximum velocity during postgrasprotate the trajectory generated by the ha for postgrasprotate for a grasped real cucumber includes a part in which the wam is moving at least one of its joints quite fast this results in the max velocity safety being triggered and in case its not it simply looks scary adding a parameter to handarm parameters py to control this behavior would be great something worth trying out is if setting a lower vmax in the ha is already solving this issue would still need to adjustable from the parameter file | 0 |
93,642 | 8,440,323,481 | IssuesEvent | 2018-10-18 06:49:44 | xcat2/xcat2-task-management | https://api.github.com/repos/xcat2/xcat2-task-management | closed | Discovery manul regression for 2.14.4 release | sprint1 test | what to do:
* [ ] switch based switch discovery, please refer to [Switch Based Switch discovery Manual Test Steps](https://github.ibm.com/xcat2/team_process/blob/master/FVT/docs/Switch_Based_Switch_discovery_Manual_Test_Steps.md)
* [ ] switch based node discovery, please refer to [Switch Based Node discovery Manual Test Steps](https://github.ibm.com/xcat2/team_process/blob/master/FVT/docs/Switch_Based_Node_discovery_Manual_Test_Steps.md)
* [ ] MTMs based node discovery please refer to [MTMs Based Node discovery Manual Test Steps](https://github.ibm.com/xcat2/team_process/blob/master/FVT/docs/MTMS_Based_Node_discovery_Manual_Test_Steps.md)
Environment: p9
The expected run time is 2018/10/16 (CDL time) | 1.0 | Discovery manul regression for 2.14.4 release - what to do:
* [ ] switch based switch discovery, please refer to [Switch Based Switch discovery Manual Test Steps](https://github.ibm.com/xcat2/team_process/blob/master/FVT/docs/Switch_Based_Switch_discovery_Manual_Test_Steps.md)
* [ ] switch based node discovery, please refer to [Switch Based Node discovery Manual Test Steps](https://github.ibm.com/xcat2/team_process/blob/master/FVT/docs/Switch_Based_Node_discovery_Manual_Test_Steps.md)
* [ ] MTMs based node discovery please refer to [MTMs Based Node discovery Manual Test Steps](https://github.ibm.com/xcat2/team_process/blob/master/FVT/docs/MTMS_Based_Node_discovery_Manual_Test_Steps.md)
Environment: p9
The expected run time is 2018/10/16 (CDL time) | non_main | discovery manul regression for release what to do switch based switch discovery please refer to switch based node discovery please refer to mtms based node discovery please refer to environment the expected run time is cdl time | 0 |
506,601 | 14,668,453,010 | IssuesEvent | 2020-12-29 21:18:07 | microsoft/PowerToys | https://api.github.com/repos/microsoft/PowerToys | closed | [KBM] Remap Shortcuts/Remap Keyboard window does not always open on the same monitor as Settings | Priority-1 Product-Keyboard Shortcut Manager Resolution-Fix-Committed | The windows have been set to start at the center of the screen, however the exact monitor where it will load need not be same as Settings right now. | 1.0 | [KBM] Remap Shortcuts/Remap Keyboard window does not always open on the same monitor as Settings - The windows have been set to start at the center of the screen, however the exact monitor where it will load need not be same as Settings right now. | non_main | remap shortcuts remap keyboard window does not always open on the same monitor as settings the windows have been set to start at the center of the screen however the exact monitor where it will load need not be same as settings right now | 0 |
553,805 | 16,382,964,925 | IssuesEvent | 2021-05-17 06:50:04 | momentum-mod/game | https://api.github.com/repos/momentum-mod/game | closed | VGUI uses wrong fonts on Linux | OS: Linux Outcome: Resolved Priority: High Size: Medium Type: Bug Where: Engine | **Describe the bug**
Instead of using Bebas Neue / Source Code Pro, the game defaults to Verdana (?), resulting in cutoff UI components
**To Reproduce**
- Launch the game on Linux
**Expected behavior**
The game should use the appropriate font
**Screenshots**

**Desktop/Branch:**
- OS: Linux
- Branch: Latest 0.8.8-rc (fresh install)
**Additional context**
You can also see it here https://discord.com/channels/235111289435717633/733230720079757332/819714644859944970 at around 4 seconds.
According to strace, the game finds the appropriate files and loads them without issue. `vgui_spew_fonts` also lists them (but I have no idea if that has any further meaning). The Momentum symbol font even displays correctly on the weapon switch menu and the Valve logo shows up in the credits.
I also tried: renaming the files to be lowercase and replacing the fonts with the .ttf variants from 0.8.7 to no avail.
| 1.0 | VGUI uses wrong fonts on Linux - **Describe the bug**
Instead of using Bebas Neue / Source Code Pro, the game defaults to Verdana (?), resulting in cutoff UI components
**To Reproduce**
- Launch the game on Linux
**Expected behavior**
The game should use the appropriate font
**Screenshots**

**Desktop/Branch:**
- OS: Linux
- Branch: Latest 0.8.8-rc (fresh install)
**Additional context**
You can also see it here https://discord.com/channels/235111289435717633/733230720079757332/819714644859944970 at around 4 seconds.
According to strace, the game finds the appropriate files and loads them without issue. `vgui_spew_fonts` also lists them (but I have no idea if that has any further meaning). The Momentum symbol font even displays correctly on the weapon switch menu and the Valve logo shows up in the credits.
I also tried: renaming the files to be lowercase and replacing the fonts with the .ttf variants from 0.8.7 to no avail.
| non_main | vgui uses wrong fonts on linux describe the bug instead of using bebas neue source code pro the game defaults to verdana resulting in cutoff ui components to reproduce launch the game on linux expected behavior the game should use the appropriate font screenshots desktop branch os linux branch latest rc fresh install additional context you can also see it here at around seconds according to strace the game finds the appropriate files and loads them without issue vgui spew fonts also lists them but i have no idea if that has any further meaning the momentum symbol font even displays correctly on the weapon switch menu and the valve logo shows up in the credits i also tried renaming the files to be lowercase and replacing the fonts with the ttf variants from to no avail | 0 |
283,383 | 8,719,402,752 | IssuesEvent | 2018-12-08 00:37:04 | aowen87/BAR | https://api.github.com/repos/aowen87/BAR | closed | Save session dialog woes on Windows | bug likelihood high priority reviewed severity medium | Tried to save a session on windows, received the error message and dialog as shown in the attached screen shots.
Upshot is that the drive letter for the 'VISITUSERHOME' path (should be something like C:\Users\<name>\Documents\VisIt), was stripped off the path and used as the host instead. Then path reverted to %HOME% (C:\Users\<Name>)
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2380
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Save session dialog woes on Windows
Assigned to: Brad Whitlock
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 09/18/2015
Due date:
% Done: 0
Estimated time:
Created: 09/18/2015 07:08 pm
Updated: 09/23/2015 08:15 pm
Likelihood: 4 - Common
Severity: 3 - Major Irritation
Found in version: trunk
Impact:
Expected Use:
OS: Windows
Support Group: Any
Description:
Tried to save a session on windows, received the error message and dialog as shown in the attached screen shots.
Upshot is that the drive letter for the 'VISITUSERHOME' path (should be something like C:\Users\<name>\Documents\VisIt), was stripped off the path and used as the host instead. Then path reverted to %HOME% (C:\Users\<Name>)
Comments:
| 1.0 | Save session dialog woes on Windows - Tried to save a session on windows, received the error message and dialog as shown in the attached screen shots.
Upshot is that the drive letter for the 'VISITUSERHOME' path (should be something like C:\Users\<name>\Documents\VisIt), was stripped off the path and used as the host instead. Then path reverted to %HOME% (C:\Users\<Name>)
-----------------------REDMINE MIGRATION-----------------------
This ticket was migrated from Redmine. As such, not all
information was able to be captured in the transition. Below is
a complete record of the original redmine ticket.
Ticket number: 2380
Status: Resolved
Project: VisIt
Tracker: Bug
Priority: High
Subject: Save session dialog woes on Windows
Assigned to: Brad Whitlock
Category:
Target version: 2.10
Author: Kathleen Biagas
Start: 09/18/2015
Due date:
% Done: 0
Estimated time:
Created: 09/18/2015 07:08 pm
Updated: 09/23/2015 08:15 pm
Likelihood: 4 - Common
Severity: 3 - Major Irritation
Found in version: trunk
Impact:
Expected Use:
OS: Windows
Support Group: Any
Description:
Tried to save a session on windows, received the error message and dialog as shown in the attached screen shots.
Upshot is that the drive letter for the 'VISITUSERHOME' path (should be something like C:\Users\<name>\Documents\VisIt), was stripped off the path and used as the host instead. Then path reverted to %HOME% (C:\Users\<Name>)
Comments:
| non_main | save session dialog woes on windows tried to save a session on windows received the error message and dialog as shown in the attached screen shots upshot is that the drive letter for the visituserhome path should be something like c users documents visit was stripped off the path and used as the host instead then path reverted to home c users redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority high subject save session dialog woes on windows assigned to brad whitlock category target version author kathleen biagas start due date done estimated time created pm updated pm likelihood common severity major irritation found in version trunk impact expected use os windows support group any description tried to save a session on windows received the error message and dialog as shown in the attached screen shots upshot is that the drive letter for the visituserhome path should be something like c users documents visit was stripped off the path and used as the host instead then path reverted to home c users comments | 0 |
5,068 | 25,959,107,651 | IssuesEvent | 2022-12-18 16:49:47 | BioArchLinux/Packages | https://api.github.com/repos/BioArchLinux/Packages | reopened | [MAINTAIN] java related packages polishment | maintain upstream | <!--
Please report the error of one package in one issue! Use multi issues to report multi bugs.
Thanks!
-->
**List of Pacakges**
<details>
- [x] aliview
- [x] artemis
- [x] bbmap
- [x] beast
- [x] beast2
- [x] canu
- [x] #101
- [x] emboss
- [x] fastqc
- [x] figtree
- [x] gatk
- [x] insight-toolkit
- [x] jane
- [x] libsbml
- [x] mcscanx
- [x] mesquite
- [x] morphoj
- [x] open-delta
- [x] phyde
- [x] phylip
- [x] pibuss
- [x] phylonet
- [x] picard
- [x] sequencematrix
- [x] spread3
- [x] spread
- [x] tempest
- [x] tracer
- [x] treemap
- [x] trimmomatic
</details>
**Description**
These packages need to be checked the following points
- [x] use `java-runtime` and `java-enviroment` instead of jdk, jre, etc.
- [x] confirm the java version and limit the enviroment in PKGBUILD
- [x] check if they can be compiled from the source code [may be difficult, considering some software use quite low version java-enviroment]
- [x] check if PKGBUILD style can be auto updated
| True | [MAINTAIN] java related packages polishment - <!--
Please report the error of one package in one issue! Use multi issues to report multi bugs.
Thanks!
-->
**List of Pacakges**
<details>
- [x] aliview
- [x] artemis
- [x] bbmap
- [x] beast
- [x] beast2
- [x] canu
- [x] #101
- [x] emboss
- [x] fastqc
- [x] figtree
- [x] gatk
- [x] insight-toolkit
- [x] jane
- [x] libsbml
- [x] mcscanx
- [x] mesquite
- [x] morphoj
- [x] open-delta
- [x] phyde
- [x] phylip
- [x] pibuss
- [x] phylonet
- [x] picard
- [x] sequencematrix
- [x] spread3
- [x] spread
- [x] tempest
- [x] tracer
- [x] treemap
- [x] trimmomatic
</details>
**Description**
These packages need to be checked the following points
- [x] use `java-runtime` and `java-enviroment` instead of jdk, jre, etc.
- [x] confirm the java version and limit the enviroment in PKGBUILD
- [x] check if they can be compiled from the source code [may be difficult, considering some software use quite low version java-enviroment]
- [x] check if PKGBUILD style can be auto updated
| main | java related packages polishment please report the error of one package in one issue use multi issues to report multi bugs thanks list of pacakges aliview artemis bbmap beast canu emboss fastqc figtree gatk insight toolkit jane libsbml mcscanx mesquite morphoj open delta phyde phylip pibuss phylonet picard sequencematrix spread tempest tracer treemap trimmomatic description these packages need to be checked the following points use java runtime and java enviroment instead of jdk jre etc confirm the java version and limit the enviroment in pkgbuild check if they can be compiled from the source code check if pkgbuild style can be auto updated | 1 |
114,482 | 24,609,583,458 | IssuesEvent | 2022-10-14 19:50:20 | celestiaorg/rollmint | https://api.github.com/repos/celestiaorg/rollmint | closed | Linter: local imports | T:code-hygiene C:ci | Another linting feature is missing: check for local imports.
_Originally posted by @tzdybal in https://github.com/celestiaorg/rollmint/pull/553#pullrequestreview-1138777729_
| 1.0 | Linter: local imports - Another linting feature is missing: check for local imports.
_Originally posted by @tzdybal in https://github.com/celestiaorg/rollmint/pull/553#pullrequestreview-1138777729_
| non_main | linter local imports another linting feature is missing check for local imports originally posted by tzdybal in | 0 |
4,866 | 25,020,177,401 | IssuesEvent | 2022-11-03 23:16:24 | aws/serverless-application-model | https://api.github.com/repos/aws/serverless-application-model | closed | AWS::Serverless::Api - Auth property doesn't allow intrinsic functionality | area/intrinsics stage/pm-review maintainer/need-response intrinsics/not-resolvable-predeployment | The intrinsic mapping for Auth on APIs does not appear to have been added in v1.16 or v1.20.0 from what I can see. I can see intrinsic mapping for function deployment preferences was in (https://github.com/awslabs/serverless-application-model/issues/1196)
Could this please be added to the roadmap? | True | AWS::Serverless::Api - Auth property doesn't allow intrinsic functionality - The intrinsic mapping for Auth on APIs does not appear to have been added in v1.16 or v1.20.0 from what I can see. I can see intrinsic mapping for function deployment preferences was in (https://github.com/awslabs/serverless-application-model/issues/1196)
Could this please be added to the roadmap? | main | aws serverless api auth property doesn t allow intrinsic functionality the intrinsic mapping for auth on apis does not appear to have been added in or from what i can see i can see intrinsic mapping for function deployment preferences was in could this please be added to the roadmap | 1 |
209,227 | 7,166,540,756 | IssuesEvent | 2018-01-29 17:32:44 | syl20bnr/spacemacs | https://api.github.com/repos/syl20bnr/spacemacs | closed | org-mode doesn't load properly | - Bug tracker - High Priority Loading Org | #### Description :octocat:
org mode does not load properly. Many functions like `org-export-latex-to-pdf` are dead.
#### Reproduction guide :beetle:
- cd ~
- mv .emacs.d emacs_d_backup # start fresh!
- mv .spacemacs spacemacs_backup # start fresh!
- git clone https://github.com/syl20bnr/spacemacs ~/.emacs.d
- Start Emacs
- wait for packages to load
- edit .spacemacs (which emacs created by default)
- only change: in dotspacemacs-configuration-layers, uncomment `org' only
- restart Emacs
- after a bit, observe "Error (use-package): org-projectile/:config: Symbol's function definition is void: org-projectile:per-repo"
*Observed behaviour:* :eyes: :broken_heart:
Org mode sort of works, but looks ugly. I'm dead in the water because I rely on `org-latex-export-to-pdf`, and `org-babel-execute-buffer', both of which say "Symbol's function definition is void: org-link-types''
*Expected behaviour:* :heart: :smile:
`org-latex-export-to-pdf` should produce a PDF file
`org-babel-execute-buffer` should run all the code in my org buffer
#### System Info :computer:
- OS: darwin
- Emacs: 25.3.1
- Spacemacs: 0.200.13
- Spacemacs branch: master (rev. 582f9aa)
- Graphic display: t
- Distribution: spacemacs
- Editing style: emacs
- Completion: helm
- Layers:
```elisp
(helm emacs-lisp org)
```
- System configuration features: NOTIFY ACL GNUTLS LIBXML2 ZLIB TOOLKIT_SCROLL_BARS NS MODULES
#### Backtrace :paw_prints:
```
``` | 1.0 | org-mode doesn't load properly - #### Description :octocat:
org mode does not load properly. Many functions like `org-export-latex-to-pdf` are dead.
#### Reproduction guide :beetle:
- cd ~
- mv .emacs.d emacs_d_backup # start fresh!
- mv .spacemacs spacemacs_backup # start fresh!
- git clone https://github.com/syl20bnr/spacemacs ~/.emacs.d
- Start Emacs
- wait for packages to load
- edit .spacemacs (which emacs created by default)
- only change: in dotspacemacs-configuration-layers, uncomment `org' only
- restart Emacs
- after a bit, observe "Error (use-package): org-projectile/:config: Symbol's function definition is void: org-projectile:per-repo"
*Observed behaviour:* :eyes: :broken_heart:
Org mode sort of works, but looks ugly. I'm dead in the water because I rely on `org-latex-export-to-pdf`, and `org-babel-execute-buffer', both of which say "Symbol's function definition is void: org-link-types''
*Expected behaviour:* :heart: :smile:
`org-latex-export-to-pdf` should produce a PDF file
`org-babel-execute-buffer` should run all the code in my org buffer
#### System Info :computer:
- OS: darwin
- Emacs: 25.3.1
- Spacemacs: 0.200.13
- Spacemacs branch: master (rev. 582f9aa)
- Graphic display: t
- Distribution: spacemacs
- Editing style: emacs
- Completion: helm
- Layers:
```elisp
(helm emacs-lisp org)
```
- System configuration features: NOTIFY ACL GNUTLS LIBXML2 ZLIB TOOLKIT_SCROLL_BARS NS MODULES
#### Backtrace :paw_prints:
```
``` | non_main | org mode doesn t load properly description octocat org mode does not load properly many functions like org export latex to pdf are dead reproduction guide beetle cd mv emacs d emacs d backup start fresh mv spacemacs spacemacs backup start fresh git clone emacs d start emacs wait for packages to load edit spacemacs which emacs created by default only change in dotspacemacs configuration layers uncomment org only restart emacs after a bit observe error use package org projectile config symbol s function definition is void org projectile per repo observed behaviour eyes broken heart org mode sort of works but looks ugly i m dead in the water because i rely on org latex export to pdf and org babel execute buffer both of which say symbol s function definition is void org link types expected behaviour heart smile org latex export to pdf should produce a pdf file org babel execute buffer should run all the code in my org buffer system info computer os darwin emacs spacemacs spacemacs branch master rev graphic display t distribution spacemacs editing style emacs completion helm layers elisp helm emacs lisp org system configuration features notify acl gnutls zlib toolkit scroll bars ns modules backtrace paw prints | 0 |
1,785 | 6,575,859,837 | IssuesEvent | 2017-09-11 17:36:34 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | mysql_user is not idempotent with complex priv | affects_2.1 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
mysql_user
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
None.
##### OS / ENVIRONMENT
- Ansible host: Xubuntu 16.04
- Ansible target: Debian Jessie + MariaDB 10.1 (with galera module activated)
##### SUMMARY
With "complex" privs, _mysql_user_ module is not idempotent. The module should not have a state changed.
##### STEPS TO REPRODUCE
```
- hosts: tag_Db_1
vars:
c_mysql_users:
- name: "monitor"
password: "1a2b3c"
priv: '*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.user:SELECT/mysql.db:SELECT/mysql.tables_priv:SELECT'
host: "%"
- name: "monitor2"
password: "1a2b3c"
priv: '*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.*:SELECT'
host: "%"
tasks:
- block:
- name: MYSQL_USERS | Create users
mysql_user: >
name="{{ item.name }}"
password="{{ item.password }}"
host="{{ item.host }}"
priv='{{ item.priv }}'
with_items: "{{ c_mysql_users }}"
run_once: true
```
##### EXPECTED RESULTS
On second launch, mysql_user module should not change.
##### ACTUAL RESULTS
User "monitor2" is not changed but user "monitor" have changed state every time
```
PLAYBOOK: bug.yml **************************************************************
1 plays in bug.yml
PLAY [tag_Db_1] ****************************************************************
TASK [setup] *******************************************************************
ok: [db2]
ok: [db1]
TASK [MYSQL_USERS | Create users] **********************************************
task path: /home/hanx/dev/na/ansible/affil/bug.yml:15
changed: [db1] => (item={u'host': u'%', u'password': u'1a2b3c', u'name': u'monitor', u'priv': u'*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.user:SELECT/mysql.db:SELECT/mysql.tables_priv:SELECT'}) => {"changed": true, "item": {"host": "%", "name": "monitor", "password": "1a2b3c", "priv": "*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.user:SELECT/mysql.db:SELECT/mysql.tables_priv:SELECT"}, "user": "monitor"}
ok: [db1] => (item={u'host': u'%', u'password': u'1a2b3c', u'name': u'monitor2', u'priv': u'*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.*:SELECT'}) => {"changed": false, "item": {"host": "%", "name": "monitor2", "password": "1a2b3c", "priv": "*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.*:SELECT"}, "user": "monitor2"}
PLAY RECAP *********************************************************************
db1 : ok=2 changed=1 unreachable=0 failed=0
db2 : ok=1 changed=0 unreachable=0 failed=0
```
There is grant output on my server:
```
MariaDB [(none)]> show grants for 'monitor';
+-------------------------------------------------------------------------------------------------------------------------------------+
| Grants for monitor@% |
+-------------------------------------------------------------------------------------------------------------------------------------+
| GRANT SHOW DATABASES, REPLICATION CLIENT ON *.* TO 'monitor'@'%' IDENTIFIED BY PASSWORD '*09D6E9E852C3A9F6BDC1C52834B361317225F914' |
| GRANT SELECT ON `mysql`.`db` TO 'monitor'@'%' |
| GRANT SELECT ON `mysql`.`user` TO 'monitor'@'%' |
| GRANT SELECT ON `mysql`.`tables_priv` TO 'monitor'@'%' |
+-------------------------------------------------------------------------------------------------------------------------------------+
4 rows in set (0.00 sec)
MariaDB [(none)]> show grants for 'monitor2';
+--------------------------------------------------------------------------------------------------------------------------------------+
| Grants for monitor2@% |
+--------------------------------------------------------------------------------------------------------------------------------------+
| GRANT SHOW DATABASES, REPLICATION CLIENT ON *.* TO 'monitor2'@'%' IDENTIFIED BY PASSWORD '*09D6E9E852C3A9F6BDC1C52834B361317225F914' |
| GRANT SELECT ON `mysql`.* TO 'monitor2'@'%' |
+--------------------------------------------------------------------------------------------------------------------------------------+
2 rows in set (0.00 sec)
```
| True | mysql_user is not idempotent with complex priv - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
mysql_user
##### ANSIBLE VERSION
```
ansible 2.1.1.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
None.
##### OS / ENVIRONMENT
- Ansible host: Xubuntu 16.04
- Ansible target: Debian Jessie + MariaDB 10.1 (with galera module activated)
##### SUMMARY
With "complex" privs, _mysql_user_ module is not idempotent. The module should not have a state changed.
##### STEPS TO REPRODUCE
```
- hosts: tag_Db_1
vars:
c_mysql_users:
- name: "monitor"
password: "1a2b3c"
priv: '*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.user:SELECT/mysql.db:SELECT/mysql.tables_priv:SELECT'
host: "%"
- name: "monitor2"
password: "1a2b3c"
priv: '*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.*:SELECT'
host: "%"
tasks:
- block:
- name: MYSQL_USERS | Create users
mysql_user: >
name="{{ item.name }}"
password="{{ item.password }}"
host="{{ item.host }}"
priv='{{ item.priv }}'
with_items: "{{ c_mysql_users }}"
run_once: true
```
##### EXPECTED RESULTS
On second launch, mysql_user module should not change.
##### ACTUAL RESULTS
User "monitor2" is not changed but user "monitor" have changed state every time
```
PLAYBOOK: bug.yml **************************************************************
1 plays in bug.yml
PLAY [tag_Db_1] ****************************************************************
TASK [setup] *******************************************************************
ok: [db2]
ok: [db1]
TASK [MYSQL_USERS | Create users] **********************************************
task path: /home/hanx/dev/na/ansible/affil/bug.yml:15
changed: [db1] => (item={u'host': u'%', u'password': u'1a2b3c', u'name': u'monitor', u'priv': u'*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.user:SELECT/mysql.db:SELECT/mysql.tables_priv:SELECT'}) => {"changed": true, "item": {"host": "%", "name": "monitor", "password": "1a2b3c", "priv": "*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.user:SELECT/mysql.db:SELECT/mysql.tables_priv:SELECT"}, "user": "monitor"}
ok: [db1] => (item={u'host': u'%', u'password': u'1a2b3c', u'name': u'monitor2', u'priv': u'*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.*:SELECT'}) => {"changed": false, "item": {"host": "%", "name": "monitor2", "password": "1a2b3c", "priv": "*.*:SHOW DATABASES,REPLICATION CLIENT/mysql.*:SELECT"}, "user": "monitor2"}
PLAY RECAP *********************************************************************
db1 : ok=2 changed=1 unreachable=0 failed=0
db2 : ok=1 changed=0 unreachable=0 failed=0
```
There is grant output on my server:
```
MariaDB [(none)]> show grants for 'monitor';
+-------------------------------------------------------------------------------------------------------------------------------------+
| Grants for monitor@% |
+-------------------------------------------------------------------------------------------------------------------------------------+
| GRANT SHOW DATABASES, REPLICATION CLIENT ON *.* TO 'monitor'@'%' IDENTIFIED BY PASSWORD '*09D6E9E852C3A9F6BDC1C52834B361317225F914' |
| GRANT SELECT ON `mysql`.`db` TO 'monitor'@'%' |
| GRANT SELECT ON `mysql`.`user` TO 'monitor'@'%' |
| GRANT SELECT ON `mysql`.`tables_priv` TO 'monitor'@'%' |
+-------------------------------------------------------------------------------------------------------------------------------------+
4 rows in set (0.00 sec)
MariaDB [(none)]> show grants for 'monitor2';
+--------------------------------------------------------------------------------------------------------------------------------------+
| Grants for monitor2@% |
+--------------------------------------------------------------------------------------------------------------------------------------+
| GRANT SHOW DATABASES, REPLICATION CLIENT ON *.* TO 'monitor2'@'%' IDENTIFIED BY PASSWORD '*09D6E9E852C3A9F6BDC1C52834B361317225F914' |
| GRANT SELECT ON `mysql`.* TO 'monitor2'@'%' |
+--------------------------------------------------------------------------------------------------------------------------------------+
2 rows in set (0.00 sec)
```
| main | mysql user is not idempotent with complex priv issue type bug report component name mysql user ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration none os environment ansible host xubuntu ansible target debian jessie mariadb with galera module activated summary with complex privs mysql user module is not idempotent the module should not have a state changed steps to reproduce hosts tag db vars c mysql users name monitor password priv show databases replication client mysql user select mysql db select mysql tables priv select host name password priv show databases replication client mysql select host tasks block name mysql users create users mysql user name item name password item password host item host priv item priv with items c mysql users run once true expected results on second launch mysql user module should not change actual results user is not changed but user monitor have changed state every time playbook bug yml plays in bug yml play task ok ok task task path home hanx dev na ansible affil bug yml changed item u host u u password u u name u monitor u priv u show databases replication client mysql user select mysql db select mysql tables priv select changed true item host name monitor password priv show databases replication client mysql user select mysql db select mysql tables priv select user monitor ok item u host u u password u u name u u priv u show databases replication client mysql select changed false item host name password priv show databases replication client mysql select user play recap ok changed unreachable failed ok changed unreachable failed there is grant output on my server mariadb show grants for monitor grants for monitor grant show databases replication client on to monitor identified by password grant select on mysql db to monitor grant select on mysql user to monitor grant select on mysql tables priv to monitor rows in set sec mariadb show grants for grants for grant show databases replication client on to identified by password grant select on mysql to rows in set sec | 1 |
2,484 | 8,640,567,741 | IssuesEvent | 2018-11-24 07:51:33 | arcticicestudio/nord-docs | https://api.github.com/repos/arcticicestudio/nord-docs | opened | Fix overriden ESLint "import/no-extraneous-dependencies" rule | context-workflow scope-maintainability scope-quality type-bug | <p align="center"><img src="https://user-images.githubusercontent.com/7836623/48658871-4707b600-ea49-11e8-8a08-39e1a318c442.png" /></p>
The [import/no-extraneous-dependencie][] rule allows to define a array of glob pattern that are allowed to define import `devDependencies`. In #15 the rule was overriden to include the project specific path `**/.gatsby/**` which whitelists all Gatsby specific scripts. Unfortunately this removes all glob pattern defined in the used `eslint-config-arcticicestudio` (rule is defined in the `-base` package) resulting in errors in other projects paths like tests.
This can be fixed by importing the paths defined in the preset from the ``eslint-config-arcticicestudio-base` package and merge it with the additional **/.gatsby/**` path.
[import/no-extraneous-dependencie]: https://github.com/benmosher/eslint-plugin-import/blob/master/docs/rules/no-extraneous-dependencies.md | True | Fix overriden ESLint "import/no-extraneous-dependencies" rule - <p align="center"><img src="https://user-images.githubusercontent.com/7836623/48658871-4707b600-ea49-11e8-8a08-39e1a318c442.png" /></p>
The [import/no-extraneous-dependencie][] rule allows to define a array of glob pattern that are allowed to define import `devDependencies`. In #15 the rule was overriden to include the project specific path `**/.gatsby/**` which whitelists all Gatsby specific scripts. Unfortunately this removes all glob pattern defined in the used `eslint-config-arcticicestudio` (rule is defined in the `-base` package) resulting in errors in other projects paths like tests.
This can be fixed by importing the paths defined in the preset from the ``eslint-config-arcticicestudio-base` package and merge it with the additional **/.gatsby/**` path.
[import/no-extraneous-dependencie]: https://github.com/benmosher/eslint-plugin-import/blob/master/docs/rules/no-extraneous-dependencies.md | main | fix overriden eslint import no extraneous dependencies rule the rule allows to define a array of glob pattern that are allowed to define import devdependencies in the rule was overriden to include the project specific path gatsby which whitelists all gatsby specific scripts unfortunately this removes all glob pattern defined in the used eslint config arcticicestudio rule is defined in the base package resulting in errors in other projects paths like tests this can be fixed by importing the paths defined in the preset from the eslint config arcticicestudio base package and merge it with the additional gatsby path | 1 |
3,172 | 12,226,785,051 | IssuesEvent | 2020-05-03 12:32:23 | gfleetwood/asteres | https://api.github.com/repos/gfleetwood/asteres | opened | hedgehogqa/r-hedgehog (104736422) | R maintain | https://github.com/hedgehogqa/r-hedgehog
Release with confidence, state-of-the-art property testing for R. | True | hedgehogqa/r-hedgehog (104736422) - https://github.com/hedgehogqa/r-hedgehog
Release with confidence, state-of-the-art property testing for R. | main | hedgehogqa r hedgehog release with confidence state of the art property testing for r | 1 |
3,323 | 12,882,874,982 | IssuesEvent | 2020-07-12 19:04:33 | precice/precice | https://api.github.com/repos/precice/precice | opened | Test Mappings with empty Meshes | maintainability | Currently, we test the functionality of the mappings and tagging methods.
However, we never test how the mapping behaves if either the input, output or both meshes are empty.
We need to check these edge cases explicitly, which should be 3 tests per mapping. | True | Test Mappings with empty Meshes - Currently, we test the functionality of the mappings and tagging methods.
However, we never test how the mapping behaves if either the input, output or both meshes are empty.
We need to check these edge cases explicitly, which should be 3 tests per mapping. | main | test mappings with empty meshes currently we test the functionality of the mappings and tagging methods however we never test how the mapping behaves if either the input output or both meshes are empty we need to check these edge cases explicitly which should be tests per mapping | 1 |
2,903 | 10,325,346,091 | IssuesEvent | 2019-09-01 16:35:12 | frej/fast-export | https://api.github.com/repos/frej/fast-export | closed | ./hg-fast-export.sh: line 177: python2: command not found | OS X not-available-to-maintainer wintendo | Although it's likely this is a configuration issue, I am stuck
- Windows 10 64 bit
- Python 2.7.15 installed and in the path
- Mercurial 4.7 install, and in the path
- Git 2.18.0.windows.1 installed, and in the path
Running the following command in git-bash:
`./hg-fast-export.sh -r ../pathToHgRepo`
Looking at hg-fast-export.sh, line 177 is a comment, so not sure where to go from here.
Any help appreciated.
| True | ./hg-fast-export.sh: line 177: python2: command not found - Although it's likely this is a configuration issue, I am stuck
- Windows 10 64 bit
- Python 2.7.15 installed and in the path
- Mercurial 4.7 install, and in the path
- Git 2.18.0.windows.1 installed, and in the path
Running the following command in git-bash:
`./hg-fast-export.sh -r ../pathToHgRepo`
Looking at hg-fast-export.sh, line 177 is a comment, so not sure where to go from here.
Any help appreciated.
| main | hg fast export sh line command not found although it s likely this is a configuration issue i am stuck windows bit python installed and in the path mercurial install and in the path git windows installed and in the path running the following command in git bash hg fast export sh r pathtohgrepo looking at hg fast export sh line is a comment so not sure where to go from here any help appreciated | 1 |
1,532 | 6,572,225,385 | IssuesEvent | 2017-09-11 00:17:04 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | Jira transisition not working by using status="" | affects_2.0 bug_report waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
jira.py
##### ANSIBLE VERSION
```
ansible 2.0.1.0
```
##### CONFIGURATION
default
##### OS / ENVIRONMENT
Ubuntu 15.10
##### SUMMARY
I'm running a playbook with a jira role, inside
`- name: Get information for an issue
jira: uri={{jiraserver}} username={{admin_account}} password={{admin_password}}
project=TEST operation=fetch issue="{{jiraissue}}"
register: issue
- name: Close the issue
jira: uri={{jiraserver}} username={{admin_account}} password={{admin_password}}
issue={{issue.meta.key}} operation=transition status="6"
`
I thought this should work. For status i used also "Done" or/and the german values i found at
https://jiraurl/rest/api/2/issue/{{jiraissue}}/transitions
##### STEPS TO REPRODUCE
Use the showed playbook/role above
##### EXPECTED RESULTS
Ticket should be closed
##### ACTUAL RESULTS
```
TASK [jiraCommentClose : Close the issue] **************************************
fatal: [127.0.0.1]: FAILED! => {"changed": false, "failed": true, "msg": "Failed find valid transition for '6'"}
```
If i use curl to use this transisition with:
```
curl -D- -u username -X POST --data '{"transition":{"id":"6"}}' -H "Content-Type: application/json" https://jiraurl/rest/api/latest/issue/TEST-1234/transitions?expand=transitions.fields
```
it is working.
Any ideas?
| True | Jira transisition not working by using status="" - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
jira.py
##### ANSIBLE VERSION
```
ansible 2.0.1.0
```
##### CONFIGURATION
default
##### OS / ENVIRONMENT
Ubuntu 15.10
##### SUMMARY
I'm running a playbook with a jira role, inside
`- name: Get information for an issue
jira: uri={{jiraserver}} username={{admin_account}} password={{admin_password}}
project=TEST operation=fetch issue="{{jiraissue}}"
register: issue
- name: Close the issue
jira: uri={{jiraserver}} username={{admin_account}} password={{admin_password}}
issue={{issue.meta.key}} operation=transition status="6"
`
I thought this should work. For status i used also "Done" or/and the german values i found at
https://jiraurl/rest/api/2/issue/{{jiraissue}}/transitions
##### STEPS TO REPRODUCE
Use the showed playbook/role above
##### EXPECTED RESULTS
Ticket should be closed
##### ACTUAL RESULTS
```
TASK [jiraCommentClose : Close the issue] **************************************
fatal: [127.0.0.1]: FAILED! => {"changed": false, "failed": true, "msg": "Failed find valid transition for '6'"}
```
If i use curl to use this transisition with:
```
curl -D- -u username -X POST --data '{"transition":{"id":"6"}}' -H "Content-Type: application/json" https://jiraurl/rest/api/latest/issue/TEST-1234/transitions?expand=transitions.fields
```
it is working.
Any ideas?
| main | jira transisition not working by using status issue type bug report component name jira py ansible version ansible configuration default os environment ubuntu summary i m running a playbook with a jira role inside name get information for an issue jira uri jiraserver username admin account password admin password project test operation fetch issue jiraissue register issue name close the issue jira uri jiraserver username admin account password admin password issue issue meta key operation transition status i thought this should work for status i used also done or and the german values i found at steps to reproduce use the showed playbook role above expected results ticket should be closed actual results task fatal failed changed false failed true msg failed find valid transition for if i use curl to use this transisition with curl d u username x post data transition id h content type application json it is working any ideas | 1 |
414,915 | 12,121,461,958 | IssuesEvent | 2020-04-22 09:20:35 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.google.com - site is not usable | browser-firefox engine-gecko priority-critical | <!-- @browser: Firefox 76.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:76.0) Gecko/20100101 Firefox/76.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/51972 -->
**URL**: https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&uact=8&ved=2ahUKEwimpZn2h_roAhUexDgGHeTKBjcQFjAAegQICRAC&url=https%3A%2F%2F3dwarehouse.sketchup.com%2F%3Fhl%3Den&usg=AOvVaw3r_zxTadmQ_YWO3IdunfNF
**Browser / Version**: Firefox 76.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
i am continously browsing 3d warehouse but its its still loading
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/29ade5ed-f3f3-4280-bb66-2d2cbdbcc517.jpeg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200412214314</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/4/6b58ad1d-0805-4df4-bc29-7f23d9c63a99)
Submitted in the name of `@ashok`
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.google.com - site is not usable - <!-- @browser: Firefox 76.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:76.0) Gecko/20100101 Firefox/76.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/51972 -->
**URL**: https://www.google.com/url?sa=t&rct=j&q=&esrc=s&source=web&cd=1&cad=rja&uact=8&ved=2ahUKEwimpZn2h_roAhUexDgGHeTKBjcQFjAAegQICRAC&url=https%3A%2F%2F3dwarehouse.sketchup.com%2F%3Fhl%3Den&usg=AOvVaw3r_zxTadmQ_YWO3IdunfNF
**Browser / Version**: Firefox 76.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Page not loading correctly
**Steps to Reproduce**:
i am continously browsing 3d warehouse but its its still loading
<details><summary>View the screenshot</summary><img alt='Screenshot' src='https://webcompat.com/uploads/2020/4/29ade5ed-f3f3-4280-bb66-2d2cbdbcc517.jpeg'></details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200412214314</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/4/6b58ad1d-0805-4df4-bc29-7f23d9c63a99)
Submitted in the name of `@ashok`
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description page not loading correctly steps to reproduce i am continously browsing warehouse but its its still loading view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false submitted in the name of ashok from with ❤️ | 0 |
996 | 4,760,606,557 | IssuesEvent | 2016-10-25 03:55:57 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | Can't delete a VM with vmware_guest module | affects_2.2 bug_report cloud vmware waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
vmware_guest
##### ANSIBLE VERSION
```
(pip)c7:ansible-demos$ ansible --version
ansible 2.2.0.0
config file = /home/vagrant/src/ansible-demos/ansible.cfg
configured module search path = Default w/o overrides
(pip)c7:ansible-demos$ rpm -q ansible
ansible-2.2.0.0-100.git201610182035.bce9bfc.HEAD.el7.centos.noarch
```
##### CONFIGURATION
##### OS / ENVIRONMENT
(pip)c7:ansible-demos$ cat /etc/centos-release
CentOS Linux release 7.1.1503 (Core)
##### SUMMARY
Trying to delete a VM by setting 'state: absent' fails with vmware_guest module. First it fails if 'template' switch is not set, second it fails by trying to clone when 'template' is set. Also, if just `state: absent` is set before trying to power off, then nothing happens at all.
##### STEPS TO REPRODUCE
```
- name: Shutdown VM first if delete required
vmware_guest:
validate_certs: False
hostname: "{{ vcenter_hostname }}"
username: "{{ vcenter_user }}"
password: "{{ vcenter_pass }}"
name: "{{ name }}"
state: "poweredoff"
template: ""
when: state == "absent"
- name: Set VM to desired state
vmware_guest:
validate_certs: False
hostname: "{{ vcenter_hostname }}"
username: "{{ vcenter_user }}"
password: "{{ vcenter_pass }}"
name: "{{ name }}"
state: "{{ state }}"
template: ""
```
##### EXPECTED RESULTS
VM powered off and deleted.
##### ACTUAL RESULTS
Nothing if just state: absent, failure if try to power off...
```
TASK [Shutdown VM first if delete required] ************************************
fatal: [linux]: FAILED! => {"changed": true, "failed": true, "msg": "The name 'ssp.box' already exists."}
```
Template missing...
```
TASK [Shutdown VM first if delete required] ************************************
fatal: [linux]: FAILED! => {"changed": false, "failed": true, "module_stderr": "/home/vagrant/pip/lib/python2.7/site-packages/requests/packages/urllib3/connectionpool.py:789: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html\n InsecureRequestWarning)\nTraceback (most recent call last):\n File \"/tmp/ansible_5ZzRF6/ansible_module_vmware_guest.py\", line 959, in <module>\n main()\n File \"/tmp/ansible_5ZzRF6/ansible_module_vmware_guest.py\", line 937, in main\n wait_for_ip=module.params['wait_for_ip_address']\n File \"/tmp/ansible_5ZzRF6/ansible_module_vmware_guest.py\", line 494, in deploy_template\n template = get_obj(self.content, [vim.VirtualMachine], self.params['template'])\nKeyError: 'template'\n", "module_stdout": "", "msg": "MODULE FAILURE"}
``` | True | Can't delete a VM with vmware_guest module - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
vmware_guest
##### ANSIBLE VERSION
```
(pip)c7:ansible-demos$ ansible --version
ansible 2.2.0.0
config file = /home/vagrant/src/ansible-demos/ansible.cfg
configured module search path = Default w/o overrides
(pip)c7:ansible-demos$ rpm -q ansible
ansible-2.2.0.0-100.git201610182035.bce9bfc.HEAD.el7.centos.noarch
```
##### CONFIGURATION
##### OS / ENVIRONMENT
(pip)c7:ansible-demos$ cat /etc/centos-release
CentOS Linux release 7.1.1503 (Core)
##### SUMMARY
Trying to delete a VM by setting 'state: absent' fails with vmware_guest module. First it fails if 'template' switch is not set, second it fails by trying to clone when 'template' is set. Also, if just `state: absent` is set before trying to power off, then nothing happens at all.
##### STEPS TO REPRODUCE
```
- name: Shutdown VM first if delete required
vmware_guest:
validate_certs: False
hostname: "{{ vcenter_hostname }}"
username: "{{ vcenter_user }}"
password: "{{ vcenter_pass }}"
name: "{{ name }}"
state: "poweredoff"
template: ""
when: state == "absent"
- name: Set VM to desired state
vmware_guest:
validate_certs: False
hostname: "{{ vcenter_hostname }}"
username: "{{ vcenter_user }}"
password: "{{ vcenter_pass }}"
name: "{{ name }}"
state: "{{ state }}"
template: ""
```
##### EXPECTED RESULTS
VM powered off and deleted.
##### ACTUAL RESULTS
Nothing if just state: absent, failure if try to power off...
```
TASK [Shutdown VM first if delete required] ************************************
fatal: [linux]: FAILED! => {"changed": true, "failed": true, "msg": "The name 'ssp.box' already exists."}
```
Template missing...
```
TASK [Shutdown VM first if delete required] ************************************
fatal: [linux]: FAILED! => {"changed": false, "failed": true, "module_stderr": "/home/vagrant/pip/lib/python2.7/site-packages/requests/packages/urllib3/connectionpool.py:789: InsecureRequestWarning: Unverified HTTPS request is being made. Adding certificate verification is strongly advised. See: https://urllib3.readthedocs.org/en/latest/security.html\n InsecureRequestWarning)\nTraceback (most recent call last):\n File \"/tmp/ansible_5ZzRF6/ansible_module_vmware_guest.py\", line 959, in <module>\n main()\n File \"/tmp/ansible_5ZzRF6/ansible_module_vmware_guest.py\", line 937, in main\n wait_for_ip=module.params['wait_for_ip_address']\n File \"/tmp/ansible_5ZzRF6/ansible_module_vmware_guest.py\", line 494, in deploy_template\n template = get_obj(self.content, [vim.VirtualMachine], self.params['template'])\nKeyError: 'template'\n", "module_stdout": "", "msg": "MODULE FAILURE"}
``` | main | can t delete a vm with vmware guest module issue type bug report component name vmware guest ansible version pip ansible demos ansible version ansible config file home vagrant src ansible demos ansible cfg configured module search path default w o overrides pip ansible demos rpm q ansible ansible head centos noarch configuration os environment pip ansible demos cat etc centos release centos linux release core summary trying to delete a vm by setting state absent fails with vmware guest module first it fails if template switch is not set second it fails by trying to clone when template is set also if just state absent is set before trying to power off then nothing happens at all steps to reproduce name shutdown vm first if delete required vmware guest validate certs false hostname vcenter hostname username vcenter user password vcenter pass name name state poweredoff template when state absent name set vm to desired state vmware guest validate certs false hostname vcenter hostname username vcenter user password vcenter pass name name state state template expected results vm powered off and deleted actual results nothing if just state absent failure if try to power off task fatal failed changed true failed true msg the name ssp box already exists template missing task fatal failed changed false failed true module stderr home vagrant pip lib site packages requests packages connectionpool py insecurerequestwarning unverified https request is being made adding certificate verification is strongly advised see insecurerequestwarning ntraceback most recent call last n file tmp ansible ansible module vmware guest py line in n main n file tmp ansible ansible module vmware guest py line in main n wait for ip module params n file tmp ansible ansible module vmware guest py line in deploy template n template get obj self content self params nkeyerror template n module stdout msg module failure | 1 |
5,660 | 29,182,387,263 | IssuesEvent | 2023-05-19 12:57:48 | jupyter-naas/awesome-notebooks | https://api.github.com/repos/jupyter-naas/awesome-notebooks | opened | Wise - List balances for a profile | templates maintainer | This notebook will list the balances for a profile using the Wise API. It is usefull for organizations to keep track of their financials.
| True | Wise - List balances for a profile - This notebook will list the balances for a profile using the Wise API. It is usefull for organizations to keep track of their financials.
| main | wise list balances for a profile this notebook will list the balances for a profile using the wise api it is usefull for organizations to keep track of their financials | 1 |
3,539 | 13,925,894,247 | IssuesEvent | 2020-10-21 17:29:41 | grey-software/org | https://api.github.com/repos/grey-software/org | opened | 🥅 Initiative: Make our discord server verified | Domain: User Experience Role: Maintainer Role: Product Owner | ### Motivation 🏁
<!--
A clear and concise motivation for this initiative? How will this help execute the vision of the org?
-->
Acceptance for Server Verification grants our server access to a Vanity URL, invite splash background, and the Verified Server badge. | True | 🥅 Initiative: Make our discord server verified - ### Motivation 🏁
<!--
A clear and concise motivation for this initiative? How will this help execute the vision of the org?
-->
Acceptance for Server Verification grants our server access to a Vanity URL, invite splash background, and the Verified Server badge. | main | 🥅 initiative make our discord server verified motivation 🏁 a clear and concise motivation for this initiative how will this help execute the vision of the org acceptance for server verification grants our server access to a vanity url invite splash background and the verified server badge | 1 |
39,912 | 5,154,630,729 | IssuesEvent | 2017-01-15 01:21:01 | Automattic/jetpack | https://api.github.com/repos/Automattic/jetpack | closed | Manage: Strange design on landing page when enabling via Calypso | Admin Page Manage [Pri] Normal [Status] Needs Design Review [Team] Poseidon [Type] Bug | #### Steps to reproduce the issue
1. Connect a site that doesn't have Jetpack manage enabled.
2. Visit `/plugins` for the site in Calypso.
3. Click the button to enable Manage.
4. You'll be taken to this page in wp-admin:
<img width="626" alt="screen shot 2016-10-19 at 1 25 34 pm" src="https://cloud.githubusercontent.com/assets/789137/19541477/7555b89c-961b-11e6-9b4a-ac3d20956a5d.png">
#### What I expected
A nice shiny message that follows the new design standards.
#### What happened instead
I saw a very basic message in wp-admin.
Safari on macOS Sierra.
| 1.0 | Manage: Strange design on landing page when enabling via Calypso - #### Steps to reproduce the issue
1. Connect a site that doesn't have Jetpack manage enabled.
2. Visit `/plugins` for the site in Calypso.
3. Click the button to enable Manage.
4. You'll be taken to this page in wp-admin:
<img width="626" alt="screen shot 2016-10-19 at 1 25 34 pm" src="https://cloud.githubusercontent.com/assets/789137/19541477/7555b89c-961b-11e6-9b4a-ac3d20956a5d.png">
#### What I expected
A nice shiny message that follows the new design standards.
#### What happened instead
I saw a very basic message in wp-admin.
Safari on macOS Sierra.
| non_main | manage strange design on landing page when enabling via calypso steps to reproduce the issue connect a site that doesn t have jetpack manage enabled visit plugins for the site in calypso click the button to enable manage you ll be taken to this page in wp admin img width alt screen shot at pm src what i expected a nice shiny message that follows the new design standards what happened instead i saw a very basic message in wp admin safari on macos sierra | 0 |
248 | 2,997,296,281 | IssuesEvent | 2015-07-23 06:19:15 | AKST/akst.io-backend | https://api.github.com/repos/AKST/akst.io-backend | closed | 100% coverage | maintainablity | Get 100% coverage with tests
#### `Posts`
- [x] ~~`GET /post`~~
- [x] ~~`GET /post/{id}`~~
- [x] ~~`POST /posts` #15~~
- [x] ~~`DELETE /posts/{id}` #17~~
- [x] ~~`PUT /posts/{id}` #16~~
#### `Revisions`
- [x] ~~`GET /revisions/{id}` #18~~
- [x] ~~`POST /revisions` #20~~
- [x] ~~`PUT /revisions/{id}` #22~~
| True | 100% coverage - Get 100% coverage with tests
#### `Posts`
- [x] ~~`GET /post`~~
- [x] ~~`GET /post/{id}`~~
- [x] ~~`POST /posts` #15~~
- [x] ~~`DELETE /posts/{id}` #17~~
- [x] ~~`PUT /posts/{id}` #16~~
#### `Revisions`
- [x] ~~`GET /revisions/{id}` #18~~
- [x] ~~`POST /revisions` #20~~
- [x] ~~`PUT /revisions/{id}` #22~~
| main | coverage get coverage with tests posts get post get post id post posts delete posts id put posts id revisions get revisions id post revisions put revisions id | 1 |
42,348 | 10,987,217,584 | IssuesEvent | 2019-12-02 08:45:14 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | Table state demo broken in 9.0.0-rc.2-SNAPSHOT | defect pending-review | **I'm submitting a bug report**
The [TableStateDemo](http://localhost:4200/#/table/state) does not work properly any more in the current master branch.
In order to reproduce, check out the current [9.0.0-rc.2-SNAPSHOT](https://github.com/primefaces/primeng/commit/40e39f5c191a0563246120bc9f64e9344c161617) and run "ng serve" on it.
Browse to http://localhost:4200/#/table/state and change the current page in the table from 1 to 2. Then reload the page.
**Current behavior**
The current page in the table is reset to 1.
**Expected behavior**
The current page in the table should be restored to 2. | 1.0 | Table state demo broken in 9.0.0-rc.2-SNAPSHOT - **I'm submitting a bug report**
The [TableStateDemo](http://localhost:4200/#/table/state) does not work properly any more in the current master branch.
In order to reproduce, check out the current [9.0.0-rc.2-SNAPSHOT](https://github.com/primefaces/primeng/commit/40e39f5c191a0563246120bc9f64e9344c161617) and run "ng serve" on it.
Browse to http://localhost:4200/#/table/state and change the current page in the table from 1 to 2. Then reload the page.
**Current behavior**
The current page in the table is reset to 1.
**Expected behavior**
The current page in the table should be restored to 2. | non_main | table state demo broken in rc snapshot i m submitting a bug report the does not work properly any more in the current master branch in order to reproduce check out the current and run ng serve on it browse to and change the current page in the table from to then reload the page current behavior the current page in the table is reset to expected behavior the current page in the table should be restored to | 0 |
3,940 | 17,766,870,824 | IssuesEvent | 2021-08-30 08:40:11 | DLR-RM/rl-baselines3-zoo | https://api.github.com/repos/DLR-RM/rl-baselines3-zoo | closed | "not iterable" TypeError due to exp_manager.py's is_atari()/is_bullet()/is_robotics_env() assuming entry_point of type str | bug Maintainers on vacation | **Describe the bug**
In OpenAI Gym's [`register` function](https://github.com/openai/gym/blob/master/gym/envs/__init__.py), the keyword argument `entry_point` accepts a value of type `str` as well as a `gym.Env` subclass.
If a custom environment has been registered using the latter option (as a class), `exp_manager.py` produces a `TypeError: 'type' object is not iterable` when trying to check whether the string `"AtariEnv"` is contained in `gym.envs.registry.env_specs[env_id].entry_point`.
**Code example**
```python
# myexample.py
# ...assuming that some gym env class MyExample exists...
from gym.envs.registration import register
register(
id="MyExample-v0",
entry_point=MyExample,
)
```
```python
# utils/import_envs.py
import myexample
```
(as [recommended here](https://github.com/DLR-RM/rl-baselines3-zoo#custom-environment))
```sh
> python train.py --env MyExample-v0
========== MyExample-v0 ==========
Seed: 1193172183
EnvSpec(MyExample-v0)
Traceback (most recent call last):
File "train.py", line 181, in <module>
no_optim_plots=args.no_optim_plots,
File "/Users/asschude/Documents/PhD/code/rl-baselines3-zoo/utils/exp_manager.py", line 116, in __init__
self._is_atari = self.is_atari(env_id)
File "/Users/asschude/Documents/PhD/code/rl-baselines3-zoo/utils/exp_manager.py", line 426, in is_atari
return "AtariEnv" in gym.envs.registry.env_specs[env_id].entry_point
TypeError: argument of type 'type' is not iterable
```
To restore full compatibility with gym's `register`, I suggest simply changing the check to `"AtariEnv" in str(...)`, which will be using the class `__str__`/`__repr__` representation. Same for the other two checks.
I came across this when implementing a custom environment.
**System Info**
Describe the characteristic of your environment:
* cloned SB3 from GitHub (https://github.com/DLR-RM/rl-baselines3-zoo/commit/4f97b7348ccddf387462de8c14d39b1e49bf9d99)
* Python 3.7.7
* torch==1.9.0
* gym==0.18.3
| True | "not iterable" TypeError due to exp_manager.py's is_atari()/is_bullet()/is_robotics_env() assuming entry_point of type str - **Describe the bug**
In OpenAI Gym's [`register` function](https://github.com/openai/gym/blob/master/gym/envs/__init__.py), the keyword argument `entry_point` accepts a value of type `str` as well as a `gym.Env` subclass.
If a custom environment has been registered using the latter option (as a class), `exp_manager.py` produces a `TypeError: 'type' object is not iterable` when trying to check whether the string `"AtariEnv"` is contained in `gym.envs.registry.env_specs[env_id].entry_point`.
**Code example**
```python
# myexample.py
# ...assuming that some gym env class MyExample exists...
from gym.envs.registration import register
register(
id="MyExample-v0",
entry_point=MyExample,
)
```
```python
# utils/import_envs.py
import myexample
```
(as [recommended here](https://github.com/DLR-RM/rl-baselines3-zoo#custom-environment))
```sh
> python train.py --env MyExample-v0
========== MyExample-v0 ==========
Seed: 1193172183
EnvSpec(MyExample-v0)
Traceback (most recent call last):
File "train.py", line 181, in <module>
no_optim_plots=args.no_optim_plots,
File "/Users/asschude/Documents/PhD/code/rl-baselines3-zoo/utils/exp_manager.py", line 116, in __init__
self._is_atari = self.is_atari(env_id)
File "/Users/asschude/Documents/PhD/code/rl-baselines3-zoo/utils/exp_manager.py", line 426, in is_atari
return "AtariEnv" in gym.envs.registry.env_specs[env_id].entry_point
TypeError: argument of type 'type' is not iterable
```
To restore full compatibility with gym's `register`, I suggest simply changing the check to `"AtariEnv" in str(...)`, which will be using the class `__str__`/`__repr__` representation. Same for the other two checks.
I came across this when implementing a custom environment.
**System Info**
Describe the characteristic of your environment:
* cloned SB3 from GitHub (https://github.com/DLR-RM/rl-baselines3-zoo/commit/4f97b7348ccddf387462de8c14d39b1e49bf9d99)
* Python 3.7.7
* torch==1.9.0
* gym==0.18.3
| main | not iterable typeerror due to exp manager py s is atari is bullet is robotics env assuming entry point of type str describe the bug in openai gym s the keyword argument entry point accepts a value of type str as well as a gym env subclass if a custom environment has been registered using the latter option as a class exp manager py produces a typeerror type object is not iterable when trying to check whether the string atarienv is contained in gym envs registry env specs entry point code example python myexample py assuming that some gym env class myexample exists from gym envs registration import register register id myexample entry point myexample python utils import envs py import myexample as sh python train py env myexample myexample seed envspec myexample traceback most recent call last file train py line in no optim plots args no optim plots file users asschude documents phd code rl zoo utils exp manager py line in init self is atari self is atari env id file users asschude documents phd code rl zoo utils exp manager py line in is atari return atarienv in gym envs registry env specs entry point typeerror argument of type type is not iterable to restore full compatibility with gym s register i suggest simply changing the check to atarienv in str which will be using the class str repr representation same for the other two checks i came across this when implementing a custom environment system info describe the characteristic of your environment cloned from github python torch gym | 1 |
504,267 | 14,615,969,810 | IssuesEvent | 2020-12-22 12:28:41 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | www.facebook.com - site is not usable | browser-firefox engine-gecko ml-needsdiagnosis-false priority-critical | <!-- @browser: Firefox 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:85.0) Gecko/20100101 Firefox/85.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/64107 -->
**URL**: https://www.facebook.com/media/set/?vanity=claycountycatcare.auctions&set=a.111631580665415
**Browser / Version**: Firefox 85.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Missing items
**Steps to Reproduce**:
I try to look at the photos in the albums but I get a grey page with no content
Message reads This content isn't available right now
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/b26619b6-dd71-4a49-8cf8-1e04957b52f2.jpeg">
</details>
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/2700ba24-f6a4-43c8-8aba-99c72ef6db98.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201217185930</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/26edaa9b-d033-45cf-b166-c701388cca5e)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | www.facebook.com - site is not usable - <!-- @browser: Firefox 85.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:85.0) Gecko/20100101 Firefox/85.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/64107 -->
**URL**: https://www.facebook.com/media/set/?vanity=claycountycatcare.auctions&set=a.111631580665415
**Browser / Version**: Firefox 85.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Chrome
**Problem type**: Site is not usable
**Description**: Missing items
**Steps to Reproduce**:
I try to look at the photos in the albums but I get a grey page with no content
Message reads This content isn't available right now
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/b26619b6-dd71-4a49-8cf8-1e04957b52f2.jpeg">
</details>
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/12/2700ba24-f6a4-43c8-8aba-99c72ef6db98.jpg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20201217185930</li><li>channel: beta</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/12/26edaa9b-d033-45cf-b166-c701388cca5e)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_main | site is not usable url browser version firefox operating system windows tested another browser yes chrome problem type site is not usable description missing items steps to reproduce i try to look at the photos in the albums but i get a grey page with no content message reads this content isn t available right now view the screenshot img alt screenshot src view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel beta hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
89,507 | 8,205,488,280 | IssuesEvent | 2018-09-03 10:16:40 | humera987/HumTestData | https://api.github.com/repos/humera987/HumTestData | opened | humerafxtesting : api_v1_orgs_by-user_get_query_param_sql_injection_generic_page | humerafxtesting | Project : humerafxtesting
Job : UAT
Env : UAT
Region : FXLabs/US_WEST_1
Result : fail
Status Code : 200
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 03 Sep 2018 10:16:37 GMT]}
Endpoint : http://13.56.210.25/api/v1/orgs/by-user?page=
Request :
Response :
{
"requestId" : "None",
"requestTime" : "2018-09-03T10:16:38.108+0000",
"errors" : false,
"messages" : [ ],
"data" : [ {
"id" : "8a80809e659e63cf01659e9cf890228d",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:46:52.048+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:46:52.048+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80809e659e63cf01659e9cf890228c",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:46:52.048+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:46:52.048+0000",
"version" : null,
"inactive" : false,
"name" : "test2"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80809e659e63cf01659e9be8db228a",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:45:42.491+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:45:42.491+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80809e659e63cf01659e9be8da2289",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:45:42.490+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:45:42.490+0000",
"version" : null,
"inactive" : false,
"name" : "test"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80809e659e63cf01659e90fb6d11a1",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:33:46.349+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:33:46.349+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80809e659e63cf01659e90fb6611a0",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:33:46.342+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:33:46.342+0000",
"version" : null,
"inactive" : true,
"name" : "stest"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80808e659e37cc01659e465a442d3f",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T07:12:15.428+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T07:12:15.428+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80808e659e37cc01659e465a3e2d3e",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T07:12:15.421+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T07:12:15.421+0000",
"version" : null,
"inactive" : false,
"name" : "testorg"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a808060659d915401659dbbc2b206a5",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T04:40:52.658+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T04:40:52.658+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a808060659d915401659dbbc2a906a4",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T04:40:52.649+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T04:40:52.649+0000",
"version" : null,
"inactive" : false,
"name" : "testorg123"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80804d658e61e201658ea0c6784797",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-31T06:17:05.912+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-31T06:17:05.912+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80804d658e61e201658ea0c6714796",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-31T06:17:05.905+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-31T06:17:05.905+0000",
"version" : null,
"inactive" : false,
"name" : "testfx"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080886583d7bd0165840144b00b0e",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-29T04:46:40.304+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-29T04:46:40.304+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080886583d7bd0165840144850b0c",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-29T04:46:40.261+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-29T04:46:40.261+0000",
"version" : null,
"inactive" : false,
"name" : "FXL"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a808057657fff9c01658011a8950d9b",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T10:26:05.589+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T10:26:05.589+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a808057657fff9c01658011a8940d9a",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T10:26:05.588+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T10:26:05.588+0000",
"version" : null,
"inactive" : false,
"name" : "Facebook1"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a808057657fff9c0165800e5eea0c15",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T10:22:30.122+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T10:22:30.122+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a808057657fff9c0165800e5ee90c14",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T10:22:30.121+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T10:22:30.121+0000",
"version" : null,
"inactive" : false,
"name" : "Org 12"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ee657f422301657f6dd1fd000b",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T07:27:08.285+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T07:27:08.285+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ee657f422301657f6dd1fc000a",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T07:27:08.283+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T07:27:08.283+0000",
"version" : null,
"inactive" : false,
"name" : "Ola Uber"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ee657f422301657f6510590001",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T07:17:34.425+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T07:17:34.425+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ee657f422301657f65103f0000",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T07:17:34.396+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T07:17:34.396+0000",
"version" : null,
"inactive" : false,
"name" : "Cloud Space"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080e7657b6af401657b983d02002f",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T13:34:59.330+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T13:34:59.330+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080e7657b6af401657b983d00002e",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T13:34:59.328+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T13:34:59.328+0000",
"version" : null,
"inactive" : false,
"name" : "Watsapp"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080e7657b6af401657b77a37d000b",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:59:22.877+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:59:22.877+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080e7657b6af401657b77a37b000a",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:59:22.875+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:59:22.875+0000",
"version" : null,
"inactive" : false,
"name" : "Uber"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080e7657b6af401657b750cb10002",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:56:33.200+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:56:33.200+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080e7657b6af401657b750ca80001",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:56:33.192+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:56:33.192+0000",
"version" : null,
"inactive" : false,
"name" : "Ola"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ec657b1b2b01657b5d53b90af9",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:30:38.521+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:30:38.521+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ec657b1b2b01657b5d53b80af8",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:30:38.520+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:30:38.520+0000",
"version" : null,
"inactive" : false,
"name" : "Facebook"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ec657b1b2b01657b518e835900",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:17:47.139+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:17:47.139+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ec657b1b2b01657b518e8258ff",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:17:47.138+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:17:47.138+0000",
"version" : null,
"inactive" : false,
"name" : "Microsoft"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ec657b1b2b01657b49da773826",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:09:22.295+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:09:22.295+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ec657b1b2b01657b49da763825",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:09:22.294+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:09:22.294+0000",
"version" : null,
"inactive" : false,
"name" : "Google"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ec657b1b2b01657b1fab6f0675",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T11:23:17.743+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T11:23:17.743+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ec657b1b2b01657b1fab6f0674",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T11:23:17.743+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T11:23:17.743+0000",
"version" : null,
"inactive" : false,
"name" : "FXLabs_UI"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80808a657aacf801657af6998d0020",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T10:38:26.189+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T10:38:26.189+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80808a657aacf801657af6998c001f",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T10:38:26.188+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T11:27:59.360+0000",
"version" : null,
"inactive" : false,
"name" : "FXLabs_QA"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80808a657aacf801657ab17dcc0002",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:57.099+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:57.108+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80808a657aacf801657ab17dd30003",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:57.107+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:57.107+0000",
"version" : null,
"inactive" : false,
"name" : "FXLabs"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ENTERPRISE_ADMIN",
"status" : "ACTIVE"
} ],
"totalPages" : 1,
"totalElements" : 20
}
Logs :
Assertion [@StatusCode != 200] failed, not expecting [200] but found [200]
--- FX Bot --- | 1.0 | humerafxtesting : api_v1_orgs_by-user_get_query_param_sql_injection_generic_page - Project : humerafxtesting
Job : UAT
Env : UAT
Region : FXLabs/US_WEST_1
Result : fail
Status Code : 200
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 03 Sep 2018 10:16:37 GMT]}
Endpoint : http://13.56.210.25/api/v1/orgs/by-user?page=
Request :
Response :
{
"requestId" : "None",
"requestTime" : "2018-09-03T10:16:38.108+0000",
"errors" : false,
"messages" : [ ],
"data" : [ {
"id" : "8a80809e659e63cf01659e9cf890228d",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:46:52.048+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:46:52.048+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80809e659e63cf01659e9cf890228c",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:46:52.048+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:46:52.048+0000",
"version" : null,
"inactive" : false,
"name" : "test2"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80809e659e63cf01659e9be8db228a",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:45:42.491+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:45:42.491+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80809e659e63cf01659e9be8da2289",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:45:42.490+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:45:42.490+0000",
"version" : null,
"inactive" : false,
"name" : "test"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80809e659e63cf01659e90fb6d11a1",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:33:46.349+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:33:46.349+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80809e659e63cf01659e90fb6611a0",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T08:33:46.342+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T08:33:46.342+0000",
"version" : null,
"inactive" : true,
"name" : "stest"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80808e659e37cc01659e465a442d3f",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T07:12:15.428+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T07:12:15.428+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80808e659e37cc01659e465a3e2d3e",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T07:12:15.421+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T07:12:15.421+0000",
"version" : null,
"inactive" : false,
"name" : "testorg"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a808060659d915401659dbbc2b206a5",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T04:40:52.658+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T04:40:52.658+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a808060659d915401659dbbc2a906a4",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-09-03T04:40:52.649+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-09-03T04:40:52.649+0000",
"version" : null,
"inactive" : false,
"name" : "testorg123"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80804d658e61e201658ea0c6784797",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-31T06:17:05.912+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-31T06:17:05.912+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80804d658e61e201658ea0c6714796",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-31T06:17:05.905+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-31T06:17:05.905+0000",
"version" : null,
"inactive" : false,
"name" : "testfx"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080886583d7bd0165840144b00b0e",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-29T04:46:40.304+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-29T04:46:40.304+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080886583d7bd0165840144850b0c",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-29T04:46:40.261+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-29T04:46:40.261+0000",
"version" : null,
"inactive" : false,
"name" : "FXL"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a808057657fff9c01658011a8950d9b",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T10:26:05.589+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T10:26:05.589+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a808057657fff9c01658011a8940d9a",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T10:26:05.588+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T10:26:05.588+0000",
"version" : null,
"inactive" : false,
"name" : "Facebook1"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a808057657fff9c0165800e5eea0c15",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T10:22:30.122+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T10:22:30.122+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a808057657fff9c0165800e5ee90c14",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T10:22:30.121+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T10:22:30.121+0000",
"version" : null,
"inactive" : false,
"name" : "Org 12"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ee657f422301657f6dd1fd000b",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T07:27:08.285+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T07:27:08.285+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ee657f422301657f6dd1fc000a",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T07:27:08.283+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T07:27:08.283+0000",
"version" : null,
"inactive" : false,
"name" : "Ola Uber"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ee657f422301657f6510590001",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T07:17:34.425+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T07:17:34.425+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ee657f422301657f65103f0000",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-28T07:17:34.396+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-28T07:17:34.396+0000",
"version" : null,
"inactive" : false,
"name" : "Cloud Space"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080e7657b6af401657b983d02002f",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T13:34:59.330+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T13:34:59.330+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080e7657b6af401657b983d00002e",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T13:34:59.328+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T13:34:59.328+0000",
"version" : null,
"inactive" : false,
"name" : "Watsapp"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080e7657b6af401657b77a37d000b",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:59:22.877+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:59:22.877+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080e7657b6af401657b77a37b000a",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:59:22.875+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:59:22.875+0000",
"version" : null,
"inactive" : false,
"name" : "Uber"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080e7657b6af401657b750cb10002",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:56:33.200+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:56:33.200+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080e7657b6af401657b750ca80001",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:56:33.192+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:56:33.192+0000",
"version" : null,
"inactive" : false,
"name" : "Ola"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ec657b1b2b01657b5d53b90af9",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:30:38.521+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:30:38.521+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ec657b1b2b01657b5d53b80af8",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:30:38.520+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:30:38.520+0000",
"version" : null,
"inactive" : false,
"name" : "Facebook"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ec657b1b2b01657b518e835900",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:17:47.139+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:17:47.139+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ec657b1b2b01657b518e8258ff",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:17:47.138+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:17:47.138+0000",
"version" : null,
"inactive" : false,
"name" : "Microsoft"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ec657b1b2b01657b49da773826",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:09:22.295+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:09:22.295+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ec657b1b2b01657b49da763825",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T12:09:22.294+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T12:09:22.294+0000",
"version" : null,
"inactive" : false,
"name" : "Google"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a8080ec657b1b2b01657b1fab6f0675",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T11:23:17.743+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T11:23:17.743+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a8080ec657b1b2b01657b1fab6f0674",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T11:23:17.743+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T11:23:17.743+0000",
"version" : null,
"inactive" : false,
"name" : "FXLabs_UI"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80808a657aacf801657af6998d0020",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T10:38:26.189+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T10:38:26.189+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80808a657aacf801657af6998c001f",
"createdBy" : "8a80808a657aacf801657ab17ca30000",
"createdDate" : "2018-08-27T10:38:26.188+0000",
"modifiedBy" : "8a80808a657aacf801657ab17ca30000",
"modifiedDate" : "2018-08-27T11:27:59.360+0000",
"version" : null,
"inactive" : false,
"name" : "FXLabs_QA"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ADMIN",
"status" : "ACTIVE"
}, {
"id" : "8a80808a657aacf801657ab17dcc0002",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:57.099+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:57.108+0000",
"version" : null,
"inactive" : false,
"org" : {
"id" : "8a80808a657aacf801657ab17dd30003",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:57.107+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:57.107+0000",
"version" : null,
"inactive" : false,
"name" : "FXLabs"
},
"users" : {
"id" : "8a80808a657aacf801657ab17ca30000",
"createdBy" : "anonymousUser",
"createdDate" : "2018-08-27T09:22:56.800+0000",
"modifiedBy" : "anonymousUser",
"modifiedDate" : "2018-08-27T09:22:56.800+0000",
"version" : null,
"inactive" : false,
"name" : null,
"email" : "admin@fxlabs.io",
"username" : "admin",
"company" : null,
"location" : null,
"jobTitle" : null
},
"orgRole" : "ENTERPRISE_ADMIN",
"status" : "ACTIVE"
} ],
"totalPages" : 1,
"totalElements" : 20
}
Logs :
Assertion [@StatusCode != 200] failed, not expecting [200] but found [200]
--- FX Bot --- | non_main | humerafxtesting api orgs by user get query param sql injection generic page project humerafxtesting job uat env uat region fxlabs us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options content type transfer encoding date endpoint request response requestid none requesttime errors false messages data id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name test users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive true name stest users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name testorg users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name testfx users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name fxl users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name org users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name ola uber users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name cloud space users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name watsapp users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name uber users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name ola users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name facebook users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name microsoft users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name google users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name fxlabs ui users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby createddate modifiedby modifieddate version null inactive false org id createdby createddate modifiedby modifieddate version null inactive false name fxlabs qa users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole admin status active id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false org id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name fxlabs users id createdby anonymoususer createddate modifiedby anonymoususer modifieddate version null inactive false name null email admin fxlabs io username admin company null location null jobtitle null orgrole enterprise admin status active totalpages totalelements logs assertion failed not expecting but found fx bot | 0 |
184,914 | 32,069,641,643 | IssuesEvent | 2023-09-25 07:05:35 | thesaurus-linguae-aegyptiae/tla-web | https://api.github.com/repos/thesaurus-linguae-aegyptiae/tla-web | opened | AUfzählungssymbol überdenken | low priority layout/design | Aktuell: schwarzer Pfeil.
PD
> Der schwarze Pfeil erweckt den Eindruck, dass es ein Link ist. | 1.0 | AUfzählungssymbol überdenken - Aktuell: schwarzer Pfeil.
PD
> Der schwarze Pfeil erweckt den Eindruck, dass es ein Link ist. | non_main | aufzählungssymbol überdenken aktuell schwarzer pfeil pd der schwarze pfeil erweckt den eindruck dass es ein link ist | 0 |
1,781 | 6,575,830,765 | IssuesEvent | 2017-09-11 17:29:42 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | stat.exists regression | affects_2.1 bug_report waiting_on_maintainer | <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
stat
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
2.1.2.0
```
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Travis CI
##### SUMMARY
<!--- Explain the problem briefly -->
My playbook that tries to check whether file exists or not started to fails with ansible 2.1.2.0 release. It worked on 2.1.1.0 version.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Getting info about WAR file
stat:
path: 'target/mystamps.war'
get_checksum: no
get_md5: no
register: war_file
become: no
delegate_to: 127.0.0.1
- name: Ensuring that WAR file exists
assert:
that:
war_file.stat.exists
become: no
delegate_to: 127.0.0.1
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
```
TASK [Getting info about WAR file] *********************************************
ok: [my-stamps.ru -> 127.0.0.1]
TASK [Ensuring that WAR file exists] *******************************************
ok: [my-stamps.ru -> 127.0.0.1]
```
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
TASK [Getting info about WAR file] *********************************************
ok: [my-stamps.ru -> 127.0.0.1]
TASK [Ensuring that WAR file exists] *******************************************
fatal: [my-stamps.ru -> 127.0.0.1]: FAILED! => {"assertion": "war_file.stat.exists", "changed": false, "evaluated_to": false, "failed": true}
```
| True | stat.exists regression - <!--- Verify first that your issue/request is not already reported in GitHub -->
##### ISSUE TYPE
<!--- Pick one below and delete the rest: -->
- Bug Report
##### COMPONENT NAME
<!--- Name of the plugin/module/task -->
stat
##### ANSIBLE VERSION
<!--- Paste verbatim output from “ansible --version” between quotes below -->
```
2.1.2.0
```
##### OS / ENVIRONMENT
<!---
Mention the OS you are running Ansible from, and the OS you are
managing, or say “N/A” for anything that is not platform-specific.
-->
Travis CI
##### SUMMARY
<!--- Explain the problem briefly -->
My playbook that tries to check whether file exists or not started to fails with ansible 2.1.2.0 release. It worked on 2.1.1.0 version.
##### STEPS TO REPRODUCE
<!---
For bugs, show exactly how to reproduce the problem.
For new features, show how the feature would be used.
-->
<!--- Paste example playbooks or commands between quotes below -->
```
- name: Getting info about WAR file
stat:
path: 'target/mystamps.war'
get_checksum: no
get_md5: no
register: war_file
become: no
delegate_to: 127.0.0.1
- name: Ensuring that WAR file exists
assert:
that:
war_file.stat.exists
become: no
delegate_to: 127.0.0.1
```
<!--- You can also paste gist.github.com links for larger files -->
##### EXPECTED RESULTS
<!--- What did you expect to happen when running the steps above? -->
```
TASK [Getting info about WAR file] *********************************************
ok: [my-stamps.ru -> 127.0.0.1]
TASK [Ensuring that WAR file exists] *******************************************
ok: [my-stamps.ru -> 127.0.0.1]
```
##### ACTUAL RESULTS
<!--- What actually happened? If possible run with extra verbosity (-vvvv) -->
<!--- Paste verbatim command output between quotes below -->
```
TASK [Getting info about WAR file] *********************************************
ok: [my-stamps.ru -> 127.0.0.1]
TASK [Ensuring that WAR file exists] *******************************************
fatal: [my-stamps.ru -> 127.0.0.1]: FAILED! => {"assertion": "war_file.stat.exists", "changed": false, "evaluated_to": false, "failed": true}
```
| main | stat exists regression issue type bug report component name stat ansible version os environment mention the os you are running ansible from and the os you are managing or say “n a” for anything that is not platform specific travis ci summary my playbook that tries to check whether file exists or not started to fails with ansible release it worked on version steps to reproduce for bugs show exactly how to reproduce the problem for new features show how the feature would be used name getting info about war file stat path target mystamps war get checksum no get no register war file become no delegate to name ensuring that war file exists assert that war file stat exists become no delegate to expected results task ok task ok actual results task ok task fatal failed assertion war file stat exists changed false evaluated to false failed true | 1 |
217,241 | 16,684,446,412 | IssuesEvent | 2021-06-08 06:17:04 | chef/chef-web-docs | https://api.github.com/repos/chef/chef-web-docs | opened | [INTAKE] Minor restructure of Chef Desktop documentation | Documentation Status: Untriaged Type: Docs Intake | ## General details
Link to page & section that needs updated: https://docs.chef.io/desktop/
Business group: (ex. product, engineering, marketing) product
Product owner: Nischal Reddy
Team: Chef Desktop
Date update needs to be published: July 15
## Product
## Product
- [ ] Chef Automate
- [ ] Chef Habitat
- [ ] Chef Infra Client
- [ ] Chef Infra Server
- [ ] Chef InSpec
- [ ] Chef Workstation
- [x] Other: Chef Desktop
## Update details
Describe the update that needs to be made. Be as specific as possible:
Minor refactor of the Chef Desktop documentation's menu based on the changes in the content.
This update contains following:
- Restructured the documentation to give a better flow
- Added a new section for getting started through terraform scripts
## Existing resources
GitHub Repository: https://github.com/chef/desktop-config (desktop documentation), https://github.com/chef/chef-desktop-quick-setup (terraform scripts)
Issue link(s):
PR link(s): https://github.com/chef/chef-web-docs/pull/3164, https://github.com/chef/desktop-config/pull/366
AHA! link(s): https://chef.aha.io/epics/DESK-E-3
Other draft documentation:
Related design, business docs, or other information: https://progresssoftware.sharepoint.com/:w:/s/ChefCoreC/ERGd5_FwKLpNo9K28SV7fNUBXAxFSxRdtir5OTTaN69kfw?e=MvyvVl
## Review
Name of technical reviewer:
| 1.0 | [INTAKE] Minor restructure of Chef Desktop documentation - ## General details
Link to page & section that needs updated: https://docs.chef.io/desktop/
Business group: (ex. product, engineering, marketing) product
Product owner: Nischal Reddy
Team: Chef Desktop
Date update needs to be published: July 15
## Product
## Product
- [ ] Chef Automate
- [ ] Chef Habitat
- [ ] Chef Infra Client
- [ ] Chef Infra Server
- [ ] Chef InSpec
- [ ] Chef Workstation
- [x] Other: Chef Desktop
## Update details
Describe the update that needs to be made. Be as specific as possible:
Minor refactor of the Chef Desktop documentation's menu based on the changes in the content.
This update contains following:
- Restructured the documentation to give a better flow
- Added a new section for getting started through terraform scripts
## Existing resources
GitHub Repository: https://github.com/chef/desktop-config (desktop documentation), https://github.com/chef/chef-desktop-quick-setup (terraform scripts)
Issue link(s):
PR link(s): https://github.com/chef/chef-web-docs/pull/3164, https://github.com/chef/desktop-config/pull/366
AHA! link(s): https://chef.aha.io/epics/DESK-E-3
Other draft documentation:
Related design, business docs, or other information: https://progresssoftware.sharepoint.com/:w:/s/ChefCoreC/ERGd5_FwKLpNo9K28SV7fNUBXAxFSxRdtir5OTTaN69kfw?e=MvyvVl
## Review
Name of technical reviewer:
| non_main | minor restructure of chef desktop documentation general details link to page section that needs updated business group ex product engineering marketing product product owner nischal reddy team chef desktop date update needs to be published july product product chef automate chef habitat chef infra client chef infra server chef inspec chef workstation other chef desktop update details describe the update that needs to be made be as specific as possible minor refactor of the chef desktop documentation s menu based on the changes in the content this update contains following restructured the documentation to give a better flow added a new section for getting started through terraform scripts existing resources github repository desktop documentation terraform scripts issue link s pr link s aha link s other draft documentation related design business docs or other information review name of technical reviewer | 0 |
177,271 | 28,433,346,929 | IssuesEvent | 2023-04-15 02:41:56 | junpotatoes/TopChart | https://api.github.com/repos/junpotatoes/TopChart | closed | TrackDetail 컴포넌트 디자인 수정 | Design | ## 만들고자 하는 기능이 무엇인가요?
TrackDetail 컴포넌트 디자인 수정
## 해당 기능을 구현하기 위해 할 일이 무엇인가요?
- [x] 뒤로가기 버튼 모바일 렌더시 위치이동
| 1.0 | TrackDetail 컴포넌트 디자인 수정 - ## 만들고자 하는 기능이 무엇인가요?
TrackDetail 컴포넌트 디자인 수정
## 해당 기능을 구현하기 위해 할 일이 무엇인가요?
- [x] 뒤로가기 버튼 모바일 렌더시 위치이동
| non_main | trackdetail 컴포넌트 디자인 수정 만들고자 하는 기능이 무엇인가요 trackdetail 컴포넌트 디자인 수정 해당 기능을 구현하기 위해 할 일이 무엇인가요 뒤로가기 버튼 모바일 렌더시 위치이동 | 0 |
52,687 | 6,650,277,985 | IssuesEvent | 2017-09-28 15:47:33 | 18F/openFEC-web-app | https://api.github.com/repos/18F/openFEC-web-app | closed | Add process details and role documentation to research<>design process diagram | Internal Work: Design | Using the process sketch from https://github.com/18F/openFEC-web-app/issues/1352
It could be helpful for new team members, deliverable documentation, and 18F comms/outreach for us to add detail about how each step is carried out: both more detail in how to do it that step, and who is involved/responsible for completing.
Filing this issue now, but putting it in the backlog for a few sprints as we test out some of the process changes.
| 1.0 | Add process details and role documentation to research<>design process diagram - Using the process sketch from https://github.com/18F/openFEC-web-app/issues/1352
It could be helpful for new team members, deliverable documentation, and 18F comms/outreach for us to add detail about how each step is carried out: both more detail in how to do it that step, and who is involved/responsible for completing.
Filing this issue now, but putting it in the backlog for a few sprints as we test out some of the process changes.
| non_main | add process details and role documentation to research design process diagram using the process sketch from it could be helpful for new team members deliverable documentation and comms outreach for us to add detail about how each step is carried out both more detail in how to do it that step and who is involved responsible for completing filing this issue now but putting it in the backlog for a few sprints as we test out some of the process changes | 0 |
4,385 | 7,065,423,177 | IssuesEvent | 2018-01-06 19:52:28 | brcontainer/stack-exchange-notifications | https://api.github.com/repos/brcontainer/stack-exchange-notifications | closed | Does not work on Firefox v47.0 | browser-compatibility invalid | It shows:
```
Response error:
You must be logged to https://stackexchange.com/
```
I did logged in Stack Exchange. | True | Does not work on Firefox v47.0 - It shows:
```
Response error:
You must be logged to https://stackexchange.com/
```
I did logged in Stack Exchange. | non_main | does not work on firefox it shows response error you must be logged to i did logged in stack exchange | 0 |
1,303 | 5,542,101,856 | IssuesEvent | 2017-03-22 14:22:57 | ansible/ansible-modules-extras | https://api.github.com/repos/ansible/ansible-modules-extras | closed | [Enhancement] Add a "provider" parameter to snmp_facts module | affects_2.3 feature_idea networking waiting_on_maintainer | ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
snmp_facts
##### ANSIBLE VERSION
```
ansible 2.3.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
inventory = ./hosts
gathering = explicit
roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/git-Ansible/roles
private_role_vars = yes
log_path = /var/log/ansible.log
fact_caching = redis
fact_caching_timeout = 86400
retry_files_enabled = False
##### OS / ENVIRONMENT
- host: Ubuntu 16.04 4.4.0
- target: **all network devices**
##### SUMMARY
snmp_facts does not offer a provider parameter, unlike many other networking modules from Cisco, Arista, Juniper, and so on.... This feature would be very useful, allowing us to hand over only one structured variable which can be reused in other roles.
This feature should be extended to all network, system, ... modules.
##### STEPS TO REPRODUCE
Structure which could be handed over as "**provider**": snmp.ssh
``````
snmp
...
ssh:
transport: cli
host: "{{ ansible_host }}"
# Username is part of the following required group in v3
group_name: admin
# Username required if version is v3
username: admin
# New SNMP Version to use
version: v3
# community string, required if version is v1/v2c - up to 32 alphanumeric characters long. Spaces are not permitted.
community: xxxxxxxx
# Authentication level, required if version is v3
level: authPriv
# Authentication key, required if version is v3
authkey: xxxxxxxx
# Type of password
authkey_type:
# Hashing algoritm, required if version is v3
integrity: sha
# Encryption algoritm, required if level is authPriv
privacy: aes
# Encryption key, required if version is authPriv
privkey: xxxxxxxx```
``````
| True | [Enhancement] Add a "provider" parameter to snmp_facts module - ##### ISSUE TYPE
- Feature Idea
##### COMPONENT NAME
snmp_facts
##### ANSIBLE VERSION
```
ansible 2.3.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
inventory = ./hosts
gathering = explicit
roles_path = /home/actionmystique/Program-Files/Ubuntu/Ansible/git-Ansible/roles
private_role_vars = yes
log_path = /var/log/ansible.log
fact_caching = redis
fact_caching_timeout = 86400
retry_files_enabled = False
##### OS / ENVIRONMENT
- host: Ubuntu 16.04 4.4.0
- target: **all network devices**
##### SUMMARY
snmp_facts does not offer a provider parameter, unlike many other networking modules from Cisco, Arista, Juniper, and so on.... This feature would be very useful, allowing us to hand over only one structured variable which can be reused in other roles.
This feature should be extended to all network, system, ... modules.
##### STEPS TO REPRODUCE
Structure which could be handed over as "**provider**": snmp.ssh
``````
snmp
...
ssh:
transport: cli
host: "{{ ansible_host }}"
# Username is part of the following required group in v3
group_name: admin
# Username required if version is v3
username: admin
# New SNMP Version to use
version: v3
# community string, required if version is v1/v2c - up to 32 alphanumeric characters long. Spaces are not permitted.
community: xxxxxxxx
# Authentication level, required if version is v3
level: authPriv
# Authentication key, required if version is v3
authkey: xxxxxxxx
# Type of password
authkey_type:
# Hashing algoritm, required if version is v3
integrity: sha
# Encryption algoritm, required if level is authPriv
privacy: aes
# Encryption key, required if version is authPriv
privkey: xxxxxxxx```
``````
| main | add a provider parameter to snmp facts module issue type feature idea component name snmp facts ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration inventory hosts gathering explicit roles path home actionmystique program files ubuntu ansible git ansible roles private role vars yes log path var log ansible log fact caching redis fact caching timeout retry files enabled false os environment host ubuntu target all network devices summary snmp facts does not offer a provider parameter unlike many other networking modules from cisco arista juniper and so on this feature would be very useful allowing us to hand over only one structured variable which can be reused in other roles this feature should be extended to all network system modules steps to reproduce structure which could be handed over as provider snmp ssh snmp ssh transport cli host ansible host username is part of the following required group in group name admin username required if version is username admin new snmp version to use version community string required if version is up to alphanumeric characters long spaces are not permitted community xxxxxxxx authentication level required if version is level authpriv authentication key required if version is authkey xxxxxxxx type of password authkey type hashing algoritm required if version is integrity sha encryption algoritm required if level is authpriv privacy aes encryption key required if version is authpriv privkey xxxxxxxx | 1 |
5,174 | 26,345,476,548 | IssuesEvent | 2023-01-10 21:37:47 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | sam local invoke errors when 3-rd party layer is used: Invalid Layer name | type/bug area/local/invoke maintainer/need-followup | ### Description
If I add some external Lambda Layer (e.g. for monitoring and error reporting) then `sam local invoke` stops working.
### Steps to reproduce
1. Create sample applicatiion with `sam init`, choose Node.js 12.x runtime and SQS example.
2. Add 3-rd party layer into function definition in `template.yml`:
```yml
Layers:
- !Sub "arn:${AWS::Partition}:lambda:${AWS::Region}:464622532012:layer:Datadog-Node12-x:36"
```
3. Build and invoke function as usual
```sh
sam build
sam local invoke --event events/event-sqs.json
```
### Observed result
```
An error occurred (InvalidParameterValueException) when calling the GetLayerVersion operation: Invalid Layer name: arn:aws:lambda:us-east-1:464622532012:layer:Datadog-Node12-x
```
<details>
<summary>Click to expand output of `sam local invoke --event events/event-sqs.json --debug`</summary>
```
2020-11-02 22:46:12,152 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2020-11-02 22:46:12,274 | local invoke command is called
2020-11-02 22:46:12,280 | No Parameters detected in the template
2020-11-02 22:46:12,317 | 2 resources found in the template
2020-11-02 22:46:12,317 | Found Serverless function with name='SQSPayloadLogger' and CodeUri='SQSPayloadLogger'
2020-11-02 22:46:12,321 | Found one Lambda function with name 'SQSPayloadLogger'
2020-11-02 22:46:12,321 | Invoking src/handlers/sqs-payload-logger.sqsPayloadLoggerHandler (nodejs12.x)
2020-11-02 22:46:12,321 | No environment variables found for function 'SQSPayloadLogger'
2020-11-02 22:46:12,321 | Environment variables overrides data is standard format
2020-11-02 22:46:12,321 | Loading AWS credentials from session with profile 'None'
2020-11-02 22:46:12,338 | Resolving code path. Cwd=/home/envek/evl.ms/2uinc/sam-app/.aws-sam/build, CodeUri=SQSPayloadLogger
2020-11-02 22:46:12,338 | Resolved absolute path to code is /home/envek/evl.ms/2uinc/sam-app/.aws-sam/build/SQSPayloadLogger
2020-11-02 22:46:12,338 | Code /home/envek/evl.ms/2uinc/sam-app/.aws-sam/build/SQSPayloadLogger is not a zip/jar file
2020-11-02 22:46:13,619 | Sending Telemetry: {'metrics': [{'commandRun': {'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam local invoke', 'duration': 1465, 'exitReason': 'InvalidParameterValueException', 'exitCode': 255, 'requestId': 'e5532896-87fa-41e9-930f-c172b5cab3d9', 'installationId': '33ae4d3b-0f37-4b53-9f05-3b6cbb85cda6', 'sessionId': '421463e4-057e-4a74-9dcb-82621aa7c8c7', 'executionEnvironment': 'CLI', 'pyversion': '3.8.6', 'samcliVersion': '1.7.0'}}]}
2020-11-02 22:46:14,773 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
Traceback (most recent call last):
File "/home/linuxbrew/.linuxbrew/bin/sam", line 33, in <module>
sys.exit(load_entry_point('aws-sam-cli==1.7.0', 'console_scripts', 'sam')())
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/decorators.py", line 73, in new_func
return ctx.invoke(f, obj, *args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/lib/telemetry/metrics.py", line 148, in wrapped
raise exception # pylint: disable=raising-bad-type
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/lib/telemetry/metrics.py", line 114, in wrapped
return_value = func(*args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/commands/local/invoke/cli.py", line 72, in cli
do_cli(
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/commands/local/invoke/cli.py", line 152, in do_cli
context.local_lambda_runner.invoke(
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/commands/local/lib/local_lambda.py", line 103, in invoke
self.local_runtime.invoke(config, event, debug_context=self.debug_context, stdout=stdout, stderr=stderr)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/lambdafn/runtime.py", line 70, in invoke
container = LambdaContainer(
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/docker/lambda_container.py", line 73, in __init__
image = LambdaContainer._get_image(image_builder, runtime, layers, debug_options)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/docker/lambda_container.py", line 179, in _get_image
return image_builder.build(runtime, layers, is_debug)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/docker/lambda_image.py", line 107, in build
downloaded_layers = self.layer_downloader.download_all(layers, self.force_image_build)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/layers/layer_downloader.py", line 72, in download_all
layer_dirs.append(self.download(layer, force))
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/layers/layer_downloader.py", line 106, in download
layer_zip_uri = self._fetch_layer_uri(layer)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/layers/layer_downloader.py", line 155, in _fetch_layer_uri
raise e
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/layers/layer_downloader.py", line 136, in _fetch_layer_uri
layer_version_response = self.lambda_client.get_layer_version(
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 337, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 656, in _make_api_call
raise error_class(parsed_response, operation_name)
botocore.errorfactory.InvalidParameterValueException: An error occurred (InvalidParameterValueException) when calling the GetLayerVersion operation: Invalid Layer name: arn:aws:lambda:us-east-1:464622532012:layer:Datadog-Node12-x
```
</details>
### Expected result
Lambda function invoked with layer 3rd-party layer from correct region (in `~/.aws/config` was `us-west-2` was specified, but error complains about `us-east-1`).
### Additional environment details
1. OS: Ubuntu Linux 18.04
2. `sam --version`: `SAM CLI, version 1.7.0` installed from Linuxbrew (as recommended) | True | sam local invoke errors when 3-rd party layer is used: Invalid Layer name - ### Description
If I add some external Lambda Layer (e.g. for monitoring and error reporting) then `sam local invoke` stops working.
### Steps to reproduce
1. Create sample applicatiion with `sam init`, choose Node.js 12.x runtime and SQS example.
2. Add 3-rd party layer into function definition in `template.yml`:
```yml
Layers:
- !Sub "arn:${AWS::Partition}:lambda:${AWS::Region}:464622532012:layer:Datadog-Node12-x:36"
```
3. Build and invoke function as usual
```sh
sam build
sam local invoke --event events/event-sqs.json
```
### Observed result
```
An error occurred (InvalidParameterValueException) when calling the GetLayerVersion operation: Invalid Layer name: arn:aws:lambda:us-east-1:464622532012:layer:Datadog-Node12-x
```
<details>
<summary>Click to expand output of `sam local invoke --event events/event-sqs.json --debug`</summary>
```
2020-11-02 22:46:12,152 | Telemetry endpoint configured to be https://aws-serverless-tools-telemetry.us-west-2.amazonaws.com/metrics
2020-11-02 22:46:12,274 | local invoke command is called
2020-11-02 22:46:12,280 | No Parameters detected in the template
2020-11-02 22:46:12,317 | 2 resources found in the template
2020-11-02 22:46:12,317 | Found Serverless function with name='SQSPayloadLogger' and CodeUri='SQSPayloadLogger'
2020-11-02 22:46:12,321 | Found one Lambda function with name 'SQSPayloadLogger'
2020-11-02 22:46:12,321 | Invoking src/handlers/sqs-payload-logger.sqsPayloadLoggerHandler (nodejs12.x)
2020-11-02 22:46:12,321 | No environment variables found for function 'SQSPayloadLogger'
2020-11-02 22:46:12,321 | Environment variables overrides data is standard format
2020-11-02 22:46:12,321 | Loading AWS credentials from session with profile 'None'
2020-11-02 22:46:12,338 | Resolving code path. Cwd=/home/envek/evl.ms/2uinc/sam-app/.aws-sam/build, CodeUri=SQSPayloadLogger
2020-11-02 22:46:12,338 | Resolved absolute path to code is /home/envek/evl.ms/2uinc/sam-app/.aws-sam/build/SQSPayloadLogger
2020-11-02 22:46:12,338 | Code /home/envek/evl.ms/2uinc/sam-app/.aws-sam/build/SQSPayloadLogger is not a zip/jar file
2020-11-02 22:46:13,619 | Sending Telemetry: {'metrics': [{'commandRun': {'awsProfileProvided': False, 'debugFlagProvided': True, 'region': '', 'commandName': 'sam local invoke', 'duration': 1465, 'exitReason': 'InvalidParameterValueException', 'exitCode': 255, 'requestId': 'e5532896-87fa-41e9-930f-c172b5cab3d9', 'installationId': '33ae4d3b-0f37-4b53-9f05-3b6cbb85cda6', 'sessionId': '421463e4-057e-4a74-9dcb-82621aa7c8c7', 'executionEnvironment': 'CLI', 'pyversion': '3.8.6', 'samcliVersion': '1.7.0'}}]}
2020-11-02 22:46:14,773 | HTTPSConnectionPool(host='aws-serverless-tools-telemetry.us-west-2.amazonaws.com', port=443): Read timed out. (read timeout=0.1)
Traceback (most recent call last):
File "/home/linuxbrew/.linuxbrew/bin/sam", line 33, in <module>
sys.exit(load_entry_point('aws-sam-cli==1.7.0', 'console_scripts', 'sam')())
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 829, in __call__
return self.main(*args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 782, in main
rv = self.invoke(ctx)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 1259, in invoke
return _process_result(sub_ctx.command.invoke(sub_ctx))
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 1066, in invoke
return ctx.invoke(self.callback, **ctx.params)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/decorators.py", line 73, in new_func
return ctx.invoke(f, obj, *args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/click/core.py", line 610, in invoke
return callback(*args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/lib/telemetry/metrics.py", line 148, in wrapped
raise exception # pylint: disable=raising-bad-type
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/lib/telemetry/metrics.py", line 114, in wrapped
return_value = func(*args, **kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/commands/local/invoke/cli.py", line 72, in cli
do_cli(
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/commands/local/invoke/cli.py", line 152, in do_cli
context.local_lambda_runner.invoke(
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/commands/local/lib/local_lambda.py", line 103, in invoke
self.local_runtime.invoke(config, event, debug_context=self.debug_context, stdout=stdout, stderr=stderr)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/lambdafn/runtime.py", line 70, in invoke
container = LambdaContainer(
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/docker/lambda_container.py", line 73, in __init__
image = LambdaContainer._get_image(image_builder, runtime, layers, debug_options)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/docker/lambda_container.py", line 179, in _get_image
return image_builder.build(runtime, layers, is_debug)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/docker/lambda_image.py", line 107, in build
downloaded_layers = self.layer_downloader.download_all(layers, self.force_image_build)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/layers/layer_downloader.py", line 72, in download_all
layer_dirs.append(self.download(layer, force))
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/layers/layer_downloader.py", line 106, in download
layer_zip_uri = self._fetch_layer_uri(layer)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/layers/layer_downloader.py", line 155, in _fetch_layer_uri
raise e
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/samcli/local/layers/layer_downloader.py", line 136, in _fetch_layer_uri
layer_version_response = self.lambda_client.get_layer_version(
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 337, in _api_call
return self._make_api_call(operation_name, kwargs)
File "/home/linuxbrew/.linuxbrew/Cellar/aws-sam-cli/1.7.0/libexec/lib/python3.8/site-packages/botocore/client.py", line 656, in _make_api_call
raise error_class(parsed_response, operation_name)
botocore.errorfactory.InvalidParameterValueException: An error occurred (InvalidParameterValueException) when calling the GetLayerVersion operation: Invalid Layer name: arn:aws:lambda:us-east-1:464622532012:layer:Datadog-Node12-x
```
</details>
### Expected result
Lambda function invoked with layer 3rd-party layer from correct region (in `~/.aws/config` was `us-west-2` was specified, but error complains about `us-east-1`).
### Additional environment details
1. OS: Ubuntu Linux 18.04
2. `sam --version`: `SAM CLI, version 1.7.0` installed from Linuxbrew (as recommended) | main | sam local invoke errors when rd party layer is used invalid layer name description if i add some external lambda layer e g for monitoring and error reporting then sam local invoke stops working steps to reproduce create sample applicatiion with sam init choose node js x runtime and sqs example add rd party layer into function definition in template yml yml layers sub arn aws partition lambda aws region layer datadog x build and invoke function as usual sh sam build sam local invoke event events event sqs json observed result an error occurred invalidparametervalueexception when calling the getlayerversion operation invalid layer name arn aws lambda us east layer datadog x click to expand output of sam local invoke event events event sqs json debug telemetry endpoint configured to be local invoke command is called no parameters detected in the template resources found in the template found serverless function with name sqspayloadlogger and codeuri sqspayloadlogger found one lambda function with name sqspayloadlogger invoking src handlers sqs payload logger sqspayloadloggerhandler x no environment variables found for function sqspayloadlogger environment variables overrides data is standard format loading aws credentials from session with profile none resolving code path cwd home envek evl ms sam app aws sam build codeuri sqspayloadlogger resolved absolute path to code is home envek evl ms sam app aws sam build sqspayloadlogger code home envek evl ms sam app aws sam build sqspayloadlogger is not a zip jar file sending telemetry metrics httpsconnectionpool host aws serverless tools telemetry us west amazonaws com port read timed out read timeout traceback most recent call last file home linuxbrew linuxbrew bin sam line in sys exit load entry point aws sam cli console scripts sam file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages click core py line in call return self main args kwargs file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages click core py line in main rv self invoke ctx file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages click core py line in invoke return process result sub ctx command invoke sub ctx file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages click core py line in invoke return ctx invoke self callback ctx params file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages click core py line in invoke return callback args kwargs file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages click decorators py line in new func return ctx invoke f obj args kwargs file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages click core py line in invoke return callback args kwargs file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli lib telemetry metrics py line in wrapped raise exception pylint disable raising bad type file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli lib telemetry metrics py line in wrapped return value func args kwargs file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli commands local invoke cli py line in cli do cli file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli commands local invoke cli py line in do cli context local lambda runner invoke file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli commands local lib local lambda py line in invoke self local runtime invoke config event debug context self debug context stdout stdout stderr stderr file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local lambdafn runtime py line in invoke container lambdacontainer file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local docker lambda container py line in init image lambdacontainer get image image builder runtime layers debug options file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local docker lambda container py line in get image return image builder build runtime layers is debug file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local docker lambda image py line in build downloaded layers self layer downloader download all layers self force image build file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local layers layer downloader py line in download all layer dirs append self download layer force file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local layers layer downloader py line in download layer zip uri self fetch layer uri layer file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local layers layer downloader py line in fetch layer uri raise e file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages samcli local layers layer downloader py line in fetch layer uri layer version response self lambda client get layer version file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages botocore client py line in api call return self make api call operation name kwargs file home linuxbrew linuxbrew cellar aws sam cli libexec lib site packages botocore client py line in make api call raise error class parsed response operation name botocore errorfactory invalidparametervalueexception an error occurred invalidparametervalueexception when calling the getlayerversion operation invalid layer name arn aws lambda us east layer datadog x expected result lambda function invoked with layer party layer from correct region in aws config was us west was specified but error complains about us east additional environment details os ubuntu linux sam version sam cli version installed from linuxbrew as recommended | 1 |
10,817 | 2,967,676,071 | IssuesEvent | 2015-07-13 02:47:15 | ohnno/Brosa | https://api.github.com/repos/ohnno/Brosa | opened | Scroll To Top Behavior Tweak | design fix / tweak | Location: Global
I've seen the scroll to top button that appears after you scroll a certain distance clash with other elements on the screen. Can we move the trigger for it appearing further down the screen. I'm guessing (400-800px downward) would be about right. | 1.0 | Scroll To Top Behavior Tweak - Location: Global
I've seen the scroll to top button that appears after you scroll a certain distance clash with other elements on the screen. Can we move the trigger for it appearing further down the screen. I'm guessing (400-800px downward) would be about right. | non_main | scroll to top behavior tweak location global i ve seen the scroll to top button that appears after you scroll a certain distance clash with other elements on the screen can we move the trigger for it appearing further down the screen i m guessing downward would be about right | 0 |
127,012 | 5,011,558,975 | IssuesEvent | 2016-12-13 08:21:56 | AnotherCodeArtist/medien-transparenz.at | https://api.github.com/repos/AnotherCodeArtist/medien-transparenz.at | closed | Aggregate Organizations/Media to Groups | feature request high priority needs refinement | - It should be possible to combine different organizations ("Stadt Wien", "Wiener Linien", ...) to groups. As a result such a group should be rendered as if it was single organization. The same thing should also work for media
- Every single user (which does not require registration or login) should be able to create such groups (they could be store in the browser local storage so that are at least partially persistent).
- If a member (=registered user) creates such groups, it should be possible to share them with other members or the general public (might require some staging mechanism).
- Generally these groups should be available on all pages | 1.0 | Aggregate Organizations/Media to Groups - - It should be possible to combine different organizations ("Stadt Wien", "Wiener Linien", ...) to groups. As a result such a group should be rendered as if it was single organization. The same thing should also work for media
- Every single user (which does not require registration or login) should be able to create such groups (they could be store in the browser local storage so that are at least partially persistent).
- If a member (=registered user) creates such groups, it should be possible to share them with other members or the general public (might require some staging mechanism).
- Generally these groups should be available on all pages | non_main | aggregate organizations media to groups it should be possible to combine different organizations stadt wien wiener linien to groups as a result such a group should be rendered as if it was single organization the same thing should also work for media every single user which does not require registration or login should be able to create such groups they could be store in the browser local storage so that are at least partially persistent if a member registered user creates such groups it should be possible to share them with other members or the general public might require some staging mechanism generally these groups should be available on all pages | 0 |
4,924 | 25,316,239,216 | IssuesEvent | 2022-11-17 21:55:11 | ipfs/ipfs-gui | https://api.github.com/repos/ipfs/ipfs-gui | closed | IPFS Gui & Tools ownership | kind/discussion P0 need/analysis need/maintainers-input need/community-input kind/question Epic | I want to discuss defining and reducing the surface area of ownership of the GUI & Tools team.
Currently, there are a few areas where we define what the IPFS GUI / IPFS Gui & Tools / IPFS GUI Tools / @ipfs-gui team owns:
1. https://github.com/protocol/w3dt-stewards/blob/main/scripts/create-triage-links/src/repos.js
2. https://github.com/ipfs/ipfs-gui#all-projects
3. https://www.notion.so/pl-strflt/IPFS-GUI-3bc1c1bf54d74f928bf11ef59c876b74#3271edf4345e4a95b163066d5a9f5da6 (points to item 2 above)
### Questions
1. Can we drop support for any of the listed packages? Can any gui&tools packages/repos be archived/deprecated?
* We have already reduced triage-work to what is listed at https://github.com/protocol/w3dt-stewards/blob/main/scripts/create-triage-links/src/repos.js. Is that sufficient?
1. We know that ipfs-desktop, ipfs-webui, and ipfs-companion are our priorities, but what are our priorities beyond those three?
* What support priority do the main three have?
* What are the priorities of the other packages?
* What support do we need from @achingbrain , i.e. what packages from https://github.com/ipfs/js-ipfs/tree/master/packages do we need tier 1 support on?
* What are our other dependencies? What support do we need from other orgs/teams? (ipld, multiformats, etc..)
1. Should we keep a list of unmaintained, yet useful packages? If so, where?
1. Decide whether this repo (ipfs/ipfs-gui) should be the home of all IPFS GUI related efforts across teams, or the centralized repo for the IPFS GUI & Tools team
### Proposal
1. Support changes
1. Drop support for the following packages:
* ipfs-share-files - https://github.com/ipfs-shipyard/ipfs-share-files#maintainers is already asking for official maintainers
* ipfs/in-web-browsers - https://github.com/ipfs/in-web-browsers should be owned by browsers-WG team
* https://github.com/ipfs/kubo/tree/master/assets/dir-index-html - drop support, we're already not including those in triage efforts and I'm not familiar with it.
2. Debateable
* ipld/explore.ipld.io - https://github.com/ipld/explore.ipld.io should be owned by ipld, but we essentially have this entire page inside of webui/desktop via https://webui.ipfs.io/#/explore
* awesome-ipfs - there is already discussion about moving support for this repo into [ecosystems dashboard]. could drop official support if we automate more of the repo(https://ecosystem.ipfs.tech/) so PRs get automerged. Needs better stewards for curation than what IPFS GUI & Tools team can offer
1. I propose the following order of support priorities:
* ipfs-webui (consumed by kubo & desktop)
* ipfs-desktop (unique user sessions over lifetime: 33k Linux, 131k Windows, 39k macOS)
* ipfs-companion ([chrome store says 60k+ users](https://chrome.google.com/webstore/detail/ipfs-companion/nibjojkomfdiaoajekhjakgkdhaomnch))
* ipfs/ipld-explorer-components - consumed by explore.ipld.io & webui
* public-gateway-checker - somewhat useful and popular repo for checking status of ipfs gateways. Lots of opportunity here without a large burden
* ipfs-shipyard/pinning-service-compliance - useful for ensuring pinning providers are compliant, and helpful as a pre-req for adding pinning-service providers to webui pinning provider defaults
* ipld/explore.ipld.io
* ipfs-shipyard/i18n - documentation only, small burden, but required by desktop and webui or other gui&tools projects that need i18n.
* ipfs-shipyard/js-pinning-service-http-client - used only by ipfs-shipyard/pinning-service-compliance currently
* multiformats/cid-utils-website
* ipfs-shipyard/ipfs-css
* awesome-ipfs - lists a lot of ipfs related projects/tools/datasets,
1. I propose we keep a list of packages in this repo's README that include unmaintained & useful repos so they're not lost and can be taken up if the need arises.
1. I think it makes sense to keep this repo as the home for GUI projects, but we may want a similar ipfs/ipfs-tools repo for things that aren't necessarily GUI.
cc @BigLep @lidel @tinytb | True | IPFS Gui & Tools ownership - I want to discuss defining and reducing the surface area of ownership of the GUI & Tools team.
Currently, there are a few areas where we define what the IPFS GUI / IPFS Gui & Tools / IPFS GUI Tools / @ipfs-gui team owns:
1. https://github.com/protocol/w3dt-stewards/blob/main/scripts/create-triage-links/src/repos.js
2. https://github.com/ipfs/ipfs-gui#all-projects
3. https://www.notion.so/pl-strflt/IPFS-GUI-3bc1c1bf54d74f928bf11ef59c876b74#3271edf4345e4a95b163066d5a9f5da6 (points to item 2 above)
### Questions
1. Can we drop support for any of the listed packages? Can any gui&tools packages/repos be archived/deprecated?
* We have already reduced triage-work to what is listed at https://github.com/protocol/w3dt-stewards/blob/main/scripts/create-triage-links/src/repos.js. Is that sufficient?
1. We know that ipfs-desktop, ipfs-webui, and ipfs-companion are our priorities, but what are our priorities beyond those three?
* What support priority do the main three have?
* What are the priorities of the other packages?
* What support do we need from @achingbrain , i.e. what packages from https://github.com/ipfs/js-ipfs/tree/master/packages do we need tier 1 support on?
* What are our other dependencies? What support do we need from other orgs/teams? (ipld, multiformats, etc..)
1. Should we keep a list of unmaintained, yet useful packages? If so, where?
1. Decide whether this repo (ipfs/ipfs-gui) should be the home of all IPFS GUI related efforts across teams, or the centralized repo for the IPFS GUI & Tools team
### Proposal
1. Support changes
1. Drop support for the following packages:
* ipfs-share-files - https://github.com/ipfs-shipyard/ipfs-share-files#maintainers is already asking for official maintainers
* ipfs/in-web-browsers - https://github.com/ipfs/in-web-browsers should be owned by browsers-WG team
* https://github.com/ipfs/kubo/tree/master/assets/dir-index-html - drop support, we're already not including those in triage efforts and I'm not familiar with it.
2. Debateable
* ipld/explore.ipld.io - https://github.com/ipld/explore.ipld.io should be owned by ipld, but we essentially have this entire page inside of webui/desktop via https://webui.ipfs.io/#/explore
* awesome-ipfs - there is already discussion about moving support for this repo into [ecosystems dashboard]. could drop official support if we automate more of the repo(https://ecosystem.ipfs.tech/) so PRs get automerged. Needs better stewards for curation than what IPFS GUI & Tools team can offer
1. I propose the following order of support priorities:
* ipfs-webui (consumed by kubo & desktop)
* ipfs-desktop (unique user sessions over lifetime: 33k Linux, 131k Windows, 39k macOS)
* ipfs-companion ([chrome store says 60k+ users](https://chrome.google.com/webstore/detail/ipfs-companion/nibjojkomfdiaoajekhjakgkdhaomnch))
* ipfs/ipld-explorer-components - consumed by explore.ipld.io & webui
* public-gateway-checker - somewhat useful and popular repo for checking status of ipfs gateways. Lots of opportunity here without a large burden
* ipfs-shipyard/pinning-service-compliance - useful for ensuring pinning providers are compliant, and helpful as a pre-req for adding pinning-service providers to webui pinning provider defaults
* ipld/explore.ipld.io
* ipfs-shipyard/i18n - documentation only, small burden, but required by desktop and webui or other gui&tools projects that need i18n.
* ipfs-shipyard/js-pinning-service-http-client - used only by ipfs-shipyard/pinning-service-compliance currently
* multiformats/cid-utils-website
* ipfs-shipyard/ipfs-css
* awesome-ipfs - lists a lot of ipfs related projects/tools/datasets,
1. I propose we keep a list of packages in this repo's README that include unmaintained & useful repos so they're not lost and can be taken up if the need arises.
1. I think it makes sense to keep this repo as the home for GUI projects, but we may want a similar ipfs/ipfs-tools repo for things that aren't necessarily GUI.
cc @BigLep @lidel @tinytb | main | ipfs gui tools ownership i want to discuss defining and reducing the surface area of ownership of the gui tools team currently there are a few areas where we define what the ipfs gui ipfs gui tools ipfs gui tools ipfs gui team owns points to item above questions can we drop support for any of the listed packages can any gui tools packages repos be archived deprecated we have already reduced triage work to what is listed at is that sufficient we know that ipfs desktop ipfs webui and ipfs companion are our priorities but what are our priorities beyond those three what support priority do the main three have what are the priorities of the other packages what support do we need from achingbrain i e what packages from do we need tier support on what are our other dependencies what support do we need from other orgs teams ipld multiformats etc should we keep a list of unmaintained yet useful packages if so where decide whether this repo ipfs ipfs gui should be the home of all ipfs gui related efforts across teams or the centralized repo for the ipfs gui tools team proposal support changes drop support for the following packages ipfs share files is already asking for official maintainers ipfs in web browsers should be owned by browsers wg team drop support we re already not including those in triage efforts and i m not familiar with it debateable ipld explore ipld io should be owned by ipld but we essentially have this entire page inside of webui desktop via awesome ipfs there is already discussion about moving support for this repo into could drop official support if we automate more of the repo so prs get automerged needs better stewards for curation than what ipfs gui tools team can offer i propose the following order of support priorities ipfs webui consumed by kubo desktop ipfs desktop unique user sessions over lifetime linux windows macos ipfs companion ipfs ipld explorer components consumed by explore ipld io webui public gateway checker somewhat useful and popular repo for checking status of ipfs gateways lots of opportunity here without a large burden ipfs shipyard pinning service compliance useful for ensuring pinning providers are compliant and helpful as a pre req for adding pinning service providers to webui pinning provider defaults ipld explore ipld io ipfs shipyard documentation only small burden but required by desktop and webui or other gui tools projects that need ipfs shipyard js pinning service http client used only by ipfs shipyard pinning service compliance currently multiformats cid utils website ipfs shipyard ipfs css awesome ipfs lists a lot of ipfs related projects tools datasets i propose we keep a list of packages in this repo s readme that include unmaintained useful repos so they re not lost and can be taken up if the need arises i think it makes sense to keep this repo as the home for gui projects but we may want a similar ipfs ipfs tools repo for things that aren t necessarily gui cc biglep lidel tinytb | 1 |
4,050 | 18,921,712,339 | IssuesEvent | 2021-11-17 03:01:48 | carbon-design-system/carbon | https://api.github.com/repos/carbon-design-system/carbon | closed | [Bug]: Not getting the selected value from row in datatable in React | type: question ❓ status: needs triage 🕵️♀️ status: waiting for maintainer response 💬 | ### Package
carbon-components-react
### Browser
Chrome
### Package version
7.25.0
### Description
Am just exploring react with carbon and got stuck with some trigger event. Here is my piece of code (ChildComponent)
```js
<DataTable rows={rows} headers={headers} size='tall' zebra>
{({ rows, headers, getTableProps, getHeaderProps, getRowProps }) => (
<Table {...getTableProps()}>
<TableHead>
<TableRow>
{headers.map((header, i) => (
<TableHeader key={i} {...getHeaderProps({ header })}>
<React.Fragment>{header.header}</React.Fragment>
</TableHeader>
))}
</TableRow>
</TableHead>
<TableBody>
{rows.map((row, j) => (
<div key={row.id}>
<TableRow key={row.package} {...getRowProps({ row })}>
<>
{row.cells.map((cell) => (
<>
<TableCell key={cell.id}>{cell.value}</TableCell>
</>
))}
<Button color='primary' id={row.package} onClick={this.props.redirectToDesigner(row.package)}>
Open
</Button>
</>
</TableRow>
<br />
</div>
))}
</TableBody>
</Table>
)}
</DataTable>
```
This lies on child component and Parent component looks like below
```js
import React from "react";
import ChildComponent from "./ChildComponent";
class ParentComponent extends React.Component {
redirectToDesigner = (id) => {
alert(1);
console.log("TRYYYY", id);
};
render() {
return <ChildComponent redirectToDesigner={this.redirectToDesigner} />;
}
}
export default ParentComponent;
```
Am getting the alert and the console as undefined for some reason
Here you can see one onclick event , its triggering but am not getting value(row.id) from it. I need the value for further development.
### CodeSandbox example
https://codesandbox.io/s/gifted-shirley-t8zcc?file=/src/index.js:63-240
### Steps to reproduce
Click on open button from any row ans its not poping up with name , coming as undefined
```js
<DataTable rows={rows} headers={headers} size='tall' zebra>
{({ rows, headers, getTableProps, getHeaderProps, getRowProps }) => (
<Table {...getTableProps()}>
<TableHead>
<TableRow>
{headers.map((header, i) => (
<TableHeader key={i} {...getHeaderProps({ header })}>
<React.Fragment>{header.header}</React.Fragment>
</TableHeader>
))}
</TableRow>
</TableHead>
<TableBody>
{rows.map((row, j) => (
<div key={row.id}>
<TableRow key={row.package} {...getRowProps({ row })}>
<>
{row.cells.map((cell) => (
<>
<TableCell key={cell.id}>{cell.value}</TableCell>
</>
))}
<Button color='primary' id={row.package} onClick={this.props.redirectToDesigner(row.package)}>
Open
</Button>
</>
</TableRow>
<br />
</div>
))}
</TableBody>
</Table>
)}
</DataTable>
```
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | True | [Bug]: Not getting the selected value from row in datatable in React - ### Package
carbon-components-react
### Browser
Chrome
### Package version
7.25.0
### Description
Am just exploring react with carbon and got stuck with some trigger event. Here is my piece of code (ChildComponent)
```js
<DataTable rows={rows} headers={headers} size='tall' zebra>
{({ rows, headers, getTableProps, getHeaderProps, getRowProps }) => (
<Table {...getTableProps()}>
<TableHead>
<TableRow>
{headers.map((header, i) => (
<TableHeader key={i} {...getHeaderProps({ header })}>
<React.Fragment>{header.header}</React.Fragment>
</TableHeader>
))}
</TableRow>
</TableHead>
<TableBody>
{rows.map((row, j) => (
<div key={row.id}>
<TableRow key={row.package} {...getRowProps({ row })}>
<>
{row.cells.map((cell) => (
<>
<TableCell key={cell.id}>{cell.value}</TableCell>
</>
))}
<Button color='primary' id={row.package} onClick={this.props.redirectToDesigner(row.package)}>
Open
</Button>
</>
</TableRow>
<br />
</div>
))}
</TableBody>
</Table>
)}
</DataTable>
```
This lies on child component and Parent component looks like below
```js
import React from "react";
import ChildComponent from "./ChildComponent";
class ParentComponent extends React.Component {
redirectToDesigner = (id) => {
alert(1);
console.log("TRYYYY", id);
};
render() {
return <ChildComponent redirectToDesigner={this.redirectToDesigner} />;
}
}
export default ParentComponent;
```
Am getting the alert and the console as undefined for some reason
Here you can see one onclick event , its triggering but am not getting value(row.id) from it. I need the value for further development.
### CodeSandbox example
https://codesandbox.io/s/gifted-shirley-t8zcc?file=/src/index.js:63-240
### Steps to reproduce
Click on open button from any row ans its not poping up with name , coming as undefined
```js
<DataTable rows={rows} headers={headers} size='tall' zebra>
{({ rows, headers, getTableProps, getHeaderProps, getRowProps }) => (
<Table {...getTableProps()}>
<TableHead>
<TableRow>
{headers.map((header, i) => (
<TableHeader key={i} {...getHeaderProps({ header })}>
<React.Fragment>{header.header}</React.Fragment>
</TableHeader>
))}
</TableRow>
</TableHead>
<TableBody>
{rows.map((row, j) => (
<div key={row.id}>
<TableRow key={row.package} {...getRowProps({ row })}>
<>
{row.cells.map((cell) => (
<>
<TableCell key={cell.id}>{cell.value}</TableCell>
</>
))}
<Button color='primary' id={row.package} onClick={this.props.redirectToDesigner(row.package)}>
Open
</Button>
</>
</TableRow>
<br />
</div>
))}
</TableBody>
</Table>
)}
</DataTable>
```
### Code of Conduct
- [X] I agree to follow this project's [Code of Conduct](https://github.com/carbon-design-system/carbon/blob/f555616971a03fd454c0f4daea184adf41fff05b/.github/CODE_OF_CONDUCT.md)
- [X] I checked the [current issues](https://github.com/carbon-design-system/carbon/issues) for duplicate problems | main | not getting the selected value from row in datatable in react package carbon components react browser chrome package version description am just exploring react with carbon and got stuck with some trigger event here is my piece of code childcomponent js rows headers gettableprops getheaderprops getrowprops headers map header i header header rows map row j row cells map cell cell value open this lies on child component and parent component looks like below js import react from react import childcomponent from childcomponent class parentcomponent extends react component redirecttodesigner id alert console log tryyyy id render return export default parentcomponent am getting the alert and the console as undefined for some reason here you can see one onclick event its triggering but am not getting value row id from it i need the value for further development codesandbox example steps to reproduce click on open button from any row ans its not poping up with name coming as undefined js rows headers gettableprops getheaderprops getrowprops headers map header i header header rows map row j row cells map cell cell value open code of conduct i agree to follow this project s i checked the for duplicate problems | 1 |
4,611 | 23,878,728,550 | IssuesEvent | 2022-09-07 21:57:12 | aws/aws-sam-cli | https://api.github.com/repos/aws/aws-sam-cli | closed | Bug: `sam build` continues even if Typsecript compilation fails | blocked/more-info-needed maintainer/need-followup | ### Description:
After adding an intentional TS error (e.g. renamed a variable without refactoring), and running `sam build`, I'm told "Build Succeeded". Running `tsc` directly shows the error correctly.
My directory structure is as follows:
```
<root>
|--> app
| |--> node_modules
| |--> src
| | |--> app.ts # and other TS files
| |--> package.json # and other config files
| |--> tsconfig.json
|--> node_modules # for building e.g. rimraf
|--> package.json # ^^
|--> samconfig.toml
|--> template.yaml
```
I run `sam build` from the root dir. In `template.yml` I have this (partial example):
```yaml
Function:
Type: AWS::Serverless::Function
# ...
Metadata:
BuildMethod: esbuild
BuildProperties:
Minify: false
Target: es2020
Sourcemap: true
EntryPoints:
- src/app.ts
```
To work around this, I have a wrapper shell script in which I call the `build` function to run `tsc` before `sam build` (partial example):
```shell
#!/bin/bash
set -e
function build() {
clean
(cd app; npx tsc) # `sam build` seems to continue even if TSC fails ... ?
sam build "$@"
validate
}
"$@"
```
So I start the build via `./scripts.sh build`.
### Steps to reproduce:
Create a simple Typescript / SAM project. Add a line of code which causes `tsc` to fail. Run `sam build` and observe that it _does not_ indicate failure.
### Observed result:
`sam build` does not report `tsc` errors / does not abort the build process on such errors.
### Expected result:
`sam build` should abort the process if `tsc` fails.
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: Manjaro Linux
2. `sam --version`: 1.55.0
3. AWS region: N/A | True | Bug: `sam build` continues even if Typsecript compilation fails - ### Description:
After adding an intentional TS error (e.g. renamed a variable without refactoring), and running `sam build`, I'm told "Build Succeeded". Running `tsc` directly shows the error correctly.
My directory structure is as follows:
```
<root>
|--> app
| |--> node_modules
| |--> src
| | |--> app.ts # and other TS files
| |--> package.json # and other config files
| |--> tsconfig.json
|--> node_modules # for building e.g. rimraf
|--> package.json # ^^
|--> samconfig.toml
|--> template.yaml
```
I run `sam build` from the root dir. In `template.yml` I have this (partial example):
```yaml
Function:
Type: AWS::Serverless::Function
# ...
Metadata:
BuildMethod: esbuild
BuildProperties:
Minify: false
Target: es2020
Sourcemap: true
EntryPoints:
- src/app.ts
```
To work around this, I have a wrapper shell script in which I call the `build` function to run `tsc` before `sam build` (partial example):
```shell
#!/bin/bash
set -e
function build() {
clean
(cd app; npx tsc) # `sam build` seems to continue even if TSC fails ... ?
sam build "$@"
validate
}
"$@"
```
So I start the build via `./scripts.sh build`.
### Steps to reproduce:
Create a simple Typescript / SAM project. Add a line of code which causes `tsc` to fail. Run `sam build` and observe that it _does not_ indicate failure.
### Observed result:
`sam build` does not report `tsc` errors / does not abort the build process on such errors.
### Expected result:
`sam build` should abort the process if `tsc` fails.
### Additional environment details (Ex: Windows, Mac, Amazon Linux etc)
1. OS: Manjaro Linux
2. `sam --version`: 1.55.0
3. AWS region: N/A | main | bug sam build continues even if typsecript compilation fails description after adding an intentional ts error e g renamed a variable without refactoring and running sam build i m told build succeeded running tsc directly shows the error correctly my directory structure is as follows app node modules src app ts and other ts files package json and other config files tsconfig json node modules for building e g rimraf package json samconfig toml template yaml i run sam build from the root dir in template yml i have this partial example yaml function type aws serverless function metadata buildmethod esbuild buildproperties minify false target sourcemap true entrypoints src app ts to work around this i have a wrapper shell script in which i call the build function to run tsc before sam build partial example shell bin bash set e function build clean cd app npx tsc sam build seems to continue even if tsc fails sam build validate so i start the build via scripts sh build steps to reproduce create a simple typescript sam project add a line of code which causes tsc to fail run sam build and observe that it does not indicate failure observed result sam build does not report tsc errors does not abort the build process on such errors expected result sam build should abort the process if tsc fails additional environment details ex windows mac amazon linux etc os manjaro linux sam version aws region n a | 1 |
62,766 | 15,357,406,766 | IssuesEvent | 2021-03-01 13:39:44 | scikit-learn/scikit-learn | https://api.github.com/repos/scikit-learn/scikit-learn | opened | Random failure in fetch_openml tests | Build / CI module:test-suite | It appended once on the CI of the unrelated PR #19558 when running `test_fetch_openml_iris[True]`:
```python-traceback
../1/s/sklearn/externals/_arff.py:911:
[...]
E sklearn.externals._arff.BadLayout: Invalid layout of the ARFF file, at line 0.
[...]
During handling of the above exception, another exception occurred:
[...]
FileNotFoundError: [Errno 2] No such file or directory: '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz
```
Full traceback:
<details>
```
2021-03-01T10:30:29.5661420Z [gw1] darwin -- Python 3.9.2 /usr/local/miniconda/envs/testvenv/bin/python
2021-03-01T10:30:29.5662060Z
2021-03-01T10:30:29.5663160Z args = ('data/v1/download/61', '/Users/runner/scikit_learn_data/openml')
2021-03-01T10:30:29.5664990Z kw = {'encode_nominal': True, 'md5_checksum': 'ad484452702105cbf3d30f8deaba39a9', 'parse_arff': <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>, 'return_type': 3}
2021-03-01T10:30:29.5666510Z local_path = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.5666930Z
2021-03-01T10:30:29.5667290Z @wraps(f)
2021-03-01T10:30:29.5667710Z def wrapper(*args, **kw):
2021-03-01T10:30:29.5668240Z if data_home is None:
2021-03-01T10:30:29.5668770Z return f(*args, **kw)
2021-03-01T10:30:29.5669240Z try:
2021-03-01T10:30:29.5669710Z > return f(*args, **kw)
2021-03-01T10:30:29.5669990Z
2021-03-01T10:30:29.5670890Z args = ('data/v1/download/61', '/Users/runner/scikit_learn_data/openml')
2021-03-01T10:30:29.5671930Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.5672580Z f = <function _load_arff_response at 0x1230f0430>
2021-03-01T10:30:29.5673960Z kw = {'encode_nominal': True, 'md5_checksum': 'ad484452702105cbf3d30f8deaba39a9', 'parse_arff': <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>, 'return_type': 3}
2021-03-01T10:30:29.5675430Z local_path = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.5676500Z openml_path = 'data/v1/download/61'
2021-03-01T10:30:29.5676850Z
2021-03-01T10:30:29.5677280Z ../1/s/sklearn/datasets/_openml.py:62:
2021-03-01T10:30:29.5677820Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5678840Z
2021-03-01T10:30:29.5679690Z url = 'data/v1/download/61'
2021-03-01T10:30:29.5680710Z data_home = '/Users/runner/scikit_learn_data/openml', return_type = 3
2021-03-01T10:30:29.5681370Z encode_nominal = True
2021-03-01T10:30:29.5682010Z parse_arff = <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>
2021-03-01T10:30:29.5683050Z md5_checksum = 'ad484452702105cbf3d30f8deaba39a9'
2021-03-01T10:30:29.5683450Z
2021-03-01T10:30:29.5683950Z def _load_arff_response(
2021-03-01T10:30:29.5684480Z url: str,
2021-03-01T10:30:29.5685020Z data_home: Optional[str],
2021-03-01T10:30:29.5685620Z return_type, encode_nominal: bool,
2021-03-01T10:30:29.5686390Z parse_arff: Callable[[ArffContainerType], Tuple],
2021-03-01T10:30:29.5687020Z md5_checksum: str
2021-03-01T10:30:29.5687860Z ) -> Tuple:
2021-03-01T10:30:29.5688520Z """Load arff data with url and parses arff response with parse_arff"""
2021-03-01T10:30:29.5689230Z response = _open_openml_url(url, data_home)
2021-03-01T10:30:29.5689670Z
2021-03-01T10:30:29.5690080Z with closing(response):
2021-03-01T10:30:29.5690680Z # Note that if the data is dense, no reading is done until the data
2021-03-01T10:30:29.5691850Z # generator is iterated.
2021-03-01T10:30:29.5692460Z actual_md5_checksum = hashlib.md5()
2021-03-01T10:30:29.5692990Z
2021-03-01T10:30:29.5693540Z def _stream_checksum_generator(response):
2021-03-01T10:30:29.5694180Z for line in response:
2021-03-01T10:30:29.5694810Z actual_md5_checksum.update(line)
2021-03-01T10:30:29.5695780Z yield line.decode('utf-8')
2021-03-01T10:30:29.5696390Z
2021-03-01T10:30:29.5696950Z stream = _stream_checksum_generator(response)
2021-03-01T10:30:29.5697500Z
2021-03-01T10:30:29.5698000Z > arff = _arff.load(stream,
2021-03-01T10:30:29.5698620Z return_type=return_type,
2021-03-01T10:30:29.5699250Z encode_nominal=encode_nominal)
2021-03-01T10:30:29.5699650Z
2021-03-01T10:30:29.5700270Z _stream_checksum_generator = <function _load_arff_response.<locals>._stream_checksum_generator at 0x12f4c2a60>
2021-03-01T10:30:29.5701030Z actual_md5_checksum = <md5 _hashlib.HASH object @ 0x12f55b230>
2021-03-01T10:30:29.5702040Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.5702640Z encode_nominal = True
2021-03-01T10:30:29.5703580Z md5_checksum = 'ad484452702105cbf3d30f8deaba39a9'
2021-03-01T10:30:29.5704300Z parse_arff = <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>
2021-03-01T10:30:29.5704960Z response = <gzip on 0x12f4decd0>
2021-03-01T10:30:29.5705490Z return_type = 3
2021-03-01T10:30:29.5706150Z stream = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5707180Z url = 'data/v1/download/61'
2021-03-01T10:30:29.5707570Z
2021-03-01T10:30:29.5708090Z ../1/s/sklearn/datasets/_openml.py:518:
2021-03-01T10:30:29.5708750Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5709180Z
2021-03-01T10:30:29.5709780Z fp = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5710480Z encode_nominal = True, return_type = 3
2021-03-01T10:30:29.5710740Z
2021-03-01T10:30:29.5711190Z def load(fp, encode_nominal=False, return_type=DENSE):
2021-03-01T10:30:29.5712280Z '''Load a file-like object containing the ARFF document and convert it into
2021-03-01T10:30:29.5712980Z a Python object.
2021-03-01T10:30:29.5713460Z
2021-03-01T10:30:29.5714330Z :param fp: a file-like object.
2021-03-01T10:30:29.5715040Z :param encode_nominal: boolean, if True perform a label encoding
2021-03-01T10:30:29.5715770Z while reading the .arff file.
2021-03-01T10:30:29.5716880Z :param return_type: determines the data structure used to store the
2021-03-01T10:30:29.5717610Z dataset. Can be one of `arff.DENSE`, `arff.COO`, `arff.LOD`,
2021-03-01T10:30:29.5718280Z `arff.DENSE_GEN` or `arff.LOD_GEN`.
2021-03-01T10:30:29.5718930Z Consult the sections on `working with sparse data`_ and `loading
2021-03-01T10:30:29.5719560Z progressively`_.
2021-03-01T10:30:29.5720120Z :return: a dictionary.
2021-03-01T10:30:29.5720980Z '''
2021-03-01T10:30:29.5721560Z decoder = ArffDecoder()
2021-03-01T10:30:29.5722170Z > return decoder.decode(fp, encode_nominal=encode_nominal,
2021-03-01T10:30:29.5722830Z return_type=return_type)
2021-03-01T10:30:29.5723220Z
2021-03-01T10:30:29.5723750Z decoder = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5724360Z encode_nominal = True
2021-03-01T10:30:29.5725020Z fp = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5725680Z return_type = 3
2021-03-01T10:30:29.5726010Z
2021-03-01T10:30:29.5726510Z ../1/s/sklearn/externals/_arff.py:1078:
2021-03-01T10:30:29.5727520Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5727970Z
2021-03-01T10:30:29.5728520Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5729230Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5729920Z encode_nominal = True, return_type = 3
2021-03-01T10:30:29.5730270Z
2021-03-01T10:30:29.5730840Z def decode(self, s, encode_nominal=False, return_type=DENSE):
2021-03-01T10:30:29.5731920Z '''Returns the Python representation of a given ARFF file.
2021-03-01T10:30:29.5732540Z
2021-03-01T10:30:29.5733130Z When a file object is passed as an argument, this method reads lines
2021-03-01T10:30:29.5733860Z iteratively, avoiding to load unnecessary information to the memory.
2021-03-01T10:30:29.5734440Z
2021-03-01T10:30:29.5735000Z :param s: a string or file object with the ARFF file.
2021-03-01T10:30:29.5735710Z :param encode_nominal: boolean, if True perform a label encoding
2021-03-01T10:30:29.5736380Z while reading the .arff file.
2021-03-01T10:30:29.5737040Z :param return_type: determines the data structure used to store the
2021-03-01T10:30:29.5737770Z dataset. Can be one of `arff.DENSE`, `arff.COO`, `arff.LOD`,
2021-03-01T10:30:29.5738420Z `arff.DENSE_GEN` or `arff.LOD_GEN`.
2021-03-01T10:30:29.5739080Z Consult the sections on `working with sparse data`_ and `loading
2021-03-01T10:30:29.5739700Z progressively`_.
2021-03-01T10:30:29.5740550Z '''
2021-03-01T10:30:29.5741070Z try:
2021-03-01T10:30:29.5741650Z return self._decode(s, encode_nominal=encode_nominal,
2021-03-01T10:30:29.5742320Z matrix_type=return_type)
2021-03-01T10:30:29.5742940Z except ArffException as e:
2021-03-01T10:30:29.5743530Z e.line = self._current_line
2021-03-01T10:30:29.5744070Z > raise e
2021-03-01T10:30:29.5744380Z
2021-03-01T10:30:29.5744840Z encode_nominal = True
2021-03-01T10:30:29.5745350Z return_type = 3
2021-03-01T10:30:29.5745990Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5746750Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5747170Z
2021-03-01T10:30:29.5747670Z ../1/s/sklearn/externals/_arff.py:915:
2021-03-01T10:30:29.5748820Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5749320Z
2021-03-01T10:30:29.5759090Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5759880Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5760860Z encode_nominal = True, return_type = 3
2021-03-01T10:30:29.5761220Z
2021-03-01T10:30:29.5761810Z def decode(self, s, encode_nominal=False, return_type=DENSE):
2021-03-01T10:30:29.5763070Z '''Returns the Python representation of a given ARFF file.
2021-03-01T10:30:29.5763700Z
2021-03-01T10:30:29.5764290Z When a file object is passed as an argument, this method reads lines
2021-03-01T10:30:29.5765020Z iteratively, avoiding to load unnecessary information to the memory.
2021-03-01T10:30:29.5765650Z
2021-03-01T10:30:29.5766240Z :param s: a string or file object with the ARFF file.
2021-03-01T10:30:29.5766950Z :param encode_nominal: boolean, if True perform a label encoding
2021-03-01T10:30:29.5767610Z while reading the .arff file.
2021-03-01T10:30:29.5768350Z :param return_type: determines the data structure used to store the
2021-03-01T10:30:29.5769090Z dataset. Can be one of `arff.DENSE`, `arff.COO`, `arff.LOD`,
2021-03-01T10:30:29.5769750Z `arff.DENSE_GEN` or `arff.LOD_GEN`.
2021-03-01T10:30:29.5770400Z Consult the sections on `working with sparse data`_ and `loading
2021-03-01T10:30:29.5771440Z progressively`_.
2021-03-01T10:30:29.5772330Z '''
2021-03-01T10:30:29.5772870Z try:
2021-03-01T10:30:29.5773460Z > return self._decode(s, encode_nominal=encode_nominal,
2021-03-01T10:30:29.5774130Z matrix_type=return_type)
2021-03-01T10:30:29.5774510Z
2021-03-01T10:30:29.5774970Z encode_nominal = True
2021-03-01T10:30:29.5775470Z return_type = 3
2021-03-01T10:30:29.5776180Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5776940Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5777360Z
2021-03-01T10:30:29.5777860Z ../1/s/sklearn/externals/_arff.py:911:
2021-03-01T10:30:29.5778580Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5779030Z
2021-03-01T10:30:29.5779580Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5780320Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5781000Z encode_nominal = True, matrix_type = 3
2021-03-01T10:30:29.5781350Z
2021-03-01T10:30:29.5781920Z def _decode(self, s, encode_nominal=False, matrix_type=DENSE):
2021-03-01T10:30:29.5782950Z '''Do the job the ``encode``.'''
2021-03-01T10:30:29.5783510Z
2021-03-01T10:30:29.5784050Z # Make sure this method is idempotent
2021-03-01T10:30:29.5784620Z self._current_line = 0
2021-03-01T10:30:29.5785130Z
2021-03-01T10:30:29.5785660Z # If string, convert to a list of lines
2021-03-01T10:30:29.5786290Z if isinstance(s, basestring):
2021-03-01T10:30:29.5787290Z s = s.strip('\r\n ').replace('\r\n', '\n').split('\n')
2021-03-01T10:30:29.5787910Z
2021-03-01T10:30:29.5788420Z # Create the return object
2021-03-01T10:30:29.5789170Z obj: ArffContainerType = {
2021-03-01T10:30:29.5790120Z u'description': u'',
2021-03-01T10:30:29.5791070Z u'relation': u'',
2021-03-01T10:30:29.5792000Z u'attributes': [],
2021-03-01T10:30:29.5792930Z u'data': []
2021-03-01T10:30:29.5793490Z }
2021-03-01T10:30:29.5794000Z attribute_names = {}
2021-03-01T10:30:29.5794500Z
2021-03-01T10:30:29.5795010Z # Create the data helper object
2021-03-01T10:30:29.5795610Z data = _get_data_object_for_decoding(matrix_type)
2021-03-01T10:30:29.5796180Z
2021-03-01T10:30:29.5796670Z # Read all lines
2021-03-01T10:30:29.5797210Z STATE = _TK_DESCRIPTION
2021-03-01T10:30:29.5797740Z s = iter(s)
2021-03-01T10:30:29.5798260Z for row in s:
2021-03-01T10:30:29.5799010Z self._current_line += 1
2021-03-01T10:30:29.5799580Z # Ignore empty lines
2021-03-01T10:30:29.5800500Z row = row.strip(' \r\n')
2021-03-01T10:30:29.5801130Z if not row: continue
2021-03-01T10:30:29.5801650Z
2021-03-01T10:30:29.5802130Z u_row = row.upper()
2021-03-01T10:30:29.5802640Z
2021-03-01T10:30:29.5803600Z # DESCRIPTION -----------------------------------------------------
2021-03-01T10:30:29.5804410Z if u_row.startswith(_TK_DESCRIPTION) and STATE == _TK_DESCRIPTION:
2021-03-01T10:30:29.5805540Z obj['description'] += self._decode_comment(row) + '\n'
2021-03-01T10:30:29.5806670Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5807310Z
2021-03-01T10:30:29.5808260Z # RELATION --------------------------------------------------------
2021-03-01T10:30:29.5809010Z elif u_row.startswith(_TK_RELATION):
2021-03-01T10:30:29.5809660Z if STATE != _TK_DESCRIPTION:
2021-03-01T10:30:29.5810280Z raise BadLayout()
2021-03-01T10:30:29.5810790Z
2021-03-01T10:30:29.5811300Z STATE = _TK_RELATION
2021-03-01T10:30:29.5812870Z obj['relation'] = self._decode_relation(row)
2021-03-01T10:30:29.5813990Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5814690Z
2021-03-01T10:30:29.5815760Z # ATTRIBUTE -------------------------------------------------------
2021-03-01T10:30:29.5816510Z elif u_row.startswith(_TK_ATTRIBUTE):
2021-03-01T10:30:29.5817220Z if STATE != _TK_RELATION and STATE != _TK_ATTRIBUTE:
2021-03-01T10:30:29.5817880Z raise BadLayout()
2021-03-01T10:30:29.5818380Z
2021-03-01T10:30:29.5818890Z STATE = _TK_ATTRIBUTE
2021-03-01T10:30:29.5819400Z
2021-03-01T10:30:29.5819930Z attr = self._decode_attribute(row)
2021-03-01T10:30:29.5820580Z if attr[0] in attribute_names:
2021-03-01T10:30:29.5821260Z raise BadAttributeName(attr[0], attribute_names[attr[0]])
2021-03-01T10:30:29.5821890Z else:
2021-03-01T10:30:29.5822490Z attribute_names[attr[0]] = self._current_line
2021-03-01T10:30:29.5823490Z obj['attributes'].append(attr)
2021-03-01T10:30:29.5824070Z
2021-03-01T10:30:29.5824870Z if isinstance(attr[1], (list, tuple)):
2021-03-01T10:30:29.5825590Z if encode_nominal:
2021-03-01T10:30:29.5826250Z conversor = EncodedNominalConversor(attr[1])
2021-03-01T10:30:29.5826860Z else:
2021-03-01T10:30:29.5827480Z conversor = NominalConversor(attr[1])
2021-03-01T10:30:29.5828080Z else:
2021-03-01T10:30:29.5829070Z CONVERSOR_MAP = {'STRING': unicode,
2021-03-01T10:30:29.5830200Z 'INTEGER': lambda x: int(float(x)),
2021-03-01T10:30:29.5831320Z 'NUMERIC': float,
2021-03-01T10:30:29.5832380Z 'REAL': float}
2021-03-01T10:30:29.5833060Z conversor = CONVERSOR_MAP[attr[1]]
2021-03-01T10:30:29.5833630Z
2021-03-01T10:30:29.5834170Z self._conversors.append(conversor)
2021-03-01T10:30:29.5835210Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5835860Z
2021-03-01T10:30:29.5836830Z # DATA ------------------------------------------------------------
2021-03-01T10:30:29.5837550Z elif u_row.startswith(_TK_DATA):
2021-03-01T10:30:29.5838190Z if STATE != _TK_ATTRIBUTE:
2021-03-01T10:30:29.5838790Z raise BadLayout()
2021-03-01T10:30:29.5839310Z
2021-03-01T10:30:29.5839790Z break
2021-03-01T10:30:29.5840780Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5841640Z
2021-03-01T10:30:29.5842600Z # COMMENT ---------------------------------------------------------
2021-03-01T10:30:29.5843340Z elif u_row.startswith(_TK_COMMENT):
2021-03-01T10:30:29.5843940Z pass
2021-03-01T10:30:29.5844930Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5845590Z else:
2021-03-01T10:30:29.5846130Z # Never found @DATA
2021-03-01T10:30:29.5846660Z > raise BadLayout()
2021-03-01T10:30:29.5847340Z E sklearn.externals._arff.BadLayout: Invalid layout of the ARFF file, at line 0.
2021-03-01T10:30:29.5847820Z
2021-03-01T10:30:29.5848620Z STATE = '%'
2021-03-01T10:30:29.5849150Z attribute_names = {}
2021-03-01T10:30:29.5849790Z data = <sklearn.externals._arff.DenseGeneratorData object at 0x12f4debe0>
2021-03-01T10:30:29.5850420Z encode_nominal = True
2021-03-01T10:30:29.5850930Z matrix_type = 3
2021-03-01T10:30:29.5851950Z obj = {'attributes': [], 'data': [], 'description': '', 'relation': ''}
2021-03-01T10:30:29.5852800Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5853930Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5854360Z
2021-03-01T10:30:29.5855010Z ../1/s/sklearn/externals/_arff.py:878: BadLayout
2021-03-01T10:30:29.5855390Z
2021-03-01T10:30:29.5856050Z During handling of the above exception, another exception occurred:
2021-03-01T10:30:29.5856480Z
2021-03-01T10:30:29.5857040Z monkeypatch = <_pytest.monkeypatch.MonkeyPatch object at 0x12f4de8e0>
2021-03-01T10:30:29.5857650Z gzip_response = True
2021-03-01T10:30:29.5858060Z
2021-03-01T10:30:29.5858990Z @pytest.mark.parametrize('gzip_response', [True, False])
2021-03-01T10:30:29.5859730Z def test_fetch_openml_iris(monkeypatch, gzip_response):
2021-03-01T10:30:29.5860420Z # classification dataset with numeric only columns
2021-03-01T10:30:29.5861010Z data_id = 61
2021-03-01T10:30:29.5861870Z data_name = 'iris'
2021-03-01T10:30:29.5862400Z
2021-03-01T10:30:29.5862990Z _monkey_patch_webbased_functions(monkeypatch, data_id, gzip_response)
2021-03-01T10:30:29.5863640Z > assert_warns_message(
2021-03-01T10:30:29.5864180Z UserWarning,
2021-03-01T10:30:29.5864790Z "Multiple active versions of the dataset matching the name"
2021-03-01T10:30:29.5865490Z " iris exist. Versions may be fundamentally different, "
2021-03-01T10:30:29.5866130Z "returning version 1.",
2021-03-01T10:30:29.5866670Z fetch_openml,
2021-03-01T10:30:29.5867210Z name=data_name,
2021-03-01T10:30:29.5867740Z as_frame=False
2021-03-01T10:30:29.5868240Z )
2021-03-01T10:30:29.5868530Z
2021-03-01T10:30:29.5868990Z data_id = 61
2021-03-01T10:30:29.5869820Z data_name = 'iris'
2021-03-01T10:30:29.5870390Z gzip_response = True
2021-03-01T10:30:29.5870990Z monkeypatch = <_pytest.monkeypatch.MonkeyPatch object at 0x12f4de8e0>
2021-03-01T10:30:29.5871390Z
2021-03-01T10:30:29.5871920Z ../1/s/sklearn/datasets/tests/test_openml.py:775:
2021-03-01T10:30:29.5872610Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5873300Z ../1/s/sklearn/utils/_testing.py:155: in assert_warns_message
2021-03-01T10:30:29.5873920Z result = func(*args, **kw)
2021-03-01T10:30:29.5874460Z args = ()
2021-03-01T10:30:29.5875050Z func = <function fetch_openml at 0x1230f0790>
2021-03-01T10:30:29.5876090Z kw = {'as_frame': False, 'name': 'iris'}
2021-03-01T10:30:29.5877430Z message = 'Multiple active versions of the dataset matching the name iris exist. Versions may be fundamentally different, returning version 1.'
2021-03-01T10:30:29.5878600Z w = [<warnings.WarningMessage object at 0x12f4deb50>, <warnings.WarningMessage object at 0x12f4de1c0>, <warnings.WarningMessage object at 0x12f4deb20>, <warnings.WarningMessage object at 0x12f4de490>]
2021-03-01T10:30:29.5880110Z warning_class = <class 'UserWarning'>
2021-03-01T10:30:29.5880790Z ../1/s/sklearn/utils/validation.py:63: in inner_f
2021-03-01T10:30:29.5881370Z return f(*args, **kwargs)
2021-03-01T10:30:29.5882250Z all_args = ['name']
2021-03-01T10:30:29.5882840Z args = ()
2021-03-01T10:30:29.5883700Z extra_args = -1
2021-03-01T10:30:29.5884330Z f = <function fetch_openml at 0x1230f0670>
2021-03-01T10:30:29.5885370Z kwargs = {'as_frame': False, 'name': 'iris'}
2021-03-01T10:30:29.5886550Z kwonly_args = ['version', 'data_id', 'data_home', 'target_column', 'cache', 'return_X_y', ...]
2021-03-01T10:30:29.5888310Z sig = <Signature (name: Optional[str] = None, *, version: Union[str, int] = 'active', data_id: Optional[int] = None, data_ho...List, NoneType] = 'default-target', cache: bool = True, return_X_y: bool = False, as_frame: Union[str, bool] = 'auto')>
2021-03-01T10:30:29.5890050Z version = '1.0 (renaming of 0.25)'
2021-03-01T10:30:29.5890750Z ../1/s/sklearn/datasets/_openml.py:919: in fetch_openml
2021-03-01T10:30:29.5891800Z bunch = _download_data_to_bunch(url, return_sparse, data_home,
2021-03-01T10:30:29.5892420Z as_frame = False
2021-03-01T10:30:29.5892960Z cache = True
2021-03-01T10:30:29.5893990Z data_columns = ['sepallength', 'sepalwidth', 'petallength', 'petalwidth']
2021-03-01T10:30:29.5895690Z data_description = {'collection_date': '1936', 'creator': 'R.A. Fisher', 'default_target_attribute': 'class', 'description': '**Author**:.... petal width in cm\n 5. class: \n -- Iris Setosa\n -- Iris Versicolour\n -- Iris Virginica', ...}
2021-03-01T10:30:29.5897260Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.5897910Z data_id = 61
2021-03-01T10:30:29.5898980Z data_info = {'did': 61, 'file_id': 61, 'format': 'ARFF', 'name': 'iris', ...}
2021-03-01T10:30:29.5900740Z data_qualities = [{'name': 'AutoCorrelation', 'value': '0.9865771812080537'}, {'name': 'CfsSubsetEval_DecisionStumpAUC', 'value': '0.95...AUC', 'value': '0.9565333333333332'}, {'name': 'CfsSubsetEval_NaiveBayesErrRate', 'value': '0.06666666666666667'}, ...]
2021-03-01T10:30:29.5902490Z feature = {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}
2021-03-01T10:30:29.5904320Z features_list = [{'data_type': 'numeric', 'index': '0', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}, {'data_type': 'numer...ifier': 'false', ...}, {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}]
2021-03-01T10:30:29.5905780Z name = 'iris'
2021-03-01T10:30:29.5906370Z return_X_y = False
2021-03-01T10:30:29.5906930Z return_sparse = False
2021-03-01T10:30:29.5907480Z shape = (150, 5)
2021-03-01T10:30:29.5908400Z target_column = 'default-target'
2021-03-01T10:30:29.5909350Z target_columns = ['class']
2021-03-01T10:30:29.5910300Z url = 'data/v1/download/61'
2021-03-01T10:30:29.5911260Z version = 'active'
2021-03-01T10:30:29.5911930Z ../1/s/sklearn/datasets/_openml.py:637: in _download_data_to_bunch
2021-03-01T10:30:29.5912600Z out = _retry_with_clean_cache(url, data_home)(
2021-03-01T10:30:29.5913180Z as_frame = False
2021-03-01T10:30:29.5913710Z col_idx = 4
2021-03-01T10:30:29.5914250Z col_slice_x = [0, 1, 2, 3]
2021-03-01T10:30:29.5914780Z col_slice_y = [4]
2021-03-01T10:30:29.5915880Z data_columns = ['sepallength', 'sepalwidth', 'petallength', 'petalwidth']
2021-03-01T10:30:29.5916990Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.6891310Z feat = {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}
2021-03-01T10:30:29.6894370Z features_dict = {'class': {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}, 'petallengt...}, 'sepallength': {'data_type': 'numeric', 'index': '0', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}, ...}
2021-03-01T10:30:29.6897320Z features_list = [{'data_type': 'numeric', 'index': '0', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}, {'data_type': 'numer...ifier': 'false', ...}, {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}]
2021-03-01T10:30:29.6898490Z frame = None
2021-03-01T10:30:29.6899440Z md5_checksum = 'ad484452702105cbf3d30f8deaba39a9'
2021-03-01T10:30:29.6900000Z nominal_attributes = None
2021-03-01T10:30:29.6900500Z nr_missing = 0
2021-03-01T10:30:29.6901080Z parse_arff = <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>
2021-03-01T10:30:29.6901870Z postprocess = <function _download_data_to_bunch.<locals>.postprocess at 0x12f4c2dc0>
2021-03-01T10:30:29.6902530Z return_type = 3
2021-03-01T10:30:29.6903060Z shape = (150, 5)
2021-03-01T10:30:29.6903590Z sparse = False
2021-03-01T10:30:29.6904990Z target_columns = ['class']
2021-03-01T10:30:29.6905950Z url = 'data/v1/download/61'
2021-03-01T10:30:29.6906610Z ../1/s/sklearn/datasets/_openml.py:70: in wrapper
2021-03-01T10:30:29.6907200Z return f(*args, **kw)
2021-03-01T10:30:29.6908230Z args = ('data/v1/download/61', '/Users/runner/scikit_learn_data/openml')
2021-03-01T10:30:29.6909330Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.6910040Z f = <function _load_arff_response at 0x1230f0430>
2021-03-01T10:30:29.6911530Z kw = {'encode_nominal': True, 'md5_checksum': 'ad484452702105cbf3d30f8deaba39a9', 'parse_arff': <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>, 'return_type': 3}
2021-03-01T10:30:29.6913030Z local_path = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6914100Z openml_path = 'data/v1/download/61'
2021-03-01T10:30:29.6914780Z ../1/s/sklearn/datasets/_openml.py:504: in _load_arff_response
2021-03-01T10:30:29.6915550Z response = _open_openml_url(url, data_home)
2021-03-01T10:30:29.6916580Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.6917230Z encode_nominal = True
2021-03-01T10:30:29.6918190Z md5_checksum = 'ad484452702105cbf3d30f8deaba39a9'
2021-03-01T10:30:29.6918940Z parse_arff = <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>
2021-03-01T10:30:29.6919590Z return_type = 3
2021-03-01T10:30:29.6920480Z url = 'data/v1/download/61'
2021-03-01T10:30:29.6921150Z ../1/s/sklearn/datasets/_openml.py:130: in _open_openml_url
2021-03-01T10:30:29.6922130Z return gzip.GzipFile(local_path, 'rb')
2021-03-01T10:30:29.6923150Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.6924360Z fdst = <_io.BufferedWriter name='/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'>
2021-03-01T10:30:29.6925250Z fsrc = <sklearn.datasets.tests.test_openml._MockHTTPResponse object at 0x12f4de580>
2021-03-01T10:30:29.6926050Z is_gzip_encoded = <function _open_openml_url.<locals>.is_gzip_encoded at 0x12f4c2c10>
2021-03-01T10:30:29.6927370Z local_path = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6928440Z opener = <built-in function open>
2021-03-01T10:30:29.6929420Z openml_path = 'data/v1/download/61'
2021-03-01T10:30:29.6930100Z req = <urllib.request.Request object at 0x12f4de0a0>
2021-03-01T10:30:29.6930820Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.6931440Z
2021-03-01T10:30:29.6932530Z self = <[AttributeError("'GzipFile' object has no attribute 'fileobj'") raised in repr()] GzipFile object at 0x12f4defa0>
2021-03-01T10:30:29.6933770Z filename = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6934890Z mode = 'rb', compresslevel = 9, fileobj = None, mtime = None
2021-03-01T10:30:29.6935320Z
2021-03-01T10:30:29.6935850Z def __init__(self, filename=None, mode=None,
2021-03-01T10:30:29.6936560Z compresslevel=_COMPRESS_LEVEL_BEST, fileobj=None, mtime=None):
2021-03-01T10:30:29.6937260Z """Constructor for the GzipFile class.
2021-03-01T10:30:29.6937790Z
2021-03-01T10:30:29.6938340Z At least one of fileobj and filename must be given a
2021-03-01T10:30:29.6939350Z non-trivial value.
2021-03-01T10:30:29.6939870Z
2021-03-01T10:30:29.6940440Z The new class instance is based on fileobj, which can be a regular
2021-03-01T10:30:29.6941160Z file, an io.BytesIO object, or any other object which simulates a file.
2021-03-01T10:30:29.6941880Z It defaults to None, in which case filename is opened to provide
2021-03-01T10:30:29.6942500Z a file object.
2021-03-01T10:30:29.6942980Z
2021-03-01T10:30:29.6943890Z When fileobj is not None, the filename argument is only used to be
2021-03-01T10:30:29.6944610Z included in the gzip file header, which may include the original
2021-03-01T10:30:29.6945310Z filename of the uncompressed file. It defaults to the filename of
2021-03-01T10:30:29.6946030Z fileobj, if discernible; otherwise, it defaults to the empty string,
2021-03-01T10:30:29.6946750Z and in this case the original filename is not included in the header.
2021-03-01T10:30:29.6947340Z
2021-03-01T10:30:29.6948320Z The mode argument can be any of 'r', 'rb', 'a', 'ab', 'w', 'wb', 'x', or
2021-03-01T10:30:29.6949460Z 'xb' depending on whether the file will be read or written. The default
2021-03-01T10:30:29.6950510Z is the mode of fileobj if discernible; otherwise, the default is 'rb'.
2021-03-01T10:30:29.6951530Z A mode of 'r' is equivalent to one of 'rb', and similarly for 'w' and
2021-03-01T10:30:29.6952520Z 'wb', 'a' and 'ab', and 'x' and 'xb'.
2021-03-01T10:30:29.6953050Z
2021-03-01T10:30:29.6953540Z The compresslevel argument is an integer from 0 to 9 controlling the
2021-03-01T10:30:29.6954200Z level of compression; 1 is fastest and produces the least compression,
2021-03-01T10:30:29.6954920Z and 9 is slowest and produces the most compression. 0 is no compression
2021-03-01T10:30:29.6955550Z at all. The default is 9.
2021-03-01T10:30:29.6956070Z
2021-03-01T10:30:29.6956640Z The mtime argument is an optional numeric timestamp to be written
2021-03-01T10:30:29.6957350Z to the last modification time field in the stream when compressing.
2021-03-01T10:30:29.6958020Z If omitted or None, the current time is used.
2021-03-01T10:30:29.6958560Z
2021-03-01T10:30:29.6959020Z """
2021-03-01T10:30:29.6959470Z
2021-03-01T10:30:29.6960370Z if mode and ('t' in mode or 'U' in mode):
2021-03-01T10:30:29.6961100Z raise ValueError("Invalid mode: {!r}".format(mode))
2021-03-01T10:30:29.6962100Z if mode and 'b' not in mode:
2021-03-01T10:30:29.6963030Z mode += 'b'
2021-03-01T10:30:29.6963610Z if fileobj is None:
2021-03-01T10:30:29.6964600Z > fileobj = self.myfileobj = builtins.open(filename, mode or 'rb')
2021-03-01T10:30:29.6965920Z E FileNotFoundError: [Errno 2] No such file or directory: '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6966540Z
2021-03-01T10:30:29.6967010Z compresslevel = 9
2021-03-01T10:30:29.6968010Z filename = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6968680Z fileobj = None
2021-03-01T10:30:29.6969500Z mode = 'rb'
2021-03-01T10:30:29.6970250Z mtime = None
2021-03-01T10:30:29.6971380Z self = <[AttributeError("'GzipFile' object has no attribute 'fileobj'") raised in repr()] GzipFile object at 0x12f4defa0>
2021-03-01T10:30:29.6971920Z
2021-03-01T10:30:29.6972520Z /usr/local/miniconda/envs/testvenv/lib/python3.9/gzip.py:173: FileNotFoundError
```
</details>
https://dev.azure.com/scikit-learn/scikit-learn/_build/results?buildId=27014&view=logs&j=97641769-79fb-5590-9088-a30ce9b850b9&t=4745baa1-36b5-56c8-9a8e-6480742db1a6&l=796
Not sure if this is related to the retry mechanism of #12526 or not. | 1.0 | Random failure in fetch_openml tests - It appended once on the CI of the unrelated PR #19558 when running `test_fetch_openml_iris[True]`:
```python-traceback
../1/s/sklearn/externals/_arff.py:911:
[...]
E sklearn.externals._arff.BadLayout: Invalid layout of the ARFF file, at line 0.
[...]
During handling of the above exception, another exception occurred:
[...]
FileNotFoundError: [Errno 2] No such file or directory: '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz
```
Full traceback:
<details>
```
2021-03-01T10:30:29.5661420Z [gw1] darwin -- Python 3.9.2 /usr/local/miniconda/envs/testvenv/bin/python
2021-03-01T10:30:29.5662060Z
2021-03-01T10:30:29.5663160Z args = ('data/v1/download/61', '/Users/runner/scikit_learn_data/openml')
2021-03-01T10:30:29.5664990Z kw = {'encode_nominal': True, 'md5_checksum': 'ad484452702105cbf3d30f8deaba39a9', 'parse_arff': <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>, 'return_type': 3}
2021-03-01T10:30:29.5666510Z local_path = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.5666930Z
2021-03-01T10:30:29.5667290Z @wraps(f)
2021-03-01T10:30:29.5667710Z def wrapper(*args, **kw):
2021-03-01T10:30:29.5668240Z if data_home is None:
2021-03-01T10:30:29.5668770Z return f(*args, **kw)
2021-03-01T10:30:29.5669240Z try:
2021-03-01T10:30:29.5669710Z > return f(*args, **kw)
2021-03-01T10:30:29.5669990Z
2021-03-01T10:30:29.5670890Z args = ('data/v1/download/61', '/Users/runner/scikit_learn_data/openml')
2021-03-01T10:30:29.5671930Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.5672580Z f = <function _load_arff_response at 0x1230f0430>
2021-03-01T10:30:29.5673960Z kw = {'encode_nominal': True, 'md5_checksum': 'ad484452702105cbf3d30f8deaba39a9', 'parse_arff': <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>, 'return_type': 3}
2021-03-01T10:30:29.5675430Z local_path = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.5676500Z openml_path = 'data/v1/download/61'
2021-03-01T10:30:29.5676850Z
2021-03-01T10:30:29.5677280Z ../1/s/sklearn/datasets/_openml.py:62:
2021-03-01T10:30:29.5677820Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5678840Z
2021-03-01T10:30:29.5679690Z url = 'data/v1/download/61'
2021-03-01T10:30:29.5680710Z data_home = '/Users/runner/scikit_learn_data/openml', return_type = 3
2021-03-01T10:30:29.5681370Z encode_nominal = True
2021-03-01T10:30:29.5682010Z parse_arff = <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>
2021-03-01T10:30:29.5683050Z md5_checksum = 'ad484452702105cbf3d30f8deaba39a9'
2021-03-01T10:30:29.5683450Z
2021-03-01T10:30:29.5683950Z def _load_arff_response(
2021-03-01T10:30:29.5684480Z url: str,
2021-03-01T10:30:29.5685020Z data_home: Optional[str],
2021-03-01T10:30:29.5685620Z return_type, encode_nominal: bool,
2021-03-01T10:30:29.5686390Z parse_arff: Callable[[ArffContainerType], Tuple],
2021-03-01T10:30:29.5687020Z md5_checksum: str
2021-03-01T10:30:29.5687860Z ) -> Tuple:
2021-03-01T10:30:29.5688520Z """Load arff data with url and parses arff response with parse_arff"""
2021-03-01T10:30:29.5689230Z response = _open_openml_url(url, data_home)
2021-03-01T10:30:29.5689670Z
2021-03-01T10:30:29.5690080Z with closing(response):
2021-03-01T10:30:29.5690680Z # Note that if the data is dense, no reading is done until the data
2021-03-01T10:30:29.5691850Z # generator is iterated.
2021-03-01T10:30:29.5692460Z actual_md5_checksum = hashlib.md5()
2021-03-01T10:30:29.5692990Z
2021-03-01T10:30:29.5693540Z def _stream_checksum_generator(response):
2021-03-01T10:30:29.5694180Z for line in response:
2021-03-01T10:30:29.5694810Z actual_md5_checksum.update(line)
2021-03-01T10:30:29.5695780Z yield line.decode('utf-8')
2021-03-01T10:30:29.5696390Z
2021-03-01T10:30:29.5696950Z stream = _stream_checksum_generator(response)
2021-03-01T10:30:29.5697500Z
2021-03-01T10:30:29.5698000Z > arff = _arff.load(stream,
2021-03-01T10:30:29.5698620Z return_type=return_type,
2021-03-01T10:30:29.5699250Z encode_nominal=encode_nominal)
2021-03-01T10:30:29.5699650Z
2021-03-01T10:30:29.5700270Z _stream_checksum_generator = <function _load_arff_response.<locals>._stream_checksum_generator at 0x12f4c2a60>
2021-03-01T10:30:29.5701030Z actual_md5_checksum = <md5 _hashlib.HASH object @ 0x12f55b230>
2021-03-01T10:30:29.5702040Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.5702640Z encode_nominal = True
2021-03-01T10:30:29.5703580Z md5_checksum = 'ad484452702105cbf3d30f8deaba39a9'
2021-03-01T10:30:29.5704300Z parse_arff = <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>
2021-03-01T10:30:29.5704960Z response = <gzip on 0x12f4decd0>
2021-03-01T10:30:29.5705490Z return_type = 3
2021-03-01T10:30:29.5706150Z stream = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5707180Z url = 'data/v1/download/61'
2021-03-01T10:30:29.5707570Z
2021-03-01T10:30:29.5708090Z ../1/s/sklearn/datasets/_openml.py:518:
2021-03-01T10:30:29.5708750Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5709180Z
2021-03-01T10:30:29.5709780Z fp = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5710480Z encode_nominal = True, return_type = 3
2021-03-01T10:30:29.5710740Z
2021-03-01T10:30:29.5711190Z def load(fp, encode_nominal=False, return_type=DENSE):
2021-03-01T10:30:29.5712280Z '''Load a file-like object containing the ARFF document and convert it into
2021-03-01T10:30:29.5712980Z a Python object.
2021-03-01T10:30:29.5713460Z
2021-03-01T10:30:29.5714330Z :param fp: a file-like object.
2021-03-01T10:30:29.5715040Z :param encode_nominal: boolean, if True perform a label encoding
2021-03-01T10:30:29.5715770Z while reading the .arff file.
2021-03-01T10:30:29.5716880Z :param return_type: determines the data structure used to store the
2021-03-01T10:30:29.5717610Z dataset. Can be one of `arff.DENSE`, `arff.COO`, `arff.LOD`,
2021-03-01T10:30:29.5718280Z `arff.DENSE_GEN` or `arff.LOD_GEN`.
2021-03-01T10:30:29.5718930Z Consult the sections on `working with sparse data`_ and `loading
2021-03-01T10:30:29.5719560Z progressively`_.
2021-03-01T10:30:29.5720120Z :return: a dictionary.
2021-03-01T10:30:29.5720980Z '''
2021-03-01T10:30:29.5721560Z decoder = ArffDecoder()
2021-03-01T10:30:29.5722170Z > return decoder.decode(fp, encode_nominal=encode_nominal,
2021-03-01T10:30:29.5722830Z return_type=return_type)
2021-03-01T10:30:29.5723220Z
2021-03-01T10:30:29.5723750Z decoder = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5724360Z encode_nominal = True
2021-03-01T10:30:29.5725020Z fp = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5725680Z return_type = 3
2021-03-01T10:30:29.5726010Z
2021-03-01T10:30:29.5726510Z ../1/s/sklearn/externals/_arff.py:1078:
2021-03-01T10:30:29.5727520Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5727970Z
2021-03-01T10:30:29.5728520Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5729230Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5729920Z encode_nominal = True, return_type = 3
2021-03-01T10:30:29.5730270Z
2021-03-01T10:30:29.5730840Z def decode(self, s, encode_nominal=False, return_type=DENSE):
2021-03-01T10:30:29.5731920Z '''Returns the Python representation of a given ARFF file.
2021-03-01T10:30:29.5732540Z
2021-03-01T10:30:29.5733130Z When a file object is passed as an argument, this method reads lines
2021-03-01T10:30:29.5733860Z iteratively, avoiding to load unnecessary information to the memory.
2021-03-01T10:30:29.5734440Z
2021-03-01T10:30:29.5735000Z :param s: a string or file object with the ARFF file.
2021-03-01T10:30:29.5735710Z :param encode_nominal: boolean, if True perform a label encoding
2021-03-01T10:30:29.5736380Z while reading the .arff file.
2021-03-01T10:30:29.5737040Z :param return_type: determines the data structure used to store the
2021-03-01T10:30:29.5737770Z dataset. Can be one of `arff.DENSE`, `arff.COO`, `arff.LOD`,
2021-03-01T10:30:29.5738420Z `arff.DENSE_GEN` or `arff.LOD_GEN`.
2021-03-01T10:30:29.5739080Z Consult the sections on `working with sparse data`_ and `loading
2021-03-01T10:30:29.5739700Z progressively`_.
2021-03-01T10:30:29.5740550Z '''
2021-03-01T10:30:29.5741070Z try:
2021-03-01T10:30:29.5741650Z return self._decode(s, encode_nominal=encode_nominal,
2021-03-01T10:30:29.5742320Z matrix_type=return_type)
2021-03-01T10:30:29.5742940Z except ArffException as e:
2021-03-01T10:30:29.5743530Z e.line = self._current_line
2021-03-01T10:30:29.5744070Z > raise e
2021-03-01T10:30:29.5744380Z
2021-03-01T10:30:29.5744840Z encode_nominal = True
2021-03-01T10:30:29.5745350Z return_type = 3
2021-03-01T10:30:29.5745990Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5746750Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5747170Z
2021-03-01T10:30:29.5747670Z ../1/s/sklearn/externals/_arff.py:915:
2021-03-01T10:30:29.5748820Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5749320Z
2021-03-01T10:30:29.5759090Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5759880Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5760860Z encode_nominal = True, return_type = 3
2021-03-01T10:30:29.5761220Z
2021-03-01T10:30:29.5761810Z def decode(self, s, encode_nominal=False, return_type=DENSE):
2021-03-01T10:30:29.5763070Z '''Returns the Python representation of a given ARFF file.
2021-03-01T10:30:29.5763700Z
2021-03-01T10:30:29.5764290Z When a file object is passed as an argument, this method reads lines
2021-03-01T10:30:29.5765020Z iteratively, avoiding to load unnecessary information to the memory.
2021-03-01T10:30:29.5765650Z
2021-03-01T10:30:29.5766240Z :param s: a string or file object with the ARFF file.
2021-03-01T10:30:29.5766950Z :param encode_nominal: boolean, if True perform a label encoding
2021-03-01T10:30:29.5767610Z while reading the .arff file.
2021-03-01T10:30:29.5768350Z :param return_type: determines the data structure used to store the
2021-03-01T10:30:29.5769090Z dataset. Can be one of `arff.DENSE`, `arff.COO`, `arff.LOD`,
2021-03-01T10:30:29.5769750Z `arff.DENSE_GEN` or `arff.LOD_GEN`.
2021-03-01T10:30:29.5770400Z Consult the sections on `working with sparse data`_ and `loading
2021-03-01T10:30:29.5771440Z progressively`_.
2021-03-01T10:30:29.5772330Z '''
2021-03-01T10:30:29.5772870Z try:
2021-03-01T10:30:29.5773460Z > return self._decode(s, encode_nominal=encode_nominal,
2021-03-01T10:30:29.5774130Z matrix_type=return_type)
2021-03-01T10:30:29.5774510Z
2021-03-01T10:30:29.5774970Z encode_nominal = True
2021-03-01T10:30:29.5775470Z return_type = 3
2021-03-01T10:30:29.5776180Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5776940Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5777360Z
2021-03-01T10:30:29.5777860Z ../1/s/sklearn/externals/_arff.py:911:
2021-03-01T10:30:29.5778580Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5779030Z
2021-03-01T10:30:29.5779580Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5780320Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5781000Z encode_nominal = True, matrix_type = 3
2021-03-01T10:30:29.5781350Z
2021-03-01T10:30:29.5781920Z def _decode(self, s, encode_nominal=False, matrix_type=DENSE):
2021-03-01T10:30:29.5782950Z '''Do the job the ``encode``.'''
2021-03-01T10:30:29.5783510Z
2021-03-01T10:30:29.5784050Z # Make sure this method is idempotent
2021-03-01T10:30:29.5784620Z self._current_line = 0
2021-03-01T10:30:29.5785130Z
2021-03-01T10:30:29.5785660Z # If string, convert to a list of lines
2021-03-01T10:30:29.5786290Z if isinstance(s, basestring):
2021-03-01T10:30:29.5787290Z s = s.strip('\r\n ').replace('\r\n', '\n').split('\n')
2021-03-01T10:30:29.5787910Z
2021-03-01T10:30:29.5788420Z # Create the return object
2021-03-01T10:30:29.5789170Z obj: ArffContainerType = {
2021-03-01T10:30:29.5790120Z u'description': u'',
2021-03-01T10:30:29.5791070Z u'relation': u'',
2021-03-01T10:30:29.5792000Z u'attributes': [],
2021-03-01T10:30:29.5792930Z u'data': []
2021-03-01T10:30:29.5793490Z }
2021-03-01T10:30:29.5794000Z attribute_names = {}
2021-03-01T10:30:29.5794500Z
2021-03-01T10:30:29.5795010Z # Create the data helper object
2021-03-01T10:30:29.5795610Z data = _get_data_object_for_decoding(matrix_type)
2021-03-01T10:30:29.5796180Z
2021-03-01T10:30:29.5796670Z # Read all lines
2021-03-01T10:30:29.5797210Z STATE = _TK_DESCRIPTION
2021-03-01T10:30:29.5797740Z s = iter(s)
2021-03-01T10:30:29.5798260Z for row in s:
2021-03-01T10:30:29.5799010Z self._current_line += 1
2021-03-01T10:30:29.5799580Z # Ignore empty lines
2021-03-01T10:30:29.5800500Z row = row.strip(' \r\n')
2021-03-01T10:30:29.5801130Z if not row: continue
2021-03-01T10:30:29.5801650Z
2021-03-01T10:30:29.5802130Z u_row = row.upper()
2021-03-01T10:30:29.5802640Z
2021-03-01T10:30:29.5803600Z # DESCRIPTION -----------------------------------------------------
2021-03-01T10:30:29.5804410Z if u_row.startswith(_TK_DESCRIPTION) and STATE == _TK_DESCRIPTION:
2021-03-01T10:30:29.5805540Z obj['description'] += self._decode_comment(row) + '\n'
2021-03-01T10:30:29.5806670Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5807310Z
2021-03-01T10:30:29.5808260Z # RELATION --------------------------------------------------------
2021-03-01T10:30:29.5809010Z elif u_row.startswith(_TK_RELATION):
2021-03-01T10:30:29.5809660Z if STATE != _TK_DESCRIPTION:
2021-03-01T10:30:29.5810280Z raise BadLayout()
2021-03-01T10:30:29.5810790Z
2021-03-01T10:30:29.5811300Z STATE = _TK_RELATION
2021-03-01T10:30:29.5812870Z obj['relation'] = self._decode_relation(row)
2021-03-01T10:30:29.5813990Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5814690Z
2021-03-01T10:30:29.5815760Z # ATTRIBUTE -------------------------------------------------------
2021-03-01T10:30:29.5816510Z elif u_row.startswith(_TK_ATTRIBUTE):
2021-03-01T10:30:29.5817220Z if STATE != _TK_RELATION and STATE != _TK_ATTRIBUTE:
2021-03-01T10:30:29.5817880Z raise BadLayout()
2021-03-01T10:30:29.5818380Z
2021-03-01T10:30:29.5818890Z STATE = _TK_ATTRIBUTE
2021-03-01T10:30:29.5819400Z
2021-03-01T10:30:29.5819930Z attr = self._decode_attribute(row)
2021-03-01T10:30:29.5820580Z if attr[0] in attribute_names:
2021-03-01T10:30:29.5821260Z raise BadAttributeName(attr[0], attribute_names[attr[0]])
2021-03-01T10:30:29.5821890Z else:
2021-03-01T10:30:29.5822490Z attribute_names[attr[0]] = self._current_line
2021-03-01T10:30:29.5823490Z obj['attributes'].append(attr)
2021-03-01T10:30:29.5824070Z
2021-03-01T10:30:29.5824870Z if isinstance(attr[1], (list, tuple)):
2021-03-01T10:30:29.5825590Z if encode_nominal:
2021-03-01T10:30:29.5826250Z conversor = EncodedNominalConversor(attr[1])
2021-03-01T10:30:29.5826860Z else:
2021-03-01T10:30:29.5827480Z conversor = NominalConversor(attr[1])
2021-03-01T10:30:29.5828080Z else:
2021-03-01T10:30:29.5829070Z CONVERSOR_MAP = {'STRING': unicode,
2021-03-01T10:30:29.5830200Z 'INTEGER': lambda x: int(float(x)),
2021-03-01T10:30:29.5831320Z 'NUMERIC': float,
2021-03-01T10:30:29.5832380Z 'REAL': float}
2021-03-01T10:30:29.5833060Z conversor = CONVERSOR_MAP[attr[1]]
2021-03-01T10:30:29.5833630Z
2021-03-01T10:30:29.5834170Z self._conversors.append(conversor)
2021-03-01T10:30:29.5835210Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5835860Z
2021-03-01T10:30:29.5836830Z # DATA ------------------------------------------------------------
2021-03-01T10:30:29.5837550Z elif u_row.startswith(_TK_DATA):
2021-03-01T10:30:29.5838190Z if STATE != _TK_ATTRIBUTE:
2021-03-01T10:30:29.5838790Z raise BadLayout()
2021-03-01T10:30:29.5839310Z
2021-03-01T10:30:29.5839790Z break
2021-03-01T10:30:29.5840780Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5841640Z
2021-03-01T10:30:29.5842600Z # COMMENT ---------------------------------------------------------
2021-03-01T10:30:29.5843340Z elif u_row.startswith(_TK_COMMENT):
2021-03-01T10:30:29.5843940Z pass
2021-03-01T10:30:29.5844930Z # -----------------------------------------------------------------
2021-03-01T10:30:29.5845590Z else:
2021-03-01T10:30:29.5846130Z # Never found @DATA
2021-03-01T10:30:29.5846660Z > raise BadLayout()
2021-03-01T10:30:29.5847340Z E sklearn.externals._arff.BadLayout: Invalid layout of the ARFF file, at line 0.
2021-03-01T10:30:29.5847820Z
2021-03-01T10:30:29.5848620Z STATE = '%'
2021-03-01T10:30:29.5849150Z attribute_names = {}
2021-03-01T10:30:29.5849790Z data = <sklearn.externals._arff.DenseGeneratorData object at 0x12f4debe0>
2021-03-01T10:30:29.5850420Z encode_nominal = True
2021-03-01T10:30:29.5850930Z matrix_type = 3
2021-03-01T10:30:29.5851950Z obj = {'attributes': [], 'data': [], 'description': '', 'relation': ''}
2021-03-01T10:30:29.5852800Z s = <generator object _load_arff_response.<locals>._stream_checksum_generator at 0x12f0b30b0>
2021-03-01T10:30:29.5853930Z self = <sklearn.externals._arff.ArffDecoder object at 0x12f4de6a0>
2021-03-01T10:30:29.5854360Z
2021-03-01T10:30:29.5855010Z ../1/s/sklearn/externals/_arff.py:878: BadLayout
2021-03-01T10:30:29.5855390Z
2021-03-01T10:30:29.5856050Z During handling of the above exception, another exception occurred:
2021-03-01T10:30:29.5856480Z
2021-03-01T10:30:29.5857040Z monkeypatch = <_pytest.monkeypatch.MonkeyPatch object at 0x12f4de8e0>
2021-03-01T10:30:29.5857650Z gzip_response = True
2021-03-01T10:30:29.5858060Z
2021-03-01T10:30:29.5858990Z @pytest.mark.parametrize('gzip_response', [True, False])
2021-03-01T10:30:29.5859730Z def test_fetch_openml_iris(monkeypatch, gzip_response):
2021-03-01T10:30:29.5860420Z # classification dataset with numeric only columns
2021-03-01T10:30:29.5861010Z data_id = 61
2021-03-01T10:30:29.5861870Z data_name = 'iris'
2021-03-01T10:30:29.5862400Z
2021-03-01T10:30:29.5862990Z _monkey_patch_webbased_functions(monkeypatch, data_id, gzip_response)
2021-03-01T10:30:29.5863640Z > assert_warns_message(
2021-03-01T10:30:29.5864180Z UserWarning,
2021-03-01T10:30:29.5864790Z "Multiple active versions of the dataset matching the name"
2021-03-01T10:30:29.5865490Z " iris exist. Versions may be fundamentally different, "
2021-03-01T10:30:29.5866130Z "returning version 1.",
2021-03-01T10:30:29.5866670Z fetch_openml,
2021-03-01T10:30:29.5867210Z name=data_name,
2021-03-01T10:30:29.5867740Z as_frame=False
2021-03-01T10:30:29.5868240Z )
2021-03-01T10:30:29.5868530Z
2021-03-01T10:30:29.5868990Z data_id = 61
2021-03-01T10:30:29.5869820Z data_name = 'iris'
2021-03-01T10:30:29.5870390Z gzip_response = True
2021-03-01T10:30:29.5870990Z monkeypatch = <_pytest.monkeypatch.MonkeyPatch object at 0x12f4de8e0>
2021-03-01T10:30:29.5871390Z
2021-03-01T10:30:29.5871920Z ../1/s/sklearn/datasets/tests/test_openml.py:775:
2021-03-01T10:30:29.5872610Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.5873300Z ../1/s/sklearn/utils/_testing.py:155: in assert_warns_message
2021-03-01T10:30:29.5873920Z result = func(*args, **kw)
2021-03-01T10:30:29.5874460Z args = ()
2021-03-01T10:30:29.5875050Z func = <function fetch_openml at 0x1230f0790>
2021-03-01T10:30:29.5876090Z kw = {'as_frame': False, 'name': 'iris'}
2021-03-01T10:30:29.5877430Z message = 'Multiple active versions of the dataset matching the name iris exist. Versions may be fundamentally different, returning version 1.'
2021-03-01T10:30:29.5878600Z w = [<warnings.WarningMessage object at 0x12f4deb50>, <warnings.WarningMessage object at 0x12f4de1c0>, <warnings.WarningMessage object at 0x12f4deb20>, <warnings.WarningMessage object at 0x12f4de490>]
2021-03-01T10:30:29.5880110Z warning_class = <class 'UserWarning'>
2021-03-01T10:30:29.5880790Z ../1/s/sklearn/utils/validation.py:63: in inner_f
2021-03-01T10:30:29.5881370Z return f(*args, **kwargs)
2021-03-01T10:30:29.5882250Z all_args = ['name']
2021-03-01T10:30:29.5882840Z args = ()
2021-03-01T10:30:29.5883700Z extra_args = -1
2021-03-01T10:30:29.5884330Z f = <function fetch_openml at 0x1230f0670>
2021-03-01T10:30:29.5885370Z kwargs = {'as_frame': False, 'name': 'iris'}
2021-03-01T10:30:29.5886550Z kwonly_args = ['version', 'data_id', 'data_home', 'target_column', 'cache', 'return_X_y', ...]
2021-03-01T10:30:29.5888310Z sig = <Signature (name: Optional[str] = None, *, version: Union[str, int] = 'active', data_id: Optional[int] = None, data_ho...List, NoneType] = 'default-target', cache: bool = True, return_X_y: bool = False, as_frame: Union[str, bool] = 'auto')>
2021-03-01T10:30:29.5890050Z version = '1.0 (renaming of 0.25)'
2021-03-01T10:30:29.5890750Z ../1/s/sklearn/datasets/_openml.py:919: in fetch_openml
2021-03-01T10:30:29.5891800Z bunch = _download_data_to_bunch(url, return_sparse, data_home,
2021-03-01T10:30:29.5892420Z as_frame = False
2021-03-01T10:30:29.5892960Z cache = True
2021-03-01T10:30:29.5893990Z data_columns = ['sepallength', 'sepalwidth', 'petallength', 'petalwidth']
2021-03-01T10:30:29.5895690Z data_description = {'collection_date': '1936', 'creator': 'R.A. Fisher', 'default_target_attribute': 'class', 'description': '**Author**:.... petal width in cm\n 5. class: \n -- Iris Setosa\n -- Iris Versicolour\n -- Iris Virginica', ...}
2021-03-01T10:30:29.5897260Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.5897910Z data_id = 61
2021-03-01T10:30:29.5898980Z data_info = {'did': 61, 'file_id': 61, 'format': 'ARFF', 'name': 'iris', ...}
2021-03-01T10:30:29.5900740Z data_qualities = [{'name': 'AutoCorrelation', 'value': '0.9865771812080537'}, {'name': 'CfsSubsetEval_DecisionStumpAUC', 'value': '0.95...AUC', 'value': '0.9565333333333332'}, {'name': 'CfsSubsetEval_NaiveBayesErrRate', 'value': '0.06666666666666667'}, ...]
2021-03-01T10:30:29.5902490Z feature = {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}
2021-03-01T10:30:29.5904320Z features_list = [{'data_type': 'numeric', 'index': '0', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}, {'data_type': 'numer...ifier': 'false', ...}, {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}]
2021-03-01T10:30:29.5905780Z name = 'iris'
2021-03-01T10:30:29.5906370Z return_X_y = False
2021-03-01T10:30:29.5906930Z return_sparse = False
2021-03-01T10:30:29.5907480Z shape = (150, 5)
2021-03-01T10:30:29.5908400Z target_column = 'default-target'
2021-03-01T10:30:29.5909350Z target_columns = ['class']
2021-03-01T10:30:29.5910300Z url = 'data/v1/download/61'
2021-03-01T10:30:29.5911260Z version = 'active'
2021-03-01T10:30:29.5911930Z ../1/s/sklearn/datasets/_openml.py:637: in _download_data_to_bunch
2021-03-01T10:30:29.5912600Z out = _retry_with_clean_cache(url, data_home)(
2021-03-01T10:30:29.5913180Z as_frame = False
2021-03-01T10:30:29.5913710Z col_idx = 4
2021-03-01T10:30:29.5914250Z col_slice_x = [0, 1, 2, 3]
2021-03-01T10:30:29.5914780Z col_slice_y = [4]
2021-03-01T10:30:29.5915880Z data_columns = ['sepallength', 'sepalwidth', 'petallength', 'petalwidth']
2021-03-01T10:30:29.5916990Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.6891310Z feat = {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}
2021-03-01T10:30:29.6894370Z features_dict = {'class': {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}, 'petallengt...}, 'sepallength': {'data_type': 'numeric', 'index': '0', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}, ...}
2021-03-01T10:30:29.6897320Z features_list = [{'data_type': 'numeric', 'index': '0', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}, {'data_type': 'numer...ifier': 'false', ...}, {'data_type': 'nominal', 'index': '4', 'is_ignore': 'false', 'is_row_identifier': 'false', ...}]
2021-03-01T10:30:29.6898490Z frame = None
2021-03-01T10:30:29.6899440Z md5_checksum = 'ad484452702105cbf3d30f8deaba39a9'
2021-03-01T10:30:29.6900000Z nominal_attributes = None
2021-03-01T10:30:29.6900500Z nr_missing = 0
2021-03-01T10:30:29.6901080Z parse_arff = <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>
2021-03-01T10:30:29.6901870Z postprocess = <function _download_data_to_bunch.<locals>.postprocess at 0x12f4c2dc0>
2021-03-01T10:30:29.6902530Z return_type = 3
2021-03-01T10:30:29.6903060Z shape = (150, 5)
2021-03-01T10:30:29.6903590Z sparse = False
2021-03-01T10:30:29.6904990Z target_columns = ['class']
2021-03-01T10:30:29.6905950Z url = 'data/v1/download/61'
2021-03-01T10:30:29.6906610Z ../1/s/sklearn/datasets/_openml.py:70: in wrapper
2021-03-01T10:30:29.6907200Z return f(*args, **kw)
2021-03-01T10:30:29.6908230Z args = ('data/v1/download/61', '/Users/runner/scikit_learn_data/openml')
2021-03-01T10:30:29.6909330Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.6910040Z f = <function _load_arff_response at 0x1230f0430>
2021-03-01T10:30:29.6911530Z kw = {'encode_nominal': True, 'md5_checksum': 'ad484452702105cbf3d30f8deaba39a9', 'parse_arff': <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>, 'return_type': 3}
2021-03-01T10:30:29.6913030Z local_path = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6914100Z openml_path = 'data/v1/download/61'
2021-03-01T10:30:29.6914780Z ../1/s/sklearn/datasets/_openml.py:504: in _load_arff_response
2021-03-01T10:30:29.6915550Z response = _open_openml_url(url, data_home)
2021-03-01T10:30:29.6916580Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.6917230Z encode_nominal = True
2021-03-01T10:30:29.6918190Z md5_checksum = 'ad484452702105cbf3d30f8deaba39a9'
2021-03-01T10:30:29.6918940Z parse_arff = <function _download_data_to_bunch.<locals>.parse_arff at 0x12f4c2d30>
2021-03-01T10:30:29.6919590Z return_type = 3
2021-03-01T10:30:29.6920480Z url = 'data/v1/download/61'
2021-03-01T10:30:29.6921150Z ../1/s/sklearn/datasets/_openml.py:130: in _open_openml_url
2021-03-01T10:30:29.6922130Z return gzip.GzipFile(local_path, 'rb')
2021-03-01T10:30:29.6923150Z data_home = '/Users/runner/scikit_learn_data/openml'
2021-03-01T10:30:29.6924360Z fdst = <_io.BufferedWriter name='/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'>
2021-03-01T10:30:29.6925250Z fsrc = <sklearn.datasets.tests.test_openml._MockHTTPResponse object at 0x12f4de580>
2021-03-01T10:30:29.6926050Z is_gzip_encoded = <function _open_openml_url.<locals>.is_gzip_encoded at 0x12f4c2c10>
2021-03-01T10:30:29.6927370Z local_path = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6928440Z opener = <built-in function open>
2021-03-01T10:30:29.6929420Z openml_path = 'data/v1/download/61'
2021-03-01T10:30:29.6930100Z req = <urllib.request.Request object at 0x12f4de0a0>
2021-03-01T10:30:29.6930820Z _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
2021-03-01T10:30:29.6931440Z
2021-03-01T10:30:29.6932530Z self = <[AttributeError("'GzipFile' object has no attribute 'fileobj'") raised in repr()] GzipFile object at 0x12f4defa0>
2021-03-01T10:30:29.6933770Z filename = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6934890Z mode = 'rb', compresslevel = 9, fileobj = None, mtime = None
2021-03-01T10:30:29.6935320Z
2021-03-01T10:30:29.6935850Z def __init__(self, filename=None, mode=None,
2021-03-01T10:30:29.6936560Z compresslevel=_COMPRESS_LEVEL_BEST, fileobj=None, mtime=None):
2021-03-01T10:30:29.6937260Z """Constructor for the GzipFile class.
2021-03-01T10:30:29.6937790Z
2021-03-01T10:30:29.6938340Z At least one of fileobj and filename must be given a
2021-03-01T10:30:29.6939350Z non-trivial value.
2021-03-01T10:30:29.6939870Z
2021-03-01T10:30:29.6940440Z The new class instance is based on fileobj, which can be a regular
2021-03-01T10:30:29.6941160Z file, an io.BytesIO object, or any other object which simulates a file.
2021-03-01T10:30:29.6941880Z It defaults to None, in which case filename is opened to provide
2021-03-01T10:30:29.6942500Z a file object.
2021-03-01T10:30:29.6942980Z
2021-03-01T10:30:29.6943890Z When fileobj is not None, the filename argument is only used to be
2021-03-01T10:30:29.6944610Z included in the gzip file header, which may include the original
2021-03-01T10:30:29.6945310Z filename of the uncompressed file. It defaults to the filename of
2021-03-01T10:30:29.6946030Z fileobj, if discernible; otherwise, it defaults to the empty string,
2021-03-01T10:30:29.6946750Z and in this case the original filename is not included in the header.
2021-03-01T10:30:29.6947340Z
2021-03-01T10:30:29.6948320Z The mode argument can be any of 'r', 'rb', 'a', 'ab', 'w', 'wb', 'x', or
2021-03-01T10:30:29.6949460Z 'xb' depending on whether the file will be read or written. The default
2021-03-01T10:30:29.6950510Z is the mode of fileobj if discernible; otherwise, the default is 'rb'.
2021-03-01T10:30:29.6951530Z A mode of 'r' is equivalent to one of 'rb', and similarly for 'w' and
2021-03-01T10:30:29.6952520Z 'wb', 'a' and 'ab', and 'x' and 'xb'.
2021-03-01T10:30:29.6953050Z
2021-03-01T10:30:29.6953540Z The compresslevel argument is an integer from 0 to 9 controlling the
2021-03-01T10:30:29.6954200Z level of compression; 1 is fastest and produces the least compression,
2021-03-01T10:30:29.6954920Z and 9 is slowest and produces the most compression. 0 is no compression
2021-03-01T10:30:29.6955550Z at all. The default is 9.
2021-03-01T10:30:29.6956070Z
2021-03-01T10:30:29.6956640Z The mtime argument is an optional numeric timestamp to be written
2021-03-01T10:30:29.6957350Z to the last modification time field in the stream when compressing.
2021-03-01T10:30:29.6958020Z If omitted or None, the current time is used.
2021-03-01T10:30:29.6958560Z
2021-03-01T10:30:29.6959020Z """
2021-03-01T10:30:29.6959470Z
2021-03-01T10:30:29.6960370Z if mode and ('t' in mode or 'U' in mode):
2021-03-01T10:30:29.6961100Z raise ValueError("Invalid mode: {!r}".format(mode))
2021-03-01T10:30:29.6962100Z if mode and 'b' not in mode:
2021-03-01T10:30:29.6963030Z mode += 'b'
2021-03-01T10:30:29.6963610Z if fileobj is None:
2021-03-01T10:30:29.6964600Z > fileobj = self.myfileobj = builtins.open(filename, mode or 'rb')
2021-03-01T10:30:29.6965920Z E FileNotFoundError: [Errno 2] No such file or directory: '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6966540Z
2021-03-01T10:30:29.6967010Z compresslevel = 9
2021-03-01T10:30:29.6968010Z filename = '/Users/runner/scikit_learn_data/openml/openml.org/data/v1/download/61.gz'
2021-03-01T10:30:29.6968680Z fileobj = None
2021-03-01T10:30:29.6969500Z mode = 'rb'
2021-03-01T10:30:29.6970250Z mtime = None
2021-03-01T10:30:29.6971380Z self = <[AttributeError("'GzipFile' object has no attribute 'fileobj'") raised in repr()] GzipFile object at 0x12f4defa0>
2021-03-01T10:30:29.6971920Z
2021-03-01T10:30:29.6972520Z /usr/local/miniconda/envs/testvenv/lib/python3.9/gzip.py:173: FileNotFoundError
```
</details>
https://dev.azure.com/scikit-learn/scikit-learn/_build/results?buildId=27014&view=logs&j=97641769-79fb-5590-9088-a30ce9b850b9&t=4745baa1-36b5-56c8-9a8e-6480742db1a6&l=796
Not sure if this is related to the retry mechanism of #12526 or not. | non_main | random failure in fetch openml tests it appended once on the ci of the unrelated pr when running test fetch openml iris python traceback s sklearn externals arff py e sklearn externals arff badlayout invalid layout of the arff file at line during handling of the above exception another exception occurred filenotfounderror no such file or directory users runner scikit learn data openml openml org data download gz full traceback darwin python usr local miniconda envs testvenv bin python args data download users runner scikit learn data openml kw encode nominal true checksum parse arff parse arff at return type local path users runner scikit learn data openml openml org data download gz wraps f def wrapper args kw if data home is none return f args kw try return f args kw args data download users runner scikit learn data openml data home users runner scikit learn data openml f kw encode nominal true checksum parse arff parse arff at return type local path users runner scikit learn data openml openml org data download gz openml path data download s sklearn datasets openml py url data download data home users runner scikit learn data openml return type encode nominal true parse arff parse arff at checksum def load arff response url str data home optional return type encode nominal bool parse arff callable tuple checksum str tuple load arff data with url and parses arff response with parse arff response open openml url url data home with closing response note that if the data is dense no reading is done until the data generator is iterated actual checksum hashlib def stream checksum generator response for line in response actual checksum update line yield line decode utf stream stream checksum generator response arff arff load stream return type return type encode nominal encode nominal stream checksum generator stream checksum generator at actual checksum data home users runner scikit learn data openml encode nominal true checksum parse arff parse arff at response return type stream stream checksum generator at url data download s sklearn datasets openml py fp stream checksum generator at encode nominal true return type def load fp encode nominal false return type dense load a file like object containing the arff document and convert it into a python object param fp a file like object param encode nominal boolean if true perform a label encoding while reading the arff file param return type determines the data structure used to store the dataset can be one of arff dense arff coo arff lod arff dense gen or arff lod gen consult the sections on working with sparse data and loading progressively return a dictionary decoder arffdecoder return decoder decode fp encode nominal encode nominal return type return type decoder encode nominal true fp stream checksum generator at return type s sklearn externals arff py self s stream checksum generator at encode nominal true return type def decode self s encode nominal false return type dense returns the python representation of a given arff file when a file object is passed as an argument this method reads lines iteratively avoiding to load unnecessary information to the memory param s a string or file object with the arff file param encode nominal boolean if true perform a label encoding while reading the arff file param return type determines the data structure used to store the dataset can be one of arff dense arff coo arff lod arff dense gen or arff lod gen consult the sections on working with sparse data and loading progressively try return self decode s encode nominal encode nominal matrix type return type except arffexception as e e line self current line raise e encode nominal true return type s stream checksum generator at self s sklearn externals arff py self s stream checksum generator at encode nominal true return type def decode self s encode nominal false return type dense returns the python representation of a given arff file when a file object is passed as an argument this method reads lines iteratively avoiding to load unnecessary information to the memory param s a string or file object with the arff file param encode nominal boolean if true perform a label encoding while reading the arff file param return type determines the data structure used to store the dataset can be one of arff dense arff coo arff lod arff dense gen or arff lod gen consult the sections on working with sparse data and loading progressively try return self decode s encode nominal encode nominal matrix type return type encode nominal true return type s stream checksum generator at self s sklearn externals arff py self s stream checksum generator at encode nominal true matrix type def decode self s encode nominal false matrix type dense do the job the encode make sure this method is idempotent self current line if string convert to a list of lines if isinstance s basestring s s strip r n replace r n n split n create the return object obj arffcontainertype u description u u relation u u attributes u data attribute names create the data helper object data get data object for decoding matrix type read all lines state tk description s iter s for row in s self current line ignore empty lines row row strip r n if not row continue u row row upper description if u row startswith tk description and state tk description obj self decode comment row n relation elif u row startswith tk relation if state tk description raise badlayout state tk relation obj self decode relation row attribute elif u row startswith tk attribute if state tk relation and state tk attribute raise badlayout state tk attribute attr self decode attribute row if attr in attribute names raise badattributename attr attribute names else attribute names self current line obj append attr if isinstance attr list tuple if encode nominal conversor encodednominalconversor attr else conversor nominalconversor attr else conversor map string unicode integer lambda x int float x numeric float real float conversor conversor map self conversors append conversor data elif u row startswith tk data if state tk attribute raise badlayout break comment elif u row startswith tk comment pass else never found data raise badlayout e sklearn externals arff badlayout invalid layout of the arff file at line state attribute names data encode nominal true matrix type obj attributes data description relation s stream checksum generator at self s sklearn externals arff py badlayout during handling of the above exception another exception occurred monkeypatch gzip response true pytest mark parametrize gzip response def test fetch openml iris monkeypatch gzip response classification dataset with numeric only columns data id data name iris monkey patch webbased functions monkeypatch data id gzip response assert warns message userwarning multiple active versions of the dataset matching the name iris exist versions may be fundamentally different returning version fetch openml name data name as frame false data id data name iris gzip response true monkeypatch s sklearn datasets tests test openml py s sklearn utils testing py in assert warns message result func args kw args func kw as frame false name iris message multiple active versions of the dataset matching the name iris exist versions may be fundamentally different returning version w warning class s sklearn utils validation py in inner f return f args kwargs all args args extra args f kwargs as frame false name iris kwonly args sig version renaming of s sklearn datasets openml py in fetch openml bunch download data to bunch url return sparse data home as frame false cache true data columns data description collection date creator r a fisher default target attribute class description author petal width in cm n class n iris setosa n iris versicolour n iris virginica data home users runner scikit learn data openml data id data info did file id format arff name iris data qualities feature data type nominal index is ignore false is row identifier false features list name iris return x y false return sparse false shape target column default target target columns url data download version active s sklearn datasets openml py in download data to bunch out retry with clean cache url data home as frame false col idx col slice x col slice y data columns data home users runner scikit learn data openml feat data type nominal index is ignore false is row identifier false features dict class data type nominal index is ignore false is row identifier false petallengt sepallength data type numeric index is ignore false is row identifier false features list frame none checksum nominal attributes none nr missing parse arff parse arff at postprocess postprocess at return type shape sparse false target columns url data download s sklearn datasets openml py in wrapper return f args kw args data download users runner scikit learn data openml data home users runner scikit learn data openml f kw encode nominal true checksum parse arff parse arff at return type local path users runner scikit learn data openml openml org data download gz openml path data download s sklearn datasets openml py in load arff response response open openml url url data home data home users runner scikit learn data openml encode nominal true checksum parse arff parse arff at return type url data download s sklearn datasets openml py in open openml url return gzip gzipfile local path rb data home users runner scikit learn data openml fdst fsrc is gzip encoded is gzip encoded at local path users runner scikit learn data openml openml org data download gz opener openml path data download req self filename users runner scikit learn data openml openml org data download gz mode rb compresslevel fileobj none mtime none def init self filename none mode none compresslevel compress level best fileobj none mtime none constructor for the gzipfile class at least one of fileobj and filename must be given a non trivial value the new class instance is based on fileobj which can be a regular file an io bytesio object or any other object which simulates a file it defaults to none in which case filename is opened to provide a file object when fileobj is not none the filename argument is only used to be included in the gzip file header which may include the original filename of the uncompressed file it defaults to the filename of fileobj if discernible otherwise it defaults to the empty string and in this case the original filename is not included in the header the mode argument can be any of r rb a ab w wb x or xb depending on whether the file will be read or written the default is the mode of fileobj if discernible otherwise the default is rb a mode of r is equivalent to one of rb and similarly for w and wb a and ab and x and xb the compresslevel argument is an integer from to controlling the level of compression is fastest and produces the least compression and is slowest and produces the most compression is no compression at all the default is the mtime argument is an optional numeric timestamp to be written to the last modification time field in the stream when compressing if omitted or none the current time is used if mode and t in mode or u in mode raise valueerror invalid mode r format mode if mode and b not in mode mode b if fileobj is none fileobj self myfileobj builtins open filename mode or rb e filenotfounderror no such file or directory users runner scikit learn data openml openml org data download gz compresslevel filename users runner scikit learn data openml openml org data download gz fileobj none mode rb mtime none self usr local miniconda envs testvenv lib gzip py filenotfounderror not sure if this is related to the retry mechanism of or not | 0 |
3,464 | 13,265,897,906 | IssuesEvent | 2020-08-21 07:29:34 | pace/bricks | https://api.github.com/repos/pace/bricks | opened | Fix current linter issues | S::Ready T::Maintainance | Running `make lint` throws a lot of warnings. We probably want to fix/address them.
```
$ make lint
/Users/juliusfoitzik/go/bin/golangci-lint run
maintenance/log/handler.go:88:5: `noXid` is unused (deadcode)
var noXid = errors.New("no xid")
^
http/jsonapi/generator/generate_helper.go:16:7: `pkgRuntime` is unused (deadcode)
const pkgRuntime = "github.com/pace/bricks/http/jsonapi/runtime"
^
http/jsonapi/errors_test.go:53:18: Error return value of `json.Unmarshal` is not checked (errcheck)
json.Unmarshal(buffer.Bytes(), &output)
^
http/jsonapi/request.go:277:32: Error return value of `(*encoding/json.Encoder).Encode` is not checked (errcheck)
json.NewEncoder(buf).Encode(data.Relationships[args[1]])
^
http/jsonapi/request.go:278:32: Error return value of `(*encoding/json.Decoder).Decode` is not checked (errcheck)
json.NewDecoder(buf).Decode(relationship)
^
http/jsonapi/request.go:305:32: Error return value of `(*encoding/json.Encoder).Encode` is not checked (errcheck)
json.NewEncoder(buf).Encode(
^
http/jsonapi/request.go:308:32: Error return value of `(*encoding/json.Decoder).Decode` is not checked (errcheck)
json.NewDecoder(buf).Decode(relationship)
^
http/jsonapi/request_test.go:364:28: Error return value of `(*encoding/json.Encoder).Encode` is not checked (errcheck)
json.NewEncoder(in).Encode(payload)
^
http/jsonapi/request_test.go:1097:16: Error return value of `MarshalPayload` is not checked (errcheck)
MarshalPayload(out, testModel)
^
http/jsonapi/request_test.go:1104:27: Error return value of `MarshalOnePayloadEmbedded` is not checked (errcheck)
MarshalOnePayloadEmbedded(out, testModel())
^
http/jsonapi/request_test.go:1107:18: Error return value of `UnmarshalPayload` is not checked (errcheck)
UnmarshalPayload(out, blog)
^
http/jsonapi/response_test.go:30:16: Error return value of `MarshalPayload` is not checked (errcheck)
MarshalPayload(out1, book)
^
http/jsonapi/response_test.go:46:16: Error return value of `MarshalPayload` is not checked (errcheck)
MarshalPayload(out2, books)
^
http/jsonapi/runtime.go:83:18: Error return value of `r.instrumentCall` is not checked (errcheck)
r.instrumentCall(UnmarshalStart, UnmarshalStop, func() error {
^
maintenance/errors/raven/client.go:374:15: Error return value of `client.SetDSN` is not checked (errcheck)
client.SetDSN(os.Getenv("SENTRY_DSN"))
^
maintenance/errors/raven/client.go:956:9: Error return value of `io.Copy` is not checked (errcheck)
io.Copy(ioutil.Discard, res.Body)
^
maintenance/errors/raven/client.go:976:16: Error return value of `deflate.Write` is not checked (errcheck)
deflate.Write(packetJSON)
^
http/jsonapi/generator/generate_security.go:115:20: Error return value of `json.Unmarshal` is not checked (errcheck)
json.Unmarshal(data, &url)
^
http/transport/chainable_test.go:37:14: Error return value of `client.Get` is not checked (errcheck)
client.Get(server.URL + "/test001")
^
http/transport/chainable_test.go:42:13: Error return value of `client.Get` is not checked (errcheck)
client.Get(server.URL + "/test002")
^
http/transport/locale_round_tripper_test.go:35:15: Error return value of `lrt.RoundTrip` is not checked (errcheck)
lrt.RoundTrip(r.WithContext(locale.WithLocale(context.Background(), l)))
^
http/jsonapi/runtime/standart_params_test.go:100:14: Error return value of `db.DropTable` is not checked (errcheck)
db.DropTable(&TestModel{}, &orm.DropTableOptions{
^
maintenance/errors/raven/client.go:341:3: ineffectual assignment to `i` (ineffassign)
i++
^
http/jsonapi/runtime/standart_params_test.go:69:13: ineffectual assignment to `err` (ineffassign)
urlParams, err = runtime.ReadURLQueryParameters(r, mapper, &testValueSanitizer{})
^
http/jsonapi/runtime/standart_params_test.go:79:13: ineffectual assignment to `err` (ineffassign)
urlParams, err = runtime.ReadURLQueryParameters(r, mapper, &testValueSanitizer{})
^
http/jsonapi/runtime/standart_params_test.go:92:13: ineffectual assignment to `err` (ineffassign)
urlParams, err = runtime.ReadURLQueryParameters(r, mapper, &testValueSanitizer{})
^
maintenance/errors/raven/stacktrace.go:31:6: S1002: should omit comparison to bool constant, can be simplified to `frame.InApp` (gosimple)
if frame.InApp == true && frame.Module != "" && frame.Function != "" {
^
http/jsonapi/response_test.go:42:14: S1030: should use out1.String() instead of string(out1.Bytes()) (gosimple)
fmt.Println(string(out1.Bytes()))
^
http/longpoll/longpoll_test.go:17:41: S1024: should use time.Until instead of t.Sub(time.Now()) (gosimple)
assert.Equal(t, time.Millisecond*999, budget.Sub(time.Now()).Truncate(time.Millisecond))
^
http/longpoll/longpoll_test.go:29:35: S1024: should use time.Until instead of t.Sub(time.Now()) (gosimple)
assert.Equal(t, time.Second*59, budget.Sub(time.Now()).Truncate(time.Second))
^
maintenance/errors/context_test.go:21:19: lostcancel: the cancel function returned by context.WithTimeout should be called, not discarded, to avoid a context leak (govet)
exceededContext, _ := context.WithTimeout(backgroundContext, time.Millisecond)
^
http/jsonapi/runtime/error.go:119:4: SA4006: this value of `ev` is never used (staticcheck)
ev = ev.Fields(*source)
^
http/jsonapi/request_test.go:205:7: type `FooStruct` is unused (unused)
make: *** [lint] Error 1
``` | True | Fix current linter issues - Running `make lint` throws a lot of warnings. We probably want to fix/address them.
```
$ make lint
/Users/juliusfoitzik/go/bin/golangci-lint run
maintenance/log/handler.go:88:5: `noXid` is unused (deadcode)
var noXid = errors.New("no xid")
^
http/jsonapi/generator/generate_helper.go:16:7: `pkgRuntime` is unused (deadcode)
const pkgRuntime = "github.com/pace/bricks/http/jsonapi/runtime"
^
http/jsonapi/errors_test.go:53:18: Error return value of `json.Unmarshal` is not checked (errcheck)
json.Unmarshal(buffer.Bytes(), &output)
^
http/jsonapi/request.go:277:32: Error return value of `(*encoding/json.Encoder).Encode` is not checked (errcheck)
json.NewEncoder(buf).Encode(data.Relationships[args[1]])
^
http/jsonapi/request.go:278:32: Error return value of `(*encoding/json.Decoder).Decode` is not checked (errcheck)
json.NewDecoder(buf).Decode(relationship)
^
http/jsonapi/request.go:305:32: Error return value of `(*encoding/json.Encoder).Encode` is not checked (errcheck)
json.NewEncoder(buf).Encode(
^
http/jsonapi/request.go:308:32: Error return value of `(*encoding/json.Decoder).Decode` is not checked (errcheck)
json.NewDecoder(buf).Decode(relationship)
^
http/jsonapi/request_test.go:364:28: Error return value of `(*encoding/json.Encoder).Encode` is not checked (errcheck)
json.NewEncoder(in).Encode(payload)
^
http/jsonapi/request_test.go:1097:16: Error return value of `MarshalPayload` is not checked (errcheck)
MarshalPayload(out, testModel)
^
http/jsonapi/request_test.go:1104:27: Error return value of `MarshalOnePayloadEmbedded` is not checked (errcheck)
MarshalOnePayloadEmbedded(out, testModel())
^
http/jsonapi/request_test.go:1107:18: Error return value of `UnmarshalPayload` is not checked (errcheck)
UnmarshalPayload(out, blog)
^
http/jsonapi/response_test.go:30:16: Error return value of `MarshalPayload` is not checked (errcheck)
MarshalPayload(out1, book)
^
http/jsonapi/response_test.go:46:16: Error return value of `MarshalPayload` is not checked (errcheck)
MarshalPayload(out2, books)
^
http/jsonapi/runtime.go:83:18: Error return value of `r.instrumentCall` is not checked (errcheck)
r.instrumentCall(UnmarshalStart, UnmarshalStop, func() error {
^
maintenance/errors/raven/client.go:374:15: Error return value of `client.SetDSN` is not checked (errcheck)
client.SetDSN(os.Getenv("SENTRY_DSN"))
^
maintenance/errors/raven/client.go:956:9: Error return value of `io.Copy` is not checked (errcheck)
io.Copy(ioutil.Discard, res.Body)
^
maintenance/errors/raven/client.go:976:16: Error return value of `deflate.Write` is not checked (errcheck)
deflate.Write(packetJSON)
^
http/jsonapi/generator/generate_security.go:115:20: Error return value of `json.Unmarshal` is not checked (errcheck)
json.Unmarshal(data, &url)
^
http/transport/chainable_test.go:37:14: Error return value of `client.Get` is not checked (errcheck)
client.Get(server.URL + "/test001")
^
http/transport/chainable_test.go:42:13: Error return value of `client.Get` is not checked (errcheck)
client.Get(server.URL + "/test002")
^
http/transport/locale_round_tripper_test.go:35:15: Error return value of `lrt.RoundTrip` is not checked (errcheck)
lrt.RoundTrip(r.WithContext(locale.WithLocale(context.Background(), l)))
^
http/jsonapi/runtime/standart_params_test.go:100:14: Error return value of `db.DropTable` is not checked (errcheck)
db.DropTable(&TestModel{}, &orm.DropTableOptions{
^
maintenance/errors/raven/client.go:341:3: ineffectual assignment to `i` (ineffassign)
i++
^
http/jsonapi/runtime/standart_params_test.go:69:13: ineffectual assignment to `err` (ineffassign)
urlParams, err = runtime.ReadURLQueryParameters(r, mapper, &testValueSanitizer{})
^
http/jsonapi/runtime/standart_params_test.go:79:13: ineffectual assignment to `err` (ineffassign)
urlParams, err = runtime.ReadURLQueryParameters(r, mapper, &testValueSanitizer{})
^
http/jsonapi/runtime/standart_params_test.go:92:13: ineffectual assignment to `err` (ineffassign)
urlParams, err = runtime.ReadURLQueryParameters(r, mapper, &testValueSanitizer{})
^
maintenance/errors/raven/stacktrace.go:31:6: S1002: should omit comparison to bool constant, can be simplified to `frame.InApp` (gosimple)
if frame.InApp == true && frame.Module != "" && frame.Function != "" {
^
http/jsonapi/response_test.go:42:14: S1030: should use out1.String() instead of string(out1.Bytes()) (gosimple)
fmt.Println(string(out1.Bytes()))
^
http/longpoll/longpoll_test.go:17:41: S1024: should use time.Until instead of t.Sub(time.Now()) (gosimple)
assert.Equal(t, time.Millisecond*999, budget.Sub(time.Now()).Truncate(time.Millisecond))
^
http/longpoll/longpoll_test.go:29:35: S1024: should use time.Until instead of t.Sub(time.Now()) (gosimple)
assert.Equal(t, time.Second*59, budget.Sub(time.Now()).Truncate(time.Second))
^
maintenance/errors/context_test.go:21:19: lostcancel: the cancel function returned by context.WithTimeout should be called, not discarded, to avoid a context leak (govet)
exceededContext, _ := context.WithTimeout(backgroundContext, time.Millisecond)
^
http/jsonapi/runtime/error.go:119:4: SA4006: this value of `ev` is never used (staticcheck)
ev = ev.Fields(*source)
^
http/jsonapi/request_test.go:205:7: type `FooStruct` is unused (unused)
make: *** [lint] Error 1
``` | main | fix current linter issues running make lint throws a lot of warnings we probably want to fix address them make lint users juliusfoitzik go bin golangci lint run maintenance log handler go noxid is unused deadcode var noxid errors new no xid http jsonapi generator generate helper go pkgruntime is unused deadcode const pkgruntime github com pace bricks http jsonapi runtime http jsonapi errors test go error return value of json unmarshal is not checked errcheck json unmarshal buffer bytes output http jsonapi request go error return value of encoding json encoder encode is not checked errcheck json newencoder buf encode data relationships http jsonapi request go error return value of encoding json decoder decode is not checked errcheck json newdecoder buf decode relationship http jsonapi request go error return value of encoding json encoder encode is not checked errcheck json newencoder buf encode http jsonapi request go error return value of encoding json decoder decode is not checked errcheck json newdecoder buf decode relationship http jsonapi request test go error return value of encoding json encoder encode is not checked errcheck json newencoder in encode payload http jsonapi request test go error return value of marshalpayload is not checked errcheck marshalpayload out testmodel http jsonapi request test go error return value of marshalonepayloadembedded is not checked errcheck marshalonepayloadembedded out testmodel http jsonapi request test go error return value of unmarshalpayload is not checked errcheck unmarshalpayload out blog http jsonapi response test go error return value of marshalpayload is not checked errcheck marshalpayload book http jsonapi response test go error return value of marshalpayload is not checked errcheck marshalpayload books http jsonapi runtime go error return value of r instrumentcall is not checked errcheck r instrumentcall unmarshalstart unmarshalstop func error maintenance errors raven client go error return value of client setdsn is not checked errcheck client setdsn os getenv sentry dsn maintenance errors raven client go error return value of io copy is not checked errcheck io copy ioutil discard res body maintenance errors raven client go error return value of deflate write is not checked errcheck deflate write packetjson http jsonapi generator generate security go error return value of json unmarshal is not checked errcheck json unmarshal data url http transport chainable test go error return value of client get is not checked errcheck client get server url http transport chainable test go error return value of client get is not checked errcheck client get server url http transport locale round tripper test go error return value of lrt roundtrip is not checked errcheck lrt roundtrip r withcontext locale withlocale context background l http jsonapi runtime standart params test go error return value of db droptable is not checked errcheck db droptable testmodel orm droptableoptions maintenance errors raven client go ineffectual assignment to i ineffassign i http jsonapi runtime standart params test go ineffectual assignment to err ineffassign urlparams err runtime readurlqueryparameters r mapper testvaluesanitizer http jsonapi runtime standart params test go ineffectual assignment to err ineffassign urlparams err runtime readurlqueryparameters r mapper testvaluesanitizer http jsonapi runtime standart params test go ineffectual assignment to err ineffassign urlparams err runtime readurlqueryparameters r mapper testvaluesanitizer maintenance errors raven stacktrace go should omit comparison to bool constant can be simplified to frame inapp gosimple if frame inapp true frame module frame function http jsonapi response test go should use string instead of string bytes gosimple fmt println string bytes http longpoll longpoll test go should use time until instead of t sub time now gosimple assert equal t time millisecond budget sub time now truncate time millisecond http longpoll longpoll test go should use time until instead of t sub time now gosimple assert equal t time second budget sub time now truncate time second maintenance errors context test go lostcancel the cancel function returned by context withtimeout should be called not discarded to avoid a context leak govet exceededcontext context withtimeout backgroundcontext time millisecond http jsonapi runtime error go this value of ev is never used staticcheck ev ev fields source http jsonapi request test go type foostruct is unused unused make error | 1 |
2,074 | 7,025,597,339 | IssuesEvent | 2017-12-23 13:12:56 | sapcc/hermes | https://api.github.com/repos/sapcc/hermes | closed | EventTime Format doesn't match spec | enhancement maintainability | current - 2017-07-20T13:43:38.368933+0000
desired - 2017-07-20T13:43:38.368933+00:00
check what the value is in kibana, if it's fine in kiabana, problem is likely in json translation to type.
if it's not fine in kibana, then either it needs changed on openstack middleware side, or just replaced as the return value for the api. | True | EventTime Format doesn't match spec - current - 2017-07-20T13:43:38.368933+0000
desired - 2017-07-20T13:43:38.368933+00:00
check what the value is in kibana, if it's fine in kiabana, problem is likely in json translation to type.
if it's not fine in kibana, then either it needs changed on openstack middleware side, or just replaced as the return value for the api. | main | eventtime format doesn t match spec current desired check what the value is in kibana if it s fine in kiabana problem is likely in json translation to type if it s not fine in kibana then either it needs changed on openstack middleware side or just replaced as the return value for the api | 1 |
3,560 | 14,243,597,562 | IssuesEvent | 2020-11-19 04:54:26 | adda-team/adda | https://api.github.com/repos/adda-team/adda | opened | Test ocl-icd on cluster | OS-Linux OpenCL comp-Docs comp-Library maintainability pri-Medium | Current description of OpenCL usage is already extensive - https://github.com/adda-team/adda/wiki/OpenCL . However, for Linux the simplest options are through the packages, which are not necessarily available in the cluster environment (without root access).
In particular, currently we use standard Nvidia drivers as a source of opencl.so, which causes problems with clFFT linking (although there is a viable workaround).
So the idea is to test ocl-icd on cluster (probably, compile from source at `$HOME`) and update the documentation.
| True | Test ocl-icd on cluster - Current description of OpenCL usage is already extensive - https://github.com/adda-team/adda/wiki/OpenCL . However, for Linux the simplest options are through the packages, which are not necessarily available in the cluster environment (without root access).
In particular, currently we use standard Nvidia drivers as a source of opencl.so, which causes problems with clFFT linking (although there is a viable workaround).
So the idea is to test ocl-icd on cluster (probably, compile from source at `$HOME`) and update the documentation.
| main | test ocl icd on cluster current description of opencl usage is already extensive however for linux the simplest options are through the packages which are not necessarily available in the cluster environment without root access in particular currently we use standard nvidia drivers as a source of opencl so which causes problems with clfft linking although there is a viable workaround so the idea is to test ocl icd on cluster probably compile from source at home and update the documentation | 1 |
94,066 | 19,475,241,811 | IssuesEvent | 2021-12-24 10:51:34 | joomla/joomla-cms | https://api.github.com/repos/joomla/joomla-cms | closed | Odd functionality when I upload an image | No Code Attached Yet | ### Steps to reproduce the issue
In the media manager add "svg" to supported file extensions and "image/svg+xml" to the allowed mime types both.
Now go to System > templates > site template styles > Cassiopeia - Default
Upload an SVG as the logo
### Expected result
SVG URL and logo loaded
### Actual result
Logo loads but this URL arrives
/images/brand-assets/lovelaser_logo_sr.svg#joomlaImage://local-images/brand-assets/lovelaser_logo_sr.svg?width=0&height=0
### System information (as much as possible)
[systeminfo-2021-12-23T12_28_24+00_00.txt](https://issues.joomla.org/uploads/1/9ddf1dbb48599e00fa7fef82408bd491.txt)
### Additional comments
| 1.0 | Odd functionality when I upload an image - ### Steps to reproduce the issue
In the media manager add "svg" to supported file extensions and "image/svg+xml" to the allowed mime types both.
Now go to System > templates > site template styles > Cassiopeia - Default
Upload an SVG as the logo
### Expected result
SVG URL and logo loaded
### Actual result
Logo loads but this URL arrives
/images/brand-assets/lovelaser_logo_sr.svg#joomlaImage://local-images/brand-assets/lovelaser_logo_sr.svg?width=0&height=0
### System information (as much as possible)
[systeminfo-2021-12-23T12_28_24+00_00.txt](https://issues.joomla.org/uploads/1/9ddf1dbb48599e00fa7fef82408bd491.txt)
### Additional comments
| non_main | odd functionality when i upload an image steps to reproduce the issue in the media manager add svg to supported file extensions and image svg xml to the allowed mime types both now go to system templates site template styles cassiopeia default upload an svg as the logo expected result svg url and logo loaded actual result logo loads but this url arrives images brand assets lovelaser logo sr svg joomlaimage local images brand assets lovelaser logo sr svg width height system information as much as possible additional comments | 0 |
4,775 | 24,599,053,177 | IssuesEvent | 2022-10-14 10:49:47 | obs-websocket-community-projects/obs-websocket-java | https://api.github.com/repos/obs-websocket-community-projects/obs-websocket-java | closed | Requests Unit Tests | maintainability cicd | - General
- [x] GetVersion
- [x] BroadcastCustomEvent
- [x] GetSystemStats
- [x] GetHotkeyList
- [x] TriggerHotkeyByName
- [x] TriggerHotkeyByKeySequence
- [x] GetProjectorList
- [x] OpenProjector
- [x] CloseProjector
- [x] GetStudioModeEnabled
- [x] SetStudioModeEnabled
- [x] Sleep
- Config
- [x] ~~GetGlobalPersistentData~~
- [x] ~~SetGlobalPersistentData~~
- [x] GetPersistentData
- [x] SetPersistentData
- [x] GetSceneCollectionList
- [x] SetCurrentSceneCollection
- [x] CreateSceneCollection
- [x] RemoveSceneCollection
- [x] GetProfileList
- [ ] SetCurrentProfile
- [x] GetProfileParameter
- [x] SetProfileParameter
- [x] ~~GetProfilePersistentData~~
- [x] ~~SetProfilePersistentData~~
- [ ] CreateProfile
- [ ] RemoveProfile
- [x] GetVideoSettings
- [ ] SetVideoSettings
- Sources
- [ ] GetSourceList
- [x] GetSourceActive
- [x] GetSourceScreenshot
- [x] SaveSourceScreenshot
- Scenes
- [x] GetSceneList
- [x] GetCurrentProgramScene
- [x] SetCurrentProgramScene
- [x] GetCurrentPreviewScene
- [x] SetCurrentPreviewScene
- [x] SetSceneIndex
- [x] SetSceneName
- [x] CreateScene
- [x] RemoveScene
- [x] GetSceneTransitionOverride
- [x] SetSceneTransitionOverride
- [x] DeleteSceneTransitionOverride
- Inputs
- [x] GetInputList
- [x] GetInputKindList
- [ ] GetSpecialInputNames
- [x] GetInputDefaultSettings
- [x] GetInputSettings
- [x] SetInputSettings
- [x] GetInputMute
- [x] SetInputMute
- [x] ToggleInputMute
- [x] GetInputVolume
- [x] SetInputVolume
- [ ] GetInputAudioSyncOffset
- [ ] SetInputAudioSyncOffset
- [x] GetInputTracks
- [ ] SetInputTracks
- [x] GetInputMonitorType
- [x] SetInputMonitorType
- [ ] GetInputPropertiesListPropertyItems
- [ ] PressInputPropertiesButton
- [x] SetInputName
- [x] CreateInput
- [x] RemoveInput
- Transitions
- [x] GetTransitionList
- [x] GetCurrentTransition
- [x] SetCurrentTransition
- [x] SetCurrentTransitionDuration
- [ ] GetTransitionSettings
- [x] SetTransitionSettings
- [x] ReleaseTbar
- [x] SetTbarPosition
- [x] TriggerStudioModeTransition
- [ ] CreateTransition
- [ ] RemoveTransition
- Filters
- [x] GetSourceFilterList
- [ ] GetSourceFilterDefaultSettings
- [x] GetSourceFilter
- [ ] SetSourceFilterIndex
- [x] SetSourceFilterSettings
- [x] SetSourceFilterEnabled
- [x] CreateSourceFilter
- [x] RemoveSourceFilter
- Scene Items
- [x] GetSceneItemList
- [ ] GetGroupSceneItemList
- [ ] GetSceneItemTransform
- [ ] SetSceneItemTransform
- [x] GetSceneItemEnabled
- [x] SetSceneItemEnabled
- [x] GetSceneItemLocked
- [x] SetSceneItemLocked
- [x] GetSceneItemColor
- [ ] SetSceneItemColor
- [x] SetSceneItemIndex
- [x] CreateSceneItem
- [x] RemoveSceneItem
- [x] DuplicateSceneItem
- Outputs
- [ ] GetVirtualCamStatus
- [ ] ToggleVirtualCam
- [ ] StartVirtualCam
- [ ] StopVirtualCam
- [x] GetReplayBufferStatus
- [x] ToggleReplayBuffer
- [ ] StartReplayBuffer
- [x] StopReplayBuffer
- [x] SaveReplayBuffer
- [x] GetLastReplayBufferReplay
- [ ] GetReplayBufferTime
- [ ] SetReplayBufferTime
- [x] GetOutputList
- [ ] GetOutputStatus
- [x] ToggleOutput
- [x] StartOutput
- [x] StopOutput
- [ ] GetOutputSettings
- [ ] SetOutputSettings
- Stream
- [x] GetStreamStatus
- [x] ToggleStream
- [x] StartStream
- [x] StopStream
- [x] SendStreamCaption
- [ ] GetStreamBitrateSetting
- [ ] SetStreamBitrateSetting
- [x] GetStreamServiceSettings
- [x] SetStreamServiceSettings
- Record
- [x] GetRecordStatus
- [x] ToggleRecord
- [x] StartRecord
- [x] StopRecord
- [x] ToggleRecordPause
- [x] PauseRecord
- [x] ResumeRecord
- [x] GetRecordDirectory
- [x] SetRecordDirectory
- [x] GetRecordFilenameFormatting
- [x] SetRecordFilenameFormatting
- Media Inputs
- [x] GetMediaInputStatus
- [x] OffsetMediaInputTimecode
- [x] SetMediaInputTimecode
- [x] SetMediaInputPauseState
- [x] StopMediaInput
- [x] RestartMediaInput
- [x] NextMediaInputPlaylistItem
- [x] PreviousMediaInputPlaylistItem | True | Requests Unit Tests - - General
- [x] GetVersion
- [x] BroadcastCustomEvent
- [x] GetSystemStats
- [x] GetHotkeyList
- [x] TriggerHotkeyByName
- [x] TriggerHotkeyByKeySequence
- [x] GetProjectorList
- [x] OpenProjector
- [x] CloseProjector
- [x] GetStudioModeEnabled
- [x] SetStudioModeEnabled
- [x] Sleep
- Config
- [x] ~~GetGlobalPersistentData~~
- [x] ~~SetGlobalPersistentData~~
- [x] GetPersistentData
- [x] SetPersistentData
- [x] GetSceneCollectionList
- [x] SetCurrentSceneCollection
- [x] CreateSceneCollection
- [x] RemoveSceneCollection
- [x] GetProfileList
- [ ] SetCurrentProfile
- [x] GetProfileParameter
- [x] SetProfileParameter
- [x] ~~GetProfilePersistentData~~
- [x] ~~SetProfilePersistentData~~
- [ ] CreateProfile
- [ ] RemoveProfile
- [x] GetVideoSettings
- [ ] SetVideoSettings
- Sources
- [ ] GetSourceList
- [x] GetSourceActive
- [x] GetSourceScreenshot
- [x] SaveSourceScreenshot
- Scenes
- [x] GetSceneList
- [x] GetCurrentProgramScene
- [x] SetCurrentProgramScene
- [x] GetCurrentPreviewScene
- [x] SetCurrentPreviewScene
- [x] SetSceneIndex
- [x] SetSceneName
- [x] CreateScene
- [x] RemoveScene
- [x] GetSceneTransitionOverride
- [x] SetSceneTransitionOverride
- [x] DeleteSceneTransitionOverride
- Inputs
- [x] GetInputList
- [x] GetInputKindList
- [ ] GetSpecialInputNames
- [x] GetInputDefaultSettings
- [x] GetInputSettings
- [x] SetInputSettings
- [x] GetInputMute
- [x] SetInputMute
- [x] ToggleInputMute
- [x] GetInputVolume
- [x] SetInputVolume
- [ ] GetInputAudioSyncOffset
- [ ] SetInputAudioSyncOffset
- [x] GetInputTracks
- [ ] SetInputTracks
- [x] GetInputMonitorType
- [x] SetInputMonitorType
- [ ] GetInputPropertiesListPropertyItems
- [ ] PressInputPropertiesButton
- [x] SetInputName
- [x] CreateInput
- [x] RemoveInput
- Transitions
- [x] GetTransitionList
- [x] GetCurrentTransition
- [x] SetCurrentTransition
- [x] SetCurrentTransitionDuration
- [ ] GetTransitionSettings
- [x] SetTransitionSettings
- [x] ReleaseTbar
- [x] SetTbarPosition
- [x] TriggerStudioModeTransition
- [ ] CreateTransition
- [ ] RemoveTransition
- Filters
- [x] GetSourceFilterList
- [ ] GetSourceFilterDefaultSettings
- [x] GetSourceFilter
- [ ] SetSourceFilterIndex
- [x] SetSourceFilterSettings
- [x] SetSourceFilterEnabled
- [x] CreateSourceFilter
- [x] RemoveSourceFilter
- Scene Items
- [x] GetSceneItemList
- [ ] GetGroupSceneItemList
- [ ] GetSceneItemTransform
- [ ] SetSceneItemTransform
- [x] GetSceneItemEnabled
- [x] SetSceneItemEnabled
- [x] GetSceneItemLocked
- [x] SetSceneItemLocked
- [x] GetSceneItemColor
- [ ] SetSceneItemColor
- [x] SetSceneItemIndex
- [x] CreateSceneItem
- [x] RemoveSceneItem
- [x] DuplicateSceneItem
- Outputs
- [ ] GetVirtualCamStatus
- [ ] ToggleVirtualCam
- [ ] StartVirtualCam
- [ ] StopVirtualCam
- [x] GetReplayBufferStatus
- [x] ToggleReplayBuffer
- [ ] StartReplayBuffer
- [x] StopReplayBuffer
- [x] SaveReplayBuffer
- [x] GetLastReplayBufferReplay
- [ ] GetReplayBufferTime
- [ ] SetReplayBufferTime
- [x] GetOutputList
- [ ] GetOutputStatus
- [x] ToggleOutput
- [x] StartOutput
- [x] StopOutput
- [ ] GetOutputSettings
- [ ] SetOutputSettings
- Stream
- [x] GetStreamStatus
- [x] ToggleStream
- [x] StartStream
- [x] StopStream
- [x] SendStreamCaption
- [ ] GetStreamBitrateSetting
- [ ] SetStreamBitrateSetting
- [x] GetStreamServiceSettings
- [x] SetStreamServiceSettings
- Record
- [x] GetRecordStatus
- [x] ToggleRecord
- [x] StartRecord
- [x] StopRecord
- [x] ToggleRecordPause
- [x] PauseRecord
- [x] ResumeRecord
- [x] GetRecordDirectory
- [x] SetRecordDirectory
- [x] GetRecordFilenameFormatting
- [x] SetRecordFilenameFormatting
- Media Inputs
- [x] GetMediaInputStatus
- [x] OffsetMediaInputTimecode
- [x] SetMediaInputTimecode
- [x] SetMediaInputPauseState
- [x] StopMediaInput
- [x] RestartMediaInput
- [x] NextMediaInputPlaylistItem
- [x] PreviousMediaInputPlaylistItem | main | requests unit tests general getversion broadcastcustomevent getsystemstats gethotkeylist triggerhotkeybyname triggerhotkeybykeysequence getprojectorlist openprojector closeprojector getstudiomodeenabled setstudiomodeenabled sleep config getglobalpersistentdata setglobalpersistentdata getpersistentdata setpersistentdata getscenecollectionlist setcurrentscenecollection createscenecollection removescenecollection getprofilelist setcurrentprofile getprofileparameter setprofileparameter getprofilepersistentdata setprofilepersistentdata createprofile removeprofile getvideosettings setvideosettings sources getsourcelist getsourceactive getsourcescreenshot savesourcescreenshot scenes getscenelist getcurrentprogramscene setcurrentprogramscene getcurrentpreviewscene setcurrentpreviewscene setsceneindex setscenename createscene removescene getscenetransitionoverride setscenetransitionoverride deletescenetransitionoverride inputs getinputlist getinputkindlist getspecialinputnames getinputdefaultsettings getinputsettings setinputsettings getinputmute setinputmute toggleinputmute getinputvolume setinputvolume getinputaudiosyncoffset setinputaudiosyncoffset getinputtracks setinputtracks getinputmonitortype setinputmonitortype getinputpropertieslistpropertyitems pressinputpropertiesbutton setinputname createinput removeinput transitions gettransitionlist getcurrenttransition setcurrenttransition setcurrenttransitionduration gettransitionsettings settransitionsettings releasetbar settbarposition triggerstudiomodetransition createtransition removetransition filters getsourcefilterlist getsourcefilterdefaultsettings getsourcefilter setsourcefilterindex setsourcefiltersettings setsourcefilterenabled createsourcefilter removesourcefilter scene items getsceneitemlist getgroupsceneitemlist getsceneitemtransform setsceneitemtransform getsceneitemenabled setsceneitemenabled getsceneitemlocked setsceneitemlocked getsceneitemcolor setsceneitemcolor setsceneitemindex createsceneitem removesceneitem duplicatesceneitem outputs getvirtualcamstatus togglevirtualcam startvirtualcam stopvirtualcam getreplaybufferstatus togglereplaybuffer startreplaybuffer stopreplaybuffer savereplaybuffer getlastreplaybufferreplay getreplaybuffertime setreplaybuffertime getoutputlist getoutputstatus toggleoutput startoutput stopoutput getoutputsettings setoutputsettings stream getstreamstatus togglestream startstream stopstream sendstreamcaption getstreambitratesetting setstreambitratesetting getstreamservicesettings setstreamservicesettings record getrecordstatus togglerecord startrecord stoprecord togglerecordpause pauserecord resumerecord getrecorddirectory setrecorddirectory getrecordfilenameformatting setrecordfilenameformatting media inputs getmediainputstatus offsetmediainputtimecode setmediainputtimecode setmediainputpausestate stopmediainput restartmediainput nextmediainputplaylistitem previousmediainputplaylistitem | 1 |
743 | 4,349,482,374 | IssuesEvent | 2016-07-30 16:03:59 | ansible/ansible-modules-core | https://api.github.com/repos/ansible/ansible-modules-core | closed | digital_ocean droplet creation with unique_name | bug_report cloud digital_ocean waiting_on_maintainer | ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
digital_ocean
##### ANSIBLE VERSION
```
ansible 2.0.2.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
default
##### OS / ENVIRONMENT
CentOS: 7.1
python: 2.7.5
dopy: 0.3.6
##### SUMMARY
If a DigitalOcean account has more that 20 existing droplets the unique_man=yes option doesn't match any droplets over this limit and creates new droplets as per default.
##### STEPS TO REPRODUCE
Create 20+ uniquely named droplets.
```
curl -X GET -H "Content-Type: application/json" -H "Authorization: Bearer 123" "https://api.digitalocean.com/v2/droplets
{"droplets":[{"id":
**limited to 20 droplets**
,"tags":[]}],"links":{"pages":{"last":"https://api.digitalocean.com/v2/droplets?page=2","next":"https://api.digitalocean.com/v2/droplets?page=2"}},"meta":{"total":21}}
```
Create a new droplet and re-run the command.
```
- name: ensure unique-droplet-name droplets exists
digital_ocean: >
state=present
command=droplet
name="unique-droplet-name"
unique_name=yes
size_id="512mb"
region_id="lon1"
image_id="123"
ssh_key_ids="123"
api_token="123"
```
##### EXPECTED RESULTS
The first request should have created a new droplet as per normal. The second request should have found the existing droplet and skipped the request.
##### ACTUAL RESULTS
A new droplet with the same name as the original was created.
##### SUGGESTED FIX
Make use of the links and meta section of the API response to request multiple pages.
```
"https://api.digitalocean.com/v2/droplets
{"droplets":[{"id":
**limited to 20 droplets**
,"tags":[]}],"links":{"pages":{"last":"https://api.digitalocean.com/v2/droplets?page=2","next":"https://api.digitalocean.com/v2/droplets?page=2"}},"meta":{"total":23}}
``` | True | digital_ocean droplet creation with unique_name - ##### ISSUE TYPE
- Bug Report
##### COMPONENT NAME
digital_ocean
##### ANSIBLE VERSION
```
ansible 2.0.2.0
config file = /etc/ansible/ansible.cfg
configured module search path = Default w/o overrides
```
##### CONFIGURATION
default
##### OS / ENVIRONMENT
CentOS: 7.1
python: 2.7.5
dopy: 0.3.6
##### SUMMARY
If a DigitalOcean account has more that 20 existing droplets the unique_man=yes option doesn't match any droplets over this limit and creates new droplets as per default.
##### STEPS TO REPRODUCE
Create 20+ uniquely named droplets.
```
curl -X GET -H "Content-Type: application/json" -H "Authorization: Bearer 123" "https://api.digitalocean.com/v2/droplets
{"droplets":[{"id":
**limited to 20 droplets**
,"tags":[]}],"links":{"pages":{"last":"https://api.digitalocean.com/v2/droplets?page=2","next":"https://api.digitalocean.com/v2/droplets?page=2"}},"meta":{"total":21}}
```
Create a new droplet and re-run the command.
```
- name: ensure unique-droplet-name droplets exists
digital_ocean: >
state=present
command=droplet
name="unique-droplet-name"
unique_name=yes
size_id="512mb"
region_id="lon1"
image_id="123"
ssh_key_ids="123"
api_token="123"
```
##### EXPECTED RESULTS
The first request should have created a new droplet as per normal. The second request should have found the existing droplet and skipped the request.
##### ACTUAL RESULTS
A new droplet with the same name as the original was created.
##### SUGGESTED FIX
Make use of the links and meta section of the API response to request multiple pages.
```
"https://api.digitalocean.com/v2/droplets
{"droplets":[{"id":
**limited to 20 droplets**
,"tags":[]}],"links":{"pages":{"last":"https://api.digitalocean.com/v2/droplets?page=2","next":"https://api.digitalocean.com/v2/droplets?page=2"}},"meta":{"total":23}}
``` | main | digital ocean droplet creation with unique name issue type bug report component name digital ocean ansible version ansible config file etc ansible ansible cfg configured module search path default w o overrides configuration default os environment centos python dopy summary if a digitalocean account has more that existing droplets the unique man yes option doesn t match any droplets over this limit and creates new droplets as per default steps to reproduce create uniquely named droplets curl x get h content type application json h authorization bearer droplets id limited to droplets tags links pages last create a new droplet and re run the command name ensure unique droplet name droplets exists digital ocean state present command droplet name unique droplet name unique name yes size id region id image id ssh key ids api token expected results the first request should have created a new droplet as per normal the second request should have found the existing droplet and skipped the request actual results a new droplet with the same name as the original was created suggested fix make use of the links and meta section of the api response to request multiple pages droplets id limited to droplets tags links pages last | 1 |
2,537 | 8,658,329,529 | IssuesEvent | 2018-11-28 00:29:46 | GItThatSparrow/arduino-cli-on | https://api.github.com/repos/GItThatSparrow/arduino-cli-on | opened | Read Configuration from arduino-cli | Feature: Maintain `arduino-cli` | *AS A* USER
*I NEED TO* add the proper commands to parse the results from the `arduino-cli` configuration
*SO THAT* I can use the value stored in the `arduino_data` to locate the `Arduino15` directory
| True | Read Configuration from arduino-cli - *AS A* USER
*I NEED TO* add the proper commands to parse the results from the `arduino-cli` configuration
*SO THAT* I can use the value stored in the `arduino_data` to locate the `Arduino15` directory
| main | read configuration from arduino cli as a user i need to add the proper commands to parse the results from the arduino cli configuration so that i can use the value stored in the arduino data to locate the directory | 1 |
62,429 | 25,998,836,992 | IssuesEvent | 2022-12-20 13:49:31 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | AWS Cloud WAN | enhancement new-resource service/networkmanager | <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
> Today, AWS announced the general availability of AWS Cloud WAN, a wide area networking (WAN) service that helps you build, manage, and monitor a unified global network.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* `aws_networkmanager_core_network`
* ...
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
# Copy-paste your Terraform configurations here - for large Terraform configs,
# please use a service like Dropbox and share a link to the ZIP file. For
# security, you can also encrypt the files using our GPG public key.
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/
--->
[Announcement](https://aws.amazon.com/about-aws/whats-new/2022/07/general-availability-aws-cloud-wan/).
[Blog post](https://aws.amazon.com/blogs/aws/new-cloud-wan-a-managed-wan-service/).
Related:
* https://github.com/hashicorp/terraform-provider-aws/pull/24368
| 1.0 | AWS Cloud WAN - <!--- Please keep this note for the community --->
### Community Note
* Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Description
<!--- Please leave a helpful description of the feature request here. --->
> Today, AWS announced the general availability of AWS Cloud WAN, a wide area networking (WAN) service that helps you build, manage, and monitor a unified global network.
### New or Affected Resource(s)
<!--- Please list the new or affected resources and data sources. --->
* `aws_networkmanager_core_network`
* ...
### Potential Terraform Configuration
<!--- Information about code formatting: https://help.github.com/articles/basic-writing-and-formatting-syntax/#quoting-code --->
```hcl
# Copy-paste your Terraform configurations here - for large Terraform configs,
# please use a service like Dropbox and share a link to the ZIP file. For
# security, you can also encrypt the files using our GPG public key.
```
### References
<!---
Information about referencing Github Issues: https://help.github.com/articles/basic-writing-and-formatting-syntax/#referencing-issues-and-pull-requests
Are there any other GitHub issues (open or closed) or pull requests that should be linked here? Vendor blog posts or documentation? For example:
* https://aws.amazon.com/about-aws/whats-new/2018/04/introducing-amazon-ec2-fleet/
--->
[Announcement](https://aws.amazon.com/about-aws/whats-new/2022/07/general-availability-aws-cloud-wan/).
[Blog post](https://aws.amazon.com/blogs/aws/new-cloud-wan-a-managed-wan-service/).
Related:
* https://github.com/hashicorp/terraform-provider-aws/pull/24368
| non_main | aws cloud wan community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment description today aws announced the general availability of aws cloud wan a wide area networking wan service that helps you build manage and monitor a unified global network new or affected resource s aws networkmanager core network potential terraform configuration hcl copy paste your terraform configurations here for large terraform configs please use a service like dropbox and share a link to the zip file for security you can also encrypt the files using our gpg public key references information about referencing github issues are there any other github issues open or closed or pull requests that should be linked here vendor blog posts or documentation for example related | 0 |
316,323 | 23,625,062,215 | IssuesEvent | 2022-08-25 02:26:53 | Mazurco066/band-manager-api | https://api.github.com/repos/Mazurco066/band-manager-api | closed | [Backend] Webscrapper endpoint for "Pocket Terço" Website | documentation enhancement | # Briefing
There is a Website named as [Liturgia Pocket Terço](https://pocketterco.com.br/liturgia/18/09/2022) that provides the daily liturgy.
The main objective is to create a endpoint that receives a date and scraps the daily liturgy such as: **Antífona de entrada,
Primeira Leitura, Salmo Responsorial, Segunda Leitura, Aleluia, Evangelho, Antífona do Ofertório and Antífona do Comunhão**.
## Inputs
It should receive a date that matches the desired concert
```ts
{
date: "YYYY-MM-DD"
}
```
## Integrations
It should send a GET request to **https://pocketterco.com.br/liturgia/18/09/2022**.
Notice that url receives a date, thats where date input will dynamically fit.
## Return object
It should return something like this:
```ts
{
status: {
code: 200,
message: "Daily liturgy retrieved"
},
data: [
{
title: "Antífona de Entrada",
content: "Text content here..."
},
...
]
}
``` | 1.0 | [Backend] Webscrapper endpoint for "Pocket Terço" Website - # Briefing
There is a Website named as [Liturgia Pocket Terço](https://pocketterco.com.br/liturgia/18/09/2022) that provides the daily liturgy.
The main objective is to create a endpoint that receives a date and scraps the daily liturgy such as: **Antífona de entrada,
Primeira Leitura, Salmo Responsorial, Segunda Leitura, Aleluia, Evangelho, Antífona do Ofertório and Antífona do Comunhão**.
## Inputs
It should receive a date that matches the desired concert
```ts
{
date: "YYYY-MM-DD"
}
```
## Integrations
It should send a GET request to **https://pocketterco.com.br/liturgia/18/09/2022**.
Notice that url receives a date, thats where date input will dynamically fit.
## Return object
It should return something like this:
```ts
{
status: {
code: 200,
message: "Daily liturgy retrieved"
},
data: [
{
title: "Antífona de Entrada",
content: "Text content here..."
},
...
]
}
``` | non_main | webscrapper endpoint for pocket terço website briefing there is a website named as that provides the daily liturgy the main objective is to create a endpoint that receives a date and scraps the daily liturgy such as antífona de entrada primeira leitura salmo responsorial segunda leitura aleluia evangelho antífona do ofertório and antífona do comunhão inputs it should receive a date that matches the desired concert ts date yyyy mm dd integrations it should send a get request to notice that url receives a date thats where date input will dynamically fit return object it should return something like this ts status code message daily liturgy retrieved data title antífona de entrada content text content here | 0 |
559 | 4,015,886,082 | IssuesEvent | 2016-05-15 07:34:01 | caskroom/homebrew-cask | https://api.github.com/repos/caskroom/homebrew-cask | closed | `brew cask style` fails because of dependencies on rubocop versions. | awaiting maintainer feedback | When I try to run `brew cask style`, I get the following errors:
```
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/specification.rb:2007:in `raise_if_conflicts': Unable to activate rubocop-cask-0.5.0, because rubocop-0.39.0 conflicts with rubocop (~> 0.37.0) (Gem::LoadError)
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/specification.rb:1176:in `activate'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems.rb:186:in `rescue in try_activate'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems.rb:183:in `try_activate'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:132:in `rescue in require'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:144:in `require'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_loader_resolver.rb:16:in `block in resolve_requires'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_loader_resolver.rb:12:in `each'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_loader_resolver.rb:12:in `resolve_requires'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_loader.rb:39:in `load_file'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_store.rb:24:in `options_config='
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/cli.rb:64:in `act_on_options'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/cli.rb:25:in `run'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/bin/rubocop:14:in `block in <top (required)>'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/benchmark.rb:296:in `realtime'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/bin/rubocop:13:in `<top (required)>'
from /Users/ntoone/.gem/ruby/2.0.0/bin/rubocop:23:in `load'
from /Users/ntoone/.gem/ruby/2.0.0/bin/rubocop:23:in `<main>'
Error: style check failed
```
It looks like there is a different dependency between rubocop-cask (which depends on rubocop 37) and what the base home-brew depends upon (rubocop 39). | True | `brew cask style` fails because of dependencies on rubocop versions. - When I try to run `brew cask style`, I get the following errors:
```
/System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/specification.rb:2007:in `raise_if_conflicts': Unable to activate rubocop-cask-0.5.0, because rubocop-0.39.0 conflicts with rubocop (~> 0.37.0) (Gem::LoadError)
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/specification.rb:1176:in `activate'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems.rb:186:in `rescue in try_activate'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems.rb:183:in `try_activate'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:132:in `rescue in require'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/rubygems/core_ext/kernel_require.rb:144:in `require'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_loader_resolver.rb:16:in `block in resolve_requires'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_loader_resolver.rb:12:in `each'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_loader_resolver.rb:12:in `resolve_requires'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_loader.rb:39:in `load_file'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/config_store.rb:24:in `options_config='
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/cli.rb:64:in `act_on_options'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/lib/rubocop/cli.rb:25:in `run'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/bin/rubocop:14:in `block in <top (required)>'
from /System/Library/Frameworks/Ruby.framework/Versions/2.0/usr/lib/ruby/2.0.0/benchmark.rb:296:in `realtime'
from /Users/ntoone/.gem/ruby/2.0.0/gems/rubocop-0.39.0/bin/rubocop:13:in `<top (required)>'
from /Users/ntoone/.gem/ruby/2.0.0/bin/rubocop:23:in `load'
from /Users/ntoone/.gem/ruby/2.0.0/bin/rubocop:23:in `<main>'
Error: style check failed
```
It looks like there is a different dependency between rubocop-cask (which depends on rubocop 37) and what the base home-brew depends upon (rubocop 39). | main | brew cask style fails because of dependencies on rubocop versions when i try to run brew cask style i get the following errors system library frameworks ruby framework versions usr lib ruby rubygems specification rb in raise if conflicts unable to activate rubocop cask because rubocop conflicts with rubocop gem loaderror from system library frameworks ruby framework versions usr lib ruby rubygems specification rb in activate from system library frameworks ruby framework versions usr lib ruby rubygems rb in rescue in try activate from system library frameworks ruby framework versions usr lib ruby rubygems rb in try activate from system library frameworks ruby framework versions usr lib ruby rubygems core ext kernel require rb in rescue in require from system library frameworks ruby framework versions usr lib ruby rubygems core ext kernel require rb in require from users ntoone gem ruby gems rubocop lib rubocop config loader resolver rb in block in resolve requires from users ntoone gem ruby gems rubocop lib rubocop config loader resolver rb in each from users ntoone gem ruby gems rubocop lib rubocop config loader resolver rb in resolve requires from users ntoone gem ruby gems rubocop lib rubocop config loader rb in load file from users ntoone gem ruby gems rubocop lib rubocop config store rb in options config from users ntoone gem ruby gems rubocop lib rubocop cli rb in act on options from users ntoone gem ruby gems rubocop lib rubocop cli rb in run from users ntoone gem ruby gems rubocop bin rubocop in block in from system library frameworks ruby framework versions usr lib ruby benchmark rb in realtime from users ntoone gem ruby gems rubocop bin rubocop in from users ntoone gem ruby bin rubocop in load from users ntoone gem ruby bin rubocop in error style check failed it looks like there is a different dependency between rubocop cask which depends on rubocop and what the base home brew depends upon rubocop | 1 |
5,390 | 27,082,785,005 | IssuesEvent | 2023-02-14 15:04:38 | centerofci/mathesar | https://api.github.com/repos/centerofci/mathesar | closed | Compress static assets while serving | type: enhancement work: backend status: ready restricted: maintainers | ## Description
* Our frontend static assets are being served as is, in production mode. This leads to a higher loading time.
* They need to be served with gzip compression.
* ~For User-Agents that support brotli, they need to be compressed with brotli~ (branched off to a separate issue: https://github.com/centerofci/mathesar/issues/2411).
| True | Compress static assets while serving - ## Description
* Our frontend static assets are being served as is, in production mode. This leads to a higher loading time.
* They need to be served with gzip compression.
* ~For User-Agents that support brotli, they need to be compressed with brotli~ (branched off to a separate issue: https://github.com/centerofci/mathesar/issues/2411).
| main | compress static assets while serving description our frontend static assets are being served as is in production mode this leads to a higher loading time they need to be served with gzip compression for user agents that support brotli they need to be compressed with brotli branched off to a separate issue | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.