Unnamed: 0 int64 3 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 742 | labels stringlengths 4 431 | body stringlengths 5 239k | index stringclasses 10 values | text_combine stringlengths 96 240k | label stringclasses 2 values | text stringlengths 96 200k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
281,828 | 30,888,966,783 | IssuesEvent | 2023-08-04 02:04:16 | nidhi7598/linux-4.1.15_CVE-2019-10220 | https://api.github.com/repos/nidhi7598/linux-4.1.15_CVE-2019-10220 | reopened | CVE-2017-7542 (Medium) detected in linuxlinux-4.4.302 | Mend: dependency security vulnerability | ## CVE-2017-7542 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/output_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The ip6_find_1stfragopt function in net/ipv6/output_core.c in the Linux kernel through 4.12.3 allows local users to cause a denial of service (integer overflow and infinite loop) by leveraging the ability to open a raw socket.
<p>Publish Date: 2017-07-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-7542>CVE-2017-7542</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-7542">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-7542</a></p>
<p>Release Date: 2017-07-21</p>
<p>Fix Resolution: v4.13-rc2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2017-7542 (Medium) detected in linuxlinux-4.4.302 - ## CVE-2017-7542 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.4.302</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.1.15_CVE-2019-10220/commit/6a0d304d962ca933d73f507ce02157ef2791851c">6a0d304d962ca933d73f507ce02157ef2791851c</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/ipv6/output_core.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The ip6_find_1stfragopt function in net/ipv6/output_core.c in the Linux kernel through 4.12.3 allows local users to cause a denial of service (integer overflow and infinite loop) by leveraging the ability to open a raw socket.
<p>Publish Date: 2017-07-21
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2017-7542>CVE-2017-7542</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-7542">http://web.nvd.nist.gov/view/vuln/detail?vulnId=CVE-2017-7542</a></p>
<p>Release Date: 2017-07-21</p>
<p>Fix Resolution: v4.13-rc2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_usab | cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net output core c vulnerability details the find function in net output core c in the linux kernel through allows local users to cause a denial of service integer overflow and infinite loop by leveraging the ability to open a raw socket publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
121,495 | 12,127,507,324 | IssuesEvent | 2020-04-22 18:50:54 | vmware-tanzu/velero | https://api.github.com/repos/vmware-tanzu/velero | closed | Velero site missing vSphere installation and configuration instructions | Documentation | **Describe the problem/challenge you have**
The Velero website site doesn't include installation and configuration instructions for the vShpere provider
**Describe the solution you'd like**
A link in the provider page to install and configuration instructions for vSphere
**Anything else you would like to add:**
Referencing the latest Velero page here
https://velero.io/docs/v1.3.1/supported-providers/
**Environment:**
N/A
- Velero version (use `velero version`):
- Kubernetes version (use `kubectl version`):
- Kubernetes installer & version:
- Cloud provider or hardware configuration:
- OS (e.g. from `/etc/os-release`):
| 1.0 | Velero site missing vSphere installation and configuration instructions - **Describe the problem/challenge you have**
The Velero website site doesn't include installation and configuration instructions for the vShpere provider
**Describe the solution you'd like**
A link in the provider page to install and configuration instructions for vSphere
**Anything else you would like to add:**
Referencing the latest Velero page here
https://velero.io/docs/v1.3.1/supported-providers/
**Environment:**
N/A
- Velero version (use `velero version`):
- Kubernetes version (use `kubectl version`):
- Kubernetes installer & version:
- Cloud provider or hardware configuration:
- OS (e.g. from `/etc/os-release`):
| non_usab | velero site missing vsphere installation and configuration instructions describe the problem challenge you have the velero website site doesn t include installation and configuration instructions for the vshpere provider describe the solution you d like a link in the provider page to install and configuration instructions for vsphere anything else you would like to add referencing the latest velero page here environment n a velero version use velero version kubernetes version use kubectl version kubernetes installer version cloud provider or hardware configuration os e g from etc os release | 0 |
86,526 | 17,019,954,468 | IssuesEvent | 2021-07-02 17:16:20 | TromboneDavies/PolarOps | https://api.github.com/repos/TromboneDavies/PolarOps | closed | Make collector.py store dates in better format (?) | code | Not sure if we actually want this, but it's hard on the eyes to look at the `.csv` file, or its `DataFrame` manifestation after `pd.read_csv()`'ing it, and see just a bunch of meaningless huge integers. It would be much nicer to see "5/13/2016 11:05PM" type stuff. | 1.0 | Make collector.py store dates in better format (?) - Not sure if we actually want this, but it's hard on the eyes to look at the `.csv` file, or its `DataFrame` manifestation after `pd.read_csv()`'ing it, and see just a bunch of meaningless huge integers. It would be much nicer to see "5/13/2016 11:05PM" type stuff. | non_usab | make collector py store dates in better format not sure if we actually want this but it s hard on the eyes to look at the csv file or its dataframe manifestation after pd read csv ing it and see just a bunch of meaningless huge integers it would be much nicer to see type stuff | 0 |
1,957 | 3,025,718,246 | IssuesEvent | 2015-08-03 10:35:18 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 22112410: Mail.app doesn't always release file handles of files sent as attachments. | classification:ui/usability reproducible:always status:open | #### Description
Summary:
Compose an email with an image attachment. "Edit" the image with the Markup feature. Send email. Wait a few minutes (the mail needs to send). Put attachment in trash. Try and empty trash. Just try. I dare you.
Steps to Reproduce:
1. Open Mail.app
2. Compose new message.
3. Add image attachment.
4. Edit the attachment with "Markup".
5. Send message.
6. Trash image file.
7. Wait a few minutes.
8. Attempt to empty trash.
Expected Results:
That the trash is emptied.
Actual Results:
I can't empty Trash because the files I've attached to a mail message are still in use by Mail (open files). See attached screenshots.
Version:
10.10.4 (14E46)
Notes:
Configuration:
If you don't edit the file before sending, the file handle is not leaked.
-
Product Version: 10.10.4 (14E46)
Created: 2015-08-03T10:33:37.060070
Originated: 2015-08-03T00:00:00
Open Radar Link: http://www.openradar.me/22112410 | True | 22112410: Mail.app doesn't always release file handles of files sent as attachments. - #### Description
Summary:
Compose an email with an image attachment. "Edit" the image with the Markup feature. Send email. Wait a few minutes (the mail needs to send). Put attachment in trash. Try and empty trash. Just try. I dare you.
Steps to Reproduce:
1. Open Mail.app
2. Compose new message.
3. Add image attachment.
4. Edit the attachment with "Markup".
5. Send message.
6. Trash image file.
7. Wait a few minutes.
8. Attempt to empty trash.
Expected Results:
That the trash is emptied.
Actual Results:
I can't empty Trash because the files I've attached to a mail message are still in use by Mail (open files). See attached screenshots.
Version:
10.10.4 (14E46)
Notes:
Configuration:
If you don't edit the file before sending, the file handle is not leaked.
-
Product Version: 10.10.4 (14E46)
Created: 2015-08-03T10:33:37.060070
Originated: 2015-08-03T00:00:00
Open Radar Link: http://www.openradar.me/22112410 | usab | mail app doesn t always release file handles of files sent as attachments description summary compose an email with an image attachment edit the image with the markup feature send email wait a few minutes the mail needs to send put attachment in trash try and empty trash just try i dare you steps to reproduce open mail app compose new message add image attachment edit the attachment with markup send message trash image file wait a few minutes attempt to empty trash expected results that the trash is emptied actual results i can t empty trash because the files i ve attached to a mail message are still in use by mail open files see attached screenshots version notes configuration if you don t edit the file before sending the file handle is not leaked product version created originated open radar link | 1 |
96,319 | 16,129,615,662 | IssuesEvent | 2021-04-29 01:04:02 | RG4421/ampere-centos-kernel | https://api.github.com/repos/RG4421/ampere-centos-kernel | opened | CVE-2020-27815 (High) detected in https://source.codeaurora.org/quic/kernel/agross-msm/qcom-arm64-for-4.3 | security vulnerability | ## CVE-2020-27815 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https://source.codeaurora.org/quic/kernel/agross-msm/qcom-arm64-for-4.3</b></p></summary>
<p>
<p>Library home page: <a href=https://source.codeaurora.org/quic/kernel/agross-msm/>https://source.codeaurora.org/quic/kernel/agross-msm/</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>ampere-centos-kernel/fs/jfs/jfs_dmap.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>ampere-centos-kernel/fs/jfs/jfs_dmap.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>ampere-centos-kernel/fs/jfs/jfs_dmap.h</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Linux kernel is vulnerable to Array index out of bounds access when setting extended attributes on journaling.
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27815>CVE-2020-27815</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| True | CVE-2020-27815 (High) detected in https://source.codeaurora.org/quic/kernel/agross-msm/qcom-arm64-for-4.3 - ## CVE-2020-27815 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>https://source.codeaurora.org/quic/kernel/agross-msm/qcom-arm64-for-4.3</b></p></summary>
<p>
<p>Library home page: <a href=https://source.codeaurora.org/quic/kernel/agross-msm/>https://source.codeaurora.org/quic/kernel/agross-msm/</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>ampere-centos-kernel/fs/jfs/jfs_dmap.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>ampere-centos-kernel/fs/jfs/jfs_dmap.h</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>ampere-centos-kernel/fs/jfs/jfs_dmap.h</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Linux kernel is vulnerable to Array index out of bounds access when setting extended attributes on journaling.
<p>Publish Date: 2020-10-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-27815>CVE-2020-27815</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
| non_usab | cve high detected in cve high severity vulnerability vulnerable library library home page a href found in base branch amp centos kernel vulnerable source files ampere centos kernel fs jfs jfs dmap h ampere centos kernel fs jfs jfs dmap h ampere centos kernel fs jfs jfs dmap h vulnerability details linux kernel is vulnerable to array index out of bounds access when setting extended attributes on journaling publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href | 0 |
14,005 | 8,782,785,530 | IssuesEvent | 2018-12-20 02:02:10 | dotnet/machinelearning | https://api.github.com/repos/dotnet/machinelearning | closed | Add confidence intervals to permutation feature importance | usability | `Permutation Feature Importance` (aka `PFI`) computes the importance of a feature to a model by permuting values for that feature, scoring it with the model, and comparing the new evaluation metrics to the original evaluation metrics. For speed, `PFI` uses only one permutation, and this leads to a bit of randomness in the predicted importances. For example, based on the random seed features can change orderings of importance and permutations can even end up showing to improve the model performance. These issues can be fixed by allowing the calculation of confidence intervals around the feature importance values. | True | Add confidence intervals to permutation feature importance - `Permutation Feature Importance` (aka `PFI`) computes the importance of a feature to a model by permuting values for that feature, scoring it with the model, and comparing the new evaluation metrics to the original evaluation metrics. For speed, `PFI` uses only one permutation, and this leads to a bit of randomness in the predicted importances. For example, based on the random seed features can change orderings of importance and permutations can even end up showing to improve the model performance. These issues can be fixed by allowing the calculation of confidence intervals around the feature importance values. | usab | add confidence intervals to permutation feature importance permutation feature importance aka pfi computes the importance of a feature to a model by permuting values for that feature scoring it with the model and comparing the new evaluation metrics to the original evaluation metrics for speed pfi uses only one permutation and this leads to a bit of randomness in the predicted importances for example based on the random seed features can change orderings of importance and permutations can even end up showing to improve the model performance these issues can be fixed by allowing the calculation of confidence intervals around the feature importance values | 1 |
16,134 | 11,848,786,341 | IssuesEvent | 2020-03-24 14:18:38 | dotnet/runtime | https://api.github.com/repos/dotnet/runtime | closed | Unable to build System.Runtime.WindowsRuntime | area-Infrastructure | I just did a clean, sync, and full rebuild. When I now try to build System.Runtime.WindowsRuntime, I get this error:
```
d:\repos\runtime\src\libraries\System.Runtime.WindowsRuntime\src>d:\repos\runtime\dotnet msbuild /t:rebuild
D:\repos\runtime\.dotnet
Microsoft (R) Build Engine version 16.6.0-preview-20126-02+13cfe7fc5 for .NET Core
Copyright (C) Microsoft Corporation. All rights reserved.
C:\Users\stoub\.nuget\packages\microsoft.dotnet.build.tasks.targetframework.sdk\5.0.0-beta.20167.1\build\Microsoft.DotNet.Build.Tasks.TargetFramework.Sdk.targets(84,5): error : Not able to find a compatible supported target framework for netstandard1.0 in Project System.Runtime.WindowsRuntime.csproj. The Supported Configurations are netcoreapp5.0, netstandard2.0 [d:\repos\runtime\src\libraries\System.Runtime.WindowsRuntime\src\System.Runtime.WindowsRuntime.csproj]
C:\Users\stoub\.nuget\packages\microsoft.dotnet.build.tasks.targetframework.sdk\5.0.0-beta.20167.1\build\Microsoft.DotNet.Build.Tasks.TargetFramework.Sdk.targets(84,5): error : Not able to find a compatible supported target framework for netstandard1.2 in Project System.Runtime.WindowsRuntime.csproj. The Supported Configurations are netcoreapp5.0, netstandard2.0 [d:\repos\runtime\src\libraries\System.Runtime.WindowsRuntime\src\System.Runtime.WindowsRuntime.csproj]
Restore completed in 42.75 ms for d:\repos\runtime\src\libraries\restore\winrt\winrt.depproj.
Restore completed in 8.25 ms for d:\repos\runtime\src\libraries\restore\winrt\winrt.depproj.
```
cc: @Anipik, @ViktorHofer | 1.0 | Unable to build System.Runtime.WindowsRuntime - I just did a clean, sync, and full rebuild. When I now try to build System.Runtime.WindowsRuntime, I get this error:
```
d:\repos\runtime\src\libraries\System.Runtime.WindowsRuntime\src>d:\repos\runtime\dotnet msbuild /t:rebuild
D:\repos\runtime\.dotnet
Microsoft (R) Build Engine version 16.6.0-preview-20126-02+13cfe7fc5 for .NET Core
Copyright (C) Microsoft Corporation. All rights reserved.
C:\Users\stoub\.nuget\packages\microsoft.dotnet.build.tasks.targetframework.sdk\5.0.0-beta.20167.1\build\Microsoft.DotNet.Build.Tasks.TargetFramework.Sdk.targets(84,5): error : Not able to find a compatible supported target framework for netstandard1.0 in Project System.Runtime.WindowsRuntime.csproj. The Supported Configurations are netcoreapp5.0, netstandard2.0 [d:\repos\runtime\src\libraries\System.Runtime.WindowsRuntime\src\System.Runtime.WindowsRuntime.csproj]
C:\Users\stoub\.nuget\packages\microsoft.dotnet.build.tasks.targetframework.sdk\5.0.0-beta.20167.1\build\Microsoft.DotNet.Build.Tasks.TargetFramework.Sdk.targets(84,5): error : Not able to find a compatible supported target framework for netstandard1.2 in Project System.Runtime.WindowsRuntime.csproj. The Supported Configurations are netcoreapp5.0, netstandard2.0 [d:\repos\runtime\src\libraries\System.Runtime.WindowsRuntime\src\System.Runtime.WindowsRuntime.csproj]
Restore completed in 42.75 ms for d:\repos\runtime\src\libraries\restore\winrt\winrt.depproj.
Restore completed in 8.25 ms for d:\repos\runtime\src\libraries\restore\winrt\winrt.depproj.
```
cc: @Anipik, @ViktorHofer | non_usab | unable to build system runtime windowsruntime i just did a clean sync and full rebuild when i now try to build system runtime windowsruntime i get this error d repos runtime src libraries system runtime windowsruntime src d repos runtime dotnet msbuild t rebuild d repos runtime dotnet microsoft r build engine version preview for net core copyright c microsoft corporation all rights reserved c users stoub nuget packages microsoft dotnet build tasks targetframework sdk beta build microsoft dotnet build tasks targetframework sdk targets error not able to find a compatible supported target framework for in project system runtime windowsruntime csproj the supported configurations are c users stoub nuget packages microsoft dotnet build tasks targetframework sdk beta build microsoft dotnet build tasks targetframework sdk targets error not able to find a compatible supported target framework for in project system runtime windowsruntime csproj the supported configurations are restore completed in ms for d repos runtime src libraries restore winrt winrt depproj restore completed in ms for d repos runtime src libraries restore winrt winrt depproj cc anipik viktorhofer | 0 |
30,482 | 4,622,528,399 | IssuesEvent | 2016-09-27 07:55:14 | Laravel-Backpack/Base | https://api.github.com/repos/Laravel-Backpack/Base | closed | Customise Registration Form | question testing or needs confirmation | Hi,
Just wondering what the best way to customise the auth templates would be?
I have added extra fields to my user model, so need them completed on registration - even for admin, so need to add these fields to the view and also Controller.
However because all of these live in /vendor/backpack etc they are outside of the regular version control.
Is there a standard way to handle this type of thing?
Cheers | 1.0 | Customise Registration Form - Hi,
Just wondering what the best way to customise the auth templates would be?
I have added extra fields to my user model, so need them completed on registration - even for admin, so need to add these fields to the view and also Controller.
However because all of these live in /vendor/backpack etc they are outside of the regular version control.
Is there a standard way to handle this type of thing?
Cheers | non_usab | customise registration form hi just wondering what the best way to customise the auth templates would be i have added extra fields to my user model so need them completed on registration even for admin so need to add these fields to the view and also controller however because all of these live in vendor backpack etc they are outside of the regular version control is there a standard way to handle this type of thing cheers | 0 |
6,818 | 4,553,622,423 | IssuesEvent | 2016-09-13 06:03:49 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Enhancement wish for "Create New Node" Window | enhancement topic:editor usability | **Old:**

**New:**


**Filter** is type of node.
- Node = everything
- CanvasItem = all Control node + all Node2D node
- Control = all Control node
and so on..
**Favorites** and **Recent** is like "Open a Resource/File" Window.
| True | Enhancement wish for "Create New Node" Window - **Old:**

**New:**


**Filter** is type of node.
- Node = everything
- CanvasItem = all Control node + all Node2D node
- Control = all Control node
and so on..
**Favorites** and **Recent** is like "Open a Resource/File" Window.
| usab | enhancement wish for create new node window old new filter is type of node node everything canvasitem all control node all node control all control node and so on favorites and recent is like open a resource file window | 1 |
19,081 | 13,536,133,185 | IssuesEvent | 2020-09-16 08:36:40 | topcoder-platform/qa-fun | https://api.github.com/repos/topcoder-platform/qa-fun | closed | Images right side is not aligned with each other | UX/Usability | Steps :
1) Go to https://www.topcoder.com/
2) Scroll sown to section "OUR CUSTOMERS, THEIR SUCCESS"
Observe right side image alligment
Expected Result:
Images right side should be aligned.
Actual Result:
Images right side is not aligned with each other
Screenshots:

Device/OS/Browser Information:
Laptop Windows 10,
Chrome Version 81.0.4044.129 (Official Build) (64-bit) | True | Images right side is not aligned with each other - Steps :
1) Go to https://www.topcoder.com/
2) Scroll sown to section "OUR CUSTOMERS, THEIR SUCCESS"
Observe right side image alligment
Expected Result:
Images right side should be aligned.
Actual Result:
Images right side is not aligned with each other
Screenshots:

Device/OS/Browser Information:
Laptop Windows 10,
Chrome Version 81.0.4044.129 (Official Build) (64-bit) | usab | images right side is not aligned with each other steps go to scroll sown to section our customers their success observe right side image alligment expected result images right side should be aligned actual result images right side is not aligned with each other screenshots device os browser information laptop windows chrome version official build bit | 1 |
19,676 | 14,405,553,717 | IssuesEvent | 2020-12-03 18:52:59 | briansmith/ring | https://api.github.com/repos/briansmith/ring | closed | Add Visual Studio 2017 builds to AppVeyor | good-first-bug usability | [Note that Visual Studio “15” is the successor to Visual Studio 2015 and will probably have a different name; see https://www.visualstudio.com/en-us/news/releasenotes/vs15-relnotes.]
Let's support building the C/C++ components with Visual Studio “15” and using its linker.
The AppVeyor issue for Visual Studio “15” support is https://github.com/appveyor/ci/issues/753, where they say they'll wait at least until there's a beta release. But, we'll take patches to support it in *ring* before then. | True | Add Visual Studio 2017 builds to AppVeyor - [Note that Visual Studio “15” is the successor to Visual Studio 2015 and will probably have a different name; see https://www.visualstudio.com/en-us/news/releasenotes/vs15-relnotes.]
Let's support building the C/C++ components with Visual Studio “15” and using its linker.
The AppVeyor issue for Visual Studio “15” support is https://github.com/appveyor/ci/issues/753, where they say they'll wait at least until there's a beta release. But, we'll take patches to support it in *ring* before then. | usab | add visual studio builds to appveyor let s support building the c c components with visual studio “ ” and using its linker the appveyor issue for visual studio “ ” support is where they say they ll wait at least until there s a beta release but we ll take patches to support it in ring before then | 1 |
2,480 | 3,078,999,819 | IssuesEvent | 2015-08-21 14:02:37 | rabbitmq/rabbitmq-server | https://api.github.com/repos/rabbitmq/rabbitmq-server | opened | Reduce default QI (queue index) journal size | bug effort-tiny enhancement usability | See #227 for background.
Profiling suggests that the bottleneck in #227 is sequential folding over QI journal. While we will look into parallelising it, the need for reducing I/O by having a larger journal seems a lot less relevant in 2015 than it was ~ 5 years ago.
Benchmarks and monitoring results from multiple people suggests values such as `16K` or `8K` result in much more even throughput.
We've agreed to reduce default journal size from `65536` to `8192`. | True | Reduce default QI (queue index) journal size - See #227 for background.
Profiling suggests that the bottleneck in #227 is sequential folding over QI journal. While we will look into parallelising it, the need for reducing I/O by having a larger journal seems a lot less relevant in 2015 than it was ~ 5 years ago.
Benchmarks and monitoring results from multiple people suggests values such as `16K` or `8K` result in much more even throughput.
We've agreed to reduce default journal size from `65536` to `8192`. | usab | reduce default qi queue index journal size see for background profiling suggests that the bottleneck in is sequential folding over qi journal while we will look into parallelising it the need for reducing i o by having a larger journal seems a lot less relevant in than it was years ago benchmarks and monitoring results from multiple people suggests values such as or result in much more even throughput we ve agreed to reduce default journal size from to | 1 |
27,854 | 30,545,324,094 | IssuesEvent | 2023-07-20 03:02:06 | penrose/penrose | https://api.github.com/repos/penrose/penrose | closed | Run core functions on worker threads in `editor` and `edgeworth` | kind:usability | Running costly functions like `stepUntilConvergence` single-threaded blocks the main UI experience. Adding webworkers should be an easy change that'd improve the UX a lot. | True | Run core functions on worker threads in `editor` and `edgeworth` - Running costly functions like `stepUntilConvergence` single-threaded blocks the main UI experience. Adding webworkers should be an easy change that'd improve the UX a lot. | usab | run core functions on worker threads in editor and edgeworth running costly functions like stepuntilconvergence single threaded blocks the main ui experience adding webworkers should be an easy change that d improve the ux a lot | 1 |
286,923 | 24,794,758,116 | IssuesEvent | 2022-10-24 16:17:10 | near/nearcore | https://api.github.com/repos/near/nearcore | opened | #7839 breaks bunch of NayDuck tests | A-testing P-high | For example
```shell
$ git checkout 864cc29
$ cargo test -pintegration-tests --features test_features,expensive_tests -- --exact --nocapture tests::standard_cases::rpc::test_delete_key_testnet
→ fails
$ git checkout @~
$ cargo test -pintegration-tests --features test_features,expensive_tests -- --exact --nocapture tests::standard_cases::rpc::test_delete_key_testnet
→ passes
```
There are more tests. Compare https://nayduck.near.org/#/run/2727 and https://nayduck.near.org/#/run/2722.
| 1.0 | #7839 breaks bunch of NayDuck tests - For example
```shell
$ git checkout 864cc29
$ cargo test -pintegration-tests --features test_features,expensive_tests -- --exact --nocapture tests::standard_cases::rpc::test_delete_key_testnet
→ fails
$ git checkout @~
$ cargo test -pintegration-tests --features test_features,expensive_tests -- --exact --nocapture tests::standard_cases::rpc::test_delete_key_testnet
→ passes
```
There are more tests. Compare https://nayduck.near.org/#/run/2727 and https://nayduck.near.org/#/run/2722.
| non_usab | breaks bunch of nayduck tests for example shell git checkout cargo test pintegration tests features test features expensive tests exact nocapture tests standard cases rpc test delete key testnet → fails git checkout cargo test pintegration tests features test features expensive tests exact nocapture tests standard cases rpc test delete key testnet → passes there are more tests compare and | 0 |
10,537 | 6,793,976,942 | IssuesEvent | 2017-11-01 10:06:22 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | No easy way to restore default editor settings | enhancement topic:editor usability | **Operating system or device - Godot version:**
Arch Linux, Godot 3 master
**Issue description:**
I noticed there is currently no easy way to restore the default settings (per setting or globally) within the editor and project settings other than deleting config files and reloading the editor. I think this might become especially problematic for new users, because it's easy to mess things up and get confused or lost when playing around with certain values without any way of telling what the original sane settings were.
| True | No easy way to restore default editor settings - **Operating system or device - Godot version:**
Arch Linux, Godot 3 master
**Issue description:**
I noticed there is currently no easy way to restore the default settings (per setting or globally) within the editor and project settings other than deleting config files and reloading the editor. I think this might become especially problematic for new users, because it's easy to mess things up and get confused or lost when playing around with certain values without any way of telling what the original sane settings were.
| usab | no easy way to restore default editor settings operating system or device godot version arch linux godot master issue description i noticed there is currently no easy way to restore the default settings per setting or globally within the editor and project settings other than deleting config files and reloading the editor i think this might become especially problematic for new users because it s easy to mess things up and get confused or lost when playing around with certain values without any way of telling what the original sane settings were | 1 |
12,090 | 7,693,113,246 | IssuesEvent | 2018-05-18 01:34:33 | coreos/bugs | https://api.github.com/repos/coreos/bugs | closed | Compatibility of HPE FlexFabric 10Gb 2-port 534FLB Adapter with coreos | area/usability component/kernel kind/enhancement low hanging fruit team/os | # Issue Report #
## Bug ##
For Fibre channel storage (FCoE), I am using HP FlexFabric 10Gb 2-Port 534FLB Adapter. It is a converged network Broadcom adapter.
I opened discussion on CoreOS user group about the compatibility of this adapter with CoreOS and found out that currently, CoreOS don't have the kernel driver that supports FCoE offload (config option SCSI_BNX2X_FCOE).
**Link to the discussion:** https://groups.google.com/forum/#!topic/coreos-user/5v0FPUUd2to
**Output of `lspci -v` command:** [lspci.txt](https://github.com/coreos/bugs/files/1742631/lspci.txt)
### Container Linux Version ###
```
cat /etc/os-release
PRETTY_NAME="Debian GNU/Linux 8 (jessie)"
NAME="Debian GNU/Linux"
VERSION_ID="8"
VERSION="8 (jessie)"
ID=debian
HOME_URL="http://www.debian.org/"
SUPPORT_URL="http://www.debian.org/support"
BUG_REPORT_URL="https://bugs.debian.org/"
```
```
uname -r
4.14.16-coreos
```
### Environment ###
Baremetal
| True | Compatibility of HPE FlexFabric 10Gb 2-port 534FLB Adapter with coreos - # Issue Report #
## Bug ##
For Fibre channel storage (FCoE), I am using HP FlexFabric 10Gb 2-Port 534FLB Adapter. It is a converged network Broadcom adapter.
I opened discussion on CoreOS user group about the compatibility of this adapter with CoreOS and found out that currently, CoreOS don't have the kernel driver that supports FCoE offload (config option SCSI_BNX2X_FCOE).
**Link to the discussion:** https://groups.google.com/forum/#!topic/coreos-user/5v0FPUUd2to
**Output of `lspci -v` command:** [lspci.txt](https://github.com/coreos/bugs/files/1742631/lspci.txt)
### Container Linux Version ###
```
cat /etc/os-release
PRETTY_NAME="Debian GNU/Linux 8 (jessie)"
NAME="Debian GNU/Linux"
VERSION_ID="8"
VERSION="8 (jessie)"
ID=debian
HOME_URL="http://www.debian.org/"
SUPPORT_URL="http://www.debian.org/support"
BUG_REPORT_URL="https://bugs.debian.org/"
```
```
uname -r
4.14.16-coreos
```
### Environment ###
Baremetal
| usab | compatibility of hpe flexfabric port adapter with coreos issue report bug for fibre channel storage fcoe i am using hp flexfabric port adapter it is a converged network broadcom adapter i opened discussion on coreos user group about the compatibility of this adapter with coreos and found out that currently coreos don t have the kernel driver that supports fcoe offload config option scsi fcoe link to the discussion output of lspci v command container linux version cat etc os release pretty name debian gnu linux jessie name debian gnu linux version id version jessie id debian home url support url bug report url uname r coreos environment baremetal | 1 |
104,112 | 22,591,891,983 | IssuesEvent | 2022-06-28 20:47:40 | phetsims/mean-share-and-balance | https://api.github.com/repos/phetsims/mean-share-and-balance | opened | Tandem name should match associated variable/property name | dev:code-review | For code review #41 ...
By PhET-iO convention, tandem names are typically supposed to match their associated variable/property name (with exceptions made for unusual cases). There are at least 5 violations of that in this sim, and none of them qualifies as "unusual".
For example, in MeanShareAndBalanceScreenView.ts:
```
this.syncDataButton = new RectangularPushButton( {
...
tandem: options.tandem.createTandem( 'syncRepresentationsButton' ),
```
I added comment "//REVIEW tandem name does not match" to all of the cases that I identified. (Not necessarily a complete list.)
| 1.0 | Tandem name should match associated variable/property name - For code review #41 ...
By PhET-iO convention, tandem names are typically supposed to match their associated variable/property name (with exceptions made for unusual cases). There are at least 5 violations of that in this sim, and none of them qualifies as "unusual".
For example, in MeanShareAndBalanceScreenView.ts:
```
this.syncDataButton = new RectangularPushButton( {
...
tandem: options.tandem.createTandem( 'syncRepresentationsButton' ),
```
I added comment "//REVIEW tandem name does not match" to all of the cases that I identified. (Not necessarily a complete list.)
| non_usab | tandem name should match associated variable property name for code review by phet io convention tandem names are typically supposed to match their associated variable property name with exceptions made for unusual cases there are at least violations of that in this sim and none of them qualifies as unusual for example in meanshareandbalancescreenview ts this syncdatabutton new rectangularpushbutton tandem options tandem createtandem syncrepresentationsbutton i added comment review tandem name does not match to all of the cases that i identified not necessarily a complete list | 0 |
266,047 | 28,298,879,317 | IssuesEvent | 2023-04-10 02:51:04 | nidhi7598/linux-4.19.72 | https://api.github.com/repos/nidhi7598/linux-4.19.72 | closed | CVE-2019-19079 (High) detected in linuxlinux-4.19.254 - autoclosed | Mend: dependency security vulnerability | ## CVE-2019-19079 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the qrtr_tun_write_iter() function in net/qrtr/tun.c in the Linux kernel before 5.3 allows attackers to cause a denial of service (memory consumption), aka CID-a21b7f0cff19.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19079>CVE-2019-19079</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19079">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19079</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2019-19079 (High) detected in linuxlinux-4.19.254 - autoclosed - ## CVE-2019-19079 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.254</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/nidhi7598/linux-4.19.72/commit/10a8c99e4f60044163c159867bc6f5452c1c36e5">10a8c99e4f60044163c159867bc6f5452c1c36e5</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A memory leak in the qrtr_tun_write_iter() function in net/qrtr/tun.c in the Linux kernel before 5.3 allows attackers to cause a denial of service (memory consumption), aka CID-a21b7f0cff19.
<p>Publish Date: 2019-11-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-19079>CVE-2019-19079</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19079">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-19079</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: v5.3</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_usab | cve high detected in linuxlinux autoclosed cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files vulnerability details a memory leak in the qrtr tun write iter function in net qrtr tun c in the linux kernel before allows attackers to cause a denial of service memory consumption aka cid publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
24,559 | 23,917,484,741 | IssuesEvent | 2022-09-09 13:55:56 | liquibase/liquibase | https://api.github.com/repos/liquibase/liquibase | closed | context property in maven plugin does not tolerate spaces | TypeBug Severity3 ImpactMedium ThemeUsability DBAll hacktoberfest | Sample:
`
<changeSet id="43765-1" author="gvw" context="(alvara or stest) and default">
`
works fine, but
`
<changeSet id="43765-1" author="gvw" context="( alvara or stest ) and default">
`
doesn't.
| True | context property in maven plugin does not tolerate spaces - Sample:
`
<changeSet id="43765-1" author="gvw" context="(alvara or stest) and default">
`
works fine, but
`
<changeSet id="43765-1" author="gvw" context="( alvara or stest ) and default">
`
doesn't.
| usab | context property in maven plugin does not tolerate spaces sample works fine but doesn t | 1 |
2,861 | 3,214,642,455 | IssuesEvent | 2015-10-07 04:06:31 | cloudera/ibis | https://api.github.com/repos/cloudera/ibis | closed | Convenience for specifying join keys that are shared in left and right tables | usability | From the mailing list
```python
joined = T1.join(T2, ['chrom', 'alt', 'pos'])
``` | True | Convenience for specifying join keys that are shared in left and right tables - From the mailing list
```python
joined = T1.join(T2, ['chrom', 'alt', 'pos'])
``` | usab | convenience for specifying join keys that are shared in left and right tables from the mailing list python joined join | 1 |
17,448 | 12,043,914,183 | IssuesEvent | 2020-04-14 13:14:37 | godotengine/godot | https://api.github.com/repos/godotengine/godot | opened | Can't use arrows in ItemList after pressing a letter | bug discussion topic:gui usability | **Godot version:**
3.2.1
**Issue description:**
When you click ItemList, you can move through elements with arrow keys. But you can't anymore after pressing a letter to search something. You need to click again or wait few secs.
After checking the code, seems like incremental search has a separate path for arrow keys, but I don't much understand what is it for. In the end it just blocks arrow keys after searching.
**Steps to reproduce:**
1. Create ItemList with some elements
2. Press some letter
3. Try to use arrow keys | True | Can't use arrows in ItemList after pressing a letter - **Godot version:**
3.2.1
**Issue description:**
When you click ItemList, you can move through elements with arrow keys. But you can't anymore after pressing a letter to search something. You need to click again or wait few secs.
After checking the code, seems like incremental search has a separate path for arrow keys, but I don't much understand what is it for. In the end it just blocks arrow keys after searching.
**Steps to reproduce:**
1. Create ItemList with some elements
2. Press some letter
3. Try to use arrow keys | usab | can t use arrows in itemlist after pressing a letter godot version issue description when you click itemlist you can move through elements with arrow keys but you can t anymore after pressing a letter to search something you need to click again or wait few secs after checking the code seems like incremental search has a separate path for arrow keys but i don t much understand what is it for in the end it just blocks arrow keys after searching steps to reproduce create itemlist with some elements press some letter try to use arrow keys | 1 |
40,078 | 12,746,037,998 | IssuesEvent | 2020-06-26 15:14:39 | RG4421/developers | https://api.github.com/repos/RG4421/developers | opened | CVE-2019-11358 (Medium) detected in jquery-2.1.4.min.js, jquery-1.9.1.js | security vulnerability | ## CVE-2019-11358 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.9.1.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/developers/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>Path to vulnerable library: /developers/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.9.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/developers/node_modules/tinycolor2/test/index.html</p>
<p>Path to vulnerable library: /developers/node_modules/tinycolor2/test/../demo/jquery-1.9.1.js,/developers/node_modules/tinycolor2/demo/jquery-1.9.1.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.9.1.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/RG4421/developers/commit/09bff0d3b38e28079c6c900ddef39d33d88ab428">09bff0d3b38e28079c6c900ddef39d33d88ab428</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.4","isTransitiveDependency":false,"dependencyTree":"jquery:2.1.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.9.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.9.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"vulnerabilityIdentifier":"CVE-2019-11358","vulnerabilityDetails":"jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | True | CVE-2019-11358 (Medium) detected in jquery-2.1.4.min.js, jquery-1.9.1.js - ## CVE-2019-11358 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-2.1.4.min.js</b>, <b>jquery-1.9.1.js</b></p></summary>
<p>
<details><summary><b>jquery-2.1.4.min.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.1.4/jquery.min.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/developers/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>Path to vulnerable library: /developers/node_modules/js-base64/.attic/test-moment/index.html</p>
<p>
Dependency Hierarchy:
- :x: **jquery-2.1.4.min.js** (Vulnerable Library)
</details>
<details><summary><b>jquery-1.9.1.js</b></p></summary>
<p>JavaScript library for DOM operations</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.9.1/jquery.js</a></p>
<p>Path to dependency file: /tmp/ws-scm/developers/node_modules/tinycolor2/test/index.html</p>
<p>Path to vulnerable library: /developers/node_modules/tinycolor2/test/../demo/jquery-1.9.1.js,/developers/node_modules/tinycolor2/demo/jquery-1.9.1.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery-1.9.1.js** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/RG4421/developers/commit/09bff0d3b38e28079c6c900ddef39d33d88ab428">09bff0d3b38e28079c6c900ddef39d33d88ab428</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.
<p>Publish Date: 2019-04-20
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p>
<p>Release Date: 2019-04-20</p>
<p>Fix Resolution: 3.4.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"2.1.4","isTransitiveDependency":false,"dependencyTree":"jquery:2.1.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.9.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.9.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"3.4.0"}],"vulnerabilityIdentifier":"CVE-2019-11358","vulnerabilityDetails":"jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> --> | non_usab | cve medium detected in jquery min js jquery js cve medium severity vulnerability vulnerable libraries jquery min js jquery js jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm developers node modules js attic test moment index html path to vulnerable library developers node modules js attic test moment index html dependency hierarchy x jquery min js vulnerable library jquery js javascript library for dom operations library home page a href path to dependency file tmp ws scm developers node modules test index html path to vulnerable library developers node modules test demo jquery js developers node modules demo jquery js dependency hierarchy x jquery js vulnerable library found in head commit a href vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype vulnerabilityurl | 0 |
21,842 | 17,873,079,503 | IssuesEvent | 2021-09-06 19:33:42 | bevyengine/bevy | https://api.github.com/repos/bevyengine/bevy | closed | Access Diagnostic::history field | C-Usability A-Diagnostics | ## What problem does this solve or what need does it fill?
I would like to access the history data of a diagnostic for drawing a graph using egui.
## What solution would you like?
The history field should be made public. Or a function to access the underlying history data.
## What alternative(s) have you considered?
Currently I'm storing the Diagnostic value in separate resource but this would be unnecessary if i could access the history stored with the diagnostic.
| True | Access Diagnostic::history field - ## What problem does this solve or what need does it fill?
I would like to access the history data of a diagnostic for drawing a graph using egui.
## What solution would you like?
The history field should be made public. Or a function to access the underlying history data.
## What alternative(s) have you considered?
Currently I'm storing the Diagnostic value in separate resource but this would be unnecessary if i could access the history stored with the diagnostic.
| usab | access diagnostic history field what problem does this solve or what need does it fill i would like to access the history data of a diagnostic for drawing a graph using egui what solution would you like the history field should be made public or a function to access the underlying history data what alternative s have you considered currently i m storing the diagnostic value in separate resource but this would be unnecessary if i could access the history stored with the diagnostic | 1 |
16,272 | 10,720,111,982 | IssuesEvent | 2019-10-26 15:22:22 | bronzehedwick/chrisdeluca | https://api.github.com/repos/bronzehedwick/chrisdeluca | closed | Create 2 column layout with nav | look + feel usability | Use a 2 column layout, using the new grid CSS module.
The left column should include a the navigation, and be ordered above the footer on small screens. | True | Create 2 column layout with nav - Use a 2 column layout, using the new grid CSS module.
The left column should include a the navigation, and be ordered above the footer on small screens. | usab | create column layout with nav use a column layout using the new grid css module the left column should include a the navigation and be ordered above the footer on small screens | 1 |
27,752 | 5,094,049,665 | IssuesEvent | 2017-01-03 09:49:53 | GoldenSoftwareLtd/gedemin | https://api.github.com/repos/GoldenSoftwareLtd/gedemin | closed | Товар и количество в накладной отображаются но нет цены ни покупной ни с НДС | Depot Priority-Critical Type-Defect | Originally reported on Google Code with ID 2991
```
допустим собрали мы какой то товар через акт переработки в материальном складе, поставили
цену и так далее, поставили на склад N далее при попытке отпустить этот товар через
отпуск товара на сторону (оптовая торговля) возникает следующая ошибка товар и количество
в накладной отображаются но нет цены ни покупной ни с НДС, при попытке набить цену
вручную выдает следующую ошибку "По данному тмц не достаточное количество остатков",
раньше было все нормально и операция проделывалась не один раз, но как только я сбросил
ваши настройки с сайта и обновил настройки торговли а именно склад-торговля (хранилище,
метаданные, макросы, данные) все перестало работать, если материал перемещать через
внутреннее перемещение материалов, то показываеться и цена и ндс все не обходимое,
но опять товар выбрать дает в отпуск товара на сторону(оптовая торговля). то происходит
тоже самое что описанное выше.
```
Reported by `gs1994` on 2012-12-08 13:10:45
| 1.0 | Товар и количество в накладной отображаются но нет цены ни покупной ни с НДС - Originally reported on Google Code with ID 2991
```
допустим собрали мы какой то товар через акт переработки в материальном складе, поставили
цену и так далее, поставили на склад N далее при попытке отпустить этот товар через
отпуск товара на сторону (оптовая торговля) возникает следующая ошибка товар и количество
в накладной отображаются но нет цены ни покупной ни с НДС, при попытке набить цену
вручную выдает следующую ошибку "По данному тмц не достаточное количество остатков",
раньше было все нормально и операция проделывалась не один раз, но как только я сбросил
ваши настройки с сайта и обновил настройки торговли а именно склад-торговля (хранилище,
метаданные, макросы, данные) все перестало работать, если материал перемещать через
внутреннее перемещение материалов, то показываеться и цена и ндс все не обходимое,
но опять товар выбрать дает в отпуск товара на сторону(оптовая торговля). то происходит
тоже самое что описанное выше.
```
Reported by `gs1994` on 2012-12-08 13:10:45
| non_usab | товар и количество в накладной отображаются но нет цены ни покупной ни с ндс originally reported on google code with id допустим собрали мы какой то товар через акт переработки в материальном складе поставили цену и так далее поставили на склад n далее при попытке отпустить этот товар через отпуск товара на сторону оптовая торговля возникает следующая ошибка товар и количество в накладной отображаются но нет цены ни покупной ни с ндс при попытке набить цену вручную выдает следующую ошибку по данному тмц не достаточное количество остатков раньше было все нормально и операция проделывалась не один раз но как только я сбросил ваши настройки с сайта и обновил настройки торговли а именно склад торговля хранилище метаданные макросы данные все перестало работать если материал перемещать через внутреннее перемещение материалов то показываеться и цена и ндс все не обходимое но опять товар выбрать дает в отпуск товара на сторону оптовая торговля то происходит тоже самое что описанное выше reported by on | 0 |
17,643 | 12,222,200,019 | IssuesEvent | 2020-05-02 12:05:18 | tideflow-io/tideflow | https://api.github.com/repos/tideflow-io/tideflow | closed | Tasks settings sidebar [workflow's editor] | :gift: Rewarded on Issuehunt 🎨 usability & ui 🏆 epic | <!-- Issuehunt Badges -->
[<img alt="Issuehunt badges" src="https://img.shields.io/badge/IssueHunt-%24150%20Rewarded-%237E24E3.svg" />](https://issuehunt.io/r/tideflow-io/tideflow/issues/70)
<!-- /Issuehunt Badges -->
As a user, I can edit a task settings by clicking on "edit". The settings editor will appear in a new sticky & scrollable sidebar. (instead of being shows in the task's card)
## Current settings editor
<img width="438" alt="Captura de pantalla 2020-04-05 a las 21 30 29" src="https://user-images.githubusercontent.com/6388629/78508112-acfef580-7784-11ea-8b9c-029dde49f998.png">
## Expected settings editor
<img width="762" alt="Captura de pantalla 2020-04-05 a las 21 30 38" src="https://user-images.githubusercontent.com/6388629/78508118-b4be9a00-7784-11ea-9e7a-d2afa1f95e52.png">
<!-- Issuehunt content -->
---
<details>
<summary>
<b>IssueHunt Summary</b>
</summary>
#### [<img src='https://avatars2.githubusercontent.com/u/332141?v=4' alt='bobthekingofegypt' width=24 height=24> bobthekingofegypt](https://issuehunt.io/u/bobthekingofegypt) has been rewarded.
### Backers (Total: $150.00)
- [<img src='https://avatars2.githubusercontent.com/u/6388629?v=4' alt='joseconstela' width=24 height=24> joseconstela](https://issuehunt.io/u/joseconstela) ($150.00)
### Submitted pull Requests
- [#72 Add sidebar for showing edit form](https://issuehunt.io/r/tideflow-io/tideflow/pull/72)
---
### Tips
- Checkout the [Issuehunt explorer](https://issuehunt.io/r/tideflow-io/tideflow/) to discover more funded issues.
- Need some help from other developers? [Add your repositories](https://issuehunt.io/r/new) on IssueHunt to raise funds.
</details>
<!-- /Issuehunt content--> | True | Tasks settings sidebar [workflow's editor] - <!-- Issuehunt Badges -->
[<img alt="Issuehunt badges" src="https://img.shields.io/badge/IssueHunt-%24150%20Rewarded-%237E24E3.svg" />](https://issuehunt.io/r/tideflow-io/tideflow/issues/70)
<!-- /Issuehunt Badges -->
As a user, I can edit a task settings by clicking on "edit". The settings editor will appear in a new sticky & scrollable sidebar. (instead of being shows in the task's card)
## Current settings editor
<img width="438" alt="Captura de pantalla 2020-04-05 a las 21 30 29" src="https://user-images.githubusercontent.com/6388629/78508112-acfef580-7784-11ea-8b9c-029dde49f998.png">
## Expected settings editor
<img width="762" alt="Captura de pantalla 2020-04-05 a las 21 30 38" src="https://user-images.githubusercontent.com/6388629/78508118-b4be9a00-7784-11ea-9e7a-d2afa1f95e52.png">
<!-- Issuehunt content -->
---
<details>
<summary>
<b>IssueHunt Summary</b>
</summary>
#### [<img src='https://avatars2.githubusercontent.com/u/332141?v=4' alt='bobthekingofegypt' width=24 height=24> bobthekingofegypt](https://issuehunt.io/u/bobthekingofegypt) has been rewarded.
### Backers (Total: $150.00)
- [<img src='https://avatars2.githubusercontent.com/u/6388629?v=4' alt='joseconstela' width=24 height=24> joseconstela](https://issuehunt.io/u/joseconstela) ($150.00)
### Submitted pull Requests
- [#72 Add sidebar for showing edit form](https://issuehunt.io/r/tideflow-io/tideflow/pull/72)
---
### Tips
- Checkout the [Issuehunt explorer](https://issuehunt.io/r/tideflow-io/tideflow/) to discover more funded issues.
- Need some help from other developers? [Add your repositories](https://issuehunt.io/r/new) on IssueHunt to raise funds.
</details>
<!-- /Issuehunt content--> | usab | tasks settings sidebar as a user i can edit a task settings by clicking on edit the settings editor will appear in a new sticky scrollable sidebar instead of being shows in the task s card current settings editor img width alt captura de pantalla a las src expected settings editor img width alt captura de pantalla a las src issuehunt summary has been rewarded backers total submitted pull requests tips checkout the to discover more funded issues need some help from other developers on issuehunt to raise funds | 1 |
4,862 | 3,897,238,833 | IssuesEvent | 2016-04-16 09:02:33 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 15792060: redownload already owned movies | classification:usability reproducible:always status:open | #### Description
Summary:
Dear iTunes-Store-Team,
why isn't it possible to re-download already bought movies,
like it is possible with apps or music?
Steps to Reproduce:
buy and download a movie
(watch &) delete it
try to download it again
Expected Results:
the movie downloads again, if it isn't present on my harddrive
(I only have a 512 ssd)
Actual Results:
a popup says, that maybe I like to buy the movie a second time
(see attached image)
Version:
OSX, all Versions
Notes:
Configuration:
Attachments:
'Screen Shot 2014-01-10 at 15.51.01.png' was successfully uploaded.
-
Product Version: all
Created: 2014-01-10T15:02:10.722498
Originated: 2014-01-10T00:00:00
Open Radar Link: http://www.openradar.me/15792060 | True | 15792060: redownload already owned movies - #### Description
Summary:
Dear iTunes-Store-Team,
why isn't it possible to re-download already bought movies,
like it is possible with apps or music?
Steps to Reproduce:
buy and download a movie
(watch &) delete it
try to download it again
Expected Results:
the movie downloads again, if it isn't present on my harddrive
(I only have a 512 ssd)
Actual Results:
a popup says, that maybe I like to buy the movie a second time
(see attached image)
Version:
OSX, all Versions
Notes:
Configuration:
Attachments:
'Screen Shot 2014-01-10 at 15.51.01.png' was successfully uploaded.
-
Product Version: all
Created: 2014-01-10T15:02:10.722498
Originated: 2014-01-10T00:00:00
Open Radar Link: http://www.openradar.me/15792060 | usab | redownload already owned movies description summary dear itunes store team why isn t it possible to re download already bought movies like it is possible with apps or music steps to reproduce buy and download a movie watch delete it try to download it again expected results the movie downloads again if it isn t present on my harddrive i only have a ssd actual results a popup says that maybe i like to buy the movie a second time see attached image version osx all versions notes configuration attachments screen shot at png was successfully uploaded product version all created originated open radar link | 1 |
9,932 | 6,520,083,104 | IssuesEvent | 2017-08-28 15:10:55 | raiden-network/raiden | https://api.github.com/repos/raiden-network/raiden | closed | Raiden fails on open channel with unregistered token address | usability | ## Problem Definition
Trying to open a channel with an unregisted token (a valid token/contract address, but which isn't yet returned in the list from `/api/1/tokens` endpoint) causes an exception in Raiden, and API request to return code 500.
### Relevant part of the exception:
```ERROR in app: Exception on /api/1/channels [PUT]
Traceback (most recent call last):
...
File "/apps/raiden/raiden/api/v1/resources.py", line 55, in put
return self.rest_api.open(**kwargs)
File "/apps/raiden/raiden/api/rest.py", line 183, in open
settle_timeout
File "/apps/raiden/raiden/api/python.py", line 140, in open
channel_manager = self.raiden.chain.manager_by_token(token_address)
File "/apps/raiden/raiden/network/rpc/client.py", line 360, in manager_by_token
poll_timeout=self.poll_timeout,
File "/apps/raiden/raiden/network/rpc/client.py", line 709, in __init__
'latest',
File "/usr/local/lib/python2.7/dist-packages/pyethapp/rpc_client.py", line 413, in call
raise JSONRPCClientReplyError(jsonrpc_reply.error)
JSONRPCClientReplyError: Invalid params: invalid length 0, expected a 0x-prefixed, padded, hex-encoded hash with length 40.
```
### Example requisition causing this error
```
curl -v -H "Content-Type: application/json" -X PUT -d '{"partner_address":"0x000D91Cf263a11F9BfCeE3752E5B03FC1196CE98","token_address":"0x9aBa529db3FF2D8409A1da4C9eB148879b046700","balance":0,"settle_timeout":600}' http://localhost:5001/api/1/channels
```
## Solution
Check why `manager_by_token` is failing to generate a valid `manager_address` when fed a `token_address` not yet on registry, and fix it, so it gets properly returned.
## Tasklist
- [x] Create test for opening channel with unregistered token
- [x] Fix bug so raiden acts accordingly
- [ ] Add a changelog entry
## System Description
Raiden in current master (as of 2017-07-09), running on a Linux system with Linux kernel 4.11.7, over docker. Using parity nightly, but further testing have shown it is not a parity bug, as the requisition goes empty and fails even with geth 1.6.6. | True | Raiden fails on open channel with unregistered token address - ## Problem Definition
Trying to open a channel with an unregisted token (a valid token/contract address, but which isn't yet returned in the list from `/api/1/tokens` endpoint) causes an exception in Raiden, and API request to return code 500.
### Relevant part of the exception:
```ERROR in app: Exception on /api/1/channels [PUT]
Traceback (most recent call last):
...
File "/apps/raiden/raiden/api/v1/resources.py", line 55, in put
return self.rest_api.open(**kwargs)
File "/apps/raiden/raiden/api/rest.py", line 183, in open
settle_timeout
File "/apps/raiden/raiden/api/python.py", line 140, in open
channel_manager = self.raiden.chain.manager_by_token(token_address)
File "/apps/raiden/raiden/network/rpc/client.py", line 360, in manager_by_token
poll_timeout=self.poll_timeout,
File "/apps/raiden/raiden/network/rpc/client.py", line 709, in __init__
'latest',
File "/usr/local/lib/python2.7/dist-packages/pyethapp/rpc_client.py", line 413, in call
raise JSONRPCClientReplyError(jsonrpc_reply.error)
JSONRPCClientReplyError: Invalid params: invalid length 0, expected a 0x-prefixed, padded, hex-encoded hash with length 40.
```
### Example requisition causing this error
```
curl -v -H "Content-Type: application/json" -X PUT -d '{"partner_address":"0x000D91Cf263a11F9BfCeE3752E5B03FC1196CE98","token_address":"0x9aBa529db3FF2D8409A1da4C9eB148879b046700","balance":0,"settle_timeout":600}' http://localhost:5001/api/1/channels
```
## Solution
Check why `manager_by_token` is failing to generate a valid `manager_address` when fed a `token_address` not yet on registry, and fix it, so it gets properly returned.
## Tasklist
- [x] Create test for opening channel with unregistered token
- [x] Fix bug so raiden acts accordingly
- [ ] Add a changelog entry
## System Description
Raiden in current master (as of 2017-07-09), running on a Linux system with Linux kernel 4.11.7, over docker. Using parity nightly, but further testing have shown it is not a parity bug, as the requisition goes empty and fails even with geth 1.6.6. | usab | raiden fails on open channel with unregistered token address problem definition trying to open a channel with an unregisted token a valid token contract address but which isn t yet returned in the list from api tokens endpoint causes an exception in raiden and api request to return code relevant part of the exception error in app exception on api channels traceback most recent call last file apps raiden raiden api resources py line in put return self rest api open kwargs file apps raiden raiden api rest py line in open settle timeout file apps raiden raiden api python py line in open channel manager self raiden chain manager by token token address file apps raiden raiden network rpc client py line in manager by token poll timeout self poll timeout file apps raiden raiden network rpc client py line in init latest file usr local lib dist packages pyethapp rpc client py line in call raise jsonrpcclientreplyerror jsonrpc reply error jsonrpcclientreplyerror invalid params invalid length expected a prefixed padded hex encoded hash with length example requisition causing this error curl v h content type application json x put d partner address token address balance settle timeout solution check why manager by token is failing to generate a valid manager address when fed a token address not yet on registry and fix it so it gets properly returned tasklist create test for opening channel with unregistered token fix bug so raiden acts accordingly add a changelog entry system description raiden in current master as of running on a linux system with linux kernel over docker using parity nightly but further testing have shown it is not a parity bug as the requisition goes empty and fails even with geth | 1 |
3,794 | 2,610,069,155 | IssuesEvent | 2015-02-26 18:20:17 | chrsmith/jsjsj122 | https://api.github.com/repos/chrsmith/jsjsj122 | opened | 台州割包茎到哪家医院效果好 | auto-migrated Priority-Medium Type-Defect | ```
台州割包茎到哪家医院效果好【台州五洲生殖医院】24小时健
康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:
台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104�
��108、118、198及椒江一金清公交车直达枫南小区,乘坐107、105
、109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 9:32 | 1.0 | 台州割包茎到哪家医院效果好 - ```
台州割包茎到哪家医院效果好【台州五洲生殖医院】24小时健
康咨询热线:0576-88066933-(扣扣800080609)-(微信号tzwzszyy)医院地址:
台州市椒江区枫南路229号(枫南大转盘旁)乘车线路:乘坐104�
��108、118、198及椒江一金清公交车直达枫南小区,乘坐107、105
、109、112、901、 902公交车到星星广场下车,步行即可到院。
诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,��
�精,无精。包皮包茎,精索静脉曲张,淋病等。
台州五洲生殖医院是台州最大的男科医院,权威专家在线免��
�咨询,拥有专业完善的男科检查治疗设备,严格按照国家标�
��收费。尖端医疗设备,与世界同步。权威专家,成就专业典
范。人性化服务,一切以患者为中心。
看男科就选台州五洲生殖医院,专业男科为男人。
```
-----
Original issue reported on code.google.com by `poweragr...@gmail.com` on 30 May 2014 at 9:32 | non_usab | 台州割包茎到哪家医院效果好 台州割包茎到哪家医院效果好【台州五洲生殖医院】 康咨询热线 微信号tzwzszyy 医院地址 (枫南大转盘旁)乘车线路 � �� 、 、 , 、 、 、 、 、 ,步行即可到院。 诊疗项目:阳痿,早泄,前列腺炎,前列腺增生,龟头炎,�� �精,无精。包皮包茎,精索静脉曲张,淋病等。 台州五洲生殖医院是台州最大的男科医院,权威专家在线免�� �咨询,拥有专业完善的男科检查治疗设备,严格按照国家标� ��收费。尖端医疗设备,与世界同步。权威专家,成就专业典 范。人性化服务,一切以患者为中心。 看男科就选台州五洲生殖医院,专业男科为男人。 original issue reported on code google com by poweragr gmail com on may at | 0 |
162,812 | 12,692,149,728 | IssuesEvent | 2020-06-21 20:47:58 | M0nica/ambition-fund-website | https://api.github.com/repos/M0nica/ambition-fund-website | opened | [Increase Test Coverage] Navigation | test-coverage | - jest/react-testing-library unit tests should be added to the nav component to confirm that the link to homepage, mission and opportunities appear as expected.
- this functionality should be tested both within the context of the index page and non-index page (i.e., testing https://github.com/M0nica/ambition-fund-website/blob/2d388b90541c4493423e6285ed3285d7ae5e8261/src/components/common/navigation/navigation.js#L54) | 1.0 | [Increase Test Coverage] Navigation - - jest/react-testing-library unit tests should be added to the nav component to confirm that the link to homepage, mission and opportunities appear as expected.
- this functionality should be tested both within the context of the index page and non-index page (i.e., testing https://github.com/M0nica/ambition-fund-website/blob/2d388b90541c4493423e6285ed3285d7ae5e8261/src/components/common/navigation/navigation.js#L54) | non_usab | navigation jest react testing library unit tests should be added to the nav component to confirm that the link to homepage mission and opportunities appear as expected this functionality should be tested both within the context of the index page and non index page i e testing | 0 |
7,755 | 5,191,825,948 | IssuesEvent | 2017-01-22 00:34:46 | piwik/piwik | https://api.github.com/repos/piwik/piwik | closed | In Users management, on hover, display a cursor over green/red icons | c: Usability Easy pick Enhancement | Currently one cannot guess that these icons are clickable.
- Let's add a cursor on hover
- Maybe even a tooltip
| True | In Users management, on hover, display a cursor over green/red icons - Currently one cannot guess that these icons are clickable.
- Let's add a cursor on hover
- Maybe even a tooltip
| usab | in users management on hover display a cursor over green red icons currently one cannot guess that these icons are clickable let s add a cursor on hover maybe even a tooltip | 1 |
12,870 | 8,140,308,908 | IssuesEvent | 2018-08-20 20:39:15 | NCAR/VAPOR | https://api.github.com/repos/NCAR/VAPOR | closed | Barb renderer: weird default rendering | Fixed Usability | To reproduce:
1. Open VAPOR
2. Import NetCDF-CF data: `/home/shaomeng/data/Source/POP/BryanASP/one_degree/monthly`
2. Create a Barb renderer. Enable it.
4. The default rendering looks weird, as the screenshot shows.

| True | Barb renderer: weird default rendering - To reproduce:
1. Open VAPOR
2. Import NetCDF-CF data: `/home/shaomeng/data/Source/POP/BryanASP/one_degree/monthly`
2. Create a Barb renderer. Enable it.
4. The default rendering looks weird, as the screenshot shows.

| usab | barb renderer weird default rendering to reproduce open vapor import netcdf cf data home shaomeng data source pop bryanasp one degree monthly create a barb renderer enable it the default rendering looks weird as the screenshot shows | 1 |
67,656 | 21,042,358,152 | IssuesEvent | 2022-03-31 13:24:14 | vector-im/element-android | https://api.github.com/repos/vector-im/element-android | closed | No loading spinner is shown while threads list is loading | T-Defect S-Major O-Frequent A-Threads Z-ThreadsInternalTesting Z-Labs | ### Steps to reproduce
1. Go to a room with lots of threads (e.g. Threads internal)
2. View the thread list on a fresh login
### Outcome
#### What did you expect?
To either see the thread list or a loading spinner telling me whats up
#### What happened instead?
The threads list looked empty for about 20 seconds, I assumed it was broken somehow
### Your phone model
Galaxy S20
### Operating system version
Android 11
### Application version and app store
1.4.6-dev [40104060] (G-b6874) feature/aris/thread_live_thread_list
### Homeserver
matrix.org
### Will you send logs?
No | 1.0 | No loading spinner is shown while threads list is loading - ### Steps to reproduce
1. Go to a room with lots of threads (e.g. Threads internal)
2. View the thread list on a fresh login
### Outcome
#### What did you expect?
To either see the thread list or a loading spinner telling me whats up
#### What happened instead?
The threads list looked empty for about 20 seconds, I assumed it was broken somehow
### Your phone model
Galaxy S20
### Operating system version
Android 11
### Application version and app store
1.4.6-dev [40104060] (G-b6874) feature/aris/thread_live_thread_list
### Homeserver
matrix.org
### Will you send logs?
No | non_usab | no loading spinner is shown while threads list is loading steps to reproduce go to a room with lots of threads e g threads internal view the thread list on a fresh login outcome what did you expect to either see the thread list or a loading spinner telling me whats up what happened instead the threads list looked empty for about seconds i assumed it was broken somehow your phone model galaxy operating system version android application version and app store dev g feature aris thread live thread list homeserver matrix org will you send logs no | 0 |
288,667 | 31,864,056,067 | IssuesEvent | 2023-09-15 13:00:33 | ITISFoundation/osparc-ops-environments | https://api.github.com/repos/ITISFoundation/osparc-ops-environments | closed | Rotate SSH Keys for autoscaled machines | p:mid-prio SECURITY | We use ec2 instance template to spin up autoscaled machines on cloud deployments when needed. We need to rotate SSH keys for these machines (i.e. update ec2 instance template).
We alredy have new SSH keys created by @YuryHrytsuk.
@mguidon should have more accurate information about the ec2 instance template | True | Rotate SSH Keys for autoscaled machines - We use ec2 instance template to spin up autoscaled machines on cloud deployments when needed. We need to rotate SSH keys for these machines (i.e. update ec2 instance template).
We alredy have new SSH keys created by @YuryHrytsuk.
@mguidon should have more accurate information about the ec2 instance template | non_usab | rotate ssh keys for autoscaled machines we use instance template to spin up autoscaled machines on cloud deployments when needed we need to rotate ssh keys for these machines i e update instance template we alredy have new ssh keys created by yuryhrytsuk mguidon should have more accurate information about the instance template | 0 |
10,322 | 6,671,047,814 | IssuesEvent | 2017-10-04 04:25:02 | loconomics/loconomics | https://api.github.com/repos/loconomics/loconomics | closed | Meet 1.4.2 - Audio Control & 2.2.2 - Pause, Stop, Hide | C: Usability F: Accessbility | ## Summary
Provides that there is a way to stop, pause, mute, or adjust volume with audio that plays automatically. Provides for user control over moving, blinking, scrolling, and information that updates automatically
**Conformance Level:** A
**Existing 508 Corresponding Provision:** None & 1194.21(h)
### Provide controls for content that starts automatically
Provide visible controls to allow users to stop any animations or auto-playing sound. This applies to carousels, image sliders, background sound, and videos.

For more information
WCAG
[Audio Control 1.4.2 ](https://www.w3.org/WAI/WCAG20/quickref/#qr-visual-audio-contrast-dis-audio)([Understanding 1.4.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/visual-audio-contrast-dis-audio.html))
[Pause, Stop, Hide 2.2.2](https://www.w3.org/WAI/WCAG20/quickref/#qr-time-limits-pause) ([Understanding 2.2.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/time-limits-pause.html))
Tutorial
[Carousel Concepts](https://www.w3.org/WAI/tutorials/carousels/)
| True | Meet 1.4.2 - Audio Control & 2.2.2 - Pause, Stop, Hide - ## Summary
Provides that there is a way to stop, pause, mute, or adjust volume with audio that plays automatically. Provides for user control over moving, blinking, scrolling, and information that updates automatically
**Conformance Level:** A
**Existing 508 Corresponding Provision:** None & 1194.21(h)
### Provide controls for content that starts automatically
Provide visible controls to allow users to stop any animations or auto-playing sound. This applies to carousels, image sliders, background sound, and videos.

For more information
WCAG
[Audio Control 1.4.2 ](https://www.w3.org/WAI/WCAG20/quickref/#qr-visual-audio-contrast-dis-audio)([Understanding 1.4.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/visual-audio-contrast-dis-audio.html))
[Pause, Stop, Hide 2.2.2](https://www.w3.org/WAI/WCAG20/quickref/#qr-time-limits-pause) ([Understanding 2.2.2](https://www.w3.org/TR/UNDERSTANDING-WCAG20/time-limits-pause.html))
Tutorial
[Carousel Concepts](https://www.w3.org/WAI/tutorials/carousels/)
| usab | meet audio control pause stop hide summary provides that there is a way to stop pause mute or adjust volume with audio that plays automatically provides for user control over moving blinking scrolling and information that updates automatically conformance level a existing corresponding provision none h provide controls for content that starts automatically provide visible controls to allow users to stop any animations or auto playing sound this applies to carousels image sliders background sound and videos for more information wcag tutorial | 1 |
774,083 | 27,182,625,491 | IssuesEvent | 2023-02-18 20:38:17 | AUBGTheHUB/spa-website-2022 | https://api.github.com/repos/AUBGTheHUB/spa-website-2022 | closed | Disable Team Section when website is viewed from an in-app browser. | high priority frontend | Section's animations stutter when viewed from an in-app browser such as Facebook's one. We should also remove the nav anchor when this happens -> set global state in `localstorage` or something so that the nav in jobs can reflect the fact that team section is missing. | 1.0 | Disable Team Section when website is viewed from an in-app browser. - Section's animations stutter when viewed from an in-app browser such as Facebook's one. We should also remove the nav anchor when this happens -> set global state in `localstorage` or something so that the nav in jobs can reflect the fact that team section is missing. | non_usab | disable team section when website is viewed from an in app browser section s animations stutter when viewed from an in app browser such as facebook s one we should also remove the nav anchor when this happens set global state in localstorage or something so that the nav in jobs can reflect the fact that team section is missing | 0 |
160,423 | 13,788,562,220 | IssuesEvent | 2020-10-09 07:26:15 | hackforla/climate-collabathon | https://api.github.com/repos/hackforla/climate-collabathon | opened | Define MVP | documentation role: product management | ### Overview
We need an MVP for team and org visibility into process.
### Dependencies
- [ ] Persona development #11
### Action Items
- [ ] Review personas
- [ ] Gather requirements from manual finding process
- [ ] Gather requirements from onboarding process
- [ ] Gather details around friction during the manual processes (anti-goals)
- [ ] Define ideal feature set
- [ ] Prioritize features by proof of concept, MVP and iterations
### Resources/Instructions
Insert link to personas here
| 1.0 | Define MVP - ### Overview
We need an MVP for team and org visibility into process.
### Dependencies
- [ ] Persona development #11
### Action Items
- [ ] Review personas
- [ ] Gather requirements from manual finding process
- [ ] Gather requirements from onboarding process
- [ ] Gather details around friction during the manual processes (anti-goals)
- [ ] Define ideal feature set
- [ ] Prioritize features by proof of concept, MVP and iterations
### Resources/Instructions
Insert link to personas here
| non_usab | define mvp overview we need an mvp for team and org visibility into process dependencies persona development action items review personas gather requirements from manual finding process gather requirements from onboarding process gather details around friction during the manual processes anti goals define ideal feature set prioritize features by proof of concept mvp and iterations resources instructions insert link to personas here | 0 |
753,001 | 26,337,348,089 | IssuesEvent | 2023-01-10 15:14:26 | graphnet-team/graphnet | https://api.github.com/repos/graphnet-team/graphnet | opened | Cannot convert upgrade i3 files to sqlite anymore | bug high priority | **Describe the bug**
I used to be able to convert these files to sqlite. I then made the mistake of updating my main branch on my local fork. Now I can't convert the files. GraphNeT now incorrectly states that no gcd was given.
**To Reproduce**
Steps to reproduce the behavior:
1. Load cvmfs for upgrade
2. install graphnet main
3. run script (adjust outdir)
4. see error
**Expected behavior**
should. work.
**Full traceback**
Please include the full error message to allow for debugging
```
Traceback (most recent call last):
File "make_database.py", line 78, in <module>
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
main_icecube_upgrade(backend)
File "make_database.py", line 71, in main_icecube_upgrade
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
converter(inputs)
File "/lustre/hpc/hep/pcs557/github/graphnet/src/graphnet/data/dataconverter.py", line 215, in __call__
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
self.execute(filesets)
File "/lustre/hpc/hep/pcs557/github/graphnet/src/graphnet/data/dataconverter.py", line 286, in execute
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
pool = self._iterate_over_individual_files(filesets)
File "/lustre/hpc/hep/pcs557/github/graphnet/src/graphnet/data/dataconverter.py", line 331, in _iterate_over_individual_files
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
self._process_file, tqdm(args, unit="file(s)", colour="green")
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
File "/cvmfs/icecube.opensciencegrid.org/py3-v4.1.1/RHEL_7_x86_64/lib/python3.7/multiprocessing/pool.py", line 748, in next
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
raise value
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
RuntimeError: no frame to pop (in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
```
**Additional context**
```
import logging
import os
import logging
import os
from graphnet.utilities.logging import get_logger
from graphnet.data.extractors import (
I3FeatureExtractorIceCubeUpgrade,
I3TruthExtractor,
I3GenericExtractor,
I3PISAExtractor,
I3PulseNoiseTruthFlagIceCubeUpgrade,
)
from graphnet.data.parquet import ParquetDataConverter
from graphnet.data.sqlite import SQLiteDataConverter
logger = get_logger(level=logging.INFO)
CONVERTER_CLASS = {
"sqlite": SQLiteDataConverter,
"parquet": ParquetDataConverter,
}
def main_icecube_upgrade(backend: str):
"""Convert IceCube-Upgrade I3 files to intermediate `backend` format."""
# Check(s)
assert backend in CONVERTER_CLASS
database_name = 'dev_step4_upgrade_028_21.11.2022_noise_test'
#inputs = ["/data/sim/IceCubeUpgrade/noise/step4/880028"]
inputs = ["/data/sim/IceCubeUpgrade/genie/step4/140028",
"/data/sim/IceCubeUpgrade/genie/step4/141028",
"/data/sim/IceCubeUpgrade/genie/step4/120028",
"/data/sim/IceCubeUpgrade/genie/step4/160028",
"/data/sim/IceCubeUpgrade/muongun/step4/130028",
"/data/sim/IceCubeUpgrade/noise/step4/880028"]
outdir = f"/data/user/rorsoe/databases/{database_name}"
workers = 1
converter = CONVERTER_CLASS[backend](
[
I3TruthExtractor(),
I3PISAExtractor(),
I3FeatureExtractorIceCubeUpgrade(
"SplitInIcePulses"
),
I3PulseNoiseTruthFlagIceCubeUpgrade(
"SplitInIcePulses_TruthFlags"
),
I3FeatureExtractorIceCubeUpgrade(
"SplitIceCubePulsesTWSRT"
),
I3PulseNoiseTruthFlagIceCubeUpgrade(
"SplitIceCubePulsesTWSRT_TruthFlags"
),
I3FeatureExtractorIceCubeUpgrade(
"SplitInIcePulses_GraphSage_Pulses"
),
I3PulseNoiseTruthFlagIceCubeUpgrade(
"SplitInIcePulses_GraphSage_Predictions"
),
I3PulseNoiseTruthFlagIceCubeUpgrade(
"SplitInIcePulses_dynedge_Predictions"
),
I3FeatureExtractorIceCubeUpgrade(
"SplitInIcePulses_dynedge_Pulses"
),
],
outdir + '/tmp',
workers=workers,
icetray_verbose=1,
gcd_rescue = '/data/sim/IceCubeUpgrade/genie/step4/140028/GeoCalibDetectorStatus_ICUpgrade.v58.mixed.V0.i3.bz2'
)
os.makedirs(outdir + '/data/')
converter(inputs)
#if backend == "sqlite":
# converter.merge_files(outdir + '/data/' + database_name)
if __name__ == "__main__":
backend = "sqlite"
main_icecube_upgrade(backend)
```
| 1.0 | Cannot convert upgrade i3 files to sqlite anymore - **Describe the bug**
I used to be able to convert these files to sqlite. I then made the mistake of updating my main branch on my local fork. Now I can't convert the files. GraphNeT now incorrectly states that no gcd was given.
**To Reproduce**
Steps to reproduce the behavior:
1. Load cvmfs for upgrade
2. install graphnet main
3. run script (adjust outdir)
4. see error
**Expected behavior**
should. work.
**Full traceback**
Please include the full error message to allow for debugging
```
Traceback (most recent call last):
File "make_database.py", line 78, in <module>
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
main_icecube_upgrade(backend)
File "make_database.py", line 71, in main_icecube_upgrade
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
converter(inputs)
File "/lustre/hpc/hep/pcs557/github/graphnet/src/graphnet/data/dataconverter.py", line 215, in __call__
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
self.execute(filesets)
File "/lustre/hpc/hep/pcs557/github/graphnet/src/graphnet/data/dataconverter.py", line 286, in execute
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
pool = self._iterate_over_individual_files(filesets)
File "/lustre/hpc/hep/pcs557/github/graphnet/src/graphnet/data/dataconverter.py", line 331, in _iterate_over_individual_files
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
self._process_file, tqdm(args, unit="file(s)", colour="green")
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
File "/cvmfs/icecube.opensciencegrid.org/py3-v4.1.1/RHEL_7_x86_64/lib/python3.7/multiprocessing/pool.py", line 748, in next
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
raise value
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
RuntimeError: no frame to pop (in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
FATAL (dataio): no frame to pop (I3File.cxx:142 in I3FramePtr dataio::I3File::pop_frame(I3Frame::Stream))
graphnet: ERROR 2023-01-10 16:06:07 - I3TruthExtractor.error - No GCD file was provided and no G-frame was found. Exiting.
```
**Additional context**
```
import logging
import os
import logging
import os
from graphnet.utilities.logging import get_logger
from graphnet.data.extractors import (
I3FeatureExtractorIceCubeUpgrade,
I3TruthExtractor,
I3GenericExtractor,
I3PISAExtractor,
I3PulseNoiseTruthFlagIceCubeUpgrade,
)
from graphnet.data.parquet import ParquetDataConverter
from graphnet.data.sqlite import SQLiteDataConverter
logger = get_logger(level=logging.INFO)
CONVERTER_CLASS = {
"sqlite": SQLiteDataConverter,
"parquet": ParquetDataConverter,
}
def main_icecube_upgrade(backend: str):
"""Convert IceCube-Upgrade I3 files to intermediate `backend` format."""
# Check(s)
assert backend in CONVERTER_CLASS
database_name = 'dev_step4_upgrade_028_21.11.2022_noise_test'
#inputs = ["/data/sim/IceCubeUpgrade/noise/step4/880028"]
inputs = ["/data/sim/IceCubeUpgrade/genie/step4/140028",
"/data/sim/IceCubeUpgrade/genie/step4/141028",
"/data/sim/IceCubeUpgrade/genie/step4/120028",
"/data/sim/IceCubeUpgrade/genie/step4/160028",
"/data/sim/IceCubeUpgrade/muongun/step4/130028",
"/data/sim/IceCubeUpgrade/noise/step4/880028"]
outdir = f"/data/user/rorsoe/databases/{database_name}"
workers = 1
converter = CONVERTER_CLASS[backend](
[
I3TruthExtractor(),
I3PISAExtractor(),
I3FeatureExtractorIceCubeUpgrade(
"SplitInIcePulses"
),
I3PulseNoiseTruthFlagIceCubeUpgrade(
"SplitInIcePulses_TruthFlags"
),
I3FeatureExtractorIceCubeUpgrade(
"SplitIceCubePulsesTWSRT"
),
I3PulseNoiseTruthFlagIceCubeUpgrade(
"SplitIceCubePulsesTWSRT_TruthFlags"
),
I3FeatureExtractorIceCubeUpgrade(
"SplitInIcePulses_GraphSage_Pulses"
),
I3PulseNoiseTruthFlagIceCubeUpgrade(
"SplitInIcePulses_GraphSage_Predictions"
),
I3PulseNoiseTruthFlagIceCubeUpgrade(
"SplitInIcePulses_dynedge_Predictions"
),
I3FeatureExtractorIceCubeUpgrade(
"SplitInIcePulses_dynedge_Pulses"
),
],
outdir + '/tmp',
workers=workers,
icetray_verbose=1,
gcd_rescue = '/data/sim/IceCubeUpgrade/genie/step4/140028/GeoCalibDetectorStatus_ICUpgrade.v58.mixed.V0.i3.bz2'
)
os.makedirs(outdir + '/data/')
converter(inputs)
#if backend == "sqlite":
# converter.merge_files(outdir + '/data/' + database_name)
if __name__ == "__main__":
backend = "sqlite"
main_icecube_upgrade(backend)
```
| non_usab | cannot convert upgrade files to sqlite anymore describe the bug i used to be able to convert these files to sqlite i then made the mistake of updating my main branch on my local fork now i can t convert the files graphnet now incorrectly states that no gcd was given to reproduce steps to reproduce the behavior load cvmfs for upgrade install graphnet main run script adjust outdir see error expected behavior should work full traceback please include the full error message to allow for debugging traceback most recent call last file make database py line in fatal dataio no frame to pop cxx in dataio pop frame stream fatal dataio no frame to pop cxx in dataio pop frame stream graphnet error error no gcd file was provided and no g frame was found exiting graphnet error error no gcd file was provided and no g frame was found exiting main icecube upgrade backend file make database py line in main icecube upgrade fatal dataio no frame to pop cxx in dataio pop frame stream graphnet error error no gcd file was provided and no g frame was found exiting fatal dataio no frame to pop cxx in dataio pop frame stream converter inputs file lustre hpc hep github graphnet src graphnet data dataconverter py line in call graphnet error error no gcd file was provided and no g frame was found exiting fatal dataio no frame to pop cxx in dataio pop frame stream self execute filesets file lustre hpc hep github graphnet src graphnet data dataconverter py line in execute graphnet error error no gcd file was provided and no g frame was found exiting fatal dataio no frame to pop cxx in dataio pop frame stream graphnet error error no gcd file was provided and no g frame was found exiting pool self iterate over individual files filesets file lustre hpc hep github graphnet src graphnet data dataconverter py line in iterate over individual files fatal dataio no frame to pop cxx in dataio pop frame stream graphnet error error no gcd file was provided and no g frame was found exiting self process file tqdm args unit file s colour green fatal dataio no frame to pop cxx in dataio pop frame stream file cvmfs icecube opensciencegrid org rhel lib multiprocessing pool py line in next graphnet error error no gcd file was provided and no g frame was found exiting fatal dataio no frame to pop cxx in dataio pop frame stream raise value graphnet error error no gcd file was provided and no g frame was found exiting runtimeerror no frame to pop in dataio pop frame stream fatal dataio no frame to pop cxx in dataio pop frame stream graphnet error error no gcd file was provided and no g frame was found exiting fatal dataio no frame to pop cxx in dataio pop frame stream graphnet error error no gcd file was provided and no g frame was found exiting additional context import logging import os import logging import os from graphnet utilities logging import get logger from graphnet data extractors import from graphnet data parquet import parquetdataconverter from graphnet data sqlite import sqlitedataconverter logger get logger level logging info converter class sqlite sqlitedataconverter parquet parquetdataconverter def main icecube upgrade backend str convert icecube upgrade files to intermediate backend format check s assert backend in converter class database name dev upgrade noise test inputs inputs data sim icecubeupgrade genie data sim icecubeupgrade genie data sim icecubeupgrade genie data sim icecubeupgrade genie data sim icecubeupgrade muongun data sim icecubeupgrade noise outdir f data user rorsoe databases database name workers converter converter class splitinicepulses splitinicepulses truthflags spliticecubepulsestwsrt spliticecubepulsestwsrt truthflags splitinicepulses graphsage pulses splitinicepulses graphsage predictions splitinicepulses dynedge predictions splitinicepulses dynedge pulses outdir tmp workers workers icetray verbose gcd rescue data sim icecubeupgrade genie geocalibdetectorstatus icupgrade mixed os makedirs outdir data converter inputs if backend sqlite converter merge files outdir data database name if name main backend sqlite main icecube upgrade backend | 0 |
9,572 | 6,394,092,962 | IssuesEvent | 2017-08-04 09:21:57 | Elgg/Elgg | https://api.github.com/repos/Elgg/Elgg | closed | Wrap full entity listing body and add $vars['body_params'] | bc break dev usability | Improve dev usability of views/default/object/elements/full/body.php
https://github.com/Elgg/Elgg/pull/9944/files#r70393659
| True | Wrap full entity listing body and add $vars['body_params'] - Improve dev usability of views/default/object/elements/full/body.php
https://github.com/Elgg/Elgg/pull/9944/files#r70393659
| usab | wrap full entity listing body and add vars improve dev usability of views default object elements full body php | 1 |
10,815 | 6,933,138,242 | IssuesEvent | 2017-12-02 02:37:24 | vmware/hillview | https://api.github.com/repos/vmware/hillview | closed | Javascript types are not native | fixed usability | When running a JavaScript function the Date objects are not JavaScript Data objects but they are Java.util.Instant objects. Would be nice to automatically convert them. | True | Javascript types are not native - When running a JavaScript function the Date objects are not JavaScript Data objects but they are Java.util.Instant objects. Would be nice to automatically convert them. | usab | javascript types are not native when running a javascript function the date objects are not javascript data objects but they are java util instant objects would be nice to automatically convert them | 1 |
61,248 | 14,619,766,808 | IssuesEvent | 2020-12-22 18:26:56 | hashgraph/hedera-mirror-node | https://api.github.com/repos/hashgraph/hedera-mirror-node | opened | TimescaleDB: Update DB User Permissions | P2 database enhancement security | **Problem**
Currently the main user on the db servers as the owner and therefore has extra permissions not needed.
We want to lock this down further and ensure permissions are intentionally set only as needed.
**Solution**
- Create a new schema on init
- Create importer users as a reader/writer not owner
- Grant explicit permissions for select and connect to users and roles in new schema
- Update init job in helm chart to apply above
**Alternatives**
**Additional Context**
| True | TimescaleDB: Update DB User Permissions - **Problem**
Currently the main user on the db servers as the owner and therefore has extra permissions not needed.
We want to lock this down further and ensure permissions are intentionally set only as needed.
**Solution**
- Create a new schema on init
- Create importer users as a reader/writer not owner
- Grant explicit permissions for select and connect to users and roles in new schema
- Update init job in helm chart to apply above
**Alternatives**
**Additional Context**
| non_usab | timescaledb update db user permissions problem currently the main user on the db servers as the owner and therefore has extra permissions not needed we want to lock this down further and ensure permissions are intentionally set only as needed solution create a new schema on init create importer users as a reader writer not owner grant explicit permissions for select and connect to users and roles in new schema update init job in helm chart to apply above alternatives additional context | 0 |
5,501 | 3,930,321,645 | IssuesEvent | 2016-04-25 07:28:20 | Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith | https://api.github.com/repos/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith | opened | QA_Triaxial Test_Back to experiments_smk | Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open | Defect Description :
In the "Triaxial Test" experiment,the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in-order to view the list of experiments by the user.
Actual Result :
In the "Triaxial Test" experiment,the back to experiments link is not present in the page.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith/blob/master/test-cases/integration_test-cases/Triaxial%20Test/Triaxial%20Test_19_Back%20to%20experiments_smk.org | True | QA_Triaxial Test_Back to experiments_smk - Defect Description :
In the "Triaxial Test" experiment,the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in-order to view the list of experiments by the user.
Actual Result :
In the "Triaxial Test" experiment,the back to experiments link is not present in the page.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith/blob/master/test-cases/integration_test-cases/Triaxial%20Test/Triaxial%20Test_19_Back%20to%20experiments_smk.org | usab | qa triaxial test back to experiments smk defect description in the triaxial test experiment the back to experiments link is not present in the page instead the back to experiments link should be displayed on the screen in order to view the list of experiments by the user actual result in the triaxial test experiment the back to experiments link is not present in the page environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link | 1 |
33,491 | 2,765,746,451 | IssuesEvent | 2015-04-29 22:14:46 | NuGet/NuGetGallery | https://api.github.com/repos/NuGet/NuGetGallery | closed | Design /api/v3 for NuGet's clients such that OData isn't needed for our own client requests | Priority - 2 |
<!---
@huboard:{"order":53.873046875}
-->
| 1.0 | Design /api/v3 for NuGet's clients such that OData isn't needed for our own client requests -
<!---
@huboard:{"order":53.873046875}
-->
| non_usab | design api for nuget s clients such that odata isn t needed for our own client requests huboard order | 0 |
18,075 | 12,529,147,683 | IssuesEvent | 2020-06-04 10:48:20 | the-tale/the-tale | https://api.github.com/repos/the-tale/the-tale | opened | Discord: явно сообщать игроку, что началась обработка его команды и что она завершилась | comp_discord cont_usability est_medium good first issue type_improvement | Сейчас некоторые команды возвращают управление раньше, чем будут завершены инициированные ими действия.
Само по себе это правильно, с точки зрения текущей архитектуры, но неудобно с точки зрения пользователей.
Надо что-нибудь придумать.
Варианты:
- Слать сообщение о начале операций для каждой (?) команды.
- Скорректировать сообщения бота так, чтобы его поведение стало понятным.
- Явно дожидаться завершения всех операций и слать отдельное сообщение по их окончании. Этот вариант я считаю плохим, так как он чрезмерно увеличит связанность логики.
- Поменять архитектуру так, чтобы проблема исчезла. Слишком много работы с неясным итогом (архитектура сейчас нормальная, менять её ради мелочи нет смысла).
- Что-то ещё. | True | Discord: явно сообщать игроку, что началась обработка его команды и что она завершилась - Сейчас некоторые команды возвращают управление раньше, чем будут завершены инициированные ими действия.
Само по себе это правильно, с точки зрения текущей архитектуры, но неудобно с точки зрения пользователей.
Надо что-нибудь придумать.
Варианты:
- Слать сообщение о начале операций для каждой (?) команды.
- Скорректировать сообщения бота так, чтобы его поведение стало понятным.
- Явно дожидаться завершения всех операций и слать отдельное сообщение по их окончании. Этот вариант я считаю плохим, так как он чрезмерно увеличит связанность логики.
- Поменять архитектуру так, чтобы проблема исчезла. Слишком много работы с неясным итогом (архитектура сейчас нормальная, менять её ради мелочи нет смысла).
- Что-то ещё. | usab | discord явно сообщать игроку что началась обработка его команды и что она завершилась сейчас некоторые команды возвращают управление раньше чем будут завершены инициированные ими действия само по себе это правильно с точки зрения текущей архитектуры но неудобно с точки зрения пользователей надо что нибудь придумать варианты слать сообщение о начале операций для каждой команды скорректировать сообщения бота так чтобы его поведение стало понятным явно дожидаться завершения всех операций и слать отдельное сообщение по их окончании этот вариант я считаю плохим так как он чрезмерно увеличит связанность логики поменять архитектуру так чтобы проблема исчезла слишком много работы с неясным итогом архитектура сейчас нормальная менять её ради мелочи нет смысла что то ещё | 1 |
350,811 | 31,932,343,522 | IssuesEvent | 2023-09-19 08:17:25 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | reopened | Fix jax_numpy_linalg.test_jax_tensorinv | JAX Frontend Sub Task Failing Test | | | |
|---|---|
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix jax_numpy_linalg.test_jax_tensorinv - | | |
|---|---|
|paddle|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-failure-red></a>
|jax|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-success-success></a>
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="https://github.com/unifyai/ivy/actions/runs/6230446511/job/16910475139"><img src=https://img.shields.io/badge/-success-success></a>
| non_usab | fix jax numpy linalg test jax tensorinv paddle a href src numpy a href src jax a href src tensorflow a href src torch a href src | 0 |
43,247 | 5,614,961,069 | IssuesEvent | 2017-04-03 13:38:20 | GlotPress/GlotPress-WP | https://api.github.com/repos/GlotPress/GlotPress-WP | closed | Consider splitting some templates | enhancement question/discussion _design | While exploring some design ideas, I realized how hard it is do customize the translation row (preview & editor) in a custom theme because (among other things) the templates are so large.
While we're probably going to change a lot of this with [the new design](https://github.com/GlotPress/GlotPress-WP/projects/2), I'd like to propose starting by splitting some of our templates where possible, especially the translation row.
Thoughts? | 1.0 | Consider splitting some templates - While exploring some design ideas, I realized how hard it is do customize the translation row (preview & editor) in a custom theme because (among other things) the templates are so large.
While we're probably going to change a lot of this with [the new design](https://github.com/GlotPress/GlotPress-WP/projects/2), I'd like to propose starting by splitting some of our templates where possible, especially the translation row.
Thoughts? | non_usab | consider splitting some templates while exploring some design ideas i realized how hard it is do customize the translation row preview editor in a custom theme because among other things the templates are so large while we re probably going to change a lot of this with i d like to propose starting by splitting some of our templates where possible especially the translation row thoughts | 0 |
767,253 | 26,916,750,706 | IssuesEvent | 2023-02-07 07:16:39 | ballerina-platform/ballerina-dev-website | https://api.github.com/repos/ballerina-platform/ballerina-dev-website | closed | Adding Back the Docs Related to "Extending with Compiler Extensions" in SLP6 | Priority/Low Type/Task | **Description:**
We have temporarily removed the Learn page[1] related to "Extending with Compiler Extensions" from Bio for SLP5 via [2].
Need to add them back in SLP6 after the issue [3] is fixed.
Also, it is added in 1.2 in [4].
[1] http://dev.ballerina.io/swan-lake/learn/extending-with-compiler-extensions/
[2] https://github.com/ballerina-platform/ballerina-dev-website/pull/1418
[3] https://github.com/ballerina-platform/ballerina-lang/issues/26503
[4] https://ballerina.io/1.2/learn/extending-with-compiler-extensions/
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, Browser, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> | 1.0 | Adding Back the Docs Related to "Extending with Compiler Extensions" in SLP6 - **Description:**
We have temporarily removed the Learn page[1] related to "Extending with Compiler Extensions" from Bio for SLP5 via [2].
Need to add them back in SLP6 after the issue [3] is fixed.
Also, it is added in 1.2 in [4].
[1] http://dev.ballerina.io/swan-lake/learn/extending-with-compiler-extensions/
[2] https://github.com/ballerina-platform/ballerina-dev-website/pull/1418
[3] https://github.com/ballerina-platform/ballerina-lang/issues/26503
[4] https://ballerina.io/1.2/learn/extending-with-compiler-extensions/
**Suggested Labels:**
<!-- Optional comma separated list of suggested labels. Non committers can’t assign labels to issues, so this will help issue creators who are not a committer to suggest possible labels-->
**Suggested Assignees:**
<!--Optional comma separated list of suggested team members who should attend the issue. Non committers can’t assign issues to assignees, so this will help issue creators who are not a committer to suggest possible assignees-->
**Affected Product Version:**
**OS, Browser, other environment details and versions:**
**Steps to reproduce:**
**Related Issues:**
<!-- Any related issues such as sub tasks, issues reported in other repositories (e.g component repositories), similar problems, etc. --> | non_usab | adding back the docs related to extending with compiler extensions in description we have temporarily removed the learn page related to extending with compiler extensions from bio for via need to add them back in after the issue is fixed also it is added in in suggested labels suggested assignees affected product version os browser other environment details and versions steps to reproduce related issues | 0 |
452,730 | 32,066,489,936 | IssuesEvent | 2023-09-25 03:45:46 | apache/incubator-opendal | https://api.github.com/repos/apache/incubator-opendal | closed | docs: Update the announcement email template to ensure that disclaimers have been added | documentation good first issue help wanted | As mentioned in the discussion at https://lists.apache.org/thread/0d8dxxbhb28m7w7kqxs1d2n0hsdybocg, Sebb reminded us to include disclaimers in our announcement.
We should add the following content:
```
---
Apache OpenDAL (incubating) is an effort undergoing incubation at the Apache
Software Foundation (ASF), sponsored by the Apache Incubator PMC.
Incubation is required of all newly accepted projects until a further review
indicates that the infrastructure, communications, and decision making process
have stabilized in a manner consistent with other successful ASF projects.
While incubation status is not necessarily a reflection of the completeness
or stability of the code, it does indicate that the project has yet to be
fully endorsed by the ASF.
``` | 1.0 | docs: Update the announcement email template to ensure that disclaimers have been added - As mentioned in the discussion at https://lists.apache.org/thread/0d8dxxbhb28m7w7kqxs1d2n0hsdybocg, Sebb reminded us to include disclaimers in our announcement.
We should add the following content:
```
---
Apache OpenDAL (incubating) is an effort undergoing incubation at the Apache
Software Foundation (ASF), sponsored by the Apache Incubator PMC.
Incubation is required of all newly accepted projects until a further review
indicates that the infrastructure, communications, and decision making process
have stabilized in a manner consistent with other successful ASF projects.
While incubation status is not necessarily a reflection of the completeness
or stability of the code, it does indicate that the project has yet to be
fully endorsed by the ASF.
``` | non_usab | docs update the announcement email template to ensure that disclaimers have been added as mentioned in the discussion at sebb reminded us to include disclaimers in our announcement we should add the following content apache opendal incubating is an effort undergoing incubation at the apache software foundation asf sponsored by the apache incubator pmc incubation is required of all newly accepted projects until a further review indicates that the infrastructure communications and decision making process have stabilized in a manner consistent with other successful asf projects while incubation status is not necessarily a reflection of the completeness or stability of the code it does indicate that the project has yet to be fully endorsed by the asf | 0 |
66,748 | 3,257,769,294 | IssuesEvent | 2015-10-20 19:17:59 | ucla/ohmageX | https://api.github.com/repos/ucla/ohmageX | opened | docx gets 'file format is not supported' | bug High Priority | I'm online and viewing the artifact with docx attachment that I just uploaded successfully to the server, but the app shows error "file format is not supported" | 1.0 | docx gets 'file format is not supported' - I'm online and viewing the artifact with docx attachment that I just uploaded successfully to the server, but the app shows error "file format is not supported" | non_usab | docx gets file format is not supported i m online and viewing the artifact with docx attachment that i just uploaded successfully to the server but the app shows error file format is not supported | 0 |
6,515 | 4,322,980,252 | IssuesEvent | 2016-07-25 15:37:17 | ff36/halo-gui | https://api.github.com/repos/ff36/halo-gui | closed | Incident table spacing | change enhancement usability | The incident table hight always seems to cut off the last entry. I know you can scroll to it but it would be better if the table was padded slightly so all the rows are properly displayed.

| True | Incident table spacing - The incident table hight always seems to cut off the last entry. I know you can scroll to it but it would be better if the table was padded slightly so all the rows are properly displayed.

| usab | incident table spacing the incident table hight always seems to cut off the last entry i know you can scroll to it but it would be better if the table was padded slightly so all the rows are properly displayed | 1 |
24,052 | 23,277,235,228 | IssuesEvent | 2022-08-05 08:27:41 | zaproxy/zaproxy | https://api.github.com/repos/zaproxy/zaproxy | closed | Persist column configuration in History tab | enhancement Component-UI Usability | First: Thank you very much for your work on ZAP!
**Describe the bug**
The HTTP request history tab has several columns. Although by default most columns are shown, users can configure which columns are displayed and in which order. This information is lost when ZAP is quit, so users have to do this configuration every time.
**To Reproduce**
Steps to reproduce the behavior:
1. Start ZAP
1. Click on the History tab
1. Configure the columns by clicking in the header row on the rightmost icon
1. Disable the "Id" column, it isn't shown any more
1. Move the Request column to the front
1. Quit ZAP
1. Start ZAP again
1. "Id" column is shown again, Method column is third column, all configuration for the columns was reset
**Expected behavior**
Column configuration for History tab should be the same after restart.
**Software versions**
- ZAP: 2.7.0
- OS: Debian stable 9.4 ("stretch")
- Java: openjdk 1.9, Debian version number: `9~b181-4~bpo9+1`, package `openjdk-9-jre`
| True | Persist column configuration in History tab - First: Thank you very much for your work on ZAP!
**Describe the bug**
The HTTP request history tab has several columns. Although by default most columns are shown, users can configure which columns are displayed and in which order. This information is lost when ZAP is quit, so users have to do this configuration every time.
**To Reproduce**
Steps to reproduce the behavior:
1. Start ZAP
1. Click on the History tab
1. Configure the columns by clicking in the header row on the rightmost icon
1. Disable the "Id" column, it isn't shown any more
1. Move the Request column to the front
1. Quit ZAP
1. Start ZAP again
1. "Id" column is shown again, Method column is third column, all configuration for the columns was reset
**Expected behavior**
Column configuration for History tab should be the same after restart.
**Software versions**
- ZAP: 2.7.0
- OS: Debian stable 9.4 ("stretch")
- Java: openjdk 1.9, Debian version number: `9~b181-4~bpo9+1`, package `openjdk-9-jre`
| usab | persist column configuration in history tab first thank you very much for your work on zap describe the bug the http request history tab has several columns although by default most columns are shown users can configure which columns are displayed and in which order this information is lost when zap is quit so users have to do this configuration every time to reproduce steps to reproduce the behavior start zap click on the history tab configure the columns by clicking in the header row on the rightmost icon disable the id column it isn t shown any more move the request column to the front quit zap start zap again id column is shown again method column is third column all configuration for the columns was reset expected behavior column configuration for history tab should be the same after restart software versions zap os debian stable stretch java openjdk debian version number package openjdk jre | 1 |
10,100 | 3,086,296,724 | IssuesEvent | 2015-08-25 01:47:47 | balderdashy/waterline-cursor | https://api.github.com/repos/balderdashy/waterline-cursor | closed | waterline-cursor@0.0.5 causing undefined attributes to be added when populating association | needs failing test | After my most recent deploy of my application, I started running into issues with my app in a staging environment. I traced it back to a version update in the waterline-cursor library. In my development env I'm using 0.0.3, and in staging I see the issue because npm is using 0.0.5 (it goes away if I set a hard dependency on 0.0.3)
Basically I noticed that after a `findOne` query, with a `populate` on a collection, I have an additional `undefined` attribute that points to an empty array. When I comment out the `populate` call, there's no extra attribute. This is occurring with the `sails-postgresql` adapter.
| 1.0 | waterline-cursor@0.0.5 causing undefined attributes to be added when populating association - After my most recent deploy of my application, I started running into issues with my app in a staging environment. I traced it back to a version update in the waterline-cursor library. In my development env I'm using 0.0.3, and in staging I see the issue because npm is using 0.0.5 (it goes away if I set a hard dependency on 0.0.3)
Basically I noticed that after a `findOne` query, with a `populate` on a collection, I have an additional `undefined` attribute that points to an empty array. When I comment out the `populate` call, there's no extra attribute. This is occurring with the `sails-postgresql` adapter.
| non_usab | waterline cursor causing undefined attributes to be added when populating association after my most recent deploy of my application i started running into issues with my app in a staging environment i traced it back to a version update in the waterline cursor library in my development env i m using and in staging i see the issue because npm is using it goes away if i set a hard dependency on basically i noticed that after a findone query with a populate on a collection i have an additional undefined attribute that points to an empty array when i comment out the populate call there s no extra attribute this is occurring with the sails postgresql adapter | 0 |
4,825 | 3,896,897,190 | IssuesEvent | 2016-04-16 03:02:06 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 16332995: Apple TV should stream iTunes purchases faster like the iTunes Music Festival channel does | classification:ui/usability reproducible:always status:open | #### Description
Summary:
The Apple TV has no user accessable storage. In order to stream from iTunes in the Cloud the Apple TV essentially has to download a given media item (movie or TV show) before it can start playing it. Thankfully the entire item doesn't have to be downloaded to start viewing, but on lower bandwidth connections the time from starting to viewing an iTunes media item can be quite a while (15-30 minutes depending on the length of the program for TV shows).
The Apple TV also includes (as of today) an iTunes Music Festival channel for the SWSW iTunes Music Festival. Concerts are presented in HD and stream to the Apple TV. The quality is high and the time from starting playback to actual viewing (stream buffering) is very short, even on a lower bandwidth link.
Steps to Reproduce:
A: Slow to start stream
1. Purchase a TV show episode on iTunes. We shall use Chozen season 1 episode 2 as our example.
2. Place an Apple TV on a local network that is connected to a lower bandwidth connection (3.9Mbps down in this case)
3. Navigate on the Apple TV to TV Shows > Purchased > All > Chozen > Season 1 > Episode 2 and select "Play"
4. See the episode start loading, after a few moments the screen warns the viewer that loading will take an additional 15 minutes
5. After waiting 15 minutes the episode is ready to play.
B: Fast to start stream
1. Goto the iTunes SXSW Music Festival channel
2. Navigate to March 13's Soundgarden concert
3. Select "Play"
4. The HD stream on the same internet link mentioned above starts nearly instantly
Expected Results:
Movies and TV Shows purchased via iTunes should start streaming as fast as the iTunes Music Festival app, especially for content that's less than one third the length of the Soundgarden concert. If a slow link can handle HD streams for the iTunes Music Festival, Netflix, and PBS channels without issue I'd expect the same level of performance from the marquee Movies and TV Shows channels.
I know how to switch the resolution in the iTunes Store settings on the Apple TV, but even 720 resolutions incur a large buffering penalty. I shouldn't have to decrease my resolution when HD streams play fine in other channels.
Actual Results:
Movies and TV Shows take far too long to buffer and start playing.
Version:
- Apple TV 3rd generation
- OS version 6.1
- wired ethernet connection
- Sonic.net DSL Internet service (at 3.9Mbps down/ 0.85Mbps up)
-
Product Version: 6.1
Created: 2014-03-15T00:21:20.126957
Originated: 2014-03-15T00:20:21+00:00
Open Radar Link: http://www.openradar.me/16332995 | True | 16332995: Apple TV should stream iTunes purchases faster like the iTunes Music Festival channel does - #### Description
Summary:
The Apple TV has no user accessable storage. In order to stream from iTunes in the Cloud the Apple TV essentially has to download a given media item (movie or TV show) before it can start playing it. Thankfully the entire item doesn't have to be downloaded to start viewing, but on lower bandwidth connections the time from starting to viewing an iTunes media item can be quite a while (15-30 minutes depending on the length of the program for TV shows).
The Apple TV also includes (as of today) an iTunes Music Festival channel for the SWSW iTunes Music Festival. Concerts are presented in HD and stream to the Apple TV. The quality is high and the time from starting playback to actual viewing (stream buffering) is very short, even on a lower bandwidth link.
Steps to Reproduce:
A: Slow to start stream
1. Purchase a TV show episode on iTunes. We shall use Chozen season 1 episode 2 as our example.
2. Place an Apple TV on a local network that is connected to a lower bandwidth connection (3.9Mbps down in this case)
3. Navigate on the Apple TV to TV Shows > Purchased > All > Chozen > Season 1 > Episode 2 and select "Play"
4. See the episode start loading, after a few moments the screen warns the viewer that loading will take an additional 15 minutes
5. After waiting 15 minutes the episode is ready to play.
B: Fast to start stream
1. Goto the iTunes SXSW Music Festival channel
2. Navigate to March 13's Soundgarden concert
3. Select "Play"
4. The HD stream on the same internet link mentioned above starts nearly instantly
Expected Results:
Movies and TV Shows purchased via iTunes should start streaming as fast as the iTunes Music Festival app, especially for content that's less than one third the length of the Soundgarden concert. If a slow link can handle HD streams for the iTunes Music Festival, Netflix, and PBS channels without issue I'd expect the same level of performance from the marquee Movies and TV Shows channels.
I know how to switch the resolution in the iTunes Store settings on the Apple TV, but even 720 resolutions incur a large buffering penalty. I shouldn't have to decrease my resolution when HD streams play fine in other channels.
Actual Results:
Movies and TV Shows take far too long to buffer and start playing.
Version:
- Apple TV 3rd generation
- OS version 6.1
- wired ethernet connection
- Sonic.net DSL Internet service (at 3.9Mbps down/ 0.85Mbps up)
-
Product Version: 6.1
Created: 2014-03-15T00:21:20.126957
Originated: 2014-03-15T00:20:21+00:00
Open Radar Link: http://www.openradar.me/16332995 | usab | apple tv should stream itunes purchases faster like the itunes music festival channel does description summary the apple tv has no user accessable storage in order to stream from itunes in the cloud the apple tv essentially has to download a given media item movie or tv show before it can start playing it thankfully the entire item doesn t have to be downloaded to start viewing but on lower bandwidth connections the time from starting to viewing an itunes media item can be quite a while minutes depending on the length of the program for tv shows the apple tv also includes as of today an itunes music festival channel for the swsw itunes music festival concerts are presented in hd and stream to the apple tv the quality is high and the time from starting playback to actual viewing stream buffering is very short even on a lower bandwidth link steps to reproduce a slow to start stream purchase a tv show episode on itunes we shall use chozen season episode as our example place an apple tv on a local network that is connected to a lower bandwidth connection down in this case navigate on the apple tv to tv shows purchased all chozen season episode and select play see the episode start loading after a few moments the screen warns the viewer that loading will take an additional minutes after waiting minutes the episode is ready to play b fast to start stream goto the itunes sxsw music festival channel navigate to march s soundgarden concert select play the hd stream on the same internet link mentioned above starts nearly instantly expected results movies and tv shows purchased via itunes should start streaming as fast as the itunes music festival app especially for content that s less than one third the length of the soundgarden concert if a slow link can handle hd streams for the itunes music festival netflix and pbs channels without issue i d expect the same level of performance from the marquee movies and tv shows channels i know how to switch the resolution in the itunes store settings on the apple tv but even resolutions incur a large buffering penalty i shouldn t have to decrease my resolution when hd streams play fine in other channels actual results movies and tv shows take far too long to buffer and start playing version apple tv generation os version wired ethernet connection sonic net dsl internet service at down up product version created originated open radar link | 1 |
9,945 | 6,524,941,677 | IssuesEvent | 2017-08-29 14:22:04 | loconomics/loconomics | https://api.github.com/repos/loconomics/loconomics | closed | Usability research: edition toolbar at Work Photos | C: Usability F: Accessbility F: Listings | ***This is a request for user testing on this topic***
From comments at #95, we need to **research** the usability of the edition toolbar that is available at Work Photos under an over/tap on each photo. Since they are hidden, user may not discover them and so not figuring out how to remove, move, rotate them. This can specially increase the problem described at #95 of user seeing a bad oriented photo and not guessing that can be rotated.
It *seems* worse at mobile, when need to be *tapped*, rather than desktop/laptop, since using mouse the user may discover them by chance just moving the pointer over the photos.
Usability comment from @situatedbit:
> generally we should avoid hiding something like this behind a swipe or a tap if it's related to the primary task on a screen/page.
Proposed alternative by @iagosrl:
> We can move buttons to below as a toolbar on each photo (only on mobile or the feature feels intuitive at desktop too?). Some user testing on this or not? | True | Usability research: edition toolbar at Work Photos - ***This is a request for user testing on this topic***
From comments at #95, we need to **research** the usability of the edition toolbar that is available at Work Photos under an over/tap on each photo. Since they are hidden, user may not discover them and so not figuring out how to remove, move, rotate them. This can specially increase the problem described at #95 of user seeing a bad oriented photo and not guessing that can be rotated.
It *seems* worse at mobile, when need to be *tapped*, rather than desktop/laptop, since using mouse the user may discover them by chance just moving the pointer over the photos.
Usability comment from @situatedbit:
> generally we should avoid hiding something like this behind a swipe or a tap if it's related to the primary task on a screen/page.
Proposed alternative by @iagosrl:
> We can move buttons to below as a toolbar on each photo (only on mobile or the feature feels intuitive at desktop too?). Some user testing on this or not? | usab | usability research edition toolbar at work photos this is a request for user testing on this topic from comments at we need to research the usability of the edition toolbar that is available at work photos under an over tap on each photo since they are hidden user may not discover them and so not figuring out how to remove move rotate them this can specially increase the problem described at of user seeing a bad oriented photo and not guessing that can be rotated it seems worse at mobile when need to be tapped rather than desktop laptop since using mouse the user may discover them by chance just moving the pointer over the photos usability comment from situatedbit generally we should avoid hiding something like this behind a swipe or a tap if it s related to the primary task on a screen page proposed alternative by iagosrl we can move buttons to below as a toolbar on each photo only on mobile or the feature feels intuitive at desktop too some user testing on this or not | 1 |
11,208 | 7,107,661,255 | IssuesEvent | 2018-01-16 20:48:48 | matomo-org/matomo | https://api.github.com/repos/matomo-org/matomo | closed | Remove word "website" from website selector | c: Usability help wanted | As mentioned here https://github.com/piwik/piwik/issues/8712 I do now think as well that we should remove the word "Website" from the sites selector eg as suggested here: https://github.com/piwik/piwik/issues/8712#issuecomment-148390762
All other selectors use a proper icon instead but the selector doesn't yet.

Instead of the word "website" I'd rather see an icon, maybe one similar to the one used in Piwik Mobile eg http://fortawesome.github.io/Font-Awesome/icon/globe/ but could be also a different one since the globe could be also used for language etc.
Because of the word website one cannot see much of the website's name which can be especially tricky if one has many websites with similar names, maybe even many websites that start with the same name.
| True | Remove word "website" from website selector - As mentioned here https://github.com/piwik/piwik/issues/8712 I do now think as well that we should remove the word "Website" from the sites selector eg as suggested here: https://github.com/piwik/piwik/issues/8712#issuecomment-148390762
All other selectors use a proper icon instead but the selector doesn't yet.

Instead of the word "website" I'd rather see an icon, maybe one similar to the one used in Piwik Mobile eg http://fortawesome.github.io/Font-Awesome/icon/globe/ but could be also a different one since the globe could be also used for language etc.
Because of the word website one cannot see much of the website's name which can be especially tricky if one has many websites with similar names, maybe even many websites that start with the same name.
| usab | remove word website from website selector as mentioned here i do now think as well that we should remove the word website from the sites selector eg as suggested here all other selectors use a proper icon instead but the selector doesn t yet instead of the word website i d rather see an icon maybe one similar to the one used in piwik mobile eg but could be also a different one since the globe could be also used for language etc because of the word website one cannot see much of the website s name which can be especially tricky if one has many websites with similar names maybe even many websites that start with the same name | 1 |
197,412 | 14,923,141,072 | IssuesEvent | 2021-01-23 17:41:23 | jakeegan/Blink-Controlled-Speech-Synthesizer | https://api.github.com/repos/jakeegan/Blink-Controlled-Speech-Synthesizer | closed | Test GUI | Test | Test graphical user interface by measuring the response time for interacting with various elements. | 1.0 | Test GUI - Test graphical user interface by measuring the response time for interacting with various elements. | non_usab | test gui test graphical user interface by measuring the response time for interacting with various elements | 0 |
421,414 | 28,315,289,604 | IssuesEvent | 2023-04-10 19:02:04 | vercel/next.js | https://api.github.com/repos/vercel/next.js | opened | Docs: Fix incorrect instrucstions in docs about absolute imports and module path aliases. | template: documentation | ### What is the improvement or update you wish to see?
I was trying to configure my app to use absolute imports and module path aliases following the docs about this topic.
### Is there any context that might help us understand?
I was using next with version `13.2.4` and following the docs instructions wasn't working because of the new `/app` directory, so I needed to change the `baseUrl` in the tsconfig.json to make it work.

### Does the docs page already exist? Please link to it.
https://nextjs.org/docs/advanced-features/module-path-aliases | 1.0 | Docs: Fix incorrect instrucstions in docs about absolute imports and module path aliases. - ### What is the improvement or update you wish to see?
I was trying to configure my app to use absolute imports and module path aliases following the docs about this topic.
### Is there any context that might help us understand?
I was using next with version `13.2.4` and following the docs instructions wasn't working because of the new `/app` directory, so I needed to change the `baseUrl` in the tsconfig.json to make it work.

### Does the docs page already exist? Please link to it.
https://nextjs.org/docs/advanced-features/module-path-aliases | non_usab | docs fix incorrect instrucstions in docs about absolute imports and module path aliases what is the improvement or update you wish to see i was trying to configure my app to use absolute imports and module path aliases following the docs about this topic is there any context that might help us understand i was using next with version and following the docs instructions wasn t working because of the new app directory so i needed to change the baseurl in the tsconfig json to make it work does the docs page already exist please link to it | 0 |
73,001 | 31,822,405,178 | IssuesEvent | 2023-09-14 04:11:55 | MicrosoftDocs/azure-docs | https://api.github.com/repos/MicrosoftDocs/azure-docs | closed | Add clarification regarding the (im)possibility of setting up different nodeOS upgrade policy for the different nodepools under the cluster | triaged cxp product-question Pri1 awaiting-customer-response azure-kubernetes-service/svc |
Hello,
The nodeOS automatic upgrade policy setting is a very nice and useful feature that would had enabled us to get rid of Kured as a way to deal with potentially necessary restarts after nodeOS nightly updates being deployed.
However, there is a show stopper that prevents us from using it yet - it does not allow defining separate nodeOS upgrade policy for the different node pools we have on our cluster. Are you planning to add such a feature in the future? Please, no matter if yes or now, put some comment on the documentation about it.
We can acutally live with a single nodeOS upgrade policy throughout the cluster, provided that we have a way to taint/annotate/whatever some of the nodes (or an entire nodepool) to be excluded from the process and let us perform the nodeOS upgrade manually in full control. Is there a setting that we can use to override the policy per-node/nodepool? Please, if such a way exist can you add it on the documentation?
Best Regards
M. Stavrev
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: d1945647-afb8-2283-0f92-e1744f51eb79
* Version Independent ID: 6894c5c4-a100-7349-8f85-e03f3e1adfe0
* Content: [Automatically upgrade Azure Kubernetes Service (AKS) cluster node operating system images - Azure Kubernetes Service](https://learn.microsoft.com/en-us/azure/aks/auto-upgrade-node-image)
* Content Source: [articles/aks/auto-upgrade-node-image.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/aks/auto-upgrade-node-image.md)
* Service: **azure-kubernetes-service**
* GitHub Login: @Nickomang
* Microsoft Alias: **nickoman** | 1.0 | Add clarification regarding the (im)possibility of setting up different nodeOS upgrade policy for the different nodepools under the cluster -
Hello,
The nodeOS automatic upgrade policy setting is a very nice and useful feature that would had enabled us to get rid of Kured as a way to deal with potentially necessary restarts after nodeOS nightly updates being deployed.
However, there is a show stopper that prevents us from using it yet - it does not allow defining separate nodeOS upgrade policy for the different node pools we have on our cluster. Are you planning to add such a feature in the future? Please, no matter if yes or now, put some comment on the documentation about it.
We can acutally live with a single nodeOS upgrade policy throughout the cluster, provided that we have a way to taint/annotate/whatever some of the nodes (or an entire nodepool) to be excluded from the process and let us perform the nodeOS upgrade manually in full control. Is there a setting that we can use to override the policy per-node/nodepool? Please, if such a way exist can you add it on the documentation?
Best Regards
M. Stavrev
---
#### Document Details
⚠ *Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.*
* ID: d1945647-afb8-2283-0f92-e1744f51eb79
* Version Independent ID: 6894c5c4-a100-7349-8f85-e03f3e1adfe0
* Content: [Automatically upgrade Azure Kubernetes Service (AKS) cluster node operating system images - Azure Kubernetes Service](https://learn.microsoft.com/en-us/azure/aks/auto-upgrade-node-image)
* Content Source: [articles/aks/auto-upgrade-node-image.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/aks/auto-upgrade-node-image.md)
* Service: **azure-kubernetes-service**
* GitHub Login: @Nickomang
* Microsoft Alias: **nickoman** | non_usab | add clarification regarding the im possibility of setting up different nodeos upgrade policy for the different nodepools under the cluster hello the nodeos automatic upgrade policy setting is a very nice and useful feature that would had enabled us to get rid of kured as a way to deal with potentially necessary restarts after nodeos nightly updates being deployed however there is a show stopper that prevents us from using it yet it does not allow defining separate nodeos upgrade policy for the different node pools we have on our cluster are you planning to add such a feature in the future please no matter if yes or now put some comment on the documentation about it we can acutally live with a single nodeos upgrade policy throughout the cluster provided that we have a way to taint annotate whatever some of the nodes or an entire nodepool to be excluded from the process and let us perform the nodeos upgrade manually in full control is there a setting that we can use to override the policy per node nodepool please if such a way exist can you add it on the documentation best regards m stavrev document details ⚠ do not edit this section it is required for learn microsoft com ➟ github issue linking id version independent id content content source service azure kubernetes service github login nickomang microsoft alias nickoman | 0 |
576,744 | 17,093,626,913 | IssuesEvent | 2021-07-08 21:13:33 | airshipit/treasuremap | https://api.github.com/repos/airshipit/treasuremap | closed | Generate and Encrypt BMC Credentials for ViNO CR | 2-Manifests enhancement priority/medium size s | **Problem description**
Currently the BMC credentials specified in the ViNO CR and used for creating BMH resources in the workload-config manifest function are hard-coded to clear text values. These credentials need to be properly secured and managed during multi-tenant type deployments.
**Proposed change**
As part of multi-tenant site deployment, provide capability to generate and encrypt the BMC credentials to be used for BMH resources generated by ViNO CR so that these credentials are properly secured and are never stored unencrypted in manifests.
| 1.0 | Generate and Encrypt BMC Credentials for ViNO CR - **Problem description**
Currently the BMC credentials specified in the ViNO CR and used for creating BMH resources in the workload-config manifest function are hard-coded to clear text values. These credentials need to be properly secured and managed during multi-tenant type deployments.
**Proposed change**
As part of multi-tenant site deployment, provide capability to generate and encrypt the BMC credentials to be used for BMH resources generated by ViNO CR so that these credentials are properly secured and are never stored unencrypted in manifests.
| non_usab | generate and encrypt bmc credentials for vino cr problem description currently the bmc credentials specified in the vino cr and used for creating bmh resources in the workload config manifest function are hard coded to clear text values these credentials need to be properly secured and managed during multi tenant type deployments proposed change as part of multi tenant site deployment provide capability to generate and encrypt the bmc credentials to be used for bmh resources generated by vino cr so that these credentials are properly secured and are never stored unencrypted in manifests | 0 |
623,154 | 19,661,955,894 | IssuesEvent | 2022-01-10 17:57:10 | GameFreedomGG/Sindragosa | https://api.github.com/repos/GameFreedomGG/Sindragosa | closed | [Quest][NPC] Teron Gorefiend, I am... - Karsius the Ancient Watcher | Priority: Low NPC Quest Status: Confirmed Bug (Awaiting Development) | Decription: Quest name: Teron Gorefiend, I am...
Location: Shadowmoon Valley
Npc who gives quest: Ancient Shadowmoon Spirit
How it works: At the beginning we need to put on your glasses (Spectrecles). After that let's talk to Ancient Shadowmoon Spirit and take his quest. Then you will change form into Teron Gorefiend and your task is: kill Karsius the Ancient Watcher. When i want to use any spell being into form to attack him or use melee attack on him, Karsius just disappears.
How it should work: Krasius the Ancient Watcher should fight with us and he shouldn't disappear during fight with us, when we use any spells into form(Teron Gorefiend) and melee attack too.
Source (you should point out proofs of your report, please give us some source):
Here is my proof what happen after take quest from Ancient Shadowmoon Spirit - just watch it, link below.
https://youtu.be/8TwpDu2bLb4
| 1.0 | [Quest][NPC] Teron Gorefiend, I am... - Karsius the Ancient Watcher - Decription: Quest name: Teron Gorefiend, I am...
Location: Shadowmoon Valley
Npc who gives quest: Ancient Shadowmoon Spirit
How it works: At the beginning we need to put on your glasses (Spectrecles). After that let's talk to Ancient Shadowmoon Spirit and take his quest. Then you will change form into Teron Gorefiend and your task is: kill Karsius the Ancient Watcher. When i want to use any spell being into form to attack him or use melee attack on him, Karsius just disappears.
How it should work: Krasius the Ancient Watcher should fight with us and he shouldn't disappear during fight with us, when we use any spells into form(Teron Gorefiend) and melee attack too.
Source (you should point out proofs of your report, please give us some source):
Here is my proof what happen after take quest from Ancient Shadowmoon Spirit - just watch it, link below.
https://youtu.be/8TwpDu2bLb4
| non_usab | teron gorefiend i am karsius the ancient watcher decription quest name teron gorefiend i am location shadowmoon valley npc who gives quest ancient shadowmoon spirit how it works at the beginning we need to put on your glasses spectrecles after that let s talk to ancient shadowmoon spirit and take his quest then you will change form into teron gorefiend and your task is kill karsius the ancient watcher when i want to use any spell being into form to attack him or use melee attack on him karsius just disappears how it should work krasius the ancient watcher should fight with us and he shouldn t disappear during fight with us when we use any spells into form teron gorefiend and melee attack too source you should point out proofs of your report please give us some source here is my proof what happen after take quest from ancient shadowmoon spirit just watch it link below | 0 |
21,590 | 17,352,145,748 | IssuesEvent | 2021-07-29 10:03:13 | rulojuka/sbking | https://api.github.com/repos/rulojuka/sbking | closed | Change strain descriptions to suit icons | good-first-issue quick-win usability | Use descriptions like :hearts: instead of "Positive Hearts" when appropriate. | True | Change strain descriptions to suit icons - Use descriptions like :hearts: instead of "Positive Hearts" when appropriate. | usab | change strain descriptions to suit icons use descriptions like hearts instead of positive hearts when appropriate | 1 |
60,769 | 8,461,337,008 | IssuesEvent | 2018-10-22 21:30:10 | segmentio/evergreen | https://api.github.com/repos/segmentio/evergreen | closed | Use @reactions/component for storybook and docs examples | Priority: Low Status: Proposal Type: Documentation | I want move over the examples and stories to use https://reactions.github.io/component/ for state and component management. This will allow examples to contain all state and data in the example, instead of referring to outside scope.
## Why?
- Allows all the context to be in the examples. No referring to outside scope.
- All examples will use one single convention.
- Easier copy-paste/migration between stories and examples.
- The https://reactions.github.io/component/ package is documented.
## How examples will look
Let's take the `SegmentedControl` component as an example. On the docs page you will see the following snippet of code with all of the state and data visible.
```jsx
<Component
initialState={{
options: [
{ label: 'Hourly', value: 'hourly' },
{ label: 'Daily', value: 'daily' },
{ label: 'Monthly', value: 'monthly' },
],
value: 'hourly'
}}
>
{({ state, setState }) => (
<SegmentedControl
width={240}
options={state.options}
value={state.value}
onChange={value => setState({ value })}
/>
)}
</Component>
```
## What we are doing now in some cases
As a side note, this component is almost the same to using the Manager that is being used in some components. The reason for moving over to https://reactions.github.io/component/ is a single documented way of doing this (instead of reinventing the wheel).
```jsx
import React from 'react'
import PropTypes from 'prop-types'
export default class Manager extends React.Component {
static propTypes = {
children: PropTypes.func,
}
constructor(props) {
super(props)
this.state = {
...props,
}
}
render() {
return this.props.children({
setState: (...args) => {
this.setState(...args)
},
state: this.state,
})
}
}
```
| 1.0 | Use @reactions/component for storybook and docs examples - I want move over the examples and stories to use https://reactions.github.io/component/ for state and component management. This will allow examples to contain all state and data in the example, instead of referring to outside scope.
## Why?
- Allows all the context to be in the examples. No referring to outside scope.
- All examples will use one single convention.
- Easier copy-paste/migration between stories and examples.
- The https://reactions.github.io/component/ package is documented.
## How examples will look
Let's take the `SegmentedControl` component as an example. On the docs page you will see the following snippet of code with all of the state and data visible.
```jsx
<Component
initialState={{
options: [
{ label: 'Hourly', value: 'hourly' },
{ label: 'Daily', value: 'daily' },
{ label: 'Monthly', value: 'monthly' },
],
value: 'hourly'
}}
>
{({ state, setState }) => (
<SegmentedControl
width={240}
options={state.options}
value={state.value}
onChange={value => setState({ value })}
/>
)}
</Component>
```
## What we are doing now in some cases
As a side note, this component is almost the same to using the Manager that is being used in some components. The reason for moving over to https://reactions.github.io/component/ is a single documented way of doing this (instead of reinventing the wheel).
```jsx
import React from 'react'
import PropTypes from 'prop-types'
export default class Manager extends React.Component {
static propTypes = {
children: PropTypes.func,
}
constructor(props) {
super(props)
this.state = {
...props,
}
}
render() {
return this.props.children({
setState: (...args) => {
this.setState(...args)
},
state: this.state,
})
}
}
```
| non_usab | use reactions component for storybook and docs examples i want move over the examples and stories to use for state and component management this will allow examples to contain all state and data in the example instead of referring to outside scope why allows all the context to be in the examples no referring to outside scope all examples will use one single convention easier copy paste migration between stories and examples the package is documented how examples will look let s take the segmentedcontrol component as an example on the docs page you will see the following snippet of code with all of the state and data visible jsx component initialstate options label hourly value hourly label daily value daily label monthly value monthly value hourly state setstate segmentedcontrol width options state options value state value onchange value setstate value what we are doing now in some cases as a side note this component is almost the same to using the manager that is being used in some components the reason for moving over to is a single documented way of doing this instead of reinventing the wheel jsx import react from react import proptypes from prop types export default class manager extends react component static proptypes children proptypes func constructor props super props this state props render return this props children setstate args this setstate args state this state | 0 |
14,856 | 9,546,799,220 | IssuesEvent | 2019-05-01 21:01:04 | meedan/montage-components | https://api.github.com/repos/meedan/montage-components | closed | Orange timeline marker should display time while being moved. | Basic Usability Standards Video Page | Suggest we do it the same way as montage.meedan.com with the expanding bubble at the top. | True | Orange timeline marker should display time while being moved. - Suggest we do it the same way as montage.meedan.com with the expanding bubble at the top. | usab | orange timeline marker should display time while being moved suggest we do it the same way as montage meedan com with the expanding bubble at the top | 1 |
357,597 | 25,176,409,778 | IssuesEvent | 2022-11-11 09:39:17 | Qiaoran-M/pe | https://api.github.com/repos/Qiaoran-M/pe | opened | Inconsistent usage of persons and contacts | severity.VeryLow type.DocumentationBug | ### description
in UG and DG, feature description, sometimes `person` is used and sometimes `contact` is used, it's better to conform to the same word `person` or `contact`

E.g. in the picture above: `adding a person`, `listing all contacts`, person and contact are actually refer to the same thing, different notation may confuse the readers
<!--session: 1668153218351-08d9e7e8-bd9e-40af-a032-8809984b1a55-->
<!--Version: Web v3.4.4--> | 1.0 | Inconsistent usage of persons and contacts - ### description
in UG and DG, feature description, sometimes `person` is used and sometimes `contact` is used, it's better to conform to the same word `person` or `contact`

E.g. in the picture above: `adding a person`, `listing all contacts`, person and contact are actually refer to the same thing, different notation may confuse the readers
<!--session: 1668153218351-08d9e7e8-bd9e-40af-a032-8809984b1a55-->
<!--Version: Web v3.4.4--> | non_usab | inconsistent usage of persons and contacts description in ug and dg feature description sometimes person is used and sometimes contact is used it s better to conform to the same word person or contact e g in the picture above adding a person listing all contacts person and contact are actually refer to the same thing different notation may confuse the readers | 0 |
71,276 | 8,641,048,087 | IssuesEvent | 2018-11-24 13:40:43 | ubuntu/yaru | https://api.github.com/repos/ubuntu/yaru | closed | Inconsistent focus rings, dashed vs. solid | design discussion | yaru-theme-* 18.10.4, Ubuntu Cosmic
Some buttons, menus, and tabs have a dashed focus ring…
 
…while other buttons, and all text fields, have a solid focus ring.


It’s not clear whether this is deliberate, but it looks inconsistent.
In the forum, [frederik-f wrote](https://discourse.ubuntu.com/t/call-for-participation-an-ubuntu-default-theme-lead-by-the-community/1545/1752):
> This one comes from a limitation in gnome-shell. A dashed focus ring is not possible there. So either we change the gtk focus ring to be solid, or we leave it as it is. There are several differences between gnome shell and the gtk theme which we accepted as a design limitation and/or a possibility to divide the the shell from the apps. :man_shrugging:
However:
- That does not explain why the focus rings for text fields and buttons are inconsistent within the same window.

- [The wiki explains that for sliders](https://github.com/ubuntu/yaru/wiki/%233-GTK-Theme-Design#gtkscale), “we had to use circles for the knob, because Gnome Shell does not support squ[ir]cles. Although this limitation only applies to the shell, we wanted the same style in both GTK and the shell.” That is solid logic — and it applies much more to focus rings than it does to sliders, because focus rings are seen much more often.
[[Originally reported in the Yaru forum](https://discourse.ubuntu.com/t/call-for-participation-an-ubuntu-default-theme-lead-by-the-community/1545/1750).]
| 1.0 | Inconsistent focus rings, dashed vs. solid - yaru-theme-* 18.10.4, Ubuntu Cosmic
Some buttons, menus, and tabs have a dashed focus ring…
 
…while other buttons, and all text fields, have a solid focus ring.


It’s not clear whether this is deliberate, but it looks inconsistent.
In the forum, [frederik-f wrote](https://discourse.ubuntu.com/t/call-for-participation-an-ubuntu-default-theme-lead-by-the-community/1545/1752):
> This one comes from a limitation in gnome-shell. A dashed focus ring is not possible there. So either we change the gtk focus ring to be solid, or we leave it as it is. There are several differences between gnome shell and the gtk theme which we accepted as a design limitation and/or a possibility to divide the the shell from the apps. :man_shrugging:
However:
- That does not explain why the focus rings for text fields and buttons are inconsistent within the same window.

- [The wiki explains that for sliders](https://github.com/ubuntu/yaru/wiki/%233-GTK-Theme-Design#gtkscale), “we had to use circles for the knob, because Gnome Shell does not support squ[ir]cles. Although this limitation only applies to the shell, we wanted the same style in both GTK and the shell.” That is solid logic — and it applies much more to focus rings than it does to sliders, because focus rings are seen much more often.
[[Originally reported in the Yaru forum](https://discourse.ubuntu.com/t/call-for-participation-an-ubuntu-default-theme-lead-by-the-community/1545/1750).]
| non_usab | inconsistent focus rings dashed vs solid yaru theme ubuntu cosmic some buttons menus and tabs have a dashed focus ring… …while other buttons and all text fields have a solid focus ring it’s not clear whether this is deliberate but it looks inconsistent in the forum this one comes from a limitation in gnome shell a dashed focus ring is not possible there so either we change the gtk focus ring to be solid or we leave it as it is there are several differences between gnome shell and the gtk theme which we accepted as a design limitation and or a possibility to divide the the shell from the apps man shrugging however that does not explain why the focus rings for text fields and buttons are inconsistent within the same window “we had to use circles for the knob because gnome shell does not support squ cles although this limitation only applies to the shell we wanted the same style in both gtk and the shell ” that is solid logic — and it applies much more to focus rings than it does to sliders because focus rings are seen much more often | 0 |
7,249 | 4,836,815,325 | IssuesEvent | 2016-11-08 20:43:16 | mitodl/micromasters | https://api.github.com/repos/mitodl/micromasters | closed | Sign up: Change label for "preferred name" and pre-populate it with given name | usability | - [ ] Change the label to "nickname / preferred name"
- [ ] Prepopulate it with user's given name
| True | Sign up: Change label for "preferred name" and pre-populate it with given name - - [ ] Change the label to "nickname / preferred name"
- [ ] Prepopulate it with user's given name
| usab | sign up change label for preferred name and pre populate it with given name change the label to nickname preferred name prepopulate it with user s given name | 1 |
18,674 | 13,157,264,526 | IssuesEvent | 2020-08-10 12:26:09 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | Theme editor opens itself on scene save | bug confirmed topic:editor usability | **Godot version:**
4ee8ecd
**Issue description:**
When a Control node has a Theme resource and the resource is unfolded in the inspector, the Theme Editor dock will open each time you save the scene. IMO it should only open when interacting with the theme. I had it open in the script editor until I realized I need to fold the Theme resource.
**Steps to reproduce:**
1. Create a Control node
2. Create some Theme resource for the node
3. Unfold the new resource
4. Close the "Theme" dock and go somewhere else (like Console, Script Editor etc.)
5. Save
6. Theme dock opens itself | True | Theme editor opens itself on scene save - **Godot version:**
4ee8ecd
**Issue description:**
When a Control node has a Theme resource and the resource is unfolded in the inspector, the Theme Editor dock will open each time you save the scene. IMO it should only open when interacting with the theme. I had it open in the script editor until I realized I need to fold the Theme resource.
**Steps to reproduce:**
1. Create a Control node
2. Create some Theme resource for the node
3. Unfold the new resource
4. Close the "Theme" dock and go somewhere else (like Console, Script Editor etc.)
5. Save
6. Theme dock opens itself | usab | theme editor opens itself on scene save godot version issue description when a control node has a theme resource and the resource is unfolded in the inspector the theme editor dock will open each time you save the scene imo it should only open when interacting with the theme i had it open in the script editor until i realized i need to fold the theme resource steps to reproduce create a control node create some theme resource for the node unfold the new resource close the theme dock and go somewhere else like console script editor etc save theme dock opens itself | 1 |
22,963 | 20,755,981,033 | IssuesEvent | 2022-03-15 12:15:31 | pulumi/pulumi | https://api.github.com/repos/pulumi/pulumi | closed | `pulumi import` in the cli not working with certain Dynamo DB tables | kind/bug impact/usability area/providers area/core resolution/fixed area/import | <!-- Please provide a general summary of the issue. -->
Trying to import a dynamo DB using `pulumi import` can sometime result in the following error: `All attributes must be indexed. Unused attributes: ["<name of attribute>"]`
We can import a table with no secondary indexes and no sort key, but on one table we get the error on the secondary index key, and on the other we get the error on the sort key.
## Expected behavior
<!-- Please tell us what should happen -->
It should import the table and provide the code to recreate it in Typescript/NodeJS
## Current behavior
<!-- Please tell us what happens instead of the expected behavior -->
Running the command `pulumi import aws:dynamodb/table:Table <pulumi resource name> <aws resource name>` gives us an error stating
```
aws:dynamodb:Table (foo):
error: Preview failed: diffing urn:pulumi:local::foobar::aws:dynamodb/table:Table::foo: 1 error occurred:
* All attributes must be indexed. Unused attributes: ["bar"]
```
## Steps to reproduce
<!-- Provide a link to a live example, or an unambiguous set of steps to reproduce this bug. Include code to reproduce, if relevant. -->
1. Create a dynamodb table with a secondary index, or a sort key, and populate it with some data
2. run `pulumi import` on that table
## Context (Environment)
<!-- How has this issue affected you? What are you trying to accomplish? Providing context helps us come up with a solution that is most useful in the real world -->
We're working on moving our CloudFormation IAC fully to Pulumi
## Affected feature
`pulumi import` in the cli
<!-- If you know that your bug is happening in a specific feature, please list it here. And, if you see a matching label with the `area/` prefix, please feel free to apply it. -->
| True | `pulumi import` in the cli not working with certain Dynamo DB tables - <!-- Please provide a general summary of the issue. -->
Trying to import a dynamo DB using `pulumi import` can sometime result in the following error: `All attributes must be indexed. Unused attributes: ["<name of attribute>"]`
We can import a table with no secondary indexes and no sort key, but on one table we get the error on the secondary index key, and on the other we get the error on the sort key.
## Expected behavior
<!-- Please tell us what should happen -->
It should import the table and provide the code to recreate it in Typescript/NodeJS
## Current behavior
<!-- Please tell us what happens instead of the expected behavior -->
Running the command `pulumi import aws:dynamodb/table:Table <pulumi resource name> <aws resource name>` gives us an error stating
```
aws:dynamodb:Table (foo):
error: Preview failed: diffing urn:pulumi:local::foobar::aws:dynamodb/table:Table::foo: 1 error occurred:
* All attributes must be indexed. Unused attributes: ["bar"]
```
## Steps to reproduce
<!-- Provide a link to a live example, or an unambiguous set of steps to reproduce this bug. Include code to reproduce, if relevant. -->
1. Create a dynamodb table with a secondary index, or a sort key, and populate it with some data
2. run `pulumi import` on that table
## Context (Environment)
<!-- How has this issue affected you? What are you trying to accomplish? Providing context helps us come up with a solution that is most useful in the real world -->
We're working on moving our CloudFormation IAC fully to Pulumi
## Affected feature
`pulumi import` in the cli
<!-- If you know that your bug is happening in a specific feature, please list it here. And, if you see a matching label with the `area/` prefix, please feel free to apply it. -->
| usab | pulumi import in the cli not working with certain dynamo db tables trying to import a dynamo db using pulumi import can sometime result in the following error all attributes must be indexed unused attributes we can import a table with no secondary indexes and no sort key but on one table we get the error on the secondary index key and on the other we get the error on the sort key expected behavior it should import the table and provide the code to recreate it in typescript nodejs current behavior running the command pulumi import aws dynamodb table table gives us an error stating aws dynamodb table foo error preview failed diffing urn pulumi local foobar aws dynamodb table table foo error occurred all attributes must be indexed unused attributes steps to reproduce create a dynamodb table with a secondary index or a sort key and populate it with some data run pulumi import on that table context environment we re working on moving our cloudformation iac fully to pulumi affected feature pulumi import in the cli | 1 |
25,336 | 24,993,611,484 | IssuesEvent | 2022-11-02 21:11:12 | mixxxdj/mixxx | https://api.github.com/repos/mixxxdj/mixxx | reopened | show ControlPotmeter and EffectParameter values in skins | bug confirmed usability | Reported by: **[Be-ing](https://github.com/Be-ing)**
Date: 2017-12-28T08:46:33Z
Status: Confirmed
Importance: Medium
Launchpad Issue: [lp1740372](https://bugs.launchpad.net/bugs/1740372)
Tags: usability
---
We describe the scales of effect parameters in parameter tooltips, but it is impossible to know exactly what value a knob is at. It would be really helpful to show this beside the knob. I think Traktor does this quite well. In Traktor, the value of a knob replaces its label while the knob is being adjusted. Watch this video for an example: https://www.youtube.com/watch?v=rRg0fXlYYDI&t=4m40s
Related: Bug #⁠1419949 | True | show ControlPotmeter and EffectParameter values in skins - Reported by: **[Be-ing](https://github.com/Be-ing)**
Date: 2017-12-28T08:46:33Z
Status: Confirmed
Importance: Medium
Launchpad Issue: [lp1740372](https://bugs.launchpad.net/bugs/1740372)
Tags: usability
---
We describe the scales of effect parameters in parameter tooltips, but it is impossible to know exactly what value a knob is at. It would be really helpful to show this beside the knob. I think Traktor does this quite well. In Traktor, the value of a knob replaces its label while the knob is being adjusted. Watch this video for an example: https://www.youtube.com/watch?v=rRg0fXlYYDI&t=4m40s
Related: Bug #⁠1419949 | usab | show controlpotmeter and effectparameter values in skins reported by date status confirmed importance medium launchpad issue tags usability we describe the scales of effect parameters in parameter tooltips but it is impossible to know exactly what value a knob is at it would be really helpful to show this beside the knob i think traktor does this quite well in traktor the value of a knob replaces its label while the knob is being adjusted watch this video for an example related bug | 1 |
298,066 | 9,195,487,793 | IssuesEvent | 2019-03-07 02:38:07 | clearlinux/clr-installer | https://api.github.com/repos/clearlinux/clr-installer | closed | Installer permits setting a static IP. Setting is not used. | UX bug high priority | **Describe the bug**
Installer permits setting a static IP. Setting is not used.
**To Reproduce**
Boot installer. Go to advanced settings. Set a static IP. See (in wireshark) how installer tries to obtain a dhcp lease.
**Expected behavior**
Upon setting a static IP address, I would expect the networking requirements to be fulfilled and the installer able to continue installing.
**Environment (please complete the following information):**
NAME="Clear Linux OS"
VERSION=1
ID=clear-linux-os
ID_LIKE=clear-linux-os
VERSION_ID=27910
PRETTY_NAME="Clear Linux OS"
ANSI_COLOR="1;35"
HOME_URL="https://clearlinux.org"
SUPPORT_URL="https://clearlinux.org"
BUG_REPORT_URL="mailto:dev@lists.clearlinux.org"
PRIVACY_POLICY_URL="http://www.intel.com/privacy"
| 1.0 | Installer permits setting a static IP. Setting is not used. - **Describe the bug**
Installer permits setting a static IP. Setting is not used.
**To Reproduce**
Boot installer. Go to advanced settings. Set a static IP. See (in wireshark) how installer tries to obtain a dhcp lease.
**Expected behavior**
Upon setting a static IP address, I would expect the networking requirements to be fulfilled and the installer able to continue installing.
**Environment (please complete the following information):**
NAME="Clear Linux OS"
VERSION=1
ID=clear-linux-os
ID_LIKE=clear-linux-os
VERSION_ID=27910
PRETTY_NAME="Clear Linux OS"
ANSI_COLOR="1;35"
HOME_URL="https://clearlinux.org"
SUPPORT_URL="https://clearlinux.org"
BUG_REPORT_URL="mailto:dev@lists.clearlinux.org"
PRIVACY_POLICY_URL="http://www.intel.com/privacy"
| non_usab | installer permits setting a static ip setting is not used describe the bug installer permits setting a static ip setting is not used to reproduce boot installer go to advanced settings set a static ip see in wireshark how installer tries to obtain a dhcp lease expected behavior upon setting a static ip address i would expect the networking requirements to be fulfilled and the installer able to continue installing environment please complete the following information name clear linux os version id clear linux os id like clear linux os version id pretty name clear linux os ansi color home url support url bug report url mailto dev lists clearlinux org privacy policy url | 0 |
14,438 | 9,186,989,566 | IssuesEvent | 2019-03-06 00:56:30 | uber/hudi | https://api.github.com/repos/uber/hudi | closed | Tool to convert Hive tables into hoodie dataset | usability | Look at HoodieDataImporter --source-type. Currently it supports "hdfs" (reading from parquet files as is). Scope for this bug is to implement "hive" [source-type] to read from registered hive tables using HiveContext. | True | Tool to convert Hive tables into hoodie dataset - Look at HoodieDataImporter --source-type. Currently it supports "hdfs" (reading from parquet files as is). Scope for this bug is to implement "hive" [source-type] to read from registered hive tables using HiveContext. | usab | tool to convert hive tables into hoodie dataset look at hoodiedataimporter source type currently it supports hdfs reading from parquet files as is scope for this bug is to implement hive to read from registered hive tables using hivecontext | 1 |
90,555 | 26,136,306,720 | IssuesEvent | 2022-12-29 12:34:49 | prusa3d/Prusa-Firmware | https://api.github.com/repos/prusa3d/Prusa-Firmware | closed | [QUESTION] Building under linux | question build | Hello,
I'm pretty sure "sudo" is not needed to compile the firmware (except for package installation), hence I believe doc should be edited to remove the "sudo" in front of the build.sh. | 1.0 | [QUESTION] Building under linux - Hello,
I'm pretty sure "sudo" is not needed to compile the firmware (except for package installation), hence I believe doc should be edited to remove the "sudo" in front of the build.sh. | non_usab | building under linux hello i m pretty sure sudo is not needed to compile the firmware except for package installation hence i believe doc should be edited to remove the sudo in front of the build sh | 0 |
551,999 | 16,192,771,296 | IssuesEvent | 2021-05-04 10:51:26 | WaterGOV/WaterCore | https://api.github.com/repos/WaterGOV/WaterCore | opened | Scoreboard | priority: low type: feature | - Customisable scoreboard module
- Multiple different boards depending on user
- Togglable
- Animations (this is a big feature - might be best left to another plugin?) | 1.0 | Scoreboard - - Customisable scoreboard module
- Multiple different boards depending on user
- Togglable
- Animations (this is a big feature - might be best left to another plugin?) | non_usab | scoreboard customisable scoreboard module multiple different boards depending on user togglable animations this is a big feature might be best left to another plugin | 0 |
10,656 | 6,845,439,925 | IssuesEvent | 2017-11-13 08:15:26 | apinf/platform | https://api.github.com/repos/apinf/platform | closed | Misplaced ":"s and wrong translations in footer | in progress Team DLG Usability Issue | **Steps to Reproduce**:
1. go to page: https://seinajoki.apinf.io/
2. scroll down to footer
**Outcome**:
In Finnish lang version, the ":"s are in wrong place and the translations are not right.
<img width="1113" alt="screenshot at marraskuuta 08 21-17-04" src="https://user-images.githubusercontent.com/1927158/32569718-91ccd062-c4ca-11e7-8ac3-6ae61a0f2f09.png">
**Expected Behavior**:
Correct would be:
- Remove the ":"s from all Finnish translation strings
- Correct the transtions to "Rajapintaa" (APIs), "Organisaatiota" (Organizations), "Käyttäjää" (Users)
**Environment**:
MacOS, FF (56.0.2 ), APInf customer instance (https://seinajoki.apinf.io/)
| True | Misplaced ":"s and wrong translations in footer - **Steps to Reproduce**:
1. go to page: https://seinajoki.apinf.io/
2. scroll down to footer
**Outcome**:
In Finnish lang version, the ":"s are in wrong place and the translations are not right.
<img width="1113" alt="screenshot at marraskuuta 08 21-17-04" src="https://user-images.githubusercontent.com/1927158/32569718-91ccd062-c4ca-11e7-8ac3-6ae61a0f2f09.png">
**Expected Behavior**:
Correct would be:
- Remove the ":"s from all Finnish translation strings
- Correct the transtions to "Rajapintaa" (APIs), "Organisaatiota" (Organizations), "Käyttäjää" (Users)
**Environment**:
MacOS, FF (56.0.2 ), APInf customer instance (https://seinajoki.apinf.io/)
| usab | misplaced s and wrong translations in footer steps to reproduce go to page scroll down to footer outcome in finnish lang version the s are in wrong place and the translations are not right img width alt screenshot at marraskuuta src expected behavior correct would be remove the s from all finnish translation strings correct the transtions to rajapintaa apis organisaatiota organizations käyttäjää users environment macos ff apinf customer instance | 1 |
64,189 | 3,206,071,852 | IssuesEvent | 2015-10-04 18:07:21 | WarEmu/WarBugs | https://api.github.com/repos/WarEmu/WarBugs | closed | Auction house sometimes sends items twice | Mail Medium Priority | Auction house sometimes sends items twice. You get a mail, take the item and later get the same mail again. | 1.0 | Auction house sometimes sends items twice - Auction house sometimes sends items twice. You get a mail, take the item and later get the same mail again. | non_usab | auction house sometimes sends items twice auction house sometimes sends items twice you get a mail take the item and later get the same mail again | 0 |
1,222 | 2,757,830,305 | IssuesEvent | 2015-04-27 16:51:51 | icebreaker/2dimagefilter | https://api.github.com/repos/icebreaker/2dimagefilter | closed | Associate scripts with ImageResizer | auto-migrated Type-Enhancement Usability | ```
Will it possible to add option for associating .irs Script with ImageResizer,
therefore the script can become executable by double clicking it. Just like
.bat file associated with CMD.exe or .py file associated with Python.exe.
```
Original issue reported on code.google.com by `ryanb...@gmail.com` on 6 Apr 2013 at 1:23 | True | Associate scripts with ImageResizer - ```
Will it possible to add option for associating .irs Script with ImageResizer,
therefore the script can become executable by double clicking it. Just like
.bat file associated with CMD.exe or .py file associated with Python.exe.
```
Original issue reported on code.google.com by `ryanb...@gmail.com` on 6 Apr 2013 at 1:23 | usab | associate scripts with imageresizer will it possible to add option for associating irs script with imageresizer therefore the script can become executable by double clicking it just like bat file associated with cmd exe or py file associated with python exe original issue reported on code google com by ryanb gmail com on apr at | 1 |
22,703 | 20,018,239,509 | IssuesEvent | 2022-02-01 14:11:07 | informalsystems/apalache | https://api.github.com/repos/informalsystems/apalache | opened | [FEATURE] Detect uninitialized `CONSTANT`s in the `ConfigPass` | usability | Currently, uninitialized constants trigger an exception during the BMC pass (after the spec has gone through several transformations), for example:
```
PASS #13: BoundedChecker I@15:04:01.473
This error may show up when CONSTANTS are not initialized. E@15:04:02.443
Check the manual: https://apalache.informal.systems/docs/apalache/parameters.html E@15:04:02.444
Input error (see the manual): SubstRule: Variable N is not assigned a value E@15:04:02.446
```
This should be detected and handled much earlier, in the configuration pass | True | [FEATURE] Detect uninitialized `CONSTANT`s in the `ConfigPass` - Currently, uninitialized constants trigger an exception during the BMC pass (after the spec has gone through several transformations), for example:
```
PASS #13: BoundedChecker I@15:04:01.473
This error may show up when CONSTANTS are not initialized. E@15:04:02.443
Check the manual: https://apalache.informal.systems/docs/apalache/parameters.html E@15:04:02.444
Input error (see the manual): SubstRule: Variable N is not assigned a value E@15:04:02.446
```
This should be detected and handled much earlier, in the configuration pass | usab | detect uninitialized constant s in the configpass currently uninitialized constants trigger an exception during the bmc pass after the spec has gone through several transformations for example pass boundedchecker i this error may show up when constants are not initialized e check the manual e input error see the manual substrule variable n is not assigned a value e this should be detected and handled much earlier in the configuration pass | 1 |
215,783 | 7,297,761,411 | IssuesEvent | 2018-02-26 15:09:06 | fourkitchens/emulsify-gulp | https://api.github.com/repos/fourkitchens/emulsify-gulp | opened | Can't use "Crimson Text" Google font CSS due to gulp-clean-css bug | Priority: Low | From @nsciacca [here](https://github.com/fourkitchens/emulsify/issues/201):
> I'm trying to use Google's font "Crimson Text", however the compiled CSS is replacing "Crimson" with the hexcode "#dc143c". I started poking around the gulp-css.js file and found if I removed the call to "cleanCSS" (gulp-clean-css) then it works fine.
>
> I tried setting up a simple node project with gulp-clean-css and replicated the error, apparently in the default optimization (level 1) there's a shorten-hex.js file that maps these color keywords to their hex values.
>
> I've logged a new issue on that page. Hopefully they'll fix it on their end but for the moment I'm leaving the cleanCSS commented out. Maybe the config for emulsify should be updated to allow passing options to cleanCSS, in which we could specify the level of optimization.
Posted a [workaround](https://github.com/jakubpawlowicz/clean-css#how-to-keep-a-css-fragment-intact), but we could maybe look into controlling options via Emulsify Gulp. | 1.0 | Can't use "Crimson Text" Google font CSS due to gulp-clean-css bug - From @nsciacca [here](https://github.com/fourkitchens/emulsify/issues/201):
> I'm trying to use Google's font "Crimson Text", however the compiled CSS is replacing "Crimson" with the hexcode "#dc143c". I started poking around the gulp-css.js file and found if I removed the call to "cleanCSS" (gulp-clean-css) then it works fine.
>
> I tried setting up a simple node project with gulp-clean-css and replicated the error, apparently in the default optimization (level 1) there's a shorten-hex.js file that maps these color keywords to their hex values.
>
> I've logged a new issue on that page. Hopefully they'll fix it on their end but for the moment I'm leaving the cleanCSS commented out. Maybe the config for emulsify should be updated to allow passing options to cleanCSS, in which we could specify the level of optimization.
Posted a [workaround](https://github.com/jakubpawlowicz/clean-css#how-to-keep-a-css-fragment-intact), but we could maybe look into controlling options via Emulsify Gulp. | non_usab | can t use crimson text google font css due to gulp clean css bug from nsciacca i m trying to use google s font crimson text however the compiled css is replacing crimson with the hexcode i started poking around the gulp css js file and found if i removed the call to cleancss gulp clean css then it works fine i tried setting up a simple node project with gulp clean css and replicated the error apparently in the default optimization level there s a shorten hex js file that maps these color keywords to their hex values i ve logged a new issue on that page hopefully they ll fix it on their end but for the moment i m leaving the cleancss commented out maybe the config for emulsify should be updated to allow passing options to cleancss in which we could specify the level of optimization posted a but we could maybe look into controlling options via emulsify gulp | 0 |
1,297 | 2,787,810,260 | IssuesEvent | 2015-05-08 09:10:11 | grafana/grafana | https://api.github.com/repos/grafana/grafana | closed | Feature Request: Add Table of Contents and hierarchy features for navigation | feature request ui:usability | The ability to search for pages, graphs, and metrics is very nice, but there is a point where this will not scale well. It would be very helpful to have an easy way to setup a hierarchy with links to pages. For example, the ability to create parent and children pages where the links to children are auto-populated on the parent pages. And/Or a menu at the top of the pages for navigation. | True | Feature Request: Add Table of Contents and hierarchy features for navigation - The ability to search for pages, graphs, and metrics is very nice, but there is a point where this will not scale well. It would be very helpful to have an easy way to setup a hierarchy with links to pages. For example, the ability to create parent and children pages where the links to children are auto-populated on the parent pages. And/Or a menu at the top of the pages for navigation. | usab | feature request add table of contents and hierarchy features for navigation the ability to search for pages graphs and metrics is very nice but there is a point where this will not scale well it would be very helpful to have an easy way to setup a hierarchy with links to pages for example the ability to create parent and children pages where the links to children are auto populated on the parent pages and or a menu at the top of the pages for navigation | 1 |
168,590 | 26,666,702,484 | IssuesEvent | 2023-01-26 05:13:05 | bigbomio/bigbom-marketplace | https://api.github.com/repos/bigbomio/bigbom-marketplace | closed | Build in-app wallet feature | enhancement design-proposal | In order to have better UX for users, we came up with a solution to build an in-app wallet management system. This app will have these features:
1. Able to create and restore wallet from seed phrase. Password-protected encrypted key vault.
2. Support remote storage sync with Gdrive, Dropbox, Blockstacks or IPFSCloud
3. Able to manage ETH and Token balance. At first it will support ETH, BBO, DAI, USDC and TUSD.
4. Able sign & broadcast transaction to different network. This will provide the capabilities of offloading some transactions to different network than Ethereum Mainnet, while keeping all the payment-related on the Ethereum Mainnet.
5. Integrates into the app like as web component, so user can sign & confirm transaction like normal applications.
The idea was from seeing the use of eth-lightwallet, and we feel that we can use part of Metamask source code to build our own wallet management system. | 1.0 | Build in-app wallet feature - In order to have better UX for users, we came up with a solution to build an in-app wallet management system. This app will have these features:
1. Able to create and restore wallet from seed phrase. Password-protected encrypted key vault.
2. Support remote storage sync with Gdrive, Dropbox, Blockstacks or IPFSCloud
3. Able to manage ETH and Token balance. At first it will support ETH, BBO, DAI, USDC and TUSD.
4. Able sign & broadcast transaction to different network. This will provide the capabilities of offloading some transactions to different network than Ethereum Mainnet, while keeping all the payment-related on the Ethereum Mainnet.
5. Integrates into the app like as web component, so user can sign & confirm transaction like normal applications.
The idea was from seeing the use of eth-lightwallet, and we feel that we can use part of Metamask source code to build our own wallet management system. | non_usab | build in app wallet feature in order to have better ux for users we came up with a solution to build an in app wallet management system this app will have these features able to create and restore wallet from seed phrase password protected encrypted key vault support remote storage sync with gdrive dropbox blockstacks or ipfscloud able to manage eth and token balance at first it will support eth bbo dai usdc and tusd able sign broadcast transaction to different network this will provide the capabilities of offloading some transactions to different network than ethereum mainnet while keeping all the payment related on the ethereum mainnet integrates into the app like as web component so user can sign confirm transaction like normal applications the idea was from seeing the use of eth lightwallet and we feel that we can use part of metamask source code to build our own wallet management system | 0 |
27,419 | 28,433,691,020 | IssuesEvent | 2023-04-15 03:46:18 | enquirer/enquirer | https://api.github.com/repos/enquirer/enquirer | opened | AutoComplete prompt - improve highlighting | usability | Currently the `highlight` function only highlights the first matching character in each item. This looks suboptimal when more than one character matches and additional input causes the next match to be highlighted, resulting in the highlighted text "jumping" from one character to later characters on the same line.
**Solution**
Use something like the following to highlight all matching sequences:
```js
const highlight = (input, color) => {
const val = input.toLowerCase();
return str => str.toLowerCase().split(val).join(color(val));
};
``` | True | AutoComplete prompt - improve highlighting - Currently the `highlight` function only highlights the first matching character in each item. This looks suboptimal when more than one character matches and additional input causes the next match to be highlighted, resulting in the highlighted text "jumping" from one character to later characters on the same line.
**Solution**
Use something like the following to highlight all matching sequences:
```js
const highlight = (input, color) => {
const val = input.toLowerCase();
return str => str.toLowerCase().split(val).join(color(val));
};
``` | usab | autocomplete prompt improve highlighting currently the highlight function only highlights the first matching character in each item this looks suboptimal when more than one character matches and additional input causes the next match to be highlighted resulting in the highlighted text jumping from one character to later characters on the same line solution use something like the following to highlight all matching sequences js const highlight input color const val input tolowercase return str str tolowercase split val join color val | 1 |
329,090 | 24,208,409,508 | IssuesEvent | 2022-09-25 15:04:58 | KendallDoesCoding/penny-lane | https://api.github.com/repos/KendallDoesCoding/penny-lane | closed | [DOCS] Write comments | documentation good first issue EddieHub:good-first-issue | Try to understand what the code does and write a comment on how it works in HTML, CSS, JS files.
| 1.0 | [DOCS] Write comments - Try to understand what the code does and write a comment on how it works in HTML, CSS, JS files.
| non_usab | write comments try to understand what the code does and write a comment on how it works in html css js files | 0 |
384,857 | 26,605,384,835 | IssuesEvent | 2023-01-23 18:54:03 | MiguelAngelTorres/quini-utils | https://api.github.com/repos/MiguelAngelTorres/quini-utils | closed | Improve readme | documentation | Give more details about the installation process and write a couple of examples more | 1.0 | Improve readme - Give more details about the installation process and write a couple of examples more | non_usab | improve readme give more details about the installation process and write a couple of examples more | 0 |
14,712 | 9,437,767,226 | IssuesEvent | 2019-04-13 17:41:59 | bird-get/terminal-hud | https://api.github.com/repos/bird-get/terminal-hud | opened | Improve region change handling | enhancement usability | When changing region, the display resets and the history is gone. It would be preferable if this doesn't happen. | True | Improve region change handling - When changing region, the display resets and the history is gone. It would be preferable if this doesn't happen. | usab | improve region change handling when changing region the display resets and the history is gone it would be preferable if this doesn t happen | 1 |
261,068 | 19,696,913,121 | IssuesEvent | 2022-01-12 13:07:49 | POEC-DOTNET-CLERMONT-2022/Fantastic3D | https://api.github.com/repos/POEC-DOTNET-CLERMONT-2022/Fantastic3D | opened | Readme.md du projet | documentation | - Déplacer les éléments en bon endroit
- Ajouter des screenshots
- Faire des readme.md spécifique aux app déjà OK (ConsoleApp) | 1.0 | Readme.md du projet - - Déplacer les éléments en bon endroit
- Ajouter des screenshots
- Faire des readme.md spécifique aux app déjà OK (ConsoleApp) | non_usab | readme md du projet déplacer les éléments en bon endroit ajouter des screenshots faire des readme md spécifique aux app déjà ok consoleapp | 0 |
9,141 | 6,151,832,062 | IssuesEvent | 2017-06-28 04:34:56 | Virtual-Labs/structural-dynamics-iiith | https://api.github.com/repos/Virtual-Labs/structural-dynamics-iiith | closed | QA_Impulse Response of S.D.O.F System_Procedure_Content-needs-to-be-updated | Category: Usability Developed by: VLEAD Open-Edx Resolved Severity: S2 Status: Open | Defect Description :
In the Procedure page of Impulse Response of S.D.O.F System experiment, the content says "The section will be updated soon". Need to either include the content or remove the section.
Actual Result :
In the Procedure page of Impulse Response of S.D.O.F System experiment, the content says "The section will be updated soon".
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM
Processor:i5
Attachment

| True | QA_Impulse Response of S.D.O.F System_Procedure_Content-needs-to-be-updated - Defect Description :
In the Procedure page of Impulse Response of S.D.O.F System experiment, the content says "The section will be updated soon". Need to either include the content or remove the section.
Actual Result :
In the Procedure page of Impulse Response of S.D.O.F System experiment, the content says "The section will be updated soon".
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM
Processor:i5
Attachment

| usab | qa impulse response of s d o f system procedure content needs to be updated defect description in the procedure page of impulse response of s d o f system experiment the content says the section will be updated soon need to either include the content or remove the section actual result in the procedure page of impulse response of s d o f system experiment the content says the section will be updated soon environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor attachment | 1 |
117,687 | 17,512,679,196 | IssuesEvent | 2021-08-11 01:04:58 | harrinry/pulsar | https://api.github.com/repos/harrinry/pulsar | opened | CVE-2020-24616 (High) detected in jackson-databind-2.8.11.4.jar, jackson-databind-2.6.5.jar | security vulnerability | ## CVE-2020-24616 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.11.4.jar</b>, <b>jackson-databind-2.6.5.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.11.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: pulsar/pulsar-sql/presto-distribution/pom.xml</p>
<p>Path to vulnerable library: 0150316_LVRAMP/downloadResource_AEDNMT/20210810150945/jackson-databind-2.8.11.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.11.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: pulsar/examples/spark/pom.xml</p>
<p>Path to vulnerable library: 0150316_LVRAMP/downloadResource_AEDNMT/20210810150943/jackson-databind-2.6.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.6.5.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).
<p>Publish Date: 2020-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-08-25</p>
<p>Fix Resolution: 2.9.10.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.4","packageFilePaths":["/pulsar-sql/presto-distribution/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.11.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.5","packageFilePaths":["/examples/spark/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24616","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-24616 (High) detected in jackson-databind-2.8.11.4.jar, jackson-databind-2.6.5.jar - ## CVE-2020-24616 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.11.4.jar</b>, <b>jackson-databind-2.6.5.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.11.4.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: pulsar/pulsar-sql/presto-distribution/pom.xml</p>
<p>Path to vulnerable library: 0150316_LVRAMP/downloadResource_AEDNMT/20210810150945/jackson-databind-2.8.11.4.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.11.4.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.6.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: pulsar/examples/spark/pom.xml</p>
<p>Path to vulnerable library: 0150316_LVRAMP/downloadResource_AEDNMT/20210810150943/jackson-databind-2.6.5.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.6.5.jar** (Vulnerable Library)
</details>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).
<p>Publish Date: 2020-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616>CVE-2020-24616</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-24616</a></p>
<p>Release Date: 2020-08-25</p>
<p>Fix Resolution: 2.9.10.6</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.11.4","packageFilePaths":["/pulsar-sql/presto-distribution/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.11.4","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.6.5","packageFilePaths":["/examples/spark/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.6.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"2.9.10.6"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-24616","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.6 mishandles the interaction between serialization gadgets and typing, related to br.com.anteros.dbcp.AnterosDBCPDataSource (aka Anteros-DBCP).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-24616","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_usab | cve high detected in jackson databind jar jackson databind jar cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pulsar pulsar sql presto distribution pom xml path to vulnerable library lvramp downloadresource aednmt jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file pulsar examples spark pom xml path to vulnerable library lvramp downloadresource aednmt jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to br com anteros dbcp anterosdbcpdatasource aka anteros dbcp vulnerabilityurl | 0 |
764,640 | 26,809,690,181 | IssuesEvent | 2023-02-01 21:13:40 | bcgov/entity | https://api.github.com/repos/bcgov/entity | closed | NAMEX outputs - Update Consent Processing time from 1 day to 2 days | bug NameX Priority2 ENTITY SRE | **Describe the bug in current situation**
This ticket originated with OPS ticket #1832.
Consent processing is done on a priority basis – 2 business days. The output on the NR approvals is stating one day. Please see the output attached. Can we get this updated?

[https://app.zenhub.com/files/157936592/673bebd5-3c66-4f3c-ac67-47953a76a5f6/download](https://app.zenhub.com/files/157936592/673bebd5-3c66-4f3c-ac67-47953a76a5f6/download)
**Link bug to the User Story**
**Impact of this bug**
Information on Name Request output is not accurate and clients may be led to believe transactions will be processed quicker than they really will.
**Chance of Occurring (high/medium/low/very low)**
High
**Pre Conditions: which Env, any pre-requesites or assumptions to execute steps?**
PROD
**Steps to Reproduce**
Steps to reproduce the behavior:
1. Request name with consent required
2. Approve name with consent required
3. View Name Request output document
**Actual/ observed behavior/ results**
Output indicates:
Consent letters will be processed within one business day of receipt.
**Expected behavior**
Output indicates:
Consent letters will be processed within two business days of receipt.
**Screenshots/ Visual Reference/ Source**
If applicable, add screenshots to help explain your problem. You an use screengrab.
| 1.0 | NAMEX outputs - Update Consent Processing time from 1 day to 2 days - **Describe the bug in current situation**
This ticket originated with OPS ticket #1832.
Consent processing is done on a priority basis – 2 business days. The output on the NR approvals is stating one day. Please see the output attached. Can we get this updated?

[https://app.zenhub.com/files/157936592/673bebd5-3c66-4f3c-ac67-47953a76a5f6/download](https://app.zenhub.com/files/157936592/673bebd5-3c66-4f3c-ac67-47953a76a5f6/download)
**Link bug to the User Story**
**Impact of this bug**
Information on Name Request output is not accurate and clients may be led to believe transactions will be processed quicker than they really will.
**Chance of Occurring (high/medium/low/very low)**
High
**Pre Conditions: which Env, any pre-requesites or assumptions to execute steps?**
PROD
**Steps to Reproduce**
Steps to reproduce the behavior:
1. Request name with consent required
2. Approve name with consent required
3. View Name Request output document
**Actual/ observed behavior/ results**
Output indicates:
Consent letters will be processed within one business day of receipt.
**Expected behavior**
Output indicates:
Consent letters will be processed within two business days of receipt.
**Screenshots/ Visual Reference/ Source**
If applicable, add screenshots to help explain your problem. You an use screengrab.
| non_usab | namex outputs update consent processing time from day to days describe the bug in current situation this ticket originated with ops ticket consent processing is done on a priority basis – business days the output on the nr approvals is stating one day please see the output attached can we get this updated link bug to the user story impact of this bug information on name request output is not accurate and clients may be led to believe transactions will be processed quicker than they really will chance of occurring high medium low very low high pre conditions which env any pre requesites or assumptions to execute steps prod steps to reproduce steps to reproduce the behavior request name with consent required approve name with consent required view name request output document actual observed behavior results output indicates consent letters will be processed within one business day of receipt expected behavior output indicates consent letters will be processed within two business days of receipt screenshots visual reference source if applicable add screenshots to help explain your problem you an use screengrab | 0 |
317,871 | 27,272,355,513 | IssuesEvent | 2023-02-22 23:54:43 | rancher/rancher | https://api.github.com/repos/rancher/rancher | opened | rancher-logging kube-audit pods should only run on CP nodes | kind/bug internal area/logging [zube]: To Test | ### Issue description:
rancher-logging 101.0.0+up3.17.7
If you enable kubeAudit logging in the rancher-logging app, the rancher-logging-kube-audit-fluentd DaemonSet pods should only run on CP nodes.
### Business impact:
Why the ds would run on nodes without the audit logs? It consumes resources that are not needed on worker nodes.
### Repro steps:
* on cluster with a mix of master nodes and worker only nodes install rancher-logging 101.0.0+up3.17.7
* check which nodes the rancher-logging-kube-audit-fluentd pods run on
### Actual behavior:
kube-audit pods runs on all node roles
### Expected behavior:
kube-audit pods run only on CP nodes | 1.0 | rancher-logging kube-audit pods should only run on CP nodes - ### Issue description:
rancher-logging 101.0.0+up3.17.7
If you enable kubeAudit logging in the rancher-logging app, the rancher-logging-kube-audit-fluentd DaemonSet pods should only run on CP nodes.
### Business impact:
Why the ds would run on nodes without the audit logs? It consumes resources that are not needed on worker nodes.
### Repro steps:
* on cluster with a mix of master nodes and worker only nodes install rancher-logging 101.0.0+up3.17.7
* check which nodes the rancher-logging-kube-audit-fluentd pods run on
### Actual behavior:
kube-audit pods runs on all node roles
### Expected behavior:
kube-audit pods run only on CP nodes | non_usab | rancher logging kube audit pods should only run on cp nodes issue description rancher logging if you enable kubeaudit logging in the rancher logging app the rancher logging kube audit fluentd daemonset pods should only run on cp nodes business impact why the ds would run on nodes without the audit logs it consumes resources that are not needed on worker nodes repro steps on cluster with a mix of master nodes and worker only nodes install rancher logging check which nodes the rancher logging kube audit fluentd pods run on actual behavior kube audit pods runs on all node roles expected behavior kube audit pods run only on cp nodes | 0 |
16,531 | 11,030,495,698 | IssuesEvent | 2019-12-06 15:51:39 | opentx/opentx | https://api.github.com/repos/opentx/opentx | closed | Companion crashes when loading bitmap files in telemetry LUA scripts | Bug OpSys-Linux Simulator Usability | Hi,
I am testing a telemetry script for X9D+ that I found on the web. The script visualize the RSSI through a widget.
The script tuns just fine on the radio, but it hangs companion when I run it with the Telemetry Simulator enabled. The same script without calling the lcd.drawPixmap function runs perfectly. On the other hand, when I call the lcd.drawPixmap the script runs for few seconds, then companion crashes. Some times the script runs a little-bit longer, but it ends crashing anyway.
I am using companion on Debian GNU/Linux, 64 bit platform.
I am attaching here:
o) the model setup
[setups_scripts.zip](https://github.com/opentx/opentx/files/2311688/setups_scripts.zip)
o) the script (with the .bmp images)
[telemetry_script.zip](https://github.com/opentx/opentx/files/2311686/telemetry_script.zip)
o) the standard error output
[companion_output.zip](https://github.com/opentx/opentx/files/2311702/companion_output.zip)
I am available to provide further informations
Thanks
Marco
| True | Companion crashes when loading bitmap files in telemetry LUA scripts - Hi,
I am testing a telemetry script for X9D+ that I found on the web. The script visualize the RSSI through a widget.
The script tuns just fine on the radio, but it hangs companion when I run it with the Telemetry Simulator enabled. The same script without calling the lcd.drawPixmap function runs perfectly. On the other hand, when I call the lcd.drawPixmap the script runs for few seconds, then companion crashes. Some times the script runs a little-bit longer, but it ends crashing anyway.
I am using companion on Debian GNU/Linux, 64 bit platform.
I am attaching here:
o) the model setup
[setups_scripts.zip](https://github.com/opentx/opentx/files/2311688/setups_scripts.zip)
o) the script (with the .bmp images)
[telemetry_script.zip](https://github.com/opentx/opentx/files/2311686/telemetry_script.zip)
o) the standard error output
[companion_output.zip](https://github.com/opentx/opentx/files/2311702/companion_output.zip)
I am available to provide further informations
Thanks
Marco
| usab | companion crashes when loading bitmap files in telemetry lua scripts hi i am testing a telemetry script for that i found on the web the script visualize the rssi through a widget the script tuns just fine on the radio but it hangs companion when i run it with the telemetry simulator enabled the same script without calling the lcd drawpixmap function runs perfectly on the other hand when i call the lcd drawpixmap the script runs for few seconds then companion crashes some times the script runs a little bit longer but it ends crashing anyway i am using companion on debian gnu linux bit platform i am attaching here o the model setup o the script with the bmp images o the standard error output i am available to provide further informations thanks marco | 1 |
25,721 | 25,771,095,819 | IssuesEvent | 2022-12-09 08:02:38 | mageshravi/lyricfront-mobile-app | https://api.github.com/repos/mageshravi/lyricfront-mobile-app | closed | Sets can be removed from bookmarks in offline mode | bug usability | In offline mode, user can click on "Bookmark" icon to un-save a worship-set. Gives a false impression that set has been removed from the bookmarks.
Hide the bookmark icon in offline mode.
| True | Sets can be removed from bookmarks in offline mode - In offline mode, user can click on "Bookmark" icon to un-save a worship-set. Gives a false impression that set has been removed from the bookmarks.
Hide the bookmark icon in offline mode.
| usab | sets can be removed from bookmarks in offline mode in offline mode user can click on bookmark icon to un save a worship set gives a false impression that set has been removed from the bookmarks hide the bookmark icon in offline mode | 1 |
17,809 | 12,344,850,870 | IssuesEvent | 2020-05-15 07:51:35 | gluster/glusterfs | https://api.github.com/repos/gluster/glusterfs | closed | RFE: Tool to fetch Faulty logs from all Geo-rep Master nodes | CB: geo-replication FA: Usability & Supportability wontfix | **Ansible based tool to report the Faulty logs**
- Run Volume info and collect the list of master nodes and brick paths
- To each of the nodes collected from the previous step, Copy a script
which parses the georep log files from the last(If session is
Faulty), and collects the output/traceback lines from logs
- Post process and format the collected output and display the report
- Output should clearly say if a worker is passive/active and the
current status. Order of worker should be same as Volinfo
Example output(Output format can be tunned based on usage and feedback),
```
Session: gv1 => root@192.168.122.171::gv2
Master Node: 192.168.122.171 Status: Active Brick Path: /exports/bricks/gv1/brick1/brick
-------------------------------------------------------------------------------------------
Master Node: 192.168.122.172 Status: Faulty Brick Path: /exports/bricks/gv1/brick2/brick
[2017-01-07 08:30:34.106961] E [syncdutils(/exports/bricks/gv1/brick2/brick):296:log_raise_exception] <top>: FAIL:
Traceback (most recent call last):
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 204, in main
main_i()
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 757, in main_i
local.service_loop(*[r for r in [remote] if r])
File "/usr/libexec/glusterfs/python/syncdaemon/resource.py", line 1555, in service_loop
g2.crawlwrap()
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 571, in crawlwrap
self.crawl()
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1134, in crawl
self.changelogs_batch_process(changes)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1109, in changelogs_batch_process
self.process(batch)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 992, in process
self.process_change(change, done, retry)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 933, in process_change
failures = self.slave.server.entry_ops(entries)
File "/usr/libexec/glusterfs/python/syncdaemon/repce.py", line 226, in __call__
return self.ins(self.meth, *a)
File "/usr/libexec/glusterfs/python/syncdaemon/repce.py", line 208, in __call__
raise res
OSError: [Errno 5] Input/output error
[2017-01-07 08:30:34.218364] I [syncdutils(/exports/bricks/gv1/brick2/brick):237:finalize] <top>: exiting.
-------------------------------------------------------------------------------------------
Master Node: 192.168.122.173 Status: Passive Brick Path: /exports/bricks/gv1/brick3/brick
-------------------------------------------------------------------------------------------
```
| True | RFE: Tool to fetch Faulty logs from all Geo-rep Master nodes - **Ansible based tool to report the Faulty logs**
- Run Volume info and collect the list of master nodes and brick paths
- To each of the nodes collected from the previous step, Copy a script
which parses the georep log files from the last(If session is
Faulty), and collects the output/traceback lines from logs
- Post process and format the collected output and display the report
- Output should clearly say if a worker is passive/active and the
current status. Order of worker should be same as Volinfo
Example output(Output format can be tunned based on usage and feedback),
```
Session: gv1 => root@192.168.122.171::gv2
Master Node: 192.168.122.171 Status: Active Brick Path: /exports/bricks/gv1/brick1/brick
-------------------------------------------------------------------------------------------
Master Node: 192.168.122.172 Status: Faulty Brick Path: /exports/bricks/gv1/brick2/brick
[2017-01-07 08:30:34.106961] E [syncdutils(/exports/bricks/gv1/brick2/brick):296:log_raise_exception] <top>: FAIL:
Traceback (most recent call last):
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 204, in main
main_i()
File "/usr/libexec/glusterfs/python/syncdaemon/gsyncd.py", line 757, in main_i
local.service_loop(*[r for r in [remote] if r])
File "/usr/libexec/glusterfs/python/syncdaemon/resource.py", line 1555, in service_loop
g2.crawlwrap()
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 571, in crawlwrap
self.crawl()
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1134, in crawl
self.changelogs_batch_process(changes)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 1109, in changelogs_batch_process
self.process(batch)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 992, in process
self.process_change(change, done, retry)
File "/usr/libexec/glusterfs/python/syncdaemon/master.py", line 933, in process_change
failures = self.slave.server.entry_ops(entries)
File "/usr/libexec/glusterfs/python/syncdaemon/repce.py", line 226, in __call__
return self.ins(self.meth, *a)
File "/usr/libexec/glusterfs/python/syncdaemon/repce.py", line 208, in __call__
raise res
OSError: [Errno 5] Input/output error
[2017-01-07 08:30:34.218364] I [syncdutils(/exports/bricks/gv1/brick2/brick):237:finalize] <top>: exiting.
-------------------------------------------------------------------------------------------
Master Node: 192.168.122.173 Status: Passive Brick Path: /exports/bricks/gv1/brick3/brick
-------------------------------------------------------------------------------------------
```
| usab | rfe tool to fetch faulty logs from all geo rep master nodes ansible based tool to report the faulty logs run volume info and collect the list of master nodes and brick paths to each of the nodes collected from the previous step copy a script which parses the georep log files from the last if session is faulty and collects the output traceback lines from logs post process and format the collected output and display the report output should clearly say if a worker is passive active and the current status order of worker should be same as volinfo example output output format can be tunned based on usage and feedback session root master node status active brick path exports bricks brick master node status faulty brick path exports bricks brick e fail traceback most recent call last file usr libexec glusterfs python syncdaemon gsyncd py line in main main i file usr libexec glusterfs python syncdaemon gsyncd py line in main i local service loop if r file usr libexec glusterfs python syncdaemon resource py line in service loop crawlwrap file usr libexec glusterfs python syncdaemon master py line in crawlwrap self crawl file usr libexec glusterfs python syncdaemon master py line in crawl self changelogs batch process changes file usr libexec glusterfs python syncdaemon master py line in changelogs batch process self process batch file usr libexec glusterfs python syncdaemon master py line in process self process change change done retry file usr libexec glusterfs python syncdaemon master py line in process change failures self slave server entry ops entries file usr libexec glusterfs python syncdaemon repce py line in call return self ins self meth a file usr libexec glusterfs python syncdaemon repce py line in call raise res oserror input output error i exiting master node status passive brick path exports bricks brick | 1 |
414,378 | 27,985,458,978 | IssuesEvent | 2023-03-26 16:42:27 | Sars9588/mywebclass-simulation | https://api.github.com/repos/Sars9588/mywebclass-simulation | closed | Privacy Policy -- Section mentioning how we use the information that is shared with us | documentation | Section that mentions the use of the information that is presented to us
| 1.0 | Privacy Policy -- Section mentioning how we use the information that is shared with us - Section that mentions the use of the information that is presented to us
| non_usab | privacy policy section mentioning how we use the information that is shared with us section that mentions the use of the information that is presented to us | 0 |
306,771 | 9,403,857,912 | IssuesEvent | 2019-04-09 03:19:21 | ClangBuiltLinux/linux | https://api.github.com/repos/ClangBuiltLinux/linux | closed | -Wparentheses-equality in arch/arm/lib/uaccess_with_memcpy.c | -Wparentheses-equality [ARCH] arm32 [BUG] linux good first issue low priority | from linaro tcwg ci:
```
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:93:45: warning: equality comparison with extraneous parentheses [-Wparentheses-equality]
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:93:45: note: remove extraneous parentheses around the comparison to silence this warning
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ~ ^ ~
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:93:45: note: use '=' to turn this equality comparison into an assignment
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ^~
00:00:35 =
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:166:45: warning: equality comparison with extraneous parentheses [-Wparentheses-equality]
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:166:45: note: remove extraneous parentheses around the comparison to silence this warning
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ~ ^ ~
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:166:45: note: use '=' to turn this equality comparison into an assignment
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ^~
00:00:35 =
00:00:35 2 warnings generated.
``` | 1.0 | -Wparentheses-equality in arch/arm/lib/uaccess_with_memcpy.c - from linaro tcwg ci:
```
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:93:45: warning: equality comparison with extraneous parentheses [-Wparentheses-equality]
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:93:45: note: remove extraneous parentheses around the comparison to silence this warning
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ~ ^ ~
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:93:45: note: use '=' to turn this equality comparison into an assignment
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ^~
00:00:35 =
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:166:45: warning: equality comparison with extraneous parentheses [-Wparentheses-equality]
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:166:45: note: remove extraneous parentheses around the comparison to silence this warning
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ~ ^ ~
00:00:35 arch/arm/lib/uaccess_with_memcpy.c:166:45: note: use '=' to turn this equality comparison into an assignment
00:00:35 if ((((current_thread_info()->addr_limit)) == (0x00000000))) {
00:00:35 ^~
00:00:35 =
00:00:35 2 warnings generated.
``` | non_usab | wparentheses equality in arch arm lib uaccess with memcpy c from linaro tcwg ci arch arm lib uaccess with memcpy c warning equality comparison with extraneous parentheses if current thread info addr limit arch arm lib uaccess with memcpy c note remove extraneous parentheses around the comparison to silence this warning if current thread info addr limit arch arm lib uaccess with memcpy c note use to turn this equality comparison into an assignment if current thread info addr limit arch arm lib uaccess with memcpy c warning equality comparison with extraneous parentheses if current thread info addr limit arch arm lib uaccess with memcpy c note remove extraneous parentheses around the comparison to silence this warning if current thread info addr limit arch arm lib uaccess with memcpy c note use to turn this equality comparison into an assignment if current thread info addr limit warnings generated | 0 |
29,378 | 23,960,913,034 | IssuesEvent | 2022-09-12 19:01:31 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | [SOCKS / Named] Duplicate DNS records prevent named from starting | operations devops needs-grooming infrastructure | ## Description
SOCKS access can be affected when there are problems with internal DNS, resulting in an error like:
`open failed: administratively prohibited
`
## Background/context
- Automation exists to create DNS records needed for review instances (RIs)
- Duplicate DNS records are sometimes created in the zone: vetsgov-internal
- These duplicates prevent named from starting, due to config errors
## Technical notes
_Notes around work that is happening, if applicable (optional, please delete if unused)_
---
## Tasks
- Investigate potential fixes
- Implement fixes
## Acceptance Criteria
- [x] _What are the specific things that must be true in order for this to be accepted?_
| 1.0 | [SOCKS / Named] Duplicate DNS records prevent named from starting - ## Description
SOCKS access can be affected when there are problems with internal DNS, resulting in an error like:
`open failed: administratively prohibited
`
## Background/context
- Automation exists to create DNS records needed for review instances (RIs)
- Duplicate DNS records are sometimes created in the zone: vetsgov-internal
- These duplicates prevent named from starting, due to config errors
## Technical notes
_Notes around work that is happening, if applicable (optional, please delete if unused)_
---
## Tasks
- Investigate potential fixes
- Implement fixes
## Acceptance Criteria
- [x] _What are the specific things that must be true in order for this to be accepted?_
| non_usab | duplicate dns records prevent named from starting description socks access can be affected when there are problems with internal dns resulting in an error like open failed administratively prohibited background context automation exists to create dns records needed for review instances ris duplicate dns records are sometimes created in the zone vetsgov internal these duplicates prevent named from starting due to config errors technical notes notes around work that is happening if applicable optional please delete if unused tasks investigate potential fixes implement fixes acceptance criteria what are the specific things that must be true in order for this to be accepted | 0 |
337,401 | 10,217,379,445 | IssuesEvent | 2019-08-15 13:31:40 | OkunaOrg/okuna-app | https://api.github.com/repos/OkunaOrg/okuna-app | closed | Ability to translate post/comment text in the app | feature priority:high | We should show a translate this text button next to posts and comments to users.
- The user should be able to set an app language through preferences, and the translations will be done to this app language. Store this on user model.
- Detect comment/post language and store it on respective model,(update this on edit flow)
- Detect if we can translate from post/comment lang to user app lang using amazon APIs. And if yes, show the translate button. Otherwise do not show it.
- Clicking translate should use Amazon APIs to fetch and show the translated result.
- Do not show translate button if post/comment is already in user preferred app language.
| 1.0 | Ability to translate post/comment text in the app - We should show a translate this text button next to posts and comments to users.
- The user should be able to set an app language through preferences, and the translations will be done to this app language. Store this on user model.
- Detect comment/post language and store it on respective model,(update this on edit flow)
- Detect if we can translate from post/comment lang to user app lang using amazon APIs. And if yes, show the translate button. Otherwise do not show it.
- Clicking translate should use Amazon APIs to fetch and show the translated result.
- Do not show translate button if post/comment is already in user preferred app language.
| non_usab | ability to translate post comment text in the app we should show a translate this text button next to posts and comments to users the user should be able to set an app language through preferences and the translations will be done to this app language store this on user model detect comment post language and store it on respective model update this on edit flow detect if we can translate from post comment lang to user app lang using amazon apis and if yes show the translate button otherwise do not show it clicking translate should use amazon apis to fetch and show the translated result do not show translate button if post comment is already in user preferred app language | 0 |
65,820 | 3,244,263,974 | IssuesEvent | 2015-10-16 00:35:31 | parallaxinc/Parallax-IDE | https://api.github.com/repos/parallaxinc/Parallax-IDE | closed | Debug Terminal: Echo Off control | Priority 3 UX/UI | A UI feature needs to be added to allow the user to set or clear the Echo Off feature. This is preferably accessible from the main screen (near the Debug Terminal, in the vicinity of the Tx/Rx indicators).
This relates to Issue #184. | 1.0 | Debug Terminal: Echo Off control - A UI feature needs to be added to allow the user to set or clear the Echo Off feature. This is preferably accessible from the main screen (near the Debug Terminal, in the vicinity of the Tx/Rx indicators).
This relates to Issue #184. | non_usab | debug terminal echo off control a ui feature needs to be added to allow the user to set or clear the echo off feature this is preferably accessible from the main screen near the debug terminal in the vicinity of the tx rx indicators this relates to issue | 0 |
27,511 | 29,496,548,420 | IssuesEvent | 2023-06-02 17:29:29 | JuliaReach/LazySets.jl | https://api.github.com/repos/JuliaReach/LazySets.jl | closed | Fallback tovrep implementation | feature :heavy_plus_sign: usability :computer_mouse: simplification :baby: | I think that it would be useful to add `tovrep(::AbstractPolytope)` and `tovrep(::AbstractPolyhedron, check_boundedness::Bool=true)`. This allows to handle efficiently the cases when the v-rep is needed. Ref: #1093. | True | Fallback tovrep implementation - I think that it would be useful to add `tovrep(::AbstractPolytope)` and `tovrep(::AbstractPolyhedron, check_boundedness::Bool=true)`. This allows to handle efficiently the cases when the v-rep is needed. Ref: #1093. | usab | fallback tovrep implementation i think that it would be useful to add tovrep abstractpolytope and tovrep abstractpolyhedron check boundedness bool true this allows to handle efficiently the cases when the v rep is needed ref | 1 |
17,733 | 12,297,228,349 | IssuesEvent | 2020-05-11 08:28:45 | raiden-network/raiden | https://api.github.com/repos/raiden-network/raiden | closed | Add API Endpoint to shutdown raiden | Component / API Effort / 2 Flag / Usability Type / Enhancement | ## Abstract
If Raiden is started other than the CLI it might be difficult to shutdown Raiden easily (especially for non-technical users). This could happen if Raiden is started by the wizard.
## Motivation
Non-technical users are controling raiden by some UI (e.g. the WebUI). The user should be able to shut raiden down via an exposed endpoint in order to have a controlled shutdown.
## Specification
- [ ] Provide an RestAPI Endpoint to shut raiden down gracefully
- ~~Take security measures into account that not everybody can shutdown the node~~ | True | Add API Endpoint to shutdown raiden - ## Abstract
If Raiden is started other than the CLI it might be difficult to shutdown Raiden easily (especially for non-technical users). This could happen if Raiden is started by the wizard.
## Motivation
Non-technical users are controling raiden by some UI (e.g. the WebUI). The user should be able to shut raiden down via an exposed endpoint in order to have a controlled shutdown.
## Specification
- [ ] Provide an RestAPI Endpoint to shut raiden down gracefully
- ~~Take security measures into account that not everybody can shutdown the node~~ | usab | add api endpoint to shutdown raiden abstract if raiden is started other than the cli it might be difficult to shutdown raiden easily especially for non technical users this could happen if raiden is started by the wizard motivation non technical users are controling raiden by some ui e g the webui the user should be able to shut raiden down via an exposed endpoint in order to have a controlled shutdown specification provide an restapi endpoint to shut raiden down gracefully take security measures into account that not everybody can shutdown the node | 1 |
385,646 | 11,423,956,163 | IssuesEvent | 2020-02-03 16:48:25 | radical-cybertools/radical.pilot | https://api.github.com/repos/radical-cybertools/radical.pilot | closed | Failure to write radical.pilot.sandbox is not fatal | comp:pmgr:launcher priority:low topic:resource type:bug | Resource: Summit
Stack:
```
$ radical-stack
/autofs/nccs-svm1_home1/mturilli1/experiments/ve/jsrun_prte/lib/python2.7/site-packages/radical/utils
/autofs/nccs-svm1_home1/mturilli1/experiments/ve/jsrun_prte/lib/python2.7/site-packages/radical/saga
/autofs/nccs-svm1_home1/mturilli1/experiments/ve/jsrun_prte/lib/python2.7/site-packages/radical/pilot
python : 2.7.15
pythonpath : /autofs/nccs-svm1_sw/summit/.swci/0-core/opt/spack/20180914/linux-rhel7-ppc64le/gcc-4.8.5/py-setuptools-40.4.3-rc56sxgpafwvs5eyrvc3uxiaqoc6oe2f/lib/python2.7/site-packages:/autofs/nccs-svm1_sw/summit/.swci/0-core/opt/spack/20180914/linux-rhel7-ppc64le/gcc-4.8.5/py-pip-10.0.1-2gr5x7tsnuxwissqhzapdbmlpheove3i/lib/python2.7/site-packages:/autofs/nccs-svm1_sw/summit/.swci/0-core/opt/spack/20180914/linux-rhel7-ppc64le/gcc-4.8.5/py-virtualenv-16.0.0-phcok3x4eyd36qfh5ptv66isyol4ui4b/lib/python2.7/site-packages:/sw/summit/xalt/1.1.3/site:/sw/summit/xalt/1.1.3/libexec
virtualenv : /autofs/nccs-svm1_home1/mturilli1/experiments/ve/jsrun_prte
radical.pilot : 0.62.0-v0.62.0-346-g8b122b6@experiment-periscope
radical.saga : 0.62.0-v0.62.0-26-gea22176@devel
radical.utils : 0.60.2-v0.60.2-5-g87e3d67@devel
```
Log:
```
2019-07-01 15:12:32,605: radical.saga.pty : MainProcess : MainThread : DEBUG : write: [ 16] [ 362] (cd /gpfs/alpine/scratch/mturilli1/bip178/radical.pilot.sandbox/rp.session.login1.mturilli1.018078.0000/pilot.0000/ && test -d '/gpfs/alpine/scratch/mturilli1/bip178/radical.pilot.sandbox/rp.session.login1.mturilli1.018078.0000/pilot.0000/' && cd '/gpfs/alpine/scratch/mturilli1/bip178/radical.pilot.sandbox/rp.session.login1.mturilli1.018078.0000/pilot.0000/'\n)
2019-07-01 15:12:32,606: radical.saga.pty : MainProcess : MainThread : DEBUG : read : [ 16] [ 148] (sh: cd: /gpfs/alpine/scratch/mturilli1/bip178/radical.pilot.sandbox/rp.session.login1.mturilli1.018078.0000/pilot.0000/: No such file or directory\n)
2019-07-01 15:12:32,606: radical.saga.pty : MainProcess : MainThread : DEBUG : read : [ 16] [ 10] (PROMPT-1->)
2019-07-01 15:12:32,606: radical.saga.cpi : MainProcess : MainThread : DEBUG : PTYShell del <radical.saga.utils.pty_shell.PTYShell object at 0x7fff8ef39390>
2019-07-01 15:12:32,707: radical.saga.pty : MainProcess : MainThread : DEBUG : PTYProcess del <radical.saga.utils.pty_process.PTYProcess object at 0x7fff8ef39610>
```
This results in the script to indefinitely hang in a seemingly alive status. | 1.0 | Failure to write radical.pilot.sandbox is not fatal - Resource: Summit
Stack:
```
$ radical-stack
/autofs/nccs-svm1_home1/mturilli1/experiments/ve/jsrun_prte/lib/python2.7/site-packages/radical/utils
/autofs/nccs-svm1_home1/mturilli1/experiments/ve/jsrun_prte/lib/python2.7/site-packages/radical/saga
/autofs/nccs-svm1_home1/mturilli1/experiments/ve/jsrun_prte/lib/python2.7/site-packages/radical/pilot
python : 2.7.15
pythonpath : /autofs/nccs-svm1_sw/summit/.swci/0-core/opt/spack/20180914/linux-rhel7-ppc64le/gcc-4.8.5/py-setuptools-40.4.3-rc56sxgpafwvs5eyrvc3uxiaqoc6oe2f/lib/python2.7/site-packages:/autofs/nccs-svm1_sw/summit/.swci/0-core/opt/spack/20180914/linux-rhel7-ppc64le/gcc-4.8.5/py-pip-10.0.1-2gr5x7tsnuxwissqhzapdbmlpheove3i/lib/python2.7/site-packages:/autofs/nccs-svm1_sw/summit/.swci/0-core/opt/spack/20180914/linux-rhel7-ppc64le/gcc-4.8.5/py-virtualenv-16.0.0-phcok3x4eyd36qfh5ptv66isyol4ui4b/lib/python2.7/site-packages:/sw/summit/xalt/1.1.3/site:/sw/summit/xalt/1.1.3/libexec
virtualenv : /autofs/nccs-svm1_home1/mturilli1/experiments/ve/jsrun_prte
radical.pilot : 0.62.0-v0.62.0-346-g8b122b6@experiment-periscope
radical.saga : 0.62.0-v0.62.0-26-gea22176@devel
radical.utils : 0.60.2-v0.60.2-5-g87e3d67@devel
```
Log:
```
2019-07-01 15:12:32,605: radical.saga.pty : MainProcess : MainThread : DEBUG : write: [ 16] [ 362] (cd /gpfs/alpine/scratch/mturilli1/bip178/radical.pilot.sandbox/rp.session.login1.mturilli1.018078.0000/pilot.0000/ && test -d '/gpfs/alpine/scratch/mturilli1/bip178/radical.pilot.sandbox/rp.session.login1.mturilli1.018078.0000/pilot.0000/' && cd '/gpfs/alpine/scratch/mturilli1/bip178/radical.pilot.sandbox/rp.session.login1.mturilli1.018078.0000/pilot.0000/'\n)
2019-07-01 15:12:32,606: radical.saga.pty : MainProcess : MainThread : DEBUG : read : [ 16] [ 148] (sh: cd: /gpfs/alpine/scratch/mturilli1/bip178/radical.pilot.sandbox/rp.session.login1.mturilli1.018078.0000/pilot.0000/: No such file or directory\n)
2019-07-01 15:12:32,606: radical.saga.pty : MainProcess : MainThread : DEBUG : read : [ 16] [ 10] (PROMPT-1->)
2019-07-01 15:12:32,606: radical.saga.cpi : MainProcess : MainThread : DEBUG : PTYShell del <radical.saga.utils.pty_shell.PTYShell object at 0x7fff8ef39390>
2019-07-01 15:12:32,707: radical.saga.pty : MainProcess : MainThread : DEBUG : PTYProcess del <radical.saga.utils.pty_process.PTYProcess object at 0x7fff8ef39610>
```
This results in the script to indefinitely hang in a seemingly alive status. | non_usab | failure to write radical pilot sandbox is not fatal resource summit stack radical stack autofs nccs experiments ve jsrun prte lib site packages radical utils autofs nccs experiments ve jsrun prte lib site packages radical saga autofs nccs experiments ve jsrun prte lib site packages radical pilot python pythonpath autofs nccs sw summit swci core opt spack linux gcc py setuptools lib site packages autofs nccs sw summit swci core opt spack linux gcc py pip lib site packages autofs nccs sw summit swci core opt spack linux gcc py virtualenv lib site packages sw summit xalt site sw summit xalt libexec virtualenv autofs nccs experiments ve jsrun prte radical pilot experiment periscope radical saga devel radical utils devel log radical saga pty mainprocess mainthread debug write cd gpfs alpine scratch radical pilot sandbox rp session pilot test d gpfs alpine scratch radical pilot sandbox rp session pilot cd gpfs alpine scratch radical pilot sandbox rp session pilot n radical saga pty mainprocess mainthread debug read sh cd gpfs alpine scratch radical pilot sandbox rp session pilot no such file or directory n radical saga pty mainprocess mainthread debug read prompt radical saga cpi mainprocess mainthread debug ptyshell del radical saga pty mainprocess mainthread debug ptyprocess del this results in the script to indefinitely hang in a seemingly alive status | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.