Unnamed: 0 int64 3 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 7 112 | repo_url stringlengths 36 141 | action stringclasses 3 values | title stringlengths 2 742 | labels stringlengths 4 431 | body stringlengths 5 239k | index stringclasses 10 values | text_combine stringlengths 96 240k | label stringclasses 2 values | text stringlengths 96 200k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
32,270 | 6,756,696,827 | IssuesEvent | 2017-10-24 08:10:12 | primefaces/primeng | https://api.github.com/repos/primefaces/primeng | closed | MegaMenu doesn't compile with TypeScript 2.4 | confirmed defect | **I'm submitting a ...**
```
[X] bug report
```
**Test case**
You can use the following demo app as test case:
https://github.com/ova2/angular-development-with-primeng/tree/master/chapter7/megamenu
**Current behavior**
If you run the showcase for the MegaMenu with TypeScript 2.4 or run the demo app linked above, you will get a compilation error like
```
Type '{ label: string; items: { label: string; }[]; }[]' has no properties in common with type 'MenuItem'.
```
**Expected behavior**
There should be no compilation error, as when compiling with TypeScript 2.3.
**Minimal reproduction of the problem with instructions**
* Install the above app
* or install the current master of PrimeNG and change the requirements in package.json to Angular 4.3, Angular-Cli 1.3, TypeScript 2.4 (lower Angular/Cli versions require TypeScript 2.3, so you need to test with Angular 4.3 and Cli 1.3)
* Run `npm` install and `npm start` and check the MegaMenu
* **Angular version:** 4.3.3
* **PrimeNG version:** 4.1.2
* **Browser:** all
* **Language:** TypeScript 2.4
* **Node (for AoT issues):** 8.1.4
* **Analysis of the problem:**
In the `MenuItem` interface, the `items` property is defined as of type `MenuItem[]`. But in the MegaMenu, you can have arrays of arrays of MenuItems as items, not just arrays of MenuItems.
In TypeScript 2.4, it’s now an error to assign anything to a weak type when there’s no overlap in properties (see [here](https://www.typescriptlang.org/docs/handbook/release-notes/typescript-2-4.html#weak-type-detection)).
Note that in the `MenuItem` interface all properties are marked as optional. Therefore, this is considered a "weak type".
* **Proposed solution:**
The `items` property in the `MenutItem` interface should be defined as follows:
```
items?: MenuItem[]|MenuItem[][];
```
| 1.0 | MegaMenu doesn't compile with TypeScript 2.4 - **I'm submitting a ...**
```
[X] bug report
```
**Test case**
You can use the following demo app as test case:
https://github.com/ova2/angular-development-with-primeng/tree/master/chapter7/megamenu
**Current behavior**
If you run the showcase for the MegaMenu with TypeScript 2.4 or run the demo app linked above, you will get a compilation error like
```
Type '{ label: string; items: { label: string; }[]; }[]' has no properties in common with type 'MenuItem'.
```
**Expected behavior**
There should be no compilation error, as when compiling with TypeScript 2.3.
**Minimal reproduction of the problem with instructions**
* Install the above app
* or install the current master of PrimeNG and change the requirements in package.json to Angular 4.3, Angular-Cli 1.3, TypeScript 2.4 (lower Angular/Cli versions require TypeScript 2.3, so you need to test with Angular 4.3 and Cli 1.3)
* Run `npm` install and `npm start` and check the MegaMenu
* **Angular version:** 4.3.3
* **PrimeNG version:** 4.1.2
* **Browser:** all
* **Language:** TypeScript 2.4
* **Node (for AoT issues):** 8.1.4
* **Analysis of the problem:**
In the `MenuItem` interface, the `items` property is defined as of type `MenuItem[]`. But in the MegaMenu, you can have arrays of arrays of MenuItems as items, not just arrays of MenuItems.
In TypeScript 2.4, it’s now an error to assign anything to a weak type when there’s no overlap in properties (see [here](https://www.typescriptlang.org/docs/handbook/release-notes/typescript-2-4.html#weak-type-detection)).
Note that in the `MenuItem` interface all properties are marked as optional. Therefore, this is considered a "weak type".
* **Proposed solution:**
The `items` property in the `MenutItem` interface should be defined as follows:
```
items?: MenuItem[]|MenuItem[][];
```
| non_usab | megamenu doesn t compile with typescript i m submitting a bug report test case you can use the following demo app as test case current behavior if you run the showcase for the megamenu with typescript or run the demo app linked above you will get a compilation error like type label string items label string has no properties in common with type menuitem expected behavior there should be no compilation error as when compiling with typescript minimal reproduction of the problem with instructions install the above app or install the current master of primeng and change the requirements in package json to angular angular cli typescript lower angular cli versions require typescript so you need to test with angular and cli run npm install and npm start and check the megamenu angular version primeng version browser all language typescript node for aot issues analysis of the problem in the menuitem interface the items property is defined as of type menuitem but in the megamenu you can have arrays of arrays of menuitems as items not just arrays of menuitems in typescript it’s now an error to assign anything to a weak type when there’s no overlap in properties see note that in the menuitem interface all properties are marked as optional therefore this is considered a weak type proposed solution the items property in the menutitem interface should be defined as follows items menuitem menuitem | 0 |
8,566 | 5,825,824,655 | IssuesEvent | 2017-05-08 00:51:39 | bronzehedwick/chrisdeluca | https://api.github.com/repos/bronzehedwick/chrisdeluca | closed | Add more links to navigation | usability | The main navigation should have the following links:
* Now
* Contact
* Sections (this is categories. Maybe rename?)
* RSS
In addition, there should be a form to subscribe to email notifications on site updates.
Depends on #61 #60 #59 #57 | True | Add more links to navigation - The main navigation should have the following links:
* Now
* Contact
* Sections (this is categories. Maybe rename?)
* RSS
In addition, there should be a form to subscribe to email notifications on site updates.
Depends on #61 #60 #59 #57 | usab | add more links to navigation the main navigation should have the following links now contact sections this is categories maybe rename rss in addition there should be a form to subscribe to email notifications on site updates depends on | 1 |
325,413 | 27,876,174,470 | IssuesEvent | 2023-03-21 16:09:19 | airbytehq/airbyte | https://api.github.com/repos/airbytehq/airbyte | closed | E2E stream table tests - stream details panel | team/platform-move area/frontend ui/connection ui/tests e2e-testing-tool | ## Tell us about the problem you're trying to solve
Add E2E tests for stream details panel functionality in the new stream table
### Tasks
- [x] Check panel opening (this case should be covered in https://github.com/airbytehq/airbyte/issues/22640)
- [x] Check stream details info: sync state, namespace(if exist), stream name, chosen sync mode
- [x] Check amount of displayed fields(depends on tests data we will use)
- [x] Check that user is able to scroll a long list of fields()scrolling and desired field is visible
- [x] Check that each field has info: source field name, data type, Cursor(optional), PK(optional), destination field name
- [x] Check that user is able to select only one cursor value (appropriate sync mode should be chosen)
- [x] Check that user is NOT able to select only one cursor value if it's source-defined (appropriate sync mode should be chosen)
- [x] Check that user is able to select only multiple PK values (appropriate sync mode should be chosen)
- [x] Check that user is NOT able to select PK value if it's source-defined (appropriate sync mode should be chosen)
| 2.0 | E2E stream table tests - stream details panel - ## Tell us about the problem you're trying to solve
Add E2E tests for stream details panel functionality in the new stream table
### Tasks
- [x] Check panel opening (this case should be covered in https://github.com/airbytehq/airbyte/issues/22640)
- [x] Check stream details info: sync state, namespace(if exist), stream name, chosen sync mode
- [x] Check amount of displayed fields(depends on tests data we will use)
- [x] Check that user is able to scroll a long list of fields()scrolling and desired field is visible
- [x] Check that each field has info: source field name, data type, Cursor(optional), PK(optional), destination field name
- [x] Check that user is able to select only one cursor value (appropriate sync mode should be chosen)
- [x] Check that user is NOT able to select only one cursor value if it's source-defined (appropriate sync mode should be chosen)
- [x] Check that user is able to select only multiple PK values (appropriate sync mode should be chosen)
- [x] Check that user is NOT able to select PK value if it's source-defined (appropriate sync mode should be chosen)
| non_usab | stream table tests stream details panel tell us about the problem you re trying to solve add tests for stream details panel functionality in the new stream table tasks check panel opening this case should be covered in check stream details info sync state namespace if exist stream name chosen sync mode check amount of displayed fields depends on tests data we will use check that user is able to scroll a long list of fields scrolling and desired field is visible check that each field has info source field name data type cursor optional pk optional destination field name check that user is able to select only one cursor value appropriate sync mode should be chosen check that user is not able to select only one cursor value if it s source defined appropriate sync mode should be chosen check that user is able to select only multiple pk values appropriate sync mode should be chosen check that user is not able to select pk value if it s source defined appropriate sync mode should be chosen | 0 |
269,063 | 28,959,988,826 | IssuesEvent | 2023-05-10 01:06:12 | dpteam/RK3188_TABLET | https://api.github.com/repos/dpteam/RK3188_TABLET | reopened | CVE-2011-2699 (High) detected in linuxv3.0 | Mend: dependency security vulnerability | ## CVE-2011-2699 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/net/inetpeer.h</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The IPv6 implementation in the Linux kernel before 3.1 does not generate Fragment Identification values separately for each destination, which makes it easier for remote attackers to cause a denial of service (disrupted networking) by predicting these values and sending crafted packets.
<p>Publish Date: 2012-05-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-2699>CVE-2011-2699</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-2699">https://nvd.nist.gov/vuln/detail/CVE-2011-2699</a></p>
<p>Release Date: 2012-05-24</p>
<p>Fix Resolution: 3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2011-2699 (High) detected in linuxv3.0 - ## CVE-2011-2699 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary>
<p>
<p>Linux kernel source tree</p>
<p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/net/inetpeer.h</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
The IPv6 implementation in the Linux kernel before 3.1 does not generate Fragment Identification values separately for each destination, which makes it easier for remote attackers to cause a denial of service (disrupted networking) by predicting these values and sending crafted packets.
<p>Publish Date: 2012-05-24
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-2699>CVE-2011-2699</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-2699">https://nvd.nist.gov/vuln/detail/CVE-2011-2699</a></p>
<p>Release Date: 2012-05-24</p>
<p>Fix Resolution: 3.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_usab | cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files include net inetpeer h vulnerability details the implementation in the linux kernel before does not generate fragment identification values separately for each destination which makes it easier for remote attackers to cause a denial of service disrupted networking by predicting these values and sending crafted packets publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
117,536 | 17,496,281,368 | IssuesEvent | 2021-08-10 01:00:40 | billmcchesney1/vulnerable-rust | https://api.github.com/repos/billmcchesney1/vulnerable-rust | opened | CVE-2020-36471 (Medium) detected in generator-0.6.21.crate | security vulnerability | ## CVE-2020-36471 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>generator-0.6.21.crate</b></p></summary>
<p>Stackfull Generator Library in Rust</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/generator/0.6.21/download">https://crates.io/api/v1/crates/generator/0.6.21/download</a></p>
<p>
Dependency Hierarchy:
- hyper-0.13.5.crate (Root Library)
- tokio-0.2.21.crate
- bytes-0.5.5.crate
- loom-0.3.4.crate
- :x: **generator-0.6.21.crate** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the generator crate before 0.7.0 for Rust. It does not ensure that a function (for yielding values) has Send bounds.
<p>Publish Date: 2021-08-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36471>CVE-2020-36471</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2020-0151.html">https://rustsec.org/advisories/RUSTSEC-2020-0151.html</a></p>
<p>Release Date: 2021-08-08</p>
<p>Fix Resolution: 0.7.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Crate","packageName":"generator","packageVersion":"0.6.21","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"hyper:0.13.5;tokio:0.2.21;bytes:0.5.5;loom:0.3.4;generator:0.6.21","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.7.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36471","vulnerabilityDetails":"An issue was discovered in the generator crate before 0.7.0 for Rust. It does not ensure that a function (for yielding values) has Send bounds.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36471","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-36471 (Medium) detected in generator-0.6.21.crate - ## CVE-2020-36471 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>generator-0.6.21.crate</b></p></summary>
<p>Stackfull Generator Library in Rust</p>
<p>Library home page: <a href="https://crates.io/api/v1/crates/generator/0.6.21/download">https://crates.io/api/v1/crates/generator/0.6.21/download</a></p>
<p>
Dependency Hierarchy:
- hyper-0.13.5.crate (Root Library)
- tokio-0.2.21.crate
- bytes-0.5.5.crate
- loom-0.3.4.crate
- :x: **generator-0.6.21.crate** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the generator crate before 0.7.0 for Rust. It does not ensure that a function (for yielding values) has Send bounds.
<p>Publish Date: 2021-08-08
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36471>CVE-2020-36471</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: N/A
- Attack Complexity: N/A
- Privileges Required: N/A
- User Interaction: N/A
- Scope: N/A
- Impact Metrics:
- Confidentiality Impact: N/A
- Integrity Impact: N/A
- Availability Impact: N/A
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2020-0151.html">https://rustsec.org/advisories/RUSTSEC-2020-0151.html</a></p>
<p>Release Date: 2021-08-08</p>
<p>Fix Resolution: 0.7.0</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Crate","packageName":"generator","packageVersion":"0.6.21","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"hyper:0.13.5;tokio:0.2.21;bytes:0.5.5;loom:0.3.4;generator:0.6.21","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.7.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36471","vulnerabilityDetails":"An issue was discovered in the generator crate before 0.7.0 for Rust. It does not ensure that a function (for yielding values) has Send bounds.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36471","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> --> | non_usab | cve medium detected in generator crate cve medium severity vulnerability vulnerable library generator crate stackfull generator library in rust library home page a href dependency hierarchy hyper crate root library tokio crate bytes crate loom crate x generator crate vulnerable library found in base branch master vulnerability details an issue was discovered in the generator crate before for rust it does not ensure that a function for yielding values has send bounds publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree hyper tokio bytes loom generator isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in the generator crate before for rust it does not ensure that a function for yielding values has send bounds vulnerabilityurl | 0 |
371,169 | 25,939,851,565 | IssuesEvent | 2022-12-16 17:16:26 | projf/projf-explore | https://api.github.com/repos/projf/projf-explore | closed | Feedback on endianness | documentation | On page https://projectf.io/posts/numbers-in-verilog/ you mention:
> Say you’ve got a bit-endian ...
You probably mean big
> ... byte from I2C and want to convert it to little-endian. Alas, you can’t mix big and little-endian vectors, so the following won’t work:
I would pay attention to endianness in this context. In my experience endianness is mostly known in data consisting of multiple bytes. I think you are talking about the endianness of bits, which seem to be called *bit endianness* ([example](https://stackoverflow.com/questions/6043483/why-bit-endianness-is-an-issue-in-bitfields).
---
Another issue I stumbled upon:
> wire [0:7] i2c_byte; // 8-bit wire (big-endian)
> reg [7:0] le_byte; // 8-bit reg (little-endian)
> always_ff @(posedge clk) le_byte <= i2c_byte; // Won't work :(
I tried:
```verilog
`define LED_COUNT 16
module led_sw_all_reversed(
input [`LED_COUNT-1:0] sw,
output [0:`LED_COUNT-1] led // reversed bit-endianness
);
assign led = sw;
endmodule
```
Which worked fine using Vivado on my hardware with 16 switches and LEDs. Last but not least, I am new to Verilog.
PS: 😎 way of teaching FPGAs using graphics! | 1.0 | Feedback on endianness - On page https://projectf.io/posts/numbers-in-verilog/ you mention:
> Say you’ve got a bit-endian ...
You probably mean big
> ... byte from I2C and want to convert it to little-endian. Alas, you can’t mix big and little-endian vectors, so the following won’t work:
I would pay attention to endianness in this context. In my experience endianness is mostly known in data consisting of multiple bytes. I think you are talking about the endianness of bits, which seem to be called *bit endianness* ([example](https://stackoverflow.com/questions/6043483/why-bit-endianness-is-an-issue-in-bitfields).
---
Another issue I stumbled upon:
> wire [0:7] i2c_byte; // 8-bit wire (big-endian)
> reg [7:0] le_byte; // 8-bit reg (little-endian)
> always_ff @(posedge clk) le_byte <= i2c_byte; // Won't work :(
I tried:
```verilog
`define LED_COUNT 16
module led_sw_all_reversed(
input [`LED_COUNT-1:0] sw,
output [0:`LED_COUNT-1] led // reversed bit-endianness
);
assign led = sw;
endmodule
```
Which worked fine using Vivado on my hardware with 16 switches and LEDs. Last but not least, I am new to Verilog.
PS: 😎 way of teaching FPGAs using graphics! | non_usab | feedback on endianness on page you mention say you’ve got a bit endian you probably mean big byte from and want to convert it to little endian alas you can’t mix big and little endian vectors so the following won’t work i would pay attention to endianness in this context in my experience endianness is mostly known in data consisting of multiple bytes i think you are talking about the endianness of bits which seem to be called bit endianness another issue i stumbled upon wire byte bit wire big endian reg le byte bit reg little endian always ff posedge clk le byte byte won t work i tried verilog define led count module led sw all reversed input sw output led reversed bit endianness assign led sw endmodule which worked fine using vivado on my hardware with switches and leds last but not least i am new to verilog ps 😎 way of teaching fpgas using graphics | 0 |
6,746 | 4,534,191,130 | IssuesEvent | 2016-09-08 14:01:03 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 28206321: iMessage extension auto rotation is completely broken | classification:ui/usability reproducible:always status:open | #### Description
Steps to Reproduce:
- Open the IceCreamBuilder sample iMessage extension.
- Expand
- Rotate from portrait to landscape
- Collapse
Looks like in the attached screenshot.
-
Product Version: 10.0 GM
Created: 2016-09-08T13:47:18.862640
Originated: 2016-09-08T15:47:00
Open Radar Link: http://www.openradar.me/28206321 | True | 28206321: iMessage extension auto rotation is completely broken - #### Description
Steps to Reproduce:
- Open the IceCreamBuilder sample iMessage extension.
- Expand
- Rotate from portrait to landscape
- Collapse
Looks like in the attached screenshot.
-
Product Version: 10.0 GM
Created: 2016-09-08T13:47:18.862640
Originated: 2016-09-08T15:47:00
Open Radar Link: http://www.openradar.me/28206321 | usab | imessage extension auto rotation is completely broken description steps to reproduce open the icecreambuilder sample imessage extension expand rotate from portrait to landscape collapse looks like in the attached screenshot product version gm created originated open radar link | 1 |
21,947 | 18,149,869,788 | IssuesEvent | 2021-09-26 04:37:42 | tailscale/tailscale | https://api.github.com/repos/tailscale/tailscale | closed | Synology DSM6 x86_64 package (1.14.3) stops working after 1 minute | L1 Very few P3 Can't get started T6 Major usability OS-synology | Hello,
I installed the new package 1.14.3 (tailscale-x86_64-1.14.3-008-dsm6.spk) over the existing 1.12.1 version on my DS3617xs. That worked, NAS showed as "online" on the Tailscale website and after approx. 1 minuted the status went to "offline". I checked the NAS and the package was stopped - restarted the package, and it worked for roughly 1 minute again.
Please let me can provide any logs. I deleted it and installed 1.12.1 again.
Thank you!
<img src="https://frontapp.com/assets/img/favicons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_3kykx) | True | Synology DSM6 x86_64 package (1.14.3) stops working after 1 minute - Hello,
I installed the new package 1.14.3 (tailscale-x86_64-1.14.3-008-dsm6.spk) over the existing 1.12.1 version on my DS3617xs. That worked, NAS showed as "online" on the Tailscale website and after approx. 1 minuted the status went to "offline". I checked the NAS and the package was stopped - restarted the package, and it worked for roughly 1 minute again.
Please let me can provide any logs. I deleted it and installed 1.12.1 again.
Thank you!
<img src="https://frontapp.com/assets/img/favicons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_3kykx) | usab | synology package stops working after minute hello i installed the new package tailscale spk over the existing version on my that worked nas showed as online on the tailscale website and after approx minuted the status went to offline i checked the nas and the package was stopped restarted the package and it worked for roughly minute again please let me can provide any logs i deleted it and installed again thank you | 1 |
24,794 | 12,403,651,756 | IssuesEvent | 2020-05-21 14:16:14 | tgstation/tgstation-server | https://api.github.com/repos/tgstation/tgstation-server | closed | Database Commit step of Automatic Deployments can hang for... 30 MINUTES?!? (MySql/Linux only maybe) | Area: Jobs Backlog Database Issue Help Wanted Performance Reproduction Required | VORE-station experiences this regularly. The deployment process runs perfectly fine, except commiting the compile job to the database can hang for 30 minutes. This not only makes the chat bots liars when they say `Deployment Complete` it's obviously bad for bitcoin.
This doesn't seem to happen with manual deployments even though they follow the same code path. | True | Database Commit step of Automatic Deployments can hang for... 30 MINUTES?!? (MySql/Linux only maybe) - VORE-station experiences this regularly. The deployment process runs perfectly fine, except commiting the compile job to the database can hang for 30 minutes. This not only makes the chat bots liars when they say `Deployment Complete` it's obviously bad for bitcoin.
This doesn't seem to happen with manual deployments even though they follow the same code path. | non_usab | database commit step of automatic deployments can hang for minutes mysql linux only maybe vore station experiences this regularly the deployment process runs perfectly fine except commiting the compile job to the database can hang for minutes this not only makes the chat bots liars when they say deployment complete it s obviously bad for bitcoin this doesn t seem to happen with manual deployments even though they follow the same code path | 0 |
15,877 | 3,488,321,709 | IssuesEvent | 2016-01-02 21:11:10 | SemanticMediaWiki/SemanticMediaWiki | https://api.github.com/repos/SemanticMediaWiki/SemanticMediaWiki | opened | Setting multiple values #set/#subobject using `|` | question requires test wikidocu missing | [0] wrote "with some old templates of mine I found out that ... can be used to store multiple values for one property. Since I did not find any mention of it as I wanted to confirm this"
```
{{#set:
|property1=value1|value2|value3
|property2=value1|value2|value3
...
}}
```
Above is codified in [1] but we are missing an integration test [2, 3] and it would be great if someone could send a PR to cover this in order to avoid any regression in future.
[0] https://www.semantic-mediawiki.org/wiki/Thread:Help_talk:Setting_values/Setting_multiple_values_in_one_turn
[1] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/src/ParserParameterProcessor.php#L205-L212
[2] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/tests/phpunit/Integration/ByJsonScript/README.md
[3] https://github.com/SemanticMediaWiki/SemanticMediaWiki/tree/master/tests#write-integration-tests-using-json-script | 1.0 | Setting multiple values #set/#subobject using `|` - [0] wrote "with some old templates of mine I found out that ... can be used to store multiple values for one property. Since I did not find any mention of it as I wanted to confirm this"
```
{{#set:
|property1=value1|value2|value3
|property2=value1|value2|value3
...
}}
```
Above is codified in [1] but we are missing an integration test [2, 3] and it would be great if someone could send a PR to cover this in order to avoid any regression in future.
[0] https://www.semantic-mediawiki.org/wiki/Thread:Help_talk:Setting_values/Setting_multiple_values_in_one_turn
[1] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/src/ParserParameterProcessor.php#L205-L212
[2] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/tests/phpunit/Integration/ByJsonScript/README.md
[3] https://github.com/SemanticMediaWiki/SemanticMediaWiki/tree/master/tests#write-integration-tests-using-json-script | non_usab | setting multiple values set subobject using wrote with some old templates of mine i found out that can be used to store multiple values for one property since i did not find any mention of it as i wanted to confirm this set above is codified in but we are missing an integration test and it would be great if someone could send a pr to cover this in order to avoid any regression in future | 0 |
16,937 | 11,495,617,675 | IssuesEvent | 2020-02-12 05:29:26 | the-tale/the-tale | https://api.github.com/repos/the-tale/the-tale | closed | Гильдии: изменить или убрать ограничение на длину описания гильдии | comp_general cont_community cont_usability est_simple good first issue type_improvement | Сейчас слишком короткое. Сделать по аналогии с описаниями Хранителя/героя? | True | Гильдии: изменить или убрать ограничение на длину описания гильдии - Сейчас слишком короткое. Сделать по аналогии с описаниями Хранителя/героя? | usab | гильдии изменить или убрать ограничение на длину описания гильдии сейчас слишком короткое сделать по аналогии с описаниями хранителя героя | 1 |
23,113 | 6,369,372,045 | IssuesEvent | 2017-08-01 11:42:36 | Yoast/wordpress-seo | https://api.github.com/repos/Yoast/wordpress-seo | closed | Updated Premium page with My Yoast integration | needs-code-review | To work with the upcoming My Yoast, the Premium page in the plugin needs some changes.
- [x] Remove the `Licenses` tab completely (this functionality will be merged into the `Extensions` tab)
- [x] On the `Extensions` tab, add a second status label to all plugins shown:
- [x] When you don't own a plugin, show the `Buy` button with a link that says `More Information`.
- [x] When you have a plugin installed, show the `INSTALLED` label and the `NOT ACTIVATED` label, with a link to My Yoast that says `Activate your license on My Yoast >>`.
- [x] When the plugin is installed and activated in My Yoast, show the `INSTALLED` and `ACTIVATED` labels with a link that says `Manage this license on My Yoast >>`
- [x] Change the warning when a plugin is not activated to `Warning! You have not yet activated [PLUGIN NAME] in My Yoast. If you want to do so now, click here. Otherwise, you will not receive updates or support.` Which links to My Yoast.
- [x] Remove the page title and tab navigation at the top.
- [x] Increase the spacing between the upsell checklists and the buy button/status labels.

Fixes #6561. | 1.0 | Updated Premium page with My Yoast integration - To work with the upcoming My Yoast, the Premium page in the plugin needs some changes.
- [x] Remove the `Licenses` tab completely (this functionality will be merged into the `Extensions` tab)
- [x] On the `Extensions` tab, add a second status label to all plugins shown:
- [x] When you don't own a plugin, show the `Buy` button with a link that says `More Information`.
- [x] When you have a plugin installed, show the `INSTALLED` label and the `NOT ACTIVATED` label, with a link to My Yoast that says `Activate your license on My Yoast >>`.
- [x] When the plugin is installed and activated in My Yoast, show the `INSTALLED` and `ACTIVATED` labels with a link that says `Manage this license on My Yoast >>`
- [x] Change the warning when a plugin is not activated to `Warning! You have not yet activated [PLUGIN NAME] in My Yoast. If you want to do so now, click here. Otherwise, you will not receive updates or support.` Which links to My Yoast.
- [x] Remove the page title and tab navigation at the top.
- [x] Increase the spacing between the upsell checklists and the buy button/status labels.

Fixes #6561. | non_usab | updated premium page with my yoast integration to work with the upcoming my yoast the premium page in the plugin needs some changes remove the licenses tab completely this functionality will be merged into the extensions tab on the extensions tab add a second status label to all plugins shown when you don t own a plugin show the buy button with a link that says more information when you have a plugin installed show the installed label and the not activated label with a link to my yoast that says activate your license on my yoast when the plugin is installed and activated in my yoast show the installed and activated labels with a link that says manage this license on my yoast change the warning when a plugin is not activated to warning you have not yet activated in my yoast if you want to do so now click here otherwise you will not receive updates or support which links to my yoast remove the page title and tab navigation at the top increase the spacing between the upsell checklists and the buy button status labels fixes | 0 |
45,678 | 7,195,418,698 | IssuesEvent | 2018-02-04 16:58:10 | golang/go | https://api.github.com/repos/golang/go | opened | strconv: Unquote example looks like a unit test instead of an example | Documentation NeedsFix help wanted | The Unquote example (https://golang.org/pkg/strconv/#Unquote) looks like a unit test instead of an example.
That is a sea of backslashes and quotes. I think we could make a more readable example.
| 1.0 | strconv: Unquote example looks like a unit test instead of an example - The Unquote example (https://golang.org/pkg/strconv/#Unquote) looks like a unit test instead of an example.
That is a sea of backslashes and quotes. I think we could make a more readable example.
| non_usab | strconv unquote example looks like a unit test instead of an example the unquote example looks like a unit test instead of an example that is a sea of backslashes and quotes i think we could make a more readable example | 0 |
465,265 | 13,369,622,130 | IssuesEvent | 2020-09-01 09:06:00 | webcompat/web-bugs | https://api.github.com/repos/webcompat/web-bugs | closed | manabi-gakushu.benesse.ne.jp - site is not usable | browser-firefox engine-gecko priority-normal | <!-- @browser: Firefox 82.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:82.0) Gecko/20100101 Firefox/82.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/57348 -->
**URL**: https://manabi-gakushu.benesse.ne.jp/gakushu/typing/nihongonyuryoku.html
**Browser / Version**: Firefox 82.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Site is not usable
**Description**: Unable to type
**Steps to Reproduce**:
does not work type " - " key
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/8/50ab0f6d-d6ef-4297-b9a3-c997f39c6fd5.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200828153126</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/8/765ca612-fb1f-4421-a199-e5a9b85ab6e6)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | 1.0 | manabi-gakushu.benesse.ne.jp - site is not usable - <!-- @browser: Firefox 82.0 -->
<!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:82.0) Gecko/20100101 Firefox/82.0 -->
<!-- @reported_with: desktop-reporter -->
<!-- @public_url: https://github.com/webcompat/web-bugs/issues/57348 -->
**URL**: https://manabi-gakushu.benesse.ne.jp/gakushu/typing/nihongonyuryoku.html
**Browser / Version**: Firefox 82.0
**Operating System**: Windows 10
**Tested Another Browser**: Yes Edge
**Problem type**: Site is not usable
**Description**: Unable to type
**Steps to Reproduce**:
does not work type " - " key
<details>
<summary>View the screenshot</summary>
<img alt="Screenshot" src="https://webcompat.com/uploads/2020/8/50ab0f6d-d6ef-4297-b9a3-c997f39c6fd5.jpeg">
</details>
<details>
<summary>Browser Configuration</summary>
<ul>
<li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200828153126</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li>
</ul>
</details>
[View console log messages](https://webcompat.com/console_logs/2020/8/765ca612-fb1f-4421-a199-e5a9b85ab6e6)
_From [webcompat.com](https://webcompat.com/) with ❤️_ | non_usab | manabi gakushu benesse ne jp site is not usable url browser version firefox operating system windows tested another browser yes edge problem type site is not usable description unable to type steps to reproduce does not work type key view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️ | 0 |
82,662 | 16,011,575,303 | IssuesEvent | 2021-04-20 11:17:23 | wazuh/wazuh | https://api.github.com/repos/wazuh/wazuh | closed | <program_name> extracting an empty string | ruleset threatintel threatintel/decoders threatintel/rules | ```
Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun)
**Phase 1: Completed pre-decoding.
full event: 'Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun)'
timestamp: 'Nov 20 06:35:00'
hostname: 'PUNGDC-FDCS-DIST_SF'
program_name: ''
log: '(root) CMD ( /usr/libexec/atrun)'
**Phase 2: Completed decoding.
No decoder matched.
```
As you can notice below the result for that log gives program_name an empty string and not (null), I've tried with a pre-match entry but the decoder hasn't been detected.
The only way to get it work was to use the program_name in that way :
```
<decoder name="Junos-dist-switch-info4">
<program_name></program_name>
</decoder>
<decoder name="Junos-dist-info-cron">
<parent>Junos-dist-switch-info4</parent>
<regex>(\(\S+\)) \S+ (\(\s+\S+\))</regex>
<order>cron.user, cron.action</order>
</decoder>
```
the result is :
```
Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun)
**Phase 1: Completed pre-decoding.
full event: 'Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun) '
timestamp: 'Nov 20 06:35:00'
hostname: 'PUNGDC-FDCS-DIST_SF'
program_name: ''
log: '(root) CMD ( /usr/libexec/atrun) '
**Phase 2: Completed decoding.
decoder: 'Junos-dist-switch-info4'
cron.user: '(root)'
cron.action: '( /usr/libexec/atrun)'
```
That may make many conflicts especially if we have many logs from different sources where the program_name is extracted as an empty string like in this case.
| 1.0 | <program_name> extracting an empty string - ```
Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun)
**Phase 1: Completed pre-decoding.
full event: 'Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun)'
timestamp: 'Nov 20 06:35:00'
hostname: 'PUNGDC-FDCS-DIST_SF'
program_name: ''
log: '(root) CMD ( /usr/libexec/atrun)'
**Phase 2: Completed decoding.
No decoder matched.
```
As you can notice below the result for that log gives program_name an empty string and not (null), I've tried with a pre-match entry but the decoder hasn't been detected.
The only way to get it work was to use the program_name in that way :
```
<decoder name="Junos-dist-switch-info4">
<program_name></program_name>
</decoder>
<decoder name="Junos-dist-info-cron">
<parent>Junos-dist-switch-info4</parent>
<regex>(\(\S+\)) \S+ (\(\s+\S+\))</regex>
<order>cron.user, cron.action</order>
</decoder>
```
the result is :
```
Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun)
**Phase 1: Completed pre-decoding.
full event: 'Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun) '
timestamp: 'Nov 20 06:35:00'
hostname: 'PUNGDC-FDCS-DIST_SF'
program_name: ''
log: '(root) CMD ( /usr/libexec/atrun) '
**Phase 2: Completed decoding.
decoder: 'Junos-dist-switch-info4'
cron.user: '(root)'
cron.action: '( /usr/libexec/atrun)'
```
That may make many conflicts especially if we have many logs from different sources where the program_name is extracted as an empty string like in this case.
| non_usab | extracting an empty string nov pungdc fdcs dist sf root cmd usr libexec atrun phase completed pre decoding full event nov pungdc fdcs dist sf root cmd usr libexec atrun timestamp nov hostname pungdc fdcs dist sf program name log root cmd usr libexec atrun phase completed decoding no decoder matched as you can notice below the result for that log gives program name an empty string and not null i ve tried with a pre match entry but the decoder hasn t been detected the only way to get it work was to use the program name in that way junos dist switch s s s s cron user cron action the result is nov pungdc fdcs dist sf root cmd usr libexec atrun phase completed pre decoding full event nov pungdc fdcs dist sf root cmd usr libexec atrun timestamp nov hostname pungdc fdcs dist sf program name log root cmd usr libexec atrun phase completed decoding decoder junos dist switch cron user root cron action usr libexec atrun that may make many conflicts especially if we have many logs from different sources where the program name is extracted as an empty string like in this case | 0 |
9,461 | 6,307,560,505 | IssuesEvent | 2017-07-22 02:30:55 | coreos/bugs | https://api.github.com/repos/coreos/bugs | closed | iptables kernel panic | area/usability component/kernel kind/bug team/os | # Issue Report #
Bug
## Bug ##
When iptables-restore is used to perform replacements of very large tables, `ipt_alloc_initial_table` hangs, and when it does, the linux kernel panics.
The following screenshot was captured from the console of a virtual machine, but the issue also occurs on bare-metal machines.

### Container Linux Version ###
```
cat /etc/os-release
NAME=CoreOS
ID=coreos
VERSION=1185.5.0
VERSION_ID=1185.5.0
BUILD_ID=2016-12-07-0937
PRETTY_NAME="CoreOS 1185.5.0 (MoreOS)"
ANSI_COLOR="1;32"
HOME_URL="https://coreos.com/"
BUG_REPORT_URL="https://github.com/coreos/bugs/issues"
cat /proc/version
Linux version 4.7.3-coreos-r3 (jenkins@jenkins-os-executor-1.c.coreos-gce-testing.internal) (gcc version 4.9.3 (Gentoo Hardened 4.9.3 p1.5, pie-0.6.4) ) #1 SMP Wed Dec 7 09:29:55 UTC 2016
```
### Environment ###
replicated on:
bare metal machines.
also on virtual machines.
occurs when iptables directly on the host machine.
also occurs when iptables is executed from within a docker container.
### Expected Behavior ###
a kernel panic does not occur
### Actual Behavior ###
linux kernel panics
### Reproduction Steps ###
1. generate an iptables ruleset with lots and lots of rules.
2. run `iptables-restore -T nat --noflush --counters saved.log`
3. the kernel will panic
or here's a shell script:
```
# with sudo:
~ # for i in {1..100000}; do iptables-save -t nat > heh.txt; lines=$(cat heh.txt | wc -l); echo -n "#$i - saved $lines rules... "; iptables-restore -T nat --noflush --counters heh.txt ; date +%s; done
#1 - saved 259 rules... 1490325227
#2 - saved 262 rules... 1490325227
#3 - saved 268 rules... 1490325227
#4 - saved 280 rules... 1490325227
#5 - saved 304 rules... 1490325227
#6 - saved 352 rules... 1490325227
#7 - saved 448 rules... 1490325227
#8 - saved 640 rules... 1490325227
#9 - saved 1024 rules... 1490325227
#10 - saved 1792 rules... 1490325227
#11 - saved 3328 rules... 1490325228 <------------- NOTE: it starts to get sluggish here
#12 - saved 6400 rules... 1490325230
#13 - saved 12544 rules... 1490325237
#14 - saved 24832 rules... <------------- Can't exceed 32768?
```
| True | iptables kernel panic - # Issue Report #
Bug
## Bug ##
When iptables-restore is used to perform replacements of very large tables, `ipt_alloc_initial_table` hangs, and when it does, the linux kernel panics.
The following screenshot was captured from the console of a virtual machine, but the issue also occurs on bare-metal machines.

### Container Linux Version ###
```
cat /etc/os-release
NAME=CoreOS
ID=coreos
VERSION=1185.5.0
VERSION_ID=1185.5.0
BUILD_ID=2016-12-07-0937
PRETTY_NAME="CoreOS 1185.5.0 (MoreOS)"
ANSI_COLOR="1;32"
HOME_URL="https://coreos.com/"
BUG_REPORT_URL="https://github.com/coreos/bugs/issues"
cat /proc/version
Linux version 4.7.3-coreos-r3 (jenkins@jenkins-os-executor-1.c.coreos-gce-testing.internal) (gcc version 4.9.3 (Gentoo Hardened 4.9.3 p1.5, pie-0.6.4) ) #1 SMP Wed Dec 7 09:29:55 UTC 2016
```
### Environment ###
replicated on:
bare metal machines.
also on virtual machines.
occurs when iptables directly on the host machine.
also occurs when iptables is executed from within a docker container.
### Expected Behavior ###
a kernel panic does not occur
### Actual Behavior ###
linux kernel panics
### Reproduction Steps ###
1. generate an iptables ruleset with lots and lots of rules.
2. run `iptables-restore -T nat --noflush --counters saved.log`
3. the kernel will panic
or here's a shell script:
```
# with sudo:
~ # for i in {1..100000}; do iptables-save -t nat > heh.txt; lines=$(cat heh.txt | wc -l); echo -n "#$i - saved $lines rules... "; iptables-restore -T nat --noflush --counters heh.txt ; date +%s; done
#1 - saved 259 rules... 1490325227
#2 - saved 262 rules... 1490325227
#3 - saved 268 rules... 1490325227
#4 - saved 280 rules... 1490325227
#5 - saved 304 rules... 1490325227
#6 - saved 352 rules... 1490325227
#7 - saved 448 rules... 1490325227
#8 - saved 640 rules... 1490325227
#9 - saved 1024 rules... 1490325227
#10 - saved 1792 rules... 1490325227
#11 - saved 3328 rules... 1490325228 <------------- NOTE: it starts to get sluggish here
#12 - saved 6400 rules... 1490325230
#13 - saved 12544 rules... 1490325237
#14 - saved 24832 rules... <------------- Can't exceed 32768?
```
| usab | iptables kernel panic issue report bug bug when iptables restore is used to perform replacements of very large tables ipt alloc initial table hangs and when it does the linux kernel panics the following screenshot was captured from the console of a virtual machine but the issue also occurs on bare metal machines container linux version cat etc os release name coreos id coreos version version id build id pretty name coreos moreos ansi color home url bug report url cat proc version linux version coreos jenkins jenkins os executor c coreos gce testing internal gcc version gentoo hardened pie smp wed dec utc environment replicated on bare metal machines also on virtual machines occurs when iptables directly on the host machine also occurs when iptables is executed from within a docker container expected behavior a kernel panic does not occur actual behavior linux kernel panics reproduction steps generate an iptables ruleset with lots and lots of rules run iptables restore t nat noflush counters saved log the kernel will panic or here s a shell script with sudo for i in do iptables save t nat heh txt lines cat heh txt wc l echo n i saved lines rules iptables restore t nat noflush counters heh txt date s done saved rules saved rules saved rules saved rules saved rules saved rules saved rules saved rules saved rules saved rules saved rules note it starts to get sluggish here saved rules saved rules saved rules can t exceed | 1 |
5,104 | 3,900,324,251 | IssuesEvent | 2016-04-18 05:02:10 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 12551140: Smart Banners require an App Store installation | classification:ui/usability reproducible:always status:open | #### Description
Summary:
iOS 6 Smart Banners only work if the device has an App Store installation of the app in question. To-be-released products, and even Xcode builds of existing products, are not recognized by the smart banner machinery.
Steps to Reproduce:
0) Uninstall any copies of the app to be tested
1) Build and run your smart-banner-supporting app from Xcode
2) Go to your website with smart banner metadata
3) Note the VIEW button, which takes you to the App Store, rather than OPEN
4) Install the app from the App Store
5) Repeat 1)
Expected Results:
Expect smart banners to work all the time.
Actual Results:
Smart banners ONLY work after step 5 — specifically after building from Xcode on top of an App Store installation.
Regression:
iOS 6.0 (10A403)
Notes:
As far as I can tell, this makes Smart Banner testing completely impossible for pre-1.0 App Store apps. Similarly painful for existing products because the tester now needs the Xcode project and codesign / provisioning. This is a nonstarter for end-user beta testing, but even in many workplaces a QA team may not have source access. IPA installations from the Xcode Organizer do not work. Third-party OTA tools like TestFlight and Hockey are similarly out of luck.
Please revise this behavior to work with Debug and Ad-Hoc builds. This should ultimately be linked to the App's Bundle ID on the device side:
1) Safari reads the app-id parameter from the meta tag
2) Safari contacts the App Store for the related bundle ID
3) Safari asks the system if an app with said bundle ID exists
These steps can simply be used as a fallback against the existing current mechanism, rather than replacing it completely.
-
Product Version: 10A403
Created: 2012-10-22T21:31:55.921885
Originated: 2012-10-22T00:00:00
Open Radar Link: http://www.openradar.me/12551140 | True | 12551140: Smart Banners require an App Store installation - #### Description
Summary:
iOS 6 Smart Banners only work if the device has an App Store installation of the app in question. To-be-released products, and even Xcode builds of existing products, are not recognized by the smart banner machinery.
Steps to Reproduce:
0) Uninstall any copies of the app to be tested
1) Build and run your smart-banner-supporting app from Xcode
2) Go to your website with smart banner metadata
3) Note the VIEW button, which takes you to the App Store, rather than OPEN
4) Install the app from the App Store
5) Repeat 1)
Expected Results:
Expect smart banners to work all the time.
Actual Results:
Smart banners ONLY work after step 5 — specifically after building from Xcode on top of an App Store installation.
Regression:
iOS 6.0 (10A403)
Notes:
As far as I can tell, this makes Smart Banner testing completely impossible for pre-1.0 App Store apps. Similarly painful for existing products because the tester now needs the Xcode project and codesign / provisioning. This is a nonstarter for end-user beta testing, but even in many workplaces a QA team may not have source access. IPA installations from the Xcode Organizer do not work. Third-party OTA tools like TestFlight and Hockey are similarly out of luck.
Please revise this behavior to work with Debug and Ad-Hoc builds. This should ultimately be linked to the App's Bundle ID on the device side:
1) Safari reads the app-id parameter from the meta tag
2) Safari contacts the App Store for the related bundle ID
3) Safari asks the system if an app with said bundle ID exists
These steps can simply be used as a fallback against the existing current mechanism, rather than replacing it completely.
-
Product Version: 10A403
Created: 2012-10-22T21:31:55.921885
Originated: 2012-10-22T00:00:00
Open Radar Link: http://www.openradar.me/12551140 | usab | smart banners require an app store installation description summary ios smart banners only work if the device has an app store installation of the app in question to be released products and even xcode builds of existing products are not recognized by the smart banner machinery steps to reproduce uninstall any copies of the app to be tested build and run your smart banner supporting app from xcode go to your website with smart banner metadata note the view button which takes you to the app store rather than open install the app from the app store repeat expected results expect smart banners to work all the time actual results smart banners only work after step — specifically after building from xcode on top of an app store installation regression ios notes as far as i can tell this makes smart banner testing completely impossible for pre app store apps similarly painful for existing products because the tester now needs the xcode project and codesign provisioning this is a nonstarter for end user beta testing but even in many workplaces a qa team may not have source access ipa installations from the xcode organizer do not work third party ota tools like testflight and hockey are similarly out of luck please revise this behavior to work with debug and ad hoc builds this should ultimately be linked to the app s bundle id on the device side safari reads the app id parameter from the meta tag safari contacts the app store for the related bundle id safari asks the system if an app with said bundle id exists these steps can simply be used as a fallback against the existing current mechanism rather than replacing it completely product version created originated open radar link | 1 |
34,362 | 12,269,416,894 | IssuesEvent | 2020-05-07 14:02:52 | logzio/jmx2logzio | https://api.github.com/repos/logzio/jmx2logzio | closed | CVE-2020-9547 (Medium) detected in jackson-databind-2.9.10.2.jar | security vulnerability | ## CVE-2020-9547 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/jmx2logzio/pom.xml</p>
<p>Path to vulnerable library: downloadResource_c762fbf8-ee9b-4076-944f-b22607c0cecb/20200210155926/jackson-databind-2.9.10.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.2.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).
<p>Publish Date: 2020-03-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547>CVE-2020-9547</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p>
<p>Release Date: 2020-03-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.2","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.3"}],"vulnerabilityIdentifier":"CVE-2020-9547","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> --> | True | CVE-2020-9547 (Medium) detected in jackson-databind-2.9.10.2.jar - ## CVE-2020-9547 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.2.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: /tmp/ws-scm/jmx2logzio/pom.xml</p>
<p>Path to vulnerable library: downloadResource_c762fbf8-ee9b-4076-944f-b22607c0cecb/20200210155926/jackson-databind-2.9.10.2.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.10.2.jar** (Vulnerable Library)
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).
<p>Publish Date: 2020-03-02
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547>CVE-2020-9547</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p>
<p>Release Date: 2020-03-02</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p>
</p>
</details>
<p></p>
***
<!-- REMEDIATE-OPEN-PR-START -->
- [ ] Check this box to open an automated fix PR
<!-- REMEDIATE-OPEN-PR-END -->
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.2","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.3"}],"vulnerabilityIdentifier":"CVE-2020-9547","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> --> | non_usab | cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm pom xml path to vulnerable library downloadresource jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap vulnerabilityurl | 0 |
11,020 | 7,027,700,807 | IssuesEvent | 2017-12-25 01:42:33 | godotengine/godot | https://api.github.com/repos/godotengine/godot | closed | "F" key Viewport focus problem | bug junior job topic:editor usability | version -> http://godot3builds.digitecnology.com/builds/2017-12-06.12:52:48.a8ae46e14/godot-v3.0-osx.fat.zip
**Issue description:**
When i press F key, wrong viewport is active.
**Steps to reproduce:**
https://www.youtube.com/watch?v=y5K5xv7UBuo&feature=youtu.be
| True | "F" key Viewport focus problem - version -> http://godot3builds.digitecnology.com/builds/2017-12-06.12:52:48.a8ae46e14/godot-v3.0-osx.fat.zip
**Issue description:**
When i press F key, wrong viewport is active.
**Steps to reproduce:**
https://www.youtube.com/watch?v=y5K5xv7UBuo&feature=youtu.be
| usab | f key viewport focus problem version issue description when i press f key wrong viewport is active steps to reproduce | 1 |
24,613 | 24,033,901,064 | IssuesEvent | 2022-09-15 17:17:02 | pulumi/pulumi | https://api.github.com/repos/pulumi/pulumi | opened | pulumi new project already exists error does not respect the default org | kind/bug impact/usability | [internal slack convo](https://pulumi.slack.com/archives/C9SGX9QA1/p1663254061919969)
repro steps
- create a project in your individual account with name X
- set your default org to an org that does not have a project with the name X
- use pulumi new
- when it asks for a project name, set it to X
- receive an error saying that project already exists | True | pulumi new project already exists error does not respect the default org - [internal slack convo](https://pulumi.slack.com/archives/C9SGX9QA1/p1663254061919969)
repro steps
- create a project in your individual account with name X
- set your default org to an org that does not have a project with the name X
- use pulumi new
- when it asks for a project name, set it to X
- receive an error saying that project already exists | usab | pulumi new project already exists error does not respect the default org repro steps create a project in your individual account with name x set your default org to an org that does not have a project with the name x use pulumi new when it asks for a project name set it to x receive an error saying that project already exists | 1 |
8,769 | 5,957,741,646 | IssuesEvent | 2017-05-29 04:21:36 | Virtual-Labs/circular-dichronism-spectroscopy-iiith | https://api.github.com/repos/Virtual-Labs/circular-dichronism-spectroscopy-iiith | opened | QA_Circular Dichroism Spectroscopy_Introduction_Target Audiance_Spelling-mistakes | Category:Usability Developed by: VLEAD Open-Edx Severity:S2 Severity:S3 | Defect Description :
In the Sub section "Target Audience" of Introduction section of this lab, found spelling mistakes.
Actual Result :
In the Sub section "Target Audience" of Introduction section of this lab, found spelling mistakes.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers:Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Attachments:

| True | QA_Circular Dichroism Spectroscopy_Introduction_Target Audiance_Spelling-mistakes - Defect Description :
In the Sub section "Target Audience" of Introduction section of this lab, found spelling mistakes.
Actual Result :
In the Sub section "Target Audience" of Introduction section of this lab, found spelling mistakes.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers:Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5
Attachments:

| usab | qa circular dichroism spectroscopy introduction target audiance spelling mistakes defect description in the sub section target audience of introduction section of this lab found spelling mistakes actual result in the sub section target audience of introduction section of this lab found spelling mistakes environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor attachments | 1 |
14,444 | 9,194,935,033 | IssuesEvent | 2019-03-07 00:16:25 | pulumi/pulumi | https://api.github.com/repos/pulumi/pulumi | opened | RunErrors aren't concisely reported anymore | area/cli impact/usability kind/bug | It used to be the case that `RunError`s led to concise error reporting, including omission of the stack trace. The idea here was that these errors want to instruct the user to do something differently -- like run `pulumi config set ...`, and that the extra noise of a stack trace is simply confusing.
It appears this behavior has regressed semi-recently:

| True | RunErrors aren't concisely reported anymore - It used to be the case that `RunError`s led to concise error reporting, including omission of the stack trace. The idea here was that these errors want to instruct the user to do something differently -- like run `pulumi config set ...`, and that the extra noise of a stack trace is simply confusing.
It appears this behavior has regressed semi-recently:

| usab | runerrors aren t concisely reported anymore it used to be the case that runerror s led to concise error reporting including omission of the stack trace the idea here was that these errors want to instruct the user to do something differently like run pulumi config set and that the extra noise of a stack trace is simply confusing it appears this behavior has regressed semi recently | 1 |
272,800 | 29,795,090,404 | IssuesEvent | 2023-06-16 01:10:11 | billmcchesney1/hadoop | https://api.github.com/repos/billmcchesney1/hadoop | closed | WS-2017-0234 (Medium) detected in jquery.dataTables-1.10.7.min.js - autoclosed | Mend: dependency security vulnerability | ## WS-2017-0234 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.dataTables-1.10.7.min.js</b></p></summary>
<p>DataTables enhances HTML tables with the ability to sort, filter and page the data in the table very easily. It provides a comprehensive API and set of configuration options, allowing you to consume data from virtually any data source.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/datatables/1.10.7/js/jquery.dataTables.min.js">https://cdnjs.cloudflare.com/ajax/libs/datatables/1.10.7/js/jquery.dataTables.min.js</a></p>
<p>Path to vulnerable library: /hadoop-hdfs-project/hadoop-hdfs/target/webapps/static/jquery.dataTables.min.js,/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/jquery.dataTables.min.js,/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/static/jquery.dataTables.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery.dataTables-1.10.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p>
<p>Found in base branch: <b>trunk</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Affected versions of the package are vulnerable to Cross-site Scripting (XSS).
<p>Publish Date: 2015-11-06
<p>URL: <a href=https://github.com/DataTables/DataTables/commit/6f67df2d21f9858ec40a6e9565c3a653cdb691a6>WS-2017-0234</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2015-11-06</p>
<p>Fix Resolution: datatables - 1.10.10;org.webjars:datatables:1.10.10</p>
</p>
</details>
<p></p>
| True | WS-2017-0234 (Medium) detected in jquery.dataTables-1.10.7.min.js - autoclosed - ## WS-2017-0234 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.dataTables-1.10.7.min.js</b></p></summary>
<p>DataTables enhances HTML tables with the ability to sort, filter and page the data in the table very easily. It provides a comprehensive API and set of configuration options, allowing you to consume data from virtually any data source.</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/datatables/1.10.7/js/jquery.dataTables.min.js">https://cdnjs.cloudflare.com/ajax/libs/datatables/1.10.7/js/jquery.dataTables.min.js</a></p>
<p>Path to vulnerable library: /hadoop-hdfs-project/hadoop-hdfs/target/webapps/static/jquery.dataTables.min.js,/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/jquery.dataTables.min.js,/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/static/jquery.dataTables.min.js</p>
<p>
Dependency Hierarchy:
- :x: **jquery.dataTables-1.10.7.min.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p>
<p>Found in base branch: <b>trunk</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
Affected versions of the package are vulnerable to Cross-site Scripting (XSS).
<p>Publish Date: 2015-11-06
<p>URL: <a href=https://github.com/DataTables/DataTables/commit/6f67df2d21f9858ec40a6e9565c3a653cdb691a6>WS-2017-0234</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Release Date: 2015-11-06</p>
<p>Fix Resolution: datatables - 1.10.10;org.webjars:datatables:1.10.10</p>
</p>
</details>
<p></p>
| non_usab | ws medium detected in jquery datatables min js autoclosed ws medium severity vulnerability vulnerable library jquery datatables min js datatables enhances html tables with the ability to sort filter and page the data in the table very easily it provides a comprehensive api and set of configuration options allowing you to consume data from virtually any data source library home page a href path to vulnerable library hadoop hdfs project hadoop hdfs target webapps static jquery datatables min js hadoop hdfs project hadoop hdfs target test classes webapps static jquery datatables min js hadoop hdfs project hadoop hdfs src main webapps static jquery datatables min js dependency hierarchy x jquery datatables min js vulnerable library found in head commit a href found in base branch trunk vulnerability details affected versions of the package are vulnerable to cross site scripting xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution datatables org webjars datatables | 0 |
10,457 | 6,733,274,596 | IssuesEvent | 2017-10-18 14:22:22 | elastic/rally | https://api.github.com/repos/elastic/rally | closed | Provide a clear error message for non-existing releases | :Benchmark Candidate Management :Usability enhancement | If users have a typo in their distribution specification, then they get a very unclear error message, e.g.:
```
tarfile.ReadError: file could not be opened successfully
```
We should detect this condition and provide a clear error message.
| True | Provide a clear error message for non-existing releases - If users have a typo in their distribution specification, then they get a very unclear error message, e.g.:
```
tarfile.ReadError: file could not be opened successfully
```
We should detect this condition and provide a clear error message.
| usab | provide a clear error message for non existing releases if users have a typo in their distribution specification then they get a very unclear error message e g tarfile readerror file could not be opened successfully we should detect this condition and provide a clear error message | 1 |
25,829 | 25,961,444,150 | IssuesEvent | 2022-12-18 23:32:10 | kubernetes/test-infra | https://api.github.com/repos/kubernetes/test-infra | closed | Using Ephemeral Volume Snapshot with Prow Job | kind/feature lifecycle/rotten sig/usability | <!-- Please only use this template for submitting enhancement requests -->
**What would you like to be added**:
Seems like a presubmit prow job is not able to create a volume by using "volume claim template" described in the presubmit job yaml. I am trying to use the volume claim template to create a PV from volume snapshot "data source" in the yaml. The mount paths come up empty. The setup works fine when tested with a POD but not with prowjob yaml.
Eg:
```
volumes:
- name: try-vol
ephemeral:
volumeClaimTemplate:
spec:
accessModes:
- ReadWriteOnce
storageClassName: "xxx-xxx-x-xx"
resources:
requests:
storage: 60Gi
dataSource:
name: snapshot
kind: VolumeSnapshot
apiGroup: snapshot.storage.k8s.io
```
**Why is this needed**:
Trying to mount cached artefacts on the fly for PROW presubmit job to make CI faster. | True | Using Ephemeral Volume Snapshot with Prow Job - <!-- Please only use this template for submitting enhancement requests -->
**What would you like to be added**:
Seems like a presubmit prow job is not able to create a volume by using "volume claim template" described in the presubmit job yaml. I am trying to use the volume claim template to create a PV from volume snapshot "data source" in the yaml. The mount paths come up empty. The setup works fine when tested with a POD but not with prowjob yaml.
Eg:
```
volumes:
- name: try-vol
ephemeral:
volumeClaimTemplate:
spec:
accessModes:
- ReadWriteOnce
storageClassName: "xxx-xxx-x-xx"
resources:
requests:
storage: 60Gi
dataSource:
name: snapshot
kind: VolumeSnapshot
apiGroup: snapshot.storage.k8s.io
```
**Why is this needed**:
Trying to mount cached artefacts on the fly for PROW presubmit job to make CI faster. | usab | using ephemeral volume snapshot with prow job what would you like to be added seems like a presubmit prow job is not able to create a volume by using volume claim template described in the presubmit job yaml i am trying to use the volume claim template to create a pv from volume snapshot data source in the yaml the mount paths come up empty the setup works fine when tested with a pod but not with prowjob yaml eg volumes name try vol ephemeral volumeclaimtemplate spec accessmodes readwriteonce storageclassname xxx xxx x xx resources requests storage datasource name snapshot kind volumesnapshot apigroup snapshot storage io why is this needed trying to mount cached artefacts on the fly for prow presubmit job to make ci faster | 1 |
18,940 | 13,486,582,078 | IssuesEvent | 2020-09-11 09:39:43 | CARTAvis/carta-frontend | https://api.github.com/repos/CARTAvis/carta-frontend | closed | Spectral Profile opened automatically | usability review | From Usability Review, Region Analytics:
I certainly think that the spectral profile should be opened automatically (much more often required than the X or Y). I think you could also probably update the button pictures to make them a little more obvious? | True | Spectral Profile opened automatically - From Usability Review, Region Analytics:
I certainly think that the spectral profile should be opened automatically (much more often required than the X or Y). I think you could also probably update the button pictures to make them a little more obvious? | usab | spectral profile opened automatically from usability review region analytics i certainly think that the spectral profile should be opened automatically much more often required than the x or y i think you could also probably update the button pictures to make them a little more obvious | 1 |
672,141 | 22,792,889,882 | IssuesEvent | 2022-07-10 09:15:55 | crestic-urca/remotelabz | https://api.github.com/repos/crestic-urca/remotelabz | closed | The forward seems no to be activate from install | normal priority needs more info | Verify if the net.ipv4.ip_forward=1 is ok on the front and the worker in the install process | 1.0 | The forward seems no to be activate from install - Verify if the net.ipv4.ip_forward=1 is ok on the front and the worker in the install process | non_usab | the forward seems no to be activate from install verify if the net ip forward is ok on the front and the worker in the install process | 0 |
10,547 | 6,794,678,069 | IssuesEvent | 2017-11-01 13:11:27 | jamesmillerburgess/customer-management | https://api.github.com/repos/jamesmillerburgess/customer-management | closed | Screen Resolution Iphone 6+ versus SamsungA5 2017 | Minor Issue Usability | Just made a quick comparison Iphone versus Samsung. Noticed that the display of the fields are different when looking at a contact
Reasons ?
- Screen resolution ?
- Length of the fields ?

| True | Screen Resolution Iphone 6+ versus SamsungA5 2017 - Just made a quick comparison Iphone versus Samsung. Noticed that the display of the fields are different when looking at a contact
Reasons ?
- Screen resolution ?
- Length of the fields ?

| usab | screen resolution iphone versus just made a quick comparison iphone versus samsung noticed that the display of the fields are different when looking at a contact reasons screen resolution length of the fields | 1 |
34,291 | 29,190,851,222 | IssuesEvent | 2023-05-19 19:50:25 | woocommerce/woocommerce | https://api.github.com/repos/woocommerce/woocommerce | closed | Optimize React Admin's `package.json` Scripts | type: task plugin: woocommerce tool: monorepo infrastructure focus: react admin [team:Ghidorah] | <!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. -->
**Prerequisites (mark completed items with an [x]):**
- [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose
- [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is).
**Issue Description:**
There are a large number of different workflows supported by the current `package.json` script. As a consequence of our tooling changes, however, there is a non-trivial amount of extra work being done in these scripts. Once https://github.com/woocommerce/woocommerce/issues/33907 lands, we should review all of the scripts and look for areas of performance deficit:
- `build:packages` is no longer necessary, `turbo` builds all dependencies.
- Check for cases where `turbo` is being bypassed entirely.
- Review supported workflows and discuss flattening them where appropriate with their respective teams. | 1.0 | Optimize React Admin's `package.json` Scripts - <!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. -->
**Prerequisites (mark completed items with an [x]):**
- [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose
- [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is).
**Issue Description:**
There are a large number of different workflows supported by the current `package.json` script. As a consequence of our tooling changes, however, there is a non-trivial amount of extra work being done in these scripts. Once https://github.com/woocommerce/woocommerce/issues/33907 lands, we should review all of the scripts and look for areas of performance deficit:
- `build:packages` is no longer necessary, `turbo` builds all dependencies.
- Check for cases where `turbo` is being bypassed entirely.
- Review supported workflows and discuss flattening them where appropriate with their respective teams. | non_usab | optimize react admin s package json scripts prerequisites mark completed items with an i have checked that my issue type is not listed here my issue is not a security issue support request bug report enhancement or feature request please use the link above if it is issue description there are a large number of different workflows supported by the current package json script as a consequence of our tooling changes however there is a non trivial amount of extra work being done in these scripts once lands we should review all of the scripts and look for areas of performance deficit build packages is no longer necessary turbo builds all dependencies check for cases where turbo is being bypassed entirely review supported workflows and discuss flattening them where appropriate with their respective teams | 0 |
8,859 | 6,008,439,179 | IssuesEvent | 2017-06-06 07:51:09 | bwsw/cloudstack-ui | https://api.github.com/repos/bwsw/cloudstack-ui | closed | Vm/Disk/other pane actions (...) bar | enhancement proposal usability | Currently actions bar is displayed top-middle of VM/Disk/etc bar, I think it's may be better to display it in the middle of the bar. | True | Vm/Disk/other pane actions (...) bar - Currently actions bar is displayed top-middle of VM/Disk/etc bar, I think it's may be better to display it in the middle of the bar. | usab | vm disk other pane actions bar currently actions bar is displayed top middle of vm disk etc bar i think it s may be better to display it in the middle of the bar | 1 |
18,328 | 12,795,613,162 | IssuesEvent | 2020-07-02 09:02:38 | WordPress/gutenberg | https://api.github.com/repos/WordPress/gutenberg | closed | Block appender appears when a reusable Heading block is selected | [Feature] Reusable Blocks [Type] Bug | **Describe the bug**
A modified Heading block saved as a Reusable block appears frozen, it neither accepts text or adds a new block when the + is pressed.
**To reproduce**
Steps to reproduce the behavior:
1. In the editor, click the + and add a Heading block
2. Change the Heading block to be H3, and Align Text Center.
3. Click on Add to reusable blocks
4. Change Resuable Block name to H3C and hit Save
5. Click on the Write heading... text to enter text, but nothing happens. No text can be entered.
6. Click on the + inside the Reusable block, and find that it does nothing.
**Expected behavior**
I expect in step 5 to see the text I'm trying to enter appear. I expect in step 6 to have a new block appear when I hit the +.
**Screenshots**

The upper box is the non-working saved reusable block, the lower asdf-text is the text inside a regular Heading block.
**Desktop (please complete the following information):**
- OS: Windows 10 Pro x64
- Browser: Chrome
- Version: 80.0.3987.163
**Additional context**
- Please add the version of Gutenberg you are using in the description: Version 7.8.1
This is the only thing in the console:
load-scripts.php?c=0&load[chunk_0]=jquery-core,jquery-migrate,utils&ver=5.4:8 JQMIGRATE: Migrate is installed, version 1.4.1
post.php?post=25380&action=edit:1 [DOM] Found 2 elements with non-unique id #_wpnonce: (More info: https://goo.gl/9p2vKq) <input type="hidden" id="_wpnonce" name="_wpnonce" value="8cf45b233c"> <input type="hidden" id="_wpnonce" name="_wpnonce" value="a13fa8130a"> | True | Block appender appears when a reusable Heading block is selected - **Describe the bug**
A modified Heading block saved as a Reusable block appears frozen, it neither accepts text or adds a new block when the + is pressed.
**To reproduce**
Steps to reproduce the behavior:
1. In the editor, click the + and add a Heading block
2. Change the Heading block to be H3, and Align Text Center.
3. Click on Add to reusable blocks
4. Change Resuable Block name to H3C and hit Save
5. Click on the Write heading... text to enter text, but nothing happens. No text can be entered.
6. Click on the + inside the Reusable block, and find that it does nothing.
**Expected behavior**
I expect in step 5 to see the text I'm trying to enter appear. I expect in step 6 to have a new block appear when I hit the +.
**Screenshots**

The upper box is the non-working saved reusable block, the lower asdf-text is the text inside a regular Heading block.
**Desktop (please complete the following information):**
- OS: Windows 10 Pro x64
- Browser: Chrome
- Version: 80.0.3987.163
**Additional context**
- Please add the version of Gutenberg you are using in the description: Version 7.8.1
This is the only thing in the console:
load-scripts.php?c=0&load[chunk_0]=jquery-core,jquery-migrate,utils&ver=5.4:8 JQMIGRATE: Migrate is installed, version 1.4.1
post.php?post=25380&action=edit:1 [DOM] Found 2 elements with non-unique id #_wpnonce: (More info: https://goo.gl/9p2vKq) <input type="hidden" id="_wpnonce" name="_wpnonce" value="8cf45b233c"> <input type="hidden" id="_wpnonce" name="_wpnonce" value="a13fa8130a"> | usab | block appender appears when a reusable heading block is selected describe the bug a modified heading block saved as a reusable block appears frozen it neither accepts text or adds a new block when the is pressed to reproduce steps to reproduce the behavior in the editor click the and add a heading block change the heading block to be and align text center click on add to reusable blocks change resuable block name to and hit save click on the write heading text to enter text but nothing happens no text can be entered click on the inside the reusable block and find that it does nothing expected behavior i expect in step to see the text i m trying to enter appear i expect in step to have a new block appear when i hit the screenshots the upper box is the non working saved reusable block the lower asdf text is the text inside a regular heading block desktop please complete the following information os windows pro browser chrome version additional context please add the version of gutenberg you are using in the description version this is the only thing in the console load scripts php c load jquery core jquery migrate utils ver jqmigrate migrate is installed version post php post action edit found elements with non unique id wpnonce more info | 1 |
20,216 | 15,148,043,958 | IssuesEvent | 2021-02-11 10:02:43 | Elgg/Elgg | https://api.github.com/repos/Elgg/Elgg | closed | site_notification FR: remove notification if content gets removed | feature notifications usability | It would be nice if the site notifications about content were removed when the content is removed.
eg: user a writes a blog, users B & C get a notification about this. User A removes the blog before B&C read the notification. User B&C check their notifications and click on the link, but it no longer works.
If the notification was related to the content, it would be easy to remove the notification if the content was removed.
| True | site_notification FR: remove notification if content gets removed - It would be nice if the site notifications about content were removed when the content is removed.
eg: user a writes a blog, users B & C get a notification about this. User A removes the blog before B&C read the notification. User B&C check their notifications and click on the link, but it no longer works.
If the notification was related to the content, it would be easy to remove the notification if the content was removed.
| usab | site notification fr remove notification if content gets removed it would be nice if the site notifications about content were removed when the content is removed eg user a writes a blog users b c get a notification about this user a removes the blog before b c read the notification user b c check their notifications and click on the link but it no longer works if the notification was related to the content it would be easy to remove the notification if the content was removed | 1 |
178,387 | 14,668,372,224 | IssuesEvent | 2020-12-29 21:04:26 | melezhik/sparky | https://api.github.com/repos/melezhik/sparky | closed | Problem loading configs hash | documentation enhancement | Hello,
`raku bin/sparky-runner.raku --dir=~/.sparky/projects/hello/` is not working because the `~` is not translated to absolute path at this [line](https://github.com/melezhik/sparky/blob/master/bin/sparky-runner.raku#L281).
You end up with this empty configuration because we never enter the if block :
`merged sparrowdo configuration: { }`
| 1.0 | Problem loading configs hash - Hello,
`raku bin/sparky-runner.raku --dir=~/.sparky/projects/hello/` is not working because the `~` is not translated to absolute path at this [line](https://github.com/melezhik/sparky/blob/master/bin/sparky-runner.raku#L281).
You end up with this empty configuration because we never enter the if block :
`merged sparrowdo configuration: { }`
| non_usab | problem loading configs hash hello raku bin sparky runner raku dir sparky projects hello is not working because the is not translated to absolute path at this you end up with this empty configuration because we never enter the if block merged sparrowdo configuration | 0 |
26,133 | 26,466,017,754 | IssuesEvent | 2023-01-16 23:41:53 | hashicorp/consul | https://api.github.com/repos/hashicorp/consul | closed | Should error on validation when upstream port the same as envoy public listener | type/bug theme/operator-usability theme/connect | #### Overview of the Issue
This registration doesn't error:
```hcl
service {
name = "frontend"
port = 6060
connect {
sidecar_service {
port = 20000
proxy {
upstreams = [
{
destination_name = "backend"
local_bind_port = 20000
}
]
}
}
}
}
```
But it results in listeners:
```json
{
"listener_statuses": [
{
"name": "public_listener:0.0.0.0:20000",
"local_address": {
"socket_address": {
"address": "0.0.0.0",
"port_value": 20000
}
}
},
{
"name": "backend:127.0.0.1:20000",
"local_address": {
"socket_address": {
"address": "127.0.0.1",
"port_value": 20000
}
}
}
]
}
```
Such that in one reproduction upstream requests couldn't be made.
When I tried to reproduce, I got hundreds of these errors in the logs, 20/second:
```
2021-11-15T00:21:43.511Z [ERROR] agent.envoy.xds: got error response from envoy proxy: service_id=backend-sidecar-proxy typeUrl=type.googleapis.com/envoy.config.listener.v3.Listener xdsVersion=v3 nonce=00001ed2 error="rpc error: code = Internal desc = Error adding/updating listener(s) frontend:127.0.0.1:21000: cannot bind '127.0.0.1:21000': Address already in use
"
```
But I never saw both listeners bound to the same port in the xds config, so not sure what the difference between my repro and the situation I saw in the wild.
Regardless if it's possibly then I think we could safely error out on service registration in this case?
### Consul info for both Client and Server
Consul 1.10.3
| True | Should error on validation when upstream port the same as envoy public listener - #### Overview of the Issue
This registration doesn't error:
```hcl
service {
name = "frontend"
port = 6060
connect {
sidecar_service {
port = 20000
proxy {
upstreams = [
{
destination_name = "backend"
local_bind_port = 20000
}
]
}
}
}
}
```
But it results in listeners:
```json
{
"listener_statuses": [
{
"name": "public_listener:0.0.0.0:20000",
"local_address": {
"socket_address": {
"address": "0.0.0.0",
"port_value": 20000
}
}
},
{
"name": "backend:127.0.0.1:20000",
"local_address": {
"socket_address": {
"address": "127.0.0.1",
"port_value": 20000
}
}
}
]
}
```
Such that in one reproduction upstream requests couldn't be made.
When I tried to reproduce, I got hundreds of these errors in the logs, 20/second:
```
2021-11-15T00:21:43.511Z [ERROR] agent.envoy.xds: got error response from envoy proxy: service_id=backend-sidecar-proxy typeUrl=type.googleapis.com/envoy.config.listener.v3.Listener xdsVersion=v3 nonce=00001ed2 error="rpc error: code = Internal desc = Error adding/updating listener(s) frontend:127.0.0.1:21000: cannot bind '127.0.0.1:21000': Address already in use
"
```
But I never saw both listeners bound to the same port in the xds config, so not sure what the difference between my repro and the situation I saw in the wild.
Regardless if it's possibly then I think we could safely error out on service registration in this case?
### Consul info for both Client and Server
Consul 1.10.3
| usab | should error on validation when upstream port the same as envoy public listener overview of the issue this registration doesn t error hcl service name frontend port connect sidecar service port proxy upstreams destination name backend local bind port but it results in listeners json listener statuses name public listener local address socket address address port value name backend local address socket address address port value such that in one reproduction upstream requests couldn t be made when i tried to reproduce i got hundreds of these errors in the logs second agent envoy xds got error response from envoy proxy service id backend sidecar proxy typeurl type googleapis com envoy config listener listener xdsversion nonce error rpc error code internal desc error adding updating listener s frontend cannot bind address already in use but i never saw both listeners bound to the same port in the xds config so not sure what the difference between my repro and the situation i saw in the wild regardless if it s possibly then i think we could safely error out on service registration in this case consul info for both client and server consul | 1 |
70,052 | 9,370,477,304 | IssuesEvent | 2019-04-03 13:33:28 | NervanaSystems/coach | https://api.github.com/repos/NervanaSystems/coach | closed | Two typos in coach.py | bug documentation priority/p3 | In line 263, 269, 276 of coach.py of master branch, what does the `--experiment_type` means? And the `--environemnt_type` should be `--environment_type`, just a typo.
One another problem: the file says
> "If no preset is defined, a preset can be set from the command-line by combining settings "
"which are set by using --agent_type, --experiment_type, --environemnt_type"
but it autually shows this thing when I run `coach -at ddpg_agent -et control_suite_environment -ept e-greedy` :
> Please choose a preset using the -p flag or use the --play flag together with choosing an environment type (-et) in order to play the game.
The arguments help information is outdated? | 1.0 | Two typos in coach.py - In line 263, 269, 276 of coach.py of master branch, what does the `--experiment_type` means? And the `--environemnt_type` should be `--environment_type`, just a typo.
One another problem: the file says
> "If no preset is defined, a preset can be set from the command-line by combining settings "
"which are set by using --agent_type, --experiment_type, --environemnt_type"
but it autually shows this thing when I run `coach -at ddpg_agent -et control_suite_environment -ept e-greedy` :
> Please choose a preset using the -p flag or use the --play flag together with choosing an environment type (-et) in order to play the game.
The arguments help information is outdated? | non_usab | two typos in coach py in line of coach py of master branch what does the experiment type means and the environemnt type should be environment type just a typo one another problem the file says if no preset is defined a preset can be set from the command line by combining settings which are set by using agent type experiment type environemnt type but it autually shows this thing when i run coach at ddpg agent et control suite environment ept e greedy please choose a preset using the p flag or use the play flag together with choosing an environment type et in order to play the game the arguments help information is outdated | 0 |
11,305 | 7,156,472,485 | IssuesEvent | 2018-01-26 16:24:04 | textpattern/textpattern | https://api.github.com/repos/textpattern/textpattern | closed | 1x pophelp icon addition: Write -> Status | Usability | I had a client look around the 4.7 dev interface and took some feedback. A lot of it is already in hand, but one thing was something I had overlooked: article statuses from the Write panel. There's no pophelp icon.
The docs work has already been done, if we steal the content from `default_publish_status` pophelp and change the heading/intro text, so it's an easy win from that side of things.
How reasonable is it to have a pophelp inline'd on `Status`, please. I can take dibs on the glue from the https://github.com/textpattern/pophelp end. | True | 1x pophelp icon addition: Write -> Status - I had a client look around the 4.7 dev interface and took some feedback. A lot of it is already in hand, but one thing was something I had overlooked: article statuses from the Write panel. There's no pophelp icon.
The docs work has already been done, if we steal the content from `default_publish_status` pophelp and change the heading/intro text, so it's an easy win from that side of things.
How reasonable is it to have a pophelp inline'd on `Status`, please. I can take dibs on the glue from the https://github.com/textpattern/pophelp end. | usab | pophelp icon addition write status i had a client look around the dev interface and took some feedback a lot of it is already in hand but one thing was something i had overlooked article statuses from the write panel there s no pophelp icon the docs work has already been done if we steal the content from default publish status pophelp and change the heading intro text so it s an easy win from that side of things how reasonable is it to have a pophelp inline d on status please i can take dibs on the glue from the end | 1 |
27,216 | 27,856,139,088 | IssuesEvent | 2023-03-20 23:09:45 | googleads/google-ads-dotnet | https://api.github.com/repos/googleads/google-ads-dotnet | closed | Provide a custom compilation of v15.0.0 with lowered dependency on Microsoft.Extensions.* | usability dependencies | When we moved from v13.02 to v14.0.0, we upgraded our Microsoft.Extensions.* dependencies from v3.1.9 to v6.0.0. This was done because v3.1.9 was a very old dependency. However, this has made things difficult for some of our users when upgrading to v13 of the API.
So we are using this issue to distribute a custom set of Nuget packages with lowered dependency requirements as a temporary workaround. Use this to complete your v13 upgrade, then continue to upgrade the rest of your code to use v6.0.0 or above of Microsoft.Extensions.*.
To use these dependencies,
1. Download the for-release.zip and extract it to a local folder.
2. Remove your project's dependency on Google.Ads.GoogleAds.
3. Add a local package source in Visual Studio: https://learn.microsoft.com/en-us/nuget/consume-packages/install-use-packages-visual-studio#package-sources
4. Install Google.Ads.GoogleAds v15.0.0 from the local folder source.
[for-release.zip](https://github.com/googleads/google-ads-dotnet/files/10984175/for-release.zip)
| True | Provide a custom compilation of v15.0.0 with lowered dependency on Microsoft.Extensions.* - When we moved from v13.02 to v14.0.0, we upgraded our Microsoft.Extensions.* dependencies from v3.1.9 to v6.0.0. This was done because v3.1.9 was a very old dependency. However, this has made things difficult for some of our users when upgrading to v13 of the API.
So we are using this issue to distribute a custom set of Nuget packages with lowered dependency requirements as a temporary workaround. Use this to complete your v13 upgrade, then continue to upgrade the rest of your code to use v6.0.0 or above of Microsoft.Extensions.*.
To use these dependencies,
1. Download the for-release.zip and extract it to a local folder.
2. Remove your project's dependency on Google.Ads.GoogleAds.
3. Add a local package source in Visual Studio: https://learn.microsoft.com/en-us/nuget/consume-packages/install-use-packages-visual-studio#package-sources
4. Install Google.Ads.GoogleAds v15.0.0 from the local folder source.
[for-release.zip](https://github.com/googleads/google-ads-dotnet/files/10984175/for-release.zip)
| usab | provide a custom compilation of with lowered dependency on microsoft extensions when we moved from to we upgraded our microsoft extensions dependencies from to this was done because was a very old dependency however this has made things difficult for some of our users when upgrading to of the api so we are using this issue to distribute a custom set of nuget packages with lowered dependency requirements as a temporary workaround use this to complete your upgrade then continue to upgrade the rest of your code to use or above of microsoft extensions to use these dependencies download the for release zip and extract it to a local folder remove your project s dependency on google ads googleads add a local package source in visual studio install google ads googleads from the local folder source | 1 |
11,630 | 7,330,864,810 | IssuesEvent | 2018-03-05 11:24:56 | nerdalize/nerd | https://api.github.com/repos/nerdalize/nerd | closed | Fail the command when too many arguments are given | P0 usability | Currently when a command is given too many argument it just takes the first. Instead we want to fail because we're running something that the user didn't intend. Exception is the `nerd job run` | True | Fail the command when too many arguments are given - Currently when a command is given too many argument it just takes the first. Instead we want to fail because we're running something that the user didn't intend. Exception is the `nerd job run` | usab | fail the command when too many arguments are given currently when a command is given too many argument it just takes the first instead we want to fail because we re running something that the user didn t intend exception is the nerd job run | 1 |
260,215 | 22,601,353,561 | IssuesEvent | 2022-06-29 09:24:32 | JanusGraph/janusgraph-dotnet | https://api.github.com/repos/JanusGraph/janusgraph-dotnet | opened | Migrate to new official testcontainers-dotnet library | area/testing | The version we just migrated to in #96 is now the official .NET version of Testcontainers: testcontainers/testcontainers-dotnet#476.
The migration should therefore be straightforward. | 1.0 | Migrate to new official testcontainers-dotnet library - The version we just migrated to in #96 is now the official .NET version of Testcontainers: testcontainers/testcontainers-dotnet#476.
The migration should therefore be straightforward. | non_usab | migrate to new official testcontainers dotnet library the version we just migrated to in is now the official net version of testcontainers testcontainers testcontainers dotnet the migration should therefore be straightforward | 0 |
10,060 | 6,552,280,467 | IssuesEvent | 2017-09-05 17:41:16 | gluster/glusterfs | https://api.github.com/repos/gluster/glusterfs | closed | Gluster threads should be named | CB: libglusterfs FA: Debug-ability & Quality FA: Usability & Supportability Type:Enhancement | A gluster process has many threads and being able to see the list of threads with the cpu consumption is a very useful feature. | True | Gluster threads should be named - A gluster process has many threads and being able to see the list of threads with the cpu consumption is a very useful feature. | usab | gluster threads should be named a gluster process has many threads and being able to see the list of threads with the cpu consumption is a very useful feature | 1 |
21,924 | 10,699,780,129 | IssuesEvent | 2019-10-23 21:47:49 | aspnet/AspNetCore | https://api.github.com/repos/aspnet/AspNetCore | closed | AllowAnonymousAttribute not working as expected with CORS request. | area-security | I'm Running ASP.NET Core 3 (Azure Web App) and have a CORS policy set up like so, in Startup.cs:
services.AddCors(options =>
{
options.AddPolicy("TestCors",
builder =>
builder.WithOrigins(allowedDomains.ToArray())
.AllowAnyHeader()
.AllowAnyMethod()
.AllowCredentials());
});
I have an AJAX endpoint in a controller defined as such:
[EnableCors("TestCors")]
[AllowAnonymous]
public async Task<IActionResult> MyAjaxEndpoint(int parm1)
{
}
When I attempt to call this controller method using JQuery AJAX method like so:
$.ajax(
{
url: 'otherdomain.com/MyAjaxEndpoint',
cache: false,
type: "POST",
crossDomain: true,
data: {parm1: 0},
xhrFields:
{
withCredentials: true
},
async: false
});
, I get a 302 redirect to the login page. This does not happen if I happen to be authenticated through the standard Identity cookie.
I have some tokens (JWT) stored in a http-only cookie that I want sent with AJAX requests, and I was under the impression that requirement necessitated `withCredentials: true` in the AJAX options (is that correct?). I'm not sure if that's what's causing the problem.
At any rate, I assumed that `[AllowAnonymous]` would have bypassed any authentication requirements by the ASP.NET pipeline.
Is there an issue with the AllowAnonymousAttribute or am I (ab)using it wrong? | True | AllowAnonymousAttribute not working as expected with CORS request. - I'm Running ASP.NET Core 3 (Azure Web App) and have a CORS policy set up like so, in Startup.cs:
services.AddCors(options =>
{
options.AddPolicy("TestCors",
builder =>
builder.WithOrigins(allowedDomains.ToArray())
.AllowAnyHeader()
.AllowAnyMethod()
.AllowCredentials());
});
I have an AJAX endpoint in a controller defined as such:
[EnableCors("TestCors")]
[AllowAnonymous]
public async Task<IActionResult> MyAjaxEndpoint(int parm1)
{
}
When I attempt to call this controller method using JQuery AJAX method like so:
$.ajax(
{
url: 'otherdomain.com/MyAjaxEndpoint',
cache: false,
type: "POST",
crossDomain: true,
data: {parm1: 0},
xhrFields:
{
withCredentials: true
},
async: false
});
, I get a 302 redirect to the login page. This does not happen if I happen to be authenticated through the standard Identity cookie.
I have some tokens (JWT) stored in a http-only cookie that I want sent with AJAX requests, and I was under the impression that requirement necessitated `withCredentials: true` in the AJAX options (is that correct?). I'm not sure if that's what's causing the problem.
At any rate, I assumed that `[AllowAnonymous]` would have bypassed any authentication requirements by the ASP.NET pipeline.
Is there an issue with the AllowAnonymousAttribute or am I (ab)using it wrong? | non_usab | allowanonymousattribute not working as expected with cors request i m running asp net core azure web app and have a cors policy set up like so in startup cs services addcors options options addpolicy testcors builder builder withorigins alloweddomains toarray allowanyheader allowanymethod allowcredentials i have an ajax endpoint in a controller defined as such public async task myajaxendpoint int when i attempt to call this controller method using jquery ajax method like so ajax url otherdomain com myajaxendpoint cache false type post crossdomain true data xhrfields withcredentials true async false i get a redirect to the login page this does not happen if i happen to be authenticated through the standard identity cookie i have some tokens jwt stored in a http only cookie that i want sent with ajax requests and i was under the impression that requirement necessitated withcredentials true in the ajax options is that correct i m not sure if that s what s causing the problem at any rate i assumed that would have bypassed any authentication requirements by the asp net pipeline is there an issue with the allowanonymousattribute or am i ab using it wrong | 0 |
665,096 | 22,299,386,705 | IssuesEvent | 2022-06-13 07:13:29 | openforis/arena | https://api.github.com/repos/openforis/arena | closed | Chain: Sampling design - new elements | Priority_1 | 1. weight, always TRUE
- Arena UI: no actions
- If "Sampling design" set in Arena, always weighted
- If no weighting, set all base units' weights to 1
2. clustering: Clustered / No-clustered
- Arena UI: no actions
- if clustering entity selected, then TRUE
- a clustering entity can be any parent entity for the base unit entity
3. sampling strategy:
- Arena: add a new dropdown list ('samplingStrategy')
0. --not specified-- (default)
1. SIMPLE RANDOM SAMPLING
2. SYSTEMATIC SAMPLING
3. STRATIFIED RANDOM SAMPLING
4. STRATIFIED SYSTEMATIC SAMPLING
5. DOUBLE PHASE SAMPLING, requires 1 phase table (coming later) * in Arena: hide still
4. stratification:
- Arena UI: move under "sampling strategy"
- if method 1 or 2, hide/inenable dropdown
- if method 3, 4, or 5, show dropdown list to select 'stratumAttribute', -
- methods 3 and 4 must have a selection, method 5 can be --not specified--
- 'stratumAttribute' can be any categorical attribute of the base unit entity or its parents
NOTE: 'stratumAttribute' can also be an active computed categorical result variable
5. post-stratification
- Arena UI: add a new dropdown list, default: --not specified-
- post-stratification can be applied with methods 1-4
- select a categorical attribute ('post_stratumAttribute') from the dropdown list.
This can be any categorical attribute of the base unit entity or its parents, but not the same as 'stratumAttribute'
NOTE: 'post_stratumAttribute' can also be an active computed categorical result variable
6. p-value <- REQUEST CANCELLED!! (This will be added to another page)
- Arena UI: add a new dropdown list ('arena_p_value')
- list items: (0.99, 0.98, 0.95, 0.90, 0.80). Default: 0.95
7. submit new information into json: 'samplingStrategy' (index), 'post_stratumAttribute' (name), 'arena_p_value' (numeric)
______________________
NEW EDIT:
8. Add a check box with label text: "Nonresponse bias correction" (default FALSE)
9. Add parameter "nonresponse_correction" (FALSE/TRUE) into the json file.
| 1.0 | Chain: Sampling design - new elements - 1. weight, always TRUE
- Arena UI: no actions
- If "Sampling design" set in Arena, always weighted
- If no weighting, set all base units' weights to 1
2. clustering: Clustered / No-clustered
- Arena UI: no actions
- if clustering entity selected, then TRUE
- a clustering entity can be any parent entity for the base unit entity
3. sampling strategy:
- Arena: add a new dropdown list ('samplingStrategy')
0. --not specified-- (default)
1. SIMPLE RANDOM SAMPLING
2. SYSTEMATIC SAMPLING
3. STRATIFIED RANDOM SAMPLING
4. STRATIFIED SYSTEMATIC SAMPLING
5. DOUBLE PHASE SAMPLING, requires 1 phase table (coming later) * in Arena: hide still
4. stratification:
- Arena UI: move under "sampling strategy"
- if method 1 or 2, hide/inenable dropdown
- if method 3, 4, or 5, show dropdown list to select 'stratumAttribute', -
- methods 3 and 4 must have a selection, method 5 can be --not specified--
- 'stratumAttribute' can be any categorical attribute of the base unit entity or its parents
NOTE: 'stratumAttribute' can also be an active computed categorical result variable
5. post-stratification
- Arena UI: add a new dropdown list, default: --not specified-
- post-stratification can be applied with methods 1-4
- select a categorical attribute ('post_stratumAttribute') from the dropdown list.
This can be any categorical attribute of the base unit entity or its parents, but not the same as 'stratumAttribute'
NOTE: 'post_stratumAttribute' can also be an active computed categorical result variable
6. p-value <- REQUEST CANCELLED!! (This will be added to another page)
- Arena UI: add a new dropdown list ('arena_p_value')
- list items: (0.99, 0.98, 0.95, 0.90, 0.80). Default: 0.95
7. submit new information into json: 'samplingStrategy' (index), 'post_stratumAttribute' (name), 'arena_p_value' (numeric)
______________________
NEW EDIT:
8. Add a check box with label text: "Nonresponse bias correction" (default FALSE)
9. Add parameter "nonresponse_correction" (FALSE/TRUE) into the json file.
| non_usab | chain sampling design new elements weight always true arena ui no actions if sampling design set in arena always weighted if no weighting set all base units weights to clustering clustered no clustered arena ui no actions if clustering entity selected then true a clustering entity can be any parent entity for the base unit entity sampling strategy arena add a new dropdown list samplingstrategy not specified default simple random sampling systematic sampling stratified random sampling stratified systematic sampling double phase sampling requires phase table coming later in arena hide still stratification arena ui move under sampling strategy if method or hide inenable dropdown if method or show dropdown list to select stratumattribute methods and must have a selection method can be not specified stratumattribute can be any categorical attribute of the base unit entity or its parents note stratumattribute can also be an active computed categorical result variable post stratification arena ui add a new dropdown list default not specified post stratification can be applied with methods select a categorical attribute post stratumattribute from the dropdown list this can be any categorical attribute of the base unit entity or its parents but not the same as stratumattribute note post stratumattribute can also be an active computed categorical result variable p value request cancelled this will be added to another page arena ui add a new dropdown list arena p value list items default submit new information into json samplingstrategy index post stratumattribute name arena p value numeric new edit add a check box with label text nonresponse bias correction default false add parameter nonresponse correction false true into the json file | 0 |
2,889 | 3,952,480,630 | IssuesEvent | 2016-04-29 09:01:10 | ScottLogic/StockFlux | https://api.github.com/repos/ScottLogic/StockFlux | closed | Creating a contributing document | infrastructure | We should have a brief document that describes how to contribute, as per: https://github.com/blog/1184-contributing-guidelines
It should detail the following:
- [ ] Which issue labels we use and what they mean (removing or adding as necessary)
- [ ] Our approach to branching (removing any dead branches)
- [ ] How we develop / test / release
This can be a very brief document, but just needs putting in writing! | 1.0 | Creating a contributing document - We should have a brief document that describes how to contribute, as per: https://github.com/blog/1184-contributing-guidelines
It should detail the following:
- [ ] Which issue labels we use and what they mean (removing or adding as necessary)
- [ ] Our approach to branching (removing any dead branches)
- [ ] How we develop / test / release
This can be a very brief document, but just needs putting in writing! | non_usab | creating a contributing document we should have a brief document that describes how to contribute as per it should detail the following which issue labels we use and what they mean removing or adding as necessary our approach to branching removing any dead branches how we develop test release this can be a very brief document but just needs putting in writing | 0 |
129,026 | 5,082,318,255 | IssuesEvent | 2016-12-29 15:04:48 | wireservice/csvkit | https://api.github.com/repos/wireservice/csvkit | opened | csvlook: 'NoneType' object has no attribute 'groupdict' | bug High Priority | Using test file from #631. Also reported via Twitter by another user.
```
Traceback (most recent call last):
File "/Users/cgroskopf/.virtualenvs/csvkit/bin/csvlook", line 9, in <module>
load_entry_point('csvkit==1.0.1', 'console_scripts', 'csvlook')()
File "/Users/cgroskopf/src/csvkit/csvkit/utilities/csvlook.py", line 44, in launch_new_instance
utility.run()
File "/Users/cgroskopf/src/csvkit/csvkit/cli.py", line 113, in run
self.main()
File "/Users/cgroskopf/src/csvkit/csvkit/utilities/csvlook.py", line 31, in main
**self.reader_kwargs
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/table/from_csv.py", line 88, in from_csv
return Table(rows, column_names, column_types, row_names=row_names)
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/table/__init__.py", line 132, in __init__
self._column_types = column_types.run(rows, self._column_names)
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/type_tester.py", line 110, in run
if len(row) > i and not column_type.test(row[i]):
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/data_types/base.py", line 29, in test
self.cast(d)
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/data_types/time_delta.py", line 35, in cast
seconds = pytimeparse.parse(d)
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/pytimeparse/timeparse.py", line 147, in timeparse
sign = -1 if match.groupdict()['sign'] == '-' else 1
AttributeError: 'NoneType' object has no attribute 'groupdict'
``` | 1.0 | csvlook: 'NoneType' object has no attribute 'groupdict' - Using test file from #631. Also reported via Twitter by another user.
```
Traceback (most recent call last):
File "/Users/cgroskopf/.virtualenvs/csvkit/bin/csvlook", line 9, in <module>
load_entry_point('csvkit==1.0.1', 'console_scripts', 'csvlook')()
File "/Users/cgroskopf/src/csvkit/csvkit/utilities/csvlook.py", line 44, in launch_new_instance
utility.run()
File "/Users/cgroskopf/src/csvkit/csvkit/cli.py", line 113, in run
self.main()
File "/Users/cgroskopf/src/csvkit/csvkit/utilities/csvlook.py", line 31, in main
**self.reader_kwargs
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/table/from_csv.py", line 88, in from_csv
return Table(rows, column_names, column_types, row_names=row_names)
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/table/__init__.py", line 132, in __init__
self._column_types = column_types.run(rows, self._column_names)
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/type_tester.py", line 110, in run
if len(row) > i and not column_type.test(row[i]):
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/data_types/base.py", line 29, in test
self.cast(d)
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/data_types/time_delta.py", line 35, in cast
seconds = pytimeparse.parse(d)
File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/pytimeparse/timeparse.py", line 147, in timeparse
sign = -1 if match.groupdict()['sign'] == '-' else 1
AttributeError: 'NoneType' object has no attribute 'groupdict'
``` | non_usab | csvlook nonetype object has no attribute groupdict using test file from also reported via twitter by another user traceback most recent call last file users cgroskopf virtualenvs csvkit bin csvlook line in load entry point csvkit console scripts csvlook file users cgroskopf src csvkit csvkit utilities csvlook py line in launch new instance utility run file users cgroskopf src csvkit csvkit cli py line in run self main file users cgroskopf src csvkit csvkit utilities csvlook py line in main self reader kwargs file users cgroskopf virtualenvs csvkit lib site packages agate table from csv py line in from csv return table rows column names column types row names row names file users cgroskopf virtualenvs csvkit lib site packages agate table init py line in init self column types column types run rows self column names file users cgroskopf virtualenvs csvkit lib site packages agate type tester py line in run if len row i and not column type test row file users cgroskopf virtualenvs csvkit lib site packages agate data types base py line in test self cast d file users cgroskopf virtualenvs csvkit lib site packages agate data types time delta py line in cast seconds pytimeparse parse d file users cgroskopf virtualenvs csvkit lib site packages pytimeparse timeparse py line in timeparse sign if match groupdict else attributeerror nonetype object has no attribute groupdict | 0 |
22,470 | 19,489,894,867 | IssuesEvent | 2021-12-27 03:11:57 | Leafwing-Studios/leafwing-input-manager | https://api.github.com/repos/Leafwing-Studios/leafwing-input-manager | closed | Add derive macro for Actionlike | usability | ## Which feature is frustrating to use or confusing?
Actionlike needs to manually implemented. Manual impls always suck when they don't contain any logic.
## Expectation
Derive macro | True | Add derive macro for Actionlike - ## Which feature is frustrating to use or confusing?
Actionlike needs to manually implemented. Manual impls always suck when they don't contain any logic.
## Expectation
Derive macro | usab | add derive macro for actionlike which feature is frustrating to use or confusing actionlike needs to manually implemented manual impls always suck when they don t contain any logic expectation derive macro | 1 |
208,242 | 7,137,472,793 | IssuesEvent | 2018-01-23 11:07:23 | openshift/origin | https://api.github.com/repos/openshift/origin | closed | Cannot oc set image/env for CronJob | kind/bug priority/P2 | Hi,
I can get a scheduledJob working using the example in https://docs.openshift.com/container-platform/3.3/dev_guide/scheduled_jobs.html, but I am not able to use an IS in current project unless I specify the absolute image path in the format <docker_registry_service_ip>:5000/\<project\>/<image_name>. Note I cannot even use the public docker registry host name. This is different from how image can be specified in other resource types such as DeploymentConfig. It makes the template containing this resource not portable.
##### Version
```
C:\Users\x>oc version
oc v1.5.0-alpha.2+e4b43ee
kubernetes v1.5.2+43a9be4
features: Basic-Auth
Server https://<console_url>:8443
openshift v3.3.1.7
kubernetes v1.3.0+52492b4
```
| 1.0 | Cannot oc set image/env for CronJob - Hi,
I can get a scheduledJob working using the example in https://docs.openshift.com/container-platform/3.3/dev_guide/scheduled_jobs.html, but I am not able to use an IS in current project unless I specify the absolute image path in the format <docker_registry_service_ip>:5000/\<project\>/<image_name>. Note I cannot even use the public docker registry host name. This is different from how image can be specified in other resource types such as DeploymentConfig. It makes the template containing this resource not portable.
##### Version
```
C:\Users\x>oc version
oc v1.5.0-alpha.2+e4b43ee
kubernetes v1.5.2+43a9be4
features: Basic-Auth
Server https://<console_url>:8443
openshift v3.3.1.7
kubernetes v1.3.0+52492b4
```
| non_usab | cannot oc set image env for cronjob hi i can get a scheduledjob working using the example in but i am not able to use an is in current project unless i specify the absolute image path in the format note i cannot even use the public docker registry host name this is different from how image can be specified in other resource types such as deploymentconfig it makes the template containing this resource not portable version c users x oc version oc alpha kubernetes features basic auth server openshift kubernetes | 0 |
28,074 | 31,561,308,613 | IssuesEvent | 2023-09-03 09:38:01 | git-cola/git-cola | https://api.github.com/repos/git-cola/git-cola | closed | UI is broken in macOS dark mode | usability macos qt bug | On macOS Mojave with the new dark theme, Cola seems to not pick up the text and background colors properly, it is probably overriding one without overriding the other.
I'm attaching two screenshots. In one nothing is visible, in the other I have selected some text elements to show that they are indeed there.
## Steps to reproduce:
1. Enable dark theme in macOS Mojave (Settings->General)
2. Open git cola in a modified git directory (must be opened after theme is selected)
3. Notice that text (diff, filenames, commit headline, commit body) is invisible
## Expected Result
Text is visible (either white-on-black or black-on-white)
## Screenshots
<img width="1265" alt="screenshot 2018-11-14 at 01 36 07 pm" src="https://user-images.githubusercontent.com/3298/48504446-c2335700-e812-11e8-8628-305c624ac0b1.png">
<img width="1265" alt="screenshot 2018-11-14 at 01 36 26 pm" src="https://user-images.githubusercontent.com/3298/48504452-c8293800-e812-11e8-802d-7c0c75827d24.png">
| True | UI is broken in macOS dark mode - On macOS Mojave with the new dark theme, Cola seems to not pick up the text and background colors properly, it is probably overriding one without overriding the other.
I'm attaching two screenshots. In one nothing is visible, in the other I have selected some text elements to show that they are indeed there.
## Steps to reproduce:
1. Enable dark theme in macOS Mojave (Settings->General)
2. Open git cola in a modified git directory (must be opened after theme is selected)
3. Notice that text (diff, filenames, commit headline, commit body) is invisible
## Expected Result
Text is visible (either white-on-black or black-on-white)
## Screenshots
<img width="1265" alt="screenshot 2018-11-14 at 01 36 07 pm" src="https://user-images.githubusercontent.com/3298/48504446-c2335700-e812-11e8-8628-305c624ac0b1.png">
<img width="1265" alt="screenshot 2018-11-14 at 01 36 26 pm" src="https://user-images.githubusercontent.com/3298/48504452-c8293800-e812-11e8-802d-7c0c75827d24.png">
| usab | ui is broken in macos dark mode on macos mojave with the new dark theme cola seems to not pick up the text and background colors properly it is probably overriding one without overriding the other i m attaching two screenshots in one nothing is visible in the other i have selected some text elements to show that they are indeed there steps to reproduce enable dark theme in macos mojave settings general open git cola in a modified git directory must be opened after theme is selected notice that text diff filenames commit headline commit body is invisible expected result text is visible either white on black or black on white screenshots img width alt screenshot at pm src img width alt screenshot at pm src | 1 |
10,160 | 6,608,328,081 | IssuesEvent | 2017-09-19 10:31:42 | zcash/zcash | https://api.github.com/repos/zcash/zcash | closed | Benchmark wallet load time in realistic and/or worst cases. | benchmarking performance usability | Anecdotally the Zcash wallet load time is frustrating for users. I've heard at least two occasions of operators changing their behavior to work around having to restart due to the downtime.
Let's benchmark wallet loading. | True | Benchmark wallet load time in realistic and/or worst cases. - Anecdotally the Zcash wallet load time is frustrating for users. I've heard at least two occasions of operators changing their behavior to work around having to restart due to the downtime.
Let's benchmark wallet loading. | usab | benchmark wallet load time in realistic and or worst cases anecdotally the zcash wallet load time is frustrating for users i ve heard at least two occasions of operators changing their behavior to work around having to restart due to the downtime let s benchmark wallet loading | 1 |
26,450 | 26,874,714,006 | IssuesEvent | 2023-02-04 22:22:43 | bevyengine/bevy | https://api.github.com/repos/bevyengine/bevy | opened | `AssetServer::get_load_state()` returns `NotLoaded` right after a call to `load()` | C-Bug C-Docs A-Assets C-Usability | ## Bevy version
0.9.1
## What you did
Load a bunch of assets at once, which will increase the chance that an asset is not "immediately" loaded, that is there's an observable delay between the call to `AssetServer::load()` and the time the internal async load is started.
## What went wrong
`AssetServer::get_load_state()` returns `NotLoaded` instead of the expected `Loading`.
## Additional information
It looks like a race condition. `AssetServer::load()` enqueues the loading task via the `IoTaskPool`, but doesn't immediately update the `asset_sources` before returning. Later, when the loading task runs, the internal `load_async()` actually updates `asset_sources`. In the meantime, there's a short period of time where calling `AssetServer::get_load_state()` will query `asset_sources` and not find the asset, and therefore return `NotLoaded`.
From a user perspective, I believe the `AssetServer` should return `Loading` immediately after a call to `load()` returned. The fact the internal async loading task didn't yet start is an implementation detail that the user doesn't care about. `NotLoaded` makes it look like the asset is not even scheduled for loading, like the call to `load()` had somehow failed.
See also:
- #6135
- #7479
| True | `AssetServer::get_load_state()` returns `NotLoaded` right after a call to `load()` - ## Bevy version
0.9.1
## What you did
Load a bunch of assets at once, which will increase the chance that an asset is not "immediately" loaded, that is there's an observable delay between the call to `AssetServer::load()` and the time the internal async load is started.
## What went wrong
`AssetServer::get_load_state()` returns `NotLoaded` instead of the expected `Loading`.
## Additional information
It looks like a race condition. `AssetServer::load()` enqueues the loading task via the `IoTaskPool`, but doesn't immediately update the `asset_sources` before returning. Later, when the loading task runs, the internal `load_async()` actually updates `asset_sources`. In the meantime, there's a short period of time where calling `AssetServer::get_load_state()` will query `asset_sources` and not find the asset, and therefore return `NotLoaded`.
From a user perspective, I believe the `AssetServer` should return `Loading` immediately after a call to `load()` returned. The fact the internal async loading task didn't yet start is an implementation detail that the user doesn't care about. `NotLoaded` makes it look like the asset is not even scheduled for loading, like the call to `load()` had somehow failed.
See also:
- #6135
- #7479
| usab | assetserver get load state returns notloaded right after a call to load bevy version what you did load a bunch of assets at once which will increase the chance that an asset is not immediately loaded that is there s an observable delay between the call to assetserver load and the time the internal async load is started what went wrong assetserver get load state returns notloaded instead of the expected loading additional information it looks like a race condition assetserver load enqueues the loading task via the iotaskpool but doesn t immediately update the asset sources before returning later when the loading task runs the internal load async actually updates asset sources in the meantime there s a short period of time where calling assetserver get load state will query asset sources and not find the asset and therefore return notloaded from a user perspective i believe the assetserver should return loading immediately after a call to load returned the fact the internal async loading task didn t yet start is an implementation detail that the user doesn t care about notloaded makes it look like the asset is not even scheduled for loading like the call to load had somehow failed see also | 1 |
621,555 | 19,591,232,457 | IssuesEvent | 2022-01-05 13:13:55 | o3de/o3de | https://api.github.com/repos/o3de/o3de | opened | Feature Request: Add an option to apply rendering "filters" to a camera | needs-triage sig/graphics-audio priority/minor kind/story | **Is your feature request related to a problem? Please describe.**
Currently there are some components available that allow manipulating the way light is displayed (e.g. Look Modification or Display Mapper), but there is no simple similar way of adding some filters that change the way the image is drawn on screen. For example in the Select Color window available for the color related parameters (such as Shape Color on the Box Shape component), there is a color picker that shows an overview of colors that the color points to, dividing the viewed pixels into larger pixels with "averaged out" colors - this gives the effect of a "pixel filter". Similar thing can be noticed on using the debug Heatmap on the Exposure Control component, where adjusting it gives a peculiar look of blue and red bleeding into normal colors depending on the set Minimum/Maximum Exposure values. Please refer to the attached video where the described effects are shown.
**Describe the solution you'd like**
Add a component with some camera filters included (e.g. "pixel filter" similar to the described color picker effect) and a way to add more.
**Describe alternatives you've considered**
Add any way of applying some of the described "filters" to e.g. Camera component directly, or any other related components.
**Additional context**
https://user-images.githubusercontent.com/86952082/148221183-d8f70ac0-f43a-4e77-9aa3-f4b458334083.mp4 | 1.0 | Feature Request: Add an option to apply rendering "filters" to a camera - **Is your feature request related to a problem? Please describe.**
Currently there are some components available that allow manipulating the way light is displayed (e.g. Look Modification or Display Mapper), but there is no simple similar way of adding some filters that change the way the image is drawn on screen. For example in the Select Color window available for the color related parameters (such as Shape Color on the Box Shape component), there is a color picker that shows an overview of colors that the color points to, dividing the viewed pixels into larger pixels with "averaged out" colors - this gives the effect of a "pixel filter". Similar thing can be noticed on using the debug Heatmap on the Exposure Control component, where adjusting it gives a peculiar look of blue and red bleeding into normal colors depending on the set Minimum/Maximum Exposure values. Please refer to the attached video where the described effects are shown.
**Describe the solution you'd like**
Add a component with some camera filters included (e.g. "pixel filter" similar to the described color picker effect) and a way to add more.
**Describe alternatives you've considered**
Add any way of applying some of the described "filters" to e.g. Camera component directly, or any other related components.
**Additional context**
https://user-images.githubusercontent.com/86952082/148221183-d8f70ac0-f43a-4e77-9aa3-f4b458334083.mp4 | non_usab | feature request add an option to apply rendering filters to a camera is your feature request related to a problem please describe currently there are some components available that allow manipulating the way light is displayed e g look modification or display mapper but there is no simple similar way of adding some filters that change the way the image is drawn on screen for example in the select color window available for the color related parameters such as shape color on the box shape component there is a color picker that shows an overview of colors that the color points to dividing the viewed pixels into larger pixels with averaged out colors this gives the effect of a pixel filter similar thing can be noticed on using the debug heatmap on the exposure control component where adjusting it gives a peculiar look of blue and red bleeding into normal colors depending on the set minimum maximum exposure values please refer to the attached video where the described effects are shown describe the solution you d like add a component with some camera filters included e g pixel filter similar to the described color picker effect and a way to add more describe alternatives you ve considered add any way of applying some of the described filters to e g camera component directly or any other related components additional context | 0 |
295,951 | 25,517,026,258 | IssuesEvent | 2022-11-28 17:07:17 | vegaprotocol/frontend-monorepo | https://api.github.com/repos/vegaprotocol/frontend-monorepo | closed | Migrate trading e2e tests to use vegawallet v2 | Trading Testing 🧪 chore | ## The Chore
We need to switch to vegawallet v2, tests need to be updated and verified if no issues are caused by using the new wallet | 1.0 | Migrate trading e2e tests to use vegawallet v2 - ## The Chore
We need to switch to vegawallet v2, tests need to be updated and verified if no issues are caused by using the new wallet | non_usab | migrate trading tests to use vegawallet the chore we need to switch to vegawallet tests need to be updated and verified if no issues are caused by using the new wallet | 0 |
329 | 2,499,836,505 | IssuesEvent | 2015-01-08 06:42:39 | fossology/fossology | https://api.github.com/repos/fossology/fossology | opened | bucket tag link not active | Category: Buckets Component: Rank Component: Tester Priority: Normal Status: New Tracker: Bug | ---
Author Name: **Bob Gobeille**
Original Redmine Issue: 1029, http://www.fossology.org/issues/1029
Original Date: 2012/02/14
Original Assignee: Bob Gobeille
---
The bucket browser has a tag link in the histogram, but the link is not active (it's just text). The idea is to make it easy to tag all the items in a bucket.
| 1.0 | bucket tag link not active - ---
Author Name: **Bob Gobeille**
Original Redmine Issue: 1029, http://www.fossology.org/issues/1029
Original Date: 2012/02/14
Original Assignee: Bob Gobeille
---
The bucket browser has a tag link in the histogram, but the link is not active (it's just text). The idea is to make it easy to tag all the items in a bucket.
| non_usab | bucket tag link not active author name bob gobeille original redmine issue original date original assignee bob gobeille the bucket browser has a tag link in the histogram but the link is not active it s just text the idea is to make it easy to tag all the items in a bucket | 0 |
4,566 | 3,872,475,777 | IssuesEvent | 2016-04-11 14:01:19 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 22542457: Deleting a Photo in Photos app is Unbearably Slow | classification:ui/usability reproducible:always status:open | #### Description
Summary:
Photos takes a very long time to delete photos. After a deletion action is initiated, several seconds can sometimes pass before the UI is updated. This causes issues where the UI is jarringly updated when it is not expected to be updated. Changing selection status (when the UI is actually updated) can cause confusion and unintended deletion (if attempting to delete multiple photos in quick succession).
Rather than photos actually being deleted on the main thread (as seems to be the case), perhaps it would be better to immediately update the UI and execute the deletion in the background. If a failure occurs (hopefully the non-normal case), a message could be shown to the user and the UI could be reverted.
This issue happens in Photos Version 1.1 (330.21.0)
Steps to Reproduce:
+ Open Photos
+ Delete a photo
Expected Results:
Photo is immediately removed from the UI, and scheduled for deletion in the background
Actual Results:
Photos UI takes a very long time (several seconds sometimes) to remove the photo from the interface
-
Product Version: 10.11 Beta (15A279b)
Created: 2015-09-02T18:34:55.374520
Originated: 2015-09-02T14:34:00
Open Radar Link: http://www.openradar.me/22542457 | True | 22542457: Deleting a Photo in Photos app is Unbearably Slow - #### Description
Summary:
Photos takes a very long time to delete photos. After a deletion action is initiated, several seconds can sometimes pass before the UI is updated. This causes issues where the UI is jarringly updated when it is not expected to be updated. Changing selection status (when the UI is actually updated) can cause confusion and unintended deletion (if attempting to delete multiple photos in quick succession).
Rather than photos actually being deleted on the main thread (as seems to be the case), perhaps it would be better to immediately update the UI and execute the deletion in the background. If a failure occurs (hopefully the non-normal case), a message could be shown to the user and the UI could be reverted.
This issue happens in Photos Version 1.1 (330.21.0)
Steps to Reproduce:
+ Open Photos
+ Delete a photo
Expected Results:
Photo is immediately removed from the UI, and scheduled for deletion in the background
Actual Results:
Photos UI takes a very long time (several seconds sometimes) to remove the photo from the interface
-
Product Version: 10.11 Beta (15A279b)
Created: 2015-09-02T18:34:55.374520
Originated: 2015-09-02T14:34:00
Open Radar Link: http://www.openradar.me/22542457 | usab | deleting a photo in photos app is unbearably slow description summary photos takes a very long time to delete photos after a deletion action is initiated several seconds can sometimes pass before the ui is updated this causes issues where the ui is jarringly updated when it is not expected to be updated changing selection status when the ui is actually updated can cause confusion and unintended deletion if attempting to delete multiple photos in quick succession rather than photos actually being deleted on the main thread as seems to be the case perhaps it would be better to immediately update the ui and execute the deletion in the background if a failure occurs hopefully the non normal case a message could be shown to the user and the ui could be reverted this issue happens in photos version steps to reproduce open photos delete a photo expected results photo is immediately removed from the ui and scheduled for deletion in the background actual results photos ui takes a very long time several seconds sometimes to remove the photo from the interface product version beta created originated open radar link | 1 |
8,979 | 7,522,647,259 | IssuesEvent | 2018-04-12 21:08:43 | istio/istio | https://api.github.com/repos/istio/istio | opened | Pilot fetches and distributes JWT public key (to sidecar) | area/security | In 0.7, envoy sidecar need to fetch public key from `JwksURI` (specified as part of JWT filter config). This requires:
* Pilot need to create outbound cluster(s) for JwksURI, which have some conflicts with v2+v1alpha3 (particularly, concept of `ExternalService`)
* In term of performance, each sidecar will need to fetch the keys independently.
An alternative solution is to have pilot fetch the keys and distribute the them to sidecars; sidecars can use the keys directly, without any extra communication to external services.
To make this works, following assumptions/requirements need to be met:
1. Pilot is allowed to talk to external web services. (does pilot sidecar needs to be configured somehow?)
1. Implementation for cache and key rotation in pilot (i.e if key is updated, it should trigger pilot to push new configs to services).
1. Security concerns: is this acceptable if control plane doesn't use mTLS.
Related issues:
* [Proxy support public key in filter config](https://github.com/istio/proxy/issues/1373)
* [Support OpenID Discovery](https://github.com/istio/istio/issues/4719)
* https://github.com/istio/istio/issues/4633
| True | Pilot fetches and distributes JWT public key (to sidecar) - In 0.7, envoy sidecar need to fetch public key from `JwksURI` (specified as part of JWT filter config). This requires:
* Pilot need to create outbound cluster(s) for JwksURI, which have some conflicts with v2+v1alpha3 (particularly, concept of `ExternalService`)
* In term of performance, each sidecar will need to fetch the keys independently.
An alternative solution is to have pilot fetch the keys and distribute the them to sidecars; sidecars can use the keys directly, without any extra communication to external services.
To make this works, following assumptions/requirements need to be met:
1. Pilot is allowed to talk to external web services. (does pilot sidecar needs to be configured somehow?)
1. Implementation for cache and key rotation in pilot (i.e if key is updated, it should trigger pilot to push new configs to services).
1. Security concerns: is this acceptable if control plane doesn't use mTLS.
Related issues:
* [Proxy support public key in filter config](https://github.com/istio/proxy/issues/1373)
* [Support OpenID Discovery](https://github.com/istio/istio/issues/4719)
* https://github.com/istio/istio/issues/4633
| non_usab | pilot fetches and distributes jwt public key to sidecar in envoy sidecar need to fetch public key from jwksuri specified as part of jwt filter config this requires pilot need to create outbound cluster s for jwksuri which have some conflicts with particularly concept of externalservice in term of performance each sidecar will need to fetch the keys independently an alternative solution is to have pilot fetch the keys and distribute the them to sidecars sidecars can use the keys directly without any extra communication to external services to make this works following assumptions requirements need to be met pilot is allowed to talk to external web services does pilot sidecar needs to be configured somehow implementation for cache and key rotation in pilot i e if key is updated it should trigger pilot to push new configs to services security concerns is this acceptable if control plane doesn t use mtls related issues | 0 |
18,767 | 13,203,945,164 | IssuesEvent | 2020-08-14 15:01:51 | ONRR/nrrd | https://api.github.com/repos/ONRR/nrrd | opened | When you select a new toggle on mobile it turns gray instead of blue | Home Page P3: Medium bug usability visual | On both iPhone and Android
 | True | When you select a new toggle on mobile it turns gray instead of blue - On both iPhone and Android
 | usab | when you select a new toggle on mobile it turns gray instead of blue on both iphone and android | 1 |
22,108 | 18,684,143,369 | IssuesEvent | 2021-11-01 10:11:52 | idiotWu/smooth-scrollbar | https://api.github.com/repos/idiotWu/smooth-scrollbar | opened | [Idea/v9] An easier way to customize the scrollbars' appearance | idea usability accepted | ### Discussed in https://github.com/idiotWu/smooth-scrollbar/discussions/399
<div type='discussions-op-text'>
<sup>Originally posted by **blitzve0** October 31, 2021</sup>
Currently if we want to change the styling of the scroll bar thumb and tracks we have to overwrite the default styling from node modules. If basics settings like color transparency tracks color width padding etc could be set through config would be awesome</div> | True | [Idea/v9] An easier way to customize the scrollbars' appearance - ### Discussed in https://github.com/idiotWu/smooth-scrollbar/discussions/399
<div type='discussions-op-text'>
<sup>Originally posted by **blitzve0** October 31, 2021</sup>
Currently if we want to change the styling of the scroll bar thumb and tracks we have to overwrite the default styling from node modules. If basics settings like color transparency tracks color width padding etc could be set through config would be awesome</div> | usab | an easier way to customize the scrollbars appearance discussed in originally posted by october currently if we want to change the styling of the scroll bar thumb and tracks we have to overwrite the default styling from node modules if basics settings like color transparency tracks color width padding etc could be set through config would be awesome | 1 |
861 | 2,643,838,010 | IssuesEvent | 2015-03-12 13:55:37 | TEAMMATES/repo | https://api.github.com/repos/TEAMMATES/repo | closed | Instructor: Add a confirmation dialog to 'delete student' button | a-FaultTolarence a-Usability d.Moderate f.Courses p.Medium s.Accepted t.Enhancement | This action is irreversible. We should get a confirmation first. | True | Instructor: Add a confirmation dialog to 'delete student' button - This action is irreversible. We should get a confirmation first. | usab | instructor add a confirmation dialog to delete student button this action is irreversible we should get a confirmation first | 1 |
315,715 | 9,631,264,840 | IssuesEvent | 2019-05-15 13:53:54 | yjacques940/customerManager | https://api.github.com/repos/yjacques940/customerManager | closed | Points à corriger pour l'épreuve synthèse | High priority bug Épreuve synthèse | - [x] Erreur sur la page d'accueil dans d'autres navigateurs
- [x] Variable session quand on se déconnecte
- [x] Pouvoir modifier son adresse email (user)
- [x] Effacer les champs quand on ajoute un rendez-vous
- [x] Conflits lors des rendez-vous (peut prendre plus d'un rendez-vous en même temps)
- [x] Ne pas pouvoir prendre de rendez-vous pour une date dans le passé
- [ ] Avoir la traduction en anglais dans toutes les pages, titres et textes | 1.0 | Points à corriger pour l'épreuve synthèse - - [x] Erreur sur la page d'accueil dans d'autres navigateurs
- [x] Variable session quand on se déconnecte
- [x] Pouvoir modifier son adresse email (user)
- [x] Effacer les champs quand on ajoute un rendez-vous
- [x] Conflits lors des rendez-vous (peut prendre plus d'un rendez-vous en même temps)
- [x] Ne pas pouvoir prendre de rendez-vous pour une date dans le passé
- [ ] Avoir la traduction en anglais dans toutes les pages, titres et textes | non_usab | points à corriger pour l épreuve synthèse erreur sur la page d accueil dans d autres navigateurs variable session quand on se déconnecte pouvoir modifier son adresse email user effacer les champs quand on ajoute un rendez vous conflits lors des rendez vous peut prendre plus d un rendez vous en même temps ne pas pouvoir prendre de rendez vous pour une date dans le passé avoir la traduction en anglais dans toutes les pages titres et textes | 0 |
586,064 | 17,568,945,639 | IssuesEvent | 2021-08-14 09:07:50 | cpp-lln-lab/CPP_SPM | https://api.github.com/repos/cpp-lln-lab/CPP_SPM | closed | rename output from segmentation to comply to BIDS | priority 2 | To comply with the BIDS specs on derivatives
https://bids-specification.readthedocs.io/en/stable/05-derivatives/03-imaging.html#probabilistic-segmentations
---
Try to comply with the coming BIDS extension proposal on space and mapping
https://docs.google.com/document/d/11gCzXOPUbYyuQx8fErtMO9tnOKC3kTWiL9axWkkILNE/edit#heading=h.mqkmyp254xh6
For example fMRIprep outputs this:
```
sub-<subject_label>_from-T1w_to-MNI152NLin2009cAsym_mode-image_xfm.h5
``` | 1.0 | rename output from segmentation to comply to BIDS - To comply with the BIDS specs on derivatives
https://bids-specification.readthedocs.io/en/stable/05-derivatives/03-imaging.html#probabilistic-segmentations
---
Try to comply with the coming BIDS extension proposal on space and mapping
https://docs.google.com/document/d/11gCzXOPUbYyuQx8fErtMO9tnOKC3kTWiL9axWkkILNE/edit#heading=h.mqkmyp254xh6
For example fMRIprep outputs this:
```
sub-<subject_label>_from-T1w_to-MNI152NLin2009cAsym_mode-image_xfm.h5
``` | non_usab | rename output from segmentation to comply to bids to comply with the bids specs on derivatives try to comply with the coming bids extension proposal on space and mapping for example fmriprep outputs this sub from to mode image xfm | 0 |
22,301 | 18,981,402,250 | IssuesEvent | 2021-11-21 00:04:14 | bevyengine/bevy | https://api.github.com/repos/bevyengine/bevy | closed | Split current `World::clear_entities` into separate methods for clearing entities and resources | A-ECS C-Usability | ## What problem does this solve or what need does it fill?
[World::clear_entities](https://github.com/bevyengine/bevy/blob/f4776f2ec413bb80ff4ee51e04bb57f0a44d9983/crates/bevy_ecs/src/world/mod.rs#L1139) currently wipes out all entities and resources (please correct me if my understanding of storages is horribly flawed).
This is unintuitive, and not as useful for testing (or advanced ECS architectures) as it could be; where treating resources and entities separately is common.
## What solution would you like?
1. Keep the existing behaviour, but rename to `clear_all`.
2. Add `clear_resources`.
3. Add `clear_entities`. | True | Split current `World::clear_entities` into separate methods for clearing entities and resources - ## What problem does this solve or what need does it fill?
[World::clear_entities](https://github.com/bevyengine/bevy/blob/f4776f2ec413bb80ff4ee51e04bb57f0a44d9983/crates/bevy_ecs/src/world/mod.rs#L1139) currently wipes out all entities and resources (please correct me if my understanding of storages is horribly flawed).
This is unintuitive, and not as useful for testing (or advanced ECS architectures) as it could be; where treating resources and entities separately is common.
## What solution would you like?
1. Keep the existing behaviour, but rename to `clear_all`.
2. Add `clear_resources`.
3. Add `clear_entities`. | usab | split current world clear entities into separate methods for clearing entities and resources what problem does this solve or what need does it fill currently wipes out all entities and resources please correct me if my understanding of storages is horribly flawed this is unintuitive and not as useful for testing or advanced ecs architectures as it could be where treating resources and entities separately is common what solution would you like keep the existing behaviour but rename to clear all add clear resources add clear entities | 1 |
5,515 | 3,930,325,810 | IssuesEvent | 2016-04-25 07:29:29 | Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith | https://api.github.com/repos/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith | opened | QA_Water Content_Prerequisites_p1 | Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open | Defect Description :
In the "Water Content" experiment, the minimum requirement to run the experiment is not displayed in the page instead a page or Scrolling should appear providing information on minimum requirement to run this experiment, information like Bandwidth,Device Resolution,Hardware Configuration and Software Required.
Actual Result :
In the "Water Content" experiment, the minimum requirement to run the experiment is not displayed in the page.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith/blob/master/test-cases/integration_test-cases/Water%20Content/Water%20Content_22_Prerequisites_p1.org | True | QA_Water Content_Prerequisites_p1 - Defect Description :
In the "Water Content" experiment, the minimum requirement to run the experiment is not displayed in the page instead a page or Scrolling should appear providing information on minimum requirement to run this experiment, information like Bandwidth,Device Resolution,Hardware Configuration and Software Required.
Actual Result :
In the "Water Content" experiment, the minimum requirement to run the experiment is not displayed in the page.
Environment :
OS: Windows 7, Ubuntu-16.04,Centos-6
Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM
Processor:i5
Test Step Link:
https://github.com/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith/blob/master/test-cases/integration_test-cases/Water%20Content/Water%20Content_22_Prerequisites_p1.org | usab | qa water content prerequisites defect description in the water content experiment the minimum requirement to run the experiment is not displayed in the page instead a page or scrolling should appear providing information on minimum requirement to run this experiment information like bandwidth device resolution hardware configuration and software required actual result in the water content experiment the minimum requirement to run the experiment is not displayed in the page environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link | 1 |
4,447 | 3,869,680,906 | IssuesEvent | 2016-04-10 19:01:20 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 23310964: Step between font sizes is too large | classification:ui/usability reproducible:always status:open | #### Description
I’ve noticed this many times over the past few years, so I don’t think it’s anything new. My gripe is that there is often no font-size setting in iBooks that is neither too big nor too small. It’s not that I want the text to be at some specific size; I just want the increments between sizes to be finer so that I don’t end up with no options that seem acceptable.
I’m attaching a representative page at two adjacent sizes. They look more clearly distinguished on the iPad screen than on a desktop. I’d call the one with the smaller font size readable, though perhaps a bit uncomfortably small with the backlight turned low. The next size up looks like a large-print edition for geriatrics. There’s simply nothing in between, at least with this font (Iowan).
-
Product Version: iOS 9.1
Created: 2015-10-29T07:48:29.279410
Originated: 2015-10-29T00:48:00
Open Radar Link: http://www.openradar.me/23310964 | True | 23310964: Step between font sizes is too large - #### Description
I’ve noticed this many times over the past few years, so I don’t think it’s anything new. My gripe is that there is often no font-size setting in iBooks that is neither too big nor too small. It’s not that I want the text to be at some specific size; I just want the increments between sizes to be finer so that I don’t end up with no options that seem acceptable.
I’m attaching a representative page at two adjacent sizes. They look more clearly distinguished on the iPad screen than on a desktop. I’d call the one with the smaller font size readable, though perhaps a bit uncomfortably small with the backlight turned low. The next size up looks like a large-print edition for geriatrics. There’s simply nothing in between, at least with this font (Iowan).
-
Product Version: iOS 9.1
Created: 2015-10-29T07:48:29.279410
Originated: 2015-10-29T00:48:00
Open Radar Link: http://www.openradar.me/23310964 | usab | step between font sizes is too large description i’ve noticed this many times over the past few years so i don’t think it’s anything new my gripe is that there is often no font size setting in ibooks that is neither too big nor too small it’s not that i want the text to be at some specific size i just want the increments between sizes to be finer so that i don’t end up with no options that seem acceptable i’m attaching a representative page at two adjacent sizes they look more clearly distinguished on the ipad screen than on a desktop i’d call the one with the smaller font size readable though perhaps a bit uncomfortably small with the backlight turned low the next size up looks like a large print edition for geriatrics there’s simply nothing in between at least with this font iowan product version ios created originated open radar link | 1 |
902 | 2,769,975,631 | IssuesEvent | 2015-05-01 09:19:43 | piwik/piwik | https://api.github.com/repos/piwik/piwik | closed | UNION query in findVisitor() is regularly 100ms+, involves temporary tables and scans | c: Performance Task | As noted in #7440, commit 0abac3b312ff9560fb43012192e37c2369bab3ed introduced a UNION into findVisitor() within the visit code (now in core/Tracker/Model.php), which [results in the creation of a temporary table on each call](http://dev.mysql.com/doc/refman/5.5/en/internal-temporary-tables.html) ("UNION queries use temporary tables.").
Separately, the idvisitor query normally takes 1-2ms and the config_id query takes around 0.5ms. The combined query is normally around 5ms. However, the number of rows examined may reach the hundreds, and it then starts taking significantly longer - up to 200ms on a dedicated machine.
The impact of these queries vary depending on their size and may not be visible on test or low-visit production websites.
This is what happened when I switched it to checking just the idvisitor part (see last hour):

[The other temporary tables which are disk tables [appeared more recently](https://github.com/piwik/piwik/issues/7416#issuecomment-81170546)]

When I switched the code around to just use just the select on config_id, the decrease in activity remained. This suggests to me that it is the UNION of the two selects that is causing objects and scans.
An example of a slow query from the log (there are several 100ms+ ones per minute):
````
# Query_time: 0.203814 Lock_time: 0.048132 Rows_sent: 1 Rows_examined: 116
SET timestamp=1426562315;
( SELECT visit_last_action_time, visit_first_action_time, idvisitor, idvisit, user_id, visit_exit_idaction_url, visit_exit_idaction_name, visitor_returning, visitor_days_since_first, visitor_days_since_order, visitor_count_visits, visit_goal_buyer, location_country, location_region, location_city, location_latitude, location_longitude, referer_name, referer_keyword, referer_type, location_ip, location_browser_lang, visit_total_events, visit_total_time, visit_total_actions, visit_total_searches ,
0 as priority
FROM piwik_log_visit
WHERE visit_last_action_time >= '2015-03-17 02:48:34' AND visit_last_action_time <= '2015-03-17 03:48:34' AND idsite = '1' AND config_id = '^Z?^\? D?' AND user_id IS NULL
ORDER BY visit_last_action_time DESC
LIMIT 1
)
UNION
( SELECT visit_last_action_time, visit_first_action_time, idvisitor, idvisit, user_id, visit_exit_idaction_url, visit_exit_idaction_name, visitor_returning, visitor_days_since_first, visitor_days_since_order, visitor_count_visits, visit_goal_buyer, location_country, location_region, location_city, location_latitude, location_longitude, referer_name, referer_keyword, referer_type, location_ip, location_browser_lang, visit_total_events, visit_total_time, visit_total_actions, visit_total_searches ,
1 as priority
FROM piwik_log_visit
WHERE visit_last_action_time >= '2015-03-17 02:48:34' AND visit_last_action_time <= '2015-03-17 03:48:34' AND idsite = '1' AND idvisitor = '^U\"?^^0^M?U'
ORDER BY visit_last_action_time DESC
LIMIT 1
)
ORDER BY priority DESC
LIMIT 1;
````
An EXPLAIN of the query (the last entry involves the temporary table created by the UNION):
````
| id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra |
+----+--------------+-----------------+-------+---------------------------------------------------------------------------+------------------------------+---------+-------------+------+-----------------------------+
| 1 | PRIMARY | piwik_log_visit | range | index_idsite_config_datetime,index_idsite_datetime,index_idsite_idvisitor | index_idsite_config_datetime | 20 | NULL | 1 | Using where |
| 2 | UNION | piwik_log_visit | ref | index_idsite_config_datetime,index_idsite_datetime,index_idsite_idvisitor | index_idsite_idvisitor | 12 | const,const | 1 | Using where; Using filesort |
| NULL | UNION RESULT | <union1,2> | ALL | NULL | NULL | NULL | NULL | NULL | Using filesort |
````
The UNION is intended to check visits against both idvisitor and config_id and return the first if it is present, other than the second. I propose that two queries be performed - check idvisitor first, and if that does not return a result, do the config_id query. | True | UNION query in findVisitor() is regularly 100ms+, involves temporary tables and scans - As noted in #7440, commit 0abac3b312ff9560fb43012192e37c2369bab3ed introduced a UNION into findVisitor() within the visit code (now in core/Tracker/Model.php), which [results in the creation of a temporary table on each call](http://dev.mysql.com/doc/refman/5.5/en/internal-temporary-tables.html) ("UNION queries use temporary tables.").
Separately, the idvisitor query normally takes 1-2ms and the config_id query takes around 0.5ms. The combined query is normally around 5ms. However, the number of rows examined may reach the hundreds, and it then starts taking significantly longer - up to 200ms on a dedicated machine.
The impact of these queries vary depending on their size and may not be visible on test or low-visit production websites.
This is what happened when I switched it to checking just the idvisitor part (see last hour):

[The other temporary tables which are disk tables [appeared more recently](https://github.com/piwik/piwik/issues/7416#issuecomment-81170546)]

When I switched the code around to just use just the select on config_id, the decrease in activity remained. This suggests to me that it is the UNION of the two selects that is causing objects and scans.
An example of a slow query from the log (there are several 100ms+ ones per minute):
````
# Query_time: 0.203814 Lock_time: 0.048132 Rows_sent: 1 Rows_examined: 116
SET timestamp=1426562315;
( SELECT visit_last_action_time, visit_first_action_time, idvisitor, idvisit, user_id, visit_exit_idaction_url, visit_exit_idaction_name, visitor_returning, visitor_days_since_first, visitor_days_since_order, visitor_count_visits, visit_goal_buyer, location_country, location_region, location_city, location_latitude, location_longitude, referer_name, referer_keyword, referer_type, location_ip, location_browser_lang, visit_total_events, visit_total_time, visit_total_actions, visit_total_searches ,
0 as priority
FROM piwik_log_visit
WHERE visit_last_action_time >= '2015-03-17 02:48:34' AND visit_last_action_time <= '2015-03-17 03:48:34' AND idsite = '1' AND config_id = '^Z?^\? D?' AND user_id IS NULL
ORDER BY visit_last_action_time DESC
LIMIT 1
)
UNION
( SELECT visit_last_action_time, visit_first_action_time, idvisitor, idvisit, user_id, visit_exit_idaction_url, visit_exit_idaction_name, visitor_returning, visitor_days_since_first, visitor_days_since_order, visitor_count_visits, visit_goal_buyer, location_country, location_region, location_city, location_latitude, location_longitude, referer_name, referer_keyword, referer_type, location_ip, location_browser_lang, visit_total_events, visit_total_time, visit_total_actions, visit_total_searches ,
1 as priority
FROM piwik_log_visit
WHERE visit_last_action_time >= '2015-03-17 02:48:34' AND visit_last_action_time <= '2015-03-17 03:48:34' AND idsite = '1' AND idvisitor = '^U\"?^^0^M?U'
ORDER BY visit_last_action_time DESC
LIMIT 1
)
ORDER BY priority DESC
LIMIT 1;
````
An EXPLAIN of the query (the last entry involves the temporary table created by the UNION):
````
| id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra |
+----+--------------+-----------------+-------+---------------------------------------------------------------------------+------------------------------+---------+-------------+------+-----------------------------+
| 1 | PRIMARY | piwik_log_visit | range | index_idsite_config_datetime,index_idsite_datetime,index_idsite_idvisitor | index_idsite_config_datetime | 20 | NULL | 1 | Using where |
| 2 | UNION | piwik_log_visit | ref | index_idsite_config_datetime,index_idsite_datetime,index_idsite_idvisitor | index_idsite_idvisitor | 12 | const,const | 1 | Using where; Using filesort |
| NULL | UNION RESULT | <union1,2> | ALL | NULL | NULL | NULL | NULL | NULL | Using filesort |
````
The UNION is intended to check visits against both idvisitor and config_id and return the first if it is present, other than the second. I propose that two queries be performed - check idvisitor first, and if that does not return a result, do the config_id query. | non_usab | union query in findvisitor is regularly involves temporary tables and scans as noted in commit introduced a union into findvisitor within the visit code now in core tracker model php which union queries use temporary tables separately the idvisitor query normally takes and the config id query takes around the combined query is normally around however the number of rows examined may reach the hundreds and it then starts taking significantly longer up to on a dedicated machine the impact of these queries vary depending on their size and may not be visible on test or low visit production websites this is what happened when i switched it to checking just the idvisitor part see last hour when i switched the code around to just use just the select on config id the decrease in activity remained this suggests to me that it is the union of the two selects that is causing objects and scans an example of a slow query from the log there are several ones per minute query time lock time rows sent rows examined set timestamp select visit last action time visit first action time idvisitor idvisit user id visit exit idaction url visit exit idaction name visitor returning visitor days since first visitor days since order visitor count visits visit goal buyer location country location region location city location latitude location longitude referer name referer keyword referer type location ip location browser lang visit total events visit total time visit total actions visit total searches as priority from piwik log visit where visit last action time and visit last action time and idsite and config id z d and user id is null order by visit last action time desc limit union select visit last action time visit first action time idvisitor idvisit user id visit exit idaction url visit exit idaction name visitor returning visitor days since first visitor days since order visitor count visits visit goal buyer location country location region location city location latitude location longitude referer name referer keyword referer type location ip location browser lang visit total events visit total time visit total actions visit total searches as priority from piwik log visit where visit last action time and visit last action time and idsite and idvisitor u m u order by visit last action time desc limit order by priority desc limit an explain of the query the last entry involves the temporary table created by the union id select type table type possible keys key key len ref rows extra primary piwik log visit range index idsite config datetime index idsite datetime index idsite idvisitor index idsite config datetime null using where union piwik log visit ref index idsite config datetime index idsite datetime index idsite idvisitor index idsite idvisitor const const using where using filesort null union result all null null null null null using filesort the union is intended to check visits against both idvisitor and config id and return the first if it is present other than the second i propose that two queries be performed check idvisitor first and if that does not return a result do the config id query | 0 |
507,411 | 14,679,980,206 | IssuesEvent | 2020-12-31 08:40:54 | k8smeetup/website-tasks | https://api.github.com/repos/k8smeetup/website-tasks | opened | /docs/setup/best-practices/_index.md | lang/zh priority/P0 sync/update version/master welcome | Source File: [/docs/setup/best-practices/_index.md](https://github.com/kubernetes/website/blob/master/content/en/docs/setup/best-practices/_index.md)
Diff 命令参考:
```bash
# 查看原始文档与翻译文档更新差异
git diff --no-index -- content/en/docs/setup/best-practices/_index.md content/zh/docs/setup/best-practices/_index.md
# 跨分支持查看原始文档更新差异
git diff release-1.19 master -- content/en/docs/setup/best-practices/_index.md
``` | 1.0 | /docs/setup/best-practices/_index.md - Source File: [/docs/setup/best-practices/_index.md](https://github.com/kubernetes/website/blob/master/content/en/docs/setup/best-practices/_index.md)
Diff 命令参考:
```bash
# 查看原始文档与翻译文档更新差异
git diff --no-index -- content/en/docs/setup/best-practices/_index.md content/zh/docs/setup/best-practices/_index.md
# 跨分支持查看原始文档更新差异
git diff release-1.19 master -- content/en/docs/setup/best-practices/_index.md
``` | non_usab | docs setup best practices index md source file diff 命令参考 bash 查看原始文档与翻译文档更新差异 git diff no index content en docs setup best practices index md content zh docs setup best practices index md 跨分支持查看原始文档更新差异 git diff release master content en docs setup best practices index md | 0 |
115,277 | 14,707,978,789 | IssuesEvent | 2021-01-04 22:38:09 | AlaskaAirlines/auro-table | https://api.github.com/repos/AlaskaAirlines/auro-table | closed | Table: symbolize | Project: Version Zero Type: Design Type: Feature help wanted | # User story
As a designer, I want to be able to reuse the correct table symbol, so I can quickly repeat UIs for my team.
## Outline subtasks or tasks
> List specific steps required to deliver this work.
- [ ] Document the currently developed v1 of the auro-table (see attached image)
- [ ] Add to MASTER branch of UI Kit project
- [ ] Communicate out to designers of update
## Definition of “Done” (exit criteria)
> The story is generally “done” when the user can complete the outlined task, but make sure to define what that is.
This story will be considered complete once the developed asset has been properly documented as a design asset and communicated out to the team.
## Epic
https://github.com/AlaskaAirlines/auro_docs/issues/98
| 1.0 | Table: symbolize - # User story
As a designer, I want to be able to reuse the correct table symbol, so I can quickly repeat UIs for my team.
## Outline subtasks or tasks
> List specific steps required to deliver this work.
- [ ] Document the currently developed v1 of the auro-table (see attached image)
- [ ] Add to MASTER branch of UI Kit project
- [ ] Communicate out to designers of update
## Definition of “Done” (exit criteria)
> The story is generally “done” when the user can complete the outlined task, but make sure to define what that is.
This story will be considered complete once the developed asset has been properly documented as a design asset and communicated out to the team.
## Epic
https://github.com/AlaskaAirlines/auro_docs/issues/98
| non_usab | table symbolize user story as a designer i want to be able to reuse the correct table symbol so i can quickly repeat uis for my team outline subtasks or tasks list specific steps required to deliver this work document the currently developed of the auro table see attached image add to master branch of ui kit project communicate out to designers of update definition of “done” exit criteria the story is generally “done” when the user can complete the outlined task but make sure to define what that is this story will be considered complete once the developed asset has been properly documented as a design asset and communicated out to the team epic | 0 |
19,181 | 13,635,223,381 | IssuesEvent | 2020-09-25 02:15:19 | rubyforgood/circulate | https://api.github.com/repos/rubyforgood/circulate | closed | Public nav should use off-canvas on mobile | :zap: Power :zap: Help Wanted Ruby For Good 🎃 Fall 2020 UX / Usability bug | At a small width, the public navigation breaks pretty badly:
<img width="501" alt="Screen Shot 2020-09-17 at 2 21 38 PM" src="https://user-images.githubusercontent.com/3331/93517641-5f6ce400-f8f1-11ea-9a5e-d14108972a3f.png">
We should be able to use the [off-canvas menu provided by the Spectre](https://picturepan2.github.io/spectre/experimentals/off-canvas.html) to solve this. We're already using that component in the admin UI.
See also #199 . | True | Public nav should use off-canvas on mobile - At a small width, the public navigation breaks pretty badly:
<img width="501" alt="Screen Shot 2020-09-17 at 2 21 38 PM" src="https://user-images.githubusercontent.com/3331/93517641-5f6ce400-f8f1-11ea-9a5e-d14108972a3f.png">
We should be able to use the [off-canvas menu provided by the Spectre](https://picturepan2.github.io/spectre/experimentals/off-canvas.html) to solve this. We're already using that component in the admin UI.
See also #199 . | usab | public nav should use off canvas on mobile at a small width the public navigation breaks pretty badly img width alt screen shot at pm src we should be able to use the to solve this we re already using that component in the admin ui see also | 1 |
408,234 | 11,943,660,294 | IssuesEvent | 2020-04-02 23:57:28 | wc-work/why | https://api.github.com/repos/wc-work/why | closed | Contact Form | High Priority | - [x] Add Contact Form
- [x] Set up form to automatically send W.H.Y. email notification | 1.0 | Contact Form - - [x] Add Contact Form
- [x] Set up form to automatically send W.H.Y. email notification | non_usab | contact form add contact form set up form to automatically send w h y email notification | 0 |
125,802 | 4,964,835,814 | IssuesEvent | 2016-12-04 00:08:13 | kubernetes/kubernetes | https://api.github.com/repos/kubernetes/kubernetes | closed | Pods with PVCs stuck in ContainerCreating after Node upgrade | area/storage area/upgrade priority/P0 release-blocker team/cluster | <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.-->
**Is this a request for help?** (If yes, you should use our troubleshooting guide and community support channels, see http://kubernetes.io/docs/troubleshooting/.):
No.
**What keywords did you search in Kubernetes issues before filing this one?** (If you have found any duplicates, you should instead reply there.):
* upgrade volume
* mount timeout
* ContainerCreating
---
**Is this a BUG REPORT or FEATURE REQUEST?** (choose one):
BUG REPORT
<!--
If this is a BUG REPORT, please:
- Fill in as much of the template below as you can. If you leave out
information, we can't help you as well.
If this is a FEATURE REQUEST, please:
- Describe *in detail* the feature/behavior/change you'd like to see.
In both cases, be ready for followup questions, and please respond in a timely
manner. If we can't reproduce a bug or think a feature already exists, we
might close your issue. If we're wrong, PLEASE feel free to reopen it and
explain why.
-->
**Kubernetes version** (use `kubectl version`):
```
Server Version: version.Info{Major:"1", Minor:"5+", GitVersion:"v1.5.0-beta.2", GitCommit:"0776eab45fe28f02bbeac0f05ae1a203051a21eb", GitTreeState:"clean", BuildDate:"2016-11-24T22:30:23Z", GoVersion:"go1.7.3", Compiler:"gc", Platform:"linux/amd64"}
```
**Environment**:
- **Cloud provider or hardware configuration**: GCE
- **OS** (e.g. from /etc/os-release):
- master:
```
BUILD_ID=8977.0.0
NAME="Container-VM Image"
GOOGLE_CRASH_ID=Lakitu
VERSION_ID=56
BUG_REPORT_URL=https://crbug.com/new
PRETTY_NAME="Google Container-VM Image"
VERSION=56
GOOGLE_METRICS_PRODUCT_ID=26
HOME_URL="https://cloud.google.com/compute/docs/containers/vm-image/"
ID=gci
```
- minion:
```
PRETTY_NAME="Debian GNU/Linux 7 (wheezy)"
NAME="Debian GNU/Linux"
VERSION_ID="7"
VERSION="7 (wheezy)"
ID=debian
ANSI_COLOR="1;31"
HOME_URL="http://www.debian.org/"
SUPPORT_URL="http://www.debian.org/support/"
BUG_REPORT_URL="http://bugs.debian.org/"
```
- **Kernel** (e.g. `uname -a`):
- master: `Linux kubernetes-master 4.4.21+ #1 SMP Thu Nov 10 02:50:15 PST 2016 x86_64 Intel(R) Xeon(R) CPU @ 2.30GHz GenuineIntel GNU/Linux`
- minion: `Linux kubernetes-minion-group-4klk 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u2 (2016-10-19) x86_64 GNU/Linux`
- **Install tools**: `cluster/gce/upgrade.sh`
- **Others**:
**What happened**:
Pods with PVCs get stuck in ContainerCreating after Node gets upgraded. The error is `timeout expired waiting for volumes to attach/mount` as shown below:
```console
$ kubectl describe pod web-1
Name: web-1
Namespace: default
Node: kubernetes-minion-group-54hw/10.128.0.4
Start Time: Mon, 28 Nov 2016 11:52:11 -0800
Labels: app=nginx
Status: Pending
IP:
Controllers: PetSet/web
Containers:
nginx:
Container ID:
Image: gcr.io/google_containers/nginx-slim:0.8
Image ID:
Port: 80/TCP
Requests:
cpu: 100m
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Volume Mounts:
/usr/share/nginx/html from www (rw)
/var/run/secrets/kubernetes.io/serviceaccount from default-token-xim26 (ro)
Environment Variables: <none>
Conditions:
Type Status
Initialized True
Ready False
PodScheduled True
Volumes:
www:
Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace)
ClaimName: www-web-1
ReadOnly: false
default-token-xim26:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-xim26
QoS Class: Burstable
Tolerations: <none>
Events:
FirstSeen LastSeen Count From SubObjectPath Type Reason Message
--------- -------- ----- ---- ------------- -------- ------ -------
40m 40m 1 {default-scheduler } Normal Scheduled Successfully assigned web-1 to kubernetes-minion-group-54hw
40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Pulling pulling image "gcr.io/google_containers/nginx-slim:0.8"
40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Pulled Successfully pulled image "gcr.io/google_containers/nginx-slim:0.8"
40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Created Created container with docker id d2d15140a74a; Security:[seccomp=unconfined]
40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Started Started container with docker id d2d15140a74a
16m 16m 4 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp: lookup kubernetes-master on 169.254.169.254:53: no such host
16m 16m 1 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp 10.128.0.2:443: i/o timeout
15m 15m 3 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp 10.128.0.2:443: getsockopt: connection refused
9m 9m 1 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: network is not ready: [Kubenet does not have netConfig. This is most likely due to lack of PodCIDR]
9m 9m 3 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: network is not ready: [Kubenet does not have netConfig. This is most likely due to lack of PodCIDR]
6m 14s 4 {kubelet kubernetes-minion-group-54hw} Warning FailedMount Unable to mount volumes for pod "web-1_default(270e4f19-b5a4-11e6-8922-42010a800002)": timeout expired waiting for volumes to attach/mount for pod "web-1"/"default". list of unattached/unmounted volumes=[www]
6m 14s 4 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: timeout expired waiting for volumes to attach/mount for pod "web-1"/"default". list of unattached/unmounted volumes=[www]
```
**What you expected to happen**:
Upon Node upgrade, Pods should either be terminated (so a controller can recreate them) or they should come back to Running on their own.
**How to reproduce it** (as minimally and precisely as possible):
1. Start a v1.4 cluster on GCE.
```shell
git checkout release-1.4
git pull upstream release-1.4
make clean quick-release
cluster/kube-up.sh
```
1. Create the following example PetSet and wait for all Pods to be Running.
```shell
kubectl create -f petset.yaml
```
```yaml
# A headless service to create DNS records
apiVersion: v1
kind: Service
metadata:
name: nginx
labels:
app: nginx
spec:
ports:
- port: 80
name: web
# *.nginx.default.svc.cluster.local
clusterIP: None
selector:
app: nginx
---
apiVersion: apps/v1alpha1
kind: PetSet
metadata:
name: web
spec:
serviceName: "nginx"
replicas: 5
template:
metadata:
labels:
app: nginx
annotations:
pod.alpha.kubernetes.io/initialized: "true"
spec:
containers:
- name: nginx
image: gcr.io/google_containers/nginx-slim:0.8
ports:
- containerPort: 80
name: web
volumeMounts:
- name: www
mountPath: /usr/share/nginx/html
volumeClaimTemplates:
- metadata:
name: www
annotations:
volume.alpha.kubernetes.io/storage-class: anything
spec:
accessModes: [ "ReadWriteOnce" ]
resources:
requests:
storage: 1Gi
```
1. Delete PetSet without cascading (keep the Pods).
```shell
kubectl delete petset web --cascade=false
```
1. Upgrade master and nodes to `v1.5.0-beta.2` with `cluster/gce/upgrade.sh`.
```shell
git checkout release-1.5
git pull upstream release-1.5
cluster/gce/upgrade.sh v1.5.0-beta.2
```
1. Check for Pods stuck in ContainerCreating.
```
NAME READY STATUS RESTARTS AGE
web-0 0/1 ContainerCreating 0 42m
web-4 0/1 ContainerCreating 0 42m
```
**Anything else do we need to know**:
After the above steps, some Pods may end up terminating. It seems to be random which ones terminate and which ones get stuck in ContainerCreating.
Manually deleting and recreating all the stuck Pods after the Nodes are upgraded allows them to correctly attach to the preexisting PVCs and become Running.
| 1.0 | Pods with PVCs stuck in ContainerCreating after Node upgrade - <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.-->
**Is this a request for help?** (If yes, you should use our troubleshooting guide and community support channels, see http://kubernetes.io/docs/troubleshooting/.):
No.
**What keywords did you search in Kubernetes issues before filing this one?** (If you have found any duplicates, you should instead reply there.):
* upgrade volume
* mount timeout
* ContainerCreating
---
**Is this a BUG REPORT or FEATURE REQUEST?** (choose one):
BUG REPORT
<!--
If this is a BUG REPORT, please:
- Fill in as much of the template below as you can. If you leave out
information, we can't help you as well.
If this is a FEATURE REQUEST, please:
- Describe *in detail* the feature/behavior/change you'd like to see.
In both cases, be ready for followup questions, and please respond in a timely
manner. If we can't reproduce a bug or think a feature already exists, we
might close your issue. If we're wrong, PLEASE feel free to reopen it and
explain why.
-->
**Kubernetes version** (use `kubectl version`):
```
Server Version: version.Info{Major:"1", Minor:"5+", GitVersion:"v1.5.0-beta.2", GitCommit:"0776eab45fe28f02bbeac0f05ae1a203051a21eb", GitTreeState:"clean", BuildDate:"2016-11-24T22:30:23Z", GoVersion:"go1.7.3", Compiler:"gc", Platform:"linux/amd64"}
```
**Environment**:
- **Cloud provider or hardware configuration**: GCE
- **OS** (e.g. from /etc/os-release):
- master:
```
BUILD_ID=8977.0.0
NAME="Container-VM Image"
GOOGLE_CRASH_ID=Lakitu
VERSION_ID=56
BUG_REPORT_URL=https://crbug.com/new
PRETTY_NAME="Google Container-VM Image"
VERSION=56
GOOGLE_METRICS_PRODUCT_ID=26
HOME_URL="https://cloud.google.com/compute/docs/containers/vm-image/"
ID=gci
```
- minion:
```
PRETTY_NAME="Debian GNU/Linux 7 (wheezy)"
NAME="Debian GNU/Linux"
VERSION_ID="7"
VERSION="7 (wheezy)"
ID=debian
ANSI_COLOR="1;31"
HOME_URL="http://www.debian.org/"
SUPPORT_URL="http://www.debian.org/support/"
BUG_REPORT_URL="http://bugs.debian.org/"
```
- **Kernel** (e.g. `uname -a`):
- master: `Linux kubernetes-master 4.4.21+ #1 SMP Thu Nov 10 02:50:15 PST 2016 x86_64 Intel(R) Xeon(R) CPU @ 2.30GHz GenuineIntel GNU/Linux`
- minion: `Linux kubernetes-minion-group-4klk 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u2 (2016-10-19) x86_64 GNU/Linux`
- **Install tools**: `cluster/gce/upgrade.sh`
- **Others**:
**What happened**:
Pods with PVCs get stuck in ContainerCreating after Node gets upgraded. The error is `timeout expired waiting for volumes to attach/mount` as shown below:
```console
$ kubectl describe pod web-1
Name: web-1
Namespace: default
Node: kubernetes-minion-group-54hw/10.128.0.4
Start Time: Mon, 28 Nov 2016 11:52:11 -0800
Labels: app=nginx
Status: Pending
IP:
Controllers: PetSet/web
Containers:
nginx:
Container ID:
Image: gcr.io/google_containers/nginx-slim:0.8
Image ID:
Port: 80/TCP
Requests:
cpu: 100m
State: Waiting
Reason: ContainerCreating
Ready: False
Restart Count: 0
Volume Mounts:
/usr/share/nginx/html from www (rw)
/var/run/secrets/kubernetes.io/serviceaccount from default-token-xim26 (ro)
Environment Variables: <none>
Conditions:
Type Status
Initialized True
Ready False
PodScheduled True
Volumes:
www:
Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace)
ClaimName: www-web-1
ReadOnly: false
default-token-xim26:
Type: Secret (a volume populated by a Secret)
SecretName: default-token-xim26
QoS Class: Burstable
Tolerations: <none>
Events:
FirstSeen LastSeen Count From SubObjectPath Type Reason Message
--------- -------- ----- ---- ------------- -------- ------ -------
40m 40m 1 {default-scheduler } Normal Scheduled Successfully assigned web-1 to kubernetes-minion-group-54hw
40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Pulling pulling image "gcr.io/google_containers/nginx-slim:0.8"
40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Pulled Successfully pulled image "gcr.io/google_containers/nginx-slim:0.8"
40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Created Created container with docker id d2d15140a74a; Security:[seccomp=unconfined]
40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Started Started container with docker id d2d15140a74a
16m 16m 4 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp: lookup kubernetes-master on 169.254.169.254:53: no such host
16m 16m 1 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp 10.128.0.2:443: i/o timeout
15m 15m 3 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp 10.128.0.2:443: getsockopt: connection refused
9m 9m 1 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: network is not ready: [Kubenet does not have netConfig. This is most likely due to lack of PodCIDR]
9m 9m 3 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: network is not ready: [Kubenet does not have netConfig. This is most likely due to lack of PodCIDR]
6m 14s 4 {kubelet kubernetes-minion-group-54hw} Warning FailedMount Unable to mount volumes for pod "web-1_default(270e4f19-b5a4-11e6-8922-42010a800002)": timeout expired waiting for volumes to attach/mount for pod "web-1"/"default". list of unattached/unmounted volumes=[www]
6m 14s 4 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: timeout expired waiting for volumes to attach/mount for pod "web-1"/"default". list of unattached/unmounted volumes=[www]
```
**What you expected to happen**:
Upon Node upgrade, Pods should either be terminated (so a controller can recreate them) or they should come back to Running on their own.
**How to reproduce it** (as minimally and precisely as possible):
1. Start a v1.4 cluster on GCE.
```shell
git checkout release-1.4
git pull upstream release-1.4
make clean quick-release
cluster/kube-up.sh
```
1. Create the following example PetSet and wait for all Pods to be Running.
```shell
kubectl create -f petset.yaml
```
```yaml
# A headless service to create DNS records
apiVersion: v1
kind: Service
metadata:
name: nginx
labels:
app: nginx
spec:
ports:
- port: 80
name: web
# *.nginx.default.svc.cluster.local
clusterIP: None
selector:
app: nginx
---
apiVersion: apps/v1alpha1
kind: PetSet
metadata:
name: web
spec:
serviceName: "nginx"
replicas: 5
template:
metadata:
labels:
app: nginx
annotations:
pod.alpha.kubernetes.io/initialized: "true"
spec:
containers:
- name: nginx
image: gcr.io/google_containers/nginx-slim:0.8
ports:
- containerPort: 80
name: web
volumeMounts:
- name: www
mountPath: /usr/share/nginx/html
volumeClaimTemplates:
- metadata:
name: www
annotations:
volume.alpha.kubernetes.io/storage-class: anything
spec:
accessModes: [ "ReadWriteOnce" ]
resources:
requests:
storage: 1Gi
```
1. Delete PetSet without cascading (keep the Pods).
```shell
kubectl delete petset web --cascade=false
```
1. Upgrade master and nodes to `v1.5.0-beta.2` with `cluster/gce/upgrade.sh`.
```shell
git checkout release-1.5
git pull upstream release-1.5
cluster/gce/upgrade.sh v1.5.0-beta.2
```
1. Check for Pods stuck in ContainerCreating.
```
NAME READY STATUS RESTARTS AGE
web-0 0/1 ContainerCreating 0 42m
web-4 0/1 ContainerCreating 0 42m
```
**Anything else do we need to know**:
After the above steps, some Pods may end up terminating. It seems to be random which ones terminate and which ones get stuck in ContainerCreating.
Manually deleting and recreating all the stuck Pods after the Nodes are upgraded allows them to correctly attach to the preexisting PVCs and become Running.
| non_usab | pods with pvcs stuck in containercreating after node upgrade is this a request for help if yes you should use our troubleshooting guide and community support channels see no what keywords did you search in kubernetes issues before filing this one if you have found any duplicates you should instead reply there upgrade volume mount timeout containercreating is this a bug report or feature request choose one bug report if this is a bug report please fill in as much of the template below as you can if you leave out information we can t help you as well if this is a feature request please describe in detail the feature behavior change you d like to see in both cases be ready for followup questions and please respond in a timely manner if we can t reproduce a bug or think a feature already exists we might close your issue if we re wrong please feel free to reopen it and explain why kubernetes version use kubectl version server version version info major minor gitversion beta gitcommit gittreestate clean builddate goversion compiler gc platform linux environment cloud provider or hardware configuration gce os e g from etc os release master build id name container vm image google crash id lakitu version id bug report url pretty name google container vm image version google metrics product id home url id gci minion pretty name debian gnu linux wheezy name debian gnu linux version id version wheezy id debian ansi color home url support url bug report url kernel e g uname a master linux kubernetes master smp thu nov pst intel r xeon r cpu genuineintel gnu linux minion linux kubernetes minion group smp debian gnu linux install tools cluster gce upgrade sh others what happened pods with pvcs get stuck in containercreating after node gets upgraded the error is timeout expired waiting for volumes to attach mount as shown below console kubectl describe pod web name web namespace default node kubernetes minion group start time mon nov labels app nginx status pending ip controllers petset web containers nginx container id image gcr io google containers nginx slim image id port tcp requests cpu state waiting reason containercreating ready false restart count volume mounts usr share nginx html from www rw var run secrets kubernetes io serviceaccount from default token ro environment variables conditions type status initialized true ready false podscheduled true volumes www type persistentvolumeclaim a reference to a persistentvolumeclaim in the same namespace claimname www web readonly false default token type secret a volume populated by a secret secretname default token qos class burstable tolerations events firstseen lastseen count from subobjectpath type reason message default scheduler normal scheduled successfully assigned web to kubernetes minion group kubelet kubernetes minion group spec containers nginx normal pulling pulling image gcr io google containers nginx slim kubelet kubernetes minion group spec containers nginx normal pulled successfully pulled image gcr io google containers nginx slim kubelet kubernetes minion group spec containers nginx normal created created container with docker id security kubelet kubernetes minion group spec containers nginx normal started started container with docker id kubelet kubernetes minion group warning failedmount mountvolume setup failed for volume kubernetes io secret default token spec name default token pod uid with get dial tcp lookup kubernetes master on no such host kubelet kubernetes minion group warning failedmount mountvolume setup failed for volume kubernetes io secret default token spec name default token pod uid with get dial tcp i o timeout kubelet kubernetes minion group warning failedmount mountvolume setup failed for volume kubernetes io secret default token spec name default token pod uid with get dial tcp getsockopt connection refused kubelet kubernetes minion group warning failedsync error syncing pod skipping network is not ready kubelet kubernetes minion group warning failedsync error syncing pod skipping network is not ready kubelet kubernetes minion group warning failedmount unable to mount volumes for pod web default timeout expired waiting for volumes to attach mount for pod web default list of unattached unmounted volumes kubelet kubernetes minion group warning failedsync error syncing pod skipping timeout expired waiting for volumes to attach mount for pod web default list of unattached unmounted volumes what you expected to happen upon node upgrade pods should either be terminated so a controller can recreate them or they should come back to running on their own how to reproduce it as minimally and precisely as possible start a cluster on gce shell git checkout release git pull upstream release make clean quick release cluster kube up sh create the following example petset and wait for all pods to be running shell kubectl create f petset yaml yaml a headless service to create dns records apiversion kind service metadata name nginx labels app nginx spec ports port name web nginx default svc cluster local clusterip none selector app nginx apiversion apps kind petset metadata name web spec servicename nginx replicas template metadata labels app nginx annotations pod alpha kubernetes io initialized true spec containers name nginx image gcr io google containers nginx slim ports containerport name web volumemounts name www mountpath usr share nginx html volumeclaimtemplates metadata name www annotations volume alpha kubernetes io storage class anything spec accessmodes resources requests storage delete petset without cascading keep the pods shell kubectl delete petset web cascade false upgrade master and nodes to beta with cluster gce upgrade sh shell git checkout release git pull upstream release cluster gce upgrade sh beta check for pods stuck in containercreating name ready status restarts age web containercreating web containercreating anything else do we need to know after the above steps some pods may end up terminating it seems to be random which ones terminate and which ones get stuck in containercreating manually deleting and recreating all the stuck pods after the nodes are upgraded allows them to correctly attach to the preexisting pvcs and become running | 0 |
4,182 | 3,757,107,614 | IssuesEvent | 2016-03-13 19:51:01 | ironjan/klausurtool-ror | https://api.github.com/repos/ironjan/klausurtool-ror | opened | Remove useless button | help wanted usability | Go to some folder's details, e.g. http://0.0.0.0:3000/admin/old_folders/235
The button "Neues Exemplar" links to a anchor on the same page; this anchor is for a inline-form which is directly below the instance table. It's therefore useless and should be removed. | True | Remove useless button - Go to some folder's details, e.g. http://0.0.0.0:3000/admin/old_folders/235
The button "Neues Exemplar" links to a anchor on the same page; this anchor is for a inline-form which is directly below the instance table. It's therefore useless and should be removed. | usab | remove useless button go to some folder s details e g the button neues exemplar links to a anchor on the same page this anchor is for a inline form which is directly below the instance table it s therefore useless and should be removed | 1 |
16,874 | 11,449,212,627 | IssuesEvent | 2020-02-06 06:27:07 | pombase/canto | https://api.github.com/repos/pombase/canto | closed | Rename the 'Back' button on gene entry page | usability | Related: #1624
When you first start a session and reach the gene entry page, the only navigation button available is a 'Continue' button, which is disabled until you've entered at least one gene ID or specified that the paper has no genes:
<img width="400" src="https://user-images.githubusercontent.com/37659591/72737529-a9050380-3b97-11ea-81d8-c4e1cffa3abf.png">
(Note the image shows PHI-Canto's version of the gene entry page, but I believe it works the same way for all versions.)
Then, when you return to the gene entry page after having entered genes, a second navigation button appears, labelled 'Back':
<img width="400" src="https://user-images.githubusercontent.com/37659591/72737882-59730780-3b98-11ea-8de9-dc0d27f39456.png">
The Back button is always enabled, but the Continue button doesn't enable until the user has entered a gene (or picked a host organism in the case of PHI-Canto).
The problem is that the Back button does almost exactly the same thing as the Continue button: they both navigate to the same page; the only difference is that the Continue button stores your changes, whereas the Back button doesn't. If the Back button truly navigated the user backwards in the workflow, it would take them to the page where they entered their curator details. So, the Back button is redundant from a navigational perspective. I think the page would make more sense if the Back button was replaced with a Cancel button (as shown below), since this matches the intended behaviour of the button.
<img width="400" src="https://user-images.githubusercontent.com/37659591/72738304-31d06f00-3b99-11ea-9758-726160e702e9.png"> | True | Rename the 'Back' button on gene entry page - Related: #1624
When you first start a session and reach the gene entry page, the only navigation button available is a 'Continue' button, which is disabled until you've entered at least one gene ID or specified that the paper has no genes:
<img width="400" src="https://user-images.githubusercontent.com/37659591/72737529-a9050380-3b97-11ea-81d8-c4e1cffa3abf.png">
(Note the image shows PHI-Canto's version of the gene entry page, but I believe it works the same way for all versions.)
Then, when you return to the gene entry page after having entered genes, a second navigation button appears, labelled 'Back':
<img width="400" src="https://user-images.githubusercontent.com/37659591/72737882-59730780-3b98-11ea-8de9-dc0d27f39456.png">
The Back button is always enabled, but the Continue button doesn't enable until the user has entered a gene (or picked a host organism in the case of PHI-Canto).
The problem is that the Back button does almost exactly the same thing as the Continue button: they both navigate to the same page; the only difference is that the Continue button stores your changes, whereas the Back button doesn't. If the Back button truly navigated the user backwards in the workflow, it would take them to the page where they entered their curator details. So, the Back button is redundant from a navigational perspective. I think the page would make more sense if the Back button was replaced with a Cancel button (as shown below), since this matches the intended behaviour of the button.
<img width="400" src="https://user-images.githubusercontent.com/37659591/72738304-31d06f00-3b99-11ea-9758-726160e702e9.png"> | usab | rename the back button on gene entry page related when you first start a session and reach the gene entry page the only navigation button available is a continue button which is disabled until you ve entered at least one gene id or specified that the paper has no genes img width src note the image shows phi canto s version of the gene entry page but i believe it works the same way for all versions then when you return to the gene entry page after having entered genes a second navigation button appears labelled back img width src the back button is always enabled but the continue button doesn t enable until the user has entered a gene or picked a host organism in the case of phi canto the problem is that the back button does almost exactly the same thing as the continue button they both navigate to the same page the only difference is that the continue button stores your changes whereas the back button doesn t if the back button truly navigated the user backwards in the workflow it would take them to the page where they entered their curator details so the back button is redundant from a navigational perspective i think the page would make more sense if the back button was replaced with a cancel button as shown below since this matches the intended behaviour of the button img width src | 1 |
2,157 | 3,051,939,420 | IssuesEvent | 2015-08-12 12:01:58 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 15349586: Events in Forecast in Notification Center should be tapable | classification:ui/usability reproducible:always status:open | #### Description
Summary:
When you open Notification Center's Today View & you have events scheduled for the next day, you can't tap them to open the Calendar.app (https://pbs.twimg.com/media/BXznBrRCAAAFEUs.jpg:large)
Steps to Reproduce:
1. Create an event for tomorrow
2. Open Notification Center
3. Swipe to the Today Tab
4. Try to tap on the scheduled event
Expected Results:
Calendar.app should open and display the next day.
Actual Results:
Nothing happens
-
Product Version: 7.0.3
Created: 2013-10-30T08:22:44.441039
Originated: 2013-10-30T09:22:00
Open Radar Link: http://www.openradar.me/15349586 | True | 15349586: Events in Forecast in Notification Center should be tapable - #### Description
Summary:
When you open Notification Center's Today View & you have events scheduled for the next day, you can't tap them to open the Calendar.app (https://pbs.twimg.com/media/BXznBrRCAAAFEUs.jpg:large)
Steps to Reproduce:
1. Create an event for tomorrow
2. Open Notification Center
3. Swipe to the Today Tab
4. Try to tap on the scheduled event
Expected Results:
Calendar.app should open and display the next day.
Actual Results:
Nothing happens
-
Product Version: 7.0.3
Created: 2013-10-30T08:22:44.441039
Originated: 2013-10-30T09:22:00
Open Radar Link: http://www.openradar.me/15349586 | usab | events in forecast in notification center should be tapable description summary when you open notification center s today view you have events scheduled for the next day you can t tap them to open the calendar app steps to reproduce create an event for tomorrow open notification center swipe to the today tab try to tap on the scheduled event expected results calendar app should open and display the next day actual results nothing happens product version created originated open radar link | 1 |
491,352 | 14,149,839,780 | IssuesEvent | 2020-11-11 01:52:56 | StrangeLoopGames/EcoIssues | https://api.github.com/repos/StrangeLoopGames/EcoIssues | closed | Test self-hosted world | Priority: High QA | We have a special system for teachers that want citizen accounts, but dont want to buy a cloud world., but instead want to self-host an eco server.
This system is not used by very many users. but it is being reported as broken.
To test:
create user with a citizen.
in user table, set self_hosted = 1 for the parent user (citizen owner)
Make sure the user does not own any cloud worlds in worlds table.
You should now be able to login to parent user, go to citizen page, and enter a server ip.
Use a non cloud world eco server to test. can be self hosted or third party hosted.
Then Login to game with citizen account, and check that user can enter their assigned world. and cannot enter any other world.
| 1.0 | Test self-hosted world - We have a special system for teachers that want citizen accounts, but dont want to buy a cloud world., but instead want to self-host an eco server.
This system is not used by very many users. but it is being reported as broken.
To test:
create user with a citizen.
in user table, set self_hosted = 1 for the parent user (citizen owner)
Make sure the user does not own any cloud worlds in worlds table.
You should now be able to login to parent user, go to citizen page, and enter a server ip.
Use a non cloud world eco server to test. can be self hosted or third party hosted.
Then Login to game with citizen account, and check that user can enter their assigned world. and cannot enter any other world.
| non_usab | test self hosted world we have a special system for teachers that want citizen accounts but dont want to buy a cloud world but instead want to self host an eco server this system is not used by very many users but it is being reported as broken to test create user with a citizen in user table set self hosted for the parent user citizen owner make sure the user does not own any cloud worlds in worlds table you should now be able to login to parent user go to citizen page and enter a server ip use a non cloud world eco server to test can be self hosted or third party hosted then login to game with citizen account and check that user can enter their assigned world and cannot enter any other world | 0 |
254,539 | 27,393,591,129 | IssuesEvent | 2023-02-28 17:54:44 | opensearch-project/oui | https://api.github.com/repos/opensearch-project/oui | closed | async-0.9.2.js: 1 vulnerabilities (highest severity is: 7.8) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>async-0.9.2.js</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js">https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js</a></p>
<p>Path to vulnerable library: /packages/react-datepicker/node_modules/watchpack/node_modules/async/lib/async.js</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/opensearch-project/oui/commit/bbca6f5de4b03c3c57a1333400bba2178b3aa99e">bbca6f5de4b03c3c57a1333400bba2178b3aa99e</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (async version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-43138](https://www.mend.io/vulnerability-database/CVE-2021-43138) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.8 | async-0.9.2.js | Direct | async - 2.6.4,3.2.2 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-43138</summary>
### Vulnerable Library - <b>async-0.9.2.js</b></p>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js">https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js</a></p>
<p>Path to vulnerable library: /packages/react-datepicker/node_modules/watchpack/node_modules/async/lib/async.js</p>
<p>
Dependency Hierarchy:
- :x: **async-0.9.2.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/opensearch-project/oui/commit/bbca6f5de4b03c3c57a1333400bba2178b3aa99e">bbca6f5de4b03c3c57a1333400bba2178b3aa99e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Async before 2.6.4 and 3.x before 3.2.2, a malicious user can obtain privileges via the mapValues() method, aka lib/internal/iterator.js createObjectIterator prototype pollution.
<p>Publish Date: 2022-04-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-43138>CVE-2021-43138</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p>
<p>Release Date: 2022-04-06</p>
<p>Fix Resolution: async - 2.6.4,3.2.2</p>
</p>
<p></p>
</details> | True | async-0.9.2.js: 1 vulnerabilities (highest severity is: 7.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>async-0.9.2.js</b></p></summary>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js">https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js</a></p>
<p>Path to vulnerable library: /packages/react-datepicker/node_modules/watchpack/node_modules/async/lib/async.js</p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/opensearch-project/oui/commit/bbca6f5de4b03c3c57a1333400bba2178b3aa99e">bbca6f5de4b03c3c57a1333400bba2178b3aa99e</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (async version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2021-43138](https://www.mend.io/vulnerability-database/CVE-2021-43138) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.8 | async-0.9.2.js | Direct | async - 2.6.4,3.2.2 | ❌ |
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-43138</summary>
### Vulnerable Library - <b>async-0.9.2.js</b></p>
<p>Higher-order functions and common patterns for asynchronous code</p>
<p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js">https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js</a></p>
<p>Path to vulnerable library: /packages/react-datepicker/node_modules/watchpack/node_modules/async/lib/async.js</p>
<p>
Dependency Hierarchy:
- :x: **async-0.9.2.js** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/opensearch-project/oui/commit/bbca6f5de4b03c3c57a1333400bba2178b3aa99e">bbca6f5de4b03c3c57a1333400bba2178b3aa99e</a></p>
<p>Found in base branch: <b>main</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
In Async before 2.6.4 and 3.x before 3.2.2, a malicious user can obtain privileges via the mapValues() method, aka lib/internal/iterator.js createObjectIterator prototype pollution.
<p>Publish Date: 2022-04-06
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-43138>CVE-2021-43138</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.8</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p>
<p>Release Date: 2022-04-06</p>
<p>Fix Resolution: async - 2.6.4,3.2.2</p>
</p>
<p></p>
</details> | non_usab | async js vulnerabilities highest severity is autoclosed vulnerable library async js higher order functions and common patterns for asynchronous code library home page a href path to vulnerable library packages react datepicker node modules watchpack node modules async lib async js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in async version remediation available high async js direct async details cve vulnerable library async js higher order functions and common patterns for asynchronous code library home page a href path to vulnerable library packages react datepicker node modules watchpack node modules async lib async js dependency hierarchy x async js vulnerable library found in head commit a href found in base branch main vulnerability details in async before and x before a malicious user can obtain privileges via the mapvalues method aka lib internal iterator js createobjectiterator prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution async | 0 |
203,368 | 7,060,436,598 | IssuesEvent | 2018-01-05 08:46:06 | BlueBrain/neurocurator | https://api.github.com/repos/BlueBrain/neurocurator | opened | Unable to load a PDF in Zotero even if connected to the VPN | bug high priority |
Steps:
1. Choose the paper that I wanted to annotate which has no PDF
2. After having the dialog window "This paper is not already in the curator database" and pressing "select PDF" I get no PDF
This is the paper I wanted to annotate
Postnatal development of the basolateral complex of rabbit amygdala: a stereological and histochemical study
Stack trace:
```
[sjimenez:~/neurocurator] $ ./neurocurator_0-4-0_mac_x64
Building ontological tree for PATO:0000261 ...
Building ontological tree for NIFORG:birnlex_569 ...
Building ontological tree for CHEBI:23367 ...
Building ontological tree for GO:0097458 ...
Building ontological tree for NIFGA:birnlex_1167 ...
Building ontological tree for NIFMOL:nifext_2506 ...
Building ontological tree for NIFSUB:sao914572699 ...
Building ontological tree for nifext_8054 ...
Building ontological tree for sao1813327414 ...
Loading cached Zotero data...
Cached Zotero data loaded.
2018-01-05 09:26:00.970 neurocurator_0-4-0_mac_x64[46624:1637373] modalSession has been exited prematurely - check for a reentrant call to endModalSession:
Traceback (most recent call last):
File "urllib/request.py", line 1183, in do_open
File "http/client.py", line 1137, in request
File "http/client.py", line 1182, in _send_request
File "http/client.py", line 1133, in endheaders
File "http/client.py", line 963, in _send_output
File "http/client.py", line 898, in send
File "http/client.py", line 1287, in connect
File "ssl.py", line 362, in wrap_socket
File "ssl.py", line 580, in __init__
File "ssl.py", line 807, in do_handshake
ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "site-packages/neurocurator/mainWin.py", line 1008, in paperSelectionChanged
File "site-packages/neurocurator/mainWin.py", line 1113, in importPDF
File "site-packages/nat/id.py", line 127, in checkID
File "site-packages/nat/id.py", line 134, in checkPMID
File "site-packages/nat/id.py", line 17, in getPMIDSoup
File "site-packages/nat/id.py", line 26, in getSoup
File "urllib/request.py", line 161, in urlopen
File "urllib/request.py", line 470, in open
File "urllib/request.py", line 580, in http_response
File "urllib/request.py", line 502, in error
File "urllib/request.py", line 442, in _call_chain
File "urllib/request.py", line 685, in http_error_302
File "urllib/request.py", line 464, in open
File "urllib/request.py", line 482, in _open
File "urllib/request.py", line 442, in _call_chain
File "urllib/request.py", line 1226, in https_open
File "urllib/request.py", line 1185, in do_open
urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)>
2018-01-05 09:31:29.447 neurocurator_0-4-0_mac_x64[46624:1637373] modalSession has been exited prematurely - check for a reentrant call to endModalSession:
Traceback (most recent call last):
File "urllib/request.py", line 1183, in do_open
File "http/client.py", line 1137, in request
File "http/client.py", line 1182, in _send_request
File "http/client.py", line 1133, in endheaders
File "http/client.py", line 963, in _send_output
File "http/client.py", line 898, in send
File "http/client.py", line 1287, in connect
File "ssl.py", line 362, in wrap_socket
File "ssl.py", line 580, in __init__
File "ssl.py", line 807, in do_handshake
ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "site-packages/neurocurator/mainWin.py", line 1008, in paperSelectionChanged
File "site-packages/neurocurator/mainWin.py", line 1113, in importPDF
File "site-packages/nat/id.py", line 127, in checkID
File "site-packages/nat/id.py", line 134, in checkPMID
File "site-packages/nat/id.py", line 17, in getPMIDSoup
File "site-packages/nat/id.py", line 26, in getSoup
File "urllib/request.py", line 161, in urlopen
File "urllib/request.py", line 470, in open
File "urllib/request.py", line 580, in http_response
File "urllib/request.py", line 502, in error
File "urllib/request.py", line 442, in _call_chain
File "urllib/request.py", line 685, in http_error_302
File "urllib/request.py", line 464, in open
File "urllib/request.py", line 482, in _open
File "urllib/request.py", line 442, in _call_chain
File "urllib/request.py", line 1226, in https_open
File "urllib/request.py", line 1185, in do_open
urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)>
```
| 1.0 | Unable to load a PDF in Zotero even if connected to the VPN -
Steps:
1. Choose the paper that I wanted to annotate which has no PDF
2. After having the dialog window "This paper is not already in the curator database" and pressing "select PDF" I get no PDF
This is the paper I wanted to annotate
Postnatal development of the basolateral complex of rabbit amygdala: a stereological and histochemical study
Stack trace:
```
[sjimenez:~/neurocurator] $ ./neurocurator_0-4-0_mac_x64
Building ontological tree for PATO:0000261 ...
Building ontological tree for NIFORG:birnlex_569 ...
Building ontological tree for CHEBI:23367 ...
Building ontological tree for GO:0097458 ...
Building ontological tree for NIFGA:birnlex_1167 ...
Building ontological tree for NIFMOL:nifext_2506 ...
Building ontological tree for NIFSUB:sao914572699 ...
Building ontological tree for nifext_8054 ...
Building ontological tree for sao1813327414 ...
Loading cached Zotero data...
Cached Zotero data loaded.
2018-01-05 09:26:00.970 neurocurator_0-4-0_mac_x64[46624:1637373] modalSession has been exited prematurely - check for a reentrant call to endModalSession:
Traceback (most recent call last):
File "urllib/request.py", line 1183, in do_open
File "http/client.py", line 1137, in request
File "http/client.py", line 1182, in _send_request
File "http/client.py", line 1133, in endheaders
File "http/client.py", line 963, in _send_output
File "http/client.py", line 898, in send
File "http/client.py", line 1287, in connect
File "ssl.py", line 362, in wrap_socket
File "ssl.py", line 580, in __init__
File "ssl.py", line 807, in do_handshake
ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "site-packages/neurocurator/mainWin.py", line 1008, in paperSelectionChanged
File "site-packages/neurocurator/mainWin.py", line 1113, in importPDF
File "site-packages/nat/id.py", line 127, in checkID
File "site-packages/nat/id.py", line 134, in checkPMID
File "site-packages/nat/id.py", line 17, in getPMIDSoup
File "site-packages/nat/id.py", line 26, in getSoup
File "urllib/request.py", line 161, in urlopen
File "urllib/request.py", line 470, in open
File "urllib/request.py", line 580, in http_response
File "urllib/request.py", line 502, in error
File "urllib/request.py", line 442, in _call_chain
File "urllib/request.py", line 685, in http_error_302
File "urllib/request.py", line 464, in open
File "urllib/request.py", line 482, in _open
File "urllib/request.py", line 442, in _call_chain
File "urllib/request.py", line 1226, in https_open
File "urllib/request.py", line 1185, in do_open
urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)>
2018-01-05 09:31:29.447 neurocurator_0-4-0_mac_x64[46624:1637373] modalSession has been exited prematurely - check for a reentrant call to endModalSession:
Traceback (most recent call last):
File "urllib/request.py", line 1183, in do_open
File "http/client.py", line 1137, in request
File "http/client.py", line 1182, in _send_request
File "http/client.py", line 1133, in endheaders
File "http/client.py", line 963, in _send_output
File "http/client.py", line 898, in send
File "http/client.py", line 1287, in connect
File "ssl.py", line 362, in wrap_socket
File "ssl.py", line 580, in __init__
File "ssl.py", line 807, in do_handshake
ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)
During handling of the above exception, another exception occurred:
Traceback (most recent call last):
File "site-packages/neurocurator/mainWin.py", line 1008, in paperSelectionChanged
File "site-packages/neurocurator/mainWin.py", line 1113, in importPDF
File "site-packages/nat/id.py", line 127, in checkID
File "site-packages/nat/id.py", line 134, in checkPMID
File "site-packages/nat/id.py", line 17, in getPMIDSoup
File "site-packages/nat/id.py", line 26, in getSoup
File "urllib/request.py", line 161, in urlopen
File "urllib/request.py", line 470, in open
File "urllib/request.py", line 580, in http_response
File "urllib/request.py", line 502, in error
File "urllib/request.py", line 442, in _call_chain
File "urllib/request.py", line 685, in http_error_302
File "urllib/request.py", line 464, in open
File "urllib/request.py", line 482, in _open
File "urllib/request.py", line 442, in _call_chain
File "urllib/request.py", line 1226, in https_open
File "urllib/request.py", line 1185, in do_open
urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)>
```
| non_usab | unable to load a pdf in zotero even if connected to the vpn steps choose the paper that i wanted to annotate which has no pdf after having the dialog window this paper is not already in the curator database and pressing select pdf i get no pdf this is the paper i wanted to annotate postnatal development of the basolateral complex of rabbit amygdala a stereological and histochemical study stack trace neurocurator mac building ontological tree for pato building ontological tree for niforg birnlex building ontological tree for chebi building ontological tree for go building ontological tree for nifga birnlex building ontological tree for nifmol nifext building ontological tree for nifsub building ontological tree for nifext building ontological tree for loading cached zotero data cached zotero data loaded neurocurator mac modalsession has been exited prematurely check for a reentrant call to endmodalsession traceback most recent call last file urllib request py line in do open file http client py line in request file http client py line in send request file http client py line in endheaders file http client py line in send output file http client py line in send file http client py line in connect file ssl py line in wrap socket file ssl py line in init file ssl py line in do handshake ssl sslerror certificate verify failed ssl c during handling of the above exception another exception occurred traceback most recent call last file site packages neurocurator mainwin py line in paperselectionchanged file site packages neurocurator mainwin py line in importpdf file site packages nat id py line in checkid file site packages nat id py line in checkpmid file site packages nat id py line in getpmidsoup file site packages nat id py line in getsoup file urllib request py line in urlopen file urllib request py line in open file urllib request py line in http response file urllib request py line in error file urllib request py line in call chain file urllib request py line in http error file urllib request py line in open file urllib request py line in open file urllib request py line in call chain file urllib request py line in https open file urllib request py line in do open urllib error urlerror neurocurator mac modalsession has been exited prematurely check for a reentrant call to endmodalsession traceback most recent call last file urllib request py line in do open file http client py line in request file http client py line in send request file http client py line in endheaders file http client py line in send output file http client py line in send file http client py line in connect file ssl py line in wrap socket file ssl py line in init file ssl py line in do handshake ssl sslerror certificate verify failed ssl c during handling of the above exception another exception occurred traceback most recent call last file site packages neurocurator mainwin py line in paperselectionchanged file site packages neurocurator mainwin py line in importpdf file site packages nat id py line in checkid file site packages nat id py line in checkpmid file site packages nat id py line in getpmidsoup file site packages nat id py line in getsoup file urllib request py line in urlopen file urllib request py line in open file urllib request py line in http response file urllib request py line in error file urllib request py line in call chain file urllib request py line in http error file urllib request py line in open file urllib request py line in open file urllib request py line in call chain file urllib request py line in https open file urllib request py line in do open urllib error urlerror | 0 |
8,041 | 5,355,084,868 | IssuesEvent | 2017-02-20 11:51:34 | apinf/platform | https://api.github.com/repos/apinf/platform | opened | API Profile -> Proxy tab: No identification of Mandatory fields | Usability Issue | # Reproduction steps:
1. Sign in as admin and go to an API profile -> proxy tab.
2. Add an API base path.
3. Leave Proxy base path (and other fields) empty and click Save and Publish
# Result
Although s-alert indicates to fill up mandatory fields, there is no indication which fields are mandatory and needs to be filled in.
# Expected Result
by using CSS alert class, on leaving mandatory fields empty, we can highlight the field names when they are left empty.
# Environment
nightly.apinf.io, windows 8, chrome. | True | API Profile -> Proxy tab: No identification of Mandatory fields - # Reproduction steps:
1. Sign in as admin and go to an API profile -> proxy tab.
2. Add an API base path.
3. Leave Proxy base path (and other fields) empty and click Save and Publish
# Result
Although s-alert indicates to fill up mandatory fields, there is no indication which fields are mandatory and needs to be filled in.
# Expected Result
by using CSS alert class, on leaving mandatory fields empty, we can highlight the field names when they are left empty.
# Environment
nightly.apinf.io, windows 8, chrome. | usab | api profile proxy tab no identification of mandatory fields reproduction steps sign in as admin and go to an api profile proxy tab add an api base path leave proxy base path and other fields empty and click save and publish result although s alert indicates to fill up mandatory fields there is no indication which fields are mandatory and needs to be filled in expected result by using css alert class on leaving mandatory fields empty we can highlight the field names when they are left empty environment nightly apinf io windows chrome | 1 |
10,873 | 6,967,788,052 | IssuesEvent | 2017-12-10 13:39:32 | vmware/docker-volume-vsphere | https://api.github.com/repos/vmware/docker-volume-vsphere | opened | Build with a fresh Windows VM throws error messages | component/guest-plugin component/windows-plugin kind/usability P1 | Make issue, below messages get displayed when building a windows plugin with a freshly installed windows VM. Make scripts need to check the existence of older source and then proceed to remove.
Remove-Item : Cannot find path 'C:\Users\root\go\src\github.com\vmware\docker-volume-vsphere'
because it does not exist.
At line:1 char:1
+ Remove-Item -Recurse -Force C:\Users\root\go\src\github.com\vmware\do ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : ObjectNotFound: (C:\Users\root\g...-volume-vsphere:String) [Remove-
Item], ItemNotFoundException
+ FullyQualifiedErrorId : PathNotFound,Microsoft.PowerShell.Commands.RemoveItemCommand
| True | Build with a fresh Windows VM throws error messages - Make issue, below messages get displayed when building a windows plugin with a freshly installed windows VM. Make scripts need to check the existence of older source and then proceed to remove.
Remove-Item : Cannot find path 'C:\Users\root\go\src\github.com\vmware\docker-volume-vsphere'
because it does not exist.
At line:1 char:1
+ Remove-Item -Recurse -Force C:\Users\root\go\src\github.com\vmware\do ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : ObjectNotFound: (C:\Users\root\g...-volume-vsphere:String) [Remove-
Item], ItemNotFoundException
+ FullyQualifiedErrorId : PathNotFound,Microsoft.PowerShell.Commands.RemoveItemCommand
| usab | build with a fresh windows vm throws error messages make issue below messages get displayed when building a windows plugin with a freshly installed windows vm make scripts need to check the existence of older source and then proceed to remove remove item cannot find path c users root go src github com vmware docker volume vsphere because it does not exist at line char remove item recurse force c users root go src github com vmware do categoryinfo objectnotfound c users root g volume vsphere string remove item itemnotfoundexception fullyqualifiederrorid pathnotfound microsoft powershell commands removeitemcommand | 1 |
28,001 | 30,879,163,022 | IssuesEvent | 2023-08-03 16:13:44 | ClickHouse/ClickHouse | https://api.github.com/repos/ClickHouse/ClickHouse | closed | clickhouse is unable to start because of the file system.sql.tmp | usability | In some cases after a **power outage** Clickhouse is unable to start.
It's unclear how to get to this state.
The state itself is easy to reproduce:
```
22.3
service clickhouse-server stop
mv /var/lib/clickhouse/metadata/system.sql /var/lib/clickhouse/metadata/system.sql.tmp
service clickhouse-server start
<Error> Application: DB::ErrnoException: Cannot open file /var/lib/clickhouse/metadata/system.sql.tmp, errno: 17, strerror: File exists: while loading database `system` from path /var/lib/clickhouse/metadata/system
<Information> Application: shutting down
```
It's not clear what is inside `system.sql.tmp` (no info).
I am wondering is it related to Atomic/Ordering conversion or not? It seems not.
I think `system.sql.tmp` only exists during the initial creation of the system database? This is right?
--------------------------
Another problem: if you do `rm /var/lib/clickhouse/metadata/system.sql.tmp`
Then clickhouse will create `system.sql` automatically but **Ordinary**
```
# cat /var/lib/clickhouse/metadata/system.sql
ATTACH DATABASE system
ENGINE = Ordinary
# ls -l /var/lib/clickhouse/metadata
total 32
lrwxrwxrwx 1 clickhouse clickhouse 67 Apr 4 19:36 default -> /var/lib/clickhouse/store/c0f/c0fdc790-48b7-4d04-80fd-c79048b7ad04/
-rw-r----- 1 clickhouse clickhouse 78 Apr 4 19:36 default.sql
drwxr-x--- 2 clickhouse clickhouse 4096 Apr 5 00:24 information_schema
drwxr-x--- 2 clickhouse clickhouse 4096 Apr 5 00:24 INFORMATION_SCHEMA
-rw-r----- 1 clickhouse clickhouse 51 Apr 5 00:24 information_schema.sql
-rw-r----- 1 clickhouse clickhouse 51 Apr 5 00:24 INFORMATION_SCHEMA.sql
lrwxrwxrwx 1 clickhouse clickhouse 67 Apr 5 00:30 system -> /var/lib/clickhouse/store/5e3/5e34ade0-c123-43f5-8e76-9da06b17cf70/
-rw-r----- 1 clickhouse clickhouse 41 Apr 5 00:54 system.sql
```
Which is odd and probably will lead to more issues. | True | clickhouse is unable to start because of the file system.sql.tmp - In some cases after a **power outage** Clickhouse is unable to start.
It's unclear how to get to this state.
The state itself is easy to reproduce:
```
22.3
service clickhouse-server stop
mv /var/lib/clickhouse/metadata/system.sql /var/lib/clickhouse/metadata/system.sql.tmp
service clickhouse-server start
<Error> Application: DB::ErrnoException: Cannot open file /var/lib/clickhouse/metadata/system.sql.tmp, errno: 17, strerror: File exists: while loading database `system` from path /var/lib/clickhouse/metadata/system
<Information> Application: shutting down
```
It's not clear what is inside `system.sql.tmp` (no info).
I am wondering is it related to Atomic/Ordering conversion or not? It seems not.
I think `system.sql.tmp` only exists during the initial creation of the system database? This is right?
--------------------------
Another problem: if you do `rm /var/lib/clickhouse/metadata/system.sql.tmp`
Then clickhouse will create `system.sql` automatically but **Ordinary**
```
# cat /var/lib/clickhouse/metadata/system.sql
ATTACH DATABASE system
ENGINE = Ordinary
# ls -l /var/lib/clickhouse/metadata
total 32
lrwxrwxrwx 1 clickhouse clickhouse 67 Apr 4 19:36 default -> /var/lib/clickhouse/store/c0f/c0fdc790-48b7-4d04-80fd-c79048b7ad04/
-rw-r----- 1 clickhouse clickhouse 78 Apr 4 19:36 default.sql
drwxr-x--- 2 clickhouse clickhouse 4096 Apr 5 00:24 information_schema
drwxr-x--- 2 clickhouse clickhouse 4096 Apr 5 00:24 INFORMATION_SCHEMA
-rw-r----- 1 clickhouse clickhouse 51 Apr 5 00:24 information_schema.sql
-rw-r----- 1 clickhouse clickhouse 51 Apr 5 00:24 INFORMATION_SCHEMA.sql
lrwxrwxrwx 1 clickhouse clickhouse 67 Apr 5 00:30 system -> /var/lib/clickhouse/store/5e3/5e34ade0-c123-43f5-8e76-9da06b17cf70/
-rw-r----- 1 clickhouse clickhouse 41 Apr 5 00:54 system.sql
```
Which is odd and probably will lead to more issues. | usab | clickhouse is unable to start because of the file system sql tmp in some cases after a power outage clickhouse is unable to start it s unclear how to get to this state the state itself is easy to reproduce service clickhouse server stop mv var lib clickhouse metadata system sql var lib clickhouse metadata system sql tmp service clickhouse server start application db errnoexception cannot open file var lib clickhouse metadata system sql tmp errno strerror file exists while loading database system from path var lib clickhouse metadata system application shutting down it s not clear what is inside system sql tmp no info i am wondering is it related to atomic ordering conversion or not it seems not i think system sql tmp only exists during the initial creation of the system database this is right another problem if you do rm var lib clickhouse metadata system sql tmp then clickhouse will create system sql automatically but ordinary cat var lib clickhouse metadata system sql attach database system engine ordinary ls l var lib clickhouse metadata total lrwxrwxrwx clickhouse clickhouse apr default var lib clickhouse store rw r clickhouse clickhouse apr default sql drwxr x clickhouse clickhouse apr information schema drwxr x clickhouse clickhouse apr information schema rw r clickhouse clickhouse apr information schema sql rw r clickhouse clickhouse apr information schema sql lrwxrwxrwx clickhouse clickhouse apr system var lib clickhouse store rw r clickhouse clickhouse apr system sql which is odd and probably will lead to more issues | 1 |
420,191 | 12,234,460,494 | IssuesEvent | 2020-05-04 13:27:58 | threefoldtech/3bot_wallet | https://api.github.com/repos/threefoldtech/3bot_wallet | closed | Stellar Staging - Loading spinner after sending transaction shows wrong loading spinner message. | priority_low type_bug | **Repro steps**
1) Send a transaction
2) Check the loading spinner displayed
**Expected Result**
"Executing transation"
**Actual Result**

**System Info**
| 1.0 | Stellar Staging - Loading spinner after sending transaction shows wrong loading spinner message. - **Repro steps**
1) Send a transaction
2) Check the loading spinner displayed
**Expected Result**
"Executing transation"
**Actual Result**

**System Info**
| non_usab | stellar staging loading spinner after sending transaction shows wrong loading spinner message repro steps send a transaction check the loading spinner displayed expected result executing transation actual result system info | 0 |
5,278 | 3,917,079,986 | IssuesEvent | 2016-04-21 06:27:21 | kolliSuman/issues | https://api.github.com/repos/kolliSuman/issues | closed | QA_Expression Trees_Back to experiment_smk | Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open |
Defect Description:
In the "Expression Trees" experiment , the list of experiments link is not present instead the list of experiments link should be displayed on the experiment home page, where the user can easily get back and view the experiments.
Actual Result :
In the "Expression Trees" experiment , the list of experiments link is not displayed on the experiment home page
Environment:
"OS: Windows 7, Linux
Browsers: Firefox,Chrome
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5"
Test Step Link:
https://github.com/Virtual-Labs/Data Structure/blob/master/test-cases/integration_test-cases/Expression%20Trees/Expression%20Trees_12_Back%20to%20experiment_smk.org | True | QA_Expression Trees_Back to experiment_smk -
Defect Description:
In the "Expression Trees" experiment , the list of experiments link is not present instead the list of experiments link should be displayed on the experiment home page, where the user can easily get back and view the experiments.
Actual Result :
In the "Expression Trees" experiment , the list of experiments link is not displayed on the experiment home page
Environment:
"OS: Windows 7, Linux
Browsers: Firefox,Chrome
Bandwidth : 100Mbps
Hardware Configuration:8GBRAM ,
Processor:i5"
Test Step Link:
https://github.com/Virtual-Labs/Data Structure/blob/master/test-cases/integration_test-cases/Expression%20Trees/Expression%20Trees_12_Back%20to%20experiment_smk.org | usab | qa expression trees back to experiment smk defect description in the expression trees experiment the list of experiments link is not present instead the list of experiments link should be displayed on the experiment home page where the user can easily get back and view the experiments actual result in the expression trees experiment the list of experiments link is not displayed on the experiment home page environment os windows linux browsers firefox chrome bandwidth hardware configuration processor test step link structure blob master test cases integration test cases expression expression back smk org | 1 |
763,195 | 26,747,004,563 | IssuesEvent | 2023-01-30 16:39:07 | opendatahub-io/odh-dashboard | https://api.github.com/repos/opendatahub-io/odh-dashboard | closed | BYON - Add support for reading custom notebook validation pipeline status | kind/enhancement priority/normal feature/byon | As the custom notebook is progressing through the validation pipeline, we need to be able to get the current status of the pipeline status during each step until validation has completed (un)successfully.
Reference: https://github.com/open-services-group/byon/issues/18 | 1.0 | BYON - Add support for reading custom notebook validation pipeline status - As the custom notebook is progressing through the validation pipeline, we need to be able to get the current status of the pipeline status during each step until validation has completed (un)successfully.
Reference: https://github.com/open-services-group/byon/issues/18 | non_usab | byon add support for reading custom notebook validation pipeline status as the custom notebook is progressing through the validation pipeline we need to be able to get the current status of the pipeline status during each step until validation has completed un successfully reference | 0 |
23,502 | 22,052,186,726 | IssuesEvent | 2022-05-30 09:37:38 | opentap/opentap | https://api.github.com/repos/opentap/opentap | closed | Incomplete message in log when opening resources | Usability | Originally filed November 23 2021 by Wooi Po Chang on [GitLab](https://gitlab.com/OpenTAP/opentap/-/issues/811)
The log message `Waiting for resources to open:` seems to be incomplete

May consider improving `PrintWaitingMessage` function in [TestPlanExecution.cs](https://gitlab.com/OpenTAP/opentap/-/blob/master/Engine/TestPlanExecution.cs) | True | Incomplete message in log when opening resources - Originally filed November 23 2021 by Wooi Po Chang on [GitLab](https://gitlab.com/OpenTAP/opentap/-/issues/811)
The log message `Waiting for resources to open:` seems to be incomplete

May consider improving `PrintWaitingMessage` function in [TestPlanExecution.cs](https://gitlab.com/OpenTAP/opentap/-/blob/master/Engine/TestPlanExecution.cs) | usab | incomplete message in log when opening resources originally filed november by wooi po chang on the log message waiting for resources to open seems to be incomplete may consider improving printwaitingmessage function in | 1 |
14,224 | 8,925,810,066 | IssuesEvent | 2019-01-22 00:52:11 | matomo-org/matomo | https://api.github.com/repos/matomo-org/matomo | closed | Website selector: Ctrl+Click should open the website in a new tab | Help wanted c: Usability | Steps to reproduce:
* Click on website selector
* Do Ctrl+Click on a website name
* Got: the website clicked opens in the same window
* Expected instead: the website clicked opens in a new tab
| True | Website selector: Ctrl+Click should open the website in a new tab - Steps to reproduce:
* Click on website selector
* Do Ctrl+Click on a website name
* Got: the website clicked opens in the same window
* Expected instead: the website clicked opens in a new tab
| usab | website selector ctrl click should open the website in a new tab steps to reproduce click on website selector do ctrl click on a website name got the website clicked opens in the same window expected instead the website clicked opens in a new tab | 1 |
3,529 | 3,480,796,695 | IssuesEvent | 2015-12-29 11:04:54 | postmanlabs/postman-app-support | https://api.github.com/repos/postmanlabs/postman-app-support | closed | Clicking on Chrome Takes Me to Postman (Mac) | Usability | On Mac, clicking on Chrome icon on the launcher takes me to Postman when I have both apps running. | True | Clicking on Chrome Takes Me to Postman (Mac) - On Mac, clicking on Chrome icon on the launcher takes me to Postman when I have both apps running. | usab | clicking on chrome takes me to postman mac on mac clicking on chrome icon on the launcher takes me to postman when i have both apps running | 1 |
37,705 | 12,489,940,980 | IssuesEvent | 2020-05-31 21:17:30 | the-benchmarker/web-frameworks | https://api.github.com/repos/the-benchmarker/web-frameworks | closed | WS-2019-0379 (Medium) detected in commons-codec-1.9.jar | security vulnerability | ## WS-2019-0379 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.9.jar</b></p></summary>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: /tmp/ws-scm/web-frameworks/java/rapidoid/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/commons-codec/commons-codec/1.9/commons-codec-1.9.jar</p>
<p>
Dependency Hierarchy:
- rapidoid-web-5.5.5.jar (Root Library)
- rapidoid-http-client-5.5.5.jar
- httpasyncclient-4.1.jar
- httpclient-4.4.1.jar
- :x: **commons-codec-1.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/the-benchmarker/web-frameworks/commit/ab64b8404e01abede0aa4aa810306b3705409b30">ab64b8404e01abede0aa4aa810306b3705409b30</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p>
<p>Release Date: 2019-05-12</p>
<p>Fix Resolution: 1.13-RC1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0379 (Medium) detected in commons-codec-1.9.jar - ## WS-2019-0379 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.9.jar</b></p></summary>
<p>The Apache Commons Codec package contains simple encoder and decoders for
various formats such as Base64 and Hexadecimal. In addition to these
widely used encoders and decoders, the codec package also maintains a
collection of phonetic encoding utilities.</p>
<p>Path to dependency file: /tmp/ws-scm/web-frameworks/java/rapidoid/pom.xml</p>
<p>Path to vulnerable library: /root/.m2/repository/commons-codec/commons-codec/1.9/commons-codec-1.9.jar</p>
<p>
Dependency Hierarchy:
- rapidoid-web-5.5.5.jar (Root Library)
- rapidoid-http-client-5.5.5.jar
- httpasyncclient-4.1.jar
- httpclient-4.4.1.jar
- :x: **commons-codec-1.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/the-benchmarker/web-frameworks/commit/ab64b8404e01abede0aa4aa810306b3705409b30">ab64b8404e01abede0aa4aa810306b3705409b30</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation.
<p>Publish Date: 2019-05-20
<p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p>
<p>Release Date: 2019-05-12</p>
<p>Fix Resolution: 1.13-RC1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_usab | ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file tmp ws scm web frameworks java rapidoid pom xml path to vulnerable library root repository commons codec commons codec commons codec jar dependency hierarchy rapidoid web jar root library rapidoid http client jar httpasyncclient jar httpclient jar x commons codec jar vulnerable library found in head commit a href vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource | 0 |
2,162 | 3,051,968,577 | IssuesEvent | 2015-08-12 12:14:18 | mesosphere/marathon | https://api.github.com/repos/mesosphere/marathon | opened | Docker container settings dialog needs better error handling | gui usability | The docker container settings in the app edit/creation modal dialog has only rudimentary error handling. We should do it like at the other app modal fields.
AC.
1. Container fields needs a client side validation
2. The API error response must be parsed and displayed at the related field. | True | Docker container settings dialog needs better error handling - The docker container settings in the app edit/creation modal dialog has only rudimentary error handling. We should do it like at the other app modal fields.
AC.
1. Container fields needs a client side validation
2. The API error response must be parsed and displayed at the related field. | usab | docker container settings dialog needs better error handling the docker container settings in the app edit creation modal dialog has only rudimentary error handling we should do it like at the other app modal fields ac container fields needs a client side validation the api error response must be parsed and displayed at the related field | 1 |
298,407 | 25,824,046,011 | IssuesEvent | 2022-12-12 11:38:15 | UCL/TDMS | https://api.github.com/repos/UCL/TDMS | closed | Add/extend unit tests for classes in field.h and arrays.h | testing priority:1 | Field has some tests, but these can probably be extended. Also need to extend to test functionality in arrays.h
This will be important to ensure we don't break anything when replacing MATLAB.
In particular: allocation and deallocation of instances of these classes.
## Tracking
### Basic Types
- [x] #152
- [x] #153
- [x] [`MaterialCollection`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L89-L92) -> Doesn't actually have any attributes, and only has _one_ protected method `init_xyz_vectors`. Ergo, doesn't have it's own test file.
- [x] #158
- [x] #154
- [x] #155
- [x] #156
### Classes that have base classes as attributes
- [x] [`CCollectionBase`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L94-L99) -> Doesn't have a constructor or any methods, and just stores three `XYZVectors` as attributes.
- [x] [`DCollectionBase`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L119-L123) -> Doesn't have a constructor or any methods, and just stores two `XYZVectors` as attributes.
- [x] #161
- [x] #162
- [x] #163
- [x] #166
- [x] #173
- [x] #167
### Subclasses Types
- [x] #178
- [x] #184
- [x] #179
- [x] #185
- [x] #174
- [x] #177
- [x] #175
- [x] #172
- [x] #170
- [x] `CCoefficientMatrix` (`Matrix<double>`) - No additional functionality on top of the base class.
- [x] #176
### Test classes for cleanup
- [ ] #165
## `Field.h` classes
- [x] Expand existing tests, add Docstrings where possible, add unit tests that are missing | 1.0 | Add/extend unit tests for classes in field.h and arrays.h - Field has some tests, but these can probably be extended. Also need to extend to test functionality in arrays.h
This will be important to ensure we don't break anything when replacing MATLAB.
In particular: allocation and deallocation of instances of these classes.
## Tracking
### Basic Types
- [x] #152
- [x] #153
- [x] [`MaterialCollection`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L89-L92) -> Doesn't actually have any attributes, and only has _one_ protected method `init_xyz_vectors`. Ergo, doesn't have it's own test file.
- [x] #158
- [x] #154
- [x] #155
- [x] #156
### Classes that have base classes as attributes
- [x] [`CCollectionBase`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L94-L99) -> Doesn't have a constructor or any methods, and just stores three `XYZVectors` as attributes.
- [x] [`DCollectionBase`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L119-L123) -> Doesn't have a constructor or any methods, and just stores two `XYZVectors` as attributes.
- [x] #161
- [x] #162
- [x] #163
- [x] #166
- [x] #173
- [x] #167
### Subclasses Types
- [x] #178
- [x] #184
- [x] #179
- [x] #185
- [x] #174
- [x] #177
- [x] #175
- [x] #172
- [x] #170
- [x] `CCoefficientMatrix` (`Matrix<double>`) - No additional functionality on top of the base class.
- [x] #176
### Test classes for cleanup
- [ ] #165
## `Field.h` classes
- [x] Expand existing tests, add Docstrings where possible, add unit tests that are missing | non_usab | add extend unit tests for classes in field h and arrays h field has some tests but these can probably be extended also need to extend to test functionality in arrays h this will be important to ensure we don t break anything when replacing matlab in particular allocation and deallocation of instances of these classes tracking basic types doesn t actually have any attributes and only has one protected method init xyz vectors ergo doesn t have it s own test file classes that have base classes as attributes doesn t have a constructor or any methods and just stores three xyzvectors as attributes doesn t have a constructor or any methods and just stores two xyzvectors as attributes subclasses types ccoefficientmatrix matrix no additional functionality on top of the base class test classes for cleanup field h classes expand existing tests add docstrings where possible add unit tests that are missing | 0 |
241,169 | 18,430,268,208 | IssuesEvent | 2021-10-14 06:44:35 | amzn/selling-partner-api-docs | https://api.github.com/repos/amzn/selling-partner-api-docs | opened | [BUG] Documentation | bug documentation | **If you need help troubleshooting a problem with Selling Partner APIs, please [open a support case](https://developer.amazonservices.com/support). Issues are for reporting documentation bugs and enhancements only.**
**Describe the bug**
I wanted to use the API to block Amazon Logistics, but In creating the outbound order interface, use the featureConstraints field and set FeatureSettings to prevent Amazon logistics from being invalid.
FeatureName is BLOCK_AMZL, and featureFulfillmentPolicy is Required.
**URL of the content on GitHub**
/fba/outbound/2020-07-01/fulfillmentOrders
**Description of what the documentation currently says**
FeatureSettings allows users to apply fulfillment features to an order. To block an order from being shipped using Amazon Logistics (AMZL) and an AMZL tracking number, use featureName as BLOCK_AMZL and featureFulfillmentPolicy as Required. Blocking AMZL will incur an additional fee surcharge on your MCF orders and increase the risk of some of your orders being unfulfilled or delivered late if there are no alternative carriers available. Using BLOCK_AMZL in an order request will take precedence over your Seller Central account setting.
**Screenshots**
#########This is my request data#########
{
"marketplaceId":null,
"sellerFulfillmentOrderId":"#36976-TEST5",
"displayableOrderId":"#36976-TEST5",
"displayableOrderDate":"2021-10-14",
"displayableOrderComment":"Thank you for your order!",
"shippingSpeedCategory":"STANDARD",
"deliveryWindow":null,
"destinationAddress":{
"name":"Dawn Horney",
"addressLine1":"14B Garden Ridge Ave",
"addressLine2":"",
"addressLine3":"",
"city":"Greendale",
"districtOrCounty":null,
"stateOrRegion":"IN",
"postalCode":"47025",
"countryCode":"US",
"phone":"(859) 609-2571"
},
"fulfillmentAction":"HOLD",
"fulfillmentPolicy":null,
"codSettings":null,
"shipFromCountryCode":null,
"notificationEmails":null,
"featureConstraints":[
{
"featureName":"BLOCK_AMZL",
"featureFulfillmentPolicy":"REQUIRED"
}
],
"items":[
{
"sellerSku":"Womens-XP9303ZLUS9",
"sellerFulfillmentOrderItemId":"#36976-0",
"quantity":1,
"giftMessage":null,
"displayableComment":"Thank you for your order!",
"fulfillmentNetworkSku":null,
"perUnitDeclaredValue":null,
"perUnitPrice":null,
"perUnitTax":null
}
]
}
##########This is the returned data#############
Contains in the returned data,But Amazon's multi-channel distribution background is not to be stopped。
featureConstraints: [class FeatureSettings {
featureName: BLANK_BOX
featureFulfillmentPolicy: NotRequired
}, class FeatureSettings {
featureName: BLOCK_AMZL
featureFulfillmentPolicy: Required
}]
class GetFulfillmentOrderResponse {
payload: class GetFulfillmentOrderResult {
fulfillmentOrder: class FulfillmentOrder {
sellerFulfillmentOrderId: #36976-TEST5
marketplaceId: ATVPDKIKX0DER
displayableOrderId: #36976-TEST5
displayableOrderDate: 2021-10-14T00:00:00Z
displayableOrderComment: Thank you for your order!
shippingSpeedCategory: Standard
deliveryWindow: null
destinationAddress: class Address {
name: Dawn Horney
addressLine1: 14B Garden Ridge Ave
addressLine2: null
addressLine3: null
city: Greendale
districtOrCounty: null
stateOrRegion: IN
postalCode: 47025
countryCode: US
phone: (859) 609-2571
}
fulfillmentAction: Hold
fulfillmentPolicy: FillOrKill
codSettings: null
receivedDate: 2021-10-14T06:32:42Z
fulfillmentOrderStatus: Received
statusUpdatedDate: 2021-10-14T06:32:46Z
notificationEmails: null
featureConstraints: [class FeatureSettings {
featureName: BLANK_BOX
featureFulfillmentPolicy: NotRequired
}, class FeatureSettings {
featureName: BLOCK_AMZL
featureFulfillmentPolicy: Required
}]
}
fulfillmentOrderItems: class FulfillmentOrderItemList {
[class FulfillmentOrderItem {
sellerSku: Womens-XP9303ZLUS9
sellerFulfillmentOrderItemId: #36976-0
quantity: 1
giftMessage: null
displayableComment: Thank you for your order!
fulfillmentNetworkSku: X0027JPGVL
orderItemDisposition: Sellable
cancelledQuantity: 0
unfulfillableQuantity: 0
estimatedShipDate: 2021-10-16T06:59:59Z
estimatedArrivalDate: 2021-10-21T06:59:59Z
perUnitPrice: null
perUnitTax: null
perUnitDeclaredValue: class Money {
currencyCode: USD
value: 78.85
}
}]
}
fulfillmentShipments: class FulfillmentShipmentList {
[]
}
returnItems: class ReturnItemList {
[]
}
returnAuthorizations: class ReturnAuthorizationList {
[]
}
}
errors: null
}

| 1.0 | [BUG] Documentation - **If you need help troubleshooting a problem with Selling Partner APIs, please [open a support case](https://developer.amazonservices.com/support). Issues are for reporting documentation bugs and enhancements only.**
**Describe the bug**
I wanted to use the API to block Amazon Logistics, but In creating the outbound order interface, use the featureConstraints field and set FeatureSettings to prevent Amazon logistics from being invalid.
FeatureName is BLOCK_AMZL, and featureFulfillmentPolicy is Required.
**URL of the content on GitHub**
/fba/outbound/2020-07-01/fulfillmentOrders
**Description of what the documentation currently says**
FeatureSettings allows users to apply fulfillment features to an order. To block an order from being shipped using Amazon Logistics (AMZL) and an AMZL tracking number, use featureName as BLOCK_AMZL and featureFulfillmentPolicy as Required. Blocking AMZL will incur an additional fee surcharge on your MCF orders and increase the risk of some of your orders being unfulfilled or delivered late if there are no alternative carriers available. Using BLOCK_AMZL in an order request will take precedence over your Seller Central account setting.
**Screenshots**
#########This is my request data#########
{
"marketplaceId":null,
"sellerFulfillmentOrderId":"#36976-TEST5",
"displayableOrderId":"#36976-TEST5",
"displayableOrderDate":"2021-10-14",
"displayableOrderComment":"Thank you for your order!",
"shippingSpeedCategory":"STANDARD",
"deliveryWindow":null,
"destinationAddress":{
"name":"Dawn Horney",
"addressLine1":"14B Garden Ridge Ave",
"addressLine2":"",
"addressLine3":"",
"city":"Greendale",
"districtOrCounty":null,
"stateOrRegion":"IN",
"postalCode":"47025",
"countryCode":"US",
"phone":"(859) 609-2571"
},
"fulfillmentAction":"HOLD",
"fulfillmentPolicy":null,
"codSettings":null,
"shipFromCountryCode":null,
"notificationEmails":null,
"featureConstraints":[
{
"featureName":"BLOCK_AMZL",
"featureFulfillmentPolicy":"REQUIRED"
}
],
"items":[
{
"sellerSku":"Womens-XP9303ZLUS9",
"sellerFulfillmentOrderItemId":"#36976-0",
"quantity":1,
"giftMessage":null,
"displayableComment":"Thank you for your order!",
"fulfillmentNetworkSku":null,
"perUnitDeclaredValue":null,
"perUnitPrice":null,
"perUnitTax":null
}
]
}
##########This is the returned data#############
Contains in the returned data,But Amazon's multi-channel distribution background is not to be stopped。
featureConstraints: [class FeatureSettings {
featureName: BLANK_BOX
featureFulfillmentPolicy: NotRequired
}, class FeatureSettings {
featureName: BLOCK_AMZL
featureFulfillmentPolicy: Required
}]
class GetFulfillmentOrderResponse {
payload: class GetFulfillmentOrderResult {
fulfillmentOrder: class FulfillmentOrder {
sellerFulfillmentOrderId: #36976-TEST5
marketplaceId: ATVPDKIKX0DER
displayableOrderId: #36976-TEST5
displayableOrderDate: 2021-10-14T00:00:00Z
displayableOrderComment: Thank you for your order!
shippingSpeedCategory: Standard
deliveryWindow: null
destinationAddress: class Address {
name: Dawn Horney
addressLine1: 14B Garden Ridge Ave
addressLine2: null
addressLine3: null
city: Greendale
districtOrCounty: null
stateOrRegion: IN
postalCode: 47025
countryCode: US
phone: (859) 609-2571
}
fulfillmentAction: Hold
fulfillmentPolicy: FillOrKill
codSettings: null
receivedDate: 2021-10-14T06:32:42Z
fulfillmentOrderStatus: Received
statusUpdatedDate: 2021-10-14T06:32:46Z
notificationEmails: null
featureConstraints: [class FeatureSettings {
featureName: BLANK_BOX
featureFulfillmentPolicy: NotRequired
}, class FeatureSettings {
featureName: BLOCK_AMZL
featureFulfillmentPolicy: Required
}]
}
fulfillmentOrderItems: class FulfillmentOrderItemList {
[class FulfillmentOrderItem {
sellerSku: Womens-XP9303ZLUS9
sellerFulfillmentOrderItemId: #36976-0
quantity: 1
giftMessage: null
displayableComment: Thank you for your order!
fulfillmentNetworkSku: X0027JPGVL
orderItemDisposition: Sellable
cancelledQuantity: 0
unfulfillableQuantity: 0
estimatedShipDate: 2021-10-16T06:59:59Z
estimatedArrivalDate: 2021-10-21T06:59:59Z
perUnitPrice: null
perUnitTax: null
perUnitDeclaredValue: class Money {
currencyCode: USD
value: 78.85
}
}]
}
fulfillmentShipments: class FulfillmentShipmentList {
[]
}
returnItems: class ReturnItemList {
[]
}
returnAuthorizations: class ReturnAuthorizationList {
[]
}
}
errors: null
}

| non_usab | documentation if you need help troubleshooting a problem with selling partner apis please issues are for reporting documentation bugs and enhancements only describe the bug i wanted to use the api to block amazon logistics but in creating the outbound order interface use the featureconstraints field and set featuresettings to prevent amazon logistics from being invalid featurename is block amzl and featurefulfillmentpolicy is required url of the content on github fba outbound fulfillmentorders description of what the documentation currently says featuresettings allows users to apply fulfillment features to an order to block an order from being shipped using amazon logistics amzl and an amzl tracking number use featurename as block amzl and featurefulfillmentpolicy as required blocking amzl will incur an additional fee surcharge on your mcf orders and increase the risk of some of your orders being unfulfilled or delivered late if there are no alternative carriers available using block amzl in an order request will take precedence over your seller central account setting screenshots this is my request data marketplaceid null sellerfulfillmentorderid displayableorderid displayableorderdate displayableordercomment thank you for your order shippingspeedcategory standard deliverywindow null destinationaddress name dawn horney garden ridge ave city greendale districtorcounty null stateorregion in postalcode countrycode us phone fulfillmentaction hold fulfillmentpolicy null codsettings null shipfromcountrycode null notificationemails null featureconstraints featurename block amzl featurefulfillmentpolicy required items sellersku womens sellerfulfillmentorderitemid quantity giftmessage null displayablecomment thank you for your order fulfillmentnetworksku null perunitdeclaredvalue null perunitprice null perunittax null this is the returned data contains in the returned data,but amazon s multi channel distribution background is not to be stopped。 featureconstraints class featuresettings featurename blank box featurefulfillmentpolicy notrequired class featuresettings featurename block amzl featurefulfillmentpolicy required class getfulfillmentorderresponse payload class getfulfillmentorderresult fulfillmentorder class fulfillmentorder sellerfulfillmentorderid marketplaceid displayableorderid displayableorderdate displayableordercomment thank you for your order shippingspeedcategory standard deliverywindow null destinationaddress class address name dawn horney garden ridge ave null null city greendale districtorcounty null stateorregion in postalcode countrycode us phone fulfillmentaction hold fulfillmentpolicy fillorkill codsettings null receiveddate fulfillmentorderstatus received statusupdateddate notificationemails null featureconstraints class featuresettings featurename blank box featurefulfillmentpolicy notrequired class featuresettings featurename block amzl featurefulfillmentpolicy required fulfillmentorderitems class fulfillmentorderitemlist class fulfillmentorderitem sellersku womens sellerfulfillmentorderitemid quantity giftmessage null displayablecomment thank you for your order fulfillmentnetworksku orderitemdisposition sellable cancelledquantity unfulfillablequantity estimatedshipdate estimatedarrivaldate perunitprice null perunittax null perunitdeclaredvalue class money currencycode usd value fulfillmentshipments class fulfillmentshipmentlist returnitems class returnitemlist returnauthorizations class returnauthorizationlist errors null | 0 |
1,847 | 3,025,053,800 | IssuesEvent | 2015-08-03 04:25:40 | lionheart/openradar-mirror | https://api.github.com/repos/lionheart/openradar-mirror | opened | 21745023: App Store trials! | classification:ui/usability reproducible:always status:open | #### Description
This is a duplicate of rdar://13197637
I'm sure you've got this request many times, but I'm adding my voice to it. Here’s how I could see it working:
Developers can choose whether to allow a trial of 1, 7 or 30 days, or to disallow trials all together, on a per-app basis. For those apps that allow trials, the App Store would show a “Try for 7 days” button alongside “Buy app”.
If you install the trial app, it gets flagged with a “Trial” flag over the icon, in a similar way to the “New” flag that was introduced in iOS 6.
When you run the app, it permanently runs with a double height status bar, which shows the amount of time left. Tapping this status bar opens the app store page allowing you to purchase the app.
Once a trial is used up, you can’t get a trial of that app again on the same Apple ID. When the trial is used up, the app icon still remains on your device, but the flag says “Expired” and tapping on it takes you to the App Store rather than opening the app.
-
Product Version: n/a
Created: 2015-07-09T14:08:10.128640
Originated: 2015-07-09T16:08:00
Open Radar Link: http://www.openradar.me/21745023 | True | 21745023: App Store trials! - #### Description
This is a duplicate of rdar://13197637
I'm sure you've got this request many times, but I'm adding my voice to it. Here’s how I could see it working:
Developers can choose whether to allow a trial of 1, 7 or 30 days, or to disallow trials all together, on a per-app basis. For those apps that allow trials, the App Store would show a “Try for 7 days” button alongside “Buy app”.
If you install the trial app, it gets flagged with a “Trial” flag over the icon, in a similar way to the “New” flag that was introduced in iOS 6.
When you run the app, it permanently runs with a double height status bar, which shows the amount of time left. Tapping this status bar opens the app store page allowing you to purchase the app.
Once a trial is used up, you can’t get a trial of that app again on the same Apple ID. When the trial is used up, the app icon still remains on your device, but the flag says “Expired” and tapping on it takes you to the App Store rather than opening the app.
-
Product Version: n/a
Created: 2015-07-09T14:08:10.128640
Originated: 2015-07-09T16:08:00
Open Radar Link: http://www.openradar.me/21745023 | usab | app store trials description this is a duplicate of rdar i m sure you ve got this request many times but i m adding my voice to it here’s how i could see it working developers can choose whether to allow a trial of or days or to disallow trials all together on a per app basis for those apps that allow trials the app store would show a “try for days” button alongside “buy app” if you install the trial app it gets flagged with a “trial” flag over the icon in a similar way to the “new” flag that was introduced in ios when you run the app it permanently runs with a double height status bar which shows the amount of time left tapping this status bar opens the app store page allowing you to purchase the app once a trial is used up you can’t get a trial of that app again on the same apple id when the trial is used up the app icon still remains on your device but the flag says “expired” and tapping on it takes you to the app store rather than opening the app product version n a created originated open radar link | 1 |
6,498 | 4,321,306,941 | IssuesEvent | 2016-07-25 09:40:44 | rpi-virtuell/reliwerk | https://api.github.com/repos/rpi-virtuell/reliwerk | closed | "Gruppe verlassen" | Text Übersetzung Usability | Nun ist genau das passiert, was ich immer befürchtet habe: das riesige "Gruppe verlassen" - Schild verführt einfach dazu, darauf zu klicken, wenn man die Arbeit in der Gruppe beendet hat. Warum diese Aufforderung zum endgültigen Verlassen so groß da stehen muss, erschließt sich mir nicht. Kann man das nicht klein oben rein machen? Es ist wirklich irritierend. Heute Morgen habe ich mit Müh und Not eine Kollegin in die Gruppe "geschleußt" und jetzt ist sie wieder draußen. Ich bitte sehr darum, diesen Button zu ändern. | True | "Gruppe verlassen" - Nun ist genau das passiert, was ich immer befürchtet habe: das riesige "Gruppe verlassen" - Schild verführt einfach dazu, darauf zu klicken, wenn man die Arbeit in der Gruppe beendet hat. Warum diese Aufforderung zum endgültigen Verlassen so groß da stehen muss, erschließt sich mir nicht. Kann man das nicht klein oben rein machen? Es ist wirklich irritierend. Heute Morgen habe ich mit Müh und Not eine Kollegin in die Gruppe "geschleußt" und jetzt ist sie wieder draußen. Ich bitte sehr darum, diesen Button zu ändern. | usab | gruppe verlassen nun ist genau das passiert was ich immer befürchtet habe das riesige gruppe verlassen schild verführt einfach dazu darauf zu klicken wenn man die arbeit in der gruppe beendet hat warum diese aufforderung zum endgültigen verlassen so groß da stehen muss erschließt sich mir nicht kann man das nicht klein oben rein machen es ist wirklich irritierend heute morgen habe ich mit müh und not eine kollegin in die gruppe geschleußt und jetzt ist sie wieder draußen ich bitte sehr darum diesen button zu ändern | 1 |
289,486 | 31,933,038,899 | IssuesEvent | 2023-09-19 08:42:37 | Trinadh465/linux-4.1.15_CVE-2023-4128 | https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-4128 | opened | CVE-2022-34918 (High) detected in linuxlinux-4.6 | Mend: dependency security vulnerability | ## CVE-2022-34918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel through 5.18.9. A type confusion bug in nft_set_elem_init (leading to a buffer overflow) could be used by a local attacker to escalate privileges, a different vulnerability than CVE-2022-32250. (The attacker can obtain root access, but must start with an unprivileged user namespace to obtain CAP_NET_ADMIN access.) This can be fixed in nft_setelem_parse_data in net/netfilter/nf_tables_api.c.
<p>Publish Date: 2022-07-04
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-34918>CVE-2022-34918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-34918">https://www.linuxkernelcves.com/cves/CVE-2022-34918</a></p>
<p>Release Date: 2022-07-04</p>
<p>Fix Resolution: v5.10.130,v5.15.54,v5.18.11,v5.19-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | CVE-2022-34918 (High) detected in linuxlinux-4.6 - ## CVE-2022-34918 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary>
<p>
<p>The Linux Kernel</p>
<p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p>
<p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel through 5.18.9. A type confusion bug in nft_set_elem_init (leading to a buffer overflow) could be used by a local attacker to escalate privileges, a different vulnerability than CVE-2022-32250. (The attacker can obtain root access, but must start with an unprivileged user namespace to obtain CAP_NET_ADMIN access.) This can be fixed in nft_setelem_parse_data in net/netfilter/nf_tables_api.c.
<p>Publish Date: 2022-07-04
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-34918>CVE-2022-34918</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-34918">https://www.linuxkernelcves.com/cves/CVE-2022-34918</a></p>
<p>Release Date: 2022-07-04</p>
<p>Fix Resolution: v5.10.130,v5.15.54,v5.18.11,v5.19-rc6</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_usab | cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net netfilter nf tables api c net netfilter nf tables api c vulnerability details an issue was discovered in the linux kernel through a type confusion bug in nft set elem init leading to a buffer overflow could be used by a local attacker to escalate privileges a different vulnerability than cve the attacker can obtain root access but must start with an unprivileged user namespace to obtain cap net admin access this can be fixed in nft setelem parse data in net netfilter nf tables api c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend | 0 |
5,260 | 3,194,296,432 | IssuesEvent | 2015-09-30 11:17:05 | VcDevel/Vc | https://api.github.com/repos/VcDevel/Vc | closed | drop all internal namespaces and move everything to Vc::Detail | code cleanup | Right now there are several internal namespaces hiding implementation details. All of these namespaces need to be folded to `Vc::Detail`. | 1.0 | drop all internal namespaces and move everything to Vc::Detail - Right now there are several internal namespaces hiding implementation details. All of these namespaces need to be folded to `Vc::Detail`. | non_usab | drop all internal namespaces and move everything to vc detail right now there are several internal namespaces hiding implementation details all of these namespaces need to be folded to vc detail | 0 |
17,839 | 12,373,218,495 | IssuesEvent | 2020-05-18 22:01:12 | argoproj/argo-cd | https://api.github.com/repos/argoproj/argo-cd | reopened | Display duration for the entire sync and deploy/rollout for a revision in history window | component:ui enhancement good first issue type:usability | # Summary
Display duration for the entire sync and deploy/rollout for a revision in history window. Currently the "History and Rollback" window shows revision and timing when it started but not clearly shows the entire sync/deploy duration it took to complete it.
It will be good to see duration as shown below in Red font.

What change you think needs making.
The sync history window needs to be updated to show duration for each sync which occurred in the past.
# Motivation
This will help the user to get the trend of past sync durations and see how much each of the previous sync took and make decision based on that.
Please give examples of your use case, e.g. when would you use this.
To view past sync durations and see its trend.
# Proposal
How do you think this should be implemented?
| True | Display duration for the entire sync and deploy/rollout for a revision in history window - # Summary
Display duration for the entire sync and deploy/rollout for a revision in history window. Currently the "History and Rollback" window shows revision and timing when it started but not clearly shows the entire sync/deploy duration it took to complete it.
It will be good to see duration as shown below in Red font.

What change you think needs making.
The sync history window needs to be updated to show duration for each sync which occurred in the past.
# Motivation
This will help the user to get the trend of past sync durations and see how much each of the previous sync took and make decision based on that.
Please give examples of your use case, e.g. when would you use this.
To view past sync durations and see its trend.
# Proposal
How do you think this should be implemented?
| usab | display duration for the entire sync and deploy rollout for a revision in history window summary display duration for the entire sync and deploy rollout for a revision in history window currently the history and rollback window shows revision and timing when it started but not clearly shows the entire sync deploy duration it took to complete it it will be good to see duration as shown below in red font what change you think needs making the sync history window needs to be updated to show duration for each sync which occurred in the past motivation this will help the user to get the trend of past sync durations and see how much each of the previous sync took and make decision based on that please give examples of your use case e g when would you use this to view past sync durations and see its trend proposal how do you think this should be implemented | 1 |
7,555 | 5,050,547,155 | IssuesEvent | 2016-12-20 19:05:11 | loconomics/loconomics | https://api.github.com/repos/loconomics/loconomics | closed | Cache data that's already been inputted into forms | C: Usability D: Duplicated F: Onboarding P3 R1: Ready for Development Sprint | From @jucsanch: He left to a help page and came back and data was gone. | True | Cache data that's already been inputted into forms - From @jucsanch: He left to a help page and came back and data was gone. | usab | cache data that s already been inputted into forms from jucsanch he left to a help page and came back and data was gone | 1 |
14,437 | 9,186,911,478 | IssuesEvent | 2019-03-06 00:36:40 | uber/hudi | https://api.github.com/repos/uber/hudi | closed | Reduce memory usage on WriteStatus for implicit indexes | enhancement usability | For things like BloomIndex, there is no need to keep accumulating the hoodiekey/recordlocation per IOHandle to be passed to index update..
Fixing this, can greatly ease bulk inserts/upserts with large number of keys and small record sizes | True | Reduce memory usage on WriteStatus for implicit indexes - For things like BloomIndex, there is no need to keep accumulating the hoodiekey/recordlocation per IOHandle to be passed to index update..
Fixing this, can greatly ease bulk inserts/upserts with large number of keys and small record sizes | usab | reduce memory usage on writestatus for implicit indexes for things like bloomindex there is no need to keep accumulating the hoodiekey recordlocation per iohandle to be passed to index update fixing this can greatly ease bulk inserts upserts with large number of keys and small record sizes | 1 |
11,827 | 7,477,842,513 | IssuesEvent | 2018-04-04 09:35:58 | cnr-ibf-pa/hbp-bsp-issues | https://api.github.com/repos/cnr-ibf-pa/hbp-bsp-issues | opened | Analysis notebook change for SynapticEventsFitting | Type_BUG Type_Usability UC_SynapEvFitting_UData UC_SynapEvFitting_UModel | New Analysis notebooks require:
- [ ] User documentation update required to explain the new GUI to the user
- [ ] Notebook documentation update required to show the timestamp to the user and have the user remember it for the Analysis step
- [ ] Make sure that in either case (NSG, Jureca) the correct (European) time is shown
| True | Analysis notebook change for SynapticEventsFitting - New Analysis notebooks require:
- [ ] User documentation update required to explain the new GUI to the user
- [ ] Notebook documentation update required to show the timestamp to the user and have the user remember it for the Analysis step
- [ ] Make sure that in either case (NSG, Jureca) the correct (European) time is shown
| usab | analysis notebook change for synapticeventsfitting new analysis notebooks require user documentation update required to explain the new gui to the user notebook documentation update required to show the timestamp to the user and have the user remember it for the analysis step make sure that in either case nsg jureca the correct european time is shown | 1 |
5,940 | 4,077,223,909 | IssuesEvent | 2016-05-30 06:57:04 | godotengine/godot | https://api.github.com/repos/godotengine/godot | opened | Changing the language directly in the project manager | enhancement topic:editor usability | Because we can, and since we spawn a new process for the editor anyway, it would be the right time to select another language.
In a more practical context, it would also make it very easy to find if we want to encourage people to make screenshots in English for bug reports, and not in their own locale. | True | Changing the language directly in the project manager - Because we can, and since we spawn a new process for the editor anyway, it would be the right time to select another language.
In a more practical context, it would also make it very easy to find if we want to encourage people to make screenshots in English for bug reports, and not in their own locale. | usab | changing the language directly in the project manager because we can and since we spawn a new process for the editor anyway it would be the right time to select another language in a more practical context it would also make it very easy to find if we want to encourage people to make screenshots in english for bug reports and not in their own locale | 1 |
1,767 | 3,004,048,870 | IssuesEvent | 2015-07-25 14:43:03 | MDAnalysis/mdanalysis | https://api.github.com/repos/MDAnalysis/mdanalysis | closed | MDAnalysis.analysis.distances.contact_matrix | Component-Docs usability | The docstring in `0.11` for this function incorrectly suggests that a list of coordinates is used as input, while a numpy array is clearly the correct input (list won't work).
Also, in practice I am not seeing the progress meter before or after 0.11 (mac or linux, even with a large input array), which I was testing for when trying to write some user documentation for `0.11` API updates [MDAnalysis 0.11 unifying release user guide: Suppressing contact matrix progress meter](/MDAnalysis/mdanalysis/wiki/MDAnalysis-0.11-unifying-release-user-guide#suppressing-contact-matrix-progress-meter-new-quiet-keyword). | True | MDAnalysis.analysis.distances.contact_matrix - The docstring in `0.11` for this function incorrectly suggests that a list of coordinates is used as input, while a numpy array is clearly the correct input (list won't work).
Also, in practice I am not seeing the progress meter before or after 0.11 (mac or linux, even with a large input array), which I was testing for when trying to write some user documentation for `0.11` API updates [MDAnalysis 0.11 unifying release user guide: Suppressing contact matrix progress meter](/MDAnalysis/mdanalysis/wiki/MDAnalysis-0.11-unifying-release-user-guide#suppressing-contact-matrix-progress-meter-new-quiet-keyword). | usab | mdanalysis analysis distances contact matrix the docstring in for this function incorrectly suggests that a list of coordinates is used as input while a numpy array is clearly the correct input list won t work also in practice i am not seeing the progress meter before or after mac or linux even with a large input array which i was testing for when trying to write some user documentation for api updates mdanalysis mdanalysis wiki mdanalysis unifying release user guide suppressing contact matrix progress meter new quiet keyword | 1 |
5,599 | 3,969,158,476 | IssuesEvent | 2016-05-03 22:17:31 | pipelinedb/pipelinedb | https://api.github.com/repos/pipelinedb/pipelinedb | opened | Add backfill utility | usability | It should use `COPY` to bulkload historical data into a target set of CVs. | True | Add backfill utility - It should use `COPY` to bulkload historical data into a target set of CVs. | usab | add backfill utility it should use copy to bulkload historical data into a target set of cvs | 1 |
20,952 | 16,357,306,383 | IssuesEvent | 2021-05-14 01:39:38 | microsoft/win32metadata | https://api.github.com/repos/microsoft/win32metadata | closed | Windows.Win32.winmd needs to be self-describing | usability | The winmd lacks the definitions of various attributes and types like `SupportedArchitecture` - everything in the `Windows.Win32.Interop` namespace is missing - so there's no way for tools to parse these attributes without a priori knowledge of them. WinRT metadata includes the definitions of all the attributes it defines in the `Windows.Foundation.Metadata` namespace. Win32 should do the same. | True | Windows.Win32.winmd needs to be self-describing - The winmd lacks the definitions of various attributes and types like `SupportedArchitecture` - everything in the `Windows.Win32.Interop` namespace is missing - so there's no way for tools to parse these attributes without a priori knowledge of them. WinRT metadata includes the definitions of all the attributes it defines in the `Windows.Foundation.Metadata` namespace. Win32 should do the same. | usab | windows winmd needs to be self describing the winmd lacks the definitions of various attributes and types like supportedarchitecture everything in the windows interop namespace is missing so there s no way for tools to parse these attributes without a priori knowledge of them winrt metadata includes the definitions of all the attributes it defines in the windows foundation metadata namespace should do the same | 1 |
16,010 | 10,481,504,066 | IssuesEvent | 2019-09-24 09:49:33 | pombase/canto | https://api.github.com/repos/pombase/canto | closed | Add help text for summary page genotype heading | PHI-Canto feedback_required usability | Pull request #1649 added a new heading on the summary page to separate the genotype links from the gene links. However this new heading doesn't have any help text associated with it, whereas the gene heading does. Is it worth adding any help text? | True | Add help text for summary page genotype heading - Pull request #1649 added a new heading on the summary page to separate the genotype links from the gene links. However this new heading doesn't have any help text associated with it, whereas the gene heading does. Is it worth adding any help text? | usab | add help text for summary page genotype heading pull request added a new heading on the summary page to separate the genotype links from the gene links however this new heading doesn t have any help text associated with it whereas the gene heading does is it worth adding any help text | 1 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.