Unnamed: 0
int64
3
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
2
742
labels
stringlengths
4
431
body
stringlengths
5
239k
index
stringclasses
10 values
text_combine
stringlengths
96
240k
label
stringclasses
2 values
text
stringlengths
96
200k
binary_label
int64
0
1
32,270
6,756,696,827
IssuesEvent
2017-10-24 08:10:12
primefaces/primeng
https://api.github.com/repos/primefaces/primeng
closed
MegaMenu doesn't compile with TypeScript 2.4
confirmed defect
**I'm submitting a ...** ``` [X] bug report ``` **Test case** You can use the following demo app as test case: https://github.com/ova2/angular-development-with-primeng/tree/master/chapter7/megamenu **Current behavior** If you run the showcase for the MegaMenu with TypeScript 2.4 or run the demo app linked above, you will get a compilation error like ``` Type '{ label: string; items: { label: string; }[]; }[]' has no properties in common with type 'MenuItem'. ``` **Expected behavior** There should be no compilation error, as when compiling with TypeScript 2.3. **Minimal reproduction of the problem with instructions** * Install the above app * or install the current master of PrimeNG and change the requirements in package.json to Angular 4.3, Angular-Cli 1.3, TypeScript 2.4 (lower Angular/Cli versions require TypeScript 2.3, so you need to test with Angular 4.3 and Cli 1.3) * Run `npm` install and `npm start` and check the MegaMenu * **Angular version:** 4.3.3 * **PrimeNG version:** 4.1.2 * **Browser:** all * **Language:** TypeScript 2.4 * **Node (for AoT issues):** 8.1.4 * **Analysis of the problem:** In the `MenuItem` interface, the `items` property is defined as of type `MenuItem[]`. But in the MegaMenu, you can have arrays of arrays of MenuItems as items, not just arrays of MenuItems. In TypeScript 2.4, it’s now an error to assign anything to a weak type when there’s no overlap in properties (see [here](https://www.typescriptlang.org/docs/handbook/release-notes/typescript-2-4.html#weak-type-detection)). Note that in the `MenuItem` interface all properties are marked as optional. Therefore, this is considered a "weak type". * **Proposed solution:** The `items` property in the `MenutItem` interface should be defined as follows: ``` items?: MenuItem[]|MenuItem[][]; ```
1.0
MegaMenu doesn't compile with TypeScript 2.4 - **I'm submitting a ...** ``` [X] bug report ``` **Test case** You can use the following demo app as test case: https://github.com/ova2/angular-development-with-primeng/tree/master/chapter7/megamenu **Current behavior** If you run the showcase for the MegaMenu with TypeScript 2.4 or run the demo app linked above, you will get a compilation error like ``` Type '{ label: string; items: { label: string; }[]; }[]' has no properties in common with type 'MenuItem'. ``` **Expected behavior** There should be no compilation error, as when compiling with TypeScript 2.3. **Minimal reproduction of the problem with instructions** * Install the above app * or install the current master of PrimeNG and change the requirements in package.json to Angular 4.3, Angular-Cli 1.3, TypeScript 2.4 (lower Angular/Cli versions require TypeScript 2.3, so you need to test with Angular 4.3 and Cli 1.3) * Run `npm` install and `npm start` and check the MegaMenu * **Angular version:** 4.3.3 * **PrimeNG version:** 4.1.2 * **Browser:** all * **Language:** TypeScript 2.4 * **Node (for AoT issues):** 8.1.4 * **Analysis of the problem:** In the `MenuItem` interface, the `items` property is defined as of type `MenuItem[]`. But in the MegaMenu, you can have arrays of arrays of MenuItems as items, not just arrays of MenuItems. In TypeScript 2.4, it’s now an error to assign anything to a weak type when there’s no overlap in properties (see [here](https://www.typescriptlang.org/docs/handbook/release-notes/typescript-2-4.html#weak-type-detection)). Note that in the `MenuItem` interface all properties are marked as optional. Therefore, this is considered a "weak type". * **Proposed solution:** The `items` property in the `MenutItem` interface should be defined as follows: ``` items?: MenuItem[]|MenuItem[][]; ```
non_usab
megamenu doesn t compile with typescript i m submitting a bug report test case you can use the following demo app as test case current behavior if you run the showcase for the megamenu with typescript or run the demo app linked above you will get a compilation error like type label string items label string has no properties in common with type menuitem expected behavior there should be no compilation error as when compiling with typescript minimal reproduction of the problem with instructions install the above app or install the current master of primeng and change the requirements in package json to angular angular cli typescript lower angular cli versions require typescript so you need to test with angular and cli run npm install and npm start and check the megamenu angular version primeng version browser all language typescript node for aot issues analysis of the problem in the menuitem interface the items property is defined as of type menuitem but in the megamenu you can have arrays of arrays of menuitems as items not just arrays of menuitems in typescript it’s now an error to assign anything to a weak type when there’s no overlap in properties see note that in the menuitem interface all properties are marked as optional therefore this is considered a weak type proposed solution the items property in the menutitem interface should be defined as follows items menuitem menuitem
0
8,566
5,825,824,655
IssuesEvent
2017-05-08 00:51:39
bronzehedwick/chrisdeluca
https://api.github.com/repos/bronzehedwick/chrisdeluca
closed
Add more links to navigation
usability
The main navigation should have the following links: * Now * Contact * Sections (this is categories. Maybe rename?) * RSS In addition, there should be a form to subscribe to email notifications on site updates. Depends on #61 #60 #59 #57
True
Add more links to navigation - The main navigation should have the following links: * Now * Contact * Sections (this is categories. Maybe rename?) * RSS In addition, there should be a form to subscribe to email notifications on site updates. Depends on #61 #60 #59 #57
usab
add more links to navigation the main navigation should have the following links now contact sections this is categories maybe rename rss in addition there should be a form to subscribe to email notifications on site updates depends on
1
325,413
27,876,174,470
IssuesEvent
2023-03-21 16:09:19
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
closed
E2E stream table tests - stream details panel
team/platform-move area/frontend ui/connection ui/tests e2e-testing-tool
## Tell us about the problem you're trying to solve Add E2E tests for stream details panel functionality in the new stream table ### Tasks - [x] Check panel opening (this case should be covered in https://github.com/airbytehq/airbyte/issues/22640) - [x] Check stream details info: sync state, namespace(if exist), stream name, chosen sync mode - [x] Check amount of displayed fields(depends on tests data we will use) - [x] Check that user is able to scroll a long list of fields()scrolling and desired field is visible - [x] Check that each field has info: source field name, data type, Cursor(optional), PK(optional), destination field name - [x] Check that user is able to select only one cursor value (appropriate sync mode should be chosen) - [x] Check that user is NOT able to select only one cursor value if it's source-defined (appropriate sync mode should be chosen) - [x] Check that user is able to select only multiple PK values (appropriate sync mode should be chosen) - [x] Check that user is NOT able to select PK value if it's source-defined (appropriate sync mode should be chosen)
2.0
E2E stream table tests - stream details panel - ## Tell us about the problem you're trying to solve Add E2E tests for stream details panel functionality in the new stream table ### Tasks - [x] Check panel opening (this case should be covered in https://github.com/airbytehq/airbyte/issues/22640) - [x] Check stream details info: sync state, namespace(if exist), stream name, chosen sync mode - [x] Check amount of displayed fields(depends on tests data we will use) - [x] Check that user is able to scroll a long list of fields()scrolling and desired field is visible - [x] Check that each field has info: source field name, data type, Cursor(optional), PK(optional), destination field name - [x] Check that user is able to select only one cursor value (appropriate sync mode should be chosen) - [x] Check that user is NOT able to select only one cursor value if it's source-defined (appropriate sync mode should be chosen) - [x] Check that user is able to select only multiple PK values (appropriate sync mode should be chosen) - [x] Check that user is NOT able to select PK value if it's source-defined (appropriate sync mode should be chosen)
non_usab
stream table tests stream details panel tell us about the problem you re trying to solve add tests for stream details panel functionality in the new stream table tasks check panel opening this case should be covered in check stream details info sync state namespace if exist stream name chosen sync mode check amount of displayed fields depends on tests data we will use check that user is able to scroll a long list of fields scrolling and desired field is visible check that each field has info source field name data type cursor optional pk optional destination field name check that user is able to select only one cursor value appropriate sync mode should be chosen check that user is not able to select only one cursor value if it s source defined appropriate sync mode should be chosen check that user is able to select only multiple pk values appropriate sync mode should be chosen check that user is not able to select pk value if it s source defined appropriate sync mode should be chosen
0
269,063
28,959,988,826
IssuesEvent
2023-05-10 01:06:12
dpteam/RK3188_TABLET
https://api.github.com/repos/dpteam/RK3188_TABLET
reopened
CVE-2011-2699 (High) detected in linuxv3.0
Mend: dependency security vulnerability
## CVE-2011-2699 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/net/inetpeer.h</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The IPv6 implementation in the Linux kernel before 3.1 does not generate Fragment Identification values separately for each destination, which makes it easier for remote attackers to cause a denial of service (disrupted networking) by predicting these values and sending crafted packets. <p>Publish Date: 2012-05-24 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-2699>CVE-2011-2699</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-2699">https://nvd.nist.gov/vuln/detail/CVE-2011-2699</a></p> <p>Release Date: 2012-05-24</p> <p>Fix Resolution: 3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2011-2699 (High) detected in linuxv3.0 - ## CVE-2011-2699 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxv3.0</b></p></summary> <p> <p>Linux kernel source tree</p> <p>Library home page: <a href=https://github.com/verygreen/linux.git>https://github.com/verygreen/linux.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/dpteam/RK3188_TABLET/commit/0c501f5a0fd72c7b2ac82904235363bd44fd8f9e">0c501f5a0fd72c7b2ac82904235363bd44fd8f9e</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/include/net/inetpeer.h</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> The IPv6 implementation in the Linux kernel before 3.1 does not generate Fragment Identification values separately for each destination, which makes it easier for remote attackers to cause a denial of service (disrupted networking) by predicting these values and sending crafted packets. <p>Publish Date: 2012-05-24 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2011-2699>CVE-2011-2699</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2011-2699">https://nvd.nist.gov/vuln/detail/CVE-2011-2699</a></p> <p>Release Date: 2012-05-24</p> <p>Fix Resolution: 3.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_usab
cve high detected in cve high severity vulnerability vulnerable library linux kernel source tree library home page a href found in head commit a href found in base branch master vulnerable source files include net inetpeer h vulnerability details the implementation in the linux kernel before does not generate fragment identification values separately for each destination which makes it easier for remote attackers to cause a denial of service disrupted networking by predicting these values and sending crafted packets publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
117,536
17,496,281,368
IssuesEvent
2021-08-10 01:00:40
billmcchesney1/vulnerable-rust
https://api.github.com/repos/billmcchesney1/vulnerable-rust
opened
CVE-2020-36471 (Medium) detected in generator-0.6.21.crate
security vulnerability
## CVE-2020-36471 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>generator-0.6.21.crate</b></p></summary> <p>Stackfull Generator Library in Rust</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/generator/0.6.21/download">https://crates.io/api/v1/crates/generator/0.6.21/download</a></p> <p> Dependency Hierarchy: - hyper-0.13.5.crate (Root Library) - tokio-0.2.21.crate - bytes-0.5.5.crate - loom-0.3.4.crate - :x: **generator-0.6.21.crate** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the generator crate before 0.7.0 for Rust. It does not ensure that a function (for yielding values) has Send bounds. <p>Publish Date: 2021-08-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36471>CVE-2020-36471</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2020-0151.html">https://rustsec.org/advisories/RUSTSEC-2020-0151.html</a></p> <p>Release Date: 2021-08-08</p> <p>Fix Resolution: 0.7.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Crate","packageName":"generator","packageVersion":"0.6.21","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"hyper:0.13.5;tokio:0.2.21;bytes:0.5.5;loom:0.3.4;generator:0.6.21","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.7.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36471","vulnerabilityDetails":"An issue was discovered in the generator crate before 0.7.0 for Rust. It does not ensure that a function (for yielding values) has Send bounds.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36471","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-36471 (Medium) detected in generator-0.6.21.crate - ## CVE-2020-36471 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>generator-0.6.21.crate</b></p></summary> <p>Stackfull Generator Library in Rust</p> <p>Library home page: <a href="https://crates.io/api/v1/crates/generator/0.6.21/download">https://crates.io/api/v1/crates/generator/0.6.21/download</a></p> <p> Dependency Hierarchy: - hyper-0.13.5.crate (Root Library) - tokio-0.2.21.crate - bytes-0.5.5.crate - loom-0.3.4.crate - :x: **generator-0.6.21.crate** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the generator crate before 0.7.0 for Rust. It does not ensure that a function (for yielding values) has Send bounds. <p>Publish Date: 2021-08-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36471>CVE-2020-36471</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://rustsec.org/advisories/RUSTSEC-2020-0151.html">https://rustsec.org/advisories/RUSTSEC-2020-0151.html</a></p> <p>Release Date: 2021-08-08</p> <p>Fix Resolution: 0.7.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Crate","packageName":"generator","packageVersion":"0.6.21","packageFilePaths":[],"isTransitiveDependency":true,"dependencyTree":"hyper:0.13.5;tokio:0.2.21;bytes:0.5.5;loom:0.3.4;generator:0.6.21","isMinimumFixVersionAvailable":true,"minimumFixVersion":"0.7.0"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-36471","vulnerabilityDetails":"An issue was discovered in the generator crate before 0.7.0 for Rust. It does not ensure that a function (for yielding values) has Send bounds.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36471","cvss3Severity":"medium","cvss3Score":"5.5","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_usab
cve medium detected in generator crate cve medium severity vulnerability vulnerable library generator crate stackfull generator library in rust library home page a href dependency hierarchy hyper crate root library tokio crate bytes crate loom crate x generator crate vulnerable library found in base branch master vulnerability details an issue was discovered in the generator crate before for rust it does not ensure that a function for yielding values has send bounds publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree hyper tokio bytes loom generator isminimumfixversionavailable true minimumfixversion basebranches vulnerabilityidentifier cve vulnerabilitydetails an issue was discovered in the generator crate before for rust it does not ensure that a function for yielding values has send bounds vulnerabilityurl
0
371,169
25,939,851,565
IssuesEvent
2022-12-16 17:16:26
projf/projf-explore
https://api.github.com/repos/projf/projf-explore
closed
Feedback on endianness
documentation
On page https://projectf.io/posts/numbers-in-verilog/ you mention: > Say you’ve got a bit-endian ... You probably mean big > ... byte from I2C and want to convert it to little-endian. Alas, you can’t mix big and little-endian vectors, so the following won’t work: I would pay attention to endianness in this context. In my experience endianness is mostly known in data consisting of multiple bytes. I think you are talking about the endianness of bits, which seem to be called *bit endianness* ([example](https://stackoverflow.com/questions/6043483/why-bit-endianness-is-an-issue-in-bitfields). --- Another issue I stumbled upon: > wire [0:7] i2c_byte; // 8-bit wire (big-endian) > reg [7:0] le_byte; // 8-bit reg (little-endian) > always_ff @(posedge clk) le_byte <= i2c_byte; // Won't work :( I tried: ```verilog `define LED_COUNT 16 module led_sw_all_reversed( input [`LED_COUNT-1:0] sw, output [0:`LED_COUNT-1] led // reversed bit-endianness ); assign led = sw; endmodule ``` Which worked fine using Vivado on my hardware with 16 switches and LEDs. Last but not least, I am new to Verilog. PS: 😎 way of teaching FPGAs using graphics!
1.0
Feedback on endianness - On page https://projectf.io/posts/numbers-in-verilog/ you mention: > Say you’ve got a bit-endian ... You probably mean big > ... byte from I2C and want to convert it to little-endian. Alas, you can’t mix big and little-endian vectors, so the following won’t work: I would pay attention to endianness in this context. In my experience endianness is mostly known in data consisting of multiple bytes. I think you are talking about the endianness of bits, which seem to be called *bit endianness* ([example](https://stackoverflow.com/questions/6043483/why-bit-endianness-is-an-issue-in-bitfields). --- Another issue I stumbled upon: > wire [0:7] i2c_byte; // 8-bit wire (big-endian) > reg [7:0] le_byte; // 8-bit reg (little-endian) > always_ff @(posedge clk) le_byte <= i2c_byte; // Won't work :( I tried: ```verilog `define LED_COUNT 16 module led_sw_all_reversed( input [`LED_COUNT-1:0] sw, output [0:`LED_COUNT-1] led // reversed bit-endianness ); assign led = sw; endmodule ``` Which worked fine using Vivado on my hardware with 16 switches and LEDs. Last but not least, I am new to Verilog. PS: 😎 way of teaching FPGAs using graphics!
non_usab
feedback on endianness on page you mention say you’ve got a bit endian you probably mean big byte from and want to convert it to little endian alas you can’t mix big and little endian vectors so the following won’t work i would pay attention to endianness in this context in my experience endianness is mostly known in data consisting of multiple bytes i think you are talking about the endianness of bits which seem to be called bit endianness another issue i stumbled upon wire byte bit wire big endian reg le byte bit reg little endian always ff posedge clk le byte byte won t work i tried verilog define led count module led sw all reversed input sw output led reversed bit endianness assign led sw endmodule which worked fine using vivado on my hardware with switches and leds last but not least i am new to verilog ps 😎 way of teaching fpgas using graphics
0
6,746
4,534,191,130
IssuesEvent
2016-09-08 14:01:03
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
28206321: iMessage extension auto rotation is completely broken
classification:ui/usability reproducible:always status:open
#### Description Steps to Reproduce: - Open the IceCreamBuilder sample iMessage extension. - Expand - Rotate from portrait to landscape - Collapse Looks like in the attached screenshot. - Product Version: 10.0 GM Created: 2016-09-08T13:47:18.862640 Originated: 2016-09-08T15:47:00 Open Radar Link: http://www.openradar.me/28206321
True
28206321: iMessage extension auto rotation is completely broken - #### Description Steps to Reproduce: - Open the IceCreamBuilder sample iMessage extension. - Expand - Rotate from portrait to landscape - Collapse Looks like in the attached screenshot. - Product Version: 10.0 GM Created: 2016-09-08T13:47:18.862640 Originated: 2016-09-08T15:47:00 Open Radar Link: http://www.openradar.me/28206321
usab
imessage extension auto rotation is completely broken description steps to reproduce open the icecreambuilder sample imessage extension expand rotate from portrait to landscape collapse looks like in the attached screenshot product version gm created originated open radar link
1
21,947
18,149,869,788
IssuesEvent
2021-09-26 04:37:42
tailscale/tailscale
https://api.github.com/repos/tailscale/tailscale
closed
Synology DSM6 x86_64 package (1.14.3) stops working after 1 minute
L1 Very few P3 Can't get started T6 Major usability OS-synology
Hello, I installed the new package 1.14.3 (tailscale-x86_64-1.14.3-008-dsm6.spk) over the existing 1.12.1 version on my DS3617xs. That worked, NAS showed as "online" on the Tailscale website and after approx. 1 minuted the status went to "offline". I checked the NAS and the package was stopped - restarted the package, and it worked for roughly 1 minute again. Please let me can provide any logs. I deleted it and installed 1.12.1 again. Thank you! <img src="https://frontapp.com/assets/img/favicons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_3kykx)
True
Synology DSM6 x86_64 package (1.14.3) stops working after 1 minute - Hello, I installed the new package 1.14.3 (tailscale-x86_64-1.14.3-008-dsm6.spk) over the existing 1.12.1 version on my DS3617xs. That worked, NAS showed as "online" on the Tailscale website and after approx. 1 minuted the status went to "offline". I checked the NAS and the package was stopped - restarted the package, and it worked for roughly 1 minute again. Please let me can provide any logs. I deleted it and installed 1.12.1 again. Thank you! <img src="https://frontapp.com/assets/img/favicons/favicon-32x32.png" height="16" width="16" alt="Front logo" /> [Front conversations](https://app.frontapp.com/open/top_3kykx)
usab
synology package stops working after minute hello i installed the new package tailscale spk over the existing version on my that worked nas showed as online on the tailscale website and after approx minuted the status went to offline i checked the nas and the package was stopped restarted the package and it worked for roughly minute again please let me can provide any logs i deleted it and installed again thank you
1
24,794
12,403,651,756
IssuesEvent
2020-05-21 14:16:14
tgstation/tgstation-server
https://api.github.com/repos/tgstation/tgstation-server
closed
Database Commit step of Automatic Deployments can hang for... 30 MINUTES?!? (MySql/Linux only maybe)
Area: Jobs Backlog Database Issue Help Wanted Performance Reproduction Required
VORE-station experiences this regularly. The deployment process runs perfectly fine, except commiting the compile job to the database can hang for 30 minutes. This not only makes the chat bots liars when they say `Deployment Complete` it's obviously bad for bitcoin. This doesn't seem to happen with manual deployments even though they follow the same code path.
True
Database Commit step of Automatic Deployments can hang for... 30 MINUTES?!? (MySql/Linux only maybe) - VORE-station experiences this regularly. The deployment process runs perfectly fine, except commiting the compile job to the database can hang for 30 minutes. This not only makes the chat bots liars when they say `Deployment Complete` it's obviously bad for bitcoin. This doesn't seem to happen with manual deployments even though they follow the same code path.
non_usab
database commit step of automatic deployments can hang for minutes mysql linux only maybe vore station experiences this regularly the deployment process runs perfectly fine except commiting the compile job to the database can hang for minutes this not only makes the chat bots liars when they say deployment complete it s obviously bad for bitcoin this doesn t seem to happen with manual deployments even though they follow the same code path
0
15,877
3,488,321,709
IssuesEvent
2016-01-02 21:11:10
SemanticMediaWiki/SemanticMediaWiki
https://api.github.com/repos/SemanticMediaWiki/SemanticMediaWiki
opened
Setting multiple values #set/#subobject using `|`
question requires test wikidocu missing
[0] wrote "with some old templates of mine I found out that ... can be used to store multiple values for one property. Since I did not find any mention of it as I wanted to confirm this" ``` {{#set: |property1=value1|value2|value3 |property2=value1|value2|value3 ... }} ``` Above is codified in [1] but we are missing an integration test [2, 3] and it would be great if someone could send a PR to cover this in order to avoid any regression in future. [0] https://www.semantic-mediawiki.org/wiki/Thread:Help_talk:Setting_values/Setting_multiple_values_in_one_turn [1] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/src/ParserParameterProcessor.php#L205-L212 [2] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/tests/phpunit/Integration/ByJsonScript/README.md [3] https://github.com/SemanticMediaWiki/SemanticMediaWiki/tree/master/tests#write-integration-tests-using-json-script
1.0
Setting multiple values #set/#subobject using `|` - [0] wrote "with some old templates of mine I found out that ... can be used to store multiple values for one property. Since I did not find any mention of it as I wanted to confirm this" ``` {{#set: |property1=value1|value2|value3 |property2=value1|value2|value3 ... }} ``` Above is codified in [1] but we are missing an integration test [2, 3] and it would be great if someone could send a PR to cover this in order to avoid any regression in future. [0] https://www.semantic-mediawiki.org/wiki/Thread:Help_talk:Setting_values/Setting_multiple_values_in_one_turn [1] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/src/ParserParameterProcessor.php#L205-L212 [2] https://github.com/SemanticMediaWiki/SemanticMediaWiki/blob/master/tests/phpunit/Integration/ByJsonScript/README.md [3] https://github.com/SemanticMediaWiki/SemanticMediaWiki/tree/master/tests#write-integration-tests-using-json-script
non_usab
setting multiple values set subobject using wrote with some old templates of mine i found out that can be used to store multiple values for one property since i did not find any mention of it as i wanted to confirm this set above is codified in but we are missing an integration test and it would be great if someone could send a pr to cover this in order to avoid any regression in future
0
16,937
11,495,617,675
IssuesEvent
2020-02-12 05:29:26
the-tale/the-tale
https://api.github.com/repos/the-tale/the-tale
closed
Гильдии: изменить или убрать ограничение на длину описания гильдии
comp_general cont_community cont_usability est_simple good first issue type_improvement
Сейчас слишком короткое. Сделать по аналогии с описаниями Хранителя/героя?
True
Гильдии: изменить или убрать ограничение на длину описания гильдии - Сейчас слишком короткое. Сделать по аналогии с описаниями Хранителя/героя?
usab
гильдии изменить или убрать ограничение на длину описания гильдии сейчас слишком короткое сделать по аналогии с описаниями хранителя героя
1
23,113
6,369,372,045
IssuesEvent
2017-08-01 11:42:36
Yoast/wordpress-seo
https://api.github.com/repos/Yoast/wordpress-seo
closed
Updated Premium page with My Yoast integration
needs-code-review
To work with the upcoming My Yoast, the Premium page in the plugin needs some changes. - [x] Remove the `Licenses` tab completely (this functionality will be merged into the `Extensions` tab) - [x] On the `Extensions` tab, add a second status label to all plugins shown: - [x] When you don't own a plugin, show the `Buy` button with a link that says `More Information`. - [x] When you have a plugin installed, show the `INSTALLED` label and the `NOT ACTIVATED` label, with a link to My Yoast that says `Activate your license on My Yoast >>`. - [x] When the plugin is installed and activated in My Yoast, show the `INSTALLED` and `ACTIVATED` labels with a link that says `Manage this license on My Yoast >>` - [x] Change the warning when a plugin is not activated to `Warning! You have not yet activated [PLUGIN NAME] in My Yoast. If you want to do so now, click here. Otherwise, you will not receive updates or support.` Which links to My Yoast. - [x] Remove the page title and tab navigation at the top. - [x] Increase the spacing between the upsell checklists and the buy button/status labels. ![new license manager](https://cloud.githubusercontent.com/assets/6073772/26549079/88e644c0-4477-11e7-9a6d-ffdbc930527d.png) Fixes #6561.
1.0
Updated Premium page with My Yoast integration - To work with the upcoming My Yoast, the Premium page in the plugin needs some changes. - [x] Remove the `Licenses` tab completely (this functionality will be merged into the `Extensions` tab) - [x] On the `Extensions` tab, add a second status label to all plugins shown: - [x] When you don't own a plugin, show the `Buy` button with a link that says `More Information`. - [x] When you have a plugin installed, show the `INSTALLED` label and the `NOT ACTIVATED` label, with a link to My Yoast that says `Activate your license on My Yoast >>`. - [x] When the plugin is installed and activated in My Yoast, show the `INSTALLED` and `ACTIVATED` labels with a link that says `Manage this license on My Yoast >>` - [x] Change the warning when a plugin is not activated to `Warning! You have not yet activated [PLUGIN NAME] in My Yoast. If you want to do so now, click here. Otherwise, you will not receive updates or support.` Which links to My Yoast. - [x] Remove the page title and tab navigation at the top. - [x] Increase the spacing between the upsell checklists and the buy button/status labels. ![new license manager](https://cloud.githubusercontent.com/assets/6073772/26549079/88e644c0-4477-11e7-9a6d-ffdbc930527d.png) Fixes #6561.
non_usab
updated premium page with my yoast integration to work with the upcoming my yoast the premium page in the plugin needs some changes remove the licenses tab completely this functionality will be merged into the extensions tab on the extensions tab add a second status label to all plugins shown when you don t own a plugin show the buy button with a link that says more information when you have a plugin installed show the installed label and the not activated label with a link to my yoast that says activate your license on my yoast when the plugin is installed and activated in my yoast show the installed and activated labels with a link that says manage this license on my yoast change the warning when a plugin is not activated to warning you have not yet activated in my yoast if you want to do so now click here otherwise you will not receive updates or support which links to my yoast remove the page title and tab navigation at the top increase the spacing between the upsell checklists and the buy button status labels fixes
0
45,678
7,195,418,698
IssuesEvent
2018-02-04 16:58:10
golang/go
https://api.github.com/repos/golang/go
opened
strconv: Unquote example looks like a unit test instead of an example
Documentation NeedsFix help wanted
The Unquote example (https://golang.org/pkg/strconv/#Unquote) looks like a unit test instead of an example. That is a sea of backslashes and quotes. I think we could make a more readable example.
1.0
strconv: Unquote example looks like a unit test instead of an example - The Unquote example (https://golang.org/pkg/strconv/#Unquote) looks like a unit test instead of an example. That is a sea of backslashes and quotes. I think we could make a more readable example.
non_usab
strconv unquote example looks like a unit test instead of an example the unquote example looks like a unit test instead of an example that is a sea of backslashes and quotes i think we could make a more readable example
0
465,265
13,369,622,130
IssuesEvent
2020-09-01 09:06:00
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
manabi-gakushu.benesse.ne.jp - site is not usable
browser-firefox engine-gecko priority-normal
<!-- @browser: Firefox 82.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:82.0) Gecko/20100101 Firefox/82.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/57348 --> **URL**: https://manabi-gakushu.benesse.ne.jp/gakushu/typing/nihongonyuryoku.html **Browser / Version**: Firefox 82.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Unable to type **Steps to Reproduce**: does not work type " - " key <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/8/50ab0f6d-d6ef-4297-b9a3-c997f39c6fd5.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200828153126</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/8/765ca612-fb1f-4421-a199-e5a9b85ab6e6) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
manabi-gakushu.benesse.ne.jp - site is not usable - <!-- @browser: Firefox 82.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:82.0) Gecko/20100101 Firefox/82.0 --> <!-- @reported_with: desktop-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/57348 --> **URL**: https://manabi-gakushu.benesse.ne.jp/gakushu/typing/nihongonyuryoku.html **Browser / Version**: Firefox 82.0 **Operating System**: Windows 10 **Tested Another Browser**: Yes Edge **Problem type**: Site is not usable **Description**: Unable to type **Steps to Reproduce**: does not work type " - " key <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2020/8/50ab0f6d-d6ef-4297-b9a3-c997f39c6fd5.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20200828153126</li><li>channel: nightly</li><li>hasTouchScreen: false</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2020/8/765ca612-fb1f-4421-a199-e5a9b85ab6e6) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_usab
manabi gakushu benesse ne jp site is not usable url browser version firefox operating system windows tested another browser yes edge problem type site is not usable description unable to type steps to reproduce does not work type key view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen false mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
82,662
16,011,575,303
IssuesEvent
2021-04-20 11:17:23
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
<program_name> extracting an empty string
ruleset threatintel threatintel/decoders threatintel/rules
``` Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun) **Phase 1: Completed pre-decoding. full event: 'Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun)' timestamp: 'Nov 20 06:35:00' hostname: 'PUNGDC-FDCS-DIST_SF' program_name: '' log: '(root) CMD ( /usr/libexec/atrun)' **Phase 2: Completed decoding. No decoder matched. ``` As you can notice below the result for that log gives program_name an empty string and not (null), I've tried with a pre-match entry but the decoder hasn't been detected. The only way to get it work was to use the program_name in that way : ``` <decoder name="Junos-dist-switch-info4"> <program_name></program_name> </decoder> <decoder name="Junos-dist-info-cron"> <parent>Junos-dist-switch-info4</parent> <regex>(\(\S+\)) \S+ (\(\s+\S+\))</regex> <order>cron.user, cron.action</order> </decoder> ``` the result is : ``` Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun) **Phase 1: Completed pre-decoding. full event: 'Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun) ' timestamp: 'Nov 20 06:35:00' hostname: 'PUNGDC-FDCS-DIST_SF' program_name: '' log: '(root) CMD ( /usr/libexec/atrun) ' **Phase 2: Completed decoding. decoder: 'Junos-dist-switch-info4' cron.user: '(root)' cron.action: '( /usr/libexec/atrun)' ``` That may make many conflicts especially if we have many logs from different sources where the program_name is extracted as an empty string like in this case.
1.0
<program_name> extracting an empty string - ``` Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun) **Phase 1: Completed pre-decoding. full event: 'Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun)' timestamp: 'Nov 20 06:35:00' hostname: 'PUNGDC-FDCS-DIST_SF' program_name: '' log: '(root) CMD ( /usr/libexec/atrun)' **Phase 2: Completed decoding. No decoder matched. ``` As you can notice below the result for that log gives program_name an empty string and not (null), I've tried with a pre-match entry but the decoder hasn't been detected. The only way to get it work was to use the program_name in that way : ``` <decoder name="Junos-dist-switch-info4"> <program_name></program_name> </decoder> <decoder name="Junos-dist-info-cron"> <parent>Junos-dist-switch-info4</parent> <regex>(\(\S+\)) \S+ (\(\s+\S+\))</regex> <order>cron.user, cron.action</order> </decoder> ``` the result is : ``` Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun) **Phase 1: Completed pre-decoding. full event: 'Nov 20 06:35:00 PUNGDC-FDCS-DIST_SF : (root) CMD ( /usr/libexec/atrun) ' timestamp: 'Nov 20 06:35:00' hostname: 'PUNGDC-FDCS-DIST_SF' program_name: '' log: '(root) CMD ( /usr/libexec/atrun) ' **Phase 2: Completed decoding. decoder: 'Junos-dist-switch-info4' cron.user: '(root)' cron.action: '( /usr/libexec/atrun)' ``` That may make many conflicts especially if we have many logs from different sources where the program_name is extracted as an empty string like in this case.
non_usab
extracting an empty string nov pungdc fdcs dist sf root cmd usr libexec atrun phase completed pre decoding full event nov pungdc fdcs dist sf root cmd usr libexec atrun timestamp nov hostname pungdc fdcs dist sf program name log root cmd usr libexec atrun phase completed decoding no decoder matched as you can notice below the result for that log gives program name an empty string and not null i ve tried with a pre match entry but the decoder hasn t been detected the only way to get it work was to use the program name in that way junos dist switch s s s s cron user cron action the result is nov pungdc fdcs dist sf root cmd usr libexec atrun phase completed pre decoding full event nov pungdc fdcs dist sf root cmd usr libexec atrun timestamp nov hostname pungdc fdcs dist sf program name log root cmd usr libexec atrun phase completed decoding decoder junos dist switch cron user root cron action usr libexec atrun that may make many conflicts especially if we have many logs from different sources where the program name is extracted as an empty string like in this case
0
9,461
6,307,560,505
IssuesEvent
2017-07-22 02:30:55
coreos/bugs
https://api.github.com/repos/coreos/bugs
closed
iptables kernel panic
area/usability component/kernel kind/bug team/os
# Issue Report # Bug ## Bug ## When iptables-restore is used to perform replacements of very large tables, `ipt_alloc_initial_table` hangs, and when it does, the linux kernel panics. The following screenshot was captured from the console of a virtual machine, but the issue also occurs on bare-metal machines. ![image](https://cloud.githubusercontent.com/assets/19591184/24277402/71faa9f8-1002-11e7-8f6e-50e837a53bb0.png) ### Container Linux Version ### ``` cat /etc/os-release NAME=CoreOS ID=coreos VERSION=1185.5.0 VERSION_ID=1185.5.0 BUILD_ID=2016-12-07-0937 PRETTY_NAME="CoreOS 1185.5.0 (MoreOS)" ANSI_COLOR="1;32" HOME_URL="https://coreos.com/" BUG_REPORT_URL="https://github.com/coreos/bugs/issues" cat /proc/version Linux version 4.7.3-coreos-r3 (jenkins@jenkins-os-executor-1.c.coreos-gce-testing.internal) (gcc version 4.9.3 (Gentoo Hardened 4.9.3 p1.5, pie-0.6.4) ) #1 SMP Wed Dec 7 09:29:55 UTC 2016 ``` ### Environment ### replicated on: bare metal machines. also on virtual machines. occurs when iptables directly on the host machine. also occurs when iptables is executed from within a docker container. ### Expected Behavior ### a kernel panic does not occur ### Actual Behavior ### linux kernel panics ### Reproduction Steps ### 1. generate an iptables ruleset with lots and lots of rules. 2. run `iptables-restore -T nat --noflush --counters saved.log` 3. the kernel will panic or here's a shell script: ``` # with sudo: ~ # for i in {1..100000}; do iptables-save -t nat > heh.txt; lines=$(cat heh.txt | wc -l); echo -n "#$i - saved $lines rules... "; iptables-restore -T nat --noflush --counters heh.txt ; date +%s; done #1 - saved 259 rules... 1490325227 #2 - saved 262 rules... 1490325227 #3 - saved 268 rules... 1490325227 #4 - saved 280 rules... 1490325227 #5 - saved 304 rules... 1490325227 #6 - saved 352 rules... 1490325227 #7 - saved 448 rules... 1490325227 #8 - saved 640 rules... 1490325227 #9 - saved 1024 rules... 1490325227 #10 - saved 1792 rules... 1490325227 #11 - saved 3328 rules... 1490325228 <------------- NOTE: it starts to get sluggish here #12 - saved 6400 rules... 1490325230 #13 - saved 12544 rules... 1490325237 #14 - saved 24832 rules... <------------- Can't exceed 32768? ```
True
iptables kernel panic - # Issue Report # Bug ## Bug ## When iptables-restore is used to perform replacements of very large tables, `ipt_alloc_initial_table` hangs, and when it does, the linux kernel panics. The following screenshot was captured from the console of a virtual machine, but the issue also occurs on bare-metal machines. ![image](https://cloud.githubusercontent.com/assets/19591184/24277402/71faa9f8-1002-11e7-8f6e-50e837a53bb0.png) ### Container Linux Version ### ``` cat /etc/os-release NAME=CoreOS ID=coreos VERSION=1185.5.0 VERSION_ID=1185.5.0 BUILD_ID=2016-12-07-0937 PRETTY_NAME="CoreOS 1185.5.0 (MoreOS)" ANSI_COLOR="1;32" HOME_URL="https://coreos.com/" BUG_REPORT_URL="https://github.com/coreos/bugs/issues" cat /proc/version Linux version 4.7.3-coreos-r3 (jenkins@jenkins-os-executor-1.c.coreos-gce-testing.internal) (gcc version 4.9.3 (Gentoo Hardened 4.9.3 p1.5, pie-0.6.4) ) #1 SMP Wed Dec 7 09:29:55 UTC 2016 ``` ### Environment ### replicated on: bare metal machines. also on virtual machines. occurs when iptables directly on the host machine. also occurs when iptables is executed from within a docker container. ### Expected Behavior ### a kernel panic does not occur ### Actual Behavior ### linux kernel panics ### Reproduction Steps ### 1. generate an iptables ruleset with lots and lots of rules. 2. run `iptables-restore -T nat --noflush --counters saved.log` 3. the kernel will panic or here's a shell script: ``` # with sudo: ~ # for i in {1..100000}; do iptables-save -t nat > heh.txt; lines=$(cat heh.txt | wc -l); echo -n "#$i - saved $lines rules... "; iptables-restore -T nat --noflush --counters heh.txt ; date +%s; done #1 - saved 259 rules... 1490325227 #2 - saved 262 rules... 1490325227 #3 - saved 268 rules... 1490325227 #4 - saved 280 rules... 1490325227 #5 - saved 304 rules... 1490325227 #6 - saved 352 rules... 1490325227 #7 - saved 448 rules... 1490325227 #8 - saved 640 rules... 1490325227 #9 - saved 1024 rules... 1490325227 #10 - saved 1792 rules... 1490325227 #11 - saved 3328 rules... 1490325228 <------------- NOTE: it starts to get sluggish here #12 - saved 6400 rules... 1490325230 #13 - saved 12544 rules... 1490325237 #14 - saved 24832 rules... <------------- Can't exceed 32768? ```
usab
iptables kernel panic issue report bug bug when iptables restore is used to perform replacements of very large tables ipt alloc initial table hangs and when it does the linux kernel panics the following screenshot was captured from the console of a virtual machine but the issue also occurs on bare metal machines container linux version cat etc os release name coreos id coreos version version id build id pretty name coreos moreos ansi color home url bug report url cat proc version linux version coreos jenkins jenkins os executor c coreos gce testing internal gcc version gentoo hardened pie smp wed dec utc environment replicated on bare metal machines also on virtual machines occurs when iptables directly on the host machine also occurs when iptables is executed from within a docker container expected behavior a kernel panic does not occur actual behavior linux kernel panics reproduction steps generate an iptables ruleset with lots and lots of rules run iptables restore t nat noflush counters saved log the kernel will panic or here s a shell script with sudo for i in do iptables save t nat heh txt lines cat heh txt wc l echo n i saved lines rules iptables restore t nat noflush counters heh txt date s done saved rules saved rules saved rules saved rules saved rules saved rules saved rules saved rules saved rules saved rules saved rules note it starts to get sluggish here saved rules saved rules saved rules can t exceed
1
5,104
3,900,324,251
IssuesEvent
2016-04-18 05:02:10
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
12551140: Smart Banners require an App Store installation
classification:ui/usability reproducible:always status:open
#### Description Summary: iOS 6 Smart Banners only work if the device has an App Store installation of the app in question. To-be-released products, and even Xcode builds of existing products, are not recognized by the smart banner machinery. Steps to Reproduce: 0) Uninstall any copies of the app to be tested 1) Build and run your smart-banner-supporting app from Xcode 2) Go to your website with smart banner metadata 3) Note the VIEW button, which takes you to the App Store, rather than OPEN 4) Install the app from the App Store 5) Repeat 1) Expected Results: Expect smart banners to work all the time. Actual Results: Smart banners ONLY work after step 5 — specifically after building from Xcode on top of an App Store installation. Regression: iOS 6.0 (10A403) Notes: As far as I can tell, this makes Smart Banner testing completely impossible for pre-1.0 App Store apps. Similarly painful for existing products because the tester now needs the Xcode project and codesign / provisioning. This is a nonstarter for end-user beta testing, but even in many workplaces a QA team may not have source access. IPA installations from the Xcode Organizer do not work. Third-party OTA tools like TestFlight and Hockey are similarly out of luck. Please revise this behavior to work with Debug and Ad-Hoc builds. This should ultimately be linked to the App's Bundle ID on the device side: 1) Safari reads the app-id parameter from the meta tag 2) Safari contacts the App Store for the related bundle ID 3) Safari asks the system if an app with said bundle ID exists These steps can simply be used as a fallback against the existing current mechanism, rather than replacing it completely. - Product Version: 10A403 Created: 2012-10-22T21:31:55.921885 Originated: 2012-10-22T00:00:00 Open Radar Link: http://www.openradar.me/12551140
True
12551140: Smart Banners require an App Store installation - #### Description Summary: iOS 6 Smart Banners only work if the device has an App Store installation of the app in question. To-be-released products, and even Xcode builds of existing products, are not recognized by the smart banner machinery. Steps to Reproduce: 0) Uninstall any copies of the app to be tested 1) Build and run your smart-banner-supporting app from Xcode 2) Go to your website with smart banner metadata 3) Note the VIEW button, which takes you to the App Store, rather than OPEN 4) Install the app from the App Store 5) Repeat 1) Expected Results: Expect smart banners to work all the time. Actual Results: Smart banners ONLY work after step 5 — specifically after building from Xcode on top of an App Store installation. Regression: iOS 6.0 (10A403) Notes: As far as I can tell, this makes Smart Banner testing completely impossible for pre-1.0 App Store apps. Similarly painful for existing products because the tester now needs the Xcode project and codesign / provisioning. This is a nonstarter for end-user beta testing, but even in many workplaces a QA team may not have source access. IPA installations from the Xcode Organizer do not work. Third-party OTA tools like TestFlight and Hockey are similarly out of luck. Please revise this behavior to work with Debug and Ad-Hoc builds. This should ultimately be linked to the App's Bundle ID on the device side: 1) Safari reads the app-id parameter from the meta tag 2) Safari contacts the App Store for the related bundle ID 3) Safari asks the system if an app with said bundle ID exists These steps can simply be used as a fallback against the existing current mechanism, rather than replacing it completely. - Product Version: 10A403 Created: 2012-10-22T21:31:55.921885 Originated: 2012-10-22T00:00:00 Open Radar Link: http://www.openradar.me/12551140
usab
smart banners require an app store installation description summary ios smart banners only work if the device has an app store installation of the app in question to be released products and even xcode builds of existing products are not recognized by the smart banner machinery steps to reproduce uninstall any copies of the app to be tested build and run your smart banner supporting app from xcode go to your website with smart banner metadata note the view button which takes you to the app store rather than open install the app from the app store repeat expected results expect smart banners to work all the time actual results smart banners only work after step — specifically after building from xcode on top of an app store installation regression ios notes as far as i can tell this makes smart banner testing completely impossible for pre app store apps similarly painful for existing products because the tester now needs the xcode project and codesign provisioning this is a nonstarter for end user beta testing but even in many workplaces a qa team may not have source access ipa installations from the xcode organizer do not work third party ota tools like testflight and hockey are similarly out of luck please revise this behavior to work with debug and ad hoc builds this should ultimately be linked to the app s bundle id on the device side safari reads the app id parameter from the meta tag safari contacts the app store for the related bundle id safari asks the system if an app with said bundle id exists these steps can simply be used as a fallback against the existing current mechanism rather than replacing it completely product version created originated open radar link
1
34,362
12,269,416,894
IssuesEvent
2020-05-07 14:02:52
logzio/jmx2logzio
https://api.github.com/repos/logzio/jmx2logzio
closed
CVE-2020-9547 (Medium) detected in jackson-databind-2.9.10.2.jar
security vulnerability
## CVE-2020-9547 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/jmx2logzio/pom.xml</p> <p>Path to vulnerable library: downloadResource_c762fbf8-ee9b-4076-944f-b22607c0cecb/20200210155926/jackson-databind-2.9.10.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.2.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547>CVE-2020-9547</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.2","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.3"}],"vulnerabilityIdentifier":"CVE-2020-9547","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
True
CVE-2020-9547 (Medium) detected in jackson-databind-2.9.10.2.jar - ## CVE-2020-9547 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.10.2.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: /tmp/ws-scm/jmx2logzio/pom.xml</p> <p>Path to vulnerable library: downloadResource_c762fbf8-ee9b-4076-944f-b22607c0cecb/20200210155926/jackson-databind-2.9.10.2.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.10.2.jar** (Vulnerable Library) </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap). <p>Publish Date: 2020-03-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547>CVE-2020-9547</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-9547</a></p> <p>Release Date: 2020-03-02</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.10.3</p> </p> </details> <p></p> *** <!-- REMEDIATE-OPEN-PR-START --> - [ ] Check this box to open an automated fix PR <!-- REMEDIATE-OPEN-PR-END --> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.10.2","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.10.2","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.10.3"}],"vulnerabilityIdentifier":"CVE-2020-9547","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to com.ibatis.sqlmap.engine.transaction.jta.JtaTransactionConfig (aka ibatis-sqlmap).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-9547","cvss2Severity":"medium","cvss2Score":"5.0","extraData":{}}</REMEDIATE> -->
non_usab
cve medium detected in jackson databind jar cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file tmp ws scm pom xml path to vulnerable library downloadresource jackson databind jar dependency hierarchy x jackson databind jar vulnerable library vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind check this box to open an automated fix pr isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to com ibatis sqlmap engine transaction jta jtatransactionconfig aka ibatis sqlmap vulnerabilityurl
0
11,020
7,027,700,807
IssuesEvent
2017-12-25 01:42:33
godotengine/godot
https://api.github.com/repos/godotengine/godot
closed
"F" key Viewport focus problem
bug junior job topic:editor usability
version -> http://godot3builds.digitecnology.com/builds/2017-12-06.12:52:48.a8ae46e14/godot-v3.0-osx.fat.zip **Issue description:** When i press F key, wrong viewport is active. **Steps to reproduce:** https://www.youtube.com/watch?v=y5K5xv7UBuo&feature=youtu.be
True
"F" key Viewport focus problem - version -> http://godot3builds.digitecnology.com/builds/2017-12-06.12:52:48.a8ae46e14/godot-v3.0-osx.fat.zip **Issue description:** When i press F key, wrong viewport is active. **Steps to reproduce:** https://www.youtube.com/watch?v=y5K5xv7UBuo&feature=youtu.be
usab
f key viewport focus problem version issue description when i press f key wrong viewport is active steps to reproduce
1
24,613
24,033,901,064
IssuesEvent
2022-09-15 17:17:02
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
opened
pulumi new project already exists error does not respect the default org
kind/bug impact/usability
[internal slack convo](https://pulumi.slack.com/archives/C9SGX9QA1/p1663254061919969) repro steps - create a project in your individual account with name X - set your default org to an org that does not have a project with the name X - use pulumi new - when it asks for a project name, set it to X - receive an error saying that project already exists
True
pulumi new project already exists error does not respect the default org - [internal slack convo](https://pulumi.slack.com/archives/C9SGX9QA1/p1663254061919969) repro steps - create a project in your individual account with name X - set your default org to an org that does not have a project with the name X - use pulumi new - when it asks for a project name, set it to X - receive an error saying that project already exists
usab
pulumi new project already exists error does not respect the default org repro steps create a project in your individual account with name x set your default org to an org that does not have a project with the name x use pulumi new when it asks for a project name set it to x receive an error saying that project already exists
1
8,769
5,957,741,646
IssuesEvent
2017-05-29 04:21:36
Virtual-Labs/circular-dichronism-spectroscopy-iiith
https://api.github.com/repos/Virtual-Labs/circular-dichronism-spectroscopy-iiith
opened
QA_Circular Dichroism Spectroscopy_Introduction_Target Audiance_Spelling-mistakes
Category:Usability Developed by: VLEAD Open-Edx Severity:S2 Severity:S3
Defect Description : In the Sub section "Target Audience" of Introduction section of this lab, found spelling mistakes. Actual Result : In the Sub section "Target Audience" of Introduction section of this lab, found spelling mistakes. Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers:Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Attachments: ![qa_oe_cds_i01](https://cloud.githubusercontent.com/assets/13479177/26536416/4d25232a-4454-11e7-84e5-4226cfc9f7bf.png)
True
QA_Circular Dichroism Spectroscopy_Introduction_Target Audiance_Spelling-mistakes - Defect Description : In the Sub section "Target Audience" of Introduction section of this lab, found spelling mistakes. Actual Result : In the Sub section "Target Audience" of Introduction section of this lab, found spelling mistakes. Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers:Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5 Attachments: ![qa_oe_cds_i01](https://cloud.githubusercontent.com/assets/13479177/26536416/4d25232a-4454-11e7-84e5-4226cfc9f7bf.png)
usab
qa circular dichroism spectroscopy introduction target audiance spelling mistakes defect description in the sub section target audience of introduction section of this lab found spelling mistakes actual result in the sub section target audience of introduction section of this lab found spelling mistakes environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor attachments
1
14,444
9,194,935,033
IssuesEvent
2019-03-07 00:16:25
pulumi/pulumi
https://api.github.com/repos/pulumi/pulumi
opened
RunErrors aren't concisely reported anymore
area/cli impact/usability kind/bug
It used to be the case that `RunError`s led to concise error reporting, including omission of the stack trace. The idea here was that these errors want to instruct the user to do something differently -- like run `pulumi config set ...`, and that the extra noise of a stack trace is simply confusing. It appears this behavior has regressed semi-recently: ![image](https://user-images.githubusercontent.com/3953235/53922928-4b7dcc00-4044-11e9-9bf4-1b2c9cffc3f5.png)
True
RunErrors aren't concisely reported anymore - It used to be the case that `RunError`s led to concise error reporting, including omission of the stack trace. The idea here was that these errors want to instruct the user to do something differently -- like run `pulumi config set ...`, and that the extra noise of a stack trace is simply confusing. It appears this behavior has regressed semi-recently: ![image](https://user-images.githubusercontent.com/3953235/53922928-4b7dcc00-4044-11e9-9bf4-1b2c9cffc3f5.png)
usab
runerrors aren t concisely reported anymore it used to be the case that runerror s led to concise error reporting including omission of the stack trace the idea here was that these errors want to instruct the user to do something differently like run pulumi config set and that the extra noise of a stack trace is simply confusing it appears this behavior has regressed semi recently
1
272,800
29,795,090,404
IssuesEvent
2023-06-16 01:10:11
billmcchesney1/hadoop
https://api.github.com/repos/billmcchesney1/hadoop
closed
WS-2017-0234 (Medium) detected in jquery.dataTables-1.10.7.min.js - autoclosed
Mend: dependency security vulnerability
## WS-2017-0234 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.dataTables-1.10.7.min.js</b></p></summary> <p>DataTables enhances HTML tables with the ability to sort, filter and page the data in the table very easily. It provides a comprehensive API and set of configuration options, allowing you to consume data from virtually any data source.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/datatables/1.10.7/js/jquery.dataTables.min.js">https://cdnjs.cloudflare.com/ajax/libs/datatables/1.10.7/js/jquery.dataTables.min.js</a></p> <p>Path to vulnerable library: /hadoop-hdfs-project/hadoop-hdfs/target/webapps/static/jquery.dataTables.min.js,/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/jquery.dataTables.min.js,/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/static/jquery.dataTables.min.js</p> <p> Dependency Hierarchy: - :x: **jquery.dataTables-1.10.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Affected versions of the package are vulnerable to Cross-site Scripting (XSS). <p>Publish Date: 2015-11-06 <p>URL: <a href=https://github.com/DataTables/DataTables/commit/6f67df2d21f9858ec40a6e9565c3a653cdb691a6>WS-2017-0234</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2015-11-06</p> <p>Fix Resolution: datatables - 1.10.10;org.webjars:datatables:1.10.10</p> </p> </details> <p></p>
True
WS-2017-0234 (Medium) detected in jquery.dataTables-1.10.7.min.js - autoclosed - ## WS-2017-0234 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery.dataTables-1.10.7.min.js</b></p></summary> <p>DataTables enhances HTML tables with the ability to sort, filter and page the data in the table very easily. It provides a comprehensive API and set of configuration options, allowing you to consume data from virtually any data source.</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/datatables/1.10.7/js/jquery.dataTables.min.js">https://cdnjs.cloudflare.com/ajax/libs/datatables/1.10.7/js/jquery.dataTables.min.js</a></p> <p>Path to vulnerable library: /hadoop-hdfs-project/hadoop-hdfs/target/webapps/static/jquery.dataTables.min.js,/hadoop-hdfs-project/hadoop-hdfs/target/test-classes/webapps/static/jquery.dataTables.min.js,/hadoop-hdfs-project/hadoop-hdfs/src/main/webapps/static/jquery.dataTables.min.js</p> <p> Dependency Hierarchy: - :x: **jquery.dataTables-1.10.7.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/billmcchesney1/hadoop/commit/6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a">6dcd8400219941dcbd7fb0f6b980cc2c6a2a6b0a</a></p> <p>Found in base branch: <b>trunk</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> Affected versions of the package are vulnerable to Cross-site Scripting (XSS). <p>Publish Date: 2015-11-06 <p>URL: <a href=https://github.com/DataTables/DataTables/commit/6f67df2d21f9858ec40a6e9565c3a653cdb691a6>WS-2017-0234</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Release Date: 2015-11-06</p> <p>Fix Resolution: datatables - 1.10.10;org.webjars:datatables:1.10.10</p> </p> </details> <p></p>
non_usab
ws medium detected in jquery datatables min js autoclosed ws medium severity vulnerability vulnerable library jquery datatables min js datatables enhances html tables with the ability to sort filter and page the data in the table very easily it provides a comprehensive api and set of configuration options allowing you to consume data from virtually any data source library home page a href path to vulnerable library hadoop hdfs project hadoop hdfs target webapps static jquery datatables min js hadoop hdfs project hadoop hdfs target test classes webapps static jquery datatables min js hadoop hdfs project hadoop hdfs src main webapps static jquery datatables min js dependency hierarchy x jquery datatables min js vulnerable library found in head commit a href found in base branch trunk vulnerability details affected versions of the package are vulnerable to cross site scripting xss publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version release date fix resolution datatables org webjars datatables
0
10,457
6,733,274,596
IssuesEvent
2017-10-18 14:22:22
elastic/rally
https://api.github.com/repos/elastic/rally
closed
Provide a clear error message for non-existing releases
:Benchmark Candidate Management :Usability enhancement
If users have a typo in their distribution specification, then they get a very unclear error message, e.g.: ``` tarfile.ReadError: file could not be opened successfully ``` We should detect this condition and provide a clear error message.
True
Provide a clear error message for non-existing releases - If users have a typo in their distribution specification, then they get a very unclear error message, e.g.: ``` tarfile.ReadError: file could not be opened successfully ``` We should detect this condition and provide a clear error message.
usab
provide a clear error message for non existing releases if users have a typo in their distribution specification then they get a very unclear error message e g tarfile readerror file could not be opened successfully we should detect this condition and provide a clear error message
1
25,829
25,961,444,150
IssuesEvent
2022-12-18 23:32:10
kubernetes/test-infra
https://api.github.com/repos/kubernetes/test-infra
closed
Using Ephemeral Volume Snapshot with Prow Job
kind/feature lifecycle/rotten sig/usability
<!-- Please only use this template for submitting enhancement requests --> **What would you like to be added**: Seems like a presubmit prow job is not able to create a volume by using "volume claim template" described in the presubmit job yaml. I am trying to use the volume claim template to create a PV from volume snapshot "data source" in the yaml. The mount paths come up empty. The setup works fine when tested with a POD but not with prowjob yaml. Eg: ``` volumes: - name: try-vol ephemeral: volumeClaimTemplate: spec: accessModes: - ReadWriteOnce storageClassName: "xxx-xxx-x-xx" resources: requests: storage: 60Gi dataSource: name: snapshot kind: VolumeSnapshot apiGroup: snapshot.storage.k8s.io ``` **Why is this needed**: Trying to mount cached artefacts on the fly for PROW presubmit job to make CI faster.
True
Using Ephemeral Volume Snapshot with Prow Job - <!-- Please only use this template for submitting enhancement requests --> **What would you like to be added**: Seems like a presubmit prow job is not able to create a volume by using "volume claim template" described in the presubmit job yaml. I am trying to use the volume claim template to create a PV from volume snapshot "data source" in the yaml. The mount paths come up empty. The setup works fine when tested with a POD but not with prowjob yaml. Eg: ``` volumes: - name: try-vol ephemeral: volumeClaimTemplate: spec: accessModes: - ReadWriteOnce storageClassName: "xxx-xxx-x-xx" resources: requests: storage: 60Gi dataSource: name: snapshot kind: VolumeSnapshot apiGroup: snapshot.storage.k8s.io ``` **Why is this needed**: Trying to mount cached artefacts on the fly for PROW presubmit job to make CI faster.
usab
using ephemeral volume snapshot with prow job what would you like to be added seems like a presubmit prow job is not able to create a volume by using volume claim template described in the presubmit job yaml i am trying to use the volume claim template to create a pv from volume snapshot data source in the yaml the mount paths come up empty the setup works fine when tested with a pod but not with prowjob yaml eg volumes name try vol ephemeral volumeclaimtemplate spec accessmodes readwriteonce storageclassname xxx xxx x xx resources requests storage datasource name snapshot kind volumesnapshot apigroup snapshot storage io why is this needed trying to mount cached artefacts on the fly for prow presubmit job to make ci faster
1
18,940
13,486,582,078
IssuesEvent
2020-09-11 09:39:43
CARTAvis/carta-frontend
https://api.github.com/repos/CARTAvis/carta-frontend
closed
Spectral Profile opened automatically
usability review
From Usability Review, Region Analytics: I certainly think that the spectral profile should be opened automatically (much more often required than the X or Y). I think you could also probably update the button pictures to make them a little more obvious?
True
Spectral Profile opened automatically - From Usability Review, Region Analytics: I certainly think that the spectral profile should be opened automatically (much more often required than the X or Y). I think you could also probably update the button pictures to make them a little more obvious?
usab
spectral profile opened automatically from usability review region analytics i certainly think that the spectral profile should be opened automatically much more often required than the x or y i think you could also probably update the button pictures to make them a little more obvious
1
672,141
22,792,889,882
IssuesEvent
2022-07-10 09:15:55
crestic-urca/remotelabz
https://api.github.com/repos/crestic-urca/remotelabz
closed
The forward seems no to be activate from install
normal priority needs more info
Verify if the net.ipv4.ip_forward=1 is ok on the front and the worker in the install process
1.0
The forward seems no to be activate from install - Verify if the net.ipv4.ip_forward=1 is ok on the front and the worker in the install process
non_usab
the forward seems no to be activate from install verify if the net ip forward is ok on the front and the worker in the install process
0
10,547
6,794,678,069
IssuesEvent
2017-11-01 13:11:27
jamesmillerburgess/customer-management
https://api.github.com/repos/jamesmillerburgess/customer-management
closed
Screen Resolution Iphone 6+ versus SamsungA5 2017
Minor Issue Usability
Just made a quick comparison Iphone versus Samsung. Noticed that the display of the fields are different when looking at a contact Reasons ? - Screen resolution ? - Length of the fields ? ![capture](https://user-images.githubusercontent.com/32377726/32265110-ed30b5f2-bee1-11e7-8e9e-4c2d7688a641.JPG)
True
Screen Resolution Iphone 6+ versus SamsungA5 2017 - Just made a quick comparison Iphone versus Samsung. Noticed that the display of the fields are different when looking at a contact Reasons ? - Screen resolution ? - Length of the fields ? ![capture](https://user-images.githubusercontent.com/32377726/32265110-ed30b5f2-bee1-11e7-8e9e-4c2d7688a641.JPG)
usab
screen resolution iphone versus just made a quick comparison iphone versus samsung noticed that the display of the fields are different when looking at a contact reasons screen resolution length of the fields
1
34,291
29,190,851,222
IssuesEvent
2023-05-19 19:50:25
woocommerce/woocommerce
https://api.github.com/repos/woocommerce/woocommerce
closed
Optimize React Admin's `package.json` Scripts
type: task plugin: woocommerce tool: monorepo infrastructure focus: react admin [team:Ghidorah]
<!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. --> **Prerequisites (mark completed items with an [x]):** - [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose - [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is). **Issue Description:** There are a large number of different workflows supported by the current `package.json` script. As a consequence of our tooling changes, however, there is a non-trivial amount of extra work being done in these scripts. Once https://github.com/woocommerce/woocommerce/issues/33907 lands, we should review all of the scripts and look for areas of performance deficit: - `build:packages` is no longer necessary, `turbo` builds all dependencies. - Check for cases where `turbo` is being bypassed entirely. - Review supported workflows and discuss flattening them where appropriate with their respective teams.
1.0
Optimize React Admin's `package.json` Scripts - <!-- This form is for other issue types specific to the WooCommerce plugin. This is not a support portal. --> **Prerequisites (mark completed items with an [x]):** - [x] I have checked that my issue type is not listed here https://github.com/woocommerce/woocommerce/issues/new/choose - [x] My issue is not a security issue, support request, bug report, enhancement or feature request (Please use the link above if it is). **Issue Description:** There are a large number of different workflows supported by the current `package.json` script. As a consequence of our tooling changes, however, there is a non-trivial amount of extra work being done in these scripts. Once https://github.com/woocommerce/woocommerce/issues/33907 lands, we should review all of the scripts and look for areas of performance deficit: - `build:packages` is no longer necessary, `turbo` builds all dependencies. - Check for cases where `turbo` is being bypassed entirely. - Review supported workflows and discuss flattening them where appropriate with their respective teams.
non_usab
optimize react admin s package json scripts prerequisites mark completed items with an i have checked that my issue type is not listed here my issue is not a security issue support request bug report enhancement or feature request please use the link above if it is issue description there are a large number of different workflows supported by the current package json script as a consequence of our tooling changes however there is a non trivial amount of extra work being done in these scripts once lands we should review all of the scripts and look for areas of performance deficit build packages is no longer necessary turbo builds all dependencies check for cases where turbo is being bypassed entirely review supported workflows and discuss flattening them where appropriate with their respective teams
0
8,859
6,008,439,179
IssuesEvent
2017-06-06 07:51:09
bwsw/cloudstack-ui
https://api.github.com/repos/bwsw/cloudstack-ui
closed
Vm/Disk/other pane actions (...) bar
enhancement proposal usability
Currently actions bar is displayed top-middle of VM/Disk/etc bar, I think it's may be better to display it in the middle of the bar.
True
Vm/Disk/other pane actions (...) bar - Currently actions bar is displayed top-middle of VM/Disk/etc bar, I think it's may be better to display it in the middle of the bar.
usab
vm disk other pane actions bar currently actions bar is displayed top middle of vm disk etc bar i think it s may be better to display it in the middle of the bar
1
18,328
12,795,613,162
IssuesEvent
2020-07-02 09:02:38
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
Block appender appears when a reusable Heading block is selected
[Feature] Reusable Blocks [Type] Bug
**Describe the bug** A modified Heading block saved as a Reusable block appears frozen, it neither accepts text or adds a new block when the + is pressed. **To reproduce** Steps to reproduce the behavior: 1. In the editor, click the + and add a Heading block 2. Change the Heading block to be H3, and Align Text Center. 3. Click on Add to reusable blocks 4. Change Resuable Block name to H3C and hit Save 5. Click on the Write heading... text to enter text, but nothing happens. No text can be entered. 6. Click on the + inside the Reusable block, and find that it does nothing. **Expected behavior** I expect in step 5 to see the text I'm trying to enter appear. I expect in step 6 to have a new block appear when I hit the +. **Screenshots** ![image](https://user-images.githubusercontent.com/12249954/79046963-61749d80-7c03-11ea-9925-07b1ffd99863.png) The upper box is the non-working saved reusable block, the lower asdf-text is the text inside a regular Heading block. **Desktop (please complete the following information):** - OS: Windows 10 Pro x64 - Browser: Chrome - Version: 80.0.3987.163 **Additional context** - Please add the version of Gutenberg you are using in the description: Version 7.8.1 This is the only thing in the console: load-scripts.php?c=0&load[chunk_0]=jquery-core,jquery-migrate,utils&ver=5.4:8 JQMIGRATE: Migrate is installed, version 1.4.1 post.php?post=25380&action=edit:1 [DOM] Found 2 elements with non-unique id #_wpnonce: (More info: https://goo.gl/9p2vKq) <input type=​"hidden" id=​"_wpnonce" name=​"_wpnonce" value=​"8cf45b233c">​ <input type=​"hidden" id=​"_wpnonce" name=​"_wpnonce" value=​"a13fa8130a">​
True
Block appender appears when a reusable Heading block is selected - **Describe the bug** A modified Heading block saved as a Reusable block appears frozen, it neither accepts text or adds a new block when the + is pressed. **To reproduce** Steps to reproduce the behavior: 1. In the editor, click the + and add a Heading block 2. Change the Heading block to be H3, and Align Text Center. 3. Click on Add to reusable blocks 4. Change Resuable Block name to H3C and hit Save 5. Click on the Write heading... text to enter text, but nothing happens. No text can be entered. 6. Click on the + inside the Reusable block, and find that it does nothing. **Expected behavior** I expect in step 5 to see the text I'm trying to enter appear. I expect in step 6 to have a new block appear when I hit the +. **Screenshots** ![image](https://user-images.githubusercontent.com/12249954/79046963-61749d80-7c03-11ea-9925-07b1ffd99863.png) The upper box is the non-working saved reusable block, the lower asdf-text is the text inside a regular Heading block. **Desktop (please complete the following information):** - OS: Windows 10 Pro x64 - Browser: Chrome - Version: 80.0.3987.163 **Additional context** - Please add the version of Gutenberg you are using in the description: Version 7.8.1 This is the only thing in the console: load-scripts.php?c=0&load[chunk_0]=jquery-core,jquery-migrate,utils&ver=5.4:8 JQMIGRATE: Migrate is installed, version 1.4.1 post.php?post=25380&action=edit:1 [DOM] Found 2 elements with non-unique id #_wpnonce: (More info: https://goo.gl/9p2vKq) <input type=​"hidden" id=​"_wpnonce" name=​"_wpnonce" value=​"8cf45b233c">​ <input type=​"hidden" id=​"_wpnonce" name=​"_wpnonce" value=​"a13fa8130a">​
usab
block appender appears when a reusable heading block is selected describe the bug a modified heading block saved as a reusable block appears frozen it neither accepts text or adds a new block when the is pressed to reproduce steps to reproduce the behavior in the editor click the and add a heading block change the heading block to be and align text center click on add to reusable blocks change resuable block name to and hit save click on the write heading text to enter text but nothing happens no text can be entered click on the inside the reusable block and find that it does nothing expected behavior i expect in step to see the text i m trying to enter appear i expect in step to have a new block appear when i hit the screenshots the upper box is the non working saved reusable block the lower asdf text is the text inside a regular heading block desktop please complete the following information os windows pro browser chrome version additional context please add the version of gutenberg you are using in the description version this is the only thing in the console load scripts php c load jquery core jquery migrate utils ver jqmigrate migrate is installed version post php post action edit found elements with non unique id wpnonce more info ​ ​
1
20,216
15,148,043,958
IssuesEvent
2021-02-11 10:02:43
Elgg/Elgg
https://api.github.com/repos/Elgg/Elgg
closed
site_notification FR: remove notification if content gets removed
feature notifications usability
It would be nice if the site notifications about content were removed when the content is removed. eg: user a writes a blog, users B & C get a notification about this. User A removes the blog before B&C read the notification. User B&C check their notifications and click on the link, but it no longer works. If the notification was related to the content, it would be easy to remove the notification if the content was removed.
True
site_notification FR: remove notification if content gets removed - It would be nice if the site notifications about content were removed when the content is removed. eg: user a writes a blog, users B & C get a notification about this. User A removes the blog before B&C read the notification. User B&C check their notifications and click on the link, but it no longer works. If the notification was related to the content, it would be easy to remove the notification if the content was removed.
usab
site notification fr remove notification if content gets removed it would be nice if the site notifications about content were removed when the content is removed eg user a writes a blog users b c get a notification about this user a removes the blog before b c read the notification user b c check their notifications and click on the link but it no longer works if the notification was related to the content it would be easy to remove the notification if the content was removed
1
178,387
14,668,372,224
IssuesEvent
2020-12-29 21:04:26
melezhik/sparky
https://api.github.com/repos/melezhik/sparky
closed
Problem loading configs hash
documentation enhancement
Hello, `raku bin/sparky-runner.raku --dir=~/.sparky/projects/hello/` is not working because the `~` is not translated to absolute path at this [line](https://github.com/melezhik/sparky/blob/master/bin/sparky-runner.raku#L281). You end up with this empty configuration because we never enter the if block : `merged sparrowdo configuration: { }`
1.0
Problem loading configs hash - Hello, `raku bin/sparky-runner.raku --dir=~/.sparky/projects/hello/` is not working because the `~` is not translated to absolute path at this [line](https://github.com/melezhik/sparky/blob/master/bin/sparky-runner.raku#L281). You end up with this empty configuration because we never enter the if block : `merged sparrowdo configuration: { }`
non_usab
problem loading configs hash hello raku bin sparky runner raku dir sparky projects hello is not working because the is not translated to absolute path at this you end up with this empty configuration because we never enter the if block merged sparrowdo configuration
0
26,133
26,466,017,754
IssuesEvent
2023-01-16 23:41:53
hashicorp/consul
https://api.github.com/repos/hashicorp/consul
closed
Should error on validation when upstream port the same as envoy public listener
type/bug theme/operator-usability theme/connect
#### Overview of the Issue This registration doesn't error: ```hcl service { name = "frontend" port = 6060 connect { sidecar_service { port = 20000 proxy { upstreams = [ { destination_name = "backend" local_bind_port = 20000 } ] } } } } ``` But it results in listeners: ```json { "listener_statuses": [ { "name": "public_listener:0.0.0.0:20000", "local_address": { "socket_address": { "address": "0.0.0.0", "port_value": 20000 } } }, { "name": "backend:127.0.0.1:20000", "local_address": { "socket_address": { "address": "127.0.0.1", "port_value": 20000 } } } ] } ``` Such that in one reproduction upstream requests couldn't be made. When I tried to reproduce, I got hundreds of these errors in the logs, 20/second: ``` 2021-11-15T00:21:43.511Z [ERROR] agent.envoy.xds: got error response from envoy proxy: service_id=backend-sidecar-proxy typeUrl=type.googleapis.com/envoy.config.listener.v3.Listener xdsVersion=v3 nonce=00001ed2 error="rpc error: code = Internal desc = Error adding/updating listener(s) frontend:127.0.0.1:21000: cannot bind '127.0.0.1:21000': Address already in use " ``` But I never saw both listeners bound to the same port in the xds config, so not sure what the difference between my repro and the situation I saw in the wild. Regardless if it's possibly then I think we could safely error out on service registration in this case? ### Consul info for both Client and Server Consul 1.10.3
True
Should error on validation when upstream port the same as envoy public listener - #### Overview of the Issue This registration doesn't error: ```hcl service { name = "frontend" port = 6060 connect { sidecar_service { port = 20000 proxy { upstreams = [ { destination_name = "backend" local_bind_port = 20000 } ] } } } } ``` But it results in listeners: ```json { "listener_statuses": [ { "name": "public_listener:0.0.0.0:20000", "local_address": { "socket_address": { "address": "0.0.0.0", "port_value": 20000 } } }, { "name": "backend:127.0.0.1:20000", "local_address": { "socket_address": { "address": "127.0.0.1", "port_value": 20000 } } } ] } ``` Such that in one reproduction upstream requests couldn't be made. When I tried to reproduce, I got hundreds of these errors in the logs, 20/second: ``` 2021-11-15T00:21:43.511Z [ERROR] agent.envoy.xds: got error response from envoy proxy: service_id=backend-sidecar-proxy typeUrl=type.googleapis.com/envoy.config.listener.v3.Listener xdsVersion=v3 nonce=00001ed2 error="rpc error: code = Internal desc = Error adding/updating listener(s) frontend:127.0.0.1:21000: cannot bind '127.0.0.1:21000': Address already in use " ``` But I never saw both listeners bound to the same port in the xds config, so not sure what the difference between my repro and the situation I saw in the wild. Regardless if it's possibly then I think we could safely error out on service registration in this case? ### Consul info for both Client and Server Consul 1.10.3
usab
should error on validation when upstream port the same as envoy public listener overview of the issue this registration doesn t error hcl service name frontend port connect sidecar service port proxy upstreams destination name backend local bind port but it results in listeners json listener statuses name public listener local address socket address address port value name backend local address socket address address port value such that in one reproduction upstream requests couldn t be made when i tried to reproduce i got hundreds of these errors in the logs second agent envoy xds got error response from envoy proxy service id backend sidecar proxy typeurl type googleapis com envoy config listener listener xdsversion nonce error rpc error code internal desc error adding updating listener s frontend cannot bind address already in use but i never saw both listeners bound to the same port in the xds config so not sure what the difference between my repro and the situation i saw in the wild regardless if it s possibly then i think we could safely error out on service registration in this case consul info for both client and server consul
1
70,052
9,370,477,304
IssuesEvent
2019-04-03 13:33:28
NervanaSystems/coach
https://api.github.com/repos/NervanaSystems/coach
closed
Two typos in coach.py
bug documentation priority/p3
In line 263, 269, 276 of coach.py of master branch, what does the `--experiment_type` means? And the `--environemnt_type` should be `--environment_type`, just a typo. One another problem: the file says > "If no preset is defined, a preset can be set from the command-line by combining settings " "which are set by using --agent_type, --experiment_type, --environemnt_type" but it autually shows this thing when I run `coach -at ddpg_agent -et control_suite_environment -ept e-greedy` : > Please choose a preset using the -p flag or use the --play flag together with choosing an environment type (-et) in order to play the game. The arguments help information is outdated?
1.0
Two typos in coach.py - In line 263, 269, 276 of coach.py of master branch, what does the `--experiment_type` means? And the `--environemnt_type` should be `--environment_type`, just a typo. One another problem: the file says > "If no preset is defined, a preset can be set from the command-line by combining settings " "which are set by using --agent_type, --experiment_type, --environemnt_type" but it autually shows this thing when I run `coach -at ddpg_agent -et control_suite_environment -ept e-greedy` : > Please choose a preset using the -p flag or use the --play flag together with choosing an environment type (-et) in order to play the game. The arguments help information is outdated?
non_usab
two typos in coach py in line of coach py of master branch what does the experiment type means and the environemnt type should be environment type just a typo one another problem the file says if no preset is defined a preset can be set from the command line by combining settings which are set by using agent type experiment type environemnt type but it autually shows this thing when i run coach at ddpg agent et control suite environment ept e greedy please choose a preset using the p flag or use the play flag together with choosing an environment type et in order to play the game the arguments help information is outdated
0
11,305
7,156,472,485
IssuesEvent
2018-01-26 16:24:04
textpattern/textpattern
https://api.github.com/repos/textpattern/textpattern
closed
1x pophelp icon addition: Write -> Status
Usability
I had a client look around the 4.7 dev interface and took some feedback. A lot of it is already in hand, but one thing was something I had overlooked: article statuses from the Write panel. There's no pophelp icon. The docs work has already been done, if we steal the content from `default_publish_status` pophelp and change the heading/intro text, so it's an easy win from that side of things. How reasonable is it to have a pophelp inline'd on `Status`, please. I can take dibs on the glue from the https://github.com/textpattern/pophelp end.
True
1x pophelp icon addition: Write -> Status - I had a client look around the 4.7 dev interface and took some feedback. A lot of it is already in hand, but one thing was something I had overlooked: article statuses from the Write panel. There's no pophelp icon. The docs work has already been done, if we steal the content from `default_publish_status` pophelp and change the heading/intro text, so it's an easy win from that side of things. How reasonable is it to have a pophelp inline'd on `Status`, please. I can take dibs on the glue from the https://github.com/textpattern/pophelp end.
usab
pophelp icon addition write status i had a client look around the dev interface and took some feedback a lot of it is already in hand but one thing was something i had overlooked article statuses from the write panel there s no pophelp icon the docs work has already been done if we steal the content from default publish status pophelp and change the heading intro text so it s an easy win from that side of things how reasonable is it to have a pophelp inline d on status please i can take dibs on the glue from the end
1
27,216
27,856,139,088
IssuesEvent
2023-03-20 23:09:45
googleads/google-ads-dotnet
https://api.github.com/repos/googleads/google-ads-dotnet
closed
Provide a custom compilation of v15.0.0 with lowered dependency on Microsoft.Extensions.*
usability dependencies
When we moved from v13.02 to v14.0.0, we upgraded our Microsoft.Extensions.* dependencies from v3.1.9 to v6.0.0. This was done because v3.1.9 was a very old dependency. However, this has made things difficult for some of our users when upgrading to v13 of the API. So we are using this issue to distribute a custom set of Nuget packages with lowered dependency requirements as a temporary workaround. Use this to complete your v13 upgrade, then continue to upgrade the rest of your code to use v6.0.0 or above of Microsoft.Extensions.*. To use these dependencies, 1. Download the for-release.zip and extract it to a local folder. 2. Remove your project's dependency on Google.Ads.GoogleAds. 3. Add a local package source in Visual Studio: https://learn.microsoft.com/en-us/nuget/consume-packages/install-use-packages-visual-studio#package-sources 4. Install Google.Ads.GoogleAds v15.0.0 from the local folder source. [for-release.zip](https://github.com/googleads/google-ads-dotnet/files/10984175/for-release.zip)
True
Provide a custom compilation of v15.0.0 with lowered dependency on Microsoft.Extensions.* - When we moved from v13.02 to v14.0.0, we upgraded our Microsoft.Extensions.* dependencies from v3.1.9 to v6.0.0. This was done because v3.1.9 was a very old dependency. However, this has made things difficult for some of our users when upgrading to v13 of the API. So we are using this issue to distribute a custom set of Nuget packages with lowered dependency requirements as a temporary workaround. Use this to complete your v13 upgrade, then continue to upgrade the rest of your code to use v6.0.0 or above of Microsoft.Extensions.*. To use these dependencies, 1. Download the for-release.zip and extract it to a local folder. 2. Remove your project's dependency on Google.Ads.GoogleAds. 3. Add a local package source in Visual Studio: https://learn.microsoft.com/en-us/nuget/consume-packages/install-use-packages-visual-studio#package-sources 4. Install Google.Ads.GoogleAds v15.0.0 from the local folder source. [for-release.zip](https://github.com/googleads/google-ads-dotnet/files/10984175/for-release.zip)
usab
provide a custom compilation of with lowered dependency on microsoft extensions when we moved from to we upgraded our microsoft extensions dependencies from to this was done because was a very old dependency however this has made things difficult for some of our users when upgrading to of the api so we are using this issue to distribute a custom set of nuget packages with lowered dependency requirements as a temporary workaround use this to complete your upgrade then continue to upgrade the rest of your code to use or above of microsoft extensions to use these dependencies download the for release zip and extract it to a local folder remove your project s dependency on google ads googleads add a local package source in visual studio install google ads googleads from the local folder source
1
11,630
7,330,864,810
IssuesEvent
2018-03-05 11:24:56
nerdalize/nerd
https://api.github.com/repos/nerdalize/nerd
closed
Fail the command when too many arguments are given
P0 usability
Currently when a command is given too many argument it just takes the first. Instead we want to fail because we're running something that the user didn't intend. Exception is the `nerd job run`
True
Fail the command when too many arguments are given - Currently when a command is given too many argument it just takes the first. Instead we want to fail because we're running something that the user didn't intend. Exception is the `nerd job run`
usab
fail the command when too many arguments are given currently when a command is given too many argument it just takes the first instead we want to fail because we re running something that the user didn t intend exception is the nerd job run
1
260,215
22,601,353,561
IssuesEvent
2022-06-29 09:24:32
JanusGraph/janusgraph-dotnet
https://api.github.com/repos/JanusGraph/janusgraph-dotnet
opened
Migrate to new official testcontainers-dotnet library
area/testing
The version we just migrated to in #96 is now the official .NET version of Testcontainers: testcontainers/testcontainers-dotnet#476. The migration should therefore be straightforward.
1.0
Migrate to new official testcontainers-dotnet library - The version we just migrated to in #96 is now the official .NET version of Testcontainers: testcontainers/testcontainers-dotnet#476. The migration should therefore be straightforward.
non_usab
migrate to new official testcontainers dotnet library the version we just migrated to in is now the official net version of testcontainers testcontainers testcontainers dotnet the migration should therefore be straightforward
0
10,060
6,552,280,467
IssuesEvent
2017-09-05 17:41:16
gluster/glusterfs
https://api.github.com/repos/gluster/glusterfs
closed
Gluster threads should be named
CB: libglusterfs FA: Debug-ability & Quality FA: Usability & Supportability Type:Enhancement
A gluster process has many threads and being able to see the list of threads with the cpu consumption is a very useful feature.
True
Gluster threads should be named - A gluster process has many threads and being able to see the list of threads with the cpu consumption is a very useful feature.
usab
gluster threads should be named a gluster process has many threads and being able to see the list of threads with the cpu consumption is a very useful feature
1
21,924
10,699,780,129
IssuesEvent
2019-10-23 21:47:49
aspnet/AspNetCore
https://api.github.com/repos/aspnet/AspNetCore
closed
AllowAnonymousAttribute not working as expected with CORS request.
area-security
I'm Running ASP.NET Core 3 (Azure Web App) and have a CORS policy set up like so, in Startup.cs: services.AddCors(options => { options.AddPolicy("TestCors", builder => builder.WithOrigins(allowedDomains.ToArray()) .AllowAnyHeader() .AllowAnyMethod() .AllowCredentials()); }); I have an AJAX endpoint in a controller defined as such: [EnableCors("TestCors")] [AllowAnonymous] public async Task<IActionResult> MyAjaxEndpoint(int parm1) { } When I attempt to call this controller method using JQuery AJAX method like so: $.ajax( { url: 'otherdomain.com/MyAjaxEndpoint', cache: false, type: "POST", crossDomain: true, data: {parm1: 0}, xhrFields: { withCredentials: true }, async: false }); , I get a 302 redirect to the login page. This does not happen if I happen to be authenticated through the standard Identity cookie. I have some tokens (JWT) stored in a http-only cookie that I want sent with AJAX requests, and I was under the impression that requirement necessitated `withCredentials: true` in the AJAX options (is that correct?). I'm not sure if that's what's causing the problem. At any rate, I assumed that `[AllowAnonymous]` would have bypassed any authentication requirements by the ASP.NET pipeline. Is there an issue with the AllowAnonymousAttribute or am I (ab)using it wrong?
True
AllowAnonymousAttribute not working as expected with CORS request. - I'm Running ASP.NET Core 3 (Azure Web App) and have a CORS policy set up like so, in Startup.cs: services.AddCors(options => { options.AddPolicy("TestCors", builder => builder.WithOrigins(allowedDomains.ToArray()) .AllowAnyHeader() .AllowAnyMethod() .AllowCredentials()); }); I have an AJAX endpoint in a controller defined as such: [EnableCors("TestCors")] [AllowAnonymous] public async Task<IActionResult> MyAjaxEndpoint(int parm1) { } When I attempt to call this controller method using JQuery AJAX method like so: $.ajax( { url: 'otherdomain.com/MyAjaxEndpoint', cache: false, type: "POST", crossDomain: true, data: {parm1: 0}, xhrFields: { withCredentials: true }, async: false }); , I get a 302 redirect to the login page. This does not happen if I happen to be authenticated through the standard Identity cookie. I have some tokens (JWT) stored in a http-only cookie that I want sent with AJAX requests, and I was under the impression that requirement necessitated `withCredentials: true` in the AJAX options (is that correct?). I'm not sure if that's what's causing the problem. At any rate, I assumed that `[AllowAnonymous]` would have bypassed any authentication requirements by the ASP.NET pipeline. Is there an issue with the AllowAnonymousAttribute or am I (ab)using it wrong?
non_usab
allowanonymousattribute not working as expected with cors request i m running asp net core azure web app and have a cors policy set up like so in startup cs services addcors options options addpolicy testcors builder builder withorigins alloweddomains toarray allowanyheader allowanymethod allowcredentials i have an ajax endpoint in a controller defined as such public async task myajaxendpoint int when i attempt to call this controller method using jquery ajax method like so ajax url otherdomain com myajaxendpoint cache false type post crossdomain true data xhrfields withcredentials true async false i get a redirect to the login page this does not happen if i happen to be authenticated through the standard identity cookie i have some tokens jwt stored in a http only cookie that i want sent with ajax requests and i was under the impression that requirement necessitated withcredentials true in the ajax options is that correct i m not sure if that s what s causing the problem at any rate i assumed that would have bypassed any authentication requirements by the asp net pipeline is there an issue with the allowanonymousattribute or am i ab using it wrong
0
665,096
22,299,386,705
IssuesEvent
2022-06-13 07:13:29
openforis/arena
https://api.github.com/repos/openforis/arena
closed
Chain: Sampling design - new elements
Priority_1
1. weight, always TRUE - Arena UI: no actions - If "Sampling design" set in Arena, always weighted - If no weighting, set all base units' weights to 1 2. clustering: Clustered / No-clustered - Arena UI: no actions - if clustering entity selected, then TRUE - a clustering entity can be any parent entity for the base unit entity 3. sampling strategy: - Arena: add a new dropdown list ('samplingStrategy') 0. --not specified-- (default) 1. SIMPLE RANDOM SAMPLING 2. SYSTEMATIC SAMPLING 3. STRATIFIED RANDOM SAMPLING 4. STRATIFIED SYSTEMATIC SAMPLING 5. DOUBLE PHASE SAMPLING, requires 1 phase table (coming later) * in Arena: hide still 4. stratification: - Arena UI: move under "sampling strategy" - if method 1 or 2, hide/inenable dropdown - if method 3, 4, or 5, show dropdown list to select 'stratumAttribute', - - methods 3 and 4 must have a selection, method 5 can be --not specified-- - 'stratumAttribute' can be any categorical attribute of the base unit entity or its parents NOTE: 'stratumAttribute' can also be an active computed categorical result variable 5. post-stratification - Arena UI: add a new dropdown list, default: --not specified- - post-stratification can be applied with methods 1-4 - select a categorical attribute ('post_stratumAttribute') from the dropdown list. This can be any categorical attribute of the base unit entity or its parents, but not the same as 'stratumAttribute' NOTE: 'post_stratumAttribute' can also be an active computed categorical result variable 6. p-value <- REQUEST CANCELLED!! (This will be added to another page) - Arena UI: add a new dropdown list ('arena_p_value') - list items: (0.99, 0.98, 0.95, 0.90, 0.80). Default: 0.95 7. submit new information into json: 'samplingStrategy' (index), 'post_stratumAttribute' (name), 'arena_p_value' (numeric) ______________________ NEW EDIT: 8. Add a check box with label text: "Nonresponse bias correction" (default FALSE) 9. Add parameter "nonresponse_correction" (FALSE/TRUE) into the json file.
1.0
Chain: Sampling design - new elements - 1. weight, always TRUE - Arena UI: no actions - If "Sampling design" set in Arena, always weighted - If no weighting, set all base units' weights to 1 2. clustering: Clustered / No-clustered - Arena UI: no actions - if clustering entity selected, then TRUE - a clustering entity can be any parent entity for the base unit entity 3. sampling strategy: - Arena: add a new dropdown list ('samplingStrategy') 0. --not specified-- (default) 1. SIMPLE RANDOM SAMPLING 2. SYSTEMATIC SAMPLING 3. STRATIFIED RANDOM SAMPLING 4. STRATIFIED SYSTEMATIC SAMPLING 5. DOUBLE PHASE SAMPLING, requires 1 phase table (coming later) * in Arena: hide still 4. stratification: - Arena UI: move under "sampling strategy" - if method 1 or 2, hide/inenable dropdown - if method 3, 4, or 5, show dropdown list to select 'stratumAttribute', - - methods 3 and 4 must have a selection, method 5 can be --not specified-- - 'stratumAttribute' can be any categorical attribute of the base unit entity or its parents NOTE: 'stratumAttribute' can also be an active computed categorical result variable 5. post-stratification - Arena UI: add a new dropdown list, default: --not specified- - post-stratification can be applied with methods 1-4 - select a categorical attribute ('post_stratumAttribute') from the dropdown list. This can be any categorical attribute of the base unit entity or its parents, but not the same as 'stratumAttribute' NOTE: 'post_stratumAttribute' can also be an active computed categorical result variable 6. p-value <- REQUEST CANCELLED!! (This will be added to another page) - Arena UI: add a new dropdown list ('arena_p_value') - list items: (0.99, 0.98, 0.95, 0.90, 0.80). Default: 0.95 7. submit new information into json: 'samplingStrategy' (index), 'post_stratumAttribute' (name), 'arena_p_value' (numeric) ______________________ NEW EDIT: 8. Add a check box with label text: "Nonresponse bias correction" (default FALSE) 9. Add parameter "nonresponse_correction" (FALSE/TRUE) into the json file.
non_usab
chain sampling design new elements weight always true arena ui no actions if sampling design set in arena always weighted if no weighting set all base units weights to clustering clustered no clustered arena ui no actions if clustering entity selected then true a clustering entity can be any parent entity for the base unit entity sampling strategy arena add a new dropdown list samplingstrategy not specified default simple random sampling systematic sampling stratified random sampling stratified systematic sampling double phase sampling requires phase table coming later in arena hide still stratification arena ui move under sampling strategy if method or hide inenable dropdown if method or show dropdown list to select stratumattribute methods and must have a selection method can be not specified stratumattribute can be any categorical attribute of the base unit entity or its parents note stratumattribute can also be an active computed categorical result variable post stratification arena ui add a new dropdown list default not specified post stratification can be applied with methods select a categorical attribute post stratumattribute from the dropdown list this can be any categorical attribute of the base unit entity or its parents but not the same as stratumattribute note post stratumattribute can also be an active computed categorical result variable p value request cancelled this will be added to another page arena ui add a new dropdown list arena p value list items default submit new information into json samplingstrategy index post stratumattribute name arena p value numeric new edit add a check box with label text nonresponse bias correction default false add parameter nonresponse correction false true into the json file
0
2,889
3,952,480,630
IssuesEvent
2016-04-29 09:01:10
ScottLogic/StockFlux
https://api.github.com/repos/ScottLogic/StockFlux
closed
Creating a contributing document
infrastructure
We should have a brief document that describes how to contribute, as per: https://github.com/blog/1184-contributing-guidelines It should detail the following: - [ ] Which issue labels we use and what they mean (removing or adding as necessary) - [ ] Our approach to branching (removing any dead branches) - [ ] How we develop / test / release This can be a very brief document, but just needs putting in writing!
1.0
Creating a contributing document - We should have a brief document that describes how to contribute, as per: https://github.com/blog/1184-contributing-guidelines It should detail the following: - [ ] Which issue labels we use and what they mean (removing or adding as necessary) - [ ] Our approach to branching (removing any dead branches) - [ ] How we develop / test / release This can be a very brief document, but just needs putting in writing!
non_usab
creating a contributing document we should have a brief document that describes how to contribute as per it should detail the following which issue labels we use and what they mean removing or adding as necessary our approach to branching removing any dead branches how we develop test release this can be a very brief document but just needs putting in writing
0
129,026
5,082,318,255
IssuesEvent
2016-12-29 15:04:48
wireservice/csvkit
https://api.github.com/repos/wireservice/csvkit
opened
csvlook: 'NoneType' object has no attribute 'groupdict'
bug High Priority
Using test file from #631. Also reported via Twitter by another user. ``` Traceback (most recent call last): File "/Users/cgroskopf/.virtualenvs/csvkit/bin/csvlook", line 9, in <module> load_entry_point('csvkit==1.0.1', 'console_scripts', 'csvlook')() File "/Users/cgroskopf/src/csvkit/csvkit/utilities/csvlook.py", line 44, in launch_new_instance utility.run() File "/Users/cgroskopf/src/csvkit/csvkit/cli.py", line 113, in run self.main() File "/Users/cgroskopf/src/csvkit/csvkit/utilities/csvlook.py", line 31, in main **self.reader_kwargs File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/table/from_csv.py", line 88, in from_csv return Table(rows, column_names, column_types, row_names=row_names) File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/table/__init__.py", line 132, in __init__ self._column_types = column_types.run(rows, self._column_names) File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/type_tester.py", line 110, in run if len(row) > i and not column_type.test(row[i]): File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/data_types/base.py", line 29, in test self.cast(d) File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/data_types/time_delta.py", line 35, in cast seconds = pytimeparse.parse(d) File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/pytimeparse/timeparse.py", line 147, in timeparse sign = -1 if match.groupdict()['sign'] == '-' else 1 AttributeError: 'NoneType' object has no attribute 'groupdict' ```
1.0
csvlook: 'NoneType' object has no attribute 'groupdict' - Using test file from #631. Also reported via Twitter by another user. ``` Traceback (most recent call last): File "/Users/cgroskopf/.virtualenvs/csvkit/bin/csvlook", line 9, in <module> load_entry_point('csvkit==1.0.1', 'console_scripts', 'csvlook')() File "/Users/cgroskopf/src/csvkit/csvkit/utilities/csvlook.py", line 44, in launch_new_instance utility.run() File "/Users/cgroskopf/src/csvkit/csvkit/cli.py", line 113, in run self.main() File "/Users/cgroskopf/src/csvkit/csvkit/utilities/csvlook.py", line 31, in main **self.reader_kwargs File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/table/from_csv.py", line 88, in from_csv return Table(rows, column_names, column_types, row_names=row_names) File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/table/__init__.py", line 132, in __init__ self._column_types = column_types.run(rows, self._column_names) File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/type_tester.py", line 110, in run if len(row) > i and not column_type.test(row[i]): File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/data_types/base.py", line 29, in test self.cast(d) File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/agate/data_types/time_delta.py", line 35, in cast seconds = pytimeparse.parse(d) File "/Users/cgroskopf/.virtualenvs/csvkit/lib/python3.5/site-packages/pytimeparse/timeparse.py", line 147, in timeparse sign = -1 if match.groupdict()['sign'] == '-' else 1 AttributeError: 'NoneType' object has no attribute 'groupdict' ```
non_usab
csvlook nonetype object has no attribute groupdict using test file from also reported via twitter by another user traceback most recent call last file users cgroskopf virtualenvs csvkit bin csvlook line in load entry point csvkit console scripts csvlook file users cgroskopf src csvkit csvkit utilities csvlook py line in launch new instance utility run file users cgroskopf src csvkit csvkit cli py line in run self main file users cgroskopf src csvkit csvkit utilities csvlook py line in main self reader kwargs file users cgroskopf virtualenvs csvkit lib site packages agate table from csv py line in from csv return table rows column names column types row names row names file users cgroskopf virtualenvs csvkit lib site packages agate table init py line in init self column types column types run rows self column names file users cgroskopf virtualenvs csvkit lib site packages agate type tester py line in run if len row i and not column type test row file users cgroskopf virtualenvs csvkit lib site packages agate data types base py line in test self cast d file users cgroskopf virtualenvs csvkit lib site packages agate data types time delta py line in cast seconds pytimeparse parse d file users cgroskopf virtualenvs csvkit lib site packages pytimeparse timeparse py line in timeparse sign if match groupdict else attributeerror nonetype object has no attribute groupdict
0
22,470
19,489,894,867
IssuesEvent
2021-12-27 03:11:57
Leafwing-Studios/leafwing-input-manager
https://api.github.com/repos/Leafwing-Studios/leafwing-input-manager
closed
Add derive macro for Actionlike
usability
## Which feature is frustrating to use or confusing? Actionlike needs to manually implemented. Manual impls always suck when they don't contain any logic. ## Expectation Derive macro
True
Add derive macro for Actionlike - ## Which feature is frustrating to use or confusing? Actionlike needs to manually implemented. Manual impls always suck when they don't contain any logic. ## Expectation Derive macro
usab
add derive macro for actionlike which feature is frustrating to use or confusing actionlike needs to manually implemented manual impls always suck when they don t contain any logic expectation derive macro
1
208,242
7,137,472,793
IssuesEvent
2018-01-23 11:07:23
openshift/origin
https://api.github.com/repos/openshift/origin
closed
Cannot oc set image/env for CronJob
kind/bug priority/P2
Hi, I can get a scheduledJob working using the example in https://docs.openshift.com/container-platform/3.3/dev_guide/scheduled_jobs.html, but I am not able to use an IS in current project unless I specify the absolute image path in the format <docker_registry_service_ip>:5000/\<project\>/<image_name>. Note I cannot even use the public docker registry host name. This is different from how image can be specified in other resource types such as DeploymentConfig. It makes the template containing this resource not portable. ##### Version ``` C:\Users\x>oc version oc v1.5.0-alpha.2+e4b43ee kubernetes v1.5.2+43a9be4 features: Basic-Auth Server https://<console_url>:8443 openshift v3.3.1.7 kubernetes v1.3.0+52492b4 ```
1.0
Cannot oc set image/env for CronJob - Hi, I can get a scheduledJob working using the example in https://docs.openshift.com/container-platform/3.3/dev_guide/scheduled_jobs.html, but I am not able to use an IS in current project unless I specify the absolute image path in the format <docker_registry_service_ip>:5000/\<project\>/<image_name>. Note I cannot even use the public docker registry host name. This is different from how image can be specified in other resource types such as DeploymentConfig. It makes the template containing this resource not portable. ##### Version ``` C:\Users\x>oc version oc v1.5.0-alpha.2+e4b43ee kubernetes v1.5.2+43a9be4 features: Basic-Auth Server https://<console_url>:8443 openshift v3.3.1.7 kubernetes v1.3.0+52492b4 ```
non_usab
cannot oc set image env for cronjob hi i can get a scheduledjob working using the example in but i am not able to use an is in current project unless i specify the absolute image path in the format note i cannot even use the public docker registry host name this is different from how image can be specified in other resource types such as deploymentconfig it makes the template containing this resource not portable version c users x oc version oc alpha kubernetes features basic auth server openshift kubernetes
0
28,074
31,561,308,613
IssuesEvent
2023-09-03 09:38:01
git-cola/git-cola
https://api.github.com/repos/git-cola/git-cola
closed
UI is broken in macOS dark mode
usability macos qt bug
On macOS Mojave with the new dark theme, Cola seems to not pick up the text and background colors properly, it is probably overriding one without overriding the other. I'm attaching two screenshots. In one nothing is visible, in the other I have selected some text elements to show that they are indeed there. ## Steps to reproduce: 1. Enable dark theme in macOS Mojave (Settings->General) 2. Open git cola in a modified git directory (must be opened after theme is selected) 3. Notice that text (diff, filenames, commit headline, commit body) is invisible ## Expected Result Text is visible (either white-on-black or black-on-white) ## Screenshots <img width="1265" alt="screenshot 2018-11-14 at 01 36 07 pm" src="https://user-images.githubusercontent.com/3298/48504446-c2335700-e812-11e8-8628-305c624ac0b1.png"> <img width="1265" alt="screenshot 2018-11-14 at 01 36 26 pm" src="https://user-images.githubusercontent.com/3298/48504452-c8293800-e812-11e8-802d-7c0c75827d24.png">
True
UI is broken in macOS dark mode - On macOS Mojave with the new dark theme, Cola seems to not pick up the text and background colors properly, it is probably overriding one without overriding the other. I'm attaching two screenshots. In one nothing is visible, in the other I have selected some text elements to show that they are indeed there. ## Steps to reproduce: 1. Enable dark theme in macOS Mojave (Settings->General) 2. Open git cola in a modified git directory (must be opened after theme is selected) 3. Notice that text (diff, filenames, commit headline, commit body) is invisible ## Expected Result Text is visible (either white-on-black or black-on-white) ## Screenshots <img width="1265" alt="screenshot 2018-11-14 at 01 36 07 pm" src="https://user-images.githubusercontent.com/3298/48504446-c2335700-e812-11e8-8628-305c624ac0b1.png"> <img width="1265" alt="screenshot 2018-11-14 at 01 36 26 pm" src="https://user-images.githubusercontent.com/3298/48504452-c8293800-e812-11e8-802d-7c0c75827d24.png">
usab
ui is broken in macos dark mode on macos mojave with the new dark theme cola seems to not pick up the text and background colors properly it is probably overriding one without overriding the other i m attaching two screenshots in one nothing is visible in the other i have selected some text elements to show that they are indeed there steps to reproduce enable dark theme in macos mojave settings general open git cola in a modified git directory must be opened after theme is selected notice that text diff filenames commit headline commit body is invisible expected result text is visible either white on black or black on white screenshots img width alt screenshot at pm src img width alt screenshot at pm src
1
10,160
6,608,328,081
IssuesEvent
2017-09-19 10:31:42
zcash/zcash
https://api.github.com/repos/zcash/zcash
closed
Benchmark wallet load time in realistic and/or worst cases.
benchmarking performance usability
Anecdotally the Zcash wallet load time is frustrating for users. I've heard at least two occasions of operators changing their behavior to work around having to restart due to the downtime. Let's benchmark wallet loading.
True
Benchmark wallet load time in realistic and/or worst cases. - Anecdotally the Zcash wallet load time is frustrating for users. I've heard at least two occasions of operators changing their behavior to work around having to restart due to the downtime. Let's benchmark wallet loading.
usab
benchmark wallet load time in realistic and or worst cases anecdotally the zcash wallet load time is frustrating for users i ve heard at least two occasions of operators changing their behavior to work around having to restart due to the downtime let s benchmark wallet loading
1
26,450
26,874,714,006
IssuesEvent
2023-02-04 22:22:43
bevyengine/bevy
https://api.github.com/repos/bevyengine/bevy
opened
`AssetServer::get_load_state()` returns `NotLoaded` right after a call to `load()`
C-Bug C-Docs A-Assets C-Usability
## Bevy version 0.9.1 ## What you did Load a bunch of assets at once, which will increase the chance that an asset is not "immediately" loaded, that is there's an observable delay between the call to `AssetServer::load()` and the time the internal async load is started. ## What went wrong `AssetServer::get_load_state()` returns `NotLoaded` instead of the expected `Loading`. ## Additional information It looks like a race condition. `AssetServer::load()` enqueues the loading task via the `IoTaskPool`, but doesn't immediately update the `asset_sources` before returning. Later, when the loading task runs, the internal `load_async()` actually updates `asset_sources`. In the meantime, there's a short period of time where calling `AssetServer::get_load_state()` will query `asset_sources` and not find the asset, and therefore return `NotLoaded`. From a user perspective, I believe the `AssetServer` should return `Loading` immediately after a call to `load()` returned. The fact the internal async loading task didn't yet start is an implementation detail that the user doesn't care about. `NotLoaded` makes it look like the asset is not even scheduled for loading, like the call to `load()` had somehow failed. See also: - #6135 - #7479
True
`AssetServer::get_load_state()` returns `NotLoaded` right after a call to `load()` - ## Bevy version 0.9.1 ## What you did Load a bunch of assets at once, which will increase the chance that an asset is not "immediately" loaded, that is there's an observable delay between the call to `AssetServer::load()` and the time the internal async load is started. ## What went wrong `AssetServer::get_load_state()` returns `NotLoaded` instead of the expected `Loading`. ## Additional information It looks like a race condition. `AssetServer::load()` enqueues the loading task via the `IoTaskPool`, but doesn't immediately update the `asset_sources` before returning. Later, when the loading task runs, the internal `load_async()` actually updates `asset_sources`. In the meantime, there's a short period of time where calling `AssetServer::get_load_state()` will query `asset_sources` and not find the asset, and therefore return `NotLoaded`. From a user perspective, I believe the `AssetServer` should return `Loading` immediately after a call to `load()` returned. The fact the internal async loading task didn't yet start is an implementation detail that the user doesn't care about. `NotLoaded` makes it look like the asset is not even scheduled for loading, like the call to `load()` had somehow failed. See also: - #6135 - #7479
usab
assetserver get load state returns notloaded right after a call to load bevy version what you did load a bunch of assets at once which will increase the chance that an asset is not immediately loaded that is there s an observable delay between the call to assetserver load and the time the internal async load is started what went wrong assetserver get load state returns notloaded instead of the expected loading additional information it looks like a race condition assetserver load enqueues the loading task via the iotaskpool but doesn t immediately update the asset sources before returning later when the loading task runs the internal load async actually updates asset sources in the meantime there s a short period of time where calling assetserver get load state will query asset sources and not find the asset and therefore return notloaded from a user perspective i believe the assetserver should return loading immediately after a call to load returned the fact the internal async loading task didn t yet start is an implementation detail that the user doesn t care about notloaded makes it look like the asset is not even scheduled for loading like the call to load had somehow failed see also
1
621,555
19,591,232,457
IssuesEvent
2022-01-05 13:13:55
o3de/o3de
https://api.github.com/repos/o3de/o3de
opened
Feature Request: Add an option to apply rendering "filters" to a camera
needs-triage sig/graphics-audio priority/minor kind/story
**Is your feature request related to a problem? Please describe.** Currently there are some components available that allow manipulating the way light is displayed (e.g. Look Modification or Display Mapper), but there is no simple similar way of adding some filters that change the way the image is drawn on screen. For example in the Select Color window available for the color related parameters (such as Shape Color on the Box Shape component), there is a color picker that shows an overview of colors that the color points to, dividing the viewed pixels into larger pixels with "averaged out" colors - this gives the effect of a "pixel filter". Similar thing can be noticed on using the debug Heatmap on the Exposure Control component, where adjusting it gives a peculiar look of blue and red bleeding into normal colors depending on the set Minimum/Maximum Exposure values. Please refer to the attached video where the described effects are shown. **Describe the solution you'd like** Add a component with some camera filters included (e.g. "pixel filter" similar to the described color picker effect) and a way to add more. **Describe alternatives you've considered** Add any way of applying some of the described "filters" to e.g. Camera component directly, or any other related components. **Additional context** https://user-images.githubusercontent.com/86952082/148221183-d8f70ac0-f43a-4e77-9aa3-f4b458334083.mp4
1.0
Feature Request: Add an option to apply rendering "filters" to a camera - **Is your feature request related to a problem? Please describe.** Currently there are some components available that allow manipulating the way light is displayed (e.g. Look Modification or Display Mapper), but there is no simple similar way of adding some filters that change the way the image is drawn on screen. For example in the Select Color window available for the color related parameters (such as Shape Color on the Box Shape component), there is a color picker that shows an overview of colors that the color points to, dividing the viewed pixels into larger pixels with "averaged out" colors - this gives the effect of a "pixel filter". Similar thing can be noticed on using the debug Heatmap on the Exposure Control component, where adjusting it gives a peculiar look of blue and red bleeding into normal colors depending on the set Minimum/Maximum Exposure values. Please refer to the attached video where the described effects are shown. **Describe the solution you'd like** Add a component with some camera filters included (e.g. "pixel filter" similar to the described color picker effect) and a way to add more. **Describe alternatives you've considered** Add any way of applying some of the described "filters" to e.g. Camera component directly, or any other related components. **Additional context** https://user-images.githubusercontent.com/86952082/148221183-d8f70ac0-f43a-4e77-9aa3-f4b458334083.mp4
non_usab
feature request add an option to apply rendering filters to a camera is your feature request related to a problem please describe currently there are some components available that allow manipulating the way light is displayed e g look modification or display mapper but there is no simple similar way of adding some filters that change the way the image is drawn on screen for example in the select color window available for the color related parameters such as shape color on the box shape component there is a color picker that shows an overview of colors that the color points to dividing the viewed pixels into larger pixels with averaged out colors this gives the effect of a pixel filter similar thing can be noticed on using the debug heatmap on the exposure control component where adjusting it gives a peculiar look of blue and red bleeding into normal colors depending on the set minimum maximum exposure values please refer to the attached video where the described effects are shown describe the solution you d like add a component with some camera filters included e g pixel filter similar to the described color picker effect and a way to add more describe alternatives you ve considered add any way of applying some of the described filters to e g camera component directly or any other related components additional context
0
295,951
25,517,026,258
IssuesEvent
2022-11-28 17:07:17
vegaprotocol/frontend-monorepo
https://api.github.com/repos/vegaprotocol/frontend-monorepo
closed
Migrate trading e2e tests to use vegawallet v2
Trading Testing 🧪 chore
## The Chore We need to switch to vegawallet v2, tests need to be updated and verified if no issues are caused by using the new wallet
1.0
Migrate trading e2e tests to use vegawallet v2 - ## The Chore We need to switch to vegawallet v2, tests need to be updated and verified if no issues are caused by using the new wallet
non_usab
migrate trading tests to use vegawallet the chore we need to switch to vegawallet tests need to be updated and verified if no issues are caused by using the new wallet
0
329
2,499,836,505
IssuesEvent
2015-01-08 06:42:39
fossology/fossology
https://api.github.com/repos/fossology/fossology
opened
bucket tag link not active
Category: Buckets Component: Rank Component: Tester Priority: Normal Status: New Tracker: Bug
--- Author Name: **Bob Gobeille** Original Redmine Issue: 1029, http://www.fossology.org/issues/1029 Original Date: 2012/02/14 Original Assignee: Bob Gobeille --- The bucket browser has a tag link in the histogram, but the link is not active (it's just text). The idea is to make it easy to tag all the items in a bucket.
1.0
bucket tag link not active - --- Author Name: **Bob Gobeille** Original Redmine Issue: 1029, http://www.fossology.org/issues/1029 Original Date: 2012/02/14 Original Assignee: Bob Gobeille --- The bucket browser has a tag link in the histogram, but the link is not active (it's just text). The idea is to make it easy to tag all the items in a bucket.
non_usab
bucket tag link not active author name bob gobeille original redmine issue original date original assignee bob gobeille the bucket browser has a tag link in the histogram but the link is not active it s just text the idea is to make it easy to tag all the items in a bucket
0
4,566
3,872,475,777
IssuesEvent
2016-04-11 14:01:19
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
22542457: Deleting a Photo in Photos app is Unbearably Slow
classification:ui/usability reproducible:always status:open
#### Description Summary: Photos takes a very long time to delete photos. After a deletion action is initiated, several seconds can sometimes pass before the UI is updated. This causes issues where the UI is jarringly updated when it is not expected to be updated. Changing selection status (when the UI is actually updated) can cause confusion and unintended deletion (if attempting to delete multiple photos in quick succession). Rather than photos actually being deleted on the main thread (as seems to be the case), perhaps it would be better to immediately update the UI and execute the deletion in the background. If a failure occurs (hopefully the non-normal case), a message could be shown to the user and the UI could be reverted. This issue happens in Photos Version 1.1 (330.21.0) Steps to Reproduce: + Open Photos + Delete a photo Expected Results: Photo is immediately removed from the UI, and scheduled for deletion in the background Actual Results: Photos UI takes a very long time (several seconds sometimes) to remove the photo from the interface - Product Version: 10.11 Beta (15A279b) Created: 2015-09-02T18:34:55.374520 Originated: 2015-09-02T14:34:00 Open Radar Link: http://www.openradar.me/22542457
True
22542457: Deleting a Photo in Photos app is Unbearably Slow - #### Description Summary: Photos takes a very long time to delete photos. After a deletion action is initiated, several seconds can sometimes pass before the UI is updated. This causes issues where the UI is jarringly updated when it is not expected to be updated. Changing selection status (when the UI is actually updated) can cause confusion and unintended deletion (if attempting to delete multiple photos in quick succession). Rather than photos actually being deleted on the main thread (as seems to be the case), perhaps it would be better to immediately update the UI and execute the deletion in the background. If a failure occurs (hopefully the non-normal case), a message could be shown to the user and the UI could be reverted. This issue happens in Photos Version 1.1 (330.21.0) Steps to Reproduce: + Open Photos + Delete a photo Expected Results: Photo is immediately removed from the UI, and scheduled for deletion in the background Actual Results: Photos UI takes a very long time (several seconds sometimes) to remove the photo from the interface - Product Version: 10.11 Beta (15A279b) Created: 2015-09-02T18:34:55.374520 Originated: 2015-09-02T14:34:00 Open Radar Link: http://www.openradar.me/22542457
usab
deleting a photo in photos app is unbearably slow description summary photos takes a very long time to delete photos after a deletion action is initiated several seconds can sometimes pass before the ui is updated this causes issues where the ui is jarringly updated when it is not expected to be updated changing selection status when the ui is actually updated can cause confusion and unintended deletion if attempting to delete multiple photos in quick succession rather than photos actually being deleted on the main thread as seems to be the case perhaps it would be better to immediately update the ui and execute the deletion in the background if a failure occurs hopefully the non normal case a message could be shown to the user and the ui could be reverted this issue happens in photos version steps to reproduce open photos delete a photo expected results photo is immediately removed from the ui and scheduled for deletion in the background actual results photos ui takes a very long time several seconds sometimes to remove the photo from the interface product version beta created originated open radar link
1
8,979
7,522,647,259
IssuesEvent
2018-04-12 21:08:43
istio/istio
https://api.github.com/repos/istio/istio
opened
Pilot fetches and distributes JWT public key (to sidecar)
area/security
In 0.7, envoy sidecar need to fetch public key from `JwksURI` (specified as part of JWT filter config). This requires: * Pilot need to create outbound cluster(s) for JwksURI, which have some conflicts with v2+v1alpha3 (particularly, concept of `ExternalService`) * In term of performance, each sidecar will need to fetch the keys independently. An alternative solution is to have pilot fetch the keys and distribute the them to sidecars; sidecars can use the keys directly, without any extra communication to external services. To make this works, following assumptions/requirements need to be met: 1. Pilot is allowed to talk to external web services. (does pilot sidecar needs to be configured somehow?) 1. Implementation for cache and key rotation in pilot (i.e if key is updated, it should trigger pilot to push new configs to services). 1. Security concerns: is this acceptable if control plane doesn't use mTLS. Related issues: * [Proxy support public key in filter config](https://github.com/istio/proxy/issues/1373) * [Support OpenID Discovery](https://github.com/istio/istio/issues/4719) * https://github.com/istio/istio/issues/4633
True
Pilot fetches and distributes JWT public key (to sidecar) - In 0.7, envoy sidecar need to fetch public key from `JwksURI` (specified as part of JWT filter config). This requires: * Pilot need to create outbound cluster(s) for JwksURI, which have some conflicts with v2+v1alpha3 (particularly, concept of `ExternalService`) * In term of performance, each sidecar will need to fetch the keys independently. An alternative solution is to have pilot fetch the keys and distribute the them to sidecars; sidecars can use the keys directly, without any extra communication to external services. To make this works, following assumptions/requirements need to be met: 1. Pilot is allowed to talk to external web services. (does pilot sidecar needs to be configured somehow?) 1. Implementation for cache and key rotation in pilot (i.e if key is updated, it should trigger pilot to push new configs to services). 1. Security concerns: is this acceptable if control plane doesn't use mTLS. Related issues: * [Proxy support public key in filter config](https://github.com/istio/proxy/issues/1373) * [Support OpenID Discovery](https://github.com/istio/istio/issues/4719) * https://github.com/istio/istio/issues/4633
non_usab
pilot fetches and distributes jwt public key to sidecar in envoy sidecar need to fetch public key from jwksuri specified as part of jwt filter config this requires pilot need to create outbound cluster s for jwksuri which have some conflicts with particularly concept of externalservice in term of performance each sidecar will need to fetch the keys independently an alternative solution is to have pilot fetch the keys and distribute the them to sidecars sidecars can use the keys directly without any extra communication to external services to make this works following assumptions requirements need to be met pilot is allowed to talk to external web services does pilot sidecar needs to be configured somehow implementation for cache and key rotation in pilot i e if key is updated it should trigger pilot to push new configs to services security concerns is this acceptable if control plane doesn t use mtls related issues
0
18,767
13,203,945,164
IssuesEvent
2020-08-14 15:01:51
ONRR/nrrd
https://api.github.com/repos/ONRR/nrrd
opened
When you select a new toggle on mobile it turns gray instead of blue
Home Page P3: Medium bug usability visual
On both iPhone and Android ![image.png](https://images.zenhubusercontent.com/5c589fe5e180620812ea22aa/ea546f3a-99a5-4b94-9cdf-146e8fd97ed1)
True
When you select a new toggle on mobile it turns gray instead of blue - On both iPhone and Android ![image.png](https://images.zenhubusercontent.com/5c589fe5e180620812ea22aa/ea546f3a-99a5-4b94-9cdf-146e8fd97ed1)
usab
when you select a new toggle on mobile it turns gray instead of blue on both iphone and android
1
22,108
18,684,143,369
IssuesEvent
2021-11-01 10:11:52
idiotWu/smooth-scrollbar
https://api.github.com/repos/idiotWu/smooth-scrollbar
opened
[Idea/v9] An easier way to customize the scrollbars' appearance
idea usability accepted
### Discussed in https://github.com/idiotWu/smooth-scrollbar/discussions/399 <div type='discussions-op-text'> <sup>Originally posted by **blitzve0** October 31, 2021</sup> Currently if we want to change the styling of the scroll bar thumb and tracks we have to overwrite the default styling from node modules. If basics settings like color transparency tracks color width padding etc could be set through config would be awesome</div>
True
[Idea/v9] An easier way to customize the scrollbars' appearance - ### Discussed in https://github.com/idiotWu/smooth-scrollbar/discussions/399 <div type='discussions-op-text'> <sup>Originally posted by **blitzve0** October 31, 2021</sup> Currently if we want to change the styling of the scroll bar thumb and tracks we have to overwrite the default styling from node modules. If basics settings like color transparency tracks color width padding etc could be set through config would be awesome</div>
usab
an easier way to customize the scrollbars appearance discussed in originally posted by october currently if we want to change the styling of the scroll bar thumb and tracks we have to overwrite the default styling from node modules if basics settings like color transparency tracks color width padding etc could be set through config would be awesome
1
861
2,643,838,010
IssuesEvent
2015-03-12 13:55:37
TEAMMATES/repo
https://api.github.com/repos/TEAMMATES/repo
closed
Instructor: Add a confirmation dialog to 'delete student' button
a-FaultTolarence a-Usability d.Moderate f.Courses p.Medium s.Accepted t.Enhancement
This action is irreversible. We should get a confirmation first.
True
Instructor: Add a confirmation dialog to 'delete student' button - This action is irreversible. We should get a confirmation first.
usab
instructor add a confirmation dialog to delete student button this action is irreversible we should get a confirmation first
1
315,715
9,631,264,840
IssuesEvent
2019-05-15 13:53:54
yjacques940/customerManager
https://api.github.com/repos/yjacques940/customerManager
closed
Points à corriger pour l'épreuve synthèse
High priority bug Épreuve synthèse
- [x] Erreur sur la page d'accueil dans d'autres navigateurs - [x] Variable session quand on se déconnecte - [x] Pouvoir modifier son adresse email (user) - [x] Effacer les champs quand on ajoute un rendez-vous - [x] Conflits lors des rendez-vous (peut prendre plus d'un rendez-vous en même temps) - [x] Ne pas pouvoir prendre de rendez-vous pour une date dans le passé - [ ] Avoir la traduction en anglais dans toutes les pages, titres et textes
1.0
Points à corriger pour l'épreuve synthèse - - [x] Erreur sur la page d'accueil dans d'autres navigateurs - [x] Variable session quand on se déconnecte - [x] Pouvoir modifier son adresse email (user) - [x] Effacer les champs quand on ajoute un rendez-vous - [x] Conflits lors des rendez-vous (peut prendre plus d'un rendez-vous en même temps) - [x] Ne pas pouvoir prendre de rendez-vous pour une date dans le passé - [ ] Avoir la traduction en anglais dans toutes les pages, titres et textes
non_usab
points à corriger pour l épreuve synthèse erreur sur la page d accueil dans d autres navigateurs variable session quand on se déconnecte pouvoir modifier son adresse email user effacer les champs quand on ajoute un rendez vous conflits lors des rendez vous peut prendre plus d un rendez vous en même temps ne pas pouvoir prendre de rendez vous pour une date dans le passé avoir la traduction en anglais dans toutes les pages titres et textes
0
586,064
17,568,945,639
IssuesEvent
2021-08-14 09:07:50
cpp-lln-lab/CPP_SPM
https://api.github.com/repos/cpp-lln-lab/CPP_SPM
closed
rename output from segmentation to comply to BIDS
priority 2
To comply with the BIDS specs on derivatives https://bids-specification.readthedocs.io/en/stable/05-derivatives/03-imaging.html#probabilistic-segmentations --- Try to comply with the coming BIDS extension proposal on space and mapping https://docs.google.com/document/d/11gCzXOPUbYyuQx8fErtMO9tnOKC3kTWiL9axWkkILNE/edit#heading=h.mqkmyp254xh6 For example fMRIprep outputs this: ``` sub-<subject_label>_from-T1w_to-MNI152NLin2009cAsym_mode-image_xfm.h5 ```
1.0
rename output from segmentation to comply to BIDS - To comply with the BIDS specs on derivatives https://bids-specification.readthedocs.io/en/stable/05-derivatives/03-imaging.html#probabilistic-segmentations --- Try to comply with the coming BIDS extension proposal on space and mapping https://docs.google.com/document/d/11gCzXOPUbYyuQx8fErtMO9tnOKC3kTWiL9axWkkILNE/edit#heading=h.mqkmyp254xh6 For example fMRIprep outputs this: ``` sub-<subject_label>_from-T1w_to-MNI152NLin2009cAsym_mode-image_xfm.h5 ```
non_usab
rename output from segmentation to comply to bids to comply with the bids specs on derivatives try to comply with the coming bids extension proposal on space and mapping for example fmriprep outputs this sub from to mode image xfm
0
22,301
18,981,402,250
IssuesEvent
2021-11-21 00:04:14
bevyengine/bevy
https://api.github.com/repos/bevyengine/bevy
closed
Split current `World::clear_entities` into separate methods for clearing entities and resources
A-ECS C-Usability
## What problem does this solve or what need does it fill? [World::clear_entities](https://github.com/bevyengine/bevy/blob/f4776f2ec413bb80ff4ee51e04bb57f0a44d9983/crates/bevy_ecs/src/world/mod.rs#L1139) currently wipes out all entities and resources (please correct me if my understanding of storages is horribly flawed). This is unintuitive, and not as useful for testing (or advanced ECS architectures) as it could be; where treating resources and entities separately is common. ## What solution would you like? 1. Keep the existing behaviour, but rename to `clear_all`. 2. Add `clear_resources`. 3. Add `clear_entities`.
True
Split current `World::clear_entities` into separate methods for clearing entities and resources - ## What problem does this solve or what need does it fill? [World::clear_entities](https://github.com/bevyengine/bevy/blob/f4776f2ec413bb80ff4ee51e04bb57f0a44d9983/crates/bevy_ecs/src/world/mod.rs#L1139) currently wipes out all entities and resources (please correct me if my understanding of storages is horribly flawed). This is unintuitive, and not as useful for testing (or advanced ECS architectures) as it could be; where treating resources and entities separately is common. ## What solution would you like? 1. Keep the existing behaviour, but rename to `clear_all`. 2. Add `clear_resources`. 3. Add `clear_entities`.
usab
split current world clear entities into separate methods for clearing entities and resources what problem does this solve or what need does it fill currently wipes out all entities and resources please correct me if my understanding of storages is horribly flawed this is unintuitive and not as useful for testing or advanced ecs architectures as it could be where treating resources and entities separately is common what solution would you like keep the existing behaviour but rename to clear all add clear resources add clear entities
1
5,515
3,930,325,810
IssuesEvent
2016-04-25 07:29:29
Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith
https://api.github.com/repos/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith
opened
QA_Water Content_Prerequisites_p1
Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open
Defect Description : In the "Water Content" experiment, the minimum requirement to run the experiment is not displayed in the page instead a page or Scrolling should appear providing information on minimum requirement to run this experiment, information like Bandwidth,Device Resolution,Hardware Configuration and Software Required. Actual Result : In the "Water Content" experiment, the minimum requirement to run the experiment is not displayed in the page. Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM Processor:i5 Test Step Link: https://github.com/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith/blob/master/test-cases/integration_test-cases/Water%20Content/Water%20Content_22_Prerequisites_p1.org
True
QA_Water Content_Prerequisites_p1 - Defect Description : In the "Water Content" experiment, the minimum requirement to run the experiment is not displayed in the page instead a page or Scrolling should appear providing information on minimum requirement to run this experiment, information like Bandwidth,Device Resolution,Hardware Configuration and Software Required. Actual Result : In the "Water Content" experiment, the minimum requirement to run the experiment is not displayed in the page. Environment : OS: Windows 7, Ubuntu-16.04,Centos-6 Browsers: Firefox-42.0,Chrome-47.0,chromium-45.0 Bandwidth : 100Mbps Hardware Configuration:8GBRAM Processor:i5 Test Step Link: https://github.com/Virtual-Labs/soil-mechanics-and-foundation-engineering-iiith/blob/master/test-cases/integration_test-cases/Water%20Content/Water%20Content_22_Prerequisites_p1.org
usab
qa water content prerequisites defect description in the water content experiment the minimum requirement to run the experiment is not displayed in the page instead a page or scrolling should appear providing information on minimum requirement to run this experiment information like bandwidth device resolution hardware configuration and software required actual result in the water content experiment the minimum requirement to run the experiment is not displayed in the page environment os windows ubuntu centos browsers firefox chrome chromium bandwidth hardware configuration processor test step link
1
4,447
3,869,680,906
IssuesEvent
2016-04-10 19:01:20
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
23310964: Step between font sizes is too large
classification:ui/usability reproducible:always status:open
#### Description I’ve noticed this many times over the past few years, so I don’t think it’s anything new. My gripe is that there is often no font-size setting in iBooks that is neither too big nor too small. It’s not that I want the text to be at some specific size; I just want the increments between sizes to be finer so that I don’t end up with no options that seem acceptable. I’m attaching a representative page at two adjacent sizes. They look more clearly distinguished on the iPad screen than on a desktop. I’d call the one with the smaller font size readable, though perhaps a bit uncomfortably small with the backlight turned low. The next size up looks like a large-print edition for geriatrics. There’s simply nothing in between, at least with this font (Iowan). - Product Version: iOS 9.1 Created: 2015-10-29T07:48:29.279410 Originated: 2015-10-29T00:48:00 Open Radar Link: http://www.openradar.me/23310964
True
23310964: Step between font sizes is too large - #### Description I’ve noticed this many times over the past few years, so I don’t think it’s anything new. My gripe is that there is often no font-size setting in iBooks that is neither too big nor too small. It’s not that I want the text to be at some specific size; I just want the increments between sizes to be finer so that I don’t end up with no options that seem acceptable. I’m attaching a representative page at two adjacent sizes. They look more clearly distinguished on the iPad screen than on a desktop. I’d call the one with the smaller font size readable, though perhaps a bit uncomfortably small with the backlight turned low. The next size up looks like a large-print edition for geriatrics. There’s simply nothing in between, at least with this font (Iowan). - Product Version: iOS 9.1 Created: 2015-10-29T07:48:29.279410 Originated: 2015-10-29T00:48:00 Open Radar Link: http://www.openradar.me/23310964
usab
step between font sizes is too large description i’ve noticed this many times over the past few years so i don’t think it’s anything new my gripe is that there is often no font size setting in ibooks that is neither too big nor too small it’s not that i want the text to be at some specific size i just want the increments between sizes to be finer so that i don’t end up with no options that seem acceptable i’m attaching a representative page at two adjacent sizes they look more clearly distinguished on the ipad screen than on a desktop i’d call the one with the smaller font size readable though perhaps a bit uncomfortably small with the backlight turned low the next size up looks like a large print edition for geriatrics there’s simply nothing in between at least with this font iowan product version ios created originated open radar link
1
902
2,769,975,631
IssuesEvent
2015-05-01 09:19:43
piwik/piwik
https://api.github.com/repos/piwik/piwik
closed
UNION query in findVisitor() is regularly 100ms+, involves temporary tables and scans
c: Performance Task
As noted in #7440, commit 0abac3b312ff9560fb43012192e37c2369bab3ed introduced a UNION into findVisitor() within the visit code (now in core/Tracker/Model.php), which [results in the creation of a temporary table on each call](http://dev.mysql.com/doc/refman/5.5/en/internal-temporary-tables.html) ("UNION queries use temporary tables."). Separately, the idvisitor query normally takes 1-2ms and the config_id query takes around 0.5ms. The combined query is normally around 5ms. However, the number of rows examined may reach the hundreds, and it then starts taking significantly longer - up to 200ms on a dedicated machine. The impact of these queries vary depending on their size and may not be visible on test or low-visit production websites. This is what happened when I switched it to checking just the idvisitor part (see last hour): ![mysql_tmp_tables-day-idvisitor](https://cloud.githubusercontent.com/assets/383043/6681708/2c675f9c-cc63-11e4-874c-e268ffb777c7.png) [The other temporary tables which are disk tables [appeared more recently](https://github.com/piwik/piwik/issues/7416#issuecomment-81170546)] ![mysql_select_types-day](https://cloud.githubusercontent.com/assets/383043/6681820/41a3274a-cc65-11e4-826a-a92e3bdde9fd.png) When I switched the code around to just use just the select on config_id, the decrease in activity remained. This suggests to me that it is the UNION of the two selects that is causing objects and scans. An example of a slow query from the log (there are several 100ms+ ones per minute): ```` # Query_time: 0.203814 Lock_time: 0.048132 Rows_sent: 1 Rows_examined: 116 SET timestamp=1426562315; ( SELECT visit_last_action_time, visit_first_action_time, idvisitor, idvisit, user_id, visit_exit_idaction_url, visit_exit_idaction_name, visitor_returning, visitor_days_since_first, visitor_days_since_order, visitor_count_visits, visit_goal_buyer, location_country, location_region, location_city, location_latitude, location_longitude, referer_name, referer_keyword, referer_type, location_ip, location_browser_lang, visit_total_events, visit_total_time, visit_total_actions, visit_total_searches , 0 as priority FROM piwik_log_visit WHERE visit_last_action_time >= '2015-03-17 02:48:34' AND visit_last_action_time <= '2015-03-17 03:48:34' AND idsite = '1' AND config_id = '^Z?^\? D?' AND user_id IS NULL ORDER BY visit_last_action_time DESC LIMIT 1 ) UNION ( SELECT visit_last_action_time, visit_first_action_time, idvisitor, idvisit, user_id, visit_exit_idaction_url, visit_exit_idaction_name, visitor_returning, visitor_days_since_first, visitor_days_since_order, visitor_count_visits, visit_goal_buyer, location_country, location_region, location_city, location_latitude, location_longitude, referer_name, referer_keyword, referer_type, location_ip, location_browser_lang, visit_total_events, visit_total_time, visit_total_actions, visit_total_searches , 1 as priority FROM piwik_log_visit WHERE visit_last_action_time >= '2015-03-17 02:48:34' AND visit_last_action_time <= '2015-03-17 03:48:34' AND idsite = '1' AND idvisitor = '^U\"?^^0^M?U' ORDER BY visit_last_action_time DESC LIMIT 1 ) ORDER BY priority DESC LIMIT 1; ```` An EXPLAIN of the query (the last entry involves the temporary table created by the UNION): ```` | id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra | +----+--------------+-----------------+-------+---------------------------------------------------------------------------+------------------------------+---------+-------------+------+-----------------------------+ | 1 | PRIMARY | piwik_log_visit | range | index_idsite_config_datetime,index_idsite_datetime,index_idsite_idvisitor | index_idsite_config_datetime | 20 | NULL | 1 | Using where | | 2 | UNION | piwik_log_visit | ref | index_idsite_config_datetime,index_idsite_datetime,index_idsite_idvisitor | index_idsite_idvisitor | 12 | const,const | 1 | Using where; Using filesort | | NULL | UNION RESULT | <union1,2> | ALL | NULL | NULL | NULL | NULL | NULL | Using filesort | ```` The UNION is intended to check visits against both idvisitor and config_id and return the first if it is present, other than the second. I propose that two queries be performed - check idvisitor first, and if that does not return a result, do the config_id query.
True
UNION query in findVisitor() is regularly 100ms+, involves temporary tables and scans - As noted in #7440, commit 0abac3b312ff9560fb43012192e37c2369bab3ed introduced a UNION into findVisitor() within the visit code (now in core/Tracker/Model.php), which [results in the creation of a temporary table on each call](http://dev.mysql.com/doc/refman/5.5/en/internal-temporary-tables.html) ("UNION queries use temporary tables."). Separately, the idvisitor query normally takes 1-2ms and the config_id query takes around 0.5ms. The combined query is normally around 5ms. However, the number of rows examined may reach the hundreds, and it then starts taking significantly longer - up to 200ms on a dedicated machine. The impact of these queries vary depending on their size and may not be visible on test or low-visit production websites. This is what happened when I switched it to checking just the idvisitor part (see last hour): ![mysql_tmp_tables-day-idvisitor](https://cloud.githubusercontent.com/assets/383043/6681708/2c675f9c-cc63-11e4-874c-e268ffb777c7.png) [The other temporary tables which are disk tables [appeared more recently](https://github.com/piwik/piwik/issues/7416#issuecomment-81170546)] ![mysql_select_types-day](https://cloud.githubusercontent.com/assets/383043/6681820/41a3274a-cc65-11e4-826a-a92e3bdde9fd.png) When I switched the code around to just use just the select on config_id, the decrease in activity remained. This suggests to me that it is the UNION of the two selects that is causing objects and scans. An example of a slow query from the log (there are several 100ms+ ones per minute): ```` # Query_time: 0.203814 Lock_time: 0.048132 Rows_sent: 1 Rows_examined: 116 SET timestamp=1426562315; ( SELECT visit_last_action_time, visit_first_action_time, idvisitor, idvisit, user_id, visit_exit_idaction_url, visit_exit_idaction_name, visitor_returning, visitor_days_since_first, visitor_days_since_order, visitor_count_visits, visit_goal_buyer, location_country, location_region, location_city, location_latitude, location_longitude, referer_name, referer_keyword, referer_type, location_ip, location_browser_lang, visit_total_events, visit_total_time, visit_total_actions, visit_total_searches , 0 as priority FROM piwik_log_visit WHERE visit_last_action_time >= '2015-03-17 02:48:34' AND visit_last_action_time <= '2015-03-17 03:48:34' AND idsite = '1' AND config_id = '^Z?^\? D?' AND user_id IS NULL ORDER BY visit_last_action_time DESC LIMIT 1 ) UNION ( SELECT visit_last_action_time, visit_first_action_time, idvisitor, idvisit, user_id, visit_exit_idaction_url, visit_exit_idaction_name, visitor_returning, visitor_days_since_first, visitor_days_since_order, visitor_count_visits, visit_goal_buyer, location_country, location_region, location_city, location_latitude, location_longitude, referer_name, referer_keyword, referer_type, location_ip, location_browser_lang, visit_total_events, visit_total_time, visit_total_actions, visit_total_searches , 1 as priority FROM piwik_log_visit WHERE visit_last_action_time >= '2015-03-17 02:48:34' AND visit_last_action_time <= '2015-03-17 03:48:34' AND idsite = '1' AND idvisitor = '^U\"?^^0^M?U' ORDER BY visit_last_action_time DESC LIMIT 1 ) ORDER BY priority DESC LIMIT 1; ```` An EXPLAIN of the query (the last entry involves the temporary table created by the UNION): ```` | id | select_type | table | type | possible_keys | key | key_len | ref | rows | Extra | +----+--------------+-----------------+-------+---------------------------------------------------------------------------+------------------------------+---------+-------------+------+-----------------------------+ | 1 | PRIMARY | piwik_log_visit | range | index_idsite_config_datetime,index_idsite_datetime,index_idsite_idvisitor | index_idsite_config_datetime | 20 | NULL | 1 | Using where | | 2 | UNION | piwik_log_visit | ref | index_idsite_config_datetime,index_idsite_datetime,index_idsite_idvisitor | index_idsite_idvisitor | 12 | const,const | 1 | Using where; Using filesort | | NULL | UNION RESULT | <union1,2> | ALL | NULL | NULL | NULL | NULL | NULL | Using filesort | ```` The UNION is intended to check visits against both idvisitor and config_id and return the first if it is present, other than the second. I propose that two queries be performed - check idvisitor first, and if that does not return a result, do the config_id query.
non_usab
union query in findvisitor is regularly involves temporary tables and scans as noted in commit introduced a union into findvisitor within the visit code now in core tracker model php which union queries use temporary tables separately the idvisitor query normally takes and the config id query takes around the combined query is normally around however the number of rows examined may reach the hundreds and it then starts taking significantly longer up to on a dedicated machine the impact of these queries vary depending on their size and may not be visible on test or low visit production websites this is what happened when i switched it to checking just the idvisitor part see last hour when i switched the code around to just use just the select on config id the decrease in activity remained this suggests to me that it is the union of the two selects that is causing objects and scans an example of a slow query from the log there are several ones per minute query time lock time rows sent rows examined set timestamp select visit last action time visit first action time idvisitor idvisit user id visit exit idaction url visit exit idaction name visitor returning visitor days since first visitor days since order visitor count visits visit goal buyer location country location region location city location latitude location longitude referer name referer keyword referer type location ip location browser lang visit total events visit total time visit total actions visit total searches as priority from piwik log visit where visit last action time and visit last action time and idsite and config id z d and user id is null order by visit last action time desc limit union select visit last action time visit first action time idvisitor idvisit user id visit exit idaction url visit exit idaction name visitor returning visitor days since first visitor days since order visitor count visits visit goal buyer location country location region location city location latitude location longitude referer name referer keyword referer type location ip location browser lang visit total events visit total time visit total actions visit total searches as priority from piwik log visit where visit last action time and visit last action time and idsite and idvisitor u m u order by visit last action time desc limit order by priority desc limit an explain of the query the last entry involves the temporary table created by the union id select type table type possible keys key key len ref rows extra primary piwik log visit range index idsite config datetime index idsite datetime index idsite idvisitor index idsite config datetime null using where union piwik log visit ref index idsite config datetime index idsite datetime index idsite idvisitor index idsite idvisitor const const using where using filesort null union result all null null null null null using filesort the union is intended to check visits against both idvisitor and config id and return the first if it is present other than the second i propose that two queries be performed check idvisitor first and if that does not return a result do the config id query
0
507,411
14,679,980,206
IssuesEvent
2020-12-31 08:40:54
k8smeetup/website-tasks
https://api.github.com/repos/k8smeetup/website-tasks
opened
/docs/setup/best-practices/_index.md
lang/zh priority/P0 sync/update version/master welcome
Source File: [/docs/setup/best-practices/_index.md](https://github.com/kubernetes/website/blob/master/content/en/docs/setup/best-practices/_index.md) Diff 命令参考: ```bash # 查看原始文档与翻译文档更新差异 git diff --no-index -- content/en/docs/setup/best-practices/_index.md content/zh/docs/setup/best-practices/_index.md # 跨分支持查看原始文档更新差异 git diff release-1.19 master -- content/en/docs/setup/best-practices/_index.md ```
1.0
/docs/setup/best-practices/_index.md - Source File: [/docs/setup/best-practices/_index.md](https://github.com/kubernetes/website/blob/master/content/en/docs/setup/best-practices/_index.md) Diff 命令参考: ```bash # 查看原始文档与翻译文档更新差异 git diff --no-index -- content/en/docs/setup/best-practices/_index.md content/zh/docs/setup/best-practices/_index.md # 跨分支持查看原始文档更新差异 git diff release-1.19 master -- content/en/docs/setup/best-practices/_index.md ```
non_usab
docs setup best practices index md source file diff 命令参考 bash 查看原始文档与翻译文档更新差异 git diff no index content en docs setup best practices index md content zh docs setup best practices index md 跨分支持查看原始文档更新差异 git diff release master content en docs setup best practices index md
0
115,277
14,707,978,789
IssuesEvent
2021-01-04 22:38:09
AlaskaAirlines/auro-table
https://api.github.com/repos/AlaskaAirlines/auro-table
closed
Table: symbolize
Project: Version Zero Type: Design Type: Feature help wanted
# User story As a designer, I want to be able to reuse the correct table symbol, so I can quickly repeat UIs for my team. ## Outline subtasks or tasks > List specific steps required to deliver this work. - [ ] Document the currently developed v1 of the auro-table (see attached image) - [ ] Add to MASTER branch of UI Kit project - [ ] Communicate out to designers of update ## Definition of “Done” (exit criteria) > The story is generally “done” when the user can complete the outlined task, but make sure to define what that is. This story will be considered complete once the developed asset has been properly documented as a design asset and communicated out to the team. ## Epic https://github.com/AlaskaAirlines/auro_docs/issues/98
1.0
Table: symbolize - # User story As a designer, I want to be able to reuse the correct table symbol, so I can quickly repeat UIs for my team. ## Outline subtasks or tasks > List specific steps required to deliver this work. - [ ] Document the currently developed v1 of the auro-table (see attached image) - [ ] Add to MASTER branch of UI Kit project - [ ] Communicate out to designers of update ## Definition of “Done” (exit criteria) > The story is generally “done” when the user can complete the outlined task, but make sure to define what that is. This story will be considered complete once the developed asset has been properly documented as a design asset and communicated out to the team. ## Epic https://github.com/AlaskaAirlines/auro_docs/issues/98
non_usab
table symbolize user story as a designer i want to be able to reuse the correct table symbol so i can quickly repeat uis for my team outline subtasks or tasks list specific steps required to deliver this work document the currently developed of the auro table see attached image add to master branch of ui kit project communicate out to designers of update definition of “done” exit criteria the story is generally “done” when the user can complete the outlined task but make sure to define what that is this story will be considered complete once the developed asset has been properly documented as a design asset and communicated out to the team epic
0
19,181
13,635,223,381
IssuesEvent
2020-09-25 02:15:19
rubyforgood/circulate
https://api.github.com/repos/rubyforgood/circulate
closed
Public nav should use off-canvas on mobile
:zap: Power :zap: Help Wanted Ruby For Good 🎃 Fall 2020 UX / Usability bug
At a small width, the public navigation breaks pretty badly: <img width="501" alt="Screen Shot 2020-09-17 at 2 21 38 PM" src="https://user-images.githubusercontent.com/3331/93517641-5f6ce400-f8f1-11ea-9a5e-d14108972a3f.png"> We should be able to use the [off-canvas menu provided by the Spectre](https://picturepan2.github.io/spectre/experimentals/off-canvas.html) to solve this. We're already using that component in the admin UI. See also #199 .
True
Public nav should use off-canvas on mobile - At a small width, the public navigation breaks pretty badly: <img width="501" alt="Screen Shot 2020-09-17 at 2 21 38 PM" src="https://user-images.githubusercontent.com/3331/93517641-5f6ce400-f8f1-11ea-9a5e-d14108972a3f.png"> We should be able to use the [off-canvas menu provided by the Spectre](https://picturepan2.github.io/spectre/experimentals/off-canvas.html) to solve this. We're already using that component in the admin UI. See also #199 .
usab
public nav should use off canvas on mobile at a small width the public navigation breaks pretty badly img width alt screen shot at pm src we should be able to use the to solve this we re already using that component in the admin ui see also
1
408,234
11,943,660,294
IssuesEvent
2020-04-02 23:57:28
wc-work/why
https://api.github.com/repos/wc-work/why
closed
Contact Form
High Priority
- [x] Add Contact Form - [x] Set up form to automatically send W.H.Y. email notification
1.0
Contact Form - - [x] Add Contact Form - [x] Set up form to automatically send W.H.Y. email notification
non_usab
contact form add contact form set up form to automatically send w h y email notification
0
125,802
4,964,835,814
IssuesEvent
2016-12-04 00:08:13
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Pods with PVCs stuck in ContainerCreating after Node upgrade
area/storage area/upgrade priority/P0 release-blocker team/cluster
<!-- Thanks for filing an issue! Before hitting the button, please answer these questions.--> **Is this a request for help?** (If yes, you should use our troubleshooting guide and community support channels, see http://kubernetes.io/docs/troubleshooting/.): No. **What keywords did you search in Kubernetes issues before filing this one?** (If you have found any duplicates, you should instead reply there.): * upgrade volume * mount timeout * ContainerCreating --- **Is this a BUG REPORT or FEATURE REQUEST?** (choose one): BUG REPORT <!-- If this is a BUG REPORT, please: - Fill in as much of the template below as you can. If you leave out information, we can't help you as well. If this is a FEATURE REQUEST, please: - Describe *in detail* the feature/behavior/change you'd like to see. In both cases, be ready for followup questions, and please respond in a timely manner. If we can't reproduce a bug or think a feature already exists, we might close your issue. If we're wrong, PLEASE feel free to reopen it and explain why. --> **Kubernetes version** (use `kubectl version`): ``` Server Version: version.Info{Major:"1", Minor:"5+", GitVersion:"v1.5.0-beta.2", GitCommit:"0776eab45fe28f02bbeac0f05ae1a203051a21eb", GitTreeState:"clean", BuildDate:"2016-11-24T22:30:23Z", GoVersion:"go1.7.3", Compiler:"gc", Platform:"linux/amd64"} ``` **Environment**: - **Cloud provider or hardware configuration**: GCE - **OS** (e.g. from /etc/os-release): - master: ``` BUILD_ID=8977.0.0 NAME="Container-VM Image" GOOGLE_CRASH_ID=Lakitu VERSION_ID=56 BUG_REPORT_URL=https://crbug.com/new PRETTY_NAME="Google Container-VM Image" VERSION=56 GOOGLE_METRICS_PRODUCT_ID=26 HOME_URL="https://cloud.google.com/compute/docs/containers/vm-image/" ID=gci ``` - minion: ``` PRETTY_NAME="Debian GNU/Linux 7 (wheezy)" NAME="Debian GNU/Linux" VERSION_ID="7" VERSION="7 (wheezy)" ID=debian ANSI_COLOR="1;31" HOME_URL="http://www.debian.org/" SUPPORT_URL="http://www.debian.org/support/" BUG_REPORT_URL="http://bugs.debian.org/" ``` - **Kernel** (e.g. `uname -a`): - master: `Linux kubernetes-master 4.4.21+ #1 SMP Thu Nov 10 02:50:15 PST 2016 x86_64 Intel(R) Xeon(R) CPU @ 2.30GHz GenuineIntel GNU/Linux` - minion: `Linux kubernetes-minion-group-4klk 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u2 (2016-10-19) x86_64 GNU/Linux` - **Install tools**: `cluster/gce/upgrade.sh` - **Others**: **What happened**: Pods with PVCs get stuck in ContainerCreating after Node gets upgraded. The error is `timeout expired waiting for volumes to attach/mount` as shown below: ```console $ kubectl describe pod web-1 Name: web-1 Namespace: default Node: kubernetes-minion-group-54hw/10.128.0.4 Start Time: Mon, 28 Nov 2016 11:52:11 -0800 Labels: app=nginx Status: Pending IP: Controllers: PetSet/web Containers: nginx: Container ID: Image: gcr.io/google_containers/nginx-slim:0.8 Image ID: Port: 80/TCP Requests: cpu: 100m State: Waiting Reason: ContainerCreating Ready: False Restart Count: 0 Volume Mounts: /usr/share/nginx/html from www (rw) /var/run/secrets/kubernetes.io/serviceaccount from default-token-xim26 (ro) Environment Variables: <none> Conditions: Type Status Initialized True Ready False PodScheduled True Volumes: www: Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace) ClaimName: www-web-1 ReadOnly: false default-token-xim26: Type: Secret (a volume populated by a Secret) SecretName: default-token-xim26 QoS Class: Burstable Tolerations: <none> Events: FirstSeen LastSeen Count From SubObjectPath Type Reason Message --------- -------- ----- ---- ------------- -------- ------ ------- 40m 40m 1 {default-scheduler } Normal Scheduled Successfully assigned web-1 to kubernetes-minion-group-54hw 40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Pulling pulling image "gcr.io/google_containers/nginx-slim:0.8" 40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Pulled Successfully pulled image "gcr.io/google_containers/nginx-slim:0.8" 40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Created Created container with docker id d2d15140a74a; Security:[seccomp=unconfined] 40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Started Started container with docker id d2d15140a74a 16m 16m 4 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp: lookup kubernetes-master on 169.254.169.254:53: no such host 16m 16m 1 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp 10.128.0.2:443: i/o timeout 15m 15m 3 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp 10.128.0.2:443: getsockopt: connection refused 9m 9m 1 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: network is not ready: [Kubenet does not have netConfig. This is most likely due to lack of PodCIDR] 9m 9m 3 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: network is not ready: [Kubenet does not have netConfig. This is most likely due to lack of PodCIDR] 6m 14s 4 {kubelet kubernetes-minion-group-54hw} Warning FailedMount Unable to mount volumes for pod "web-1_default(270e4f19-b5a4-11e6-8922-42010a800002)": timeout expired waiting for volumes to attach/mount for pod "web-1"/"default". list of unattached/unmounted volumes=[www] 6m 14s 4 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: timeout expired waiting for volumes to attach/mount for pod "web-1"/"default". list of unattached/unmounted volumes=[www] ``` **What you expected to happen**: Upon Node upgrade, Pods should either be terminated (so a controller can recreate them) or they should come back to Running on their own. **How to reproduce it** (as minimally and precisely as possible): 1. Start a v1.4 cluster on GCE. ```shell git checkout release-1.4 git pull upstream release-1.4 make clean quick-release cluster/kube-up.sh ``` 1. Create the following example PetSet and wait for all Pods to be Running. ```shell kubectl create -f petset.yaml ``` ```yaml # A headless service to create DNS records apiVersion: v1 kind: Service metadata: name: nginx labels: app: nginx spec: ports: - port: 80 name: web # *.nginx.default.svc.cluster.local clusterIP: None selector: app: nginx --- apiVersion: apps/v1alpha1 kind: PetSet metadata: name: web spec: serviceName: "nginx" replicas: 5 template: metadata: labels: app: nginx annotations: pod.alpha.kubernetes.io/initialized: "true" spec: containers: - name: nginx image: gcr.io/google_containers/nginx-slim:0.8 ports: - containerPort: 80 name: web volumeMounts: - name: www mountPath: /usr/share/nginx/html volumeClaimTemplates: - metadata: name: www annotations: volume.alpha.kubernetes.io/storage-class: anything spec: accessModes: [ "ReadWriteOnce" ] resources: requests: storage: 1Gi ``` 1. Delete PetSet without cascading (keep the Pods). ```shell kubectl delete petset web --cascade=false ``` 1. Upgrade master and nodes to `v1.5.0-beta.2` with `cluster/gce/upgrade.sh`. ```shell git checkout release-1.5 git pull upstream release-1.5 cluster/gce/upgrade.sh v1.5.0-beta.2 ``` 1. Check for Pods stuck in ContainerCreating. ``` NAME READY STATUS RESTARTS AGE web-0 0/1 ContainerCreating 0 42m web-4 0/1 ContainerCreating 0 42m ``` **Anything else do we need to know**: After the above steps, some Pods may end up terminating. It seems to be random which ones terminate and which ones get stuck in ContainerCreating. Manually deleting and recreating all the stuck Pods after the Nodes are upgraded allows them to correctly attach to the preexisting PVCs and become Running.
1.0
Pods with PVCs stuck in ContainerCreating after Node upgrade - <!-- Thanks for filing an issue! Before hitting the button, please answer these questions.--> **Is this a request for help?** (If yes, you should use our troubleshooting guide and community support channels, see http://kubernetes.io/docs/troubleshooting/.): No. **What keywords did you search in Kubernetes issues before filing this one?** (If you have found any duplicates, you should instead reply there.): * upgrade volume * mount timeout * ContainerCreating --- **Is this a BUG REPORT or FEATURE REQUEST?** (choose one): BUG REPORT <!-- If this is a BUG REPORT, please: - Fill in as much of the template below as you can. If you leave out information, we can't help you as well. If this is a FEATURE REQUEST, please: - Describe *in detail* the feature/behavior/change you'd like to see. In both cases, be ready for followup questions, and please respond in a timely manner. If we can't reproduce a bug or think a feature already exists, we might close your issue. If we're wrong, PLEASE feel free to reopen it and explain why. --> **Kubernetes version** (use `kubectl version`): ``` Server Version: version.Info{Major:"1", Minor:"5+", GitVersion:"v1.5.0-beta.2", GitCommit:"0776eab45fe28f02bbeac0f05ae1a203051a21eb", GitTreeState:"clean", BuildDate:"2016-11-24T22:30:23Z", GoVersion:"go1.7.3", Compiler:"gc", Platform:"linux/amd64"} ``` **Environment**: - **Cloud provider or hardware configuration**: GCE - **OS** (e.g. from /etc/os-release): - master: ``` BUILD_ID=8977.0.0 NAME="Container-VM Image" GOOGLE_CRASH_ID=Lakitu VERSION_ID=56 BUG_REPORT_URL=https://crbug.com/new PRETTY_NAME="Google Container-VM Image" VERSION=56 GOOGLE_METRICS_PRODUCT_ID=26 HOME_URL="https://cloud.google.com/compute/docs/containers/vm-image/" ID=gci ``` - minion: ``` PRETTY_NAME="Debian GNU/Linux 7 (wheezy)" NAME="Debian GNU/Linux" VERSION_ID="7" VERSION="7 (wheezy)" ID=debian ANSI_COLOR="1;31" HOME_URL="http://www.debian.org/" SUPPORT_URL="http://www.debian.org/support/" BUG_REPORT_URL="http://bugs.debian.org/" ``` - **Kernel** (e.g. `uname -a`): - master: `Linux kubernetes-master 4.4.21+ #1 SMP Thu Nov 10 02:50:15 PST 2016 x86_64 Intel(R) Xeon(R) CPU @ 2.30GHz GenuineIntel GNU/Linux` - minion: `Linux kubernetes-minion-group-4klk 3.16.0-4-amd64 #1 SMP Debian 3.16.36-1+deb8u2 (2016-10-19) x86_64 GNU/Linux` - **Install tools**: `cluster/gce/upgrade.sh` - **Others**: **What happened**: Pods with PVCs get stuck in ContainerCreating after Node gets upgraded. The error is `timeout expired waiting for volumes to attach/mount` as shown below: ```console $ kubectl describe pod web-1 Name: web-1 Namespace: default Node: kubernetes-minion-group-54hw/10.128.0.4 Start Time: Mon, 28 Nov 2016 11:52:11 -0800 Labels: app=nginx Status: Pending IP: Controllers: PetSet/web Containers: nginx: Container ID: Image: gcr.io/google_containers/nginx-slim:0.8 Image ID: Port: 80/TCP Requests: cpu: 100m State: Waiting Reason: ContainerCreating Ready: False Restart Count: 0 Volume Mounts: /usr/share/nginx/html from www (rw) /var/run/secrets/kubernetes.io/serviceaccount from default-token-xim26 (ro) Environment Variables: <none> Conditions: Type Status Initialized True Ready False PodScheduled True Volumes: www: Type: PersistentVolumeClaim (a reference to a PersistentVolumeClaim in the same namespace) ClaimName: www-web-1 ReadOnly: false default-token-xim26: Type: Secret (a volume populated by a Secret) SecretName: default-token-xim26 QoS Class: Burstable Tolerations: <none> Events: FirstSeen LastSeen Count From SubObjectPath Type Reason Message --------- -------- ----- ---- ------------- -------- ------ ------- 40m 40m 1 {default-scheduler } Normal Scheduled Successfully assigned web-1 to kubernetes-minion-group-54hw 40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Pulling pulling image "gcr.io/google_containers/nginx-slim:0.8" 40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Pulled Successfully pulled image "gcr.io/google_containers/nginx-slim:0.8" 40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Created Created container with docker id d2d15140a74a; Security:[seccomp=unconfined] 40m 40m 1 {kubelet kubernetes-minion-group-54hw} spec.containers{nginx} Normal Started Started container with docker id d2d15140a74a 16m 16m 4 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp: lookup kubernetes-master on 169.254.169.254:53: no such host 16m 16m 1 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp 10.128.0.2:443: i/o timeout 15m 15m 3 {kubelet kubernetes-minion-group-54hw} Warning FailedMount MountVolume.SetUp failed for volume "kubernetes.io/secret/270e4f19-b5a4-11e6-8922-42010a800002-default-token-xim26" (spec.Name: "default-token-xim26") pod "270e4f19-b5a4-11e6-8922-42010a800002" (UID: "270e4f19-b5a4-11e6-8922-42010a800002") with: Get https://kubernetes-master/api/v1/namespaces/default/secrets/default-token-xim26: dial tcp 10.128.0.2:443: getsockopt: connection refused 9m 9m 1 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: network is not ready: [Kubenet does not have netConfig. This is most likely due to lack of PodCIDR] 9m 9m 3 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: network is not ready: [Kubenet does not have netConfig. This is most likely due to lack of PodCIDR] 6m 14s 4 {kubelet kubernetes-minion-group-54hw} Warning FailedMount Unable to mount volumes for pod "web-1_default(270e4f19-b5a4-11e6-8922-42010a800002)": timeout expired waiting for volumes to attach/mount for pod "web-1"/"default". list of unattached/unmounted volumes=[www] 6m 14s 4 {kubelet kubernetes-minion-group-54hw} Warning FailedSync Error syncing pod, skipping: timeout expired waiting for volumes to attach/mount for pod "web-1"/"default". list of unattached/unmounted volumes=[www] ``` **What you expected to happen**: Upon Node upgrade, Pods should either be terminated (so a controller can recreate them) or they should come back to Running on their own. **How to reproduce it** (as minimally and precisely as possible): 1. Start a v1.4 cluster on GCE. ```shell git checkout release-1.4 git pull upstream release-1.4 make clean quick-release cluster/kube-up.sh ``` 1. Create the following example PetSet and wait for all Pods to be Running. ```shell kubectl create -f petset.yaml ``` ```yaml # A headless service to create DNS records apiVersion: v1 kind: Service metadata: name: nginx labels: app: nginx spec: ports: - port: 80 name: web # *.nginx.default.svc.cluster.local clusterIP: None selector: app: nginx --- apiVersion: apps/v1alpha1 kind: PetSet metadata: name: web spec: serviceName: "nginx" replicas: 5 template: metadata: labels: app: nginx annotations: pod.alpha.kubernetes.io/initialized: "true" spec: containers: - name: nginx image: gcr.io/google_containers/nginx-slim:0.8 ports: - containerPort: 80 name: web volumeMounts: - name: www mountPath: /usr/share/nginx/html volumeClaimTemplates: - metadata: name: www annotations: volume.alpha.kubernetes.io/storage-class: anything spec: accessModes: [ "ReadWriteOnce" ] resources: requests: storage: 1Gi ``` 1. Delete PetSet without cascading (keep the Pods). ```shell kubectl delete petset web --cascade=false ``` 1. Upgrade master and nodes to `v1.5.0-beta.2` with `cluster/gce/upgrade.sh`. ```shell git checkout release-1.5 git pull upstream release-1.5 cluster/gce/upgrade.sh v1.5.0-beta.2 ``` 1. Check for Pods stuck in ContainerCreating. ``` NAME READY STATUS RESTARTS AGE web-0 0/1 ContainerCreating 0 42m web-4 0/1 ContainerCreating 0 42m ``` **Anything else do we need to know**: After the above steps, some Pods may end up terminating. It seems to be random which ones terminate and which ones get stuck in ContainerCreating. Manually deleting and recreating all the stuck Pods after the Nodes are upgraded allows them to correctly attach to the preexisting PVCs and become Running.
non_usab
pods with pvcs stuck in containercreating after node upgrade is this a request for help if yes you should use our troubleshooting guide and community support channels see no what keywords did you search in kubernetes issues before filing this one if you have found any duplicates you should instead reply there upgrade volume mount timeout containercreating is this a bug report or feature request choose one bug report if this is a bug report please fill in as much of the template below as you can if you leave out information we can t help you as well if this is a feature request please describe in detail the feature behavior change you d like to see in both cases be ready for followup questions and please respond in a timely manner if we can t reproduce a bug or think a feature already exists we might close your issue if we re wrong please feel free to reopen it and explain why kubernetes version use kubectl version server version version info major minor gitversion beta gitcommit gittreestate clean builddate goversion compiler gc platform linux environment cloud provider or hardware configuration gce os e g from etc os release master build id name container vm image google crash id lakitu version id bug report url pretty name google container vm image version google metrics product id home url id gci minion pretty name debian gnu linux wheezy name debian gnu linux version id version wheezy id debian ansi color home url support url bug report url kernel e g uname a master linux kubernetes master smp thu nov pst intel r xeon r cpu genuineintel gnu linux minion linux kubernetes minion group smp debian gnu linux install tools cluster gce upgrade sh others what happened pods with pvcs get stuck in containercreating after node gets upgraded the error is timeout expired waiting for volumes to attach mount as shown below console kubectl describe pod web name web namespace default node kubernetes minion group start time mon nov labels app nginx status pending ip controllers petset web containers nginx container id image gcr io google containers nginx slim image id port tcp requests cpu state waiting reason containercreating ready false restart count volume mounts usr share nginx html from www rw var run secrets kubernetes io serviceaccount from default token ro environment variables conditions type status initialized true ready false podscheduled true volumes www type persistentvolumeclaim a reference to a persistentvolumeclaim in the same namespace claimname www web readonly false default token type secret a volume populated by a secret secretname default token qos class burstable tolerations events firstseen lastseen count from subobjectpath type reason message default scheduler normal scheduled successfully assigned web to kubernetes minion group kubelet kubernetes minion group spec containers nginx normal pulling pulling image gcr io google containers nginx slim kubelet kubernetes minion group spec containers nginx normal pulled successfully pulled image gcr io google containers nginx slim kubelet kubernetes minion group spec containers nginx normal created created container with docker id security kubelet kubernetes minion group spec containers nginx normal started started container with docker id kubelet kubernetes minion group warning failedmount mountvolume setup failed for volume kubernetes io secret default token spec name default token pod uid with get dial tcp lookup kubernetes master on no such host kubelet kubernetes minion group warning failedmount mountvolume setup failed for volume kubernetes io secret default token spec name default token pod uid with get dial tcp i o timeout kubelet kubernetes minion group warning failedmount mountvolume setup failed for volume kubernetes io secret default token spec name default token pod uid with get dial tcp getsockopt connection refused kubelet kubernetes minion group warning failedsync error syncing pod skipping network is not ready kubelet kubernetes minion group warning failedsync error syncing pod skipping network is not ready kubelet kubernetes minion group warning failedmount unable to mount volumes for pod web default timeout expired waiting for volumes to attach mount for pod web default list of unattached unmounted volumes kubelet kubernetes minion group warning failedsync error syncing pod skipping timeout expired waiting for volumes to attach mount for pod web default list of unattached unmounted volumes what you expected to happen upon node upgrade pods should either be terminated so a controller can recreate them or they should come back to running on their own how to reproduce it as minimally and precisely as possible start a cluster on gce shell git checkout release git pull upstream release make clean quick release cluster kube up sh create the following example petset and wait for all pods to be running shell kubectl create f petset yaml yaml a headless service to create dns records apiversion kind service metadata name nginx labels app nginx spec ports port name web nginx default svc cluster local clusterip none selector app nginx apiversion apps kind petset metadata name web spec servicename nginx replicas template metadata labels app nginx annotations pod alpha kubernetes io initialized true spec containers name nginx image gcr io google containers nginx slim ports containerport name web volumemounts name www mountpath usr share nginx html volumeclaimtemplates metadata name www annotations volume alpha kubernetes io storage class anything spec accessmodes resources requests storage delete petset without cascading keep the pods shell kubectl delete petset web cascade false upgrade master and nodes to beta with cluster gce upgrade sh shell git checkout release git pull upstream release cluster gce upgrade sh beta check for pods stuck in containercreating name ready status restarts age web containercreating web containercreating anything else do we need to know after the above steps some pods may end up terminating it seems to be random which ones terminate and which ones get stuck in containercreating manually deleting and recreating all the stuck pods after the nodes are upgraded allows them to correctly attach to the preexisting pvcs and become running
0
4,182
3,757,107,614
IssuesEvent
2016-03-13 19:51:01
ironjan/klausurtool-ror
https://api.github.com/repos/ironjan/klausurtool-ror
opened
Remove useless button
help wanted usability
Go to some folder's details, e.g. http://0.0.0.0:3000/admin/old_folders/235 The button "Neues Exemplar" links to a anchor on the same page; this anchor is for a inline-form which is directly below the instance table. It's therefore useless and should be removed.
True
Remove useless button - Go to some folder's details, e.g. http://0.0.0.0:3000/admin/old_folders/235 The button "Neues Exemplar" links to a anchor on the same page; this anchor is for a inline-form which is directly below the instance table. It's therefore useless and should be removed.
usab
remove useless button go to some folder s details e g the button neues exemplar links to a anchor on the same page this anchor is for a inline form which is directly below the instance table it s therefore useless and should be removed
1
16,874
11,449,212,627
IssuesEvent
2020-02-06 06:27:07
pombase/canto
https://api.github.com/repos/pombase/canto
closed
Rename the 'Back' button on gene entry page
usability
Related: #1624 When you first start a session and reach the gene entry page, the only navigation button available is a 'Continue' button, which is disabled until you've entered at least one gene ID or specified that the paper has no genes: <img width="400" src="https://user-images.githubusercontent.com/37659591/72737529-a9050380-3b97-11ea-81d8-c4e1cffa3abf.png"> (Note the image shows PHI-Canto's version of the gene entry page, but I believe it works the same way for all versions.) Then, when you return to the gene entry page after having entered genes, a second navigation button appears, labelled 'Back': <img width="400" src="https://user-images.githubusercontent.com/37659591/72737882-59730780-3b98-11ea-8de9-dc0d27f39456.png"> The Back button is always enabled, but the Continue button doesn't enable until the user has entered a gene (or picked a host organism in the case of PHI-Canto). The problem is that the Back button does almost exactly the same thing as the Continue button: they both navigate to the same page; the only difference is that the Continue button stores your changes, whereas the Back button doesn't. If the Back button truly navigated the user backwards in the workflow, it would take them to the page where they entered their curator details. So, the Back button is redundant from a navigational perspective. I think the page would make more sense if the Back button was replaced with a Cancel button (as shown below), since this matches the intended behaviour of the button. <img width="400" src="https://user-images.githubusercontent.com/37659591/72738304-31d06f00-3b99-11ea-9758-726160e702e9.png">
True
Rename the 'Back' button on gene entry page - Related: #1624 When you first start a session and reach the gene entry page, the only navigation button available is a 'Continue' button, which is disabled until you've entered at least one gene ID or specified that the paper has no genes: <img width="400" src="https://user-images.githubusercontent.com/37659591/72737529-a9050380-3b97-11ea-81d8-c4e1cffa3abf.png"> (Note the image shows PHI-Canto's version of the gene entry page, but I believe it works the same way for all versions.) Then, when you return to the gene entry page after having entered genes, a second navigation button appears, labelled 'Back': <img width="400" src="https://user-images.githubusercontent.com/37659591/72737882-59730780-3b98-11ea-8de9-dc0d27f39456.png"> The Back button is always enabled, but the Continue button doesn't enable until the user has entered a gene (or picked a host organism in the case of PHI-Canto). The problem is that the Back button does almost exactly the same thing as the Continue button: they both navigate to the same page; the only difference is that the Continue button stores your changes, whereas the Back button doesn't. If the Back button truly navigated the user backwards in the workflow, it would take them to the page where they entered their curator details. So, the Back button is redundant from a navigational perspective. I think the page would make more sense if the Back button was replaced with a Cancel button (as shown below), since this matches the intended behaviour of the button. <img width="400" src="https://user-images.githubusercontent.com/37659591/72738304-31d06f00-3b99-11ea-9758-726160e702e9.png">
usab
rename the back button on gene entry page related when you first start a session and reach the gene entry page the only navigation button available is a continue button which is disabled until you ve entered at least one gene id or specified that the paper has no genes img width src note the image shows phi canto s version of the gene entry page but i believe it works the same way for all versions then when you return to the gene entry page after having entered genes a second navigation button appears labelled back img width src the back button is always enabled but the continue button doesn t enable until the user has entered a gene or picked a host organism in the case of phi canto the problem is that the back button does almost exactly the same thing as the continue button they both navigate to the same page the only difference is that the continue button stores your changes whereas the back button doesn t if the back button truly navigated the user backwards in the workflow it would take them to the page where they entered their curator details so the back button is redundant from a navigational perspective i think the page would make more sense if the back button was replaced with a cancel button as shown below since this matches the intended behaviour of the button img width src
1
2,157
3,051,939,420
IssuesEvent
2015-08-12 12:01:58
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
15349586: Events in Forecast in Notification Center should be tapable
classification:ui/usability reproducible:always status:open
#### Description Summary: When you open Notification Center's Today View & you have events scheduled for the next day, you can't tap them to open the Calendar.app (https://pbs.twimg.com/media/BXznBrRCAAAFEUs.jpg:large) Steps to Reproduce: 1. Create an event for tomorrow 2. Open Notification Center 3. Swipe to the Today Tab 4. Try to tap on the scheduled event Expected Results: Calendar.app should open and display the next day. Actual Results: Nothing happens - Product Version: 7.0.3 Created: 2013-10-30T08:22:44.441039 Originated: 2013-10-30T09:22:00 Open Radar Link: http://www.openradar.me/15349586
True
15349586: Events in Forecast in Notification Center should be tapable - #### Description Summary: When you open Notification Center's Today View & you have events scheduled for the next day, you can't tap them to open the Calendar.app (https://pbs.twimg.com/media/BXznBrRCAAAFEUs.jpg:large) Steps to Reproduce: 1. Create an event for tomorrow 2. Open Notification Center 3. Swipe to the Today Tab 4. Try to tap on the scheduled event Expected Results: Calendar.app should open and display the next day. Actual Results: Nothing happens - Product Version: 7.0.3 Created: 2013-10-30T08:22:44.441039 Originated: 2013-10-30T09:22:00 Open Radar Link: http://www.openradar.me/15349586
usab
events in forecast in notification center should be tapable description summary when you open notification center s today view you have events scheduled for the next day you can t tap them to open the calendar app steps to reproduce create an event for tomorrow open notification center swipe to the today tab try to tap on the scheduled event expected results calendar app should open and display the next day actual results nothing happens product version created originated open radar link
1
491,352
14,149,839,780
IssuesEvent
2020-11-11 01:52:56
StrangeLoopGames/EcoIssues
https://api.github.com/repos/StrangeLoopGames/EcoIssues
closed
Test self-hosted world
Priority: High QA
We have a special system for teachers that want citizen accounts, but dont want to buy a cloud world., but instead want to self-host an eco server. This system is not used by very many users. but it is being reported as broken. To test: create user with a citizen. in user table, set self_hosted = 1 for the parent user (citizen owner) Make sure the user does not own any cloud worlds in worlds table. You should now be able to login to parent user, go to citizen page, and enter a server ip. Use a non cloud world eco server to test. can be self hosted or third party hosted. Then Login to game with citizen account, and check that user can enter their assigned world. and cannot enter any other world.
1.0
Test self-hosted world - We have a special system for teachers that want citizen accounts, but dont want to buy a cloud world., but instead want to self-host an eco server. This system is not used by very many users. but it is being reported as broken. To test: create user with a citizen. in user table, set self_hosted = 1 for the parent user (citizen owner) Make sure the user does not own any cloud worlds in worlds table. You should now be able to login to parent user, go to citizen page, and enter a server ip. Use a non cloud world eco server to test. can be self hosted or third party hosted. Then Login to game with citizen account, and check that user can enter their assigned world. and cannot enter any other world.
non_usab
test self hosted world we have a special system for teachers that want citizen accounts but dont want to buy a cloud world but instead want to self host an eco server this system is not used by very many users but it is being reported as broken to test create user with a citizen in user table set self hosted for the parent user citizen owner make sure the user does not own any cloud worlds in worlds table you should now be able to login to parent user go to citizen page and enter a server ip use a non cloud world eco server to test can be self hosted or third party hosted then login to game with citizen account and check that user can enter their assigned world and cannot enter any other world
0
254,539
27,393,591,129
IssuesEvent
2023-02-28 17:54:44
opensearch-project/oui
https://api.github.com/repos/opensearch-project/oui
closed
async-0.9.2.js: 1 vulnerabilities (highest severity is: 7.8) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>async-0.9.2.js</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js">https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js</a></p> <p>Path to vulnerable library: /packages/react-datepicker/node_modules/watchpack/node_modules/async/lib/async.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/oui/commit/bbca6f5de4b03c3c57a1333400bba2178b3aa99e">bbca6f5de4b03c3c57a1333400bba2178b3aa99e</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (async version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-43138](https://www.mend.io/vulnerability-database/CVE-2021-43138) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.8 | async-0.9.2.js | Direct | async - 2.6.4,3.2.2 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-43138</summary> ### Vulnerable Library - <b>async-0.9.2.js</b></p> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js">https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js</a></p> <p>Path to vulnerable library: /packages/react-datepicker/node_modules/watchpack/node_modules/async/lib/async.js</p> <p> Dependency Hierarchy: - :x: **async-0.9.2.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/oui/commit/bbca6f5de4b03c3c57a1333400bba2178b3aa99e">bbca6f5de4b03c3c57a1333400bba2178b3aa99e</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Async before 2.6.4 and 3.x before 3.2.2, a malicious user can obtain privileges via the mapValues() method, aka lib/internal/iterator.js createObjectIterator prototype pollution. <p>Publish Date: 2022-04-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-43138>CVE-2021-43138</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p> <p>Release Date: 2022-04-06</p> <p>Fix Resolution: async - 2.6.4,3.2.2</p> </p> <p></p> </details>
True
async-0.9.2.js: 1 vulnerabilities (highest severity is: 7.8) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>async-0.9.2.js</b></p></summary> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js">https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js</a></p> <p>Path to vulnerable library: /packages/react-datepicker/node_modules/watchpack/node_modules/async/lib/async.js</p> <p> <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/oui/commit/bbca6f5de4b03c3c57a1333400bba2178b3aa99e">bbca6f5de4b03c3c57a1333400bba2178b3aa99e</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (async version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2021-43138](https://www.mend.io/vulnerability-database/CVE-2021-43138) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.8 | async-0.9.2.js | Direct | async - 2.6.4,3.2.2 | &#10060; | ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2021-43138</summary> ### Vulnerable Library - <b>async-0.9.2.js</b></p> <p>Higher-order functions and common patterns for asynchronous code</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js">https://cdnjs.cloudflare.com/ajax/libs/async/0.9.2/async.js</a></p> <p>Path to vulnerable library: /packages/react-datepicker/node_modules/watchpack/node_modules/async/lib/async.js</p> <p> Dependency Hierarchy: - :x: **async-0.9.2.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/oui/commit/bbca6f5de4b03c3c57a1333400bba2178b3aa99e">bbca6f5de4b03c3c57a1333400bba2178b3aa99e</a></p> <p>Found in base branch: <b>main</b></p> </p> <p></p> ### Vulnerability Details <p> In Async before 2.6.4 and 3.x before 3.2.2, a malicious user can obtain privileges via the mapValues() method, aka lib/internal/iterator.js createObjectIterator prototype pollution. <p>Publish Date: 2022-04-06 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-43138>CVE-2021-43138</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.8</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2021-43138">https://nvd.nist.gov/vuln/detail/CVE-2021-43138</a></p> <p>Release Date: 2022-04-06</p> <p>Fix Resolution: async - 2.6.4,3.2.2</p> </p> <p></p> </details>
non_usab
async js vulnerabilities highest severity is autoclosed vulnerable library async js higher order functions and common patterns for asynchronous code library home page a href path to vulnerable library packages react datepicker node modules watchpack node modules async lib async js found in head commit a href vulnerabilities cve severity cvss dependency type fixed in async version remediation available high async js direct async details cve vulnerable library async js higher order functions and common patterns for asynchronous code library home page a href path to vulnerable library packages react datepicker node modules watchpack node modules async lib async js dependency hierarchy x async js vulnerable library found in head commit a href found in base branch main vulnerability details in async before and x before a malicious user can obtain privileges via the mapvalues method aka lib internal iterator js createobjectiterator prototype pollution publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution async
0
203,368
7,060,436,598
IssuesEvent
2018-01-05 08:46:06
BlueBrain/neurocurator
https://api.github.com/repos/BlueBrain/neurocurator
opened
Unable to load a PDF in Zotero even if connected to the VPN
bug high priority
Steps: 1. Choose the paper that I wanted to annotate which has no PDF 2. After having the dialog window "This paper is not already in the curator database" and pressing "select PDF" I get no PDF This is the paper I wanted to annotate Postnatal development of the basolateral complex of rabbit amygdala: a stereological and histochemical study Stack trace: ``` [sjimenez:~/neurocurator] $ ./neurocurator_0-4-0_mac_x64 Building ontological tree for PATO:0000261 ... Building ontological tree for NIFORG:birnlex_569 ... Building ontological tree for CHEBI:23367 ... Building ontological tree for GO:0097458 ... Building ontological tree for NIFGA:birnlex_1167 ... Building ontological tree for NIFMOL:nifext_2506 ... Building ontological tree for NIFSUB:sao914572699 ... Building ontological tree for nifext_8054 ... Building ontological tree for sao1813327414 ... Loading cached Zotero data... Cached Zotero data loaded. 2018-01-05 09:26:00.970 neurocurator_0-4-0_mac_x64[46624:1637373] modalSession has been exited prematurely - check for a reentrant call to endModalSession: Traceback (most recent call last): File "urllib/request.py", line 1183, in do_open File "http/client.py", line 1137, in request File "http/client.py", line 1182, in _send_request File "http/client.py", line 1133, in endheaders File "http/client.py", line 963, in _send_output File "http/client.py", line 898, in send File "http/client.py", line 1287, in connect File "ssl.py", line 362, in wrap_socket File "ssl.py", line 580, in __init__ File "ssl.py", line 807, in do_handshake ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "site-packages/neurocurator/mainWin.py", line 1008, in paperSelectionChanged File "site-packages/neurocurator/mainWin.py", line 1113, in importPDF File "site-packages/nat/id.py", line 127, in checkID File "site-packages/nat/id.py", line 134, in checkPMID File "site-packages/nat/id.py", line 17, in getPMIDSoup File "site-packages/nat/id.py", line 26, in getSoup File "urllib/request.py", line 161, in urlopen File "urllib/request.py", line 470, in open File "urllib/request.py", line 580, in http_response File "urllib/request.py", line 502, in error File "urllib/request.py", line 442, in _call_chain File "urllib/request.py", line 685, in http_error_302 File "urllib/request.py", line 464, in open File "urllib/request.py", line 482, in _open File "urllib/request.py", line 442, in _call_chain File "urllib/request.py", line 1226, in https_open File "urllib/request.py", line 1185, in do_open urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)> 2018-01-05 09:31:29.447 neurocurator_0-4-0_mac_x64[46624:1637373] modalSession has been exited prematurely - check for a reentrant call to endModalSession: Traceback (most recent call last): File "urllib/request.py", line 1183, in do_open File "http/client.py", line 1137, in request File "http/client.py", line 1182, in _send_request File "http/client.py", line 1133, in endheaders File "http/client.py", line 963, in _send_output File "http/client.py", line 898, in send File "http/client.py", line 1287, in connect File "ssl.py", line 362, in wrap_socket File "ssl.py", line 580, in __init__ File "ssl.py", line 807, in do_handshake ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "site-packages/neurocurator/mainWin.py", line 1008, in paperSelectionChanged File "site-packages/neurocurator/mainWin.py", line 1113, in importPDF File "site-packages/nat/id.py", line 127, in checkID File "site-packages/nat/id.py", line 134, in checkPMID File "site-packages/nat/id.py", line 17, in getPMIDSoup File "site-packages/nat/id.py", line 26, in getSoup File "urllib/request.py", line 161, in urlopen File "urllib/request.py", line 470, in open File "urllib/request.py", line 580, in http_response File "urllib/request.py", line 502, in error File "urllib/request.py", line 442, in _call_chain File "urllib/request.py", line 685, in http_error_302 File "urllib/request.py", line 464, in open File "urllib/request.py", line 482, in _open File "urllib/request.py", line 442, in _call_chain File "urllib/request.py", line 1226, in https_open File "urllib/request.py", line 1185, in do_open urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)> ```
1.0
Unable to load a PDF in Zotero even if connected to the VPN - Steps: 1. Choose the paper that I wanted to annotate which has no PDF 2. After having the dialog window "This paper is not already in the curator database" and pressing "select PDF" I get no PDF This is the paper I wanted to annotate Postnatal development of the basolateral complex of rabbit amygdala: a stereological and histochemical study Stack trace: ``` [sjimenez:~/neurocurator] $ ./neurocurator_0-4-0_mac_x64 Building ontological tree for PATO:0000261 ... Building ontological tree for NIFORG:birnlex_569 ... Building ontological tree for CHEBI:23367 ... Building ontological tree for GO:0097458 ... Building ontological tree for NIFGA:birnlex_1167 ... Building ontological tree for NIFMOL:nifext_2506 ... Building ontological tree for NIFSUB:sao914572699 ... Building ontological tree for nifext_8054 ... Building ontological tree for sao1813327414 ... Loading cached Zotero data... Cached Zotero data loaded. 2018-01-05 09:26:00.970 neurocurator_0-4-0_mac_x64[46624:1637373] modalSession has been exited prematurely - check for a reentrant call to endModalSession: Traceback (most recent call last): File "urllib/request.py", line 1183, in do_open File "http/client.py", line 1137, in request File "http/client.py", line 1182, in _send_request File "http/client.py", line 1133, in endheaders File "http/client.py", line 963, in _send_output File "http/client.py", line 898, in send File "http/client.py", line 1287, in connect File "ssl.py", line 362, in wrap_socket File "ssl.py", line 580, in __init__ File "ssl.py", line 807, in do_handshake ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "site-packages/neurocurator/mainWin.py", line 1008, in paperSelectionChanged File "site-packages/neurocurator/mainWin.py", line 1113, in importPDF File "site-packages/nat/id.py", line 127, in checkID File "site-packages/nat/id.py", line 134, in checkPMID File "site-packages/nat/id.py", line 17, in getPMIDSoup File "site-packages/nat/id.py", line 26, in getSoup File "urllib/request.py", line 161, in urlopen File "urllib/request.py", line 470, in open File "urllib/request.py", line 580, in http_response File "urllib/request.py", line 502, in error File "urllib/request.py", line 442, in _call_chain File "urllib/request.py", line 685, in http_error_302 File "urllib/request.py", line 464, in open File "urllib/request.py", line 482, in _open File "urllib/request.py", line 442, in _call_chain File "urllib/request.py", line 1226, in https_open File "urllib/request.py", line 1185, in do_open urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)> 2018-01-05 09:31:29.447 neurocurator_0-4-0_mac_x64[46624:1637373] modalSession has been exited prematurely - check for a reentrant call to endModalSession: Traceback (most recent call last): File "urllib/request.py", line 1183, in do_open File "http/client.py", line 1137, in request File "http/client.py", line 1182, in _send_request File "http/client.py", line 1133, in endheaders File "http/client.py", line 963, in _send_output File "http/client.py", line 898, in send File "http/client.py", line 1287, in connect File "ssl.py", line 362, in wrap_socket File "ssl.py", line 580, in __init__ File "ssl.py", line 807, in do_handshake ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600) During handling of the above exception, another exception occurred: Traceback (most recent call last): File "site-packages/neurocurator/mainWin.py", line 1008, in paperSelectionChanged File "site-packages/neurocurator/mainWin.py", line 1113, in importPDF File "site-packages/nat/id.py", line 127, in checkID File "site-packages/nat/id.py", line 134, in checkPMID File "site-packages/nat/id.py", line 17, in getPMIDSoup File "site-packages/nat/id.py", line 26, in getSoup File "urllib/request.py", line 161, in urlopen File "urllib/request.py", line 470, in open File "urllib/request.py", line 580, in http_response File "urllib/request.py", line 502, in error File "urllib/request.py", line 442, in _call_chain File "urllib/request.py", line 685, in http_error_302 File "urllib/request.py", line 464, in open File "urllib/request.py", line 482, in _open File "urllib/request.py", line 442, in _call_chain File "urllib/request.py", line 1226, in https_open File "urllib/request.py", line 1185, in do_open urllib.error.URLError: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed (_ssl.c:600)> ```
non_usab
unable to load a pdf in zotero even if connected to the vpn steps choose the paper that i wanted to annotate which has no pdf after having the dialog window this paper is not already in the curator database and pressing select pdf i get no pdf this is the paper i wanted to annotate postnatal development of the basolateral complex of rabbit amygdala a stereological and histochemical study stack trace neurocurator mac building ontological tree for pato building ontological tree for niforg birnlex building ontological tree for chebi building ontological tree for go building ontological tree for nifga birnlex building ontological tree for nifmol nifext building ontological tree for nifsub building ontological tree for nifext building ontological tree for loading cached zotero data cached zotero data loaded neurocurator mac modalsession has been exited prematurely check for a reentrant call to endmodalsession traceback most recent call last file urllib request py line in do open file http client py line in request file http client py line in send request file http client py line in endheaders file http client py line in send output file http client py line in send file http client py line in connect file ssl py line in wrap socket file ssl py line in init file ssl py line in do handshake ssl sslerror certificate verify failed ssl c during handling of the above exception another exception occurred traceback most recent call last file site packages neurocurator mainwin py line in paperselectionchanged file site packages neurocurator mainwin py line in importpdf file site packages nat id py line in checkid file site packages nat id py line in checkpmid file site packages nat id py line in getpmidsoup file site packages nat id py line in getsoup file urllib request py line in urlopen file urllib request py line in open file urllib request py line in http response file urllib request py line in error file urllib request py line in call chain file urllib request py line in http error file urllib request py line in open file urllib request py line in open file urllib request py line in call chain file urllib request py line in https open file urllib request py line in do open urllib error urlerror neurocurator mac modalsession has been exited prematurely check for a reentrant call to endmodalsession traceback most recent call last file urllib request py line in do open file http client py line in request file http client py line in send request file http client py line in endheaders file http client py line in send output file http client py line in send file http client py line in connect file ssl py line in wrap socket file ssl py line in init file ssl py line in do handshake ssl sslerror certificate verify failed ssl c during handling of the above exception another exception occurred traceback most recent call last file site packages neurocurator mainwin py line in paperselectionchanged file site packages neurocurator mainwin py line in importpdf file site packages nat id py line in checkid file site packages nat id py line in checkpmid file site packages nat id py line in getpmidsoup file site packages nat id py line in getsoup file urllib request py line in urlopen file urllib request py line in open file urllib request py line in http response file urllib request py line in error file urllib request py line in call chain file urllib request py line in http error file urllib request py line in open file urllib request py line in open file urllib request py line in call chain file urllib request py line in https open file urllib request py line in do open urllib error urlerror
0
8,041
5,355,084,868
IssuesEvent
2017-02-20 11:51:34
apinf/platform
https://api.github.com/repos/apinf/platform
opened
API Profile -> Proxy tab: No identification of Mandatory fields
Usability Issue
# Reproduction steps: 1. Sign in as admin and go to an API profile -> proxy tab. 2. Add an API base path. 3. Leave Proxy base path (and other fields) empty and click Save and Publish # Result Although s-alert indicates to fill up mandatory fields, there is no indication which fields are mandatory and needs to be filled in. # Expected Result by using CSS alert class, on leaving mandatory fields empty, we can highlight the field names when they are left empty. # Environment nightly.apinf.io, windows 8, chrome.
True
API Profile -> Proxy tab: No identification of Mandatory fields - # Reproduction steps: 1. Sign in as admin and go to an API profile -> proxy tab. 2. Add an API base path. 3. Leave Proxy base path (and other fields) empty and click Save and Publish # Result Although s-alert indicates to fill up mandatory fields, there is no indication which fields are mandatory and needs to be filled in. # Expected Result by using CSS alert class, on leaving mandatory fields empty, we can highlight the field names when they are left empty. # Environment nightly.apinf.io, windows 8, chrome.
usab
api profile proxy tab no identification of mandatory fields reproduction steps sign in as admin and go to an api profile proxy tab add an api base path leave proxy base path and other fields empty and click save and publish result although s alert indicates to fill up mandatory fields there is no indication which fields are mandatory and needs to be filled in expected result by using css alert class on leaving mandatory fields empty we can highlight the field names when they are left empty environment nightly apinf io windows chrome
1
10,873
6,967,788,052
IssuesEvent
2017-12-10 13:39:32
vmware/docker-volume-vsphere
https://api.github.com/repos/vmware/docker-volume-vsphere
opened
Build with a fresh Windows VM throws error messages
component/guest-plugin component/windows-plugin kind/usability P1
Make issue, below messages get displayed when building a windows plugin with a freshly installed windows VM. Make scripts need to check the existence of older source and then proceed to remove. Remove-Item : Cannot find path 'C:\Users\root\go\src\github.com\vmware\docker-volume-vsphere' because it does not exist. At line:1 char:1 + Remove-Item -Recurse -Force C:\Users\root\go\src\github.com\vmware\do ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : ObjectNotFound: (C:\Users\root\g...-volume-vsphere:String) [Remove- Item], ItemNotFoundException + FullyQualifiedErrorId : PathNotFound,Microsoft.PowerShell.Commands.RemoveItemCommand
True
Build with a fresh Windows VM throws error messages - Make issue, below messages get displayed when building a windows plugin with a freshly installed windows VM. Make scripts need to check the existence of older source and then proceed to remove. Remove-Item : Cannot find path 'C:\Users\root\go\src\github.com\vmware\docker-volume-vsphere' because it does not exist. At line:1 char:1 + Remove-Item -Recurse -Force C:\Users\root\go\src\github.com\vmware\do ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : ObjectNotFound: (C:\Users\root\g...-volume-vsphere:String) [Remove- Item], ItemNotFoundException + FullyQualifiedErrorId : PathNotFound,Microsoft.PowerShell.Commands.RemoveItemCommand
usab
build with a fresh windows vm throws error messages make issue below messages get displayed when building a windows plugin with a freshly installed windows vm make scripts need to check the existence of older source and then proceed to remove remove item cannot find path c users root go src github com vmware docker volume vsphere because it does not exist at line char remove item recurse force c users root go src github com vmware do categoryinfo objectnotfound c users root g volume vsphere string remove item itemnotfoundexception fullyqualifiederrorid pathnotfound microsoft powershell commands removeitemcommand
1
28,001
30,879,163,022
IssuesEvent
2023-08-03 16:13:44
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
closed
clickhouse is unable to start because of the file system.sql.tmp
usability
In some cases after a **power outage** Clickhouse is unable to start. It's unclear how to get to this state. The state itself is easy to reproduce: ``` 22.3 service clickhouse-server stop mv /var/lib/clickhouse/metadata/system.sql /var/lib/clickhouse/metadata/system.sql.tmp service clickhouse-server start <Error> Application: DB::ErrnoException: Cannot open file /var/lib/clickhouse/metadata/system.sql.tmp, errno: 17, strerror: File exists: while loading database `system` from path /var/lib/clickhouse/metadata/system <Information> Application: shutting down ``` It's not clear what is inside `system.sql.tmp` (no info). I am wondering is it related to Atomic/Ordering conversion or not? It seems not. I think `system.sql.tmp` only exists during the initial creation of the system database? This is right? -------------------------- Another problem: if you do `rm /var/lib/clickhouse/metadata/system.sql.tmp` Then clickhouse will create `system.sql` automatically but **Ordinary** ``` # cat /var/lib/clickhouse/metadata/system.sql ATTACH DATABASE system ENGINE = Ordinary # ls -l /var/lib/clickhouse/metadata total 32 lrwxrwxrwx 1 clickhouse clickhouse 67 Apr 4 19:36 default -> /var/lib/clickhouse/store/c0f/c0fdc790-48b7-4d04-80fd-c79048b7ad04/ -rw-r----- 1 clickhouse clickhouse 78 Apr 4 19:36 default.sql drwxr-x--- 2 clickhouse clickhouse 4096 Apr 5 00:24 information_schema drwxr-x--- 2 clickhouse clickhouse 4096 Apr 5 00:24 INFORMATION_SCHEMA -rw-r----- 1 clickhouse clickhouse 51 Apr 5 00:24 information_schema.sql -rw-r----- 1 clickhouse clickhouse 51 Apr 5 00:24 INFORMATION_SCHEMA.sql lrwxrwxrwx 1 clickhouse clickhouse 67 Apr 5 00:30 system -> /var/lib/clickhouse/store/5e3/5e34ade0-c123-43f5-8e76-9da06b17cf70/ -rw-r----- 1 clickhouse clickhouse 41 Apr 5 00:54 system.sql ``` Which is odd and probably will lead to more issues.
True
clickhouse is unable to start because of the file system.sql.tmp - In some cases after a **power outage** Clickhouse is unable to start. It's unclear how to get to this state. The state itself is easy to reproduce: ``` 22.3 service clickhouse-server stop mv /var/lib/clickhouse/metadata/system.sql /var/lib/clickhouse/metadata/system.sql.tmp service clickhouse-server start <Error> Application: DB::ErrnoException: Cannot open file /var/lib/clickhouse/metadata/system.sql.tmp, errno: 17, strerror: File exists: while loading database `system` from path /var/lib/clickhouse/metadata/system <Information> Application: shutting down ``` It's not clear what is inside `system.sql.tmp` (no info). I am wondering is it related to Atomic/Ordering conversion or not? It seems not. I think `system.sql.tmp` only exists during the initial creation of the system database? This is right? -------------------------- Another problem: if you do `rm /var/lib/clickhouse/metadata/system.sql.tmp` Then clickhouse will create `system.sql` automatically but **Ordinary** ``` # cat /var/lib/clickhouse/metadata/system.sql ATTACH DATABASE system ENGINE = Ordinary # ls -l /var/lib/clickhouse/metadata total 32 lrwxrwxrwx 1 clickhouse clickhouse 67 Apr 4 19:36 default -> /var/lib/clickhouse/store/c0f/c0fdc790-48b7-4d04-80fd-c79048b7ad04/ -rw-r----- 1 clickhouse clickhouse 78 Apr 4 19:36 default.sql drwxr-x--- 2 clickhouse clickhouse 4096 Apr 5 00:24 information_schema drwxr-x--- 2 clickhouse clickhouse 4096 Apr 5 00:24 INFORMATION_SCHEMA -rw-r----- 1 clickhouse clickhouse 51 Apr 5 00:24 information_schema.sql -rw-r----- 1 clickhouse clickhouse 51 Apr 5 00:24 INFORMATION_SCHEMA.sql lrwxrwxrwx 1 clickhouse clickhouse 67 Apr 5 00:30 system -> /var/lib/clickhouse/store/5e3/5e34ade0-c123-43f5-8e76-9da06b17cf70/ -rw-r----- 1 clickhouse clickhouse 41 Apr 5 00:54 system.sql ``` Which is odd and probably will lead to more issues.
usab
clickhouse is unable to start because of the file system sql tmp in some cases after a power outage clickhouse is unable to start it s unclear how to get to this state the state itself is easy to reproduce service clickhouse server stop mv var lib clickhouse metadata system sql var lib clickhouse metadata system sql tmp service clickhouse server start application db errnoexception cannot open file var lib clickhouse metadata system sql tmp errno strerror file exists while loading database system from path var lib clickhouse metadata system application shutting down it s not clear what is inside system sql tmp no info i am wondering is it related to atomic ordering conversion or not it seems not i think system sql tmp only exists during the initial creation of the system database this is right another problem if you do rm var lib clickhouse metadata system sql tmp then clickhouse will create system sql automatically but ordinary cat var lib clickhouse metadata system sql attach database system engine ordinary ls l var lib clickhouse metadata total lrwxrwxrwx clickhouse clickhouse apr default var lib clickhouse store rw r clickhouse clickhouse apr default sql drwxr x clickhouse clickhouse apr information schema drwxr x clickhouse clickhouse apr information schema rw r clickhouse clickhouse apr information schema sql rw r clickhouse clickhouse apr information schema sql lrwxrwxrwx clickhouse clickhouse apr system var lib clickhouse store rw r clickhouse clickhouse apr system sql which is odd and probably will lead to more issues
1
420,191
12,234,460,494
IssuesEvent
2020-05-04 13:27:58
threefoldtech/3bot_wallet
https://api.github.com/repos/threefoldtech/3bot_wallet
closed
Stellar Staging - Loading spinner after sending transaction shows wrong loading spinner message.
priority_low type_bug
**Repro steps** 1) Send a transaction 2) Check the loading spinner displayed **Expected Result** "Executing transation" **Actual Result** ![image](https://user-images.githubusercontent.com/59569757/80573674-b0db0c00-8a00-11ea-8148-a1b4121c454f.png) **System Info**
1.0
Stellar Staging - Loading spinner after sending transaction shows wrong loading spinner message. - **Repro steps** 1) Send a transaction 2) Check the loading spinner displayed **Expected Result** "Executing transation" **Actual Result** ![image](https://user-images.githubusercontent.com/59569757/80573674-b0db0c00-8a00-11ea-8148-a1b4121c454f.png) **System Info**
non_usab
stellar staging loading spinner after sending transaction shows wrong loading spinner message repro steps send a transaction check the loading spinner displayed expected result executing transation actual result system info
0
5,278
3,917,079,986
IssuesEvent
2016-04-21 06:27:21
kolliSuman/issues
https://api.github.com/repos/kolliSuman/issues
closed
QA_Expression Trees_Back to experiment_smk
Category: Usability Developed By: VLEAD Release Number: Production Severity: S2 Status: Open
Defect Description: In the "Expression Trees" experiment , the list of experiments link is not present instead the list of experiments link should be displayed on the experiment home page, where the user can easily get back and view the experiments. Actual Result : In the "Expression Trees" experiment , the list of experiments link is not displayed on the experiment home page Environment: "OS: Windows 7, Linux Browsers: Firefox,Chrome Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5" Test Step Link: https://github.com/Virtual-Labs/Data Structure/blob/master/test-cases/integration_test-cases/Expression%20Trees/Expression%20Trees_12_Back%20to%20experiment_smk.org
True
QA_Expression Trees_Back to experiment_smk - Defect Description: In the "Expression Trees" experiment , the list of experiments link is not present instead the list of experiments link should be displayed on the experiment home page, where the user can easily get back and view the experiments. Actual Result : In the "Expression Trees" experiment , the list of experiments link is not displayed on the experiment home page Environment: "OS: Windows 7, Linux Browsers: Firefox,Chrome Bandwidth : 100Mbps Hardware Configuration:8GBRAM , Processor:i5" Test Step Link: https://github.com/Virtual-Labs/Data Structure/blob/master/test-cases/integration_test-cases/Expression%20Trees/Expression%20Trees_12_Back%20to%20experiment_smk.org
usab
qa expression trees back to experiment smk defect description in the expression trees experiment the list of experiments link is not present instead the list of experiments link should be displayed on the experiment home page where the user can easily get back and view the experiments actual result in the expression trees experiment the list of experiments link is not displayed on the experiment home page environment os windows linux browsers firefox chrome bandwidth hardware configuration processor test step link structure blob master test cases integration test cases expression expression back smk org
1
763,195
26,747,004,563
IssuesEvent
2023-01-30 16:39:07
opendatahub-io/odh-dashboard
https://api.github.com/repos/opendatahub-io/odh-dashboard
closed
BYON - Add support for reading custom notebook validation pipeline status
kind/enhancement priority/normal feature/byon
As the custom notebook is progressing through the validation pipeline, we need to be able to get the current status of the pipeline status during each step until validation has completed (un)successfully. Reference: https://github.com/open-services-group/byon/issues/18
1.0
BYON - Add support for reading custom notebook validation pipeline status - As the custom notebook is progressing through the validation pipeline, we need to be able to get the current status of the pipeline status during each step until validation has completed (un)successfully. Reference: https://github.com/open-services-group/byon/issues/18
non_usab
byon add support for reading custom notebook validation pipeline status as the custom notebook is progressing through the validation pipeline we need to be able to get the current status of the pipeline status during each step until validation has completed un successfully reference
0
23,502
22,052,186,726
IssuesEvent
2022-05-30 09:37:38
opentap/opentap
https://api.github.com/repos/opentap/opentap
closed
Incomplete message in log when opening resources
Usability
Originally filed November 23 2021 by Wooi Po Chang on [GitLab](https://gitlab.com/OpenTAP/opentap/-/issues/811) The log message `Waiting for resources to open:` seems to be incomplete ![image](https://gitlab.com/OpenTAP/opentap/uploads/b144f00af1037baa8d7f72c308109bf4/image.png) May consider improving `PrintWaitingMessage` function in [TestPlanExecution.cs](https://gitlab.com/OpenTAP/opentap/-/blob/master/Engine/TestPlanExecution.cs)
True
Incomplete message in log when opening resources - Originally filed November 23 2021 by Wooi Po Chang on [GitLab](https://gitlab.com/OpenTAP/opentap/-/issues/811) The log message `Waiting for resources to open:` seems to be incomplete ![image](https://gitlab.com/OpenTAP/opentap/uploads/b144f00af1037baa8d7f72c308109bf4/image.png) May consider improving `PrintWaitingMessage` function in [TestPlanExecution.cs](https://gitlab.com/OpenTAP/opentap/-/blob/master/Engine/TestPlanExecution.cs)
usab
incomplete message in log when opening resources originally filed november by wooi po chang on the log message waiting for resources to open seems to be incomplete may consider improving printwaitingmessage function in
1
14,224
8,925,810,066
IssuesEvent
2019-01-22 00:52:11
matomo-org/matomo
https://api.github.com/repos/matomo-org/matomo
closed
Website selector: Ctrl+Click should open the website in a new tab
Help wanted c: Usability
Steps to reproduce: * Click on website selector * Do Ctrl+Click on a website name * Got: the website clicked opens in the same window * Expected instead: the website clicked opens in a new tab
True
Website selector: Ctrl+Click should open the website in a new tab - Steps to reproduce: * Click on website selector * Do Ctrl+Click on a website name * Got: the website clicked opens in the same window * Expected instead: the website clicked opens in a new tab
usab
website selector ctrl click should open the website in a new tab steps to reproduce click on website selector do ctrl click on a website name got the website clicked opens in the same window expected instead the website clicked opens in a new tab
1
3,529
3,480,796,695
IssuesEvent
2015-12-29 11:04:54
postmanlabs/postman-app-support
https://api.github.com/repos/postmanlabs/postman-app-support
closed
Clicking on Chrome Takes Me to Postman (Mac)
Usability
On Mac, clicking on Chrome icon on the launcher takes me to Postman when I have both apps running.
True
Clicking on Chrome Takes Me to Postman (Mac) - On Mac, clicking on Chrome icon on the launcher takes me to Postman when I have both apps running.
usab
clicking on chrome takes me to postman mac on mac clicking on chrome icon on the launcher takes me to postman when i have both apps running
1
37,705
12,489,940,980
IssuesEvent
2020-05-31 21:17:30
the-benchmarker/web-frameworks
https://api.github.com/repos/the-benchmarker/web-frameworks
closed
WS-2019-0379 (Medium) detected in commons-codec-1.9.jar
security vulnerability
## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.9.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: /tmp/ws-scm/web-frameworks/java/rapidoid/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/commons-codec/commons-codec/1.9/commons-codec-1.9.jar</p> <p> Dependency Hierarchy: - rapidoid-web-5.5.5.jar (Root Library) - rapidoid-http-client-5.5.5.jar - httpasyncclient-4.1.jar - httpclient-4.4.1.jar - :x: **commons-codec-1.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/the-benchmarker/web-frameworks/commit/ab64b8404e01abede0aa4aa810306b3705409b30">ab64b8404e01abede0aa4aa810306b3705409b30</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-12</p> <p>Fix Resolution: 1.13-RC1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0379 (Medium) detected in commons-codec-1.9.jar - ## WS-2019-0379 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>commons-codec-1.9.jar</b></p></summary> <p>The Apache Commons Codec package contains simple encoder and decoders for various formats such as Base64 and Hexadecimal. In addition to these widely used encoders and decoders, the codec package also maintains a collection of phonetic encoding utilities.</p> <p>Path to dependency file: /tmp/ws-scm/web-frameworks/java/rapidoid/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/commons-codec/commons-codec/1.9/commons-codec-1.9.jar</p> <p> Dependency Hierarchy: - rapidoid-web-5.5.5.jar (Root Library) - rapidoid-http-client-5.5.5.jar - httpasyncclient-4.1.jar - httpclient-4.4.1.jar - :x: **commons-codec-1.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/the-benchmarker/web-frameworks/commit/ab64b8404e01abede0aa4aa810306b3705409b30">ab64b8404e01abede0aa4aa810306b3705409b30</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache commons-codec before version “commons-codec-1.13-RC1” is vulnerable to information disclosure due to Improper Input validation. <p>Publish Date: 2019-05-20 <p>URL: <a href=https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113>WS-2019-0379</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113">https://github.com/apache/commons-codec/commit/48b615756d1d770091ea3322eefc08011ee8b113</a></p> <p>Release Date: 2019-05-12</p> <p>Fix Resolution: 1.13-RC1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_usab
ws medium detected in commons codec jar ws medium severity vulnerability vulnerable library commons codec jar the apache commons codec package contains simple encoder and decoders for various formats such as and hexadecimal in addition to these widely used encoders and decoders the codec package also maintains a collection of phonetic encoding utilities path to dependency file tmp ws scm web frameworks java rapidoid pom xml path to vulnerable library root repository commons codec commons codec commons codec jar dependency hierarchy rapidoid web jar root library rapidoid http client jar httpasyncclient jar httpclient jar x commons codec jar vulnerable library found in head commit a href vulnerability details apache commons codec before version “commons codec ” is vulnerable to information disclosure due to improper input validation publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
2,162
3,051,968,577
IssuesEvent
2015-08-12 12:14:18
mesosphere/marathon
https://api.github.com/repos/mesosphere/marathon
opened
Docker container settings dialog needs better error handling
gui usability
The docker container settings in the app edit/creation modal dialog has only rudimentary error handling. We should do it like at the other app modal fields. AC. 1. Container fields needs a client side validation 2. The API error response must be parsed and displayed at the related field.
True
Docker container settings dialog needs better error handling - The docker container settings in the app edit/creation modal dialog has only rudimentary error handling. We should do it like at the other app modal fields. AC. 1. Container fields needs a client side validation 2. The API error response must be parsed and displayed at the related field.
usab
docker container settings dialog needs better error handling the docker container settings in the app edit creation modal dialog has only rudimentary error handling we should do it like at the other app modal fields ac container fields needs a client side validation the api error response must be parsed and displayed at the related field
1
298,407
25,824,046,011
IssuesEvent
2022-12-12 11:38:15
UCL/TDMS
https://api.github.com/repos/UCL/TDMS
closed
Add/extend unit tests for classes in field.h and arrays.h
testing priority:1
Field has some tests, but these can probably be extended. Also need to extend to test functionality in arrays.h This will be important to ensure we don't break anything when replacing MATLAB. In particular: allocation and deallocation of instances of these classes. ## Tracking ### Basic Types - [x] #152 - [x] #153 - [x] [`MaterialCollection`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L89-L92) -> Doesn't actually have any attributes, and only has _one_ protected method `init_xyz_vectors`. Ergo, doesn't have it's own test file. - [x] #158 - [x] #154 - [x] #155 - [x] #156 ### Classes that have base classes as attributes - [x] [`CCollectionBase`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L94-L99) -> Doesn't have a constructor or any methods, and just stores three `XYZVectors` as attributes. - [x] [`DCollectionBase`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L119-L123) -> Doesn't have a constructor or any methods, and just stores two `XYZVectors` as attributes. - [x] #161 - [x] #162 - [x] #163 - [x] #166 - [x] #173 - [x] #167 ### Subclasses Types - [x] #178 - [x] #184 - [x] #179 - [x] #185 - [x] #174 - [x] #177 - [x] #175 - [x] #172 - [x] #170 - [x] `CCoefficientMatrix` (`Matrix<double>`) - No additional functionality on top of the base class. - [x] #176 ### Test classes for cleanup - [ ] #165 ## `Field.h` classes - [x] Expand existing tests, add Docstrings where possible, add unit tests that are missing
1.0
Add/extend unit tests for classes in field.h and arrays.h - Field has some tests, but these can probably be extended. Also need to extend to test functionality in arrays.h This will be important to ensure we don't break anything when replacing MATLAB. In particular: allocation and deallocation of instances of these classes. ## Tracking ### Basic Types - [x] #152 - [x] #153 - [x] [`MaterialCollection`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L89-L92) -> Doesn't actually have any attributes, and only has _one_ protected method `init_xyz_vectors`. Ergo, doesn't have it's own test file. - [x] #158 - [x] #154 - [x] #155 - [x] #156 ### Classes that have base classes as attributes - [x] [`CCollectionBase`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L94-L99) -> Doesn't have a constructor or any methods, and just stores three `XYZVectors` as attributes. - [x] [`DCollectionBase`](https://github.com/UCL/TDMS/blob/0f00b6ecb53c81ac8cad1f396b77852c971dbc48/tdms/include/arrays.h#L119-L123) -> Doesn't have a constructor or any methods, and just stores two `XYZVectors` as attributes. - [x] #161 - [x] #162 - [x] #163 - [x] #166 - [x] #173 - [x] #167 ### Subclasses Types - [x] #178 - [x] #184 - [x] #179 - [x] #185 - [x] #174 - [x] #177 - [x] #175 - [x] #172 - [x] #170 - [x] `CCoefficientMatrix` (`Matrix<double>`) - No additional functionality on top of the base class. - [x] #176 ### Test classes for cleanup - [ ] #165 ## `Field.h` classes - [x] Expand existing tests, add Docstrings where possible, add unit tests that are missing
non_usab
add extend unit tests for classes in field h and arrays h field has some tests but these can probably be extended also need to extend to test functionality in arrays h this will be important to ensure we don t break anything when replacing matlab in particular allocation and deallocation of instances of these classes tracking basic types doesn t actually have any attributes and only has one protected method init xyz vectors ergo doesn t have it s own test file classes that have base classes as attributes doesn t have a constructor or any methods and just stores three xyzvectors as attributes doesn t have a constructor or any methods and just stores two xyzvectors as attributes subclasses types ccoefficientmatrix matrix no additional functionality on top of the base class test classes for cleanup field h classes expand existing tests add docstrings where possible add unit tests that are missing
0
241,169
18,430,268,208
IssuesEvent
2021-10-14 06:44:35
amzn/selling-partner-api-docs
https://api.github.com/repos/amzn/selling-partner-api-docs
opened
[BUG] Documentation
bug documentation
**If you need help troubleshooting a problem with Selling Partner APIs, please [open a support case](https://developer.amazonservices.com/support). Issues are for reporting documentation bugs and enhancements only.** **Describe the bug** I wanted to use the API to block Amazon Logistics, but In creating the outbound order interface, use the featureConstraints field and set FeatureSettings to prevent Amazon logistics from being invalid. FeatureName is BLOCK_AMZL, and featureFulfillmentPolicy is Required. **URL of the content on GitHub** /fba/outbound/2020-07-01/fulfillmentOrders **Description of what the documentation currently says** FeatureSettings allows users to apply fulfillment features to an order. To block an order from being shipped using Amazon Logistics (AMZL) and an AMZL tracking number, use featureName as BLOCK_AMZL and featureFulfillmentPolicy as Required. Blocking AMZL will incur an additional fee surcharge on your MCF orders and increase the risk of some of your orders being unfulfilled or delivered late if there are no alternative carriers available. Using BLOCK_AMZL in an order request will take precedence over your Seller Central account setting. **Screenshots** #########This is my request data######### { "marketplaceId":null, "sellerFulfillmentOrderId":"#36976-TEST5", "displayableOrderId":"#36976-TEST5", "displayableOrderDate":"2021-10-14", "displayableOrderComment":"Thank you for your order!", "shippingSpeedCategory":"STANDARD", "deliveryWindow":null, "destinationAddress":{ "name":"Dawn Horney", "addressLine1":"14B Garden Ridge Ave", "addressLine2":"", "addressLine3":"", "city":"Greendale", "districtOrCounty":null, "stateOrRegion":"IN", "postalCode":"47025", "countryCode":"US", "phone":"(859) 609-2571" }, "fulfillmentAction":"HOLD", "fulfillmentPolicy":null, "codSettings":null, "shipFromCountryCode":null, "notificationEmails":null, "featureConstraints":[ { "featureName":"BLOCK_AMZL", "featureFulfillmentPolicy":"REQUIRED" } ], "items":[ { "sellerSku":"Womens-XP9303ZLUS9", "sellerFulfillmentOrderItemId":"#36976-0", "quantity":1, "giftMessage":null, "displayableComment":"Thank you for your order!", "fulfillmentNetworkSku":null, "perUnitDeclaredValue":null, "perUnitPrice":null, "perUnitTax":null } ] } ##########This is the returned data############# Contains in the returned data,But Amazon's multi-channel distribution background is not to be stopped。 featureConstraints: [class FeatureSettings { featureName: BLANK_BOX featureFulfillmentPolicy: NotRequired }, class FeatureSettings { featureName: BLOCK_AMZL featureFulfillmentPolicy: Required }] class GetFulfillmentOrderResponse { payload: class GetFulfillmentOrderResult { fulfillmentOrder: class FulfillmentOrder { sellerFulfillmentOrderId: #36976-TEST5 marketplaceId: ATVPDKIKX0DER displayableOrderId: #36976-TEST5 displayableOrderDate: 2021-10-14T00:00:00Z displayableOrderComment: Thank you for your order! shippingSpeedCategory: Standard deliveryWindow: null destinationAddress: class Address { name: Dawn Horney addressLine1: 14B Garden Ridge Ave addressLine2: null addressLine3: null city: Greendale districtOrCounty: null stateOrRegion: IN postalCode: 47025 countryCode: US phone: (859) 609-2571 } fulfillmentAction: Hold fulfillmentPolicy: FillOrKill codSettings: null receivedDate: 2021-10-14T06:32:42Z fulfillmentOrderStatus: Received statusUpdatedDate: 2021-10-14T06:32:46Z notificationEmails: null featureConstraints: [class FeatureSettings { featureName: BLANK_BOX featureFulfillmentPolicy: NotRequired }, class FeatureSettings { featureName: BLOCK_AMZL featureFulfillmentPolicy: Required }] } fulfillmentOrderItems: class FulfillmentOrderItemList { [class FulfillmentOrderItem { sellerSku: Womens-XP9303ZLUS9 sellerFulfillmentOrderItemId: #36976-0 quantity: 1 giftMessage: null displayableComment: Thank you for your order! fulfillmentNetworkSku: X0027JPGVL orderItemDisposition: Sellable cancelledQuantity: 0 unfulfillableQuantity: 0 estimatedShipDate: 2021-10-16T06:59:59Z estimatedArrivalDate: 2021-10-21T06:59:59Z perUnitPrice: null perUnitTax: null perUnitDeclaredValue: class Money { currencyCode: USD value: 78.85 } }] } fulfillmentShipments: class FulfillmentShipmentList { [] } returnItems: class ReturnItemList { [] } returnAuthorizations: class ReturnAuthorizationList { [] } } errors: null } ![微信图片_20211014144410](https://user-images.githubusercontent.com/59277033/137265456-62bca530-f631-4900-bd8a-f2b1d6b08220.png)
1.0
[BUG] Documentation - **If you need help troubleshooting a problem with Selling Partner APIs, please [open a support case](https://developer.amazonservices.com/support). Issues are for reporting documentation bugs and enhancements only.** **Describe the bug** I wanted to use the API to block Amazon Logistics, but In creating the outbound order interface, use the featureConstraints field and set FeatureSettings to prevent Amazon logistics from being invalid. FeatureName is BLOCK_AMZL, and featureFulfillmentPolicy is Required. **URL of the content on GitHub** /fba/outbound/2020-07-01/fulfillmentOrders **Description of what the documentation currently says** FeatureSettings allows users to apply fulfillment features to an order. To block an order from being shipped using Amazon Logistics (AMZL) and an AMZL tracking number, use featureName as BLOCK_AMZL and featureFulfillmentPolicy as Required. Blocking AMZL will incur an additional fee surcharge on your MCF orders and increase the risk of some of your orders being unfulfilled or delivered late if there are no alternative carriers available. Using BLOCK_AMZL in an order request will take precedence over your Seller Central account setting. **Screenshots** #########This is my request data######### { "marketplaceId":null, "sellerFulfillmentOrderId":"#36976-TEST5", "displayableOrderId":"#36976-TEST5", "displayableOrderDate":"2021-10-14", "displayableOrderComment":"Thank you for your order!", "shippingSpeedCategory":"STANDARD", "deliveryWindow":null, "destinationAddress":{ "name":"Dawn Horney", "addressLine1":"14B Garden Ridge Ave", "addressLine2":"", "addressLine3":"", "city":"Greendale", "districtOrCounty":null, "stateOrRegion":"IN", "postalCode":"47025", "countryCode":"US", "phone":"(859) 609-2571" }, "fulfillmentAction":"HOLD", "fulfillmentPolicy":null, "codSettings":null, "shipFromCountryCode":null, "notificationEmails":null, "featureConstraints":[ { "featureName":"BLOCK_AMZL", "featureFulfillmentPolicy":"REQUIRED" } ], "items":[ { "sellerSku":"Womens-XP9303ZLUS9", "sellerFulfillmentOrderItemId":"#36976-0", "quantity":1, "giftMessage":null, "displayableComment":"Thank you for your order!", "fulfillmentNetworkSku":null, "perUnitDeclaredValue":null, "perUnitPrice":null, "perUnitTax":null } ] } ##########This is the returned data############# Contains in the returned data,But Amazon's multi-channel distribution background is not to be stopped。 featureConstraints: [class FeatureSettings { featureName: BLANK_BOX featureFulfillmentPolicy: NotRequired }, class FeatureSettings { featureName: BLOCK_AMZL featureFulfillmentPolicy: Required }] class GetFulfillmentOrderResponse { payload: class GetFulfillmentOrderResult { fulfillmentOrder: class FulfillmentOrder { sellerFulfillmentOrderId: #36976-TEST5 marketplaceId: ATVPDKIKX0DER displayableOrderId: #36976-TEST5 displayableOrderDate: 2021-10-14T00:00:00Z displayableOrderComment: Thank you for your order! shippingSpeedCategory: Standard deliveryWindow: null destinationAddress: class Address { name: Dawn Horney addressLine1: 14B Garden Ridge Ave addressLine2: null addressLine3: null city: Greendale districtOrCounty: null stateOrRegion: IN postalCode: 47025 countryCode: US phone: (859) 609-2571 } fulfillmentAction: Hold fulfillmentPolicy: FillOrKill codSettings: null receivedDate: 2021-10-14T06:32:42Z fulfillmentOrderStatus: Received statusUpdatedDate: 2021-10-14T06:32:46Z notificationEmails: null featureConstraints: [class FeatureSettings { featureName: BLANK_BOX featureFulfillmentPolicy: NotRequired }, class FeatureSettings { featureName: BLOCK_AMZL featureFulfillmentPolicy: Required }] } fulfillmentOrderItems: class FulfillmentOrderItemList { [class FulfillmentOrderItem { sellerSku: Womens-XP9303ZLUS9 sellerFulfillmentOrderItemId: #36976-0 quantity: 1 giftMessage: null displayableComment: Thank you for your order! fulfillmentNetworkSku: X0027JPGVL orderItemDisposition: Sellable cancelledQuantity: 0 unfulfillableQuantity: 0 estimatedShipDate: 2021-10-16T06:59:59Z estimatedArrivalDate: 2021-10-21T06:59:59Z perUnitPrice: null perUnitTax: null perUnitDeclaredValue: class Money { currencyCode: USD value: 78.85 } }] } fulfillmentShipments: class FulfillmentShipmentList { [] } returnItems: class ReturnItemList { [] } returnAuthorizations: class ReturnAuthorizationList { [] } } errors: null } ![微信图片_20211014144410](https://user-images.githubusercontent.com/59277033/137265456-62bca530-f631-4900-bd8a-f2b1d6b08220.png)
non_usab
documentation if you need help troubleshooting a problem with selling partner apis please issues are for reporting documentation bugs and enhancements only describe the bug i wanted to use the api to block amazon logistics but in creating the outbound order interface use the featureconstraints field and set featuresettings to prevent amazon logistics from being invalid featurename is block amzl and featurefulfillmentpolicy is required url of the content on github fba outbound fulfillmentorders description of what the documentation currently says featuresettings allows users to apply fulfillment features to an order to block an order from being shipped using amazon logistics amzl and an amzl tracking number use featurename as block amzl and featurefulfillmentpolicy as required blocking amzl will incur an additional fee surcharge on your mcf orders and increase the risk of some of your orders being unfulfilled or delivered late if there are no alternative carriers available using block amzl in an order request will take precedence over your seller central account setting screenshots this is my request data marketplaceid null sellerfulfillmentorderid displayableorderid displayableorderdate displayableordercomment thank you for your order shippingspeedcategory standard deliverywindow null destinationaddress name dawn horney garden ridge ave city greendale districtorcounty null stateorregion in postalcode countrycode us phone fulfillmentaction hold fulfillmentpolicy null codsettings null shipfromcountrycode null notificationemails null featureconstraints featurename block amzl featurefulfillmentpolicy required items sellersku womens sellerfulfillmentorderitemid quantity giftmessage null displayablecomment thank you for your order fulfillmentnetworksku null perunitdeclaredvalue null perunitprice null perunittax null this is the returned data contains in the returned data,but amazon s multi channel distribution background is not to be stopped。 featureconstraints class featuresettings featurename blank box featurefulfillmentpolicy notrequired class featuresettings featurename block amzl featurefulfillmentpolicy required class getfulfillmentorderresponse payload class getfulfillmentorderresult fulfillmentorder class fulfillmentorder sellerfulfillmentorderid marketplaceid displayableorderid displayableorderdate displayableordercomment thank you for your order shippingspeedcategory standard deliverywindow null destinationaddress class address name dawn horney garden ridge ave null null city greendale districtorcounty null stateorregion in postalcode countrycode us phone fulfillmentaction hold fulfillmentpolicy fillorkill codsettings null receiveddate fulfillmentorderstatus received statusupdateddate notificationemails null featureconstraints class featuresettings featurename blank box featurefulfillmentpolicy notrequired class featuresettings featurename block amzl featurefulfillmentpolicy required fulfillmentorderitems class fulfillmentorderitemlist class fulfillmentorderitem sellersku womens sellerfulfillmentorderitemid quantity giftmessage null displayablecomment thank you for your order fulfillmentnetworksku orderitemdisposition sellable cancelledquantity unfulfillablequantity estimatedshipdate estimatedarrivaldate perunitprice null perunittax null perunitdeclaredvalue class money currencycode usd value fulfillmentshipments class fulfillmentshipmentlist returnitems class returnitemlist returnauthorizations class returnauthorizationlist errors null
0
1,847
3,025,053,800
IssuesEvent
2015-08-03 04:25:40
lionheart/openradar-mirror
https://api.github.com/repos/lionheart/openradar-mirror
opened
21745023: App Store trials!
classification:ui/usability reproducible:always status:open
#### Description This is a duplicate of rdar://13197637 I'm sure you've got this request many times, but I'm adding my voice to it. Here’s how I could see it working: Developers can choose whether to allow a trial of 1, 7 or 30 days, or to disallow trials all together, on a per-app basis. For those apps that allow trials, the App Store would show a “Try for 7 days” button alongside “Buy app”. If you install the trial app, it gets flagged with a “Trial” flag over the icon, in a similar way to the “New” flag that was introduced in iOS 6. When you run the app, it permanently runs with a double height status bar, which shows the amount of time left. Tapping this status bar opens the app store page allowing you to purchase the app. Once a trial is used up, you can’t get a trial of that app again on the same Apple ID. When the trial is used up, the app icon still remains on your device, but the flag says “Expired” and tapping on it takes you to the App Store rather than opening the app. - Product Version: n/a Created: 2015-07-09T14:08:10.128640 Originated: 2015-07-09T16:08:00 Open Radar Link: http://www.openradar.me/21745023
True
21745023: App Store trials! - #### Description This is a duplicate of rdar://13197637 I'm sure you've got this request many times, but I'm adding my voice to it. Here’s how I could see it working: Developers can choose whether to allow a trial of 1, 7 or 30 days, or to disallow trials all together, on a per-app basis. For those apps that allow trials, the App Store would show a “Try for 7 days” button alongside “Buy app”. If you install the trial app, it gets flagged with a “Trial” flag over the icon, in a similar way to the “New” flag that was introduced in iOS 6. When you run the app, it permanently runs with a double height status bar, which shows the amount of time left. Tapping this status bar opens the app store page allowing you to purchase the app. Once a trial is used up, you can’t get a trial of that app again on the same Apple ID. When the trial is used up, the app icon still remains on your device, but the flag says “Expired” and tapping on it takes you to the App Store rather than opening the app. - Product Version: n/a Created: 2015-07-09T14:08:10.128640 Originated: 2015-07-09T16:08:00 Open Radar Link: http://www.openradar.me/21745023
usab
app store trials description this is a duplicate of rdar i m sure you ve got this request many times but i m adding my voice to it here’s how i could see it working developers can choose whether to allow a trial of or days or to disallow trials all together on a per app basis for those apps that allow trials the app store would show a “try for days” button alongside “buy app” if you install the trial app it gets flagged with a “trial” flag over the icon in a similar way to the “new” flag that was introduced in ios when you run the app it permanently runs with a double height status bar which shows the amount of time left tapping this status bar opens the app store page allowing you to purchase the app once a trial is used up you can’t get a trial of that app again on the same apple id when the trial is used up the app icon still remains on your device but the flag says “expired” and tapping on it takes you to the app store rather than opening the app product version n a created originated open radar link
1
6,498
4,321,306,941
IssuesEvent
2016-07-25 09:40:44
rpi-virtuell/reliwerk
https://api.github.com/repos/rpi-virtuell/reliwerk
closed
"Gruppe verlassen"
Text Übersetzung Usability
Nun ist genau das passiert, was ich immer befürchtet habe: das riesige "Gruppe verlassen" - Schild verführt einfach dazu, darauf zu klicken, wenn man die Arbeit in der Gruppe beendet hat. Warum diese Aufforderung zum endgültigen Verlassen so groß da stehen muss, erschließt sich mir nicht. Kann man das nicht klein oben rein machen? Es ist wirklich irritierend. Heute Morgen habe ich mit Müh und Not eine Kollegin in die Gruppe "geschleußt" und jetzt ist sie wieder draußen. Ich bitte sehr darum, diesen Button zu ändern.
True
"Gruppe verlassen" - Nun ist genau das passiert, was ich immer befürchtet habe: das riesige "Gruppe verlassen" - Schild verführt einfach dazu, darauf zu klicken, wenn man die Arbeit in der Gruppe beendet hat. Warum diese Aufforderung zum endgültigen Verlassen so groß da stehen muss, erschließt sich mir nicht. Kann man das nicht klein oben rein machen? Es ist wirklich irritierend. Heute Morgen habe ich mit Müh und Not eine Kollegin in die Gruppe "geschleußt" und jetzt ist sie wieder draußen. Ich bitte sehr darum, diesen Button zu ändern.
usab
gruppe verlassen nun ist genau das passiert was ich immer befürchtet habe das riesige gruppe verlassen schild verführt einfach dazu darauf zu klicken wenn man die arbeit in der gruppe beendet hat warum diese aufforderung zum endgültigen verlassen so groß da stehen muss erschließt sich mir nicht kann man das nicht klein oben rein machen es ist wirklich irritierend heute morgen habe ich mit müh und not eine kollegin in die gruppe geschleußt und jetzt ist sie wieder draußen ich bitte sehr darum diesen button zu ändern
1
289,486
31,933,038,899
IssuesEvent
2023-09-19 08:42:37
Trinadh465/linux-4.1.15_CVE-2023-4128
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-4128
opened
CVE-2022-34918 (High) detected in linuxlinux-4.6
Mend: dependency security vulnerability
## CVE-2022-34918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel through 5.18.9. A type confusion bug in nft_set_elem_init (leading to a buffer overflow) could be used by a local attacker to escalate privileges, a different vulnerability than CVE-2022-32250. (The attacker can obtain root access, but must start with an unprivileged user namespace to obtain CAP_NET_ADMIN access.) This can be fixed in nft_setelem_parse_data in net/netfilter/nf_tables_api.c. <p>Publish Date: 2022-07-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-34918>CVE-2022-34918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-34918">https://www.linuxkernelcves.com/cves/CVE-2022-34918</a></p> <p>Release Date: 2022-07-04</p> <p>Fix Resolution: v5.10.130,v5.15.54,v5.18.11,v5.19-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-34918 (High) detected in linuxlinux-4.6 - ## CVE-2022-34918 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/netfilter/nf_tables_api.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel through 5.18.9. A type confusion bug in nft_set_elem_init (leading to a buffer overflow) could be used by a local attacker to escalate privileges, a different vulnerability than CVE-2022-32250. (The attacker can obtain root access, but must start with an unprivileged user namespace to obtain CAP_NET_ADMIN access.) This can be fixed in nft_setelem_parse_data in net/netfilter/nf_tables_api.c. <p>Publish Date: 2022-07-04 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-34918>CVE-2022-34918</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-34918">https://www.linuxkernelcves.com/cves/CVE-2022-34918</a></p> <p>Release Date: 2022-07-04</p> <p>Fix Resolution: v5.10.130,v5.15.54,v5.18.11,v5.19-rc6</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_usab
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net netfilter nf tables api c net netfilter nf tables api c vulnerability details an issue was discovered in the linux kernel through a type confusion bug in nft set elem init leading to a buffer overflow could be used by a local attacker to escalate privileges a different vulnerability than cve the attacker can obtain root access but must start with an unprivileged user namespace to obtain cap net admin access this can be fixed in nft setelem parse data in net netfilter nf tables api c publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
5,260
3,194,296,432
IssuesEvent
2015-09-30 11:17:05
VcDevel/Vc
https://api.github.com/repos/VcDevel/Vc
closed
drop all internal namespaces and move everything to Vc::Detail
code cleanup
Right now there are several internal namespaces hiding implementation details. All of these namespaces need to be folded to `Vc::Detail`.
1.0
drop all internal namespaces and move everything to Vc::Detail - Right now there are several internal namespaces hiding implementation details. All of these namespaces need to be folded to `Vc::Detail`.
non_usab
drop all internal namespaces and move everything to vc detail right now there are several internal namespaces hiding implementation details all of these namespaces need to be folded to vc detail
0
17,839
12,373,218,495
IssuesEvent
2020-05-18 22:01:12
argoproj/argo-cd
https://api.github.com/repos/argoproj/argo-cd
reopened
Display duration for the entire sync and deploy/rollout for a revision in history window
component:ui enhancement good first issue type:usability
# Summary Display duration for the entire sync and deploy/rollout for a revision in history window. Currently the "History and Rollback" window shows revision and timing when it started but not clearly shows the entire sync/deploy duration it took to complete it. It will be good to see duration as shown below in Red font. ![image](https://user-images.githubusercontent.com/22686326/80660481-aa6c8480-8a40-11ea-844d-3b603c13ae87.png) What change you think needs making. The sync history window needs to be updated to show duration for each sync which occurred in the past. # Motivation This will help the user to get the trend of past sync durations and see how much each of the previous sync took and make decision based on that. Please give examples of your use case, e.g. when would you use this. To view past sync durations and see its trend. # Proposal How do you think this should be implemented?
True
Display duration for the entire sync and deploy/rollout for a revision in history window - # Summary Display duration for the entire sync and deploy/rollout for a revision in history window. Currently the "History and Rollback" window shows revision and timing when it started but not clearly shows the entire sync/deploy duration it took to complete it. It will be good to see duration as shown below in Red font. ![image](https://user-images.githubusercontent.com/22686326/80660481-aa6c8480-8a40-11ea-844d-3b603c13ae87.png) What change you think needs making. The sync history window needs to be updated to show duration for each sync which occurred in the past. # Motivation This will help the user to get the trend of past sync durations and see how much each of the previous sync took and make decision based on that. Please give examples of your use case, e.g. when would you use this. To view past sync durations and see its trend. # Proposal How do you think this should be implemented?
usab
display duration for the entire sync and deploy rollout for a revision in history window summary display duration for the entire sync and deploy rollout for a revision in history window currently the history and rollback window shows revision and timing when it started but not clearly shows the entire sync deploy duration it took to complete it it will be good to see duration as shown below in red font what change you think needs making the sync history window needs to be updated to show duration for each sync which occurred in the past motivation this will help the user to get the trend of past sync durations and see how much each of the previous sync took and make decision based on that please give examples of your use case e g when would you use this to view past sync durations and see its trend proposal how do you think this should be implemented
1
7,555
5,050,547,155
IssuesEvent
2016-12-20 19:05:11
loconomics/loconomics
https://api.github.com/repos/loconomics/loconomics
closed
Cache data that's already been inputted into forms
C: Usability D: Duplicated F: Onboarding P3 R1: Ready for Development Sprint
From @jucsanch: He left to a help page and came back and data was gone.
True
Cache data that's already been inputted into forms - From @jucsanch: He left to a help page and came back and data was gone.
usab
cache data that s already been inputted into forms from jucsanch he left to a help page and came back and data was gone
1
14,437
9,186,911,478
IssuesEvent
2019-03-06 00:36:40
uber/hudi
https://api.github.com/repos/uber/hudi
closed
Reduce memory usage on WriteStatus for implicit indexes
enhancement usability
For things like BloomIndex, there is no need to keep accumulating the hoodiekey/recordlocation per IOHandle to be passed to index update.. Fixing this, can greatly ease bulk inserts/upserts with large number of keys and small record sizes
True
Reduce memory usage on WriteStatus for implicit indexes - For things like BloomIndex, there is no need to keep accumulating the hoodiekey/recordlocation per IOHandle to be passed to index update.. Fixing this, can greatly ease bulk inserts/upserts with large number of keys and small record sizes
usab
reduce memory usage on writestatus for implicit indexes for things like bloomindex there is no need to keep accumulating the hoodiekey recordlocation per iohandle to be passed to index update fixing this can greatly ease bulk inserts upserts with large number of keys and small record sizes
1
11,827
7,477,842,513
IssuesEvent
2018-04-04 09:35:58
cnr-ibf-pa/hbp-bsp-issues
https://api.github.com/repos/cnr-ibf-pa/hbp-bsp-issues
opened
Analysis notebook change for SynapticEventsFitting
Type_BUG Type_Usability UC_SynapEvFitting_UData UC_SynapEvFitting_UModel
New Analysis notebooks require: - [ ] User documentation update required to explain the new GUI to the user - [ ] Notebook documentation update required to show the timestamp to the user and have the user remember it for the Analysis step - [ ] Make sure that in either case (NSG, Jureca) the correct (European) time is shown
True
Analysis notebook change for SynapticEventsFitting - New Analysis notebooks require: - [ ] User documentation update required to explain the new GUI to the user - [ ] Notebook documentation update required to show the timestamp to the user and have the user remember it for the Analysis step - [ ] Make sure that in either case (NSG, Jureca) the correct (European) time is shown
usab
analysis notebook change for synapticeventsfitting new analysis notebooks require user documentation update required to explain the new gui to the user notebook documentation update required to show the timestamp to the user and have the user remember it for the analysis step make sure that in either case nsg jureca the correct european time is shown
1
5,940
4,077,223,909
IssuesEvent
2016-05-30 06:57:04
godotengine/godot
https://api.github.com/repos/godotengine/godot
opened
Changing the language directly in the project manager
enhancement topic:editor usability
Because we can, and since we spawn a new process for the editor anyway, it would be the right time to select another language. In a more practical context, it would also make it very easy to find if we want to encourage people to make screenshots in English for bug reports, and not in their own locale.
True
Changing the language directly in the project manager - Because we can, and since we spawn a new process for the editor anyway, it would be the right time to select another language. In a more practical context, it would also make it very easy to find if we want to encourage people to make screenshots in English for bug reports, and not in their own locale.
usab
changing the language directly in the project manager because we can and since we spawn a new process for the editor anyway it would be the right time to select another language in a more practical context it would also make it very easy to find if we want to encourage people to make screenshots in english for bug reports and not in their own locale
1
1,767
3,004,048,870
IssuesEvent
2015-07-25 14:43:03
MDAnalysis/mdanalysis
https://api.github.com/repos/MDAnalysis/mdanalysis
closed
MDAnalysis.analysis.distances.contact_matrix
Component-Docs usability
The docstring in `0.11` for this function incorrectly suggests that a list of coordinates is used as input, while a numpy array is clearly the correct input (list won't work). Also, in practice I am not seeing the progress meter before or after 0.11 (mac or linux, even with a large input array), which I was testing for when trying to write some user documentation for `0.11` API updates [MDAnalysis 0.11 unifying release user guide: Suppressing contact matrix progress meter](/MDAnalysis/mdanalysis/wiki/MDAnalysis-0.11-unifying-release-user-guide#suppressing-contact-matrix-progress-meter-new-quiet-keyword).
True
MDAnalysis.analysis.distances.contact_matrix - The docstring in `0.11` for this function incorrectly suggests that a list of coordinates is used as input, while a numpy array is clearly the correct input (list won't work). Also, in practice I am not seeing the progress meter before or after 0.11 (mac or linux, even with a large input array), which I was testing for when trying to write some user documentation for `0.11` API updates [MDAnalysis 0.11 unifying release user guide: Suppressing contact matrix progress meter](/MDAnalysis/mdanalysis/wiki/MDAnalysis-0.11-unifying-release-user-guide#suppressing-contact-matrix-progress-meter-new-quiet-keyword).
usab
mdanalysis analysis distances contact matrix the docstring in for this function incorrectly suggests that a list of coordinates is used as input while a numpy array is clearly the correct input list won t work also in practice i am not seeing the progress meter before or after mac or linux even with a large input array which i was testing for when trying to write some user documentation for api updates mdanalysis mdanalysis wiki mdanalysis unifying release user guide suppressing contact matrix progress meter new quiet keyword
1
5,599
3,969,158,476
IssuesEvent
2016-05-03 22:17:31
pipelinedb/pipelinedb
https://api.github.com/repos/pipelinedb/pipelinedb
opened
Add backfill utility
usability
It should use `COPY` to bulkload historical data into a target set of CVs.
True
Add backfill utility - It should use `COPY` to bulkload historical data into a target set of CVs.
usab
add backfill utility it should use copy to bulkload historical data into a target set of cvs
1
20,952
16,357,306,383
IssuesEvent
2021-05-14 01:39:38
microsoft/win32metadata
https://api.github.com/repos/microsoft/win32metadata
closed
Windows.Win32.winmd needs to be self-describing
usability
The winmd lacks the definitions of various attributes and types like `SupportedArchitecture` - everything in the `Windows.Win32.Interop` namespace is missing - so there's no way for tools to parse these attributes without a priori knowledge of them. WinRT metadata includes the definitions of all the attributes it defines in the `Windows.Foundation.Metadata` namespace. Win32 should do the same.
True
Windows.Win32.winmd needs to be self-describing - The winmd lacks the definitions of various attributes and types like `SupportedArchitecture` - everything in the `Windows.Win32.Interop` namespace is missing - so there's no way for tools to parse these attributes without a priori knowledge of them. WinRT metadata includes the definitions of all the attributes it defines in the `Windows.Foundation.Metadata` namespace. Win32 should do the same.
usab
windows winmd needs to be self describing the winmd lacks the definitions of various attributes and types like supportedarchitecture everything in the windows interop namespace is missing so there s no way for tools to parse these attributes without a priori knowledge of them winrt metadata includes the definitions of all the attributes it defines in the windows foundation metadata namespace should do the same
1
16,010
10,481,504,066
IssuesEvent
2019-09-24 09:49:33
pombase/canto
https://api.github.com/repos/pombase/canto
closed
Add help text for summary page genotype heading
PHI-Canto feedback_required usability
Pull request #1649 added a new heading on the summary page to separate the genotype links from the gene links. However this new heading doesn't have any help text associated with it, whereas the gene heading does. Is it worth adding any help text?
True
Add help text for summary page genotype heading - Pull request #1649 added a new heading on the summary page to separate the genotype links from the gene links. However this new heading doesn't have any help text associated with it, whereas the gene heading does. Is it worth adding any help text?
usab
add help text for summary page genotype heading pull request added a new heading on the summary page to separate the genotype links from the gene links however this new heading doesn t have any help text associated with it whereas the gene heading does is it worth adding any help text
1