Unnamed: 0 int64 0 832k | id float64 2.49B 32.1B | type stringclasses 1 value | created_at stringlengths 19 19 | repo stringlengths 4 112 | repo_url stringlengths 33 141 | action stringclasses 3 values | title stringlengths 1 1.02k | labels stringlengths 4 1.54k | body stringlengths 1 262k | index stringclasses 17 values | text_combine stringlengths 95 262k | label stringclasses 2 values | text stringlengths 96 252k | binary_label int64 0 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
8,714 | 3,004,049,298 | IssuesEvent | 2015-07-25 14:43:35 | IntellectualSites/PlotSquared | https://api.github.com/repos/IntellectualSites/PlotSquared | closed | NPE | [!] bug [?] needs testing | [23:36:03] [Server thread/WARN]: [PlotSquared] Task #6 for PlotSquared v2.12.15 generated an exception
java.lang.NullPointerException
at com.intellectualcrafters.plot.listeners.PlotPlusListener$1.run(PlotPlusListener.java:90) ~[?:?]
at org.bukkit.craftbukkit.v1_8_R3.scheduler.CraftTask.run(CraftTask.java:71) ~[creative.jar:git-PaperSpigot-4d70f42-b105298]
at org.bukkit.craftbukkit.v1_8_R3.scheduler.CraftScheduler.mainThreadHeartbeat(CraftScheduler.java:350) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at net.minecraft.server.v1_8_R3.MinecraftServer.B(MinecraftServer.java:774) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at net.minecraft.server.v1_8_R3.DedicatedServer.B(DedicatedServer.java:378) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at net.minecraft.server.v1_8_R3.MinecraftServer.A(MinecraftServer.java:705) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at net.minecraft.server.v1_8_R3.MinecraftServer.run(MinecraftServer.java:608) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_45]
| 1.0 | NPE - [23:36:03] [Server thread/WARN]: [PlotSquared] Task #6 for PlotSquared v2.12.15 generated an exception
java.lang.NullPointerException
at com.intellectualcrafters.plot.listeners.PlotPlusListener$1.run(PlotPlusListener.java:90) ~[?:?]
at org.bukkit.craftbukkit.v1_8_R3.scheduler.CraftTask.run(CraftTask.java:71) ~[creative.jar:git-PaperSpigot-4d70f42-b105298]
at org.bukkit.craftbukkit.v1_8_R3.scheduler.CraftScheduler.mainThreadHeartbeat(CraftScheduler.java:350) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at net.minecraft.server.v1_8_R3.MinecraftServer.B(MinecraftServer.java:774) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at net.minecraft.server.v1_8_R3.DedicatedServer.B(DedicatedServer.java:378) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at net.minecraft.server.v1_8_R3.MinecraftServer.A(MinecraftServer.java:705) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at net.minecraft.server.v1_8_R3.MinecraftServer.run(MinecraftServer.java:608) [creative.jar:git-PaperSpigot-4d70f42-b105298]
at java.lang.Thread.run(Thread.java:745) [?:1.8.0_45]
| test | npe task for plotsquared generated an exception java lang nullpointerexception at com intellectualcrafters plot listeners plotpluslistener run plotpluslistener java at org bukkit craftbukkit scheduler crafttask run crafttask java at org bukkit craftbukkit scheduler craftscheduler mainthreadheartbeat craftscheduler java at net minecraft server minecraftserver b minecraftserver java at net minecraft server dedicatedserver b dedicatedserver java at net minecraft server minecraftserver a minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run thread java | 1 |
128,212 | 12,367,140,713 | IssuesEvent | 2020-05-18 11:46:25 | ponylang/ponyup | https://api.github.com/repos/ponylang/ponyup | opened | Document ponyup macOS/brew libressl connection | documentation help wanted | #117 was caused by this.
macOS, it's all dynamic linking.
`brew install libressl` will periodically change the version it installs as they switch to a newer version of libressl. This is rare, but does happen. When that happens, older versions of ponyup will stop working once libressl is updated.
We need to document the expected error that users would see and tell them to reinstall ponyup via the init script as that will download the most recent nightly version. Within 24 hours of a libressl change, it will work again.
This also means, that eventually if they update to a different version of ponyup, that it might be using a different libressl than they have installed and will fail. | 1.0 | Document ponyup macOS/brew libressl connection - #117 was caused by this.
macOS, it's all dynamic linking.
`brew install libressl` will periodically change the version it installs as they switch to a newer version of libressl. This is rare, but does happen. When that happens, older versions of ponyup will stop working once libressl is updated.
We need to document the expected error that users would see and tell them to reinstall ponyup via the init script as that will download the most recent nightly version. Within 24 hours of a libressl change, it will work again.
This also means, that eventually if they update to a different version of ponyup, that it might be using a different libressl than they have installed and will fail. | non_test | document ponyup macos brew libressl connection was caused by this macos it s all dynamic linking brew install libressl will periodically change the version it installs as they switch to a newer version of libressl this is rare but does happen when that happens older versions of ponyup will stop working once libressl is updated we need to document the expected error that users would see and tell them to reinstall ponyup via the init script as that will download the most recent nightly version within hours of a libressl change it will work again this also means that eventually if they update to a different version of ponyup that it might be using a different libressl than they have installed and will fail | 0 |
214,295 | 16,580,204,319 | IssuesEvent | 2021-05-31 10:41:11 | blynkkk/blynk_Issues | https://api.github.com/repos/blynkkk/blynk_Issues | closed | Web dashboard ignores empty space | bug ready to test web | On web dashboard, when you leave a space on top of a widget, after you save the dashboard, it goes up. It does not allow white space in UI design. (Mac Os 11.3.1, Firefox 88.0.1)
This is editing page.

This is after you save.

| 1.0 | Web dashboard ignores empty space - On web dashboard, when you leave a space on top of a widget, after you save the dashboard, it goes up. It does not allow white space in UI design. (Mac Os 11.3.1, Firefox 88.0.1)
This is editing page.

This is after you save.

| test | web dashboard ignores empty space on web dashboard when you leave a space on top of a widget after you save the dashboard it goes up it does not allow white space in ui design mac os firefox this is editing page this is after you save | 1 |
78,647 | 7,657,016,347 | IssuesEvent | 2018-05-10 18:13:42 | couchbase/couchbase-lite-ios | https://api.github.com/repos/couchbase/couchbase-lite-ios | closed | latest iOS cbl 2.1.0 builds are crashing | functional-test-blocker ready | ### Version
CBL - 2.1.0-150
sg version -> 2.1.0-55
### Issue caused
1. Running ios functional tests on jenkins having crash on the app and bunch of tests failing
I ran tests with 2.1.0-126 and that looks good
If I run tests individually, it works fine.
### Logs:
[CBLTestServer-iOS_2018-04-30-154305-1_Dans-Test-MacBook-Pro.crash.zip](https://github.com/couchbase/couchbase-lite-ios/files/1968363/CBLTestServer-iOS_2018-04-30-154305-1_Dans-Test-MacBook-Pro.crash.zip)
| 1.0 | latest iOS cbl 2.1.0 builds are crashing - ### Version
CBL - 2.1.0-150
sg version -> 2.1.0-55
### Issue caused
1. Running ios functional tests on jenkins having crash on the app and bunch of tests failing
I ran tests with 2.1.0-126 and that looks good
If I run tests individually, it works fine.
### Logs:
[CBLTestServer-iOS_2018-04-30-154305-1_Dans-Test-MacBook-Pro.crash.zip](https://github.com/couchbase/couchbase-lite-ios/files/1968363/CBLTestServer-iOS_2018-04-30-154305-1_Dans-Test-MacBook-Pro.crash.zip)
| test | latest ios cbl builds are crashing version cbl sg version issue caused running ios functional tests on jenkins having crash on the app and bunch of tests failing i ran tests with and that looks good if i run tests individually it works fine logs | 1 |
231,484 | 17,690,791,027 | IssuesEvent | 2021-08-24 09:40:33 | owncloud/ocis | https://api.github.com/repos/owncloud/ocis | reopened | Write documentation for roles & permissions concept | Topic:Documentation | Write documentation for roles & permissions concept. This is intended to be a living document and it should reflect decisions made during the development process. The end result being a document in a state that mirrors the settings service functionality regarding roles and permissions. | 1.0 | Write documentation for roles & permissions concept - Write documentation for roles & permissions concept. This is intended to be a living document and it should reflect decisions made during the development process. The end result being a document in a state that mirrors the settings service functionality regarding roles and permissions. | non_test | write documentation for roles permissions concept write documentation for roles permissions concept this is intended to be a living document and it should reflect decisions made during the development process the end result being a document in a state that mirrors the settings service functionality regarding roles and permissions | 0 |
251,817 | 27,211,188,498 | IssuesEvent | 2023-02-20 16:40:23 | ZSBRybnik/frontend | https://api.github.com/repos/ZSBRybnik/frontend | closed | node-jq-2.3.3.tgz: 3 vulnerabilities (highest severity is: 7.5) - autoclosed | security vulnerability | <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-jq-2.3.3.tgz</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (node-jq version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-25881](https://www.mend.io/vulnerability-database/CVE-2022-25881) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | http-cache-semantics-3.8.1.tgz | Transitive | N/A* | ❌ |
| [CVE-2023-25166](https://www.mend.io/vulnerability-database/CVE-2023-25166) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | formula-3.0.0.tgz | Transitive | N/A* | ❌ |
| [CVE-2022-33987](https://www.mend.io/vulnerability-database/CVE-2022-33987) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-25881</summary>
### Vulnerable Library - <b>http-cache-semantics-3.8.1.tgz</b></p>
<p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz</a></p>
<p>
Dependency Hierarchy:
- node-jq-2.3.3.tgz (Root Library)
- download-8.0.0.tgz
- got-8.3.2.tgz
- cacheable-request-2.1.4.tgz
- :x: **http-cache-semantics-3.8.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library.
<p>Publish Date: 2023-01-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p>
<p>Release Date: 2023-01-31</p>
<p>Fix Resolution: http-cache-semantics - 4.1.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-25166</summary>
### Vulnerable Library - <b>formula-3.0.0.tgz</b></p>
<p>Math and string formula parser.</p>
<p>Library home page: <a href="https://registry.npmjs.org/@sideway/formula/-/formula-3.0.0.tgz">https://registry.npmjs.org/@sideway/formula/-/formula-3.0.0.tgz</a></p>
<p>
Dependency Hierarchy:
- node-jq-2.3.3.tgz (Root Library)
- joi-17.6.0.tgz
- :x: **formula-3.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
formula is a math and string formula parser. In versions prior to 3.0.1 crafted user-provided strings to formula's parser might lead to polynomial execution time and a denial of service. Users should upgrade to 3.0.1+. There are no known workarounds for this vulnerability.
<p>Publish Date: 2023-02-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25166>CVE-2023-25166</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25166">https://www.cve.org/CVERecord?id=CVE-2023-25166</a></p>
<p>Release Date: 2023-02-08</p>
<p>Fix Resolution: @sideway/formula - 3.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-33987</summary>
### Vulnerable Libraries - <b>got-8.3.2.tgz</b>, <b>got-7.1.0.tgz</b></p>
<p>
### <b>got-8.3.2.tgz</b></p>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-8.3.2.tgz">https://registry.npmjs.org/got/-/got-8.3.2.tgz</a></p>
<p>
Dependency Hierarchy:
- node-jq-2.3.3.tgz (Root Library)
- download-8.0.0.tgz
- :x: **got-8.3.2.tgz** (Vulnerable Library)
### <b>got-7.1.0.tgz</b></p>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-7.1.0.tgz">https://registry.npmjs.org/got/-/got-7.1.0.tgz</a></p>
<p>
Dependency Hierarchy:
- node-jq-2.3.3.tgz (Root Library)
- bin-build-3.0.0.tgz
- download-6.2.5.tgz
- :x: **got-7.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The got package before 12.1.0 (also fixed in 11.8.5) for Node.js allows a redirect to a UNIX socket.
<p>Publish Date: 2022-06-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-33987>CVE-2022-33987</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p>
<p>Release Date: 2022-06-18</p>
<p>Fix Resolution: got - 11.8.5,12.1.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | True | node-jq-2.3.3.tgz: 3 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-jq-2.3.3.tgz</b></p></summary>
<p></p>
<p>
<p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p></details>
## Vulnerabilities
| CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (node-jq version) | Remediation Available |
| ------------- | ------------- | ----- | ----- | ----- | ------------- | --- |
| [CVE-2022-25881](https://www.mend.io/vulnerability-database/CVE-2022-25881) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | http-cache-semantics-3.8.1.tgz | Transitive | N/A* | ❌ |
| [CVE-2023-25166](https://www.mend.io/vulnerability-database/CVE-2023-25166) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | formula-3.0.0.tgz | Transitive | N/A* | ❌ |
| [CVE-2022-33987](https://www.mend.io/vulnerability-database/CVE-2022-33987) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | N/A* | ❌ |
<p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p>
## Details
<details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-25881</summary>
### Vulnerable Library - <b>http-cache-semantics-3.8.1.tgz</b></p>
<p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p>
<p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz</a></p>
<p>
Dependency Hierarchy:
- node-jq-2.3.3.tgz (Root Library)
- download-8.0.0.tgz
- got-8.3.2.tgz
- cacheable-request-2.1.4.tgz
- :x: **http-cache-semantics-3.8.1.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library.
<p>Publish Date: 2023-01-31
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>7.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p>
<p>Release Date: 2023-01-31</p>
<p>Fix Resolution: http-cache-semantics - 4.1.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-25166</summary>
### Vulnerable Library - <b>formula-3.0.0.tgz</b></p>
<p>Math and string formula parser.</p>
<p>Library home page: <a href="https://registry.npmjs.org/@sideway/formula/-/formula-3.0.0.tgz">https://registry.npmjs.org/@sideway/formula/-/formula-3.0.0.tgz</a></p>
<p>
Dependency Hierarchy:
- node-jq-2.3.3.tgz (Root Library)
- joi-17.6.0.tgz
- :x: **formula-3.0.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
formula is a math and string formula parser. In versions prior to 3.0.1 crafted user-provided strings to formula's parser might lead to polynomial execution time and a denial of service. Users should upgrade to 3.0.1+. There are no known workarounds for this vulnerability.
<p>Publish Date: 2023-02-08
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25166>CVE-2023-25166</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>6.5</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25166">https://www.cve.org/CVERecord?id=CVE-2023-25166</a></p>
<p>Release Date: 2023-02-08</p>
<p>Fix Resolution: @sideway/formula - 3.0.1</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details><details>
<summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-33987</summary>
### Vulnerable Libraries - <b>got-8.3.2.tgz</b>, <b>got-7.1.0.tgz</b></p>
<p>
### <b>got-8.3.2.tgz</b></p>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-8.3.2.tgz">https://registry.npmjs.org/got/-/got-8.3.2.tgz</a></p>
<p>
Dependency Hierarchy:
- node-jq-2.3.3.tgz (Root Library)
- download-8.0.0.tgz
- :x: **got-8.3.2.tgz** (Vulnerable Library)
### <b>got-7.1.0.tgz</b></p>
<p>Simplified HTTP requests</p>
<p>Library home page: <a href="https://registry.npmjs.org/got/-/got-7.1.0.tgz">https://registry.npmjs.org/got/-/got-7.1.0.tgz</a></p>
<p>
Dependency Hierarchy:
- node-jq-2.3.3.tgz (Root Library)
- bin-build-3.0.0.tgz
- download-6.2.5.tgz
- :x: **got-7.1.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
<p></p>
### Vulnerability Details
<p>
The got package before 12.1.0 (also fixed in 11.8.5) for Node.js allows a redirect to a UNIX socket.
<p>Publish Date: 2022-06-18
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-33987>CVE-2022-33987</a></p>
</p>
<p></p>
### CVSS 3 Score Details (<b>5.3</b>)
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
<p></p>
### Suggested Fix
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p>
<p>Release Date: 2022-06-18</p>
<p>Fix Resolution: got - 11.8.5,12.1.0</p>
</p>
<p></p>
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
</details> | non_test | node jq tgz vulnerabilities highest severity is autoclosed vulnerable library node jq tgz found in head commit a href vulnerabilities cve severity cvss dependency type fixed in node jq version remediation available high http cache semantics tgz transitive n a medium formula tgz transitive n a medium detected in multiple dependencies transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library http cache semantics tgz parses cache control and other headers helps building correct http caches and proxies library home page a href dependency hierarchy node jq tgz root library download tgz got tgz cacheable request tgz x http cache semantics tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects versions of the package http cache semantics before the issue can be exploited via malicious request header values sent to a server when that server reads the cache policy from the request using this library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution http cache semantics step up your open source security game with mend cve vulnerable library formula tgz math and string formula parser library home page a href dependency hierarchy node jq tgz root library joi tgz x formula tgz vulnerable library found in head commit a href found in base branch master vulnerability details formula is a math and string formula parser in versions prior to crafted user provided strings to formula s parser might lead to polynomial execution time and a denial of service users should upgrade to there are no known workarounds for this vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution sideway formula step up your open source security game with mend cve vulnerable libraries got tgz got tgz got tgz simplified http requests library home page a href dependency hierarchy node jq tgz root library download tgz x got tgz vulnerable library got tgz simplified http requests library home page a href dependency hierarchy node jq tgz root library bin build tgz download tgz x got tgz vulnerable library found in head commit a href found in base branch master vulnerability details the got package before also fixed in for node js allows a redirect to a unix socket publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution got step up your open source security game with mend | 0 |
23,527 | 10,894,847,044 | IssuesEvent | 2019-11-19 09:31:21 | elikkatzgit/quantumsim | https://api.github.com/repos/elikkatzgit/quantumsim | closed | CVE-2015-0220 (Medium) detected in Django-1.3.tar.gz | security vulnerability | ## CVE-2015-0220 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.3.tar.gz</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz">https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz</a></p>
<p>Path to dependency file: /tmp/ws-scm/quantumsim/requirements.txt</p>
<p>Path to vulnerable library: /quantumsim/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-1.3.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/quantumsim/commit/d6624156203bb0fc439915ed3fc47432b9cbbeb5">d6624156203bb0fc439915ed3fc47432b9cbbeb5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The django.util.http.is_safe_url function in Django before 1.4.18, 1.6.x before 1.6.10, and 1.7.x before 1.7.3 does not properly handle leading whitespaces, which allows remote attackers to conduct cross-site scripting (XSS) attacks via a crafted URL, related to redirect URLs, as demonstrated by a "\njavascript:" URL.
<p>Publish Date: 2015-01-16
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-0220>CVE-2015-0220</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-0220">https://nvd.nist.gov/vuln/detail/CVE-2015-0220</a></p>
<p>Release Date: 2015-01-16</p>
<p>Fix Resolution: 1.4.18,1.6.10,1.7.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Django","packageVersion":"1.3","isTransitiveDependency":false,"dependencyTree":"Django:1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.18,1.6.10,1.7.3"}],"vulnerabilityIdentifier":"CVE-2015-0220","vulnerabilityDetails":"The django.util.http.is_safe_url function in Django before 1.4.18, 1.6.x before 1.6.10, and 1.7.x before 1.7.3 does not properly handle leading whitespaces, which allows remote attackers to conduct cross-site scripting (XSS) attacks via a crafted URL, related to redirect URLs, as demonstrated by a \"\\njavascript:\" URL.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2015-0220","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> --> | True | CVE-2015-0220 (Medium) detected in Django-1.3.tar.gz - ## CVE-2015-0220 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.3.tar.gz</b></p></summary>
<p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz">https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz</a></p>
<p>Path to dependency file: /tmp/ws-scm/quantumsim/requirements.txt</p>
<p>Path to vulnerable library: /quantumsim/requirements.txt</p>
<p>
Dependency Hierarchy:
- :x: **Django-1.3.tar.gz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/quantumsim/commit/d6624156203bb0fc439915ed3fc47432b9cbbeb5">d6624156203bb0fc439915ed3fc47432b9cbbeb5</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The django.util.http.is_safe_url function in Django before 1.4.18, 1.6.x before 1.6.10, and 1.7.x before 1.7.3 does not properly handle leading whitespaces, which allows remote attackers to conduct cross-site scripting (XSS) attacks via a crafted URL, related to redirect URLs, as demonstrated by a "\njavascript:" URL.
<p>Publish Date: 2015-01-16
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-0220>CVE-2015-0220</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-0220">https://nvd.nist.gov/vuln/detail/CVE-2015-0220</a></p>
<p>Release Date: 2015-01-16</p>
<p>Fix Resolution: 1.4.18,1.6.10,1.7.3</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Django","packageVersion":"1.3","isTransitiveDependency":false,"dependencyTree":"Django:1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.18,1.6.10,1.7.3"}],"vulnerabilityIdentifier":"CVE-2015-0220","vulnerabilityDetails":"The django.util.http.is_safe_url function in Django before 1.4.18, 1.6.x before 1.6.10, and 1.7.x before 1.7.3 does not properly handle leading whitespaces, which allows remote attackers to conduct cross-site scripting (XSS) attacks via a crafted URL, related to redirect URLs, as demonstrated by a \"\\njavascript:\" URL.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2015-0220","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> --> | non_test | cve medium detected in django tar gz cve medium severity vulnerability vulnerable library django tar gz a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file tmp ws scm quantumsim requirements txt path to vulnerable library quantumsim requirements txt dependency hierarchy x django tar gz vulnerable library found in head commit a href vulnerability details the django util http is safe url function in django before x before and x before does not properly handle leading whitespaces which allows remote attackers to conduct cross site scripting xss attacks via a crafted url related to redirect urls as demonstrated by a njavascript url publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the django util http is safe url function in django before x before and x before does not properly handle leading whitespaces which allows remote attackers to conduct cross site scripting xss attacks via a crafted url related to redirect urls as demonstrated by a njavascript url vulnerabilityurl | 0 |
372,395 | 11,013,717,793 | IssuesEvent | 2019-12-04 21:05:27 | dmwm/WMCore | https://api.github.com/repos/dmwm/WMCore | closed | Break closure of phedex blocks into smaller chunks | BUG High Priority Need wmagent branch WMAgent | **Impact of the bug**
WMAgent (PhEDExInjector)
**Describe the bug**
Apparently we put too much data into the same PhEDEx call to close blocks here:
https://github.com/dmwm/WMCore/blob/master/src/python/WMComponent/PhEDExInjector/PhEDExInjectorPoller.py#L375
it's one call per location, and that location might have many many blocks from multiple datasets.
The problem is, if there is a problem with any of those blocks (e.g., the current state of vocms0253 which has a **dataset** closed in PhEDEx, see [1]), the whole request would fail and no blocks can be closed...
**How to reproduce it**
none
**Expected behavior**
My suggestion is to break the http requests per location and dataset, something like
for location in locations:
for dataset in datasets:
make a phedex request
Yes, it will increase the amount of PhEDEx calls, but still it's going to be 1 call per block that we close, which should be just fine for the phedex data-service.
**Additional context and error message**
[1]
```
2019-11-27 22:23:08,945:140168429426432:WARNING:Service:The cachefile /data/srv/wmagent/v1.2.6.patch1/install/wmagent/PhEDExInjector/.wmcore_cache/.wmcore_cache_31961/requests/cmsweb.cern.ch:8443/-7685143179613178687_POST_inject does n
ot exist and the service at https://cmsweb.cern.ch:8443/phedex/datasvc/json/prod/inject is unavailable - it returned 400 because Bad Request with result: injectData error: dataset /RadionToWW_narrow_M-6500_TuneCUETP8M1_13TeV-madgraph-p
ythia8/RunIISummer16MiniAODv3-PUMoriond17_94X_mcRun2_asymptotic_v3-v1/MINIAODSIM is closed\n
2019-11-27 22:23:08,946:140168429426432:ERROR:PhEDExInjectorPoller:PhEDEx block close failed with HTTPException: 400 injectData error: dataset /RadionToWW_narrow_M-6500_TuneCUETP8M1_13TeV-madgraph-pythia8/RunIISummer16MiniAODv3-PUMorio
nd17_94X_mcRun2_asymptotic_v3-v1/MINIAODSIM is closed\n
``` | 1.0 | Break closure of phedex blocks into smaller chunks - **Impact of the bug**
WMAgent (PhEDExInjector)
**Describe the bug**
Apparently we put too much data into the same PhEDEx call to close blocks here:
https://github.com/dmwm/WMCore/blob/master/src/python/WMComponent/PhEDExInjector/PhEDExInjectorPoller.py#L375
it's one call per location, and that location might have many many blocks from multiple datasets.
The problem is, if there is a problem with any of those blocks (e.g., the current state of vocms0253 which has a **dataset** closed in PhEDEx, see [1]), the whole request would fail and no blocks can be closed...
**How to reproduce it**
none
**Expected behavior**
My suggestion is to break the http requests per location and dataset, something like
for location in locations:
for dataset in datasets:
make a phedex request
Yes, it will increase the amount of PhEDEx calls, but still it's going to be 1 call per block that we close, which should be just fine for the phedex data-service.
**Additional context and error message**
[1]
```
2019-11-27 22:23:08,945:140168429426432:WARNING:Service:The cachefile /data/srv/wmagent/v1.2.6.patch1/install/wmagent/PhEDExInjector/.wmcore_cache/.wmcore_cache_31961/requests/cmsweb.cern.ch:8443/-7685143179613178687_POST_inject does n
ot exist and the service at https://cmsweb.cern.ch:8443/phedex/datasvc/json/prod/inject is unavailable - it returned 400 because Bad Request with result: injectData error: dataset /RadionToWW_narrow_M-6500_TuneCUETP8M1_13TeV-madgraph-p
ythia8/RunIISummer16MiniAODv3-PUMoriond17_94X_mcRun2_asymptotic_v3-v1/MINIAODSIM is closed\n
2019-11-27 22:23:08,946:140168429426432:ERROR:PhEDExInjectorPoller:PhEDEx block close failed with HTTPException: 400 injectData error: dataset /RadionToWW_narrow_M-6500_TuneCUETP8M1_13TeV-madgraph-pythia8/RunIISummer16MiniAODv3-PUMorio
nd17_94X_mcRun2_asymptotic_v3-v1/MINIAODSIM is closed\n
``` | non_test | break closure of phedex blocks into smaller chunks impact of the bug wmagent phedexinjector describe the bug apparently we put too much data into the same phedex call to close blocks here it s one call per location and that location might have many many blocks from multiple datasets the problem is if there is a problem with any of those blocks e g the current state of which has a dataset closed in phedex see the whole request would fail and no blocks can be closed how to reproduce it none expected behavior my suggestion is to break the http requests per location and dataset something like for location in locations for dataset in datasets make a phedex request yes it will increase the amount of phedex calls but still it s going to be call per block that we close which should be just fine for the phedex data service additional context and error message warning service the cachefile data srv wmagent install wmagent phedexinjector wmcore cache wmcore cache requests cmsweb cern ch post inject does n ot exist and the service at is unavailable it returned because bad request with result injectdata error dataset radiontoww narrow m madgraph p asymptotic miniaodsim is closed n error phedexinjectorpoller phedex block close failed with httpexception injectdata error dataset radiontoww narrow m madgraph pumorio asymptotic miniaodsim is closed n | 0 |
249,612 | 21,179,721,071 | IssuesEvent | 2022-04-08 06:34:23 | microsoft/vscode | https://api.github.com/repos/microsoft/vscode | opened | Terminal flaky tests | smoke-test-failure | Lately terminal tests became flaky with:
```
1) VSCode Smoke Tests (Web)
Terminal
Terminal Editors
should update color of the tab:
Error: Timeout: is active element '.quick-input-widget .quick-input-box input' after 20 seconds.
at Code.poll (D:\a\_work\1\s\test\automation\src\code.ts:296:11)
at Code.waitForActiveElement (D:\a\_work\1\s\test\automation\src\code.ts:232:3)
at QuickInput.waitForQuickInputOpened (D:\a\_work\1\s\test\automation\src\quickinput.ts:20:3)
at Terminal.runCommandWithValue (D:\a\_work\1\s\test\automation\src\terminal.ts:88:3)
at Context.<anonymous> (src\areas\terminal\terminal-editors.test.ts:21:4)
```
This tests uses `runCommandWithValue` and there is a somewhat questionable line here:
https://github.com/microsoft/vscode/blob/921264bfe3ffbbfe5ec8c4b08214b88e2148fa3e/test/automation/src/terminal.ts#L83-L88
Fyi we had updated playwright to latest.
I will go ahead and skip tests for now that use this method. | 1.0 | Terminal flaky tests - Lately terminal tests became flaky with:
```
1) VSCode Smoke Tests (Web)
Terminal
Terminal Editors
should update color of the tab:
Error: Timeout: is active element '.quick-input-widget .quick-input-box input' after 20 seconds.
at Code.poll (D:\a\_work\1\s\test\automation\src\code.ts:296:11)
at Code.waitForActiveElement (D:\a\_work\1\s\test\automation\src\code.ts:232:3)
at QuickInput.waitForQuickInputOpened (D:\a\_work\1\s\test\automation\src\quickinput.ts:20:3)
at Terminal.runCommandWithValue (D:\a\_work\1\s\test\automation\src\terminal.ts:88:3)
at Context.<anonymous> (src\areas\terminal\terminal-editors.test.ts:21:4)
```
This tests uses `runCommandWithValue` and there is a somewhat questionable line here:
https://github.com/microsoft/vscode/blob/921264bfe3ffbbfe5ec8c4b08214b88e2148fa3e/test/automation/src/terminal.ts#L83-L88
Fyi we had updated playwright to latest.
I will go ahead and skip tests for now that use this method. | test | terminal flaky tests lately terminal tests became flaky with vscode smoke tests web terminal terminal editors should update color of the tab error timeout is active element quick input widget quick input box input after seconds at code poll d a work s test automation src code ts at code waitforactiveelement d a work s test automation src code ts at quickinput waitforquickinputopened d a work s test automation src quickinput ts at terminal runcommandwithvalue d a work s test automation src terminal ts at context src areas terminal terminal editors test ts this tests uses runcommandwithvalue and there is a somewhat questionable line here fyi we had updated playwright to latest i will go ahead and skip tests for now that use this method | 1 |
296,736 | 25,572,522,611 | IssuesEvent | 2022-11-30 18:56:16 | MD-Anderson-Bioinformatics/NG-CHM | https://api.github.com/repos/MD-Anderson-Bioinformatics/NG-CHM | closed | Pubmed linkouts need to be opened in new frame. | bug linkouts passed retest 2.21.3 | If linkout frame opened, Pubmed linkouts do not work.
Refused to display 'https://pubmed.ncbi.nlm.nih.gov/' in a frame because it set 'X-Frame-Options' to 'deny'. | 1.0 | Pubmed linkouts need to be opened in new frame. - If linkout frame opened, Pubmed linkouts do not work.
Refused to display 'https://pubmed.ncbi.nlm.nih.gov/' in a frame because it set 'X-Frame-Options' to 'deny'. | test | pubmed linkouts need to be opened in new frame if linkout frame opened pubmed linkouts do not work refused to display in a frame because it set x frame options to deny | 1 |
293,607 | 8,998,093,882 | IssuesEvent | 2019-02-02 18:33:57 | Beep6581/RawTherapee | https://api.github.com/repos/Beep6581/RawTherapee | opened | Segfault in lmmse demosaic | Priority-Critical bug | There's a really hard to reproduce segfault in lmmse demosaic.
Searching..... | 1.0 | Segfault in lmmse demosaic - There's a really hard to reproduce segfault in lmmse demosaic.
Searching..... | non_test | segfault in lmmse demosaic there s a really hard to reproduce segfault in lmmse demosaic searching | 0 |
58,511 | 24,468,752,529 | IssuesEvent | 2022-10-07 17:31:38 | valor-software/valor-software.github.io | https://api.github.com/repos/valor-software/valor-software.github.io | closed | Design | Service page | service page | To create a Design page that should have a path:
Home > Services > Design
It can be accessible via:
https://valor-software.com/services
https://valor-software.com/
main menu
Design: https://www.figma.com/file/StpiCGh7YZyAPRjtD5gJBo/Valor-Site-Design-2021?node-id=9450%3A82643 | 1.0 | Design | Service page - To create a Design page that should have a path:
Home > Services > Design
It can be accessible via:
https://valor-software.com/services
https://valor-software.com/
main menu
Design: https://www.figma.com/file/StpiCGh7YZyAPRjtD5gJBo/Valor-Site-Design-2021?node-id=9450%3A82643 | non_test | design service page to create a design page that should have a path home services design it can be accessible via main menu design | 0 |
272,205 | 20,737,252,094 | IssuesEvent | 2022-03-14 14:42:46 | dj-stripe/dj-stripe | https://api.github.com/repos/dj-stripe/dj-stripe | closed | Ability to create express account attached to auth user | documentation | Hi. Is does this package have a way to create an express account and attach it to the django auth user (onetoonefield) or does this have to be done manually?
| 1.0 | Ability to create express account attached to auth user - Hi. Is does this package have a way to create an express account and attach it to the django auth user (onetoonefield) or does this have to be done manually?
| non_test | ability to create express account attached to auth user hi is does this package have a way to create an express account and attach it to the django auth user onetoonefield or does this have to be done manually | 0 |
60,590 | 6,711,005,623 | IssuesEvent | 2017-10-13 00:49:29 | ansible/galaxy-issues | https://api.github.com/repos/ansible/galaxy-issues | closed | No error when adding a new role/container project that has the same name as an existing project | bug ready for testing | ## Steps to Reproduce
1. Visit 'roleadd' on Ansible Galaxy (https://galaxy.ansible.com/roleadd#/)
2. Click the 'enable' toggle next to a role named `username/xyz`
3. Click the configure widget, and verify the role name is `xyz`
4. Wait for role import to succeed (green dot next to 'Succeeded')
5. Click the 'enable' toggle next to another role named `username/something-else`
6. Click the configure widget, and change role name to `xyz`
7. Click 'Save' and observe the result.
## Expected Result
* I should see an error message alerting me that a role with that name already exists in my Galaxy namespace.
## Actual Result
The 'Running' status widget keeps spinning forever, and nothing seems to happen. After a while, if you refresh the page, you can see the project was imported as `something-else` and not `xyz` (I think, if I remember correctly).
This was discovered when I was adding a `geerlingguy.solr` ansible-container project, while I already had a `geerlingguy.solr` ansible role in my user namespace. See: https://github.com/ansible/ansible-container/issues/629 | 1.0 | No error when adding a new role/container project that has the same name as an existing project - ## Steps to Reproduce
1. Visit 'roleadd' on Ansible Galaxy (https://galaxy.ansible.com/roleadd#/)
2. Click the 'enable' toggle next to a role named `username/xyz`
3. Click the configure widget, and verify the role name is `xyz`
4. Wait for role import to succeed (green dot next to 'Succeeded')
5. Click the 'enable' toggle next to another role named `username/something-else`
6. Click the configure widget, and change role name to `xyz`
7. Click 'Save' and observe the result.
## Expected Result
* I should see an error message alerting me that a role with that name already exists in my Galaxy namespace.
## Actual Result
The 'Running' status widget keeps spinning forever, and nothing seems to happen. After a while, if you refresh the page, you can see the project was imported as `something-else` and not `xyz` (I think, if I remember correctly).
This was discovered when I was adding a `geerlingguy.solr` ansible-container project, while I already had a `geerlingguy.solr` ansible role in my user namespace. See: https://github.com/ansible/ansible-container/issues/629 | test | no error when adding a new role container project that has the same name as an existing project steps to reproduce visit roleadd on ansible galaxy click the enable toggle next to a role named username xyz click the configure widget and verify the role name is xyz wait for role import to succeed green dot next to succeeded click the enable toggle next to another role named username something else click the configure widget and change role name to xyz click save and observe the result expected result i should see an error message alerting me that a role with that name already exists in my galaxy namespace actual result the running status widget keeps spinning forever and nothing seems to happen after a while if you refresh the page you can see the project was imported as something else and not xyz i think if i remember correctly this was discovered when i was adding a geerlingguy solr ansible container project while i already had a geerlingguy solr ansible role in my user namespace see | 1 |
55,529 | 14,533,130,997 | IssuesEvent | 2020-12-14 23:51:30 | department-of-veterans-affairs/va.gov-team | https://api.github.com/repos/department-of-veterans-affairs/va.gov-team | closed | 508-defect-2 [AXE-CORE]: App Directory - Heading levels should increase by one | 508-defect-2 508-issue-headings 508/Accessibility | # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
## Feedback framework
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Definition of done
1. Review and acknowledge feedback.
1. Fix and/or document decisions made.
1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. -->
**VFS Point of Contact:** _Trevor_
## User Story or Problem Statement
<!-- Example: As a user with cognitive considerations, I expect to see a label and input pairing consistently styled as throughout the rest of the site, with the label just above the text/email/search input or to the right of a radio/checkbox input, so that I am clearly able to understand what entry is expected. -->
As an assistive tech user, I want to hear headings read out in the correct nesting order.
## Details
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
Our app headings and sub-headings inside the accordions need to be H2 and H3 headings respectively. Screen shots attached below.
## Acceptance Criteria
- [ ] Axe browser plugin doesn't report a heading nesting best practice warning on future runs
- [ ] Current visual styles are maintained
## WCAG or Vendor Guidance (optional)
* [Heading levels should only increase by one](https://dequeuniversity.com/rules/axe/4.0/heading-order)
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->

---

| 1.0 | 508-defect-2 [AXE-CORE]: App Directory - Heading levels should increase by one - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2)
<!--
Enter an issue title using the format [ERROR TYPE]: Brief description of the problem
---
[SCREENREADER]: Edit buttons need aria-label for context
[KEYBOARD]: Add another user link will not receive keyboard focus
[AXE-CORE]: Heading levels should increase by one
[COGNITION]: Error messages should be more specific
[COLOR]: Blue button on blue background does not have sufficient contrast ratio
---
-->
<!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. -->
## Feedback framework
- **❗️ Must** for if the feedback must be applied
- **⚠️ Should** if the feedback is best practice
- **✔️ Consider** for suggestions/enhancements
## Definition of done
1. Review and acknowledge feedback.
1. Fix and/or document decisions made.
1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix.
## Point of Contact
<!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. -->
**VFS Point of Contact:** _Trevor_
## User Story or Problem Statement
<!-- Example: As a user with cognitive considerations, I expect to see a label and input pairing consistently styled as throughout the rest of the site, with the label just above the text/email/search input or to the right of a radio/checkbox input, so that I am clearly able to understand what entry is expected. -->
As an assistive tech user, I want to hear headings read out in the correct nesting order.
## Details
<!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. -->
Our app headings and sub-headings inside the accordions need to be H2 and H3 headings respectively. Screen shots attached below.
## Acceptance Criteria
- [ ] Axe browser plugin doesn't report a heading nesting best practice warning on future runs
- [ ] Current visual styles are maintained
## WCAG or Vendor Guidance (optional)
* [Heading levels should only increase by one](https://dequeuniversity.com/rules/axe/4.0/heading-order)
## Screenshots or Trace Logs
<!-- Drop any screenshots or error logs that might be useful for debugging -->

---

| non_test | defect app directory heading levels should increase by one enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements definition of done review and acknowledge feedback fix and or document decisions made accessibility specialist will close ticket after reviewing documented decisions validating fix point of contact vfs point of contact trevor user story or problem statement as an assistive tech user i want to hear headings read out in the correct nesting order details our app headings and sub headings inside the accordions need to be and headings respectively screen shots attached below acceptance criteria axe browser plugin doesn t report a heading nesting best practice warning on future runs current visual styles are maintained wcag or vendor guidance optional screenshots or trace logs | 0 |
280,984 | 24,352,632,114 | IssuesEvent | 2022-10-03 02:42:31 | ECP-WarpX/WarpX | https://api.github.com/repos/ECP-WarpX/WarpX | reopened | oneAPI 2022.2.0 Hangs in CI | bug component: tests install component: third party bug: affects latest release backend: dpc++ workaround | Since the update 1 week ago from `2022.1.0` to `2022.2.0`, most CI runs using either ICX (host) or DPC++ (device) compiles hang.
It looks like this is from the linking part: https://github.com/ECP-WarpX/WarpX/pull/3421#issuecomment-1261311343
The same problem appears in the [AMReX](https://github.com/AMReX-Codes/amrex/) CI.
Open this issue for triage and tracking.
cc @rscohn2 | 2.0 | oneAPI 2022.2.0 Hangs in CI - Since the update 1 week ago from `2022.1.0` to `2022.2.0`, most CI runs using either ICX (host) or DPC++ (device) compiles hang.
It looks like this is from the linking part: https://github.com/ECP-WarpX/WarpX/pull/3421#issuecomment-1261311343
The same problem appears in the [AMReX](https://github.com/AMReX-Codes/amrex/) CI.
Open this issue for triage and tracking.
cc @rscohn2 | test | oneapi hangs in ci since the update week ago from to most ci runs using either icx host or dpc device compiles hang it looks like this is from the linking part the same problem appears in the ci open this issue for triage and tracking cc | 1 |
64,744 | 16,021,378,649 | IssuesEvent | 2021-04-21 00:14:04 | jmuelbert/jmbde-QT | https://api.github.com/repos/jmuelbert/jmbde-QT | closed | Workflow: CD: RPM - openSUSE TW | build ci dependencies github_actions no-issue-activity | ## Build the RPM is not really implemented
Here is missing the dependencies. | 1.0 | Workflow: CD: RPM - openSUSE TW - ## Build the RPM is not really implemented
Here is missing the dependencies. | non_test | workflow cd rpm opensuse tw build the rpm is not really implemented here is missing the dependencies | 0 |
318,385 | 27,300,039,562 | IssuesEvent | 2023-02-24 00:37:24 | devssa/onde-codar-em-salvador | https://api.github.com/repos/devssa/onde-codar-em-salvador | closed | [REMOTO] [JAVA] [KAFKA] [AWS] [GIT] Pessoa Desenvolvedora Java Especialista na [INVILLIA] | HOME OFFICE JAVA SPRING SQL NOSQL AWS REMOTO JENKINS KAFKA GITFLOW TESTES UNITARIOS HELP WANTED ESPECIALISTA SPRING DATA SPRING BOOT Stale | <!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Descrição da vaga
- Aproxime-se. A Invillia não apenas transformou a forma como as empresas mais revolucionárias do mundo criam e desenvolvem estratégias, negócios e produtos digitais.
- Inovou também a maneira como pessoas apaixonadas por tecnologia, de qualquer lugar do planeta, podem interagir, evoluir, mais conectados do que nunca.
- Para a Invillia, não importa onde você está. Se é um país grande. Ou uma cidade pequena. E sim a sua vontade. As suas ideias. O seu potencial.
- O tamanho do seu talento.
**Responsabilidades e atribuições:**
- O profissional será responsável em prover soluções técnicas para novas features e dar o suporte necessário as features já existentes, afinal, nem tudo são flores.
- Esperamos também que essa pessoa auxilie os outros membros do time em questões técnicas não esquecendo de fornecer a melhor solução para o negócio.
- Algo que prezamos bastante é qualidade, isso inclui um código limpo e legível (clean code).
- Também é desejável que o mesmo tenha um perfil intra-empreendedor, onde seus objetivos estejam alinhados com os objetivos da empresa, afinal, temos muito orgulho do que fazemos aqui!
## Local
- Home Office
## Benefícios
- Informações diretamente com o responsável pela vaga/recrutador.
## Requisitos
**Obrigatórios:**
- Experiência em desenvolvimento com Java;
- Definição de Arquitetura exercendo o papel de Referência Técnica;
- Experiência em desenvolvimento com Spring (Boot, Data, Cache, etc);
- Conhecimentos em Java 8 (mínimo);
- Conhecimento em Filas (Rabbit);
- Conhecimento em Kafka;
- Conhecimentos em AWS (SNS, SQS, S3);
- Conhecimentos em Git e Git-Flow;
- Experiência com bancos de dados SQL e NoSQL;
- Desenvolvimento com foco em qualidade: testes unitários e Sonar(métricas);
- Experiência em micro serviços e sistemas concorrentes;
- Contínuos delivery (Jenkins);
## Contratação
- a combinar
## Nossa empresa
- A Invillia é uma empresa global que vem revolucionando a maneira como game-changers expandem o poder de inovar, implementar tecnologias de ponta e desenvolver novas estratégias, produtos e serviços digitais.
- Nenhuma outra empresa no mundo atua como a Invillia.
- E o que torna nosso Global Growth Framework tão único e poderoso?
- Primeiro, dissolvemos os limites entre o físico e o virtual para ter em nosso time os melhores talentos do planeta.
- Criamos infinitas práticas e metodologias para que que cada squad seja super customizado e engajado na cultura e desafios de cada cliente.
- Adoramos usar ferramentas ágeis, métricas, inteligência de dados no dia-a-dia. Para que ideias e melhorias se multipliquem.
- Mas acreditamos que é na educação contínua, na abordagem mais humana e colaborativa que a mágica acontece.
- Novas oportunidades surgem. E a inovação nunca para. Infinite Digital Power.
## Como se candidatar
- [Clique aqui para se candidatar](https://invillia.gupy.io/jobs/571873?jobBoardSource=gupy_public_page)
| 1.0 | [REMOTO] [JAVA] [KAFKA] [AWS] [GIT] Pessoa Desenvolvedora Java Especialista na [INVILLIA] - <!--
==================================================
POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS!
Use: "Desenvolvedor Front-end" ao invés de
"Front-End Developer" \o/
Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]`
==================================================
-->
## Descrição da vaga
- Aproxime-se. A Invillia não apenas transformou a forma como as empresas mais revolucionárias do mundo criam e desenvolvem estratégias, negócios e produtos digitais.
- Inovou também a maneira como pessoas apaixonadas por tecnologia, de qualquer lugar do planeta, podem interagir, evoluir, mais conectados do que nunca.
- Para a Invillia, não importa onde você está. Se é um país grande. Ou uma cidade pequena. E sim a sua vontade. As suas ideias. O seu potencial.
- O tamanho do seu talento.
**Responsabilidades e atribuições:**
- O profissional será responsável em prover soluções técnicas para novas features e dar o suporte necessário as features já existentes, afinal, nem tudo são flores.
- Esperamos também que essa pessoa auxilie os outros membros do time em questões técnicas não esquecendo de fornecer a melhor solução para o negócio.
- Algo que prezamos bastante é qualidade, isso inclui um código limpo e legível (clean code).
- Também é desejável que o mesmo tenha um perfil intra-empreendedor, onde seus objetivos estejam alinhados com os objetivos da empresa, afinal, temos muito orgulho do que fazemos aqui!
## Local
- Home Office
## Benefícios
- Informações diretamente com o responsável pela vaga/recrutador.
## Requisitos
**Obrigatórios:**
- Experiência em desenvolvimento com Java;
- Definição de Arquitetura exercendo o papel de Referência Técnica;
- Experiência em desenvolvimento com Spring (Boot, Data, Cache, etc);
- Conhecimentos em Java 8 (mínimo);
- Conhecimento em Filas (Rabbit);
- Conhecimento em Kafka;
- Conhecimentos em AWS (SNS, SQS, S3);
- Conhecimentos em Git e Git-Flow;
- Experiência com bancos de dados SQL e NoSQL;
- Desenvolvimento com foco em qualidade: testes unitários e Sonar(métricas);
- Experiência em micro serviços e sistemas concorrentes;
- Contínuos delivery (Jenkins);
## Contratação
- a combinar
## Nossa empresa
- A Invillia é uma empresa global que vem revolucionando a maneira como game-changers expandem o poder de inovar, implementar tecnologias de ponta e desenvolver novas estratégias, produtos e serviços digitais.
- Nenhuma outra empresa no mundo atua como a Invillia.
- E o que torna nosso Global Growth Framework tão único e poderoso?
- Primeiro, dissolvemos os limites entre o físico e o virtual para ter em nosso time os melhores talentos do planeta.
- Criamos infinitas práticas e metodologias para que que cada squad seja super customizado e engajado na cultura e desafios de cada cliente.
- Adoramos usar ferramentas ágeis, métricas, inteligência de dados no dia-a-dia. Para que ideias e melhorias se multipliquem.
- Mas acreditamos que é na educação contínua, na abordagem mais humana e colaborativa que a mágica acontece.
- Novas oportunidades surgem. E a inovação nunca para. Infinite Digital Power.
## Como se candidatar
- [Clique aqui para se candidatar](https://invillia.gupy.io/jobs/571873?jobBoardSource=gupy_public_page)
| test | pessoa desenvolvedora java especialista na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga aproxime se a invillia não apenas transformou a forma como as empresas mais revolucionárias do mundo criam e desenvolvem estratégias negócios e produtos digitais inovou também a maneira como pessoas apaixonadas por tecnologia de qualquer lugar do planeta podem interagir evoluir mais conectados do que nunca para a invillia não importa onde você está se é um país grande ou uma cidade pequena e sim a sua vontade as suas ideias o seu potencial o tamanho do seu talento responsabilidades e atribuições o profissional será responsável em prover soluções técnicas para novas features e dar o suporte necessário as features já existentes afinal nem tudo são flores esperamos também que essa pessoa auxilie os outros membros do time em questões técnicas não esquecendo de fornecer a melhor solução para o negócio algo que prezamos bastante é qualidade isso inclui um código limpo e legível clean code também é desejável que o mesmo tenha um perfil intra empreendedor onde seus objetivos estejam alinhados com os objetivos da empresa afinal temos muito orgulho do que fazemos aqui local home office benefícios informações diretamente com o responsável pela vaga recrutador requisitos obrigatórios experiência em desenvolvimento com java definição de arquitetura exercendo o papel de referência técnica experiência em desenvolvimento com spring boot data cache etc conhecimentos em java mínimo conhecimento em filas rabbit conhecimento em kafka conhecimentos em aws sns sqs conhecimentos em git e git flow experiência com bancos de dados sql e nosql desenvolvimento com foco em qualidade testes unitários e sonar métricas experiência em micro serviços e sistemas concorrentes contínuos delivery jenkins contratação a combinar nossa empresa a invillia é uma empresa global que vem revolucionando a maneira como game changers expandem o poder de inovar implementar tecnologias de ponta e desenvolver novas estratégias produtos e serviços digitais nenhuma outra empresa no mundo atua como a invillia e o que torna nosso global growth framework tão único e poderoso primeiro dissolvemos os limites entre o físico e o virtual para ter em nosso time os melhores talentos do planeta criamos infinitas práticas e metodologias para que que cada squad seja super customizado e engajado na cultura e desafios de cada cliente adoramos usar ferramentas ágeis métricas inteligência de dados no dia a dia para que ideias e melhorias se multipliquem mas acreditamos que é na educação contínua na abordagem mais humana e colaborativa que a mágica acontece novas oportunidades surgem e a inovação nunca para infinite digital power como se candidatar | 1 |
158,802 | 24,899,512,456 | IssuesEvent | 2022-10-28 19:15:10 | vegaprotocol/vegawallet-desktop | https://api.github.com/repos/vegaprotocol/vegawallet-desktop | closed | Revoke permissions | feature desktop-wallet backend ux-and-visual-design refine | Revoking permissions between wallet and hostname should automatically shutdown the connection between these two entities, if any. | 1.0 | Revoke permissions - Revoking permissions between wallet and hostname should automatically shutdown the connection between these two entities, if any. | non_test | revoke permissions revoking permissions between wallet and hostname should automatically shutdown the connection between these two entities if any | 0 |
281,182 | 21,315,383,327 | IssuesEvent | 2022-04-16 07:15:19 | putaojuice/pe | https://api.github.com/repos/putaojuice/pe | opened | Unclear use case for sort task in DG | severity.Low type.DocumentationBug | In DG the sort task use case step 1 of MSS states that sort the task by certain property, I feel like this could have been elaborated and stated clearer because in the UG there are so many properties covered
**DG UC11**

**UG** `sort`

<!--session: 1650088126549-fc759982-4493-4e69-bd46-1702e0a9f91f-->
<!--Version: Web v3.4.2--> | 1.0 | Unclear use case for sort task in DG - In DG the sort task use case step 1 of MSS states that sort the task by certain property, I feel like this could have been elaborated and stated clearer because in the UG there are so many properties covered
**DG UC11**

**UG** `sort`

<!--session: 1650088126549-fc759982-4493-4e69-bd46-1702e0a9f91f-->
<!--Version: Web v3.4.2--> | non_test | unclear use case for sort task in dg in dg the sort task use case step of mss states that sort the task by certain property i feel like this could have been elaborated and stated clearer because in the ug there are so many properties covered dg ug sort | 0 |
252,178 | 21,561,161,074 | IssuesEvent | 2022-05-01 07:10:34 | prestodb/presto | https://api.github.com/repos/prestodb/presto | closed | Flaky TestJdbcWarnings.testLongRunningStatement | tests stale | Build failed with
```
[ERROR] testLongRunningStatement(com.facebook.presto.jdbc.TestJdbcWarnings) Time elapsed: 0.172 s <<< FAILURE!
java.lang.NullPointerException: throwable is null
at java.util.Objects.requireNonNull(Objects.java:228)
at com.facebook.presto.jdbc.TestJdbcWarnings$WarningEntry.<init>(TestJdbcWarnings.java:305)
at com.facebook.presto.jdbc.TestJdbcWarnings.testLongRunningStatement(TestJdbcWarnings.java:150)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:104)
at org.testng.internal.Invoker.invokeMethod(Invoker.java:645)
at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:851)
at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1177)
at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:129)
at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:112)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
```
The relevant part of the test is
```
while (statement.getWarnings() == null) {
Thread.sleep(100);
}
SQLWarning warning = statement.getWarnings();
Set<WarningEntry> currentWarnings = new HashSet<>();
assertTrue(currentWarnings.add(new WarningEntry(warning)));
```
It seems like somehow the warnings are getting cleared or something between the two getWarnings() calls so that they are null at the second call. | 1.0 | Flaky TestJdbcWarnings.testLongRunningStatement - Build failed with
```
[ERROR] testLongRunningStatement(com.facebook.presto.jdbc.TestJdbcWarnings) Time elapsed: 0.172 s <<< FAILURE!
java.lang.NullPointerException: throwable is null
at java.util.Objects.requireNonNull(Objects.java:228)
at com.facebook.presto.jdbc.TestJdbcWarnings$WarningEntry.<init>(TestJdbcWarnings.java:305)
at com.facebook.presto.jdbc.TestJdbcWarnings.testLongRunningStatement(TestJdbcWarnings.java:150)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:104)
at org.testng.internal.Invoker.invokeMethod(Invoker.java:645)
at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:851)
at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1177)
at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:129)
at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:112)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
```
The relevant part of the test is
```
while (statement.getWarnings() == null) {
Thread.sleep(100);
}
SQLWarning warning = statement.getWarnings();
Set<WarningEntry> currentWarnings = new HashSet<>();
assertTrue(currentWarnings.add(new WarningEntry(warning)));
```
It seems like somehow the warnings are getting cleared or something between the two getWarnings() calls so that they are null at the second call. | test | flaky testjdbcwarnings testlongrunningstatement build failed with testlongrunningstatement com facebook presto jdbc testjdbcwarnings time elapsed s failure java lang nullpointerexception throwable is null at java util objects requirenonnull objects java at com facebook presto jdbc testjdbcwarnings warningentry testjdbcwarnings java at com facebook presto jdbc testjdbcwarnings testlongrunningstatement testjdbcwarnings java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org testng internal methodinvocationhelper invokemethod methodinvocationhelper java at org testng internal invoker invokemethod invoker java at org testng internal invoker invoketestmethod invoker java at org testng internal invoker invoketestmethods invoker java at org testng internal testmethodworker invoketestmethods testmethodworker java at org testng internal testmethodworker run testmethodworker java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java the relevant part of the test is while statement getwarnings null thread sleep sqlwarning warning statement getwarnings set currentwarnings new hashset asserttrue currentwarnings add new warningentry warning it seems like somehow the warnings are getting cleared or something between the two getwarnings calls so that they are null at the second call | 1 |
168,250 | 13,067,467,057 | IssuesEvent | 2020-07-31 00:32:53 | ayumi-cloud/oc-security-module | https://api.github.com/repos/ayumi-cloud/oc-security-module | closed | Add AppEngine detector to the Google firewall module | Add to Blacklist Add to Whitelist FINSIHED Firewall Priority: Medium Testing - Passed enhancement | ### Enhancement idea
- [x] Add `AppEngine` detector to the Google firewall module.
Google App Engine is a Platform as a Service and cloud computing platform for developing and hosting web applications in Google-managed data centers. Applications are sandboxed and run across multiple servers.
| 1.0 | Add AppEngine detector to the Google firewall module - ### Enhancement idea
- [x] Add `AppEngine` detector to the Google firewall module.
Google App Engine is a Platform as a Service and cloud computing platform for developing and hosting web applications in Google-managed data centers. Applications are sandboxed and run across multiple servers.
| test | add appengine detector to the google firewall module enhancement idea add appengine detector to the google firewall module google app engine is a platform as a service and cloud computing platform for developing and hosting web applications in google managed data centers applications are sandboxed and run across multiple servers | 1 |
8,373 | 8,272,528,201 | IssuesEvent | 2018-09-16 21:07:49 | javaee/glassfish | https://api.github.com/repos/javaee/glassfish | closed | PostConstruct called twice for web service | Component: web_services ERR: Assignee Priority: Minor Type: Bug | Simple annotated web service with public constructor, private @PostConstruct and @PreDestroy methods, and a single web service method.
For each call to the web service method, I see a new service object created and two calls to the PostConstruct method, e.g.
INFO: WEB0671: Loading application [PostConstructIssue] at [/PostConstructIssue]
INFO: PostConstructIssue was successfully deployed in 959 milliseconds.
INFO: parsing WSDL...
INFO: Generating code...
INFO: Compiling code...
INFO: Invoking wsimport with [http://localhost:8080/PostConstructIssue/MyWebServiceService?WSDL](http://localhost:8080/PostConstructIssue/MyWebServiceService?WSDL)
INFO: wsimport successful
INFO: webapp.postconstructissue.MyWebService@d4d7db constructor
INFO: webapp.postconstructissue.MyWebService@d4d7db @PostConstruct init
INFO: webapp.postconstructissue.MyWebService@d4d7db @PostConstruct init
INFO: webapp.postconstructissue.MyWebService@c8d0e constructor
INFO: webapp.postconstructissue.MyWebService@c8d0e @PostConstruct init
INFO: webapp.postconstructissue.MyWebService@c8d0e @PostConstruct init
Example Maven project attached.
Aside from the double-call to the @PostConstruct method, I notice each webservice method call results in a new service object being created. This is different from GlassFish 2.1 behaviour - is it expected, and if so, is the old behaviour configurable?
#### Environment
GlassFish 3.1, Windows Vista
#### Affected Versions
[3.1] | 1.0 | PostConstruct called twice for web service - Simple annotated web service with public constructor, private @PostConstruct and @PreDestroy methods, and a single web service method.
For each call to the web service method, I see a new service object created and two calls to the PostConstruct method, e.g.
INFO: WEB0671: Loading application [PostConstructIssue] at [/PostConstructIssue]
INFO: PostConstructIssue was successfully deployed in 959 milliseconds.
INFO: parsing WSDL...
INFO: Generating code...
INFO: Compiling code...
INFO: Invoking wsimport with [http://localhost:8080/PostConstructIssue/MyWebServiceService?WSDL](http://localhost:8080/PostConstructIssue/MyWebServiceService?WSDL)
INFO: wsimport successful
INFO: webapp.postconstructissue.MyWebService@d4d7db constructor
INFO: webapp.postconstructissue.MyWebService@d4d7db @PostConstruct init
INFO: webapp.postconstructissue.MyWebService@d4d7db @PostConstruct init
INFO: webapp.postconstructissue.MyWebService@c8d0e constructor
INFO: webapp.postconstructissue.MyWebService@c8d0e @PostConstruct init
INFO: webapp.postconstructissue.MyWebService@c8d0e @PostConstruct init
Example Maven project attached.
Aside from the double-call to the @PostConstruct method, I notice each webservice method call results in a new service object being created. This is different from GlassFish 2.1 behaviour - is it expected, and if so, is the old behaviour configurable?
#### Environment
GlassFish 3.1, Windows Vista
#### Affected Versions
[3.1] | non_test | postconstruct called twice for web service simple annotated web service with public constructor private postconstruct and predestroy methods and a single web service method for each call to the web service method i see a new service object created and two calls to the postconstruct method e g info loading application at info postconstructissue was successfully deployed in milliseconds info parsing wsdl info generating code info compiling code info invoking wsimport with info wsimport successful info webapp postconstructissue mywebservice constructor info webapp postconstructissue mywebservice postconstruct init info webapp postconstructissue mywebservice postconstruct init info webapp postconstructissue mywebservice constructor info webapp postconstructissue mywebservice postconstruct init info webapp postconstructissue mywebservice postconstruct init example maven project attached aside from the double call to the postconstruct method i notice each webservice method call results in a new service object being created this is different from glassfish behaviour is it expected and if so is the old behaviour configurable environment glassfish windows vista affected versions | 0 |
10,492 | 8,584,213,515 | IssuesEvent | 2018-11-13 22:02:43 | nest/nest-simulator | https://api.github.com/repos/nest/nest-simulator | closed | Are you interested in `pip install pynest` ? | C: Infrastructure I: No breaking change P: In progess S: Low T: Discussion | After working on #844, I was wondering, if I could decouple NEST and PyNEST even further and had some interesting hacking hours. Now, I have a [prove of concept](https://github.com/tammoippen/nest-simulator/tree/pynest) to make NEST and PyNEST two independent projects, with NEST completely dependancy-free from PyNEST or Python installations and PyNEST installs with setup.py using `nest-config`. This gives:
* No python / cython discovery and handling in CMakefiles, except for `docs` generation.
* NEST does not need to link to the python lib, not even with `conngen` enabled.
* `pip install nest` ! One could think about deploying PyNEST to http://pypi.python.org/ ?
* Parallel installation of PyNEST for Python2 and Python3 with only one installation of NEST.
* Possibly have PyNEST in a separate repository?
* pytest (unrelated)
On the other hand, this would require to install NEST and PyNEST separately, something like:
```sh
cmake ...
make
make install
pip install pynest/
```
And two build chains have to be maintained!
There are still some TODO's regarding documentation and testing, and maybe some questions regarding the design decisions. I would like to know, whether there is interest in this refactoring and if i should continue? Addressing the missing parts will require some effort on my side and I am unsure if I should invest. I am happy to do so, if there is consent.
What changed in detail so far:
* Remove most Python/Cython related stuff from CMakefiles
* Extend `pynest/setup.py` to compile the cython extension, build and install PyNEST. `pip install pynest/` will install pynest to default python `site-packages` directories.
* Put the python extension from the `topology` module also into PyNEST. Right now, it is not possible, to install nest without topology, hence the python extension will also always be installed. To reflect this in this changes, I moved the topology python files.
* Remove `cythonize-pynest` option from CMakefiles: the setup.py excepts options to use rebuild python extensions. Crosscompiling still has to be tested.
* Remove libneurosim / python parts from `neststartup.h/.cpp` and put them into a new header file in PyNEST: `pynestkernel.h`
* move `pynestkernel.pyx/pxd` into the `nest` folder of pynest.
* install python examples twice: once for nest / once for the pynest installation.
* have `do_tests.sh.in` discover, whether there is a pynest installation and do tests, if they are found. (Use `pytest` instead of `nose` see #761. One could also think about putting that part into a separate PR; accept both `nose` or `pytest`; stick to `pytest` and also remove `do_tests.py.in` support altogether.)
* there is also `extras/ConnPlotter/` python package ... I did not put this into pynest, because I am not sure, how both relate. Maybe this could also be a separate package / project / repository?
* changed travis ci to reflect the install procedures. TravisCI works, but finds one python test failing ... I have to investigate: https://travis-ci.org/tammoippen/nest-simulator/builds/298469055
* The code is still not PR ready, has some debug stuff, and a ugly git history...
TODOs:
* Documentation.
* Code cleanup.
* Pytest fixing.
* git history fixing.
* cross compiling testing and documenting. see http://spscicomp.org/wordpress/wp-content/uploads/2014/05/BertrandBrelier_PythonBGQ.pdf slide 17ff)
* unforeseeables ...
| 1.0 | Are you interested in `pip install pynest` ? - After working on #844, I was wondering, if I could decouple NEST and PyNEST even further and had some interesting hacking hours. Now, I have a [prove of concept](https://github.com/tammoippen/nest-simulator/tree/pynest) to make NEST and PyNEST two independent projects, with NEST completely dependancy-free from PyNEST or Python installations and PyNEST installs with setup.py using `nest-config`. This gives:
* No python / cython discovery and handling in CMakefiles, except for `docs` generation.
* NEST does not need to link to the python lib, not even with `conngen` enabled.
* `pip install nest` ! One could think about deploying PyNEST to http://pypi.python.org/ ?
* Parallel installation of PyNEST for Python2 and Python3 with only one installation of NEST.
* Possibly have PyNEST in a separate repository?
* pytest (unrelated)
On the other hand, this would require to install NEST and PyNEST separately, something like:
```sh
cmake ...
make
make install
pip install pynest/
```
And two build chains have to be maintained!
There are still some TODO's regarding documentation and testing, and maybe some questions regarding the design decisions. I would like to know, whether there is interest in this refactoring and if i should continue? Addressing the missing parts will require some effort on my side and I am unsure if I should invest. I am happy to do so, if there is consent.
What changed in detail so far:
* Remove most Python/Cython related stuff from CMakefiles
* Extend `pynest/setup.py` to compile the cython extension, build and install PyNEST. `pip install pynest/` will install pynest to default python `site-packages` directories.
* Put the python extension from the `topology` module also into PyNEST. Right now, it is not possible, to install nest without topology, hence the python extension will also always be installed. To reflect this in this changes, I moved the topology python files.
* Remove `cythonize-pynest` option from CMakefiles: the setup.py excepts options to use rebuild python extensions. Crosscompiling still has to be tested.
* Remove libneurosim / python parts from `neststartup.h/.cpp` and put them into a new header file in PyNEST: `pynestkernel.h`
* move `pynestkernel.pyx/pxd` into the `nest` folder of pynest.
* install python examples twice: once for nest / once for the pynest installation.
* have `do_tests.sh.in` discover, whether there is a pynest installation and do tests, if they are found. (Use `pytest` instead of `nose` see #761. One could also think about putting that part into a separate PR; accept both `nose` or `pytest`; stick to `pytest` and also remove `do_tests.py.in` support altogether.)
* there is also `extras/ConnPlotter/` python package ... I did not put this into pynest, because I am not sure, how both relate. Maybe this could also be a separate package / project / repository?
* changed travis ci to reflect the install procedures. TravisCI works, but finds one python test failing ... I have to investigate: https://travis-ci.org/tammoippen/nest-simulator/builds/298469055
* The code is still not PR ready, has some debug stuff, and a ugly git history...
TODOs:
* Documentation.
* Code cleanup.
* Pytest fixing.
* git history fixing.
* cross compiling testing and documenting. see http://spscicomp.org/wordpress/wp-content/uploads/2014/05/BertrandBrelier_PythonBGQ.pdf slide 17ff)
* unforeseeables ...
| non_test | are you interested in pip install pynest after working on i was wondering if i could decouple nest and pynest even further and had some interesting hacking hours now i have a to make nest and pynest two independent projects with nest completely dependancy free from pynest or python installations and pynest installs with setup py using nest config this gives no python cython discovery and handling in cmakefiles except for docs generation nest does not need to link to the python lib not even with conngen enabled pip install nest one could think about deploying pynest to parallel installation of pynest for and with only one installation of nest possibly have pynest in a separate repository pytest unrelated on the other hand this would require to install nest and pynest separately something like sh cmake make make install pip install pynest and two build chains have to be maintained there are still some todo s regarding documentation and testing and maybe some questions regarding the design decisions i would like to know whether there is interest in this refactoring and if i should continue addressing the missing parts will require some effort on my side and i am unsure if i should invest i am happy to do so if there is consent what changed in detail so far remove most python cython related stuff from cmakefiles extend pynest setup py to compile the cython extension build and install pynest pip install pynest will install pynest to default python site packages directories put the python extension from the topology module also into pynest right now it is not possible to install nest without topology hence the python extension will also always be installed to reflect this in this changes i moved the topology python files remove cythonize pynest option from cmakefiles the setup py excepts options to use rebuild python extensions crosscompiling still has to be tested remove libneurosim python parts from neststartup h cpp and put them into a new header file in pynest pynestkernel h move pynestkernel pyx pxd into the nest folder of pynest install python examples twice once for nest once for the pynest installation have do tests sh in discover whether there is a pynest installation and do tests if they are found use pytest instead of nose see one could also think about putting that part into a separate pr accept both nose or pytest stick to pytest and also remove do tests py in support altogether there is also extras connplotter python package i did not put this into pynest because i am not sure how both relate maybe this could also be a separate package project repository changed travis ci to reflect the install procedures travisci works but finds one python test failing i have to investigate the code is still not pr ready has some debug stuff and a ugly git history todos documentation code cleanup pytest fixing git history fixing cross compiling testing and documenting see slide unforeseeables | 0 |
126,423 | 17,030,062,020 | IssuesEvent | 2021-07-04 11:27:54 | alan-turing-institute/sktime | https://api.github.com/repos/alan-turing-institute/sktime | reopened | Benchmarking design & implementation | API design implementing framework | ## Evaluation
- [ ] Extend evaluation API and functionality to single dataset case, currently only multiple dataset case is supported
## Performance metrics
A few implementation notes:
- Vectorised vs iterative computations
- Callable classes vs classes with methods for computation
- Use of jackknife by default for non point-wise metrics
- Computation of standard error as decorator/mix-in
- Have separate classes for point-wise metrics which can be wrapped by aggregation functions (e.g. mean)
Also see https://github.com/JuliaML/LossFunctions.jl.
## Orchestration
### Should have
- [ ] Allow orchestrator to be persisted to replicate benchmarking studies
- [ ] add unit tests for `evaluator` methods
- [ ] update all methods on evaluator to work on new internal data representation, also see https://www.statsmodels.org/stable/stats.html for some additional test implementations, e.g. the sign test, to improve readability, so that we can deprecate `_get_metrics_per_estimator_dataset` and `_get_metrics_per_estimator` methods
- [ ] for saving results inside the `orchestrator` and for loading results in results classes use `_ResultsWrapper` to simply/unify interface, `_ResultsWrapper` needs to have slots for at least: y_true, y_pred, y_proba, index, fit_time, predict_time, strategy_name, dataset_name, cv_fold, train_or_test
- [ ] No timing of fit and predict available, see https://docs.python.org/3/library/time.html#time.perf_counter, potentially have new `save_timings` and `load_timings` method
- [ ] `orchestrator` cannot make probabilistic predictions, orchestrator tries to make probabilistic predictions using `predict_proba`, but (i) this will only works for some but not all classifiers and it won't work in regression, (ii) strategies currently don't even have a `predict_proba` (not even `TSCStrategy`), and (iii) current computation of `y_proba` fails if `y_pred` contains strings instead of integers which however is an accepted output format for classification I believe, add `predict_proba` to `TSCStrategy`
- [ ] handling of probabilistic metrics in `evaluator`
- [ ] no longer sure that saving results object as a master file is a good idea, as it may cause problems when multiple processes try to update it and because it needs to reflect the state of the directory somehow, maybe better to have a method on results object that allow to infer datasets, strategies and so on, something like a `register_results` method, instead of loading a fully specified dumped result object
- [ ] separate `predict` method on `orchestrator` which loads and uses already fitted strategies
- [ ] fix UEA results class
### Could have
- [ ] allow for pre-defined cv splits in files
- [ ] allow for pre-defined tasks in files
- [ ] add `random_state` as input arg to orchestrator which is propagated to all strategies and cv
- [ ] perhaps also useful to catch exceptions and skip over them in `orchestrator` instead of breaking execution?
- [ ] currently only works for ts data input format, add other use cases
- [ ] better user feedback, logging, keeping track of progress
- [ ] many docstrings still missing or outdated
- [ ] perhaps metrics shouldn't be wrapped in classes and the evaluator should take care of it internally, working with kwargs (e.g. `pointwise=True`)
- [ ] handling of multiple metrics in `evaluator`
- [ ] functionality for space filling parameter grids for large hyper-parameter search spaces (e.g. latin hypercube design), see this Python package: https://github.com/tirthajyoti/doepy
- [ ] monitoring and comparison of memory usage of different estimators
Related issues/PRs: #132 | 1.0 | Benchmarking design & implementation - ## Evaluation
- [ ] Extend evaluation API and functionality to single dataset case, currently only multiple dataset case is supported
## Performance metrics
A few implementation notes:
- Vectorised vs iterative computations
- Callable classes vs classes with methods for computation
- Use of jackknife by default for non point-wise metrics
- Computation of standard error as decorator/mix-in
- Have separate classes for point-wise metrics which can be wrapped by aggregation functions (e.g. mean)
Also see https://github.com/JuliaML/LossFunctions.jl.
## Orchestration
### Should have
- [ ] Allow orchestrator to be persisted to replicate benchmarking studies
- [ ] add unit tests for `evaluator` methods
- [ ] update all methods on evaluator to work on new internal data representation, also see https://www.statsmodels.org/stable/stats.html for some additional test implementations, e.g. the sign test, to improve readability, so that we can deprecate `_get_metrics_per_estimator_dataset` and `_get_metrics_per_estimator` methods
- [ ] for saving results inside the `orchestrator` and for loading results in results classes use `_ResultsWrapper` to simply/unify interface, `_ResultsWrapper` needs to have slots for at least: y_true, y_pred, y_proba, index, fit_time, predict_time, strategy_name, dataset_name, cv_fold, train_or_test
- [ ] No timing of fit and predict available, see https://docs.python.org/3/library/time.html#time.perf_counter, potentially have new `save_timings` and `load_timings` method
- [ ] `orchestrator` cannot make probabilistic predictions, orchestrator tries to make probabilistic predictions using `predict_proba`, but (i) this will only works for some but not all classifiers and it won't work in regression, (ii) strategies currently don't even have a `predict_proba` (not even `TSCStrategy`), and (iii) current computation of `y_proba` fails if `y_pred` contains strings instead of integers which however is an accepted output format for classification I believe, add `predict_proba` to `TSCStrategy`
- [ ] handling of probabilistic metrics in `evaluator`
- [ ] no longer sure that saving results object as a master file is a good idea, as it may cause problems when multiple processes try to update it and because it needs to reflect the state of the directory somehow, maybe better to have a method on results object that allow to infer datasets, strategies and so on, something like a `register_results` method, instead of loading a fully specified dumped result object
- [ ] separate `predict` method on `orchestrator` which loads and uses already fitted strategies
- [ ] fix UEA results class
### Could have
- [ ] allow for pre-defined cv splits in files
- [ ] allow for pre-defined tasks in files
- [ ] add `random_state` as input arg to orchestrator which is propagated to all strategies and cv
- [ ] perhaps also useful to catch exceptions and skip over them in `orchestrator` instead of breaking execution?
- [ ] currently only works for ts data input format, add other use cases
- [ ] better user feedback, logging, keeping track of progress
- [ ] many docstrings still missing or outdated
- [ ] perhaps metrics shouldn't be wrapped in classes and the evaluator should take care of it internally, working with kwargs (e.g. `pointwise=True`)
- [ ] handling of multiple metrics in `evaluator`
- [ ] functionality for space filling parameter grids for large hyper-parameter search spaces (e.g. latin hypercube design), see this Python package: https://github.com/tirthajyoti/doepy
- [ ] monitoring and comparison of memory usage of different estimators
Related issues/PRs: #132 | non_test | benchmarking design implementation evaluation extend evaluation api and functionality to single dataset case currently only multiple dataset case is supported performance metrics a few implementation notes vectorised vs iterative computations callable classes vs classes with methods for computation use of jackknife by default for non point wise metrics computation of standard error as decorator mix in have separate classes for point wise metrics which can be wrapped by aggregation functions e g mean also see orchestration should have allow orchestrator to be persisted to replicate benchmarking studies add unit tests for evaluator methods update all methods on evaluator to work on new internal data representation also see for some additional test implementations e g the sign test to improve readability so that we can deprecate get metrics per estimator dataset and get metrics per estimator methods for saving results inside the orchestrator and for loading results in results classes use resultswrapper to simply unify interface resultswrapper needs to have slots for at least y true y pred y proba index fit time predict time strategy name dataset name cv fold train or test no timing of fit and predict available see potentially have new save timings and load timings method orchestrator cannot make probabilistic predictions orchestrator tries to make probabilistic predictions using predict proba but i this will only works for some but not all classifiers and it won t work in regression ii strategies currently don t even have a predict proba not even tscstrategy and iii current computation of y proba fails if y pred contains strings instead of integers which however is an accepted output format for classification i believe add predict proba to tscstrategy handling of probabilistic metrics in evaluator no longer sure that saving results object as a master file is a good idea as it may cause problems when multiple processes try to update it and because it needs to reflect the state of the directory somehow maybe better to have a method on results object that allow to infer datasets strategies and so on something like a register results method instead of loading a fully specified dumped result object separate predict method on orchestrator which loads and uses already fitted strategies fix uea results class could have allow for pre defined cv splits in files allow for pre defined tasks in files add random state as input arg to orchestrator which is propagated to all strategies and cv perhaps also useful to catch exceptions and skip over them in orchestrator instead of breaking execution currently only works for ts data input format add other use cases better user feedback logging keeping track of progress many docstrings still missing or outdated perhaps metrics shouldn t be wrapped in classes and the evaluator should take care of it internally working with kwargs e g pointwise true handling of multiple metrics in evaluator functionality for space filling parameter grids for large hyper parameter search spaces e g latin hypercube design see this python package monitoring and comparison of memory usage of different estimators related issues prs | 0 |
304,900 | 26,344,792,989 | IssuesEvent | 2023-01-10 20:57:58 | jmarou/EventMapping | https://api.github.com/repos/jmarou/EventMapping | closed | Method using NLP: Words of interest using python library for NLP | A-feature B-Backend D-High C-development C-test | Develop method for extraction of words relating to location based on NLP
Use spacy | 1.0 | Method using NLP: Words of interest using python library for NLP - Develop method for extraction of words relating to location based on NLP
Use spacy | test | method using nlp words of interest using python library for nlp develop method for extraction of words relating to location based on nlp use spacy | 1 |
298,752 | 9,200,871,202 | IssuesEvent | 2019-03-07 18:07:54 | Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth | closed | [LOCALIZATION] | INVADER_INFAMY | :beetle: bug - localization :scroll: :exclamation: priority high | **Mod Version**
Master branch
**Please explain your issue in as much detail as possible:**
No loc for INVADER_INFAMY threat modifier
**Upload screenshots of the problem localization:**
<details>
<summary>Click to expand</summary>

</details> | 1.0 | [LOCALIZATION] | INVADER_INFAMY - **Mod Version**
Master branch
**Please explain your issue in as much detail as possible:**
No loc for INVADER_INFAMY threat modifier
**Upload screenshots of the problem localization:**
<details>
<summary>Click to expand</summary>

</details> | non_test | invader infamy mod version master branch please explain your issue in as much detail as possible no loc for invader infamy threat modifier upload screenshots of the problem localization click to expand | 0 |
328,086 | 28,100,777,339 | IssuesEvent | 2023-03-30 19:18:18 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: schemachange/mixed-versions-compat failed | C-test-failure O-robot O-roachtest branch-master release-blocker | roachtest.schemachange/mixed-versions-compat [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=artifacts#/schemachange/mixed-versions-compat) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c):
```
test artifacts and logs in: /artifacts/schemachange/mixed-versions-compat/run_1
(mixed_version_decl_schemachange_compat.go:83).fetchCorpusToTmpDir: Missing validation corpus for release-22.2 (output in run_191732.317448944_n1_gsutil-cp-gscockroac: gsutil cp gs://cockroach-corpus/corpus-release-22.2/corpus /tmp/corpus3626686228/corpus returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_191732.322819992_n1_gsutil-cp-gscockroac.log: exit status 1)
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #100164 roachtest: schemachange/mixed-versions-compat failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-23.1 release-blocker]
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions-compat.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: schemachange/mixed-versions-compat failed - roachtest.schemachange/mixed-versions-compat [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=artifacts#/schemachange/mixed-versions-compat) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c):
```
test artifacts and logs in: /artifacts/schemachange/mixed-versions-compat/run_1
(mixed_version_decl_schemachange_compat.go:83).fetchCorpusToTmpDir: Missing validation corpus for release-22.2 (output in run_191732.317448944_n1_gsutil-cp-gscockroac: gsutil cp gs://cockroach-corpus/corpus-release-22.2/corpus /tmp/corpus3626686228/corpus returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_191732.322819992_n1_gsutil-cp-gscockroac.log: exit status 1)
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
<details><summary>Same failure on other branches</summary>
<p>
- #100164 roachtest: schemachange/mixed-versions-compat failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-23.1 release-blocker]
</p>
</details>
/cc @cockroachdb/sql-schema
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions-compat.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| test | roachtest schemachange mixed versions compat failed roachtest schemachange mixed versions compat with on master test artifacts and logs in artifacts schemachange mixed versions compat run mixed version decl schemachange compat go fetchcorpustotmpdir missing validation corpus for release output in run gsutil cp gscockroac gsutil cp gs cockroach corpus corpus release corpus tmp corpus returned command problem ssh verbose log retained in ssh gsutil cp gscockroac log exit status parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest schemachange mixed versions compat failed cc cockroachdb sql schema | 1 |
374,968 | 26,141,364,963 | IssuesEvent | 2022-12-29 19:00:20 | sukesh-ak/ESP32-TUX | https://api.github.com/repos/sukesh-ak/ESP32-TUX | closed | ESP-IDF compiler error - need docs improvements. | documentation question | I'm using Windows and also mac. There are huge gabs in the readme on what to do and it seems that it was written for a veteran of ESP32 developers.
Right now I'm having trouble with the following steps:
`idf.py set-target esp32s3`
`idf.py build`
`idf.py flash monitor`
Which produce errors like the following:
```
ZacharysMBPM16:ESP32-TUX niteris$ idf.py flash monitor
Executing action: flash
Running cmake in directory /Users/niteris/dev/ESP32-TUX/build
Executing "cmake -G Ninja -DPYTHON_DEPS_CHECKED=1 -DPYTHON=/Users/niteris/.espressif/python_env/idf5.1_py3.10_env/bin/python -DESP_PLATFORM=1 -DCCACHE_ENABLE=0 /Users/niteris/dev/ESP32-TUX"...
-- IDF_TARGET not set, using default target: esp32
-- Found Git: /usr/bin/git (found version "2.37.1 (Apple Git-137.1)")
-- Component directory /Users/niteris/dev/ESP32-TUX/components/ConfigBase does not contain a CMakeLists.txt file. No component will be added
-- The C compiler identification is GNU 11.2.0
-- The CXX compiler identification is GNU 11.2.0
-- The ASM compiler identification is GNU
-- Found assembler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-gcc
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-gcc - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-g++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Building ESP-IDF components for target esp32
Processing 2 dependencies:
[1/2] espressif/fmt (9.1.0)
[2/2] idf (5.1.0)
CMake Error at /Users/niteris/esp/esp-idf/tools/cmake/build.cmake:542 (message):
ERROR:
Some components (espressif/fmt) in the "managed_components" directory were modified on the disk since the last run of the CMake. Content of this directory is managed automatically.
If you want to keep the changes, you can move the directory with the component to the "components" directory of your project.
I.E. for "espressif__fmt" run:
mv /Users/niteris/dev/ESP32-TUX/managed_components/espressif__fmt /Users/niteris/dev/ESP32-TUX/components/espressif__fmt
Or, if you want to discard the changes remove the ".component_hash" file from the component's directory.
I.E. for "espressif__fmt" run:
rm /Users/niteris/dev/ESP32-TUX/managed_components/espressif__fmt/.component_hash
Call Stack (most recent call first):
/Users/niteris/esp/esp-idf/tools/cmake/project.cmake:447 (idf_build_process)
CMakeLists.txt:28 (project)
-- Configuring incomplete, errors occurred!
See also "/Users/niteris/dev/ESP32-TUX/build/CMakeFiles/CMakeOutput.log".
cmake failed with exit code 1, output of the command is in the /Users/niteris/dev/ESP32-TUX/build/log/idf_py_stderr_output_39505 and /Users/niteris/dev/ESP32-TUX/build/log/idf_py_stdout_output_39505
``` | 1.0 | ESP-IDF compiler error - need docs improvements. - I'm using Windows and also mac. There are huge gabs in the readme on what to do and it seems that it was written for a veteran of ESP32 developers.
Right now I'm having trouble with the following steps:
`idf.py set-target esp32s3`
`idf.py build`
`idf.py flash monitor`
Which produce errors like the following:
```
ZacharysMBPM16:ESP32-TUX niteris$ idf.py flash monitor
Executing action: flash
Running cmake in directory /Users/niteris/dev/ESP32-TUX/build
Executing "cmake -G Ninja -DPYTHON_DEPS_CHECKED=1 -DPYTHON=/Users/niteris/.espressif/python_env/idf5.1_py3.10_env/bin/python -DESP_PLATFORM=1 -DCCACHE_ENABLE=0 /Users/niteris/dev/ESP32-TUX"...
-- IDF_TARGET not set, using default target: esp32
-- Found Git: /usr/bin/git (found version "2.37.1 (Apple Git-137.1)")
-- Component directory /Users/niteris/dev/ESP32-TUX/components/ConfigBase does not contain a CMakeLists.txt file. No component will be added
-- The C compiler identification is GNU 11.2.0
-- The CXX compiler identification is GNU 11.2.0
-- The ASM compiler identification is GNU
-- Found assembler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-gcc
-- Detecting C compiler ABI info
-- Detecting C compiler ABI info - done
-- Check for working C compiler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-gcc - skipped
-- Detecting C compile features
-- Detecting C compile features - done
-- Detecting CXX compiler ABI info
-- Detecting CXX compiler ABI info - done
-- Check for working CXX compiler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-g++ - skipped
-- Detecting CXX compile features
-- Detecting CXX compile features - done
-- Building ESP-IDF components for target esp32
Processing 2 dependencies:
[1/2] espressif/fmt (9.1.0)
[2/2] idf (5.1.0)
CMake Error at /Users/niteris/esp/esp-idf/tools/cmake/build.cmake:542 (message):
ERROR:
Some components (espressif/fmt) in the "managed_components" directory were modified on the disk since the last run of the CMake. Content of this directory is managed automatically.
If you want to keep the changes, you can move the directory with the component to the "components" directory of your project.
I.E. for "espressif__fmt" run:
mv /Users/niteris/dev/ESP32-TUX/managed_components/espressif__fmt /Users/niteris/dev/ESP32-TUX/components/espressif__fmt
Or, if you want to discard the changes remove the ".component_hash" file from the component's directory.
I.E. for "espressif__fmt" run:
rm /Users/niteris/dev/ESP32-TUX/managed_components/espressif__fmt/.component_hash
Call Stack (most recent call first):
/Users/niteris/esp/esp-idf/tools/cmake/project.cmake:447 (idf_build_process)
CMakeLists.txt:28 (project)
-- Configuring incomplete, errors occurred!
See also "/Users/niteris/dev/ESP32-TUX/build/CMakeFiles/CMakeOutput.log".
cmake failed with exit code 1, output of the command is in the /Users/niteris/dev/ESP32-TUX/build/log/idf_py_stderr_output_39505 and /Users/niteris/dev/ESP32-TUX/build/log/idf_py_stdout_output_39505
``` | non_test | esp idf compiler error need docs improvements i m using windows and also mac there are huge gabs in the readme on what to do and it seems that it was written for a veteran of developers right now i m having trouble with the following steps idf py set target idf py build idf py flash monitor which produce errors like the following tux niteris idf py flash monitor executing action flash running cmake in directory users niteris dev tux build executing cmake g ninja dpython deps checked dpython users niteris espressif python env env bin python desp platform dccache enable users niteris dev tux idf target not set using default target found git usr bin git found version apple git component directory users niteris dev tux components configbase does not contain a cmakelists txt file no component will be added the c compiler identification is gnu the cxx compiler identification is gnu the asm compiler identification is gnu found assembler users niteris espressif tools xtensa elf esp xtensa elf bin xtensa elf gcc detecting c compiler abi info detecting c compiler abi info done check for working c compiler users niteris espressif tools xtensa elf esp xtensa elf bin xtensa elf gcc skipped detecting c compile features detecting c compile features done detecting cxx compiler abi info detecting cxx compiler abi info done check for working cxx compiler users niteris espressif tools xtensa elf esp xtensa elf bin xtensa elf g skipped detecting cxx compile features detecting cxx compile features done building esp idf components for target processing dependencies espressif fmt idf cmake error at users niteris esp esp idf tools cmake build cmake message error some components espressif fmt in the managed components directory were modified on the disk since the last run of the cmake content of this directory is managed automatically if you want to keep the changes you can move the directory with the component to the components directory of your project i e for espressif fmt run mv users niteris dev tux managed components espressif fmt users niteris dev tux components espressif fmt or if you want to discard the changes remove the component hash file from the component s directory i e for espressif fmt run rm users niteris dev tux managed components espressif fmt component hash call stack most recent call first users niteris esp esp idf tools cmake project cmake idf build process cmakelists txt project configuring incomplete errors occurred see also users niteris dev tux build cmakefiles cmakeoutput log cmake failed with exit code output of the command is in the users niteris dev tux build log idf py stderr output and users niteris dev tux build log idf py stdout output | 0 |
10,764 | 7,302,159,490 | IssuesEvent | 2018-02-27 08:45:57 | avocode/avocode | https://api.github.com/repos/avocode/avocode | closed | App is not loading on Windows 10 | needsMoreInfo performance waitingForUser windows | <!-- Which version of avocode are you using? -->
<!-- Please make sure you are using the latest version. -->
<!-- You can always check and download the latest version from https://avocode.com/changelog.html. -->
* **Version**: 2.14.8
<!-- Are you using the web or desktop version of Avocode? -->
* **Web/Desktop**: Desktop
<!-- What OS are you using? If you use the web version, which browser are you using? -->
* **Platform**: Windows 10
<!-- Enter your issue details below this comment. -->
App is not loading. It installs fine, I can see the entry in my Application Manager saying it's installed, I can see the process running in my Task Manager but nothing is showing on screen. The app is simply not loading. | True | App is not loading on Windows 10 - <!-- Which version of avocode are you using? -->
<!-- Please make sure you are using the latest version. -->
<!-- You can always check and download the latest version from https://avocode.com/changelog.html. -->
* **Version**: 2.14.8
<!-- Are you using the web or desktop version of Avocode? -->
* **Web/Desktop**: Desktop
<!-- What OS are you using? If you use the web version, which browser are you using? -->
* **Platform**: Windows 10
<!-- Enter your issue details below this comment. -->
App is not loading. It installs fine, I can see the entry in my Application Manager saying it's installed, I can see the process running in my Task Manager but nothing is showing on screen. The app is simply not loading. | non_test | app is not loading on windows version web desktop desktop platform windows app is not loading it installs fine i can see the entry in my application manager saying it s installed i can see the process running in my task manager but nothing is showing on screen the app is simply not loading | 0 |
349,397 | 31,798,711,265 | IssuesEvent | 2023-09-13 09:40:35 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | DISABLED test_nondeterministic_resize_quantized_cpu_quint4x2 (__main__.TestTorchDeviceTypeCPU) | module: tests triaged module: flaky-tests skipped oncall: pt2 | Platforms: dynamo
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_nondeterministic_resize_quantized_cpu_quint4x2&suite=TestTorchDeviceTypeCPU) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/16740983982).
Over the past 3 hours, it has been determined flaky in 6 workflow(s) with 6 failures and 6 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_nondeterministic_resize_quantized_cpu_quint4x2`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_torch.py` | 2.0 | DISABLED test_nondeterministic_resize_quantized_cpu_quint4x2 (__main__.TestTorchDeviceTypeCPU) - Platforms: dynamo
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_nondeterministic_resize_quantized_cpu_quint4x2&suite=TestTorchDeviceTypeCPU) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/16740983982).
Over the past 3 hours, it has been determined flaky in 6 workflow(s) with 6 failures and 6 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_nondeterministic_resize_quantized_cpu_quint4x2`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_torch.py` | test | disabled test nondeterministic resize quantized cpu main testtorchdevicetypecpu platforms dynamo this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test nondeterministic resize quantized cpu there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test torch py | 1 |
87,720 | 8,120,164,393 | IssuesEvent | 2018-08-16 01:00:41 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: jepsen/register/strobe-skews failed on master | C-test-failure O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/d957838032a2dc162794efd23432f2672322f231
Parameters:
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=837166&tab=buildLog
```
test.go:494,jepsen.go:244,jepsen.go:288: /home/agent/work/.go/bin/roachprod run teamcity-837166-jepsen:6 -- bash -e -c "\
cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \
~/lein run test \
--tarball file://${PWD}/cockroach.tgz \
--username ${USER} \
--ssh-private-key ~/.ssh/id_rsa \
--os ubuntu \
--time-limit 300 \
--concurrency 30 \
--recovery-time 25 \
--test-count 1 \
-n 10.128.0.21 -n 10.128.0.5 -n 10.128.0.25 -n 10.128.0.23 -n 10.128.0.12 \
--test register --nemesis strobe-skews \
> invoke.log 2>&1 \
": exit status 1
``` | 1.0 | roachtest: jepsen/register/strobe-skews failed on master - SHA: https://github.com/cockroachdb/cockroach/commits/d957838032a2dc162794efd23432f2672322f231
Parameters:
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=837166&tab=buildLog
```
test.go:494,jepsen.go:244,jepsen.go:288: /home/agent/work/.go/bin/roachprod run teamcity-837166-jepsen:6 -- bash -e -c "\
cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \
~/lein run test \
--tarball file://${PWD}/cockroach.tgz \
--username ${USER} \
--ssh-private-key ~/.ssh/id_rsa \
--os ubuntu \
--time-limit 300 \
--concurrency 30 \
--recovery-time 25 \
--test-count 1 \
-n 10.128.0.21 -n 10.128.0.5 -n 10.128.0.25 -n 10.128.0.23 -n 10.128.0.12 \
--test register --nemesis strobe-skews \
> invoke.log 2>&1 \
": exit status 1
``` | test | roachtest jepsen register strobe skews failed on master sha parameters failed test test go jepsen go jepsen go home agent work go bin roachprod run teamcity jepsen bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test register nemesis strobe skews invoke log exit status | 1 |
75,866 | 9,895,509,828 | IssuesEvent | 2019-06-26 08:00:47 | google/draco | https://api.github.com/repos/google/draco | closed | Running the WebAssembly version of Draco in a web worker | documentation | Hey All,
I am trying to run the WebAssembly version of Draco in a web worker. I have created the following worker code based on [this example](https://github.com/google/draco/blob/master/javascript/time_draco_decode.html):
```JavaScript
importScripts("draco_wasm_wrapper.js");
var decoderModule = {};
var dracoDecoderType = {};
onmessage = function (e) {
var compressedMeshBlob = e.data;
var fileReader = new FileReader();
loadWebAssemblyDecoder(function decoderReady() {
console.log("Decoder is ready.");
fileReader.onload = function () {
var arrayBuffer = fileReader.result;
var geometryData = decodeMesh(arrayBuffer);
postMessage(geometryData);
geometryData = undefined;
close();
};
fileReader.readAsArrayBuffer(compressedMeshBlob);
});
};
function loadWebAssemblyDecoder(decoderReadyCallback) {
dracoDecoderType['wasmBinaryFile'] = 'draco_decoder.wasm';
const xhr = new XMLHttpRequest();
xhr.open('GET', 'draco_decoder.wasm', true);
xhr.responseType = 'arraybuffer';
xhr.onload = function () {
dracoDecoderType['wasmBinary'] = xhr.response;
decoderModule = DracoDecoderModule(dracoDecoderType);
decoderReadyCallback();
};
xhr.send(null)
}
function decodeMesh(arrayBuffer) {
const byteArray = new Uint8Array(arrayBuffer);
const buffer = new decoderModule.DecoderBuffer();
buffer.Init(byteArray, byteArray.length);
const decoder = new decoderModule.Decoder();
const geometryType = decoder.GetEncodedGeometryType(buffer);
var dracoGeometry;
if (geometryType === decoderModule.TRIANGULAR_MESH) {
dracoGeometry = new decoderModule.Mesh();
} else {
throw Error("Only triangular meshes are supported.");
}
decoderModule.destroy(decoder);
decoderModule.destroy(buffer);
return dracoGeometry;
}
```
The `compressedMeshBlob` is a compressed Draco mesh which is loaded as a [Blob](https://developer.mozilla.org/en-US/docs/Web/API/Blob) by the main thread and passed to the web worker through `postMessage()`. I am using the [draco_decoder.wasm](https://github.com/google/draco/blob/master/javascript/draco_decoder.wasm) and [draco_wasm_wrapper.js](https://github.com/google/draco/blob/master/javascript/draco_wasm_wrapper.js) from the respository. I am getting the following error message:

I am not sure if I am doing something incorrectly or if it is fundamentally not possible to run the WebAssembly version of Draco in a web worker. Using the JavaScript decoder works fine.
Thank you very much for your help in advance! | 1.0 | Running the WebAssembly version of Draco in a web worker - Hey All,
I am trying to run the WebAssembly version of Draco in a web worker. I have created the following worker code based on [this example](https://github.com/google/draco/blob/master/javascript/time_draco_decode.html):
```JavaScript
importScripts("draco_wasm_wrapper.js");
var decoderModule = {};
var dracoDecoderType = {};
onmessage = function (e) {
var compressedMeshBlob = e.data;
var fileReader = new FileReader();
loadWebAssemblyDecoder(function decoderReady() {
console.log("Decoder is ready.");
fileReader.onload = function () {
var arrayBuffer = fileReader.result;
var geometryData = decodeMesh(arrayBuffer);
postMessage(geometryData);
geometryData = undefined;
close();
};
fileReader.readAsArrayBuffer(compressedMeshBlob);
});
};
function loadWebAssemblyDecoder(decoderReadyCallback) {
dracoDecoderType['wasmBinaryFile'] = 'draco_decoder.wasm';
const xhr = new XMLHttpRequest();
xhr.open('GET', 'draco_decoder.wasm', true);
xhr.responseType = 'arraybuffer';
xhr.onload = function () {
dracoDecoderType['wasmBinary'] = xhr.response;
decoderModule = DracoDecoderModule(dracoDecoderType);
decoderReadyCallback();
};
xhr.send(null)
}
function decodeMesh(arrayBuffer) {
const byteArray = new Uint8Array(arrayBuffer);
const buffer = new decoderModule.DecoderBuffer();
buffer.Init(byteArray, byteArray.length);
const decoder = new decoderModule.Decoder();
const geometryType = decoder.GetEncodedGeometryType(buffer);
var dracoGeometry;
if (geometryType === decoderModule.TRIANGULAR_MESH) {
dracoGeometry = new decoderModule.Mesh();
} else {
throw Error("Only triangular meshes are supported.");
}
decoderModule.destroy(decoder);
decoderModule.destroy(buffer);
return dracoGeometry;
}
```
The `compressedMeshBlob` is a compressed Draco mesh which is loaded as a [Blob](https://developer.mozilla.org/en-US/docs/Web/API/Blob) by the main thread and passed to the web worker through `postMessage()`. I am using the [draco_decoder.wasm](https://github.com/google/draco/blob/master/javascript/draco_decoder.wasm) and [draco_wasm_wrapper.js](https://github.com/google/draco/blob/master/javascript/draco_wasm_wrapper.js) from the respository. I am getting the following error message:

I am not sure if I am doing something incorrectly or if it is fundamentally not possible to run the WebAssembly version of Draco in a web worker. Using the JavaScript decoder works fine.
Thank you very much for your help in advance! | non_test | running the webassembly version of draco in a web worker hey all i am trying to run the webassembly version of draco in a web worker i have created the following worker code based on javascript importscripts draco wasm wrapper js var decodermodule var dracodecodertype onmessage function e var compressedmeshblob e data var filereader new filereader loadwebassemblydecoder function decoderready console log decoder is ready filereader onload function var arraybuffer filereader result var geometrydata decodemesh arraybuffer postmessage geometrydata geometrydata undefined close filereader readasarraybuffer compressedmeshblob function loadwebassemblydecoder decoderreadycallback dracodecodertype draco decoder wasm const xhr new xmlhttprequest xhr open get draco decoder wasm true xhr responsetype arraybuffer xhr onload function dracodecodertype xhr response decodermodule dracodecodermodule dracodecodertype decoderreadycallback xhr send null function decodemesh arraybuffer const bytearray new arraybuffer const buffer new decodermodule decoderbuffer buffer init bytearray bytearray length const decoder new decodermodule decoder const geometrytype decoder getencodedgeometrytype buffer var dracogeometry if geometrytype decodermodule triangular mesh dracogeometry new decodermodule mesh else throw error only triangular meshes are supported decodermodule destroy decoder decodermodule destroy buffer return dracogeometry the compressedmeshblob is a compressed draco mesh which is loaded as a by the main thread and passed to the web worker through postmessage i am using the and from the respository i am getting the following error message i am not sure if i am doing something incorrectly or if it is fundamentally not possible to run the webassembly version of draco in a web worker using the javascript decoder works fine thank you very much for your help in advance | 0 |
1,360 | 2,511,938,065 | IssuesEvent | 2015-01-14 12:46:40 | transientskp/tkp | https://api.github.com/repos/transientskp/tkp | opened | Quality Control: Significant number of particular station type flagged | enhancement priority normal | Method:
1. Select all baselines between the imaged minimum and maximum baseline length (max_bl and min_bl given in the LOFAR_OBSERVATION table of the CASA image). The antenna id numbers are obtained from the LOFAR_ANTENNA table (i.e. find all pairs of antennas with a separation in the baseline range). For all the available baselines, identify the unique stations used and sort into core, remote and international.
2. Use the NVIS_USED and NVIS_TOTAL tables in the LOFAR_OBSERVATION table. Sum the maximum possible data available and sum the amount of data used for each type of station.
3. Calculate the percentage of flagged data ((1 - used/total)*100%) and flag the station as 'bad' if this exceeds the user defined maximum percentage allowed (same value as used in #3788).
4. Sum the number of 'bad' stations of a particular type and the total number of stations of a particular type. Use these to calculate the percentage of 'bad' baselines. Print the number of good and bad stations into the pipeline.log for each type of station. $type = (core remote international)
5. Reject image if the percentage of 'bad' stations in a particular category exceeds a user defined value (bad_station_percentage given in the quality_control.parset). Quality rejection reason: “Large amount of Measurement Set flagged. $value % of $type stations flagged.”
original issue:
https://support.astron.nl/lofar_issuetracker/issues/3884 | 1.0 | Quality Control: Significant number of particular station type flagged - Method:
1. Select all baselines between the imaged minimum and maximum baseline length (max_bl and min_bl given in the LOFAR_OBSERVATION table of the CASA image). The antenna id numbers are obtained from the LOFAR_ANTENNA table (i.e. find all pairs of antennas with a separation in the baseline range). For all the available baselines, identify the unique stations used and sort into core, remote and international.
2. Use the NVIS_USED and NVIS_TOTAL tables in the LOFAR_OBSERVATION table. Sum the maximum possible data available and sum the amount of data used for each type of station.
3. Calculate the percentage of flagged data ((1 - used/total)*100%) and flag the station as 'bad' if this exceeds the user defined maximum percentage allowed (same value as used in #3788).
4. Sum the number of 'bad' stations of a particular type and the total number of stations of a particular type. Use these to calculate the percentage of 'bad' baselines. Print the number of good and bad stations into the pipeline.log for each type of station. $type = (core remote international)
5. Reject image if the percentage of 'bad' stations in a particular category exceeds a user defined value (bad_station_percentage given in the quality_control.parset). Quality rejection reason: “Large amount of Measurement Set flagged. $value % of $type stations flagged.”
original issue:
https://support.astron.nl/lofar_issuetracker/issues/3884 | non_test | quality control significant number of particular station type flagged method select all baselines between the imaged minimum and maximum baseline length max bl and min bl given in the lofar observation table of the casa image the antenna id numbers are obtained from the lofar antenna table i e find all pairs of antennas with a separation in the baseline range for all the available baselines identify the unique stations used and sort into core remote and international use the nvis used and nvis total tables in the lofar observation table sum the maximum possible data available and sum the amount of data used for each type of station calculate the percentage of flagged data used total and flag the station as bad if this exceeds the user defined maximum percentage allowed same value as used in sum the number of bad stations of a particular type and the total number of stations of a particular type use these to calculate the percentage of bad baselines print the number of good and bad stations into the pipeline log for each type of station type core remote international reject image if the percentage of bad stations in a particular category exceeds a user defined value bad station percentage given in the quality control parset quality rejection reason “large amount of measurement set flagged value of type stations flagged ” original issue | 0 |
91,280 | 26,346,600,782 | IssuesEvent | 2023-01-10 22:46:26 | FashionFreedom/Seamly2D | https://api.github.com/repos/FashionFreedom/Seamly2D | closed | Workflows are referencing vulnerable actions | bug build | Hello, there!
As part of the university research we are currently doing regarding the security of Github Actions, we noticed that one or many of the workflows that are part of this repository are referencing vulnerable versions of the third-party actions. As part of a disclosure process, we decided to open issues to notify GitHub Community.
*Please note that there are could be some false positives in our methodology, thus not all of the open issues could be valid. If that is the case, please let us know, so that we can improve on our approach. You can contact me directly using an email: ikoishy [at] ncsu.edu*
Thanks in advance
1. The workflow **build-release-assets.yml** is referencing action **ilammy/msvc-dev-cmd** using references **v1.4.1**. However this reference is missing the commit [74a501b](https://github.com/ilammy/msvc-dev-cmd/commits/74a501b) which may contain fix to the vulnerability.
1. The workflow **build-weekly-release.yml** is referencing action **ilammy/msvc-dev-cmd** using references **v1.4.1**. However this reference is missing the commit [74a501b](https://github.com/ilammy/msvc-dev-cmd/commits/74a501b) which may contain fix to the vulnerability.
The vulnerability fix that is missing by actions' versions could be related to:
(1) CVE fix
(2) upgrade of vulnerable dependency
(3) fix to secret leak and others.
Please consider updating the reference to the action.
*If you end up updating the reference, please let us know. We need the stats for the paper :-)* | 1.0 | Workflows are referencing vulnerable actions - Hello, there!
As part of the university research we are currently doing regarding the security of Github Actions, we noticed that one or many of the workflows that are part of this repository are referencing vulnerable versions of the third-party actions. As part of a disclosure process, we decided to open issues to notify GitHub Community.
*Please note that there are could be some false positives in our methodology, thus not all of the open issues could be valid. If that is the case, please let us know, so that we can improve on our approach. You can contact me directly using an email: ikoishy [at] ncsu.edu*
Thanks in advance
1. The workflow **build-release-assets.yml** is referencing action **ilammy/msvc-dev-cmd** using references **v1.4.1**. However this reference is missing the commit [74a501b](https://github.com/ilammy/msvc-dev-cmd/commits/74a501b) which may contain fix to the vulnerability.
1. The workflow **build-weekly-release.yml** is referencing action **ilammy/msvc-dev-cmd** using references **v1.4.1**. However this reference is missing the commit [74a501b](https://github.com/ilammy/msvc-dev-cmd/commits/74a501b) which may contain fix to the vulnerability.
The vulnerability fix that is missing by actions' versions could be related to:
(1) CVE fix
(2) upgrade of vulnerable dependency
(3) fix to secret leak and others.
Please consider updating the reference to the action.
*If you end up updating the reference, please let us know. We need the stats for the paper :-)* | non_test | workflows are referencing vulnerable actions hello there as part of the university research we are currently doing regarding the security of github actions we noticed that one or many of the workflows that are part of this repository are referencing vulnerable versions of the third party actions as part of a disclosure process we decided to open issues to notify github community please note that there are could be some false positives in our methodology thus not all of the open issues could be valid if that is the case please let us know so that we can improve on our approach you can contact me directly using an email ikoishy ncsu edu thanks in advance the workflow build release assets yml is referencing action ilammy msvc dev cmd using references however this reference is missing the commit which may contain fix to the vulnerability the workflow build weekly release yml is referencing action ilammy msvc dev cmd using references however this reference is missing the commit which may contain fix to the vulnerability the vulnerability fix that is missing by actions versions could be related to cve fix upgrade of vulnerable dependency fix to secret leak and others please consider updating the reference to the action if you end up updating the reference please let us know we need the stats for the paper | 0 |
693,451 | 23,775,902,484 | IssuesEvent | 2022-09-01 20:54:54 | GoogleCloudPlatform/microservices-demo | https://api.github.com/repos/GoogleCloudPlatform/microservices-demo | closed | Update `skaffold` in `cloudbuild.yaml` | type: bug priority: p3 | ### Describe the bug
* Our `skaffold.yaml` uses `skaffold` "Modules" — added in https://github.com/GoogleCloudPlatform/microservices-demo/pull/609.
* But the our `cloudbuild.yaml` file in the root directory uses `skaffold` **v0.20.0** which doesn't support `skaffold` "Modules".
https://github.com/GoogleCloudPlatform/microservices-demo/blob/8cd3ae52c93b2c5be5808c47bbfba1f8002ea13e/cloudbuild.yaml#L28
* We need to:
* [ ] Update the version of `skaffold` being used to at least 1.27 (ideally, the latest) in the above `cloudbuild.yaml` file (in a new branch).
* [ ] Test that updated `cloudbuild.yaml` file by creating a Google Cloud Build Trigger targeting the new branch.
* [ ] Merge changes into `main`.
* This issue was discovered in #752.
### To Reproduce
* We would have to create a Google Cloud Build Trigger pointing to the `cloudbuild.yaml` file in `main`.
### Logs
* I have not tested this breakage manually, but I assume we should see a message similar to [this](https://github.com/GoogleCloudPlatform/bank-of-anthos/pull/514#pullrequestreview-713779181) in the Google Cloud Build Trigger's logs:
```
WARN[0000] Your Skaffold version might be too old. Download the latest version (1.28.1) from:
https://github.com/GoogleContainerTools/skaffold/releases/tag/v1.28.1
parsing skaffold config: error parsing skaffold configuration file: unknown skaffold config API version "skaffold/v2beta18". Set the config 'apiVersion' to a known value. Check https://skaffold.dev/docs/references/yaml/ for the list of valid API versions. Otherwise, check that your skaffold version is up-to-date.
```
### Screenshots
N/A
### Environment
* Google Cloud Build
### Additional context
N/A
### Exposure
* This has been broken for a while now (`skaffold` modules was [added in Oct 2021](https://github.com/GoogleCloudPlatform/microservices-demo/pull/609)).
* We don't have data on how many people are relying on our `cloudbuild.yaml` file (i.e., for learning/demo purposes) — I assume it's minimal.
| 1.0 | Update `skaffold` in `cloudbuild.yaml` - ### Describe the bug
* Our `skaffold.yaml` uses `skaffold` "Modules" — added in https://github.com/GoogleCloudPlatform/microservices-demo/pull/609.
* But the our `cloudbuild.yaml` file in the root directory uses `skaffold` **v0.20.0** which doesn't support `skaffold` "Modules".
https://github.com/GoogleCloudPlatform/microservices-demo/blob/8cd3ae52c93b2c5be5808c47bbfba1f8002ea13e/cloudbuild.yaml#L28
* We need to:
* [ ] Update the version of `skaffold` being used to at least 1.27 (ideally, the latest) in the above `cloudbuild.yaml` file (in a new branch).
* [ ] Test that updated `cloudbuild.yaml` file by creating a Google Cloud Build Trigger targeting the new branch.
* [ ] Merge changes into `main`.
* This issue was discovered in #752.
### To Reproduce
* We would have to create a Google Cloud Build Trigger pointing to the `cloudbuild.yaml` file in `main`.
### Logs
* I have not tested this breakage manually, but I assume we should see a message similar to [this](https://github.com/GoogleCloudPlatform/bank-of-anthos/pull/514#pullrequestreview-713779181) in the Google Cloud Build Trigger's logs:
```
WARN[0000] Your Skaffold version might be too old. Download the latest version (1.28.1) from:
https://github.com/GoogleContainerTools/skaffold/releases/tag/v1.28.1
parsing skaffold config: error parsing skaffold configuration file: unknown skaffold config API version "skaffold/v2beta18". Set the config 'apiVersion' to a known value. Check https://skaffold.dev/docs/references/yaml/ for the list of valid API versions. Otherwise, check that your skaffold version is up-to-date.
```
### Screenshots
N/A
### Environment
* Google Cloud Build
### Additional context
N/A
### Exposure
* This has been broken for a while now (`skaffold` modules was [added in Oct 2021](https://github.com/GoogleCloudPlatform/microservices-demo/pull/609)).
* We don't have data on how many people are relying on our `cloudbuild.yaml` file (i.e., for learning/demo purposes) — I assume it's minimal.
| non_test | update skaffold in cloudbuild yaml describe the bug our skaffold yaml uses skaffold modules — added in but the our cloudbuild yaml file in the root directory uses skaffold which doesn t support skaffold modules we need to update the version of skaffold being used to at least ideally the latest in the above cloudbuild yaml file in a new branch test that updated cloudbuild yaml file by creating a google cloud build trigger targeting the new branch merge changes into main this issue was discovered in to reproduce we would have to create a google cloud build trigger pointing to the cloudbuild yaml file in main logs i have not tested this breakage manually but i assume we should see a message similar to in the google cloud build trigger s logs warn your skaffold version might be too old download the latest version from parsing skaffold config error parsing skaffold configuration file unknown skaffold config api version skaffold set the config apiversion to a known value check for the list of valid api versions otherwise check that your skaffold version is up to date screenshots n a environment google cloud build additional context n a exposure this has been broken for a while now skaffold modules was we don t have data on how many people are relying on our cloudbuild yaml file i e for learning demo purposes — i assume it s minimal | 0 |
684,745 | 23,428,876,485 | IssuesEvent | 2022-08-14 20:20:19 | MartinXPN/profound.academy | https://api.github.com/repos/MartinXPN/profound.academy | opened | Re-evaluate is not scalable enough | bug enhancement priority/low | Re-evaluate dumps everything into a single transaction and becomes really not scalable. It can't process 200 submissions for instance. We should make it more scalable. | 1.0 | Re-evaluate is not scalable enough - Re-evaluate dumps everything into a single transaction and becomes really not scalable. It can't process 200 submissions for instance. We should make it more scalable. | non_test | re evaluate is not scalable enough re evaluate dumps everything into a single transaction and becomes really not scalable it can t process submissions for instance we should make it more scalable | 0 |
52,153 | 27,402,195,518 | IssuesEvent | 2023-03-01 02:06:26 | vmware-tanzu/velero | https://api.github.com/repos/vmware-tanzu/velero | closed | Velero backup duration using Kopia | Performance Needs investigation area/fs-backup | **Describe the problem/challenge you have**
Noticed during testing, that once uploader=kopia is configured,
cluster nodes are backup one by one instead of backup them in parallel.
It increases duration of backup for Kopia, especially that node agents are run on each node.
**Describe the solution you'd like**
Consider to run Kopia backup in parallel on all nodes or have it configurable.
**Anything else you would like to add:**
Likely VCPU/Memory for velero deployment would increase, consider more tasks to run on node agents and velero deployment to manage status of agents.
**Environment:**
- Velero version (use `velero version`): 1.10.1 with Kopia uploaded
- Kubernetes version (use `kubectl version`): AKS 1.24.6
- Kubernetes installer & version: 1.24.6
- Cloud provider or hardware configuration: Azure
- OS (e.g. from `/etc/os-release`): Ubuntu 18 as part of AKS
**Vote on this issue!**
This is an invitation to the Velero community to vote on issues, you can see the project's [top voted issues listed here](https://github.com/vmware-tanzu/velero/issues?q=is%3Aissue+is%3Aopen+sort%3Areactions-%2B1-desc).
Use the "reaction smiley face" up to the right of this comment to vote.
- :+1: for "The project would be better with this feature added"
- :-1: for "This feature will not enhance the project in a meaningful way"
| True | Velero backup duration using Kopia - **Describe the problem/challenge you have**
Noticed during testing, that once uploader=kopia is configured,
cluster nodes are backup one by one instead of backup them in parallel.
It increases duration of backup for Kopia, especially that node agents are run on each node.
**Describe the solution you'd like**
Consider to run Kopia backup in parallel on all nodes or have it configurable.
**Anything else you would like to add:**
Likely VCPU/Memory for velero deployment would increase, consider more tasks to run on node agents and velero deployment to manage status of agents.
**Environment:**
- Velero version (use `velero version`): 1.10.1 with Kopia uploaded
- Kubernetes version (use `kubectl version`): AKS 1.24.6
- Kubernetes installer & version: 1.24.6
- Cloud provider or hardware configuration: Azure
- OS (e.g. from `/etc/os-release`): Ubuntu 18 as part of AKS
**Vote on this issue!**
This is an invitation to the Velero community to vote on issues, you can see the project's [top voted issues listed here](https://github.com/vmware-tanzu/velero/issues?q=is%3Aissue+is%3Aopen+sort%3Areactions-%2B1-desc).
Use the "reaction smiley face" up to the right of this comment to vote.
- :+1: for "The project would be better with this feature added"
- :-1: for "This feature will not enhance the project in a meaningful way"
| non_test | velero backup duration using kopia describe the problem challenge you have noticed during testing that once uploader kopia is configured cluster nodes are backup one by one instead of backup them in parallel it increases duration of backup for kopia especially that node agents are run on each node describe the solution you d like consider to run kopia backup in parallel on all nodes or have it configurable anything else you would like to add likely vcpu memory for velero deployment would increase consider more tasks to run on node agents and velero deployment to manage status of agents environment velero version use velero version with kopia uploaded kubernetes version use kubectl version aks kubernetes installer version cloud provider or hardware configuration azure os e g from etc os release ubuntu as part of aks vote on this issue this is an invitation to the velero community to vote on issues you can see the project s use the reaction smiley face up to the right of this comment to vote for the project would be better with this feature added for this feature will not enhance the project in a meaningful way | 0 |
295,563 | 25,484,535,942 | IssuesEvent | 2022-11-26 07:31:39 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | pkg/ccl/logictestccl/tests/3node-tenant/3node-tenant_test: TestTenantLogic_timetz failed | C-test-failure O-robot branch-release-22.2 | pkg/ccl/logictestccl/tests/3node-tenant/3node-tenant_test.TestTenantLogic_timetz [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7698384?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7698384?buildTab=artifacts#/) on release-22.2 @ [fd133ba3af3954c93584a501cf68dfa3131e586c](https://github.com/cockroachdb/cockroach/commits/fd133ba3af3954c93584a501cf68dfa3131e586c):
```
Slow failing tests:
TestTenantLogic_timetz - 1.41s
Slow passing tests:
TestTenantLogic_collatedstring - 160.43s
TestTenantLogic_srfs - 133.88s
TestTenantLogic_pg_builtins - 128.96s
TestTenantLogic_fk - 84.38s
TestTenantLogic_fk - 76.01s
TestTenantLogic_alter_table - 70.13s
TestTenantLogic_alter_table - 67.00s
TestTenantLogic_alter_primary_key - 60.31s
TestTenantLogic_alter_primary_key - 52.40s
TestTenantLogic_create_as_non_metamorphic - 48.82s
TestTenantLogic_cascade - 45.36s
TestTenantLogic_enums - 43.74s
TestTenantLogic_cascade - 43.16s
TestTenantLogic_enums - 42.67s
TestTenantLogic_drop_type - 40.63s
TestTenantLogic_hash_sharded_index - 39.27s
TestTenantLogic_new_schema_changer - 39.11s
TestTenantLogic_new_schema_changer - 38.80s
TestTenantLogic_hash_sharded_index - 38.06s
TestTenantLogic_drop_type - 37.03s
```
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantLogic_timetz.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 1.0 | pkg/ccl/logictestccl/tests/3node-tenant/3node-tenant_test: TestTenantLogic_timetz failed - pkg/ccl/logictestccl/tests/3node-tenant/3node-tenant_test.TestTenantLogic_timetz [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7698384?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7698384?buildTab=artifacts#/) on release-22.2 @ [fd133ba3af3954c93584a501cf68dfa3131e586c](https://github.com/cockroachdb/cockroach/commits/fd133ba3af3954c93584a501cf68dfa3131e586c):
```
Slow failing tests:
TestTenantLogic_timetz - 1.41s
Slow passing tests:
TestTenantLogic_collatedstring - 160.43s
TestTenantLogic_srfs - 133.88s
TestTenantLogic_pg_builtins - 128.96s
TestTenantLogic_fk - 84.38s
TestTenantLogic_fk - 76.01s
TestTenantLogic_alter_table - 70.13s
TestTenantLogic_alter_table - 67.00s
TestTenantLogic_alter_primary_key - 60.31s
TestTenantLogic_alter_primary_key - 52.40s
TestTenantLogic_create_as_non_metamorphic - 48.82s
TestTenantLogic_cascade - 45.36s
TestTenantLogic_enums - 43.74s
TestTenantLogic_cascade - 43.16s
TestTenantLogic_enums - 42.67s
TestTenantLogic_drop_type - 40.63s
TestTenantLogic_hash_sharded_index - 39.27s
TestTenantLogic_new_schema_changer - 39.11s
TestTenantLogic_new_schema_changer - 38.80s
TestTenantLogic_hash_sharded_index - 38.06s
TestTenantLogic_drop_type - 37.03s
```
<p>Parameters: <code>TAGS=bazel,gss</code>
</p>
<details><summary>Help</summary>
<p>
See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM)
</p>
</details>
/cc @cockroachdb/sql-queries
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantLogic_timetz.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| test | pkg ccl logictestccl tests tenant tenant test testtenantlogic timetz failed pkg ccl logictestccl tests tenant tenant test testtenantlogic timetz with on release slow failing tests testtenantlogic timetz slow passing tests testtenantlogic collatedstring testtenantlogic srfs testtenantlogic pg builtins testtenantlogic fk testtenantlogic fk testtenantlogic alter table testtenantlogic alter table testtenantlogic alter primary key testtenantlogic alter primary key testtenantlogic create as non metamorphic testtenantlogic cascade testtenantlogic enums testtenantlogic cascade testtenantlogic enums testtenantlogic drop type testtenantlogic hash sharded index testtenantlogic new schema changer testtenantlogic new schema changer testtenantlogic hash sharded index testtenantlogic drop type parameters tags bazel gss help see also cc cockroachdb sql queries | 1 |
232,066 | 17,770,710,226 | IssuesEvent | 2021-08-30 13:21:35 | juliuskoskela/corewar | https://api.github.com/repos/juliuskoskela/corewar | opened | create a README in project root | documentation | Add a README that would contain the following elements, for example:
- short project description
- how to compile and run each program
- some example?
- what kind of tests exist, and how to run them
- links to documents that contain more information | 1.0 | create a README in project root - Add a README that would contain the following elements, for example:
- short project description
- how to compile and run each program
- some example?
- what kind of tests exist, and how to run them
- links to documents that contain more information | non_test | create a readme in project root add a readme that would contain the following elements for example short project description how to compile and run each program some example what kind of tests exist and how to run them links to documents that contain more information | 0 |
22,839 | 10,789,236,122 | IssuesEvent | 2019-11-05 11:27:12 | zuhlke/ManageMyBudget | https://api.github.com/repos/zuhlke/ManageMyBudget | closed | Create VPC and security group to enable inbound connection to DocumentDB cluster | provisioning security terraform | Write a Terraform script to create VPC and security group for DocumentDB cluster to enable inbound connection | True | Create VPC and security group to enable inbound connection to DocumentDB cluster - Write a Terraform script to create VPC and security group for DocumentDB cluster to enable inbound connection | non_test | create vpc and security group to enable inbound connection to documentdb cluster write a terraform script to create vpc and security group for documentdb cluster to enable inbound connection | 0 |
345,877 | 30,849,622,451 | IssuesEvent | 2023-08-02 15:48:25 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: knex failed | C-test-failure O-robot O-roachtest T-sql-foundations branch-release-22.2 | roachtest.knex [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10044657?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10044657?buildTab=artifacts#/knex) on release-22.2 @ [97259ff3650b28abde99a8174d70016c9b5daa28](https://github.com/cockroachdb/cockroach/commits/97259ff3650b28abde99a8174d70016c9b5daa28):
```
test artifacts and logs in: /artifacts/knex/run_1
(knex.go:115).func1: COMMAND_PROBLEM: ssh verbose log retained in ssh_083745.354953804_n1_cd-mntdata1knex-DBco.log: exit status 1
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-sessions
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*knex.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| 2.0 | roachtest: knex failed - roachtest.knex [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10044657?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10044657?buildTab=artifacts#/knex) on release-22.2 @ [97259ff3650b28abde99a8174d70016c9b5daa28](https://github.com/cockroachdb/cockroach/commits/97259ff3650b28abde99a8174d70016c9b5daa28):
```
test artifacts and logs in: /artifacts/knex/run_1
(knex.go:115).func1: COMMAND_PROBLEM: ssh verbose log retained in ssh_083745.354953804_n1_cd-mntdata1knex-DBco.log: exit status 1
```
<p>Parameters: <code>ROACHTEST_cloud=gce</code>
, <code>ROACHTEST_cpu=4</code>
, <code>ROACHTEST_encrypted=false</code>
, <code>ROACHTEST_ssd=0</code>
</p>
<details><summary>Help</summary>
<p>
See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md)
See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7)
</p>
</details>
/cc @cockroachdb/sql-sessions
<sub>
[This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*knex.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)
</sub>
| test | roachtest knex failed roachtest knex with on release test artifacts and logs in artifacts knex run knex go command problem ssh verbose log retained in ssh cd dbco log exit status parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql sessions | 1 |
693,446 | 23,775,756,975 | IssuesEvent | 2022-09-01 20:45:43 | kytos-ng/maintenance | https://api.github.com/repos/kytos-ng/maintenance | closed | rest endpoints don't include a version in the route | enhancement priority_low 2022.3 epic_maintenance_v1 | Currently, this NApp isn't using versioned routes in the rest endpoints, most of the NApps we maintain such as `mef_eline`, `flow_manager`, `topology`, `storehouse`, `of_lldp` and so on uses a version like `v[1-9]+` in the route path which makes easier to provide certain guarantees for clients | 1.0 | rest endpoints don't include a version in the route - Currently, this NApp isn't using versioned routes in the rest endpoints, most of the NApps we maintain such as `mef_eline`, `flow_manager`, `topology`, `storehouse`, `of_lldp` and so on uses a version like `v[1-9]+` in the route path which makes easier to provide certain guarantees for clients | non_test | rest endpoints don t include a version in the route currently this napp isn t using versioned routes in the rest endpoints most of the napps we maintain such as mef eline flow manager topology storehouse of lldp and so on uses a version like v in the route path which makes easier to provide certain guarantees for clients | 0 |
328,469 | 28,122,417,542 | IssuesEvent | 2023-03-31 15:04:20 | unifyai/ivy | https://api.github.com/repos/unifyai/ivy | closed | Fix ndarray.test_numpy_instance_copy__ | NumPy Frontend Sub Task Failing Test | | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4508140756/jobs/7936571840" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4508140756/jobs/7936571840" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
| 1.0 | Fix ndarray.test_numpy_instance_copy__ - | | |
|---|---|
|tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4508140756/jobs/7936571840" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4508140756/jobs/7936571840" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
|jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
| test | fix ndarray test numpy instance copy tensorflow img src torch img src numpy img src jax img src | 1 |
9,575 | 3,059,000,962 | IssuesEvent | 2015-08-14 12:27:18 | TypeCobolTeam/TypeCobol | https://api.github.com/repos/TypeCobolTeam/TypeCobol | closed | Unit Tests must be locale invariant | bug Tests | Whereas compiler should output message in the user's locale, we cannot allow unit tests to depend of their execution environment. In particular, current scanner tests compare numbers, separators and stuff against their string equivalents *in the french locale*.
All tests should run in an invariant culture, so their results are predictable.
Hint: add `Thread.CurrentThread.CurrentCulture = CultureInfo.InvariantCulture;` before any test runs. | 1.0 | Unit Tests must be locale invariant - Whereas compiler should output message in the user's locale, we cannot allow unit tests to depend of their execution environment. In particular, current scanner tests compare numbers, separators and stuff against their string equivalents *in the french locale*.
All tests should run in an invariant culture, so their results are predictable.
Hint: add `Thread.CurrentThread.CurrentCulture = CultureInfo.InvariantCulture;` before any test runs. | test | unit tests must be locale invariant whereas compiler should output message in the user s locale we cannot allow unit tests to depend of their execution environment in particular current scanner tests compare numbers separators and stuff against their string equivalents in the french locale all tests should run in an invariant culture so their results are predictable hint add thread currentthread currentculture cultureinfo invariantculture before any test runs | 1 |
6,589 | 2,852,613,022 | IssuesEvent | 2015-06-01 14:31:00 | NativeScript/NativeScript | https://api.github.com/repos/NativeScript/NativeScript | closed | Expose application level events as real events | 2 - Open 3 - In Progress 4 - Ready For Test 5 - Done | Currently we are exposing the application level events with `onEventName` pattern. Consider providing real events as well: `addEventListener`/`on` in order to better handle all use cases.
See [here](https://github.com/NativeScript/ios-runtime/issues/62#issuecomment-106268863) for more details. | 1.0 | Expose application level events as real events - Currently we are exposing the application level events with `onEventName` pattern. Consider providing real events as well: `addEventListener`/`on` in order to better handle all use cases.
See [here](https://github.com/NativeScript/ios-runtime/issues/62#issuecomment-106268863) for more details. | test | expose application level events as real events currently we are exposing the application level events with oneventname pattern consider providing real events as well addeventlistener on in order to better handle all use cases see for more details | 1 |
309,679 | 26,672,507,568 | IssuesEvent | 2023-01-26 11:33:55 | wazuh/wazuh-qa | https://api.github.com/repos/wazuh/wazuh-qa | opened | Test unnecessary sanity check removed on wdb_syscollector_save2 function | team/qa type/manual-testing status/not-tracked role/qa-data-pirates subteam/qa-rainbow | | Target version | Related issue | Related PR/dev branch |
|--------------------|--------------------|-----------------|
| TBD | https://github.com/wazuh/wazuh/issues/10380 | https://github.com/wazuh/wazuh/pull/16008 |
<!-- Important: No section may be left blank. If not, delete it directly (in principle only "Configurations" and "Considerations" could be left blank in case of not proceeding). -->
## Description
<!-- Description that puts into context and shows the QA tester the changes that have been implemented and have to be tested. -->
This development aims to fix a segmentation fault in Wazuh db syscollector in versions less than 4.2
## Proposed test cases
<!-- Indicate the minimum test cases proposed by the developer. -->
- Upgrade to core-branch
1. Install wazuh agent
2. Enable the syscollector config
3. Check wazuh db does not crash in the manager
4. Check that there are no errors in the ossec.log
- Fresh install core-branch
1. Install wazuh agent
2. Enable the syscollector config
3. Check wazuh db does not crash in the manager
4. Check that there are no errors in the ossec.log
| 1.0 | Test unnecessary sanity check removed on wdb_syscollector_save2 function - | Target version | Related issue | Related PR/dev branch |
|--------------------|--------------------|-----------------|
| TBD | https://github.com/wazuh/wazuh/issues/10380 | https://github.com/wazuh/wazuh/pull/16008 |
<!-- Important: No section may be left blank. If not, delete it directly (in principle only "Configurations" and "Considerations" could be left blank in case of not proceeding). -->
## Description
<!-- Description that puts into context and shows the QA tester the changes that have been implemented and have to be tested. -->
This development aims to fix a segmentation fault in Wazuh db syscollector in versions less than 4.2
## Proposed test cases
<!-- Indicate the minimum test cases proposed by the developer. -->
- Upgrade to core-branch
1. Install wazuh agent
2. Enable the syscollector config
3. Check wazuh db does not crash in the manager
4. Check that there are no errors in the ossec.log
- Fresh install core-branch
1. Install wazuh agent
2. Enable the syscollector config
3. Check wazuh db does not crash in the manager
4. Check that there are no errors in the ossec.log
| test | test unnecessary sanity check removed on wdb syscollector function target version related issue related pr dev branch tbd description this development aims to fix a segmentation fault in wazuh db syscollector in versions less than proposed test cases upgrade to core branch install wazuh agent enable the syscollector config check wazuh db does not crash in the manager check that there are no errors in the ossec log fresh install core branch install wazuh agent enable the syscollector config check wazuh db does not crash in the manager check that there are no errors in the ossec log | 1 |
95,824 | 27,628,379,722 | IssuesEvent | 2023-03-10 08:57:51 | NVIDIA/spark-rapids-jni | https://api.github.com/repos/NVIDIA/spark-rapids-jni | closed | [FEA] Add CI for nightly or weekly fuzz testing for SparkRMM retry | build | **Is your feature request related to a problem? Please describe.**
I added in a fuzz tester for Spark RMM retry
https://github.com/NVIDIA/spark-rapids-jni/pull/977
It would be nice to run it with a random seed nightly or weekly to make sure that things are still working, and possibly find some errors that need to be fixed. | 1.0 | [FEA] Add CI for nightly or weekly fuzz testing for SparkRMM retry - **Is your feature request related to a problem? Please describe.**
I added in a fuzz tester for Spark RMM retry
https://github.com/NVIDIA/spark-rapids-jni/pull/977
It would be nice to run it with a random seed nightly or weekly to make sure that things are still working, and possibly find some errors that need to be fixed. | non_test | add ci for nightly or weekly fuzz testing for sparkrmm retry is your feature request related to a problem please describe i added in a fuzz tester for spark rmm retry it would be nice to run it with a random seed nightly or weekly to make sure that things are still working and possibly find some errors that need to be fixed | 0 |
54,074 | 13,894,680,270 | IssuesEvent | 2020-10-19 14:57:50 | jgeraigery/thingworx-gitbackup-extension | https://api.github.com/repos/jgeraigery/thingworx-gitbackup-extension | opened | CVE-2020-11111 (High) detected in jackson-databind-2.9.8.jar | security vulnerability | ## CVE-2020-11111 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: thingworx-gitbackup-extension/twx-lib/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/thingworx-gitbackup-extension/commit/a274c5bf23bad29fe832613963ebac660dbc17bb">a274c5bf23bad29fe832613963ebac660dbc17bb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.activemq.* (aka activemq-jms, activemq-core, activemq-pool, and activemq-pool-jms).
<p>Publish Date: 2020-03-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11111>CVE-2020-11111</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11113">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11113</a></p>
<p>Release Date: 2020-03-31</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0"}],"vulnerabilityIdentifier":"CVE-2020-11111","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.activemq.* (aka activemq-jms, activemq-core, activemq-pool, and activemq-pool-jms).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11111","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | True | CVE-2020-11111 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-11111 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: thingworx-gitbackup-extension/twx-lib/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/jgeraigery/thingworx-gitbackup-extension/commit/a274c5bf23bad29fe832613963ebac660dbc17bb">a274c5bf23bad29fe832613963ebac660dbc17bb</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.activemq.* (aka activemq-jms, activemq-core, activemq-pool, and activemq-pool-jms).
<p>Publish Date: 2020-03-31
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11111>CVE-2020-11111</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11113">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11113</a></p>
<p>Release Date: 2020-03-31</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p>
</p>
</details>
<p></p>
***
:rescue_worker_helmet: Automatic Remediation is available for this issue
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0"}],"vulnerabilityIdentifier":"CVE-2020-11111","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.activemq.* (aka activemq-jms, activemq-core, activemq-pool, and activemq-pool-jms).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11111","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> --> | non_test | cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library thingworx gitbackup extension twx lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache activemq aka activemq jms activemq core activemq pool and activemq pool jms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache activemq aka activemq jms activemq core activemq pool and activemq pool jms vulnerabilityurl | 0 |
47,704 | 5,908,625,104 | IssuesEvent | 2017-05-19 20:56:51 | karmaradio/karma | https://api.github.com/repos/karmaradio/karma | closed | Email Verification | bug please-test T1d user-story | As a contractor/project manager,
I want to verify my email when registering,
So that my offers are more secure | 1.0 | Email Verification - As a contractor/project manager,
I want to verify my email when registering,
So that my offers are more secure | test | email verification as a contractor project manager i want to verify my email when registering so that my offers are more secure | 1 |
217,640 | 16,857,866,969 | IssuesEvent | 2021-06-21 09:09:50 | hazelcast/hazelcast | https://api.github.com/repos/hazelcast/hazelcast | closed | PhoneHomeIntegrationTest.testMapLatenciesWithMapStore test fails on Windows Oracle JDK8 | Source: Internal Team: Management Center Type: Test-Failure | **Failing test**
com.hazelcast.internal.util.phonehome.PhoneHomeIntegrationTest.testMapLatenciesWithMapStore
**Error message**
```
Expected 'mpgtlams' to be greater than or equal to 200, but was 199
```
**Stacktrace**
```
java.lang.AssertionError: Expected 'mpgtlams' to be greater than or equal to 200, but was 199
at org.junit.Assert.fail(Assert.java:89)
at org.junit.Assert.assertTrue(Assert.java:42)
at com.hazelcast.test.HazelcastTestSupport.assertGreaterOrEquals(HazelcastTestSupport.java:1398)
at com.hazelcast.internal.util.phonehome.PhoneHomeIntegrationTest.testMapLatenciesWithMapStore(PhoneHomeIntegrationTest.java:217)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
```
**Jenkins build:**
http://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-4.maintenance-Windows-OracleJDK8/311/ | 1.0 | PhoneHomeIntegrationTest.testMapLatenciesWithMapStore test fails on Windows Oracle JDK8 - **Failing test**
com.hazelcast.internal.util.phonehome.PhoneHomeIntegrationTest.testMapLatenciesWithMapStore
**Error message**
```
Expected 'mpgtlams' to be greater than or equal to 200, but was 199
```
**Stacktrace**
```
java.lang.AssertionError: Expected 'mpgtlams' to be greater than or equal to 200, but was 199
at org.junit.Assert.fail(Assert.java:89)
at org.junit.Assert.assertTrue(Assert.java:42)
at com.hazelcast.test.HazelcastTestSupport.assertGreaterOrEquals(HazelcastTestSupport.java:1398)
at com.hazelcast.internal.util.phonehome.PhoneHomeIntegrationTest.testMapLatenciesWithMapStore(PhoneHomeIntegrationTest.java:217)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59)
at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56)
at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115)
at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
```
**Jenkins build:**
http://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-4.maintenance-Windows-OracleJDK8/311/ | test | phonehomeintegrationtest testmaplatencieswithmapstore test fails on windows oracle failing test com hazelcast internal util phonehome phonehomeintegrationtest testmaplatencieswithmapstore error message expected mpgtlams to be greater than or equal to but was stacktrace java lang assertionerror expected mpgtlams to be greater than or equal to but was at org junit assert fail assert java at org junit assert asserttrue assert java at com hazelcast test hazelcasttestsupport assertgreaterorequals hazelcasttestsupport java at com hazelcast internal util phonehome phonehomeintegrationtest testmaplatencieswithmapstore phonehomeintegrationtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java jenkins build | 1 |
167,321 | 13,019,751,408 | IssuesEvent | 2020-07-27 00:25:58 | libra/libra | https://api.github.com/repos/libra/libra | closed | [cluster-test] Vault follow ups | cluster_test enhancement | Some things that are still missing from LSR/vault support
- [x] We create instances for LSR/Vault, we don't add them to `Cluster`. We need to add vault_instances and lsr_instances fields and appropriate methods on the `Cluster`. This is required if we want people to use them in experiments
- [x] Grouping API - we need some API on the cluster like `instances_for_group(ValidatorGroup)`
- [x] Support in version compatibility experiment - need to leverage grouping api and update LSR tag in the compatibility experiment
| 1.0 | [cluster-test] Vault follow ups - Some things that are still missing from LSR/vault support
- [x] We create instances for LSR/Vault, we don't add them to `Cluster`. We need to add vault_instances and lsr_instances fields and appropriate methods on the `Cluster`. This is required if we want people to use them in experiments
- [x] Grouping API - we need some API on the cluster like `instances_for_group(ValidatorGroup)`
- [x] Support in version compatibility experiment - need to leverage grouping api and update LSR tag in the compatibility experiment
| test | vault follow ups some things that are still missing from lsr vault support we create instances for lsr vault we don t add them to cluster we need to add vault instances and lsr instances fields and appropriate methods on the cluster this is required if we want people to use them in experiments grouping api we need some api on the cluster like instances for group validatorgroup support in version compatibility experiment need to leverage grouping api and update lsr tag in the compatibility experiment | 1 |
113,005 | 9,608,167,727 | IssuesEvent | 2019-05-12 03:03:52 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | roachtest: jepsen-batch3/multi-register/majority-ring failed | C-test-failure O-roachtest O-robot | SHA: https://github.com/cockroachdb/cockroach/commits/45e15e05abff25e099ca59f4c5cb40a6cf695e6d
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=jepsen-batch3/multi-register/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1285294&tab=buildLog
```
The test failed on branch=master, cloud=gce:
jepsen.go:260,jepsen.go:322,test.go:1251: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1285294-jepsen-batch3:6 -- bash -e -c "\
cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \
~/lein run test \
--tarball file://${PWD}/cockroach.tgz \
--username ${USER} \
--ssh-private-key ~/.ssh/id_rsa \
--os ubuntu \
--time-limit 300 \
--concurrency 30 \
--recovery-time 25 \
--test-count 1 \
-n 10.142.0.45 -n 10.142.0.4 -n 10.142.0.18 -n 10.142.0.25 -n 10.142.0.6 \
--test multi-register --nemesis majority-ring \
> invoke.log 2>&1 \
" returned:
stderr:
stdout:
Error: exit status 255
: exit status 1
``` | 2.0 | roachtest: jepsen-batch3/multi-register/majority-ring failed - SHA: https://github.com/cockroachdb/cockroach/commits/45e15e05abff25e099ca59f4c5cb40a6cf695e6d
Parameters:
To repro, try:
```
# Don't forget to check out a clean suitable branch and experiment with the
# stress invocation until the desired results present themselves. For example,
# using stress instead of stressrace and passing the '-p' stressflag which
# controls concurrency.
./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh
cd ~/go/src/github.com/cockroachdb/cockroach && \
stdbuf -oL -eL \
make stressrace TESTS=jepsen-batch3/multi-register/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log
```
Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1285294&tab=buildLog
```
The test failed on branch=master, cloud=gce:
jepsen.go:260,jepsen.go:322,test.go:1251: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1285294-jepsen-batch3:6 -- bash -e -c "\
cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \
~/lein run test \
--tarball file://${PWD}/cockroach.tgz \
--username ${USER} \
--ssh-private-key ~/.ssh/id_rsa \
--os ubuntu \
--time-limit 300 \
--concurrency 30 \
--recovery-time 25 \
--test-count 1 \
-n 10.142.0.45 -n 10.142.0.4 -n 10.142.0.18 -n 10.142.0.25 -n 10.142.0.6 \
--test multi-register --nemesis majority-ring \
> invoke.log 2>&1 \
" returned:
stderr:
stdout:
Error: exit status 255
: exit status 1
``` | test | roachtest jepsen multi register majority ring failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests jepsen multi register majority ring pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch master cloud gce jepsen go jepsen go test go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity jepsen bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test multi register nemesis majority ring invoke log returned stderr stdout error exit status exit status | 1 |
25,194 | 4,148,880,376 | IssuesEvent | 2016-06-15 12:45:59 | Gapminder/ddf-validation | https://api.github.com/repos/Gapminder/ddf-validation | closed | INVALID_DRILL_UP rule providing | effort1: medium (half-day) priority2: required status: done (tested) type: document type: enhancement type: feature type: rule | * **Rule name:** Should be filled if this issue depends on particular rule especially during new rule creation. This name should be a valid JS variable name and should be used for connecting with code.
`INVALID_DRILL_UP`
* **Rule test folder:** Should be filled if type of Request is `rule`. Needed for documentation auto generation.
`test/fixtures/rules-cases/invalid-drill_up`
* **Rule description:** Should be filled if type of Request is `rule`. Needed for documentation auto generation.
An issue according to this rule will be fired when drillup in concept is defined and not valid: not a set of valid concepts
* **Examples of correct data:**
`ddf--concepts.csv`
```
concept,concept_type,domain,name,drill_up
name,string,,
geo,entity_domain,,
region,entity_set,geo,Region
country,entity_set,geo,Country,['geographic_regions', 'income_groups', 'landlocked', 'g77_and_oecd_countries', 'geographic_regions_in_4_colors', 'main_religion_2008']
capital,entity_set,geo,Capital
pop,measure,geo,Population
year,time,,year
```
* **Examples of incorrect data:**
`ddf--concepts.csv`
```
concept,concept_type,domain,name,drill_up
name,string,,
geo,entity_domain,,
region,entity_set,geo,Region
country,entity_set,geo,Country,['foo', 'main_religion_2008']
capital,entity_set,geo,Capital,foo
pop,measure,geo,Population
year,time,,year
```
* **Scenarios** Should be filled if type of Request is `rule`.
```
when dataset is correct
any issue should NOT be found for this rule
```
```
when dataset is NOT correct
issues in accordance with wrong concept drillups quantity should be detected for this rule
output data for any issue should be expected
```
* **Output data format** Additional data that depends on particular issue type. Should be filled if type of Request is `rule`.
wrong drillup value
| 1.0 | INVALID_DRILL_UP rule providing - * **Rule name:** Should be filled if this issue depends on particular rule especially during new rule creation. This name should be a valid JS variable name and should be used for connecting with code.
`INVALID_DRILL_UP`
* **Rule test folder:** Should be filled if type of Request is `rule`. Needed for documentation auto generation.
`test/fixtures/rules-cases/invalid-drill_up`
* **Rule description:** Should be filled if type of Request is `rule`. Needed for documentation auto generation.
An issue according to this rule will be fired when drillup in concept is defined and not valid: not a set of valid concepts
* **Examples of correct data:**
`ddf--concepts.csv`
```
concept,concept_type,domain,name,drill_up
name,string,,
geo,entity_domain,,
region,entity_set,geo,Region
country,entity_set,geo,Country,['geographic_regions', 'income_groups', 'landlocked', 'g77_and_oecd_countries', 'geographic_regions_in_4_colors', 'main_religion_2008']
capital,entity_set,geo,Capital
pop,measure,geo,Population
year,time,,year
```
* **Examples of incorrect data:**
`ddf--concepts.csv`
```
concept,concept_type,domain,name,drill_up
name,string,,
geo,entity_domain,,
region,entity_set,geo,Region
country,entity_set,geo,Country,['foo', 'main_religion_2008']
capital,entity_set,geo,Capital,foo
pop,measure,geo,Population
year,time,,year
```
* **Scenarios** Should be filled if type of Request is `rule`.
```
when dataset is correct
any issue should NOT be found for this rule
```
```
when dataset is NOT correct
issues in accordance with wrong concept drillups quantity should be detected for this rule
output data for any issue should be expected
```
* **Output data format** Additional data that depends on particular issue type. Should be filled if type of Request is `rule`.
wrong drillup value
| test | invalid drill up rule providing rule name should be filled if this issue depends on particular rule especially during new rule creation this name should be a valid js variable name and should be used for connecting with code invalid drill up rule test folder should be filled if type of request is rule needed for documentation auto generation test fixtures rules cases invalid drill up rule description should be filled if type of request is rule needed for documentation auto generation an issue according to this rule will be fired when drillup in concept is defined and not valid not a set of valid concepts examples of correct data ddf concepts csv concept concept type domain name drill up name string geo entity domain region entity set geo region country entity set geo country capital entity set geo capital pop measure geo population year time year examples of incorrect data ddf concepts csv concept concept type domain name drill up name string geo entity domain region entity set geo region country entity set geo country capital entity set geo capital foo pop measure geo population year time year scenarios should be filled if type of request is rule when dataset is correct any issue should not be found for this rule when dataset is not correct issues in accordance with wrong concept drillups quantity should be detected for this rule output data for any issue should be expected output data format additional data that depends on particular issue type should be filled if type of request is rule wrong drillup value | 1 |
300,385 | 25,962,668,139 | IssuesEvent | 2022-12-19 02:01:58 | Ishiyuka/Kaizen | https://api.github.com/repos/Ishiyuka/Kaizen | opened | Rspecの実装 | test | **model Spec バリデーションテスト**
- [ ] user
- [ ] team
- [ ] issue
- [ ] plan
**System Spec **
- [ ] user
- [ ] team, assign
- [ ] issue
- [ ] plan
| 1.0 | Rspecの実装 - **model Spec バリデーションテスト**
- [ ] user
- [ ] team
- [ ] issue
- [ ] plan
**System Spec **
- [ ] user
- [ ] team, assign
- [ ] issue
- [ ] plan
| test | rspecの実装 model spec バリデーションテスト user team issue plan system spec user team assign issue plan | 1 |
331,126 | 24,292,676,415 | IssuesEvent | 2022-09-29 07:35:47 | DHCJS/DHCJS | https://api.github.com/repos/DHCJS/DHCJS | closed | Heading | documentation | <h1 align="center">Hartato Cipta Jaya</h1>

<h2 align="center">About Me</h2>
- 🍀Currently Learning HTML, CSS and JavaScript
- 🍮Former FnB Staff & Part-Time 📈Indonesia Stock Trader
- 🐑I Use [Medium](https://hartatociptajaya.medium.com/) To Blog My Journey, [Twitter](https://twitter.com/hartatocj) To Share My Achievements and Thoughts, and [LinkedIn](https://www.linkedin.com/in/hcj06/) To Connect. Feel Free To Connect With Me.
<h2>Programming Language</h2>

| 1.0 | Heading - <h1 align="center">Hartato Cipta Jaya</h1>

<h2 align="center">About Me</h2>
- 🍀Currently Learning HTML, CSS and JavaScript
- 🍮Former FnB Staff & Part-Time 📈Indonesia Stock Trader
- 🐑I Use [Medium](https://hartatociptajaya.medium.com/) To Blog My Journey, [Twitter](https://twitter.com/hartatocj) To Share My Achievements and Thoughts, and [LinkedIn](https://www.linkedin.com/in/hcj06/) To Connect. Feel Free To Connect With Me.
<h2>Programming Language</h2>

| non_test | heading hartato cipta jaya about me 🍀currently learning html css and javascript 🍮former fnb staff part time 📈indonesia stock trader 🐑i use to blog my journey to share my achievements and thoughts and to connect feel free to connect with me programming language | 0 |
98,458 | 12,325,454,671 | IssuesEvent | 2020-05-13 15:04:42 | codingdavinci/relaunch2018 | https://api.github.com/repos/codingdavinci/relaunch2018 | opened | Abschnitt Zeitleiste- Fußzeile text align bottom | design | The Fußzeile text is aligning top when it should align bottom, as it does for Abschnitt Box/Box-Gruppen. Plx fix, ty! | 1.0 | Abschnitt Zeitleiste- Fußzeile text align bottom - The Fußzeile text is aligning top when it should align bottom, as it does for Abschnitt Box/Box-Gruppen. Plx fix, ty! | non_test | abschnitt zeitleiste fußzeile text align bottom the fußzeile text is aligning top when it should align bottom as it does for abschnitt box box gruppen plx fix ty | 0 |
262,281 | 22,828,867,318 | IssuesEvent | 2022-07-12 11:05:23 | clash-lang/clash-compiler | https://api.github.com/repos/clash-lang/clash-compiler | opened | Re-enable Verilator simulation in clash-testsuite | bug tests | I introduced a bug so that [verilator no longer runs simulation](https://github.com/clash-lang/clash-compiler/pull/2257#discussion_r916653627).
We can also make some things more elegant, as Peter notes. | 1.0 | Re-enable Verilator simulation in clash-testsuite - I introduced a bug so that [verilator no longer runs simulation](https://github.com/clash-lang/clash-compiler/pull/2257#discussion_r916653627).
We can also make some things more elegant, as Peter notes. | test | re enable verilator simulation in clash testsuite i introduced a bug so that we can also make some things more elegant as peter notes | 1 |
171,679 | 14,342,861,395 | IssuesEvent | 2020-11-28 05:53:49 | extratone/xyz | https://api.github.com/repos/extratone/xyz | opened | Create "The Unlicense, Dave Edition" Page | documentation | # The Unlicense, Dave Edition
Considering my use of Git as a means to track revisions on my own writing, it's important to note that you should feel free to replace all instances of "software" in the statement below with "words," "writing," "expression," etc.
While I appreciate tremendously the work and contribution of [Arlo Bendiken](https://ar.to/2010/01/set-your-code-free) in the form of The Unlicense, I would like to add that I see no need, personally, to make the refutation of Intellectual Property law regarding my own work into some profound ethical statement. As it stands, I have not to my knowledge been the victim of any sort of theft of any kind, and find the suggestion highly unlikely. If I *am* made aware of such a case, there is certainly a possibility that I may change my mind on this issue, so I suppose I should highlight that **THESE TERMS ARE SUBJECT TO CHANGE**, though I suspect I would be more flattered and/or amused than actually offended.
I would also like to annotate that my use of this license should not be regarded as a suggestion that anyone else follow my example, or that I "believe in" taking such action regarding one's work, generally. While I did indeed take the time to set up [a repository](https://github.com/extratone/eoi) of John Perry Barlow's "[The Economy of Ideas](https://www.wired.com/1994/03/economy-ideas/)" in a bunch of different document formats, I do not necessarily agree with all of what he argued, especially in the context of the world 27 years later. If you *are* particularly interested in my opinion on the matter for whatever reason, here is what I have to say to you:
The principled beginning of The Open Web was a great moment in history, but - like all historical principles - it is extremely important that we consider *context* and maintain an appropriate level of criticism when looking back on old manifestos written by old white guys. In general, try your best to *be reasonable*.
***
This is free and unencumbered software released into the public domain.
Anyone is free to copy, modify, publish, use, compile, sell, or
distribute this software, either in source code form or as a compiled
binary, for any purpose, commercial or non-commercial, and by any
means.
In jurisdictions that recognize copyright laws, the author or authors
of this software dedicate any and all copyright interest in the
software to the public domain. We make this dedication for the benefit
of the public at large and to the detriment of our heirs and
successors. We intend this dedication to be an overt act of
relinquishment in perpetuity of all present and future rights to this
software under copyright law.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR
OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
OTHER DEALINGS IN THE SOFTWARE.
For more information, please refer to <https://unlicense.org>
| 1.0 | Create "The Unlicense, Dave Edition" Page - # The Unlicense, Dave Edition
Considering my use of Git as a means to track revisions on my own writing, it's important to note that you should feel free to replace all instances of "software" in the statement below with "words," "writing," "expression," etc.
While I appreciate tremendously the work and contribution of [Arlo Bendiken](https://ar.to/2010/01/set-your-code-free) in the form of The Unlicense, I would like to add that I see no need, personally, to make the refutation of Intellectual Property law regarding my own work into some profound ethical statement. As it stands, I have not to my knowledge been the victim of any sort of theft of any kind, and find the suggestion highly unlikely. If I *am* made aware of such a case, there is certainly a possibility that I may change my mind on this issue, so I suppose I should highlight that **THESE TERMS ARE SUBJECT TO CHANGE**, though I suspect I would be more flattered and/or amused than actually offended.
I would also like to annotate that my use of this license should not be regarded as a suggestion that anyone else follow my example, or that I "believe in" taking such action regarding one's work, generally. While I did indeed take the time to set up [a repository](https://github.com/extratone/eoi) of John Perry Barlow's "[The Economy of Ideas](https://www.wired.com/1994/03/economy-ideas/)" in a bunch of different document formats, I do not necessarily agree with all of what he argued, especially in the context of the world 27 years later. If you *are* particularly interested in my opinion on the matter for whatever reason, here is what I have to say to you:
The principled beginning of The Open Web was a great moment in history, but - like all historical principles - it is extremely important that we consider *context* and maintain an appropriate level of criticism when looking back on old manifestos written by old white guys. In general, try your best to *be reasonable*.
***
This is free and unencumbered software released into the public domain.
Anyone is free to copy, modify, publish, use, compile, sell, or
distribute this software, either in source code form or as a compiled
binary, for any purpose, commercial or non-commercial, and by any
means.
In jurisdictions that recognize copyright laws, the author or authors
of this software dedicate any and all copyright interest in the
software to the public domain. We make this dedication for the benefit
of the public at large and to the detriment of our heirs and
successors. We intend this dedication to be an overt act of
relinquishment in perpetuity of all present and future rights to this
software under copyright law.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.
IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR
OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
OTHER DEALINGS IN THE SOFTWARE.
For more information, please refer to <https://unlicense.org>
| non_test | create the unlicense dave edition page the unlicense dave edition considering my use of git as a means to track revisions on my own writing it s important to note that you should feel free to replace all instances of software in the statement below with words writing expression etc while i appreciate tremendously the work and contribution of in the form of the unlicense i would like to add that i see no need personally to make the refutation of intellectual property law regarding my own work into some profound ethical statement as it stands i have not to my knowledge been the victim of any sort of theft of any kind and find the suggestion highly unlikely if i am made aware of such a case there is certainly a possibility that i may change my mind on this issue so i suppose i should highlight that these terms are subject to change though i suspect i would be more flattered and or amused than actually offended i would also like to annotate that my use of this license should not be regarded as a suggestion that anyone else follow my example or that i believe in taking such action regarding one s work generally while i did indeed take the time to set up of john perry barlow s in a bunch of different document formats i do not necessarily agree with all of what he argued especially in the context of the world years later if you are particularly interested in my opinion on the matter for whatever reason here is what i have to say to you the principled beginning of the open web was a great moment in history but like all historical principles it is extremely important that we consider context and maintain an appropriate level of criticism when looking back on old manifestos written by old white guys in general try your best to be reasonable this is free and unencumbered software released into the public domain anyone is free to copy modify publish use compile sell or distribute this software either in source code form or as a compiled binary for any purpose commercial or non commercial and by any means in jurisdictions that recognize copyright laws the author or authors of this software dedicate any and all copyright interest in the software to the public domain we make this dedication for the benefit of the public at large and to the detriment of our heirs and successors we intend this dedication to be an overt act of relinquishment in perpetuity of all present and future rights to this software under copyright law the software is provided as is without warranty of any kind express or implied including but not limited to the warranties of merchantability fitness for a particular purpose and noninfringement in no event shall the authors be liable for any claim damages or other liability whether in an action of contract tort or otherwise arising from out of or in connection with the software or the use or other dealings in the software for more information please refer to | 0 |
150,426 | 11,960,386,275 | IssuesEvent | 2020-04-05 02:50:40 | tensorflow/addons | https://api.github.com/repos/tensorflow/addons | closed | TimeStopping callback is untested. | bug callbacks help wanted test-cases | We have no way of knowing if it works or not. Some minimal tests would be greatly appreciated.
See #964 | 1.0 | TimeStopping callback is untested. - We have no way of knowing if it works or not. Some minimal tests would be greatly appreciated.
See #964 | test | timestopping callback is untested we have no way of knowing if it works or not some minimal tests would be greatly appreciated see | 1 |
409,674 | 27,748,138,759 | IssuesEvent | 2023-03-15 18:31:13 | LibertyDSNP/spec | https://api.github.com/repos/LibertyDSNP/spec | closed | DIP-227 User Data Operations | documentation enhancement blocked DSNP | Create a DIP that describes the new Replace User Data and Get User Data operations, their parameters, and outputs, as well as the User Data Changed Record. | 1.0 | DIP-227 User Data Operations - Create a DIP that describes the new Replace User Data and Get User Data operations, their parameters, and outputs, as well as the User Data Changed Record. | non_test | dip user data operations create a dip that describes the new replace user data and get user data operations their parameters and outputs as well as the user data changed record | 0 |
293,177 | 25,275,304,640 | IssuesEvent | 2022-11-16 12:13:32 | wazuh/wazuh | https://api.github.com/repos/wazuh/wazuh | closed | Release 4.3.10 - Revision 1 - Release Candidate RC1 - Footprint Metrics - ALL (2.5d) | team/cicd type/release release test/4.3.10 | ## Footprint metrics information
| | |
|---------------------------------| :--:|
| **Main release candidate issue #** | #15372 |
| **Main footprint metrics issue #** | #15348 |
| **Version** | 4.3.10 |
| **Release candidate #** | RC1 |
| **Tag** | https://github.com/wazuh/wazuh/tree/4.3.10-rc1 |
## Stress test documentation
### Packages used
- Repository: `packages-dev.wazuh.com`
- Package path: `pre-release`
- Package revision: `1`
- **Jenkins build**: https://ci.wazuh.info/job/Test_stress/3655/
---
<details><summary>Manager</summary>
+ <details><summary>Plots</summary>
















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3655_manager_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/logs/ossec_Test_stress_B3655_manager_2022-11-14.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-manager-Test_stress_B3655_manager-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/monitor-manager-Test_stress_B3655_manager-pre-release.csv)
[Test_stress_B3655_manager_analysisd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/Test_stress_B3655_manager_analysisd_state.csv)
[Test_stress_B3655_manager_remoted_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/Test_stress_B3655_manager_remoted_state.csv)
</details>
</details>
<details><summary>Centos agent</summary>
+ <details><summary>Plots</summary>

















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3655_centos_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/logs/ossec_Test_stress_B3655_centos_2022-11-14.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-agent-Test_stress_B3655_centos-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/data/monitor-agent-Test_stress_B3655_centos-pre-release.csv)
[Test_stress_B3655_centos_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/data/Test_stress_B3655_centos_agentd_state.csv)
</details>
</details>
<details><summary>Ubuntu agent</summary>
+ <details><summary>Plots</summary>

















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3655_ubuntu_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/logs/ossec_Test_stress_B3655_ubuntu_2022-11-14.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-agent-Test_stress_B3655_ubuntu-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/data/monitor-agent-Test_stress_B3655_ubuntu-pre-release.csv)
[Test_stress_B3655_ubuntu_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/data/Test_stress_B3655_ubuntu_agentd_state.csv)
</details>
</details>
<details><summary>Windows agent</summary>
+ <details><summary>Plots</summary>















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3655_windows_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/logs/ossec_Test_stress_B3655_windows_2022-11-14.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-winagent-Test_stress_B3655_windows-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/data/monitor-winagent-Test_stress_B3655_windows-pre-release.csv)
[Test_stress_B3655_windows_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/data/Test_stress_B3655_windows_agentd_state.csv)
</details>
</details>
<details><summary>macOS agent</summary>
+ <details><summary>Plots</summary>
</details>
+ <details><summary>Logs and configuration</summary>
</details>
+ <details><summary>CSV</summary>
</details>
</details>
<details><summary>Solaris agent</summary>
+ <details><summary>Plots</summary>
</details>
+ <details><summary>Logs and configuration</summary>
</details>
+ <details><summary>CSV</summary>
</details>
</details> | 1.0 | Release 4.3.10 - Revision 1 - Release Candidate RC1 - Footprint Metrics - ALL (2.5d) - ## Footprint metrics information
| | |
|---------------------------------| :--:|
| **Main release candidate issue #** | #15372 |
| **Main footprint metrics issue #** | #15348 |
| **Version** | 4.3.10 |
| **Release candidate #** | RC1 |
| **Tag** | https://github.com/wazuh/wazuh/tree/4.3.10-rc1 |
## Stress test documentation
### Packages used
- Repository: `packages-dev.wazuh.com`
- Package path: `pre-release`
- Package revision: `1`
- **Jenkins build**: https://ci.wazuh.info/job/Test_stress/3655/
---
<details><summary>Manager</summary>
+ <details><summary>Plots</summary>
















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3655_manager_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/logs/ossec_Test_stress_B3655_manager_2022-11-14.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-manager-Test_stress_B3655_manager-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/monitor-manager-Test_stress_B3655_manager-pre-release.csv)
[Test_stress_B3655_manager_analysisd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/Test_stress_B3655_manager_analysisd_state.csv)
[Test_stress_B3655_manager_remoted_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/Test_stress_B3655_manager_remoted_state.csv)
</details>
</details>
<details><summary>Centos agent</summary>
+ <details><summary>Plots</summary>

















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3655_centos_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/logs/ossec_Test_stress_B3655_centos_2022-11-14.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-agent-Test_stress_B3655_centos-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/data/monitor-agent-Test_stress_B3655_centos-pre-release.csv)
[Test_stress_B3655_centos_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/data/Test_stress_B3655_centos_agentd_state.csv)
</details>
</details>
<details><summary>Ubuntu agent</summary>
+ <details><summary>Plots</summary>

















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3655_ubuntu_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/logs/ossec_Test_stress_B3655_ubuntu_2022-11-14.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-agent-Test_stress_B3655_ubuntu-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/data/monitor-agent-Test_stress_B3655_ubuntu-pre-release.csv)
[Test_stress_B3655_ubuntu_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/data/Test_stress_B3655_ubuntu_agentd_state.csv)
</details>
</details>
<details><summary>Windows agent</summary>
+ <details><summary>Plots</summary>















</details>
+ <details><summary>Logs and configuration</summary>
[ossec_Test_stress_B3655_windows_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/logs/ossec_Test_stress_B3655_windows_2022-11-14.zip)
</details>
+ <details><summary>CSV</summary>
[monitor-winagent-Test_stress_B3655_windows-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/data/monitor-winagent-Test_stress_B3655_windows-pre-release.csv)
[Test_stress_B3655_windows_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/data/Test_stress_B3655_windows_agentd_state.csv)
</details>
</details>
<details><summary>macOS agent</summary>
+ <details><summary>Plots</summary>
</details>
+ <details><summary>Logs and configuration</summary>
</details>
+ <details><summary>CSV</summary>
</details>
</details>
<details><summary>Solaris agent</summary>
+ <details><summary>Plots</summary>
</details>
+ <details><summary>Logs and configuration</summary>
</details>
+ <details><summary>CSV</summary>
</details>
</details> | test | release revision release candidate footprint metrics all footprint metrics information main release candidate issue main footprint metrics issue version release candidate tag stress test documentation packages used repository packages dev wazuh com package path pre release package revision jenkins build manager plots logs and configuration csv centos agent plots logs and configuration csv ubuntu agent plots logs and configuration csv windows agent plots logs and configuration csv macos agent plots logs and configuration csv solaris agent plots logs and configuration csv | 1 |
214,277 | 24,056,925,381 | IssuesEvent | 2022-09-16 17:52:48 | Agoric/agoric-sdk | https://api.github.com/repos/Agoric/agoric-sdk | opened | Await safety uncertain in cache/src/store.js | bug security | https://github.com/Agoric/agoric-sdk/blob/a1dedeae72908fda45afcb6038d76f8359adc8de/packages/cache/src/store.js#L82
The triage at https://github.com/Agoric/agoric-sdk/pull/6219 currently classifies this as
// TODO FIXME This code should be refactored to make this
// await checkably safe, or to remove it, or to record here
// why it is actually safe.
Git blame shows @turadg and @michaelfig as the ones who should probably investigate this, so I'm assigning to them. Feel free to reassign as appropriate of course. | True | Await safety uncertain in cache/src/store.js - https://github.com/Agoric/agoric-sdk/blob/a1dedeae72908fda45afcb6038d76f8359adc8de/packages/cache/src/store.js#L82
The triage at https://github.com/Agoric/agoric-sdk/pull/6219 currently classifies this as
// TODO FIXME This code should be refactored to make this
// await checkably safe, or to remove it, or to record here
// why it is actually safe.
Git blame shows @turadg and @michaelfig as the ones who should probably investigate this, so I'm assigning to them. Feel free to reassign as appropriate of course. | non_test | await safety uncertain in cache src store js the triage at currently classifies this as todo fixme this code should be refactored to make this await checkably safe or to remove it or to record here why it is actually safe git blame shows turadg and michaelfig as the ones who should probably investigate this so i m assigning to them feel free to reassign as appropriate of course | 0 |
279,334 | 24,217,117,906 | IssuesEvent | 2022-09-26 07:48:33 | elastic/kibana | https://api.github.com/repos/elastic/kibana | opened | [APM] Test Plan 8.5.0 | Team:APM apm:test-plan-guide v8.5.0 | # Test Plan 8.5.0
## Phase 1 - Manual Testing: Monday 26th September
- Pick an issue from the [curated list](https://github.com/elastic/kibana/issues?q=label%3Aapm%3Atest-plan-8.5.0+is%3Aclosed+-label%3Aapm%3Atest-plan-done)
- After testing apply `apm:test-plan-done` label. If a problem is found a new issue should be opened and labelled `apm:test-plan-regression` and moved to "Ready" column.
### Run latest build candidate (BC)
Using [apm-integration-testing](https://github.com/elastic/apm-integration-testing) it's trivial to start the stack for the latest BC:
```
./scripts/compose.py start 8.5.0 --bc --all-opbeans
```
#### Check to see if issues are fixed before opening a bug
Find the commit kibana was built with:
```
./scripts/compose.py versions
...
Kibana (image built: 2022-09-22 07:33:44 UTC):
Version: 8.5.0-SNAPSHOT
Branch: 8.5
Build SHA: -
Build number: -
```
Using that Build SHA, visit https://github.com/elastic/kibana/compare/2aa076a33a2d67a72152240aa69ee1d26ff2485b...8.5 to see commits since the BC was built.
### Creating users locally
```
node ./x-pack/plugins/apm/scripts/create_apm_users.js --username elastic --password changeme --kibana-url http://localhost:5601
```
Creates the following users (username / password):
- `viewer` / `changeme`
- `editor` / `changeme`
## Phase 2 - Improving Quality: Tuesday 27th September - Friday 30th September
Primarily we need more e2e (cypress) tests and api tests. In addition to this we should replace the archives containing random opbeans data with synthbeans generated data. Below is a list of the important flows that should be covered in both a unit and an api test.
### Important areas & flows
- Pending
### Migrate from opbeans archives to synthtrace scenarios
We should replace the existing opbeans-based archives with data generated at runtime with synthtrace.
### Migrate API test to `apmApiClient`
API tests should use the new `apmApiClient` instead of the legacy `supertest` clients. [Example on how to migrate.](https://github.com/elastic/kibana/issues/111308) | 1.0 | [APM] Test Plan 8.5.0 - # Test Plan 8.5.0
## Phase 1 - Manual Testing: Monday 26th September
- Pick an issue from the [curated list](https://github.com/elastic/kibana/issues?q=label%3Aapm%3Atest-plan-8.5.0+is%3Aclosed+-label%3Aapm%3Atest-plan-done)
- After testing apply `apm:test-plan-done` label. If a problem is found a new issue should be opened and labelled `apm:test-plan-regression` and moved to "Ready" column.
### Run latest build candidate (BC)
Using [apm-integration-testing](https://github.com/elastic/apm-integration-testing) it's trivial to start the stack for the latest BC:
```
./scripts/compose.py start 8.5.0 --bc --all-opbeans
```
#### Check to see if issues are fixed before opening a bug
Find the commit kibana was built with:
```
./scripts/compose.py versions
...
Kibana (image built: 2022-09-22 07:33:44 UTC):
Version: 8.5.0-SNAPSHOT
Branch: 8.5
Build SHA: -
Build number: -
```
Using that Build SHA, visit https://github.com/elastic/kibana/compare/2aa076a33a2d67a72152240aa69ee1d26ff2485b...8.5 to see commits since the BC was built.
### Creating users locally
```
node ./x-pack/plugins/apm/scripts/create_apm_users.js --username elastic --password changeme --kibana-url http://localhost:5601
```
Creates the following users (username / password):
- `viewer` / `changeme`
- `editor` / `changeme`
## Phase 2 - Improving Quality: Tuesday 27th September - Friday 30th September
Primarily we need more e2e (cypress) tests and api tests. In addition to this we should replace the archives containing random opbeans data with synthbeans generated data. Below is a list of the important flows that should be covered in both a unit and an api test.
### Important areas & flows
- Pending
### Migrate from opbeans archives to synthtrace scenarios
We should replace the existing opbeans-based archives with data generated at runtime with synthtrace.
### Migrate API test to `apmApiClient`
API tests should use the new `apmApiClient` instead of the legacy `supertest` clients. [Example on how to migrate.](https://github.com/elastic/kibana/issues/111308) | test | test plan test plan phase manual testing monday september pick an issue from the after testing apply apm test plan done label if a problem is found a new issue should be opened and labelled apm test plan regression and moved to ready column run latest build candidate bc using it s trivial to start the stack for the latest bc scripts compose py start bc all opbeans check to see if issues are fixed before opening a bug find the commit kibana was built with scripts compose py versions kibana image built utc version snapshot branch build sha build number using that build sha visit to see commits since the bc was built creating users locally node x pack plugins apm scripts create apm users js username elastic password changeme kibana url creates the following users username password viewer changeme editor changeme phase improving quality tuesday september friday september primarily we need more cypress tests and api tests in addition to this we should replace the archives containing random opbeans data with synthbeans generated data below is a list of the important flows that should be covered in both a unit and an api test important areas flows pending migrate from opbeans archives to synthtrace scenarios we should replace the existing opbeans based archives with data generated at runtime with synthtrace migrate api test to apmapiclient api tests should use the new apmapiclient instead of the legacy supertest clients | 1 |
98,815 | 8,685,473,625 | IssuesEvent | 2018-12-03 07:52:45 | humera987/FXLabs-Test-Automation | https://api.github.com/repos/humera987/FXLabs-Test-Automation | reopened | FX Testing 3 : ApiV1IssuesJobIdIdGetQueryParamPageInvalidDatatype | FX Testing 3 | Project : FX Testing 3
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=YzA4ZGZkNDItMjBkYS00NTk5LTkzMWItZDM2MDhmNWFhMjcw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 03 Dec 2018 07:45:48 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/issues/job-id/buLiWKdz?page=n2rClY&status=buLiWKdz
Request :
Response :
{
"timestamp" : "2018-12-03T07:45:49.053+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/issues/job-id/buLiWKdz"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | 1.0 | FX Testing 3 : ApiV1IssuesJobIdIdGetQueryParamPageInvalidDatatype - Project : FX Testing 3
Job : UAT
Env : UAT
Region : US_WEST
Result : fail
Status Code : 404
Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=YzA4ZGZkNDItMjBkYS00NTk5LTkzMWItZDM2MDhmNWFhMjcw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 03 Dec 2018 07:45:48 GMT]}
Endpoint : http://13.56.210.25/api/v1/api/v1/issues/job-id/buLiWKdz?page=n2rClY&status=buLiWKdz
Request :
Response :
{
"timestamp" : "2018-12-03T07:45:49.053+0000",
"status" : 404,
"error" : "Not Found",
"message" : "No message available",
"path" : "/api/v1/api/v1/issues/job-id/buLiWKdz"
}
Logs :
Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]
--- FX Bot --- | test | fx testing project fx testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api issues job id buliwkdz logs assertion resolved to result assertion resolved to result fx bot | 1 |
278,365 | 24,148,654,357 | IssuesEvent | 2022-09-21 21:22:19 | unicode-org/icu4x | https://api.github.com/repos/unicode-org/icu4x | closed | Add benches for data provider constructors | T-docs-tests A-performance C-data-infra S-small | The data provider constructors are a big part of our startup cost. We should measure them and improve them if possible.
I think it's fine for the tests to measure the testdata versions of the providers. | 1.0 | Add benches for data provider constructors - The data provider constructors are a big part of our startup cost. We should measure them and improve them if possible.
I think it's fine for the tests to measure the testdata versions of the providers. | test | add benches for data provider constructors the data provider constructors are a big part of our startup cost we should measure them and improve them if possible i think it s fine for the tests to measure the testdata versions of the providers | 1 |
63,345 | 14,656,703,165 | IssuesEvent | 2020-12-28 14:00:46 | fu1771695yongxie/react-motion | https://api.github.com/repos/fu1771695yongxie/react-motion | opened | WS-2019-0032 (Medium) detected in js-yaml-3.7.0.tgz | security vulnerability | ## WS-2019-0032 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p>
<p>Path to dependency file: react-motion/package.json</p>
<p>Path to vulnerable library: react-motion/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-1.1.2.tgz (Root Library)
- istanbul-0.4.5.tgz
- :x: **js-yaml-3.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/react-motion/commit/cc696c0e5f9cf48a30e0c7c74ea0d8811a5b1487">cc696c0e5f9cf48a30e0c7c74ea0d8811a5b1487</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.
<p>Publish Date: 2019-03-20
<p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p>
<p>Release Date: 2019-03-20</p>
<p>Fix Resolution: js-yaml - 3.13.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2019-0032 (Medium) detected in js-yaml-3.7.0.tgz - ## WS-2019-0032 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary>
<p>YAML 1.2 parser and serializer</p>
<p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p>
<p>Path to dependency file: react-motion/package.json</p>
<p>Path to vulnerable library: react-motion/node_modules/js-yaml/package.json</p>
<p>
Dependency Hierarchy:
- karma-coverage-1.1.2.tgz (Root Library)
- istanbul-0.4.5.tgz
- :x: **js-yaml-3.7.0.tgz** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/react-motion/commit/cc696c0e5f9cf48a30e0c7c74ea0d8811a5b1487">cc696c0e5f9cf48a30e0c7c74ea0d8811a5b1487</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service.
<p>Publish Date: 2019-03-20
<p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary>
<p>
Base Score Metrics not available</p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p>
<p>Release Date: 2019-03-20</p>
<p>Fix Resolution: js-yaml - 3.13.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws medium detected in js yaml tgz ws medium severity vulnerability vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file react motion package json path to vulnerable library react motion node modules js yaml package json dependency hierarchy karma coverage tgz root library istanbul tgz x js yaml tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions js yaml prior to are vulnerable to denial of service by parsing a carefully crafted yaml file the node process stalls and may exhaust system resources leading to a denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution js yaml step up your open source security game with whitesource | 0 |
165,425 | 12,842,243,984 | IssuesEvent | 2020-07-08 01:29:27 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | opened | roachtest: django failed | C-test-failure O-roachtest O-robot branch-provisional_202007071743_v20.2.0-alpha.2 release-blocker | [(roachtest).django failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2071705&tab=buildLog) on [provisional_202007071743_v20.2.0-alpha.2@0b6e118bc1bcba4cfb4fc6c660153ec5be3989e8](https://github.com/cockroachdb/cockroach/commits/0b6e118bc1bcba4cfb4fc6c660153ec5be3989e8):
```
--- FAIL: introspection.tests.IntrospectionTests.test_table_names_with_views (unexpected)
--- FAIL: queries.tests.ExcludeTest17600.test_only_orders_with_all_items_having_status_1 (unexpected)
--- FAIL: backends.tests.FkConstraintsTests.test_disable_constraint_checks_manually (unexpected)
--- FAIL: aggregation_regress.tests.AggregationTests.test_stddev (unexpected)
--- FAIL: expressions.tests.FTimeDeltaTests.test_datetime_subtraction_microseconds (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_numeric_field_keep_null_status (unexpected)
--- FAIL: schema.tests.SchemaTests.test_m2m_rename_field_in_target_model (unexpected)
--- FAIL: schema.tests.SchemaTests.test_db_table (unexpected)
--- FAIL: queries.tests.RelatedLookupTypeTests.test_values_queryset_lookup (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_order_with_respect_to (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_textual_field_keep_null_status (unexpected)
--- FAIL: auth_tests.test_views.UUIDUserTests.test_admin_password_change (unexpected)
--- FAIL: queries.tests.Ticket14056Tests.test_ticket_14056 (unexpected)
--- FAIL: queries.tests.SubqueryTests.test_slice_subquery_and_query (unexpected)
--- FAIL: many_to_one.tests.ManyToOneTests.test_fk_to_smallautofield (unexpected)
--- FAIL: aggregation_regress.tests.AggregationTests.test_ticket_11293 (unexpected)
--- FAIL: syndication_tests.tests.SyndicationFeedTest.test_latest_post_date (unexpected)
--- FAIL: queries.test_explain.ExplainTests.test_unknown_format (unexpected)
--- FAIL: db_functions.text.test_sha224.SHA224Tests.test_transform (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_not_unique_field_to_primary_key (unexpected)
--- FAIL: admin_filters.tests.ListFiltersTests.test_simplelistfilter_with_queryset_based_lookups (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_field_pk_fk (unexpected)
--- FAIL: syndication_tests.tests.SyndicationFeedTest.test_rss091_feed (unexpected)
--- FAIL: queries.tests.NullInExcludeTest.test_col_not_in_list_containing_null (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_fk_non_fk (unexpected)
--- FAIL: queries.test_bulk_update.BulkUpdateTests.test_inherited_fields (unexpected)
--- FAIL: aggregation_regress.tests.AggregationTests.test_more_more (unexpected)
--- FAIL: syndication_tests.tests.SyndicationFeedTest.test_rss2_feed (unexpected)
--- FAIL: ordering.tests.OrderingTests.test_order_by_fk_attname (unexpected)
--- FAIL: db_functions.math.test_power.PowerTests.test_integer (unexpected)
--- FAIL: backends.tests.FkConstraintsTests.test_disable_constraint_checks_context_manager (unexpected)
--- FAIL: queries.test_bulk_update.BulkUpdateNoteTests.test_multiple_fields (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_int_pk_to_int_unique (unexpected)
--- FAIL: db_functions.text.test_sha384.SHA384Tests.test_basic (unexpected)
--- FAIL: schema.tests.SchemaTests.test_add_field_temp_default (unexpected)
--- FAIL: backends.tests.FkConstraintsTests.test_check_constraints (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_field_reloads_state_on_fk_with_to_field_target_changes (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_field_reloads_state_on_fk_with_to_field_target_type_change (unexpected)
--- FAIL: introspection.tests.IntrospectionTests.test_get_table_description_types (unexpected)
--- FAIL: schema.tests.SchemaTests.test_char_field_with_db_index_to_fk (unexpected)
--- FAIL: db_functions.text.test_sha224.SHA224Tests.test_basic (unexpected)
--- FAIL: timezones.tests.LegacyDatabaseTests.test_cursor_execute_accepts_naive_datetime (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_text_field_to_time_field (unexpected)
--- FAIL: aggregation.test_filter_argument.FilteredAggregateTests.test_filtered_numerical_aggregates (unexpected)
--- FAIL: introspection.tests.IntrospectionTests.test_get_table_description_col_lengths (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_text_field_to_datetime_field (unexpected)
For a full summary look at the django artifacts
An updated blocklist (djangoBlocklist20_2) is available in the artifacts' django log
```
<details><summary>More</summary><p>
Artifacts: [/django](https://teamcity.cockroachdb.com/viewLog.html?buildId=2071705&tab=artifacts#/django)
Related:
- #51051 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #51049 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #50793 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006292135_v19.2.9](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006292135_v19.2.9) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46800 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adjango.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| 2.0 | roachtest: django failed - [(roachtest).django failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2071705&tab=buildLog) on [provisional_202007071743_v20.2.0-alpha.2@0b6e118bc1bcba4cfb4fc6c660153ec5be3989e8](https://github.com/cockroachdb/cockroach/commits/0b6e118bc1bcba4cfb4fc6c660153ec5be3989e8):
```
--- FAIL: introspection.tests.IntrospectionTests.test_table_names_with_views (unexpected)
--- FAIL: queries.tests.ExcludeTest17600.test_only_orders_with_all_items_having_status_1 (unexpected)
--- FAIL: backends.tests.FkConstraintsTests.test_disable_constraint_checks_manually (unexpected)
--- FAIL: aggregation_regress.tests.AggregationTests.test_stddev (unexpected)
--- FAIL: expressions.tests.FTimeDeltaTests.test_datetime_subtraction_microseconds (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_numeric_field_keep_null_status (unexpected)
--- FAIL: schema.tests.SchemaTests.test_m2m_rename_field_in_target_model (unexpected)
--- FAIL: schema.tests.SchemaTests.test_db_table (unexpected)
--- FAIL: queries.tests.RelatedLookupTypeTests.test_values_queryset_lookup (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_order_with_respect_to (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_textual_field_keep_null_status (unexpected)
--- FAIL: auth_tests.test_views.UUIDUserTests.test_admin_password_change (unexpected)
--- FAIL: queries.tests.Ticket14056Tests.test_ticket_14056 (unexpected)
--- FAIL: queries.tests.SubqueryTests.test_slice_subquery_and_query (unexpected)
--- FAIL: many_to_one.tests.ManyToOneTests.test_fk_to_smallautofield (unexpected)
--- FAIL: aggregation_regress.tests.AggregationTests.test_ticket_11293 (unexpected)
--- FAIL: syndication_tests.tests.SyndicationFeedTest.test_latest_post_date (unexpected)
--- FAIL: queries.test_explain.ExplainTests.test_unknown_format (unexpected)
--- FAIL: db_functions.text.test_sha224.SHA224Tests.test_transform (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_not_unique_field_to_primary_key (unexpected)
--- FAIL: admin_filters.tests.ListFiltersTests.test_simplelistfilter_with_queryset_based_lookups (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_field_pk_fk (unexpected)
--- FAIL: syndication_tests.tests.SyndicationFeedTest.test_rss091_feed (unexpected)
--- FAIL: queries.tests.NullInExcludeTest.test_col_not_in_list_containing_null (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_fk_non_fk (unexpected)
--- FAIL: queries.test_bulk_update.BulkUpdateTests.test_inherited_fields (unexpected)
--- FAIL: aggregation_regress.tests.AggregationTests.test_more_more (unexpected)
--- FAIL: syndication_tests.tests.SyndicationFeedTest.test_rss2_feed (unexpected)
--- FAIL: ordering.tests.OrderingTests.test_order_by_fk_attname (unexpected)
--- FAIL: db_functions.math.test_power.PowerTests.test_integer (unexpected)
--- FAIL: backends.tests.FkConstraintsTests.test_disable_constraint_checks_context_manager (unexpected)
--- FAIL: queries.test_bulk_update.BulkUpdateNoteTests.test_multiple_fields (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_int_pk_to_int_unique (unexpected)
--- FAIL: db_functions.text.test_sha384.SHA384Tests.test_basic (unexpected)
--- FAIL: schema.tests.SchemaTests.test_add_field_temp_default (unexpected)
--- FAIL: backends.tests.FkConstraintsTests.test_check_constraints (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_field_reloads_state_on_fk_with_to_field_target_changes (unexpected)
--- FAIL: migrations.test_operations.OperationTests.test_alter_field_reloads_state_on_fk_with_to_field_target_type_change (unexpected)
--- FAIL: introspection.tests.IntrospectionTests.test_get_table_description_types (unexpected)
--- FAIL: schema.tests.SchemaTests.test_char_field_with_db_index_to_fk (unexpected)
--- FAIL: db_functions.text.test_sha224.SHA224Tests.test_basic (unexpected)
--- FAIL: timezones.tests.LegacyDatabaseTests.test_cursor_execute_accepts_naive_datetime (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_text_field_to_time_field (unexpected)
--- FAIL: aggregation.test_filter_argument.FilteredAggregateTests.test_filtered_numerical_aggregates (unexpected)
--- FAIL: introspection.tests.IntrospectionTests.test_get_table_description_col_lengths (unexpected)
--- FAIL: schema.tests.SchemaTests.test_alter_text_field_to_datetime_field (unexpected)
For a full summary look at the django artifacts
An updated blocklist (djangoBlocklist20_2) is available in the artifacts' django log
```
<details><summary>More</summary><p>
Artifacts: [/django](https://teamcity.cockroachdb.com/viewLog.html?buildId=2071705&tab=artifacts#/django)
Related:
- #51051 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #51049 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #50793 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006292135_v19.2.9](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006292135_v19.2.9) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
- #46800 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker)
[See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adjango.%2A&sort=title&restgroup=false&display=lastcommented+project)
<sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
| test | roachtest django failed on fail introspection tests introspectiontests test table names with views unexpected fail queries tests test only orders with all items having status unexpected fail backends tests fkconstraintstests test disable constraint checks manually unexpected fail aggregation regress tests aggregationtests test stddev unexpected fail expressions tests ftimedeltatests test datetime subtraction microseconds unexpected fail schema tests schematests test alter numeric field keep null status unexpected fail schema tests schematests test rename field in target model unexpected fail schema tests schematests test db table unexpected fail queries tests relatedlookuptypetests test values queryset lookup unexpected fail schema tests schematests test alter unexpected fail migrations test operations operationtests test alter order with respect to unexpected fail schema tests schematests test alter textual field keep null status unexpected fail auth tests test views uuidusertests test admin password change unexpected fail queries tests test ticket unexpected fail queries tests subquerytests test slice subquery and query unexpected fail many to one tests manytoonetests test fk to smallautofield unexpected fail aggregation regress tests aggregationtests test ticket unexpected fail syndication tests tests syndicationfeedtest test latest post date unexpected fail queries test explain explaintests test unknown format unexpected fail db functions text test test transform unexpected fail schema tests schematests test alter not unique field to primary key unexpected fail admin filters tests listfilterstests test simplelistfilter with queryset based lookups unexpected fail migrations test operations operationtests test alter field pk fk unexpected fail syndication tests tests syndicationfeedtest test feed unexpected fail queries tests nullinexcludetest test col not in list containing null unexpected fail migrations test operations operationtests test alter fk non fk unexpected fail queries test bulk update bulkupdatetests test inherited fields unexpected fail aggregation regress tests aggregationtests test more more unexpected fail syndication tests tests syndicationfeedtest test feed unexpected fail ordering tests orderingtests test order by fk attname unexpected fail db functions math test power powertests test integer unexpected fail backends tests fkconstraintstests test disable constraint checks context manager unexpected fail queries test bulk update bulkupdatenotetests test multiple fields unexpected fail schema tests schematests test alter int pk to int unique unexpected fail db functions text test test basic unexpected fail schema tests schematests test add field temp default unexpected fail backends tests fkconstraintstests test check constraints unexpected fail migrations test operations operationtests test alter field reloads state on fk with to field target changes unexpected fail migrations test operations operationtests test alter field reloads state on fk with to field target type change unexpected fail introspection tests introspectiontests test get table description types unexpected fail schema tests schematests test char field with db index to fk unexpected fail db functions text test test basic unexpected fail timezones tests legacydatabasetests test cursor execute accepts naive datetime unexpected fail schema tests schematests test alter text field to time field unexpected fail aggregation test filter argument filteredaggregatetests test filtered numerical aggregates unexpected fail introspection tests introspectiontests test get table description col lengths unexpected fail schema tests schematests test alter text field to datetime field unexpected for a full summary look at the django artifacts an updated blocklist is available in the artifacts django log more artifacts related roachtest django failed roachtest django failed roachtest django failed roachtest django failed powered by | 1 |
149,073 | 11,881,602,227 | IssuesEvent | 2020-03-27 13:00:33 | emoncms/emoncms | https://api.github.com/repos/emoncms/emoncms | closed | Device Already Exists Error | testing required | https://github.com/emoncms/emoncms/blob/ccb3a41ed17879ff154dfcabf7fa58d010ef8dce/Modules/input/Views/input_view.js#L878
Getting these errors on a regular basis. Could be for any existing node.

It has also been reported on the community.
https://community.openenergymonitor.org/t/constant-device-already-exists-error-alert-poup/12714
Whilst clearing REDIS cache does fix it, it does not solve why it is happening.
The code has reached this point, because it does not think the device exists, yet, when it checks again, it does exist.
Something not right......
| 1.0 | Device Already Exists Error - https://github.com/emoncms/emoncms/blob/ccb3a41ed17879ff154dfcabf7fa58d010ef8dce/Modules/input/Views/input_view.js#L878
Getting these errors on a regular basis. Could be for any existing node.

It has also been reported on the community.
https://community.openenergymonitor.org/t/constant-device-already-exists-error-alert-poup/12714
Whilst clearing REDIS cache does fix it, it does not solve why it is happening.
The code has reached this point, because it does not think the device exists, yet, when it checks again, it does exist.
Something not right......
| test | device already exists error getting these errors on a regular basis could be for any existing node it has also been reported on the community whilst clearing redis cache does fix it it does not solve why it is happening the code has reached this point because it does not think the device exists yet when it checks again it does exist something not right | 1 |
433,169 | 30,316,471,410 | IssuesEvent | 2023-07-10 15:54:13 | supabase/supabase | https://api.github.com/repos/supabase/supabase | closed | Typo in Bangla README | documentation good first issue | # Improve documentation
## Link
https://github.com/supabase/supabase/blob/master/i18n/README.bn.md
## Describe the problem

Just follow the last line. The word `রিপুর` will be exchanged with `রেপোর`
## Describe the improvement
Change the spelling from `রিপুর` to `রেপোর`
## Additional context

| 1.0 | Typo in Bangla README - # Improve documentation
## Link
https://github.com/supabase/supabase/blob/master/i18n/README.bn.md
## Describe the problem

Just follow the last line. The word `রিপুর` will be exchanged with `রেপোর`
## Describe the improvement
Change the spelling from `রিপুর` to `রেপোর`
## Additional context

| non_test | typo in bangla readme improve documentation link describe the problem just follow the last line the word রিপুর will be exchanged with রেপোর describe the improvement change the spelling from রিপুর to রেপোর additional context | 0 |
281,949 | 21,315,450,923 | IssuesEvent | 2022-04-16 07:30:23 | kxshxsh/pe | https://api.github.com/repos/kxshxsh/pe | opened | Model component class diagram very complicated and labels tough to understand | severity.Medium type.DocumentationBug | 
Extemely messy class diagram. All I see is curved arrows and the words "* filered" and "* all" floating everywhere with no clear indication of which association arrow they belong to or even what purpose they add.
I don't know if * is for multiplicity or not, and it is why is it following "all" and "filtered" so closely?
<!--session: 1650087226793-f266947a-1da7-4629-8f86-32a293c5b0a1-->
<!--Version: Web v3.4.2--> | 1.0 | Model component class diagram very complicated and labels tough to understand - 
Extemely messy class diagram. All I see is curved arrows and the words "* filered" and "* all" floating everywhere with no clear indication of which association arrow they belong to or even what purpose they add.
I don't know if * is for multiplicity or not, and it is why is it following "all" and "filtered" so closely?
<!--session: 1650087226793-f266947a-1da7-4629-8f86-32a293c5b0a1-->
<!--Version: Web v3.4.2--> | non_test | model component class diagram very complicated and labels tough to understand extemely messy class diagram all i see is curved arrows and the words filered and all floating everywhere with no clear indication of which association arrow they belong to or even what purpose they add i don t know if is for multiplicity or not and it is why is it following all and filtered so closely | 0 |
124,425 | 10,311,233,361 | IssuesEvent | 2019-08-29 16:51:40 | rancher/rancher | https://api.github.com/repos/rancher/rancher | closed | Add windows node failed when rancher server url included port | [zube]: To Test area/windows kind/bug-qa team/cn | Rancher version: master-head
Step to reproduce:
1. run rancher sever with the command:
`docker run -p 8443:443 -p 8080:80 rancher/rancher:master-head`
2. enable windows
3. add windows node
4. then the agent container in windows got errors
```
INFO: https://52.197.xxx.xxx:8443 is accessible
Copy-Item : The filename, directory name, or volume label syntax is incorrect : 'C:\etc\docker\certs.d\52.197.xxx.xxx:8443\ca.crt'
At C:\etc\rancher\execute.ps1:299 char:5
+ Copy-Item -Force -Path "$sslCertDir\serverca" -Destination "$dock ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : NotSpecified: (:) [Copy-Item], IOException
+ FullyQualifiedErrorId : System.IO.IOException,Microsoft.PowerShell.Commands.CopyItemCommand
```
| 1.0 | Add windows node failed when rancher server url included port - Rancher version: master-head
Step to reproduce:
1. run rancher sever with the command:
`docker run -p 8443:443 -p 8080:80 rancher/rancher:master-head`
2. enable windows
3. add windows node
4. then the agent container in windows got errors
```
INFO: https://52.197.xxx.xxx:8443 is accessible
Copy-Item : The filename, directory name, or volume label syntax is incorrect : 'C:\etc\docker\certs.d\52.197.xxx.xxx:8443\ca.crt'
At C:\etc\rancher\execute.ps1:299 char:5
+ Copy-Item -Force -Path "$sslCertDir\serverca" -Destination "$dock ...
+ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
+ CategoryInfo : NotSpecified: (:) [Copy-Item], IOException
+ FullyQualifiedErrorId : System.IO.IOException,Microsoft.PowerShell.Commands.CopyItemCommand
```
| test | add windows node failed when rancher server url included port rancher version master head step to reproduce run rancher sever with the command docker run p p rancher rancher master head enable windows add windows node then the agent container in windows got errors info is accessible copy item the filename directory name or volume label syntax is incorrect c etc docker certs d xxx xxx ca crt at c etc rancher execute char copy item force path sslcertdir serverca destination dock categoryinfo notspecified ioexception fullyqualifiederrorid system io ioexception microsoft powershell commands copyitemcommand | 1 |
402,510 | 27,371,859,849 | IssuesEvent | 2023-02-28 00:34:29 | KA-devl/VueLovers-soen341project2023 | https://api.github.com/repos/KA-devl/VueLovers-soen341project2023 | closed | Employer Static Dashboard | documentation new feature Points: 2 | As an employer, I should have a dashboard where I can see the applicants to my job postings | 1.0 | Employer Static Dashboard - As an employer, I should have a dashboard where I can see the applicants to my job postings | non_test | employer static dashboard as an employer i should have a dashboard where i can see the applicants to my job postings | 0 |
274,022 | 23,803,865,772 | IssuesEvent | 2022-09-03 18:28:33 | doctrine/dbal | https://api.github.com/repos/doctrine/dbal | closed | DBAL-1168: Schema's getMigrateFromSql always adds CREATE SCHEMA | Missing Tests Bug PostgreSQL | Jira issue originally created by user vbence:
I originally posted this to Migrations; noticing that all the generated down() methods start with a "CREATE SCHEMA public" line.
Inspecting the return from Schema#getMigrateFromSql it indeed contains the create statement.
| 1.0 | DBAL-1168: Schema's getMigrateFromSql always adds CREATE SCHEMA - Jira issue originally created by user vbence:
I originally posted this to Migrations; noticing that all the generated down() methods start with a "CREATE SCHEMA public" line.
Inspecting the return from Schema#getMigrateFromSql it indeed contains the create statement.
| test | dbal schema s getmigratefromsql always adds create schema jira issue originally created by user vbence i originally posted this to migrations noticing that all the generated down methods start with a create schema public line inspecting the return from schema getmigratefromsql it indeed contains the create statement | 1 |
87,343 | 10,895,373,525 | IssuesEvent | 2019-11-19 10:31:33 | betagouv/demarches-simplifiees.fr | https://api.github.com/repos/betagouv/demarches-simplifiees.fr | closed | ETQ Usager, je veux voir le logo de la démarche correctement sur ie11 | a-communiquer contributions acceptées design usager | `hs #15254`
Version correcte :

Version IE11:

| 1.0 | ETQ Usager, je veux voir le logo de la démarche correctement sur ie11 - `hs #15254`
Version correcte :

Version IE11:

| non_test | etq usager je veux voir le logo de la démarche correctement sur hs version correcte version | 0 |
201,628 | 15,216,117,944 | IssuesEvent | 2021-02-17 15:09:36 | Princeton-CDH/geniza | https://api.github.com/repos/Princeton-CDH/geniza | reopened | As a global admin, I want a one-time import of the list of all languages and scripts, and their correlation, so that I can access, display and manage the information in the database. | 🆕 enhancement 🗜️ awaiting testing | ## testing notes
- Confirm that the list of languages + scripts has been correctly imported from the ontology spreadsheet
- Check that the edit history for a few records to confirm that there is a log entry showing the record was imported by script | 1.0 | As a global admin, I want a one-time import of the list of all languages and scripts, and their correlation, so that I can access, display and manage the information in the database. - ## testing notes
- Confirm that the list of languages + scripts has been correctly imported from the ontology spreadsheet
- Check that the edit history for a few records to confirm that there is a log entry showing the record was imported by script | test | as a global admin i want a one time import of the list of all languages and scripts and their correlation so that i can access display and manage the information in the database testing notes confirm that the list of languages scripts has been correctly imported from the ontology spreadsheet check that the edit history for a few records to confirm that there is a log entry showing the record was imported by script | 1 |
246,994 | 20,954,966,112 | IssuesEvent | 2022-03-27 01:31:25 | datafuselabs/databend | https://api.github.com/repos/datafuselabs/databend | closed | stateless-test `06_0002_show_create_database.sql` sometimes run slow | C-testing | In stateless-test, when running `06_0002_show_create_database.sql`, sometimes stuck.
cc @drmingdrmer | 1.0 | stateless-test `06_0002_show_create_database.sql` sometimes run slow - In stateless-test, when running `06_0002_show_create_database.sql`, sometimes stuck.
cc @drmingdrmer | test | stateless test show create database sql sometimes run slow in stateless test when running show create database sql sometimes stuck cc drmingdrmer | 1 |
137,573 | 5,311,805,337 | IssuesEvent | 2017-02-13 06:07:34 | BuiltBrokenModding/VoltzEngine | https://api.github.com/repos/BuiltBrokenModding/VoltzEngine | opened | Implement property system | enhancement High Priority | To improve the functionality of modifying common data we should switch over to using JSON files. These files will store information such as block name, localization key, hardness, resistance, and recipes. This way we do not need to worry about code implementation when handling each part. Instead, VE will read in the files and inject the data where it is needed. Additionally, this files can be used to register the block and item block without specifically the exacts for each.
Example file:
`{
"block": {
"name":"block",
"class":"BlockMeta",
"Item":"ItemBlockMeta",
"hardness":5,
"resistance":5,
"localization":"tile.#mod#.block"
"recipes":
{
"1":
{
"type":"shaped",
"output":
{
"item":"this",
"meta":0,
"nbt":"none"
},
"recipe":"abc-abc-abc",
"components":
{
"a":"apple",
"b":"bow:0",
"c": {
"item":"stone",
"meta":2
}
}
}
}
}` | 1.0 | Implement property system - To improve the functionality of modifying common data we should switch over to using JSON files. These files will store information such as block name, localization key, hardness, resistance, and recipes. This way we do not need to worry about code implementation when handling each part. Instead, VE will read in the files and inject the data where it is needed. Additionally, this files can be used to register the block and item block without specifically the exacts for each.
Example file:
`{
"block": {
"name":"block",
"class":"BlockMeta",
"Item":"ItemBlockMeta",
"hardness":5,
"resistance":5,
"localization":"tile.#mod#.block"
"recipes":
{
"1":
{
"type":"shaped",
"output":
{
"item":"this",
"meta":0,
"nbt":"none"
},
"recipe":"abc-abc-abc",
"components":
{
"a":"apple",
"b":"bow:0",
"c": {
"item":"stone",
"meta":2
}
}
}
}
}` | non_test | implement property system to improve the functionality of modifying common data we should switch over to using json files these files will store information such as block name localization key hardness resistance and recipes this way we do not need to worry about code implementation when handling each part instead ve will read in the files and inject the data where it is needed additionally this files can be used to register the block and item block without specifically the exacts for each example file block name block class blockmeta item itemblockmeta hardness resistance localization tile mod block recipes type shaped output item this meta nbt none recipe abc abc abc components a apple b bow c item stone meta | 0 |
129,787 | 10,586,620,287 | IssuesEvent | 2019-10-08 20:08:40 | eclipse/openj9 | https://api.github.com/repos/eclipse/openj9 | closed | Implement `TerminatingThreadLocal` as required by JTReg Test Failure: jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java | test excluded test failure | Failure link
------------
Rebuild Link: https://ci.adoptopenjdk.net/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&BUILD_LIST=openjdk&JenkinsFile=openjdk_x86-64_linux&CUSTOM_TARGET=jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java&TARGET=jdk_custom_0
- test category: openjdk
- OS/architecture: all
- version: 11, 13
- JVM: openj9 only
Optional info
-------------
- intermittent failure: No
- regression or new test: Regression
Failure output (captured from console output)
---------------------------------------------
```
14:54:51 java.lang.AssertionError: Expected terminated values: [42] but got: []
14:54:51 at TestTerminatingThreadLocal.ttlTest(TestTerminatingThreadLocal.java:86)
14:54:51 at TestTerminatingThreadLocal.ttlTestSet(TestTerminatingThreadLocal.java:48)
14:54:51 at TestTerminatingThreadLocal.main(TestTerminatingThreadLocal.java:41)
14:54:51 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
14:54:51 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
14:54:51 at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
14:54:51 at java.base/java.lang.reflect.Method.invoke(Method.java:566)
14:54:51 at com.sun.javatest.regtest.agent.MainActionHelper$AgentVMRunnable.run(MainActionHelper.java:298)
14:54:51 at java.base/java.lang.Thread.run(Thread.java:831)
```
After further investigation, it would seem that this method
https://github.com/ibmruntimes/openj9-openjdk-jdk11/blob/8bf2a8166185765e9c6ddbffb66d1adbee67fd16/test/jdk/jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java#L66 is not being called by any of the test cases, so `terminatedValues` remains empty and causes the test to fail.
| 2.0 | Implement `TerminatingThreadLocal` as required by JTReg Test Failure: jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java - Failure link
------------
Rebuild Link: https://ci.adoptopenjdk.net/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&BUILD_LIST=openjdk&JenkinsFile=openjdk_x86-64_linux&CUSTOM_TARGET=jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java&TARGET=jdk_custom_0
- test category: openjdk
- OS/architecture: all
- version: 11, 13
- JVM: openj9 only
Optional info
-------------
- intermittent failure: No
- regression or new test: Regression
Failure output (captured from console output)
---------------------------------------------
```
14:54:51 java.lang.AssertionError: Expected terminated values: [42] but got: []
14:54:51 at TestTerminatingThreadLocal.ttlTest(TestTerminatingThreadLocal.java:86)
14:54:51 at TestTerminatingThreadLocal.ttlTestSet(TestTerminatingThreadLocal.java:48)
14:54:51 at TestTerminatingThreadLocal.main(TestTerminatingThreadLocal.java:41)
14:54:51 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
14:54:51 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
14:54:51 at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
14:54:51 at java.base/java.lang.reflect.Method.invoke(Method.java:566)
14:54:51 at com.sun.javatest.regtest.agent.MainActionHelper$AgentVMRunnable.run(MainActionHelper.java:298)
14:54:51 at java.base/java.lang.Thread.run(Thread.java:831)
```
After further investigation, it would seem that this method
https://github.com/ibmruntimes/openj9-openjdk-jdk11/blob/8bf2a8166185765e9c6ddbffb66d1adbee67fd16/test/jdk/jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java#L66 is not being called by any of the test cases, so `terminatedValues` remains empty and causes the test to fail.
| test | implement terminatingthreadlocal as required by jtreg test failure jdk internal misc terminatingthreadlocal testterminatingthreadlocal java failure link rebuild link test category openjdk os architecture all version jvm only optional info intermittent failure no regression or new test regression failure output captured from console output java lang assertionerror expected terminated values but got at testterminatingthreadlocal ttltest testterminatingthreadlocal java at testterminatingthreadlocal ttltestset testterminatingthreadlocal java at testterminatingthreadlocal main testterminatingthreadlocal java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at com sun javatest regtest agent mainactionhelper agentvmrunnable run mainactionhelper java at java base java lang thread run thread java after further investigation it would seem that this method is not being called by any of the test cases so terminatedvalues remains empty and causes the test to fail | 1 |
45,332 | 11,633,933,681 | IssuesEvent | 2020-02-28 09:21:44 | feelpp/feelpp | https://api.github.com/repos/feelpp/feelpp | opened | Support clang-10 | module:build type:feature | clang-10 is out and should be supported asap.
here is the current issue
```
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/quickstart/qs_hdg_laplacian.cpp:24:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feel.hpp:58:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/filters.hpp:30:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/savegmshmesh.hpp:32:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/exportergmsh.hpp:41:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/exporter.hpp:39:
/home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feeldiscr/timeset.hpp:369:39: error: no viable overloaded '='
M_complex[sanitize(name)] = __s;
~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~~
/usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:378:7: note: candidate function not viable: no known conversion from 'const Feel::complex_type' (aka 'const complex<double>') to 'const std::pair<std::complex<double>, bool>' for 1st argument
operator=(typename conditional<
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:389:7: note: candidate function not viable: no known conversion from 'const Feel::complex_type' (aka 'const complex<double>') to 'std::pair<std::complex<double>, bool>' for 1st argument
operator=(typename conditional<
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:405:2: note: candidate template ignored: could not match 'pair' against 'complex'
operator=(const pair<_U1, _U2>& __p)
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:416:2: note: candidate template ignored: could not match 'pair' against 'complex'
operator=(pair<_U1, _U2>&& __p)
^
``` | 1.0 | Support clang-10 - clang-10 is out and should be supported asap.
here is the current issue
```
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/quickstart/qs_hdg_laplacian.cpp:24:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feel.hpp:58:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/filters.hpp:30:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/savegmshmesh.hpp:32:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/exportergmsh.hpp:41:
In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/exporter.hpp:39:
/home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feeldiscr/timeset.hpp:369:39: error: no viable overloaded '='
M_complex[sanitize(name)] = __s;
~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~~
/usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:378:7: note: candidate function not viable: no known conversion from 'const Feel::complex_type' (aka 'const complex<double>') to 'const std::pair<std::complex<double>, bool>' for 1st argument
operator=(typename conditional<
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:389:7: note: candidate function not viable: no known conversion from 'const Feel::complex_type' (aka 'const complex<double>') to 'std::pair<std::complex<double>, bool>' for 1st argument
operator=(typename conditional<
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:405:2: note: candidate template ignored: could not match 'pair' against 'complex'
operator=(const pair<_U1, _U2>& __p)
^
/usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:416:2: note: candidate template ignored: could not match 'pair' against 'complex'
operator=(pair<_U1, _U2>&& __p)
^
``` | non_test | support clang clang is out and should be supported asap here is the current issue in file included from home prudhomm devel feelpp clean feelpp quickstart qs hdg laplacian cpp in file included from home prudhomm devel feelpp clean feelpp feel feel feel hpp in file included from home prudhomm devel feelpp clean feelpp feel feel feelfilters filters hpp in file included from home prudhomm devel feelpp clean feelpp feel feel feelfilters savegmshmesh hpp in file included from home prudhomm devel feelpp clean feelpp feel feel feelfilters exportergmsh hpp in file included from home prudhomm devel feelpp clean feelpp feel feel feelfilters exporter hpp home prudhomm devel feelpp clean feelpp feel feel feeldiscr timeset hpp error no viable overloaded m complex s usr bin lib gcc linux gnu include c bits stl pair h note candidate function not viable no known conversion from const feel complex type aka const complex to const std pair bool for argument operator typename conditional usr bin lib gcc linux gnu include c bits stl pair h note candidate function not viable no known conversion from const feel complex type aka const complex to std pair bool for argument operator typename conditional usr bin lib gcc linux gnu include c bits stl pair h note candidate template ignored could not match pair against complex operator const pair p usr bin lib gcc linux gnu include c bits stl pair h note candidate template ignored could not match pair against complex operator pair p | 0 |
215,349 | 16,666,971,423 | IssuesEvent | 2021-06-07 06:04:11 | yyLeaves/DS-Project | https://api.github.com/repos/yyLeaves/DS-Project | opened | Tests and Debugs | tests |
# Tests
| Part | Tester1 | Tester2 |
| --------- | ------- | ------- |
| Event 1&2 | | |
| Event 3 | L | |
| Event 4 | | |
| Event 5 | | |
| Event 6 | | |
| Event 0 | | |
# TODO tests
## L
- [ ] Own part & with GUI part
- [ ] event 3
- [ ] -
- [ ] -
## N
- [ ] Own part & with GUI part
- [ ] -
- [ ] -
- [ ] -
## T
- [ ] Own part & with GUI part
- [ ] -
- [ ] -
- [ ] -
## C
- [ ] Own part & with GUI part
- [ ] -
- [ ] -
- [ ] -
| 1.0 | Tests and Debugs -
# Tests
| Part | Tester1 | Tester2 |
| --------- | ------- | ------- |
| Event 1&2 | | |
| Event 3 | L | |
| Event 4 | | |
| Event 5 | | |
| Event 6 | | |
| Event 0 | | |
# TODO tests
## L
- [ ] Own part & with GUI part
- [ ] event 3
- [ ] -
- [ ] -
## N
- [ ] Own part & with GUI part
- [ ] -
- [ ] -
- [ ] -
## T
- [ ] Own part & with GUI part
- [ ] -
- [ ] -
- [ ] -
## C
- [ ] Own part & with GUI part
- [ ] -
- [ ] -
- [ ] -
| test | tests and debugs tests part event event l event event event event todo tests l own part with gui part event n own part with gui part t own part with gui part c own part with gui part | 1 |
57,797 | 24,230,869,393 | IssuesEvent | 2022-09-26 18:09:17 | hashicorp/terraform-provider-aws | https://api.github.com/repos/hashicorp/terraform-provider-aws | closed | aws_acm_certificate - allow validation from a "validation domain" | enhancement service/acm | AWS terraform resource ```aws_acm_certificate```
Can this please be updated to allow a ```validation domain``` to be included when using email validation.
The link below shows the AWS CLI documentation regarding a ```validation domain```
https://docs.aws.amazon.com/cli/latest/reference/acm/resend-validation-email.html
Something like this in Terraform. This should allow the validation emails to be sent to @example.com rather than @my.example.com
```
resource "aws_acm_certificate" "aws_acm_certificate" {
domain_name = "my.example.com"
validation_method = "EMAIL"
validation_domain = "example.com"
}
````
| 1.0 | aws_acm_certificate - allow validation from a "validation domain" - AWS terraform resource ```aws_acm_certificate```
Can this please be updated to allow a ```validation domain``` to be included when using email validation.
The link below shows the AWS CLI documentation regarding a ```validation domain```
https://docs.aws.amazon.com/cli/latest/reference/acm/resend-validation-email.html
Something like this in Terraform. This should allow the validation emails to be sent to @example.com rather than @my.example.com
```
resource "aws_acm_certificate" "aws_acm_certificate" {
domain_name = "my.example.com"
validation_method = "EMAIL"
validation_domain = "example.com"
}
````
| non_test | aws acm certificate allow validation from a validation domain aws terraform resource aws acm certificate can this please be updated to allow a validation domain to be included when using email validation the link below shows the aws cli documentation regarding a validation domain something like this in terraform this should allow the validation emails to be sent to example com rather than my example com resource aws acm certificate aws acm certificate domain name my example com validation method email validation domain example com | 0 |
325,461 | 27,879,929,864 | IssuesEvent | 2023-03-21 18:34:24 | brave/brave-browser | https://api.github.com/repos/brave/brave-browser | closed | `brave_ads.enabled_last_profile` is `false` when only profile left has ads enabled | bug needs-investigation priority/P3 QA/Yes release-notes/exclude QA/Test-Plan-Specified feature/ads OS/Desktop | Follow up to: https://github.com/brave/brave-browser/issues/26864
This happens after deleting profile that ads were never enabled
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Profile1: Enabled ads
1. Profile2: Do not able ads
1. Run Brave
1. Delete Profile2 using Profile Manager (make sure you don't activate the window of Profile1)
1. Check `brave_ads.enabled_last_profile` in local state
## Actual result:
<!--Please add screenshots if needed-->
`brave_ads.enabled_last_profile` is `false`
Note: the state is fixed when window with Profile1 is activated
## Expected result:
`brave_ads.enabled_last_profile` is `true`, because only profile left has ads enabled
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily reproduced
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.49.91 Chromium: 110.0.5481.77 (Official Build) beta (64-bit)
-- | --
Revision | 65ed616c6e8ee3fe0ad64fe83796c020644d42af-refs/branch-heads/5481@{#839}
OS | Ubuntu 18.04 LTS
cc @tmancey @aseren @DJAndries | 1.0 | `brave_ads.enabled_last_profile` is `false` when only profile left has ads enabled - Follow up to: https://github.com/brave/brave-browser/issues/26864
This happens after deleting profile that ads were never enabled
## Steps to Reproduce
<!--Please add a series of steps to reproduce the issue-->
1. Profile1: Enabled ads
1. Profile2: Do not able ads
1. Run Brave
1. Delete Profile2 using Profile Manager (make sure you don't activate the window of Profile1)
1. Check `brave_ads.enabled_last_profile` in local state
## Actual result:
<!--Please add screenshots if needed-->
`brave_ads.enabled_last_profile` is `false`
Note: the state is fixed when window with Profile1 is activated
## Expected result:
`brave_ads.enabled_last_profile` is `true`, because only profile left has ads enabled
## Reproduces how often:
<!--[Easily reproduced/Intermittent issue/No steps to reproduce]-->
Easily reproduced
## Brave version (brave://version info)
<!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details-->
Brave | 1.49.91 Chromium: 110.0.5481.77 (Official Build) beta (64-bit)
-- | --
Revision | 65ed616c6e8ee3fe0ad64fe83796c020644d42af-refs/branch-heads/5481@{#839}
OS | Ubuntu 18.04 LTS
cc @tmancey @aseren @DJAndries | test | brave ads enabled last profile is false when only profile left has ads enabled follow up to this happens after deleting profile that ads were never enabled steps to reproduce enabled ads do not able ads run brave delete using profile manager make sure you don t activate the window of check brave ads enabled last profile in local state actual result brave ads enabled last profile is false note the state is fixed when window with is activated expected result brave ads enabled last profile is true because only profile left has ads enabled reproduces how often easily reproduced brave version brave version info brave chromium official build beta bit revision refs branch heads os ubuntu lts cc tmancey aseren djandries | 1 |
261,267 | 8,228,924,559 | IssuesEvent | 2018-09-07 07:39:57 | mozilla/addons-frontend | https://api.github.com/repos/mozilla/addons-frontend | opened | Add smooth transitions to the AMInstallButton | component: installation contrib: welcome priority: p3 | We should make the `AMInstallButton` look and feel slightly better by improving the transitions when the button switches from the "uninstalled" to "pending" to "installed" states. | 1.0 | Add smooth transitions to the AMInstallButton - We should make the `AMInstallButton` look and feel slightly better by improving the transitions when the button switches from the "uninstalled" to "pending" to "installed" states. | non_test | add smooth transitions to the aminstallbutton we should make the aminstallbutton look and feel slightly better by improving the transitions when the button switches from the uninstalled to pending to installed states | 0 |
303,724 | 26,226,239,325 | IssuesEvent | 2023-01-04 18:58:24 | rancher/dashboard | https://api.github.com/repos/rancher/dashboard | closed | [UI] - Search box - Select on Enter | area/navigation [zube]: To Test kind/enhancement good-first-issue | Rancher 2.6.4 - UI
On the Search box

In this sample, when input "data", and press enter : Nothing happends
It would be great to add "data" to the filter on Enter press, cause one item match exactly :

And also clean the filter box to let user input another filter, and so on... | 1.0 | [UI] - Search box - Select on Enter - Rancher 2.6.4 - UI
On the Search box

In this sample, when input "data", and press enter : Nothing happends
It would be great to add "data" to the filter on Enter press, cause one item match exactly :

And also clean the filter box to let user input another filter, and so on... | test | search box select on enter rancher ui on the search box in this sample when input data and press enter nothing happends it would be great to add data to the filter on enter press cause one item match exactly and also clean the filter box to let user input another filter and so on | 1 |
12,188 | 3,257,019,449 | IssuesEvent | 2015-10-20 16:03:53 | akvo/akvo-caddisfly | https://api.github.com/repos/akvo/akvo-caddisfly | closed | 2.12 Create user interface for strip test | Strip test | The user interface for the strip test should include:
1) an easy way to select the type of test to be done
2) a way to visualise the steps that need to be taken, and where the user is in the current phase
3) realtime feedback on the quality of the image (illumination, saturation, focus)
4) a way to display the test the user is currently doing, as an extra check
5) after the image is captured, a way to display the results including the calibration scale
6) an optional visual 'walk through' that the user can do to see how the process works. | 1.0 | 2.12 Create user interface for strip test - The user interface for the strip test should include:
1) an easy way to select the type of test to be done
2) a way to visualise the steps that need to be taken, and where the user is in the current phase
3) realtime feedback on the quality of the image (illumination, saturation, focus)
4) a way to display the test the user is currently doing, as an extra check
5) after the image is captured, a way to display the results including the calibration scale
6) an optional visual 'walk through' that the user can do to see how the process works. | test | create user interface for strip test the user interface for the strip test should include an easy way to select the type of test to be done a way to visualise the steps that need to be taken and where the user is in the current phase realtime feedback on the quality of the image illumination saturation focus a way to display the test the user is currently doing as an extra check after the image is captured a way to display the results including the calibration scale an optional visual walk through that the user can do to see how the process works | 1 |
21,295 | 6,133,269,612 | IssuesEvent | 2017-06-25 12:20:18 | TEAMMATES/teammates | https://api.github.com/repos/TEAMMATES/teammates | opened | Use href="javascript:;" to indicate anchor tags without link | a-CodeQuality d.Contributors t-HTML | ```html
<a href="#">Text</a>
```
The above is commonly used such that `Text` gains the appearance of a link while it does not actually link anywhere. However, this is the wrong method because `#` actually links to the top of the web page. The correct way to suppress such link is with `href="javascript:;"`.
P.S. Of course, there might be cases where the intention _is_ to go to the top of the page. This needs to be examined case-by-case. | 1.0 | Use href="javascript:;" to indicate anchor tags without link - ```html
<a href="#">Text</a>
```
The above is commonly used such that `Text` gains the appearance of a link while it does not actually link anywhere. However, this is the wrong method because `#` actually links to the top of the web page. The correct way to suppress such link is with `href="javascript:;"`.
P.S. Of course, there might be cases where the intention _is_ to go to the top of the page. This needs to be examined case-by-case. | non_test | use href javascript to indicate anchor tags without link html text the above is commonly used such that text gains the appearance of a link while it does not actually link anywhere however this is the wrong method because actually links to the top of the web page the correct way to suppress such link is with href javascript p s of course there might be cases where the intention is to go to the top of the page this needs to be examined case by case | 0 |
328,049 | 28,099,993,966 | IssuesEvent | 2023-03-30 18:41:03 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | closed | DISABLED test_variant_consistency_jit_expand_cuda_complex64 (__main__.TestJitCUDA) | triaged module: flaky-tests skipped module: unknown | Platforms: win, windows
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_variant_consistency_jit_expand_cuda_complex64&suite=TestJitCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/12024506828).
Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_variant_consistency_jit_expand_cuda_complex64`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_ops_jit.py` | 1.0 | DISABLED test_variant_consistency_jit_expand_cuda_complex64 (__main__.TestJitCUDA) - Platforms: win, windows
This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_variant_consistency_jit_expand_cuda_complex64&suite=TestJitCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/12024506828).
Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes.
**Debugging instructions (after clicking on the recent samples link):**
DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs.
To find relevant log snippets:
1. Click on the workflow logs linked above
2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work.
3. Grep for `test_variant_consistency_jit_expand_cuda_complex64`
4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs.
Test file path: `test_ops_jit.py` | test | disabled test variant consistency jit expand cuda main testjitcuda platforms win windows this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test variant consistency jit expand cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test ops jit py | 1 |
347,683 | 31,240,814,047 | IssuesEvent | 2023-08-20 21:02:31 | MohistMC/Mohist | https://api.github.com/repos/MohistMC/Mohist | closed | [1.20.1] 在更新实体速度时的Minecraft世界计时崩溃 | Wait Needs Testing 1.12.2 | <!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.-->
<!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.-->
<!-- If you don't know where to upload your logs and crash reports, you can use these websites : -->
<!-- https://gist.github.com (recommended) -->
<!-- https://mclo.gs -->
<!-- https://haste.mohistmc.com -->
<!-- https://pastebin.com -->
<!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT -->
**Minecraft Version :** 1.20.1
**Mohist Version :** 357
**Operating System :** win10
**Concerned mod / plugin** : Plugins (60): Law, PlaceholderAPI, VoidWorld, CustomPlaceholder, Yum, AntiLoop, BBSToper, Ip2region, NBTAPI, MiniMOTD, LuckPerms, DreamCleaner, CrazyAuctionsPlus, PlayerChat, CMILib, ItemNBTEdit, LoneLibs, AlwaysKeepInventory, ProtocolLib, Notbuild, LiteSignIn, Snow-BetterRTP, MiaoBoard, Vault, WTF, SbPlaceholder, BKCommonLib, PlayerPoints, EconomicBalance, ItemCommand, PlayerTitle, DemonMarket, MyCommand, WorldEdit, PlayerKits, DecentHolograms, Multiverse-Core, AuthMe, SakuraPurchasePlugin, CMI, AllMusic, CMIEInjector, CoreProtect, MysqlEcoBridge, Towny, WorldBorder, PocketGames, Residence, ItemsAdder, LockettePro, BanItem, TownyGUI, PlayerGuild, PlayerTop, TrMenu, YeqiLib, DeluxeMenus, PlugManX (PlugMan), Mohist, Mohist 模组:[14:28:55 INFO]: Mod List:
• mods_folder [前置][神秘主义][聪明大脑库]SmartBrainLib-fabric-1.20-1.11.jar : smartbrainlib (1.11) - 1
• mods_folder [前置][动态联合]CyclopsCore-1.20.1-1.18.8.jar : cyclopscore (1.18.8) - 1
• mods_folder [天空村庄] SkyVillages-1.0.2-1.20.1-forge-release.jar : skyvillages (1.0.2-1.20.1-forge) - 1
• mods_folder [前置][炼金化学]alchemylib-1.20.1-1.0.29.jar : alchemylib (1.0.29) - 1
• mods_folder [前置][合成扩展]Cucumber-1.20.1-7.0.0.jar : cucumber (7.0.0) - 1
• mods_folder [前置]geckolib-forge-1.20.1-4.2.1.jar : geckolib (4.2.1) - 1
• mods_folder [前置][铁魔法]player-animation-lib-forge-1.0.2-rc1+1.20.jar : playeranimator (1.0.2-rc1+1.20) - 1
• mods_folder [合成扩展]ExtendedCrafting-1.20.1-6.0.2.jar : extendedcrafting (6.0.2) - 1
• mods_folder [更多实用工具]scalar_utils-1.20.1-5.4.10.jar : scalar_utils (1.20.1-5.4.10) - 1
• mods_folder [网络发包修复]connectivity-1.20.1-4.4.jar : connectivity (1.20.1-4.4) - 1
• mods_folder [是,史蒂夫模型] yesstevemodel-1.20-release-1.1.5-hotfix-2.jar : yes_steve_model (1.1.5-hotfix-2) - 1
• mods_folder [前置][铁魔法]caelus-forge-3.1.0+1.20.jar : caelus (3.1.0+1.20) - 1
• mods_folder [更多结构]Structory_1.20.1_v1.3.2.jar : structory (1.3.2) - 1
• mods_folder [懒人厨房] cookingforblockheads-forge-1.20-16.0.0.jar : cookingforblockheads (16.0.0) - 1
• mods_folder [现代化修复] modernfix-forge-5.6.0+mc1.20.1.jar : modernfix (5.6.0+mc1.20.1) - 1
• mods_folder [JEI物品管理器]jei-1.20.1-forge-15.2.0.25.jar : jei (15.2.0.25) - 1
• mods_folder [服务端][前置]Placebo-1.20.1-8.3.2.jar : placebo (8.3.2) - 1
• mods_folder [自然罗盘/生物群系指南针] NaturesCompass-1.20.1-1.11.1-forge.jar : naturescompass (1.0.0) - 1
• mods_folder [大力能源]Powah-5.0.1.jar : powah (5.0.1) - 1
• JarInJar mixinextras-forge-0.2.0-beta.9.jar : mixinextras (0.2.0-beta.9) - 1
• mods_folder [前置][柠檬核心]lemon_core-1.20.1-1.0.21.jar : lemon_core (1.20.1-1.0.21) - 1
• mods_folder [魔法使]mahoutsukai-1.20.1-v1.34.53.jar : mahoutsukai (1.20.1-v1.34.53) - 1
• mods_folder [前置][懒人厨房]balm-forge-1.20-7.0.4.jar : balm (7.0.4) - 1
• mods_folder [合成表]CraftTweaker-forge-1.20.1-14.0.12.jar : crafttweaker (14.0.12) - 1
• mods_folder [更多磁盘] ExtraDisks-1.20.1-3.0.0.jar : extradisks (1.20.1-3.0.0) - 1
• mods_folder [原神仪器]genshinstrument-1.20.1-3.1.jar : genshinstrument (3.1) - 1
• mods_folder [前置]cloth-config-11.0.99-forge.jar : cloth_config (11.0.99) - 1
• minecraft forge-1.20.1-47.1.70-universal.jar : forge (47.1.70) - 2
• mods_folder [精致存储] refinedstorage-1.12.2.jar : refinedstorage (1.12.2) - 1
• mods_folder [服务端][机械动力扳手修复]windy-fixes-0.1.0.jar : windy_fixes (0.1.0) - 1
• mods_folder [更多箱子]ironchest-1.20.1-14.4.4.jar : ironchest (1.20.1-14.4.4) - 1
• mods_folder [炼金化学]alchemistry-1.20.1-2.3.3.jar : alchemistry (2.3.3) - 1
• mods_folder [动态联合][集成动力学]IntegratedCrafting-1.20.1-1.1.4.jar : integratedcrafting (1.1.4) - 1
• minecraft server-1.20.1-20230612.114412-srg.jar : minecraft (1.20.1) - 1
• mods_folder [农夫乐事] FarmersDelight-1.20.1-1.2.2a.jar : farmersdelight (1.20.1-1.2.2a) - 1
• mods_folder [农夫乐事][末地乐事] ends_delight-1.20.1-1.0.jar : ends_delight (1.0) - 1
• mods_folder [前置][集成动力]CommonCapabilities-1.20.1-2.9.0.jar : commoncapabilities (2.9.0) - 1
• mods_folder [前置][生物群系]TerraBlender-forge-1.20.1-3.0.0.169.jar : terrablender (3.0.0.169) - 1
• mods_folder [汉堡包]burgermod-2.9.4-1.20.1.jar : burgermod (2.9.4) - 1
• mods_folder [前置][妖怪之山通用库] mysterious_mountain_lib-1.2.7-1.20.jar : mysterious_mountain_lib (1.2.7-1.20) - 1
• mods_folder [更多生物群系]BiomesOPlenty-1.20.1-18.0.0.592.jar : biomesoplenty (18.0.0.592) - 1
• mods_folder [前置][创造核心]CreativeCore_FORGE_v2.10.31_mc1.20.1.jar : creativecore (2.10.31) - 1
• JarInJar spectrelib-forge-0.13.13+1.20.1.jar : spectrelib (0.13.13+1.20.1) - 1
• mods_folder [前置]supermartijn642corelib-1.1.9-forge-mc1.20.jar : supermartijn642corelib (1.1.9) - 1
• mods_folder [炼金术]theurgy-1.20.1-1.3.10.jar : theurgy (1.3.10) - 1
• mods_folder [植物魔法]Botania-1.20.1-441-FORGE-SNAPSHOT.jar : botania (1.20.1-441-FORGE-SNAPSHOT) - 1
• mods_folder [倒置火把] ceilingtorch-1.20-1.25.jar : ceilingtorch (1.25) - 1
• mods_folder [末影接口]EnderIO-1.20.1-6.0.14-alpha.jar : enderio (6.0.14-alpha) - 1
• mods_folder [服务端]spark-1.10.43-forge.jar : spark (1.10.43) - 1
• mods_folder [前置][饰品栏]curios-forge-5.2.0-beta.3+1.20.1.jar : curios (5.2.0-beta.3+1.20.1) - 1
• mods_folder [通用机械]Mekanism-1.20.1-10.3.9.homebaked-all.jar : mekanism (10.3.9) - 5
• mods_folder [赛🐎娘]umapyoi-0.2.15-1.20.jar : umapyoi (0.2.15-1.20) - 1
• mods_folder [管道]pipez-1.20.1-1.1.4.jar : pipez (1.20.1-1.1.4) - 1
• JarInJar flywheel-forge-1.20-0.6.9-4.jar : flywheel (0.6.9-4) - 1
• mods_folder [机械动力] create-1.20.1-0.5.1.d.jar : create (0.5.1.d) - 1
• mods_folder [前置][帕特抽手册]Patchouli-1.20.1-82-FORGE-SNAPSHOT.jar : patchouli (1.20.1-82-FORGE-SNAPSHOT) - 1
• mods_folder [动态联合]IntegratedDynamics-1.20.1-1.16.10.jar : integrateddynamics (1.16.10) - 2
• mods_folder [前置][化学]chemlib-1.20.1-2.0.17.jar : chemlib (2.0.17) - 1
• mods_folder [多态合成] polymorph-forge-0.49.0+1.20.1.jar : polymorph (0.49.0+1.20.1) - 1
• mods_folder [通量网络] FluxNetworks-1.20.1-7.2.0.14.jar : fluxnetworks (7.2.0.14) - 1
• mods_folder [铁魔法]irons_spellbooks-1.20.1-1.2.0.jar : irons_spellbooks (1.20.1-1.2.0) - 1
• mods_folder [前置][手册]modonomicon-1.20.1-forge-1.38.5.jar : modonomicon (1.38.5) - 1
• mods_folder [飞车奇匠] automobility-0.4.2+1.20.1-forge.jar : automobility (0.4.2+1.20.1-forge) - 1
• mods_folder [前置]architectury-9.0.8-forge.jar : architectury (9.0.8) - 1
• mods_folder [前置][葡园酒香]doapi-1.2.6-forge.jar : doapi (1.2.6) - 2
• mods_folder [葡园酒香] letsdo-vinery-forge-1.4.2.jar : vinery (1.4.2) - 1
• mods_folder [方块小镇]yuushya-1.20.1-forge-2.0.0.jar : yuushya (2.0.0) - 1
• mods_folder [农夫乐事][海洋乐事]oceansdelight-1.0.2-1.20.jar : oceansdelight (1.0.2-1.20) - 1
• mods_folder [铁氧体磁芯] ferritecore-6.0.0-forge.jar : ferritecore (6.0.0) - 1
• mods_folder [神秘主义]occultism-1.20.1-1.81.6.jar : occultism (1.81.6) - 1
• mods_folder [卡哇伊]KawaiiDishes_1.10.1-1.20.1.jar : kawaiidishes (1.10.1) - 1
• mods_folder [农夫乐事][玉米乐事] corn_delight-1.0.3-1.20.1.jar : corn_delight (1.0.3-1.20.1) - 1
• mods_folder [前置][循环]flib-1.20.1-0.0.9.jar : flib (0.0.9) - 1
• JarInJar l2library-2.4.14-slim.jar : l2library (2.4.14) - 1
• mods_folder [精致存储附属] refinedstorageaddons-0.10.0.jar : refinedstorageaddons (0.10.0) - 1
• mods_folder [农夫乐事][料理乐事]cuisinedelight-1.1.8.jar : cuisinedelight (1.1.8) - 1
• mods_folder [豆腐工艺]TofuCraftReload-1.20.1-5.4.0.0.jar : tofucraft (1.20.1-5.4.0.0) - 1
• mods_folder [循环]Cyclic-1.20.1-1.12.2.jar : cyclic (1.12.2) - 1
**Logs :**
[crash-2023-08-20_14.07.08-server.txt](https://github.com/MohistMC/Mohist/files/12387390/crash-2023-08-20_14.07.08-server.txt)
[debug.log](https://github.com/MohistMC/Mohist/files/12387391/debug.log)
**Steps to Reproduce :**
1. 正常启动Minecraft服务器
2. 在游戏中进行常规操作
3.服务器在尝试设置某个实体的速度时崩溃
**Description of issue :** 在正常的服务器操作中,服务器因与"Exception ticking world"相关的错误而崩溃。主要的错误是"java.lang.IllegalArgumentException: x not finite",这个错误似乎是在尝试设置某个实体的速度时触发的。
| 1.0 | [1.20.1] 在更新实体速度时的Minecraft世界计时崩溃 - <!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.-->
<!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.-->
<!-- If you don't know where to upload your logs and crash reports, you can use these websites : -->
<!-- https://gist.github.com (recommended) -->
<!-- https://mclo.gs -->
<!-- https://haste.mohistmc.com -->
<!-- https://pastebin.com -->
<!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT -->
**Minecraft Version :** 1.20.1
**Mohist Version :** 357
**Operating System :** win10
**Concerned mod / plugin** : Plugins (60): Law, PlaceholderAPI, VoidWorld, CustomPlaceholder, Yum, AntiLoop, BBSToper, Ip2region, NBTAPI, MiniMOTD, LuckPerms, DreamCleaner, CrazyAuctionsPlus, PlayerChat, CMILib, ItemNBTEdit, LoneLibs, AlwaysKeepInventory, ProtocolLib, Notbuild, LiteSignIn, Snow-BetterRTP, MiaoBoard, Vault, WTF, SbPlaceholder, BKCommonLib, PlayerPoints, EconomicBalance, ItemCommand, PlayerTitle, DemonMarket, MyCommand, WorldEdit, PlayerKits, DecentHolograms, Multiverse-Core, AuthMe, SakuraPurchasePlugin, CMI, AllMusic, CMIEInjector, CoreProtect, MysqlEcoBridge, Towny, WorldBorder, PocketGames, Residence, ItemsAdder, LockettePro, BanItem, TownyGUI, PlayerGuild, PlayerTop, TrMenu, YeqiLib, DeluxeMenus, PlugManX (PlugMan), Mohist, Mohist 模组:[14:28:55 INFO]: Mod List:
• mods_folder [前置][神秘主义][聪明大脑库]SmartBrainLib-fabric-1.20-1.11.jar : smartbrainlib (1.11) - 1
• mods_folder [前置][动态联合]CyclopsCore-1.20.1-1.18.8.jar : cyclopscore (1.18.8) - 1
• mods_folder [天空村庄] SkyVillages-1.0.2-1.20.1-forge-release.jar : skyvillages (1.0.2-1.20.1-forge) - 1
• mods_folder [前置][炼金化学]alchemylib-1.20.1-1.0.29.jar : alchemylib (1.0.29) - 1
• mods_folder [前置][合成扩展]Cucumber-1.20.1-7.0.0.jar : cucumber (7.0.0) - 1
• mods_folder [前置]geckolib-forge-1.20.1-4.2.1.jar : geckolib (4.2.1) - 1
• mods_folder [前置][铁魔法]player-animation-lib-forge-1.0.2-rc1+1.20.jar : playeranimator (1.0.2-rc1+1.20) - 1
• mods_folder [合成扩展]ExtendedCrafting-1.20.1-6.0.2.jar : extendedcrafting (6.0.2) - 1
• mods_folder [更多实用工具]scalar_utils-1.20.1-5.4.10.jar : scalar_utils (1.20.1-5.4.10) - 1
• mods_folder [网络发包修复]connectivity-1.20.1-4.4.jar : connectivity (1.20.1-4.4) - 1
• mods_folder [是,史蒂夫模型] yesstevemodel-1.20-release-1.1.5-hotfix-2.jar : yes_steve_model (1.1.5-hotfix-2) - 1
• mods_folder [前置][铁魔法]caelus-forge-3.1.0+1.20.jar : caelus (3.1.0+1.20) - 1
• mods_folder [更多结构]Structory_1.20.1_v1.3.2.jar : structory (1.3.2) - 1
• mods_folder [懒人厨房] cookingforblockheads-forge-1.20-16.0.0.jar : cookingforblockheads (16.0.0) - 1
• mods_folder [现代化修复] modernfix-forge-5.6.0+mc1.20.1.jar : modernfix (5.6.0+mc1.20.1) - 1
• mods_folder [JEI物品管理器]jei-1.20.1-forge-15.2.0.25.jar : jei (15.2.0.25) - 1
• mods_folder [服务端][前置]Placebo-1.20.1-8.3.2.jar : placebo (8.3.2) - 1
• mods_folder [自然罗盘/生物群系指南针] NaturesCompass-1.20.1-1.11.1-forge.jar : naturescompass (1.0.0) - 1
• mods_folder [大力能源]Powah-5.0.1.jar : powah (5.0.1) - 1
• JarInJar mixinextras-forge-0.2.0-beta.9.jar : mixinextras (0.2.0-beta.9) - 1
• mods_folder [前置][柠檬核心]lemon_core-1.20.1-1.0.21.jar : lemon_core (1.20.1-1.0.21) - 1
• mods_folder [魔法使]mahoutsukai-1.20.1-v1.34.53.jar : mahoutsukai (1.20.1-v1.34.53) - 1
• mods_folder [前置][懒人厨房]balm-forge-1.20-7.0.4.jar : balm (7.0.4) - 1
• mods_folder [合成表]CraftTweaker-forge-1.20.1-14.0.12.jar : crafttweaker (14.0.12) - 1
• mods_folder [更多磁盘] ExtraDisks-1.20.1-3.0.0.jar : extradisks (1.20.1-3.0.0) - 1
• mods_folder [原神仪器]genshinstrument-1.20.1-3.1.jar : genshinstrument (3.1) - 1
• mods_folder [前置]cloth-config-11.0.99-forge.jar : cloth_config (11.0.99) - 1
• minecraft forge-1.20.1-47.1.70-universal.jar : forge (47.1.70) - 2
• mods_folder [精致存储] refinedstorage-1.12.2.jar : refinedstorage (1.12.2) - 1
• mods_folder [服务端][机械动力扳手修复]windy-fixes-0.1.0.jar : windy_fixes (0.1.0) - 1
• mods_folder [更多箱子]ironchest-1.20.1-14.4.4.jar : ironchest (1.20.1-14.4.4) - 1
• mods_folder [炼金化学]alchemistry-1.20.1-2.3.3.jar : alchemistry (2.3.3) - 1
• mods_folder [动态联合][集成动力学]IntegratedCrafting-1.20.1-1.1.4.jar : integratedcrafting (1.1.4) - 1
• minecraft server-1.20.1-20230612.114412-srg.jar : minecraft (1.20.1) - 1
• mods_folder [农夫乐事] FarmersDelight-1.20.1-1.2.2a.jar : farmersdelight (1.20.1-1.2.2a) - 1
• mods_folder [农夫乐事][末地乐事] ends_delight-1.20.1-1.0.jar : ends_delight (1.0) - 1
• mods_folder [前置][集成动力]CommonCapabilities-1.20.1-2.9.0.jar : commoncapabilities (2.9.0) - 1
• mods_folder [前置][生物群系]TerraBlender-forge-1.20.1-3.0.0.169.jar : terrablender (3.0.0.169) - 1
• mods_folder [汉堡包]burgermod-2.9.4-1.20.1.jar : burgermod (2.9.4) - 1
• mods_folder [前置][妖怪之山通用库] mysterious_mountain_lib-1.2.7-1.20.jar : mysterious_mountain_lib (1.2.7-1.20) - 1
• mods_folder [更多生物群系]BiomesOPlenty-1.20.1-18.0.0.592.jar : biomesoplenty (18.0.0.592) - 1
• mods_folder [前置][创造核心]CreativeCore_FORGE_v2.10.31_mc1.20.1.jar : creativecore (2.10.31) - 1
• JarInJar spectrelib-forge-0.13.13+1.20.1.jar : spectrelib (0.13.13+1.20.1) - 1
• mods_folder [前置]supermartijn642corelib-1.1.9-forge-mc1.20.jar : supermartijn642corelib (1.1.9) - 1
• mods_folder [炼金术]theurgy-1.20.1-1.3.10.jar : theurgy (1.3.10) - 1
• mods_folder [植物魔法]Botania-1.20.1-441-FORGE-SNAPSHOT.jar : botania (1.20.1-441-FORGE-SNAPSHOT) - 1
• mods_folder [倒置火把] ceilingtorch-1.20-1.25.jar : ceilingtorch (1.25) - 1
• mods_folder [末影接口]EnderIO-1.20.1-6.0.14-alpha.jar : enderio (6.0.14-alpha) - 1
• mods_folder [服务端]spark-1.10.43-forge.jar : spark (1.10.43) - 1
• mods_folder [前置][饰品栏]curios-forge-5.2.0-beta.3+1.20.1.jar : curios (5.2.0-beta.3+1.20.1) - 1
• mods_folder [通用机械]Mekanism-1.20.1-10.3.9.homebaked-all.jar : mekanism (10.3.9) - 5
• mods_folder [赛🐎娘]umapyoi-0.2.15-1.20.jar : umapyoi (0.2.15-1.20) - 1
• mods_folder [管道]pipez-1.20.1-1.1.4.jar : pipez (1.20.1-1.1.4) - 1
• JarInJar flywheel-forge-1.20-0.6.9-4.jar : flywheel (0.6.9-4) - 1
• mods_folder [机械动力] create-1.20.1-0.5.1.d.jar : create (0.5.1.d) - 1
• mods_folder [前置][帕特抽手册]Patchouli-1.20.1-82-FORGE-SNAPSHOT.jar : patchouli (1.20.1-82-FORGE-SNAPSHOT) - 1
• mods_folder [动态联合]IntegratedDynamics-1.20.1-1.16.10.jar : integrateddynamics (1.16.10) - 2
• mods_folder [前置][化学]chemlib-1.20.1-2.0.17.jar : chemlib (2.0.17) - 1
• mods_folder [多态合成] polymorph-forge-0.49.0+1.20.1.jar : polymorph (0.49.0+1.20.1) - 1
• mods_folder [通量网络] FluxNetworks-1.20.1-7.2.0.14.jar : fluxnetworks (7.2.0.14) - 1
• mods_folder [铁魔法]irons_spellbooks-1.20.1-1.2.0.jar : irons_spellbooks (1.20.1-1.2.0) - 1
• mods_folder [前置][手册]modonomicon-1.20.1-forge-1.38.5.jar : modonomicon (1.38.5) - 1
• mods_folder [飞车奇匠] automobility-0.4.2+1.20.1-forge.jar : automobility (0.4.2+1.20.1-forge) - 1
• mods_folder [前置]architectury-9.0.8-forge.jar : architectury (9.0.8) - 1
• mods_folder [前置][葡园酒香]doapi-1.2.6-forge.jar : doapi (1.2.6) - 2
• mods_folder [葡园酒香] letsdo-vinery-forge-1.4.2.jar : vinery (1.4.2) - 1
• mods_folder [方块小镇]yuushya-1.20.1-forge-2.0.0.jar : yuushya (2.0.0) - 1
• mods_folder [农夫乐事][海洋乐事]oceansdelight-1.0.2-1.20.jar : oceansdelight (1.0.2-1.20) - 1
• mods_folder [铁氧体磁芯] ferritecore-6.0.0-forge.jar : ferritecore (6.0.0) - 1
• mods_folder [神秘主义]occultism-1.20.1-1.81.6.jar : occultism (1.81.6) - 1
• mods_folder [卡哇伊]KawaiiDishes_1.10.1-1.20.1.jar : kawaiidishes (1.10.1) - 1
• mods_folder [农夫乐事][玉米乐事] corn_delight-1.0.3-1.20.1.jar : corn_delight (1.0.3-1.20.1) - 1
• mods_folder [前置][循环]flib-1.20.1-0.0.9.jar : flib (0.0.9) - 1
• JarInJar l2library-2.4.14-slim.jar : l2library (2.4.14) - 1
• mods_folder [精致存储附属] refinedstorageaddons-0.10.0.jar : refinedstorageaddons (0.10.0) - 1
• mods_folder [农夫乐事][料理乐事]cuisinedelight-1.1.8.jar : cuisinedelight (1.1.8) - 1
• mods_folder [豆腐工艺]TofuCraftReload-1.20.1-5.4.0.0.jar : tofucraft (1.20.1-5.4.0.0) - 1
• mods_folder [循环]Cyclic-1.20.1-1.12.2.jar : cyclic (1.12.2) - 1
**Logs :**
[crash-2023-08-20_14.07.08-server.txt](https://github.com/MohistMC/Mohist/files/12387390/crash-2023-08-20_14.07.08-server.txt)
[debug.log](https://github.com/MohistMC/Mohist/files/12387391/debug.log)
**Steps to Reproduce :**
1. 正常启动Minecraft服务器
2. 在游戏中进行常规操作
3.服务器在尝试设置某个实体的速度时崩溃
**Description of issue :** 在正常的服务器操作中,服务器因与"Exception ticking world"相关的错误而崩溃。主要的错误是"java.lang.IllegalArgumentException: x not finite",这个错误似乎是在尝试设置某个实体的速度时触发的。
| test | 在更新实体速度时的minecraft世界计时崩溃 important do not delete this line minecraft version mohist version operating system concerned mod plugin plugins law placeholderapi voidworld customplaceholder yum antiloop bbstoper nbtapi minimotd luckperms dreamcleaner crazyauctionsplus playerchat cmilib itemnbtedit lonelibs alwayskeepinventory protocollib notbuild litesignin snow betterrtp miaoboard vault wtf sbplaceholder bkcommonlib playerpoints economicbalance itemcommand playertitle demonmarket mycommand worldedit playerkits decentholograms multiverse core authme sakurapurchaseplugin cmi allmusic cmieinjector coreprotect mysqlecobridge towny worldborder pocketgames residence itemsadder lockettepro banitem townygui playerguild playertop trmenu yeqilib deluxemenus plugmanx plugman mohist mohist 模组: mod list • mods folder smartbrainlib fabric jar smartbrainlib • mods folder cyclopscore jar cyclopscore • mods folder skyvillages forge release jar skyvillages forge • mods folder alchemylib jar alchemylib • mods folder cucumber jar cucumber • mods folder geckolib forge jar geckolib • mods folder player animation lib forge jar playeranimator • mods folder extendedcrafting jar extendedcrafting • mods folder scalar utils jar scalar utils • mods folder connectivity jar connectivity • mods folder yesstevemodel release hotfix jar yes steve model hotfix • mods folder caelus forge jar caelus • mods folder structory jar structory • mods folder cookingforblockheads forge jar cookingforblockheads • mods folder modernfix forge jar modernfix • mods folder jei forge jar jei • mods folder placebo jar placebo • mods folder naturescompass forge jar naturescompass • mods folder powah jar powah • jarinjar mixinextras forge beta jar mixinextras beta • mods folder lemon core jar lemon core • mods folder mahoutsukai jar mahoutsukai • mods folder balm forge jar balm • mods folder crafttweaker forge jar crafttweaker • mods folder extradisks jar extradisks • mods folder genshinstrument jar genshinstrument • mods folder cloth config forge jar cloth config • minecraft forge universal jar forge • mods folder refinedstorage jar refinedstorage • mods folder windy fixes jar windy fixes • mods folder ironchest jar ironchest • mods folder alchemistry jar alchemistry • mods folder integratedcrafting jar integratedcrafting • minecraft server srg jar minecraft • mods folder farmersdelight jar farmersdelight • mods folder ends delight jar ends delight • mods folder commoncapabilities jar commoncapabilities • mods folder terrablender forge jar terrablender • mods folder burgermod jar burgermod • mods folder mysterious mountain lib jar mysterious mountain lib • mods folder biomesoplenty jar biomesoplenty • mods folder creativecore forge jar creativecore • jarinjar spectrelib forge jar spectrelib • mods folder forge jar • mods folder theurgy jar theurgy • mods folder botania forge snapshot jar botania forge snapshot • mods folder ceilingtorch jar ceilingtorch • mods folder enderio alpha jar enderio alpha • mods folder spark forge jar spark • mods folder curios forge beta jar curios beta • mods folder mekanism homebaked all jar mekanism • mods folder umapyoi jar umapyoi • mods folder pipez jar pipez • jarinjar flywheel forge jar flywheel • mods folder create d jar create d • mods folder patchouli forge snapshot jar patchouli forge snapshot • mods folder integrateddynamics jar integrateddynamics • mods folder chemlib jar chemlib • mods folder polymorph forge jar polymorph • mods folder fluxnetworks jar fluxnetworks • mods folder irons spellbooks jar irons spellbooks • mods folder modonomicon forge jar modonomicon • mods folder automobility forge jar automobility forge • mods folder architectury forge jar architectury • mods folder doapi forge jar doapi • mods folder letsdo vinery forge jar vinery • mods folder yuushya forge jar yuushya • mods folder oceansdelight jar oceansdelight • mods folder ferritecore forge jar ferritecore • mods folder occultism jar occultism • mods folder kawaiidishes jar kawaiidishes • mods folder corn delight jar corn delight • mods folder flib jar flib • jarinjar slim jar • mods folder refinedstorageaddons jar refinedstorageaddons • mods folder cuisinedelight jar cuisinedelight • mods folder tofucraftreload jar tofucraft • mods folder cyclic jar cyclic logs steps to reproduce 正常启动minecraft服务器 在游戏中进行常规操作 服务器在尝试设置某个实体的速度时崩溃 description of issue 在正常的服务器操作中,服务器因与 exception ticking world 相关的错误而崩溃。主要的错误是 java lang illegalargumentexception x not finite ,这个错误似乎是在尝试设置某个实体的速度时触发的。 | 1 |
79,000 | 7,689,166,164 | IssuesEvent | 2018-05-17 11:52:31 | NativeScript/nativescript-cli | https://api.github.com/repos/NativeScript/nativescript-cli | closed | Do not track local paths in Analytics | bug ready for test | ### Did you verify this is a real problem by searching the [NativeScript Forum](http://forum.nativescript.org) and the [other open issues in this repo](https://github.com/NativeScript/nativescript/issues)?
Yes
### Tell us about the problem
Creating project from local template tracks the path in Google Analytics. The path may contain user information.
### Which platform(s) does your issue occur on?
Both
### Please provide the following version numbers that your issue occurs with:
- CLI: 4.0.1
- Cross-platform modules: Not applicable
- Runtime(s): Not applicable
- Plugin(s): Not applicable
### Please tell us how to recreate the issue in as much detail as possible.
Use local template (dir or .tgz) and try the following:
```
$ tns usage-reporting enable
$ tns create myApp --template <local path> --log trace
```
In the logs, you'll find out that the path is tracked in Google Analytics.
### Is there code involved? If so, please share the minimal amount of code needed to recreate the problem.
No
| 1.0 | Do not track local paths in Analytics - ### Did you verify this is a real problem by searching the [NativeScript Forum](http://forum.nativescript.org) and the [other open issues in this repo](https://github.com/NativeScript/nativescript/issues)?
Yes
### Tell us about the problem
Creating project from local template tracks the path in Google Analytics. The path may contain user information.
### Which platform(s) does your issue occur on?
Both
### Please provide the following version numbers that your issue occurs with:
- CLI: 4.0.1
- Cross-platform modules: Not applicable
- Runtime(s): Not applicable
- Plugin(s): Not applicable
### Please tell us how to recreate the issue in as much detail as possible.
Use local template (dir or .tgz) and try the following:
```
$ tns usage-reporting enable
$ tns create myApp --template <local path> --log trace
```
In the logs, you'll find out that the path is tracked in Google Analytics.
### Is there code involved? If so, please share the minimal amount of code needed to recreate the problem.
No
| test | do not track local paths in analytics did you verify this is a real problem by searching the and the yes tell us about the problem creating project from local template tracks the path in google analytics the path may contain user information which platform s does your issue occur on both please provide the following version numbers that your issue occurs with cli cross platform modules not applicable runtime s not applicable plugin s not applicable please tell us how to recreate the issue in as much detail as possible use local template dir or tgz and try the following tns usage reporting enable tns create myapp template log trace in the logs you ll find out that the path is tracked in google analytics is there code involved if so please share the minimal amount of code needed to recreate the problem no | 1 |
349,050 | 31,769,876,918 | IssuesEvent | 2023-09-12 11:05:00 | elastic/elasticsearch | https://api.github.com/repos/elastic/elasticsearch | closed | [CI] More multi_node.EsqlSpecIT failures | >test-failure Team:QL :Query Languages/ES|QL | ### CI Link
https://gradle-enterprise.elastic.co/s/kzhx4kcc6e3pw
### Repro line
`./gradlew ':x-pack:plugin:esql:qa:server:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT" \ -Dtests.method="test {stats.ByStringAndLongWithAlias}" \ -Dtests.seed=9EEC253CF1741A06 \ -Dtests.locale=sv \ -Dtests.timezone=Africa/Bamako \ -Druntime.java=21`
### Does it reproduce?
Didn't try
### Applicable branches
main
### Failure history
_No response_
### Failure excerpt
```
12:02:38 REPRODUCE WITH: ./gradlew ':x-pack:plugin:esql:qa:server:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT" -Dtests.method="test {stats.ByStringAndLong}" -Dtests.seed=9EEC253CF1741A06 -Dtests.locale=sv -Dtests.timezone=Africa/Bamako -Druntime.java=21
12:02:38
12:02:38 org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT > test {stats.ByStringAndLong} FAILED
12:02:38 org.junit.ComparisonFailure: expected:<[2]00000000> but was:<[3]00000000>
12:02:38 at __randomizedtesting.SeedInfo.seed([9EEC253CF1741A06:16B81AE65F8877FE]:0)
12:02:38 at org.junit.Assert.assertEquals(Assert.java:115)
12:02:38 at org.junit.Assert.assertEquals(Assert.java:144)
12:02:38 at org.elasticsearch.xpack.esql.CsvAssert.assertData(CsvAssert.java:208)
12:02:38 at org.elasticsearch.xpack.esql.qa.rest.EsqlSpecTestCase.doTest(EsqlSpecTestCase.java:103)
12:02:38 at org.elasticsearch.xpack.esql.qa.rest.EsqlSpecTestCase.test(EsqlSpecTestCase.java:84)
12:02:38 at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103)
``` | 1.0 | [CI] More multi_node.EsqlSpecIT failures - ### CI Link
https://gradle-enterprise.elastic.co/s/kzhx4kcc6e3pw
### Repro line
`./gradlew ':x-pack:plugin:esql:qa:server:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT" \ -Dtests.method="test {stats.ByStringAndLongWithAlias}" \ -Dtests.seed=9EEC253CF1741A06 \ -Dtests.locale=sv \ -Dtests.timezone=Africa/Bamako \ -Druntime.java=21`
### Does it reproduce?
Didn't try
### Applicable branches
main
### Failure history
_No response_
### Failure excerpt
```
12:02:38 REPRODUCE WITH: ./gradlew ':x-pack:plugin:esql:qa:server:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT" -Dtests.method="test {stats.ByStringAndLong}" -Dtests.seed=9EEC253CF1741A06 -Dtests.locale=sv -Dtests.timezone=Africa/Bamako -Druntime.java=21
12:02:38
12:02:38 org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT > test {stats.ByStringAndLong} FAILED
12:02:38 org.junit.ComparisonFailure: expected:<[2]00000000> but was:<[3]00000000>
12:02:38 at __randomizedtesting.SeedInfo.seed([9EEC253CF1741A06:16B81AE65F8877FE]:0)
12:02:38 at org.junit.Assert.assertEquals(Assert.java:115)
12:02:38 at org.junit.Assert.assertEquals(Assert.java:144)
12:02:38 at org.elasticsearch.xpack.esql.CsvAssert.assertData(CsvAssert.java:208)
12:02:38 at org.elasticsearch.xpack.esql.qa.rest.EsqlSpecTestCase.doTest(EsqlSpecTestCase.java:103)
12:02:38 at org.elasticsearch.xpack.esql.qa.rest.EsqlSpecTestCase.test(EsqlSpecTestCase.java:84)
12:02:38 at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103)
``` | test | more multi node esqlspecit failures ci link repro line gradlew x pack plugin esql qa server multi node javaresttest tests org elasticsearch xpack esql qa multi node esqlspecit dtests method test stats bystringandlongwithalias dtests seed dtests locale sv dtests timezone africa bamako druntime java does it reproduce didn t try applicable branches main failure history no response failure excerpt reproduce with gradlew x pack plugin esql qa server multi node javaresttest tests org elasticsearch xpack esql qa multi node esqlspecit dtests method test stats bystringandlong dtests seed dtests locale sv dtests timezone africa bamako druntime java org elasticsearch xpack esql qa multi node esqlspecit test stats bystringandlong failed org junit comparisonfailure expected but was at randomizedtesting seedinfo seed at org junit assert assertequals assert java at org junit assert assertequals assert java at org elasticsearch xpack esql csvassert assertdata csvassert java at org elasticsearch xpack esql qa rest esqlspectestcase dotest esqlspectestcase java at org elasticsearch xpack esql qa rest esqlspectestcase test esqlspectestcase java at java base jdk internal reflect directmethodhandleaccessor invoke directmethodhandleaccessor java | 1 |
245,546 | 20,776,796,842 | IssuesEvent | 2022-03-16 11:14:49 | cosmos/ibc-go | https://api.github.com/repos/cosmos/ibc-go | closed | Support validator set updates in testing package | testing | Currently the testing suite will fail on `UpdateClient` because the testing suite does not call `Endblock` and validator updates are not properly reflected in the updateclient headers.
The chain simulation must call EndBlock, test headers must be allowed to have next vals different from vals, and validator updates need to correctly get simulated in the update headers | 1.0 | Support validator set updates in testing package - Currently the testing suite will fail on `UpdateClient` because the testing suite does not call `Endblock` and validator updates are not properly reflected in the updateclient headers.
The chain simulation must call EndBlock, test headers must be allowed to have next vals different from vals, and validator updates need to correctly get simulated in the update headers | test | support validator set updates in testing package currently the testing suite will fail on updateclient because the testing suite does not call endblock and validator updates are not properly reflected in the updateclient headers the chain simulation must call endblock test headers must be allowed to have next vals different from vals and validator updates need to correctly get simulated in the update headers | 1 |
215,922 | 24,197,962,877 | IssuesEvent | 2022-09-24 06:11:41 | mkevenaar/OctoPrint-Slack | https://api.github.com/repos/mkevenaar/OctoPrint-Slack | closed | WS-2022-0249 (Medium) detected in OctoPrint-1.7.2-py2.py3-none-any.whl - autoclosed | security vulnerability | ## WS-2022-0249 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>OctoPrint-1.7.2-py2.py3-none-any.whl</b></p></summary>
<p>The snappy web interface for your 3D printer</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/85/f9/643f8de22bc50d11afee9560e0557859f0af7993c79cd4794c79bcd4d8c4/OctoPrint-1.7.2-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/85/f9/643f8de22bc50d11afee9560e0557859f0af7993c79cd4794c79bcd4d8c4/OctoPrint-1.7.2-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt,/tmp/ws-scm/OctoPrint-Slack</p>
<p>
Dependency Hierarchy:
- :x: **OctoPrint-1.7.2-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An open-redirect vulnerability was discovered in In OctoPrint before 1.8.2. The redirect get variable in login page isn't properly checked. Currently, it check if url.scheme and url.netloc are empty using urllib, which does not handle all URL currently. An attacker could redirect a user to a malicious domain.
<p>Publish Date: 2022-06-30
<p>URL: <a href=https://github.com/octoprint/octoprint/commit/dabdd40806e3409f134ab8e276f078e2a02d062b>WS-2022-0249</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/7d2e6212-7431-45c6-9f69-e5957b40fb36/">https://huntr.dev/bounties/7d2e6212-7431-45c6-9f69-e5957b40fb36/</a></p>
<p>Release Date: 2022-06-30</p>
<p>Fix Resolution: OctoPrint - 1.8.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | True | WS-2022-0249 (Medium) detected in OctoPrint-1.7.2-py2.py3-none-any.whl - autoclosed - ## WS-2022-0249 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>OctoPrint-1.7.2-py2.py3-none-any.whl</b></p></summary>
<p>The snappy web interface for your 3D printer</p>
<p>Library home page: <a href="https://files.pythonhosted.org/packages/85/f9/643f8de22bc50d11afee9560e0557859f0af7993c79cd4794c79bcd4d8c4/OctoPrint-1.7.2-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/85/f9/643f8de22bc50d11afee9560e0557859f0af7993c79cd4794c79bcd4d8c4/OctoPrint-1.7.2-py2.py3-none-any.whl</a></p>
<p>Path to dependency file: /requirements.txt</p>
<p>Path to vulnerable library: /requirements.txt,/tmp/ws-scm/OctoPrint-Slack</p>
<p>
Dependency Hierarchy:
- :x: **OctoPrint-1.7.2-py2.py3-none-any.whl** (Vulnerable Library)
<p>Found in base branch: <b>develop</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An open-redirect vulnerability was discovered in In OctoPrint before 1.8.2. The redirect get variable in login page isn't properly checked. Currently, it check if url.scheme and url.netloc are empty using urllib, which does not handle all URL currently. An attacker could redirect a user to a malicious domain.
<p>Publish Date: 2022-06-30
<p>URL: <a href=https://github.com/octoprint/octoprint/commit/dabdd40806e3409f134ab8e276f078e2a02d062b>WS-2022-0249</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: Required
- Scope: Changed
- Impact Metrics:
- Confidentiality Impact: Low
- Integrity Impact: Low
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://huntr.dev/bounties/7d2e6212-7431-45c6-9f69-e5957b40fb36/">https://huntr.dev/bounties/7d2e6212-7431-45c6-9f69-e5957b40fb36/</a></p>
<p>Release Date: 2022-06-30</p>
<p>Fix Resolution: OctoPrint - 1.8.2</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) | non_test | ws medium detected in octoprint none any whl autoclosed ws medium severity vulnerability vulnerable library octoprint none any whl the snappy web interface for your printer library home page a href path to dependency file requirements txt path to vulnerable library requirements txt tmp ws scm octoprint slack dependency hierarchy x octoprint none any whl vulnerable library found in base branch develop vulnerability details an open redirect vulnerability was discovered in in octoprint before the redirect get variable in login page isn t properly checked currently it check if url scheme and url netloc are empty using urllib which does not handle all url currently an attacker could redirect a user to a malicious domain publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution octoprint step up your open source security game with mend | 0 |
1,438 | 2,757,003,911 | IssuesEvent | 2015-04-27 12:27:32 | EricssonResearch/openwebrtc | https://api.github.com/repos/EricssonResearch/openwebrtc | closed | Versioning | All Platforms Build Issue enhancement help wanted Non-Build Issue | This issue is not to create a bikeshed about versioning, but I wanted to discuss some of the points of the versioning integration stuff in OpenWebRTC to make sure we have all bases covered.
I plan to use [semantic versioning](http://semver.org) for OpenWebRTC releases. Given our internal iterations on the codebase, I think we will start at 0.3.0 (the current 0.1 is arbitrary.)
We need to identify all points in the codebase that need to be changed when a version is set. We also need to make sure all the results are correctly versioned. What is needed here? | 2.0 | Versioning - This issue is not to create a bikeshed about versioning, but I wanted to discuss some of the points of the versioning integration stuff in OpenWebRTC to make sure we have all bases covered.
I plan to use [semantic versioning](http://semver.org) for OpenWebRTC releases. Given our internal iterations on the codebase, I think we will start at 0.3.0 (the current 0.1 is arbitrary.)
We need to identify all points in the codebase that need to be changed when a version is set. We also need to make sure all the results are correctly versioned. What is needed here? | non_test | versioning this issue is not to create a bikeshed about versioning but i wanted to discuss some of the points of the versioning integration stuff in openwebrtc to make sure we have all bases covered i plan to use for openwebrtc releases given our internal iterations on the codebase i think we will start at the current is arbitrary we need to identify all points in the codebase that need to be changed when a version is set we also need to make sure all the results are correctly versioned what is needed here | 0 |
233,836 | 17,909,306,291 | IssuesEvent | 2021-09-09 01:27:40 | ignitionrobotics/ign-gazebo | https://api.github.com/repos/ignitionrobotics/ign-gazebo | opened | Document the various ways of controlling a joint | documentation enhancement good first issue | <!-- If you're not sure on the specifics of the feature or would like a broader
discussion, please consider posting a proposal to
http://community.gazebosim.org/ instead.-->
Ignition Gazebo offers many out-of-box controllers with different features and intended use cases, for example:
* [JointController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointController.html)
* [JointPositionController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointPositionController.html)
* [JointTrajectoryController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointTrajectoryController.html)
It can be difficult for users to choose what to use. So it would be nice to have a high-level joint control tutorial explaining the features and usage of each of them, and why a user would choose one over the other.
## Desired behavior
<!-- Describe the current problem and the feature you want implemented.-->
We should have a new tutorial in https://ignitionrobotics.org/api/gazebo/5.1/tutorials.html which summarizes all ways of controlling a joint.
## Alternatives considered
<!-- Describe alternate solutions or features you've considered.-->
Instead of documenting just joint controllers, we could document all ways of moving things, including links, models, etc. This reminds me of this great Gazebo classic tutorial: [Setting Velocity on Joints and Links](http://gazebosim.org/tutorials?tut=set_velocity&cat=)
## Implementation suggestion
<!-- Provide a suggestion on how to implement this feature, which could help us
expedite this implementation.-->
Add a tutorial here: https://github.com/ignitionrobotics/ign-gazebo/tree/ign-gazebo5/tutorials
| 1.0 | Document the various ways of controlling a joint - <!-- If you're not sure on the specifics of the feature or would like a broader
discussion, please consider posting a proposal to
http://community.gazebosim.org/ instead.-->
Ignition Gazebo offers many out-of-box controllers with different features and intended use cases, for example:
* [JointController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointController.html)
* [JointPositionController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointPositionController.html)
* [JointTrajectoryController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointTrajectoryController.html)
It can be difficult for users to choose what to use. So it would be nice to have a high-level joint control tutorial explaining the features and usage of each of them, and why a user would choose one over the other.
## Desired behavior
<!-- Describe the current problem and the feature you want implemented.-->
We should have a new tutorial in https://ignitionrobotics.org/api/gazebo/5.1/tutorials.html which summarizes all ways of controlling a joint.
## Alternatives considered
<!-- Describe alternate solutions or features you've considered.-->
Instead of documenting just joint controllers, we could document all ways of moving things, including links, models, etc. This reminds me of this great Gazebo classic tutorial: [Setting Velocity on Joints and Links](http://gazebosim.org/tutorials?tut=set_velocity&cat=)
## Implementation suggestion
<!-- Provide a suggestion on how to implement this feature, which could help us
expedite this implementation.-->
Add a tutorial here: https://github.com/ignitionrobotics/ign-gazebo/tree/ign-gazebo5/tutorials
| non_test | document the various ways of controlling a joint if you re not sure on the specifics of the feature or would like a broader discussion please consider posting a proposal to instead ignition gazebo offers many out of box controllers with different features and intended use cases for example it can be difficult for users to choose what to use so it would be nice to have a high level joint control tutorial explaining the features and usage of each of them and why a user would choose one over the other desired behavior we should have a new tutorial in which summarizes all ways of controlling a joint alternatives considered instead of documenting just joint controllers we could document all ways of moving things including links models etc this reminds me of this great gazebo classic tutorial implementation suggestion provide a suggestion on how to implement this feature which could help us expedite this implementation add a tutorial here | 0 |
231,307 | 18,759,406,504 | IssuesEvent | 2021-11-05 14:49:08 | pytorch/pytorch | https://api.github.com/repos/pytorch/pytorch | opened | Rollup: ModuleInfo-based testing | module: nn module: testing | This issue tracks the progress of updating module testing from the old `ModuleTest` / `NewModuleTest` / `CriterionTest` structure to a more understandable `ModuleInfo` based framework modeled after `OpInfo`.
- [X] Design `ModuleInfo` structure
- [ ] Port module test coverage to use `ModuleInfo`
- [ ] Test module forward with inputs compared to reference functions
- [ ] Test module backward pass with gradcheck / gradgradcheck
- [ ] Test with non-contiguous inputs
- [ ] Test across devices / dtypes
- [ ] Test across memory_formats
- [ ] Test module instantiation with device / dtype kwargs
- [ ] Test module printing (i.e. `str(m)` / `m.__repr__()`)
- [ ] Test module pickling / unpickling
- [ ] Test parameter casting and movement across devices (e.g. `m.cuda()`, `m.to()`, `m.half()`, etc.)
- [ ] Test module in-place variants (i.e. `inplace=True`) vs. out-of-place variants
- [ ] Test CPU / GPU parity
- [ ] Test C++ / python parity
- [ ] Other consumers of `ModuleTest`, etc. updated to use `ModuleInfo`
- [ ] JIT tests use `ModuleInfo`
- [ ] FX tests use `ModuleInfo`
- [ ] ONNX tests use `ModuleInfo`
- [ ] Full coverage for `ModuleInfo` entries across `torch.nn`
- [ ] TODO: insert itemized list of modules here | 1.0 | Rollup: ModuleInfo-based testing - This issue tracks the progress of updating module testing from the old `ModuleTest` / `NewModuleTest` / `CriterionTest` structure to a more understandable `ModuleInfo` based framework modeled after `OpInfo`.
- [X] Design `ModuleInfo` structure
- [ ] Port module test coverage to use `ModuleInfo`
- [ ] Test module forward with inputs compared to reference functions
- [ ] Test module backward pass with gradcheck / gradgradcheck
- [ ] Test with non-contiguous inputs
- [ ] Test across devices / dtypes
- [ ] Test across memory_formats
- [ ] Test module instantiation with device / dtype kwargs
- [ ] Test module printing (i.e. `str(m)` / `m.__repr__()`)
- [ ] Test module pickling / unpickling
- [ ] Test parameter casting and movement across devices (e.g. `m.cuda()`, `m.to()`, `m.half()`, etc.)
- [ ] Test module in-place variants (i.e. `inplace=True`) vs. out-of-place variants
- [ ] Test CPU / GPU parity
- [ ] Test C++ / python parity
- [ ] Other consumers of `ModuleTest`, etc. updated to use `ModuleInfo`
- [ ] JIT tests use `ModuleInfo`
- [ ] FX tests use `ModuleInfo`
- [ ] ONNX tests use `ModuleInfo`
- [ ] Full coverage for `ModuleInfo` entries across `torch.nn`
- [ ] TODO: insert itemized list of modules here | test | rollup moduleinfo based testing this issue tracks the progress of updating module testing from the old moduletest newmoduletest criteriontest structure to a more understandable moduleinfo based framework modeled after opinfo design moduleinfo structure port module test coverage to use moduleinfo test module forward with inputs compared to reference functions test module backward pass with gradcheck gradgradcheck test with non contiguous inputs test across devices dtypes test across memory formats test module instantiation with device dtype kwargs test module printing i e str m m repr test module pickling unpickling test parameter casting and movement across devices e g m cuda m to m half etc test module in place variants i e inplace true vs out of place variants test cpu gpu parity test c python parity other consumers of moduletest etc updated to use moduleinfo jit tests use moduleinfo fx tests use moduleinfo onnx tests use moduleinfo full coverage for moduleinfo entries across torch nn todo insert itemized list of modules here | 1 |
2,847 | 2,533,570,695 | IssuesEvent | 2015-01-24 01:00:07 | golang/go | https://api.github.com/repos/golang/go | closed | cmd/gc: clang 3.3 complains about undefined behavior | accepted priority-later release-none repo-main | <pre>What steps will reproduce the problem?
1. Download clang 3.3 or later, <a href="http://llvm.org/releases/download.html#3.3">http://llvm.org/releases/download.html#3.3</a>
2. export CC=clang
3. apply CL <a href="https://golang.org/cl/10488044">https://golang.org/cl/10488044</a>
4. ./make.bash
What is the expected output? What do you see instead?
Expected, clean build
Actual, quite a few warnings, some trivial, others not.
<a href="https://gist.github.com/davecheney/5847237">https://gist.github.com/davecheney/5847237</a>
Please use labels and text to provide additional information.
See also the discussion here
<a href="https://groups.google.com/forum/#">https://groups.google.com/forum/#</a>!topic/golang-dev/SlTKGOfHjtU</pre> | 1.0 | cmd/gc: clang 3.3 complains about undefined behavior - <pre>What steps will reproduce the problem?
1. Download clang 3.3 or later, <a href="http://llvm.org/releases/download.html#3.3">http://llvm.org/releases/download.html#3.3</a>
2. export CC=clang
3. apply CL <a href="https://golang.org/cl/10488044">https://golang.org/cl/10488044</a>
4. ./make.bash
What is the expected output? What do you see instead?
Expected, clean build
Actual, quite a few warnings, some trivial, others not.
<a href="https://gist.github.com/davecheney/5847237">https://gist.github.com/davecheney/5847237</a>
Please use labels and text to provide additional information.
See also the discussion here
<a href="https://groups.google.com/forum/#">https://groups.google.com/forum/#</a>!topic/golang-dev/SlTKGOfHjtU</pre> | non_test | cmd gc clang complains about undefined behavior what steps will reproduce the problem download clang or later a href export cc clang apply cl a href make bash what is the expected output what do you see instead expected clean build actual quite a few warnings some trivial others not a href please use labels and text to provide additional information see also the discussion here a href | 0 |
18,565 | 6,623,229,979 | IssuesEvent | 2017-09-22 05:54:53 | commonality/generator-community | https://api.github.com/repos/commonality/generator-community | closed | feat(product-manifest): capture essential software configuration information | Priority: Medium Status: In Progress Type: Build Type: CI Type: Feature | 💡 __TIP:__ Select the "Preview" Tab to help read these instructions.
## 1. Issue type
- [x] __Feature:__ I'm requesting an enhancement.
## 2. User story summary
> Describe what you want to accomplish and in what role/capacity, and why it's important to you.
As a product maintainer,
I need to read and write to a product manifest
In order to manage the configuration of my product.
## 3. Acceptance criteria
> ℹ️ Initial MVPs will serialize software configurations with Node.js product manifests, i.e., the `package.json` file. If a `package.json` exists, `generator-community` will only ever __*add*__ information: `generator-community` should never replace or overwrite existing data.
>
> 📡 Future releases should detect a software product's manifest by language and dependency manager, e.g., Ruby: Gemfile; Java (Gradle): build.gradle; Python: MANIFEST; etc.
- [x] 1. A `package.json` is created if one doesn't exist already.
- [x] 2. The manifest declares a canonical name for the software product.
- [x] 3. The manifest provides a brief description of the product in order to communicate purpose and value.
- [x] 4. The manifest provides the product's [semantic version](http://semver.org).
- [x] 5. The manifest provides the primary product author's name, contact url, and contact email.
- [x] 6. The manifest provides an issue tracking URL in case consumers need to report a defect.
- [x] 7. The manifest lists all third-party production dependencies and their semantic versions.
- [x] 8. The manifest lists all third-party development dependencies and their semantic versions.
- [x] 9. The manifest provides the software product's homepage URL.
- [x] 10. The manifest provides a list of keywords that can be used to help classify the software product.
- [x] 11. The manifest provides the URI to the product's primary source code repository.
- [x] 12. The manifest declares the product's open source license.
## 4. Reference material
* [`generator-node's app` generator](https://github.com/yeoman/generator-node/blob/ca86c075176d029cf0174b31187b2e0a7b32eaa4/generators/app/index.js#L199) has good sample code that reads from and writes to a product's `package.json` file.
* [NPM's package.json documentation](https://docs.npmjs.com/files/package.json) describes every property in a Node.js product manifest (the package.json file). | 1.0 | feat(product-manifest): capture essential software configuration information - 💡 __TIP:__ Select the "Preview" Tab to help read these instructions.
## 1. Issue type
- [x] __Feature:__ I'm requesting an enhancement.
## 2. User story summary
> Describe what you want to accomplish and in what role/capacity, and why it's important to you.
As a product maintainer,
I need to read and write to a product manifest
In order to manage the configuration of my product.
## 3. Acceptance criteria
> ℹ️ Initial MVPs will serialize software configurations with Node.js product manifests, i.e., the `package.json` file. If a `package.json` exists, `generator-community` will only ever __*add*__ information: `generator-community` should never replace or overwrite existing data.
>
> 📡 Future releases should detect a software product's manifest by language and dependency manager, e.g., Ruby: Gemfile; Java (Gradle): build.gradle; Python: MANIFEST; etc.
- [x] 1. A `package.json` is created if one doesn't exist already.
- [x] 2. The manifest declares a canonical name for the software product.
- [x] 3. The manifest provides a brief description of the product in order to communicate purpose and value.
- [x] 4. The manifest provides the product's [semantic version](http://semver.org).
- [x] 5. The manifest provides the primary product author's name, contact url, and contact email.
- [x] 6. The manifest provides an issue tracking URL in case consumers need to report a defect.
- [x] 7. The manifest lists all third-party production dependencies and their semantic versions.
- [x] 8. The manifest lists all third-party development dependencies and their semantic versions.
- [x] 9. The manifest provides the software product's homepage URL.
- [x] 10. The manifest provides a list of keywords that can be used to help classify the software product.
- [x] 11. The manifest provides the URI to the product's primary source code repository.
- [x] 12. The manifest declares the product's open source license.
## 4. Reference material
* [`generator-node's app` generator](https://github.com/yeoman/generator-node/blob/ca86c075176d029cf0174b31187b2e0a7b32eaa4/generators/app/index.js#L199) has good sample code that reads from and writes to a product's `package.json` file.
* [NPM's package.json documentation](https://docs.npmjs.com/files/package.json) describes every property in a Node.js product manifest (the package.json file). | non_test | feat product manifest capture essential software configuration information 💡 tip select the preview tab to help read these instructions issue type feature i m requesting an enhancement user story summary describe what you want to accomplish and in what role capacity and why it s important to you as a product maintainer i need to read and write to a product manifest in order to manage the configuration of my product acceptance criteria ℹ️ initial mvps will serialize software configurations with node js product manifests i e the package json file if a package json exists generator community will only ever add information generator community should never replace or overwrite existing data 📡 future releases should detect a software product s manifest by language and dependency manager e g ruby gemfile java gradle build gradle python manifest etc a package json is created if one doesn t exist already the manifest declares a canonical name for the software product the manifest provides a brief description of the product in order to communicate purpose and value the manifest provides the product s the manifest provides the primary product author s name contact url and contact email the manifest provides an issue tracking url in case consumers need to report a defect the manifest lists all third party production dependencies and their semantic versions the manifest lists all third party development dependencies and their semantic versions the manifest provides the software product s homepage url the manifest provides a list of keywords that can be used to help classify the software product the manifest provides the uri to the product s primary source code repository the manifest declares the product s open source license reference material has good sample code that reads from and writes to a product s package json file describes every property in a node js product manifest the package json file | 0 |
123,802 | 10,291,644,529 | IssuesEvent | 2019-08-27 12:56:54 | cockroachdb/cockroach | https://api.github.com/repos/cockroachdb/cockroach | closed | teamcity: failed test: _size_out_of_range_direct=false | C-test-failure O-robot | The following tests appear to have failed on master (testrace): _size_out_of_range_direct=false
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+_size_out_of_range_direct=false).
[#1451983](https://teamcity.cockroachdb.com/viewLog.html?buildId=1451983):
```
_size_out_of_range_direct=false
--- FAIL: testrace/TestImportData/PGDUMP:_size_out_of_range_direct=false (0.000s)
Test ended in panic.
------- Stdout: -------
I190823 22:18:23.891262 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "create_database", target: 120, info: {DatabaseName:d34 Statement:CREATE DATABASE d34 User:root}
I190823 22:18:24.123243 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "drop_database", target: 120, info: {DatabaseName:d34 Statement:DROP DATABASE d34 User:root DroppedSchemaObjects:[]}
_size_out_of_range_direct=false
--- FAIL: testrace/TestImportData/PGCOPY:_size_out_of_range_direct=false (0.000s)
Test ended in panic.
------- Stdout: -------
I190823 22:18:16.860213 187 storage/store.go:2593 [n1,s1,r59/1:/Table/86{-/1}] removing replica r57/1
I190823 22:18:16.887111 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "create_database", target: 107, info: {DatabaseName:d27 Statement:CREATE DATABASE d27 User:root}
I190823 22:18:17.032604 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "drop_database", target: 107, info: {DatabaseName:d27 Statement:DROP DATABASE d27 User:root DroppedSchemaObjects:[]}
```
Please assign, take a look and update the issue accordingly.
| 1.0 | teamcity: failed test: _size_out_of_range_direct=false - The following tests appear to have failed on master (testrace): _size_out_of_range_direct=false
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+_size_out_of_range_direct=false).
[#1451983](https://teamcity.cockroachdb.com/viewLog.html?buildId=1451983):
```
_size_out_of_range_direct=false
--- FAIL: testrace/TestImportData/PGDUMP:_size_out_of_range_direct=false (0.000s)
Test ended in panic.
------- Stdout: -------
I190823 22:18:23.891262 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "create_database", target: 120, info: {DatabaseName:d34 Statement:CREATE DATABASE d34 User:root}
I190823 22:18:24.123243 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "drop_database", target: 120, info: {DatabaseName:d34 Statement:DROP DATABASE d34 User:root DroppedSchemaObjects:[]}
_size_out_of_range_direct=false
--- FAIL: testrace/TestImportData/PGCOPY:_size_out_of_range_direct=false (0.000s)
Test ended in panic.
------- Stdout: -------
I190823 22:18:16.860213 187 storage/store.go:2593 [n1,s1,r59/1:/Table/86{-/1}] removing replica r57/1
I190823 22:18:16.887111 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "create_database", target: 107, info: {DatabaseName:d27 Statement:CREATE DATABASE d27 User:root}
I190823 22:18:17.032604 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "drop_database", target: 107, info: {DatabaseName:d27 Statement:DROP DATABASE d27 User:root DroppedSchemaObjects:[]}
```
Please assign, take a look and update the issue accordingly.
| test | teamcity failed test size out of range direct false the following tests appear to have failed on master testrace size out of range direct false you may want to check size out of range direct false fail testrace testimportdata pgdump size out of range direct false test ended in panic stdout sql event log go event create database target info databasename statement create database user root sql event log go event drop database target info databasename statement drop database user root droppedschemaobjects size out of range direct false fail testrace testimportdata pgcopy size out of range direct false test ended in panic stdout storage store go removing replica sql event log go event create database target info databasename statement create database user root sql event log go event drop database target info databasename statement drop database user root droppedschemaobjects please assign take a look and update the issue accordingly | 1 |
226,689 | 24,994,505,288 | IssuesEvent | 2022-11-02 22:10:00 | RG4421/ampere-centos-kernel | https://api.github.com/repos/RG4421/ampere-centos-kernel | reopened | CVE-2019-15212 (Medium) detected in linux-yocto-devv5.3 | security vulnerability | ## CVE-2019-15212 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv5.3</b></p></summary>
<p>
<p>Linux Embedded Kernel - tracks the next mainline release</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.1.8. There is a double-free caused by a malicious USB device in the drivers/usb/misc/rio500.c driver.
<p>Publish Date: 2019-08-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15212>CVE-2019-15212</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212</a></p>
<p>Release Date: 2019-09-03</p>
<p>Fix Resolution: v5.2-rc3</p>
</p>
</details>
<p></p>
| True | CVE-2019-15212 (Medium) detected in linux-yocto-devv5.3 - ## CVE-2019-15212 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv5.3</b></p></summary>
<p>
<p>Linux Embedded Kernel - tracks the next mainline release</p>
<p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p>
<p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was discovered in the Linux kernel before 5.1.8. There is a double-free caused by a malicious USB device in the drivers/usb/misc/rio500.c driver.
<p>Publish Date: 2019-08-19
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15212>CVE-2019-15212</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212</a></p>
<p>Release Date: 2019-09-03</p>
<p>Fix Resolution: v5.2-rc3</p>
</p>
</details>
<p></p>
| non_test | cve medium detected in linux yocto cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel tracks the next mainline release library home page a href found in base branch amp centos kernel vulnerable source files drivers usb misc c drivers usb misc c drivers usb misc c vulnerability details an issue was discovered in the linux kernel before there is a double free caused by a malicious usb device in the drivers usb misc c driver publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution | 0 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.