Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
8,714
3,004,049,298
IssuesEvent
2015-07-25 14:43:35
IntellectualSites/PlotSquared
https://api.github.com/repos/IntellectualSites/PlotSquared
closed
NPE
[!] bug [?] needs testing
[23:36:03] [Server thread/WARN]: [PlotSquared] Task #6 for PlotSquared v2.12.15 generated an exception java.lang.NullPointerException at com.intellectualcrafters.plot.listeners.PlotPlusListener$1.run(PlotPlusListener.java:90) ~[?:?] at org.bukkit.craftbukkit.v1_8_R3.scheduler.CraftTask.run(CraftTask.java:71) ~[creative.jar:git-PaperSpigot-4d70f42-b105298] at org.bukkit.craftbukkit.v1_8_R3.scheduler.CraftScheduler.mainThreadHeartbeat(CraftScheduler.java:350) [creative.jar:git-PaperSpigot-4d70f42-b105298] at net.minecraft.server.v1_8_R3.MinecraftServer.B(MinecraftServer.java:774) [creative.jar:git-PaperSpigot-4d70f42-b105298] at net.minecraft.server.v1_8_R3.DedicatedServer.B(DedicatedServer.java:378) [creative.jar:git-PaperSpigot-4d70f42-b105298] at net.minecraft.server.v1_8_R3.MinecraftServer.A(MinecraftServer.java:705) [creative.jar:git-PaperSpigot-4d70f42-b105298] at net.minecraft.server.v1_8_R3.MinecraftServer.run(MinecraftServer.java:608) [creative.jar:git-PaperSpigot-4d70f42-b105298] at java.lang.Thread.run(Thread.java:745) [?:1.8.0_45]
1.0
NPE - [23:36:03] [Server thread/WARN]: [PlotSquared] Task #6 for PlotSquared v2.12.15 generated an exception java.lang.NullPointerException at com.intellectualcrafters.plot.listeners.PlotPlusListener$1.run(PlotPlusListener.java:90) ~[?:?] at org.bukkit.craftbukkit.v1_8_R3.scheduler.CraftTask.run(CraftTask.java:71) ~[creative.jar:git-PaperSpigot-4d70f42-b105298] at org.bukkit.craftbukkit.v1_8_R3.scheduler.CraftScheduler.mainThreadHeartbeat(CraftScheduler.java:350) [creative.jar:git-PaperSpigot-4d70f42-b105298] at net.minecraft.server.v1_8_R3.MinecraftServer.B(MinecraftServer.java:774) [creative.jar:git-PaperSpigot-4d70f42-b105298] at net.minecraft.server.v1_8_R3.DedicatedServer.B(DedicatedServer.java:378) [creative.jar:git-PaperSpigot-4d70f42-b105298] at net.minecraft.server.v1_8_R3.MinecraftServer.A(MinecraftServer.java:705) [creative.jar:git-PaperSpigot-4d70f42-b105298] at net.minecraft.server.v1_8_R3.MinecraftServer.run(MinecraftServer.java:608) [creative.jar:git-PaperSpigot-4d70f42-b105298] at java.lang.Thread.run(Thread.java:745) [?:1.8.0_45]
test
npe task for plotsquared generated an exception java lang nullpointerexception at com intellectualcrafters plot listeners plotpluslistener run plotpluslistener java at org bukkit craftbukkit scheduler crafttask run crafttask java at org bukkit craftbukkit scheduler craftscheduler mainthreadheartbeat craftscheduler java at net minecraft server minecraftserver b minecraftserver java at net minecraft server dedicatedserver b dedicatedserver java at net minecraft server minecraftserver a minecraftserver java at net minecraft server minecraftserver run minecraftserver java at java lang thread run thread java
1
128,212
12,367,140,713
IssuesEvent
2020-05-18 11:46:25
ponylang/ponyup
https://api.github.com/repos/ponylang/ponyup
opened
Document ponyup macOS/brew libressl connection
documentation help wanted
#117 was caused by this. macOS, it's all dynamic linking. `brew install libressl` will periodically change the version it installs as they switch to a newer version of libressl. This is rare, but does happen. When that happens, older versions of ponyup will stop working once libressl is updated. We need to document the expected error that users would see and tell them to reinstall ponyup via the init script as that will download the most recent nightly version. Within 24 hours of a libressl change, it will work again. This also means, that eventually if they update to a different version of ponyup, that it might be using a different libressl than they have installed and will fail.
1.0
Document ponyup macOS/brew libressl connection - #117 was caused by this. macOS, it's all dynamic linking. `brew install libressl` will periodically change the version it installs as they switch to a newer version of libressl. This is rare, but does happen. When that happens, older versions of ponyup will stop working once libressl is updated. We need to document the expected error that users would see and tell them to reinstall ponyup via the init script as that will download the most recent nightly version. Within 24 hours of a libressl change, it will work again. This also means, that eventually if they update to a different version of ponyup, that it might be using a different libressl than they have installed and will fail.
non_test
document ponyup macos brew libressl connection was caused by this macos it s all dynamic linking brew install libressl will periodically change the version it installs as they switch to a newer version of libressl this is rare but does happen when that happens older versions of ponyup will stop working once libressl is updated we need to document the expected error that users would see and tell them to reinstall ponyup via the init script as that will download the most recent nightly version within hours of a libressl change it will work again this also means that eventually if they update to a different version of ponyup that it might be using a different libressl than they have installed and will fail
0
214,295
16,580,204,319
IssuesEvent
2021-05-31 10:41:11
blynkkk/blynk_Issues
https://api.github.com/repos/blynkkk/blynk_Issues
closed
Web dashboard ignores empty space
bug ready to test web
On web dashboard, when you leave a space on top of a widget, after you save the dashboard, it goes up. It does not allow white space in UI design. (Mac Os 11.3.1, Firefox 88.0.1) This is editing page. ![Screen Shot 2021-05-26 at 18 40 47](https://user-images.githubusercontent.com/81516108/119689940-f2332a80-be51-11eb-8353-5bd347215e0b.png) This is after you save. ![Screen Shot 2021-05-26 at 18 40 55](https://user-images.githubusercontent.com/81516108/119690038-04ad6400-be52-11eb-9939-8ea769c4db4a.png)
1.0
Web dashboard ignores empty space - On web dashboard, when you leave a space on top of a widget, after you save the dashboard, it goes up. It does not allow white space in UI design. (Mac Os 11.3.1, Firefox 88.0.1) This is editing page. ![Screen Shot 2021-05-26 at 18 40 47](https://user-images.githubusercontent.com/81516108/119689940-f2332a80-be51-11eb-8353-5bd347215e0b.png) This is after you save. ![Screen Shot 2021-05-26 at 18 40 55](https://user-images.githubusercontent.com/81516108/119690038-04ad6400-be52-11eb-9939-8ea769c4db4a.png)
test
web dashboard ignores empty space on web dashboard when you leave a space on top of a widget after you save the dashboard it goes up it does not allow white space in ui design mac os firefox this is editing page this is after you save
1
78,647
7,657,016,347
IssuesEvent
2018-05-10 18:13:42
couchbase/couchbase-lite-ios
https://api.github.com/repos/couchbase/couchbase-lite-ios
closed
latest iOS cbl 2.1.0 builds are crashing
functional-test-blocker ready
### Version CBL - 2.1.0-150 sg version -> 2.1.0-55 ### Issue caused 1. Running ios functional tests on jenkins having crash on the app and bunch of tests failing I ran tests with 2.1.0-126 and that looks good If I run tests individually, it works fine. ### Logs: [CBLTestServer-iOS_2018-04-30-154305-1_Dans-Test-MacBook-Pro.crash.zip](https://github.com/couchbase/couchbase-lite-ios/files/1968363/CBLTestServer-iOS_2018-04-30-154305-1_Dans-Test-MacBook-Pro.crash.zip)
1.0
latest iOS cbl 2.1.0 builds are crashing - ### Version CBL - 2.1.0-150 sg version -> 2.1.0-55 ### Issue caused 1. Running ios functional tests on jenkins having crash on the app and bunch of tests failing I ran tests with 2.1.0-126 and that looks good If I run tests individually, it works fine. ### Logs: [CBLTestServer-iOS_2018-04-30-154305-1_Dans-Test-MacBook-Pro.crash.zip](https://github.com/couchbase/couchbase-lite-ios/files/1968363/CBLTestServer-iOS_2018-04-30-154305-1_Dans-Test-MacBook-Pro.crash.zip)
test
latest ios cbl builds are crashing version cbl sg version issue caused running ios functional tests on jenkins having crash on the app and bunch of tests failing i ran tests with and that looks good if i run tests individually it works fine logs
1
231,484
17,690,791,027
IssuesEvent
2021-08-24 09:40:33
owncloud/ocis
https://api.github.com/repos/owncloud/ocis
reopened
Write documentation for roles & permissions concept
Topic:Documentation
Write documentation for roles & permissions concept. This is intended to be a living document and it should reflect decisions made during the development process. The end result being a document in a state that mirrors the settings service functionality regarding roles and permissions.
1.0
Write documentation for roles & permissions concept - Write documentation for roles & permissions concept. This is intended to be a living document and it should reflect decisions made during the development process. The end result being a document in a state that mirrors the settings service functionality regarding roles and permissions.
non_test
write documentation for roles permissions concept write documentation for roles permissions concept this is intended to be a living document and it should reflect decisions made during the development process the end result being a document in a state that mirrors the settings service functionality regarding roles and permissions
0
251,817
27,211,188,498
IssuesEvent
2023-02-20 16:40:23
ZSBRybnik/frontend
https://api.github.com/repos/ZSBRybnik/frontend
closed
node-jq-2.3.3.tgz: 3 vulnerabilities (highest severity is: 7.5) - autoclosed
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-jq-2.3.3.tgz</b></p></summary> <p></p> <p> <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (node-jq version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-25881](https://www.mend.io/vulnerability-database/CVE-2022-25881) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | http-cache-semantics-3.8.1.tgz | Transitive | N/A* | &#10060; | | [CVE-2023-25166](https://www.mend.io/vulnerability-database/CVE-2023-25166) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | formula-3.0.0.tgz | Transitive | N/A* | &#10060; | | [CVE-2022-33987](https://www.mend.io/vulnerability-database/CVE-2022-33987) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-25881</summary> ### Vulnerable Library - <b>http-cache-semantics-3.8.1.tgz</b></p> <p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p> <p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz</a></p> <p> Dependency Hierarchy: - node-jq-2.3.3.tgz (Root Library) - download-8.0.0.tgz - got-8.3.2.tgz - cacheable-request-2.1.4.tgz - :x: **http-cache-semantics-3.8.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library. <p>Publish Date: 2023-01-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p> <p>Release Date: 2023-01-31</p> <p>Fix Resolution: http-cache-semantics - 4.1.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-25166</summary> ### Vulnerable Library - <b>formula-3.0.0.tgz</b></p> <p>Math and string formula parser.</p> <p>Library home page: <a href="https://registry.npmjs.org/@sideway/formula/-/formula-3.0.0.tgz">https://registry.npmjs.org/@sideway/formula/-/formula-3.0.0.tgz</a></p> <p> Dependency Hierarchy: - node-jq-2.3.3.tgz (Root Library) - joi-17.6.0.tgz - :x: **formula-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> formula is a math and string formula parser. In versions prior to 3.0.1 crafted user-provided strings to formula's parser might lead to polynomial execution time and a denial of service. Users should upgrade to 3.0.1+. There are no known workarounds for this vulnerability. <p>Publish Date: 2023-02-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25166>CVE-2023-25166</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25166">https://www.cve.org/CVERecord?id=CVE-2023-25166</a></p> <p>Release Date: 2023-02-08</p> <p>Fix Resolution: @sideway/formula - 3.0.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-33987</summary> ### Vulnerable Libraries - <b>got-8.3.2.tgz</b>, <b>got-7.1.0.tgz</b></p> <p> ### <b>got-8.3.2.tgz</b></p> <p>Simplified HTTP requests</p> <p>Library home page: <a href="https://registry.npmjs.org/got/-/got-8.3.2.tgz">https://registry.npmjs.org/got/-/got-8.3.2.tgz</a></p> <p> Dependency Hierarchy: - node-jq-2.3.3.tgz (Root Library) - download-8.0.0.tgz - :x: **got-8.3.2.tgz** (Vulnerable Library) ### <b>got-7.1.0.tgz</b></p> <p>Simplified HTTP requests</p> <p>Library home page: <a href="https://registry.npmjs.org/got/-/got-7.1.0.tgz">https://registry.npmjs.org/got/-/got-7.1.0.tgz</a></p> <p> Dependency Hierarchy: - node-jq-2.3.3.tgz (Root Library) - bin-build-3.0.0.tgz - download-6.2.5.tgz - :x: **got-7.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The got package before 12.1.0 (also fixed in 11.8.5) for Node.js allows a redirect to a UNIX socket. <p>Publish Date: 2022-06-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-33987>CVE-2022-33987</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p> <p>Release Date: 2022-06-18</p> <p>Fix Resolution: got - 11.8.5,12.1.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
node-jq-2.3.3.tgz: 3 vulnerabilities (highest severity is: 7.5) - autoclosed - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-jq-2.3.3.tgz</b></p></summary> <p></p> <p> <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p></details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (node-jq version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-25881](https://www.mend.io/vulnerability-database/CVE-2022-25881) | <img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> High | 7.5 | http-cache-semantics-3.8.1.tgz | Transitive | N/A* | &#10060; | | [CVE-2023-25166](https://www.mend.io/vulnerability-database/CVE-2023-25166) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 6.5 | formula-3.0.0.tgz | Transitive | N/A* | &#10060; | | [CVE-2022-33987](https://www.mend.io/vulnerability-database/CVE-2022-33987) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | detected in multiple dependencies | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> CVE-2022-25881</summary> ### Vulnerable Library - <b>http-cache-semantics-3.8.1.tgz</b></p> <p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p> <p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-3.8.1.tgz</a></p> <p> Dependency Hierarchy: - node-jq-2.3.3.tgz (Root Library) - download-8.0.0.tgz - got-8.3.2.tgz - cacheable-request-2.1.4.tgz - :x: **http-cache-semantics-3.8.1.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library. <p>Publish Date: 2023-01-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>7.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p> <p>Release Date: 2023-01-31</p> <p>Fix Resolution: http-cache-semantics - 4.1.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2023-25166</summary> ### Vulnerable Library - <b>formula-3.0.0.tgz</b></p> <p>Math and string formula parser.</p> <p>Library home page: <a href="https://registry.npmjs.org/@sideway/formula/-/formula-3.0.0.tgz">https://registry.npmjs.org/@sideway/formula/-/formula-3.0.0.tgz</a></p> <p> Dependency Hierarchy: - node-jq-2.3.3.tgz (Root Library) - joi-17.6.0.tgz - :x: **formula-3.0.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> formula is a math and string formula parser. In versions prior to 3.0.1 crafted user-provided strings to formula's parser might lead to polynomial execution time and a denial of service. Users should upgrade to 3.0.1+. There are no known workarounds for this vulnerability. <p>Publish Date: 2023-02-08 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-25166>CVE-2023-25166</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>6.5</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2023-25166">https://www.cve.org/CVERecord?id=CVE-2023-25166</a></p> <p>Release Date: 2023-02-08</p> <p>Fix Resolution: @sideway/formula - 3.0.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details><details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-33987</summary> ### Vulnerable Libraries - <b>got-8.3.2.tgz</b>, <b>got-7.1.0.tgz</b></p> <p> ### <b>got-8.3.2.tgz</b></p> <p>Simplified HTTP requests</p> <p>Library home page: <a href="https://registry.npmjs.org/got/-/got-8.3.2.tgz">https://registry.npmjs.org/got/-/got-8.3.2.tgz</a></p> <p> Dependency Hierarchy: - node-jq-2.3.3.tgz (Root Library) - download-8.0.0.tgz - :x: **got-8.3.2.tgz** (Vulnerable Library) ### <b>got-7.1.0.tgz</b></p> <p>Simplified HTTP requests</p> <p>Library home page: <a href="https://registry.npmjs.org/got/-/got-7.1.0.tgz">https://registry.npmjs.org/got/-/got-7.1.0.tgz</a></p> <p> Dependency Hierarchy: - node-jq-2.3.3.tgz (Root Library) - bin-build-3.0.0.tgz - download-6.2.5.tgz - :x: **got-7.1.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ZSBRybnik/frontend/commit/273a134394edfb54991ff74097965c8f3cac3de7">273a134394edfb54991ff74097965c8f3cac3de7</a></p> <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> The got package before 12.1.0 (also fixed in 11.8.5) for Node.js allows a redirect to a UNIX socket. <p>Publish Date: 2022-06-18 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-33987>CVE-2022-33987</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-33987</a></p> <p>Release Date: 2022-06-18</p> <p>Fix Resolution: got - 11.8.5,12.1.0</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_test
node jq tgz vulnerabilities highest severity is autoclosed vulnerable library node jq tgz found in head commit a href vulnerabilities cve severity cvss dependency type fixed in node jq version remediation available high http cache semantics tgz transitive n a medium formula tgz transitive n a medium detected in multiple dependencies transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library http cache semantics tgz parses cache control and other headers helps building correct http caches and proxies library home page a href dependency hierarchy node jq tgz root library download tgz got tgz cacheable request tgz x http cache semantics tgz vulnerable library found in head commit a href found in base branch master vulnerability details this affects versions of the package http cache semantics before the issue can be exploited via malicious request header values sent to a server when that server reads the cache policy from the request using this library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution http cache semantics step up your open source security game with mend cve vulnerable library formula tgz math and string formula parser library home page a href dependency hierarchy node jq tgz root library joi tgz x formula tgz vulnerable library found in head commit a href found in base branch master vulnerability details formula is a math and string formula parser in versions prior to crafted user provided strings to formula s parser might lead to polynomial execution time and a denial of service users should upgrade to there are no known workarounds for this vulnerability publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution sideway formula step up your open source security game with mend cve vulnerable libraries got tgz got tgz got tgz simplified http requests library home page a href dependency hierarchy node jq tgz root library download tgz x got tgz vulnerable library got tgz simplified http requests library home page a href dependency hierarchy node jq tgz root library bin build tgz download tgz x got tgz vulnerable library found in head commit a href found in base branch master vulnerability details the got package before also fixed in for node js allows a redirect to a unix socket publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution got step up your open source security game with mend
0
23,527
10,894,847,044
IssuesEvent
2019-11-19 09:31:21
elikkatzgit/quantumsim
https://api.github.com/repos/elikkatzgit/quantumsim
closed
CVE-2015-0220 (Medium) detected in Django-1.3.tar.gz
security vulnerability
## CVE-2015-0220 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.3.tar.gz</b></p></summary> <p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz">https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz</a></p> <p>Path to dependency file: /tmp/ws-scm/quantumsim/requirements.txt</p> <p>Path to vulnerable library: /quantumsim/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Django-1.3.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/quantumsim/commit/d6624156203bb0fc439915ed3fc47432b9cbbeb5">d6624156203bb0fc439915ed3fc47432b9cbbeb5</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The django.util.http.is_safe_url function in Django before 1.4.18, 1.6.x before 1.6.10, and 1.7.x before 1.7.3 does not properly handle leading whitespaces, which allows remote attackers to conduct cross-site scripting (XSS) attacks via a crafted URL, related to redirect URLs, as demonstrated by a "\njavascript:" URL. <p>Publish Date: 2015-01-16 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-0220>CVE-2015-0220</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-0220">https://nvd.nist.gov/vuln/detail/CVE-2015-0220</a></p> <p>Release Date: 2015-01-16</p> <p>Fix Resolution: 1.4.18,1.6.10,1.7.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Django","packageVersion":"1.3","isTransitiveDependency":false,"dependencyTree":"Django:1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.18,1.6.10,1.7.3"}],"vulnerabilityIdentifier":"CVE-2015-0220","vulnerabilityDetails":"The django.util.http.is_safe_url function in Django before 1.4.18, 1.6.x before 1.6.10, and 1.7.x before 1.7.3 does not properly handle leading whitespaces, which allows remote attackers to conduct cross-site scripting (XSS) attacks via a crafted URL, related to redirect URLs, as demonstrated by a \"\\njavascript:\" URL.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2015-0220","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> -->
True
CVE-2015-0220 (Medium) detected in Django-1.3.tar.gz - ## CVE-2015-0220 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.3.tar.gz</b></p></summary> <p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz">https://files.pythonhosted.org/packages/f5/d5/6722d3091946734194ffcfe8ef074f63e8acdd1ff51dfcfc87c2c194fd3f/Django-1.3.tar.gz</a></p> <p>Path to dependency file: /tmp/ws-scm/quantumsim/requirements.txt</p> <p>Path to vulnerable library: /quantumsim/requirements.txt</p> <p> Dependency Hierarchy: - :x: **Django-1.3.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/elikkatzgit/quantumsim/commit/d6624156203bb0fc439915ed3fc47432b9cbbeb5">d6624156203bb0fc439915ed3fc47432b9cbbeb5</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The django.util.http.is_safe_url function in Django before 1.4.18, 1.6.x before 1.6.10, and 1.7.x before 1.7.3 does not properly handle leading whitespaces, which allows remote attackers to conduct cross-site scripting (XSS) attacks via a crafted URL, related to redirect URLs, as demonstrated by a "\njavascript:" URL. <p>Publish Date: 2015-01-16 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2015-0220>CVE-2015-0220</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>4.3</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2015-0220">https://nvd.nist.gov/vuln/detail/CVE-2015-0220</a></p> <p>Release Date: 2015-01-16</p> <p>Fix Resolution: 1.4.18,1.6.10,1.7.3</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Python","packageName":"Django","packageVersion":"1.3","isTransitiveDependency":false,"dependencyTree":"Django:1.3","isMinimumFixVersionAvailable":true,"minimumFixVersion":"1.4.18,1.6.10,1.7.3"}],"vulnerabilityIdentifier":"CVE-2015-0220","vulnerabilityDetails":"The django.util.http.is_safe_url function in Django before 1.4.18, 1.6.x before 1.6.10, and 1.7.x before 1.7.3 does not properly handle leading whitespaces, which allows remote attackers to conduct cross-site scripting (XSS) attacks via a crafted URL, related to redirect URLs, as demonstrated by a \"\\njavascript:\" URL.","vulnerabilityUrl":"https://cve.mitre.org/cgi-bin/cvename.cgi?name\u003dCVE-2015-0220","cvss2Severity":"medium","cvss2Score":"4.3","extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in django tar gz cve medium severity vulnerability vulnerable library django tar gz a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file tmp ws scm quantumsim requirements txt path to vulnerable library quantumsim requirements txt dependency hierarchy x django tar gz vulnerable library found in head commit a href vulnerability details the django util http is safe url function in django before x before and x before does not properly handle leading whitespaces which allows remote attackers to conduct cross site scripting xss attacks via a crafted url related to redirect urls as demonstrated by a njavascript url publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution isopenpronvulnerability false ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails the django util http is safe url function in django before x before and x before does not properly handle leading whitespaces which allows remote attackers to conduct cross site scripting xss attacks via a crafted url related to redirect urls as demonstrated by a njavascript url vulnerabilityurl
0
372,395
11,013,717,793
IssuesEvent
2019-12-04 21:05:27
dmwm/WMCore
https://api.github.com/repos/dmwm/WMCore
closed
Break closure of phedex blocks into smaller chunks
BUG High Priority Need wmagent branch WMAgent
**Impact of the bug** WMAgent (PhEDExInjector) **Describe the bug** Apparently we put too much data into the same PhEDEx call to close blocks here: https://github.com/dmwm/WMCore/blob/master/src/python/WMComponent/PhEDExInjector/PhEDExInjectorPoller.py#L375 it's one call per location, and that location might have many many blocks from multiple datasets. The problem is, if there is a problem with any of those blocks (e.g., the current state of vocms0253 which has a **dataset** closed in PhEDEx, see [1]), the whole request would fail and no blocks can be closed... **How to reproduce it** none **Expected behavior** My suggestion is to break the http requests per location and dataset, something like for location in locations: for dataset in datasets: make a phedex request Yes, it will increase the amount of PhEDEx calls, but still it's going to be 1 call per block that we close, which should be just fine for the phedex data-service. **Additional context and error message** [1] ``` 2019-11-27 22:23:08,945:140168429426432:WARNING:Service:The cachefile /data/srv/wmagent/v1.2.6.patch1/install/wmagent/PhEDExInjector/.wmcore_cache/.wmcore_cache_31961/requests/cmsweb.cern.ch:8443/-7685143179613178687_POST_inject does n ot exist and the service at https://cmsweb.cern.ch:8443/phedex/datasvc/json/prod/inject is unavailable - it returned 400 because Bad Request with result: injectData error: dataset /RadionToWW_narrow_M-6500_TuneCUETP8M1_13TeV-madgraph-p ythia8/RunIISummer16MiniAODv3-PUMoriond17_94X_mcRun2_asymptotic_v3-v1/MINIAODSIM is closed\n 2019-11-27 22:23:08,946:140168429426432:ERROR:PhEDExInjectorPoller:PhEDEx block close failed with HTTPException: 400 injectData error: dataset /RadionToWW_narrow_M-6500_TuneCUETP8M1_13TeV-madgraph-pythia8/RunIISummer16MiniAODv3-PUMorio nd17_94X_mcRun2_asymptotic_v3-v1/MINIAODSIM is closed\n ```
1.0
Break closure of phedex blocks into smaller chunks - **Impact of the bug** WMAgent (PhEDExInjector) **Describe the bug** Apparently we put too much data into the same PhEDEx call to close blocks here: https://github.com/dmwm/WMCore/blob/master/src/python/WMComponent/PhEDExInjector/PhEDExInjectorPoller.py#L375 it's one call per location, and that location might have many many blocks from multiple datasets. The problem is, if there is a problem with any of those blocks (e.g., the current state of vocms0253 which has a **dataset** closed in PhEDEx, see [1]), the whole request would fail and no blocks can be closed... **How to reproduce it** none **Expected behavior** My suggestion is to break the http requests per location and dataset, something like for location in locations: for dataset in datasets: make a phedex request Yes, it will increase the amount of PhEDEx calls, but still it's going to be 1 call per block that we close, which should be just fine for the phedex data-service. **Additional context and error message** [1] ``` 2019-11-27 22:23:08,945:140168429426432:WARNING:Service:The cachefile /data/srv/wmagent/v1.2.6.patch1/install/wmagent/PhEDExInjector/.wmcore_cache/.wmcore_cache_31961/requests/cmsweb.cern.ch:8443/-7685143179613178687_POST_inject does n ot exist and the service at https://cmsweb.cern.ch:8443/phedex/datasvc/json/prod/inject is unavailable - it returned 400 because Bad Request with result: injectData error: dataset /RadionToWW_narrow_M-6500_TuneCUETP8M1_13TeV-madgraph-p ythia8/RunIISummer16MiniAODv3-PUMoriond17_94X_mcRun2_asymptotic_v3-v1/MINIAODSIM is closed\n 2019-11-27 22:23:08,946:140168429426432:ERROR:PhEDExInjectorPoller:PhEDEx block close failed with HTTPException: 400 injectData error: dataset /RadionToWW_narrow_M-6500_TuneCUETP8M1_13TeV-madgraph-pythia8/RunIISummer16MiniAODv3-PUMorio nd17_94X_mcRun2_asymptotic_v3-v1/MINIAODSIM is closed\n ```
non_test
break closure of phedex blocks into smaller chunks impact of the bug wmagent phedexinjector describe the bug apparently we put too much data into the same phedex call to close blocks here it s one call per location and that location might have many many blocks from multiple datasets the problem is if there is a problem with any of those blocks e g the current state of which has a dataset closed in phedex see the whole request would fail and no blocks can be closed how to reproduce it none expected behavior my suggestion is to break the http requests per location and dataset something like for location in locations for dataset in datasets make a phedex request yes it will increase the amount of phedex calls but still it s going to be call per block that we close which should be just fine for the phedex data service additional context and error message warning service the cachefile data srv wmagent install wmagent phedexinjector wmcore cache wmcore cache requests cmsweb cern ch post inject does n ot exist and the service at is unavailable it returned because bad request with result injectdata error dataset radiontoww narrow m madgraph p asymptotic miniaodsim is closed n error phedexinjectorpoller phedex block close failed with httpexception injectdata error dataset radiontoww narrow m madgraph pumorio asymptotic miniaodsim is closed n
0
249,612
21,179,721,071
IssuesEvent
2022-04-08 06:34:23
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
opened
Terminal flaky tests
smoke-test-failure
Lately terminal tests became flaky with: ``` 1) VSCode Smoke Tests (Web) Terminal Terminal Editors should update color of the tab: Error: Timeout: is active element '.quick-input-widget .quick-input-box input' after 20 seconds. at Code.poll (D:\a\_work\1\s\test\automation\src\code.ts:296:11) at Code.waitForActiveElement (D:\a\_work\1\s\test\automation\src\code.ts:232:3) at QuickInput.waitForQuickInputOpened (D:\a\_work\1\s\test\automation\src\quickinput.ts:20:3) at Terminal.runCommandWithValue (D:\a\_work\1\s\test\automation\src\terminal.ts:88:3) at Context.<anonymous> (src\areas\terminal\terminal-editors.test.ts:21:4) ``` This tests uses `runCommandWithValue` and there is a somewhat questionable line here: https://github.com/microsoft/vscode/blob/921264bfe3ffbbfe5ec8c4b08214b88e2148fa3e/test/automation/src/terminal.ts#L83-L88 Fyi we had updated playwright to latest. I will go ahead and skip tests for now that use this method.
1.0
Terminal flaky tests - Lately terminal tests became flaky with: ``` 1) VSCode Smoke Tests (Web) Terminal Terminal Editors should update color of the tab: Error: Timeout: is active element '.quick-input-widget .quick-input-box input' after 20 seconds. at Code.poll (D:\a\_work\1\s\test\automation\src\code.ts:296:11) at Code.waitForActiveElement (D:\a\_work\1\s\test\automation\src\code.ts:232:3) at QuickInput.waitForQuickInputOpened (D:\a\_work\1\s\test\automation\src\quickinput.ts:20:3) at Terminal.runCommandWithValue (D:\a\_work\1\s\test\automation\src\terminal.ts:88:3) at Context.<anonymous> (src\areas\terminal\terminal-editors.test.ts:21:4) ``` This tests uses `runCommandWithValue` and there is a somewhat questionable line here: https://github.com/microsoft/vscode/blob/921264bfe3ffbbfe5ec8c4b08214b88e2148fa3e/test/automation/src/terminal.ts#L83-L88 Fyi we had updated playwright to latest. I will go ahead and skip tests for now that use this method.
test
terminal flaky tests lately terminal tests became flaky with vscode smoke tests web terminal terminal editors should update color of the tab error timeout is active element quick input widget quick input box input after seconds at code poll d a work s test automation src code ts at code waitforactiveelement d a work s test automation src code ts at quickinput waitforquickinputopened d a work s test automation src quickinput ts at terminal runcommandwithvalue d a work s test automation src terminal ts at context src areas terminal terminal editors test ts this tests uses runcommandwithvalue and there is a somewhat questionable line here fyi we had updated playwright to latest i will go ahead and skip tests for now that use this method
1
296,736
25,572,522,611
IssuesEvent
2022-11-30 18:56:16
MD-Anderson-Bioinformatics/NG-CHM
https://api.github.com/repos/MD-Anderson-Bioinformatics/NG-CHM
closed
Pubmed linkouts need to be opened in new frame.
bug linkouts passed retest 2.21.3
If linkout frame opened, Pubmed linkouts do not work. Refused to display 'https://pubmed.ncbi.nlm.nih.gov/' in a frame because it set 'X-Frame-Options' to 'deny'.
1.0
Pubmed linkouts need to be opened in new frame. - If linkout frame opened, Pubmed linkouts do not work. Refused to display 'https://pubmed.ncbi.nlm.nih.gov/' in a frame because it set 'X-Frame-Options' to 'deny'.
test
pubmed linkouts need to be opened in new frame if linkout frame opened pubmed linkouts do not work refused to display in a frame because it set x frame options to deny
1
293,607
8,998,093,882
IssuesEvent
2019-02-02 18:33:57
Beep6581/RawTherapee
https://api.github.com/repos/Beep6581/RawTherapee
opened
Segfault in lmmse demosaic
Priority-Critical bug
There's a really hard to reproduce segfault in lmmse demosaic. Searching.....
1.0
Segfault in lmmse demosaic - There's a really hard to reproduce segfault in lmmse demosaic. Searching.....
non_test
segfault in lmmse demosaic there s a really hard to reproduce segfault in lmmse demosaic searching
0
58,511
24,468,752,529
IssuesEvent
2022-10-07 17:31:38
valor-software/valor-software.github.io
https://api.github.com/repos/valor-software/valor-software.github.io
closed
Design | Service page
service page
To create a Design page that should have a path: Home > Services > Design It can be accessible via: https://valor-software.com/services https://valor-software.com/ main menu Design: https://www.figma.com/file/StpiCGh7YZyAPRjtD5gJBo/Valor-Site-Design-2021?node-id=9450%3A82643
1.0
Design | Service page - To create a Design page that should have a path: Home > Services > Design It can be accessible via: https://valor-software.com/services https://valor-software.com/ main menu Design: https://www.figma.com/file/StpiCGh7YZyAPRjtD5gJBo/Valor-Site-Design-2021?node-id=9450%3A82643
non_test
design service page to create a design page that should have a path home services design it can be accessible via main menu design
0
272,205
20,737,252,094
IssuesEvent
2022-03-14 14:42:46
dj-stripe/dj-stripe
https://api.github.com/repos/dj-stripe/dj-stripe
closed
Ability to create express account attached to auth user
documentation
Hi. Is does this package have a way to create an express account and attach it to the django auth user (onetoonefield) or does this have to be done manually?
1.0
Ability to create express account attached to auth user - Hi. Is does this package have a way to create an express account and attach it to the django auth user (onetoonefield) or does this have to be done manually?
non_test
ability to create express account attached to auth user hi is does this package have a way to create an express account and attach it to the django auth user onetoonefield or does this have to be done manually
0
60,590
6,711,005,623
IssuesEvent
2017-10-13 00:49:29
ansible/galaxy-issues
https://api.github.com/repos/ansible/galaxy-issues
closed
No error when adding a new role/container project that has the same name as an existing project
bug ready for testing
## Steps to Reproduce 1. Visit 'roleadd' on Ansible Galaxy (https://galaxy.ansible.com/roleadd#/) 2. Click the 'enable' toggle next to a role named `username/xyz` 3. Click the configure widget, and verify the role name is `xyz` 4. Wait for role import to succeed (green dot next to 'Succeeded') 5. Click the 'enable' toggle next to another role named `username/something-else` 6. Click the configure widget, and change role name to `xyz` 7. Click 'Save' and observe the result. ## Expected Result * I should see an error message alerting me that a role with that name already exists in my Galaxy namespace. ## Actual Result The 'Running' status widget keeps spinning forever, and nothing seems to happen. After a while, if you refresh the page, you can see the project was imported as `something-else` and not `xyz` (I think, if I remember correctly). This was discovered when I was adding a `geerlingguy.solr` ansible-container project, while I already had a `geerlingguy.solr` ansible role in my user namespace. See: https://github.com/ansible/ansible-container/issues/629
1.0
No error when adding a new role/container project that has the same name as an existing project - ## Steps to Reproduce 1. Visit 'roleadd' on Ansible Galaxy (https://galaxy.ansible.com/roleadd#/) 2. Click the 'enable' toggle next to a role named `username/xyz` 3. Click the configure widget, and verify the role name is `xyz` 4. Wait for role import to succeed (green dot next to 'Succeeded') 5. Click the 'enable' toggle next to another role named `username/something-else` 6. Click the configure widget, and change role name to `xyz` 7. Click 'Save' and observe the result. ## Expected Result * I should see an error message alerting me that a role with that name already exists in my Galaxy namespace. ## Actual Result The 'Running' status widget keeps spinning forever, and nothing seems to happen. After a while, if you refresh the page, you can see the project was imported as `something-else` and not `xyz` (I think, if I remember correctly). This was discovered when I was adding a `geerlingguy.solr` ansible-container project, while I already had a `geerlingguy.solr` ansible role in my user namespace. See: https://github.com/ansible/ansible-container/issues/629
test
no error when adding a new role container project that has the same name as an existing project steps to reproduce visit roleadd on ansible galaxy click the enable toggle next to a role named username xyz click the configure widget and verify the role name is xyz wait for role import to succeed green dot next to succeeded click the enable toggle next to another role named username something else click the configure widget and change role name to xyz click save and observe the result expected result i should see an error message alerting me that a role with that name already exists in my galaxy namespace actual result the running status widget keeps spinning forever and nothing seems to happen after a while if you refresh the page you can see the project was imported as something else and not xyz i think if i remember correctly this was discovered when i was adding a geerlingguy solr ansible container project while i already had a geerlingguy solr ansible role in my user namespace see
1
55,529
14,533,130,997
IssuesEvent
2020-12-14 23:51:30
department-of-veterans-affairs/va.gov-team
https://api.github.com/repos/department-of-veterans-affairs/va.gov-team
closed
508-defect-2 [AXE-CORE]: App Directory - Heading levels should increase by one
508-defect-2 508-issue-headings 508/Accessibility
# [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> ## Feedback framework - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Definition of done 1. Review and acknowledge feedback. 1. Fix and/or document decisions made. 1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix. ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Trevor_ ## User Story or Problem Statement <!-- Example: As a user with cognitive considerations, I expect to see a label and input pairing consistently styled as throughout the rest of the site, with the label just above the text/email/search input or to the right of a radio/checkbox input, so that I am clearly able to understand what entry is expected. --> As an assistive tech user, I want to hear headings read out in the correct nesting order. ## Details <!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. --> Our app headings and sub-headings inside the accordions need to be H2 and H3 headings respectively. Screen shots attached below. ## Acceptance Criteria - [ ] Axe browser plugin doesn't report a heading nesting best practice warning on future runs - [ ] Current visual styles are maintained ## WCAG or Vendor Guidance (optional) * [Heading levels should only increase by one](https://dequeuniversity.com/rules/axe/4.0/heading-order) ## Screenshots or Trace Logs <!-- Drop any screenshots or error logs that might be useful for debugging --> ![Screen Shot 2020-11-19 at 2 22 53 PM](https://user-images.githubusercontent.com/934879/99720651-aacb7c00-2a73-11eb-9460-3649b9e5d32c.png) --- ![Screen Shot 2020-11-19 at 2 23 31 PM](https://user-images.githubusercontent.com/934879/99720682-b323b700-2a73-11eb-849b-f2b7171f2418.png)
1.0
508-defect-2 [AXE-CORE]: App Directory - Heading levels should increase by one - # [508-defect-2](https://github.com/department-of-veterans-affairs/va.gov-team/blob/master/platform/accessibility/guidance/defect-severity-rubric.md#508-defect-2) <!-- Enter an issue title using the format [ERROR TYPE]: Brief description of the problem --- [SCREENREADER]: Edit buttons need aria-label for context [KEYBOARD]: Add another user link will not receive keyboard focus [AXE-CORE]: Heading levels should increase by one [COGNITION]: Error messages should be more specific [COLOR]: Blue button on blue background does not have sufficient contrast ratio --- --> <!-- It's okay to delete the instructions above, but leave the link to the 508 defect severity level for your issue. --> ## Feedback framework - **❗️ Must** for if the feedback must be applied - **⚠️ Should** if the feedback is best practice - **✔️ Consider** for suggestions/enhancements ## Definition of done 1. Review and acknowledge feedback. 1. Fix and/or document decisions made. 1. Accessibility specialist will close ticket after reviewing documented decisions / validating fix. ## Point of Contact <!-- If this issue is being opened by a VFS team member, please add a point of contact. Usually this is the same person who enters the issue ticket. --> **VFS Point of Contact:** _Trevor_ ## User Story or Problem Statement <!-- Example: As a user with cognitive considerations, I expect to see a label and input pairing consistently styled as throughout the rest of the site, with the label just above the text/email/search input or to the right of a radio/checkbox input, so that I am clearly able to understand what entry is expected. --> As an assistive tech user, I want to hear headings read out in the correct nesting order. ## Details <!-- This is a detailed description of the issue. It should include a restatement of the title, and provide more background information. --> Our app headings and sub-headings inside the accordions need to be H2 and H3 headings respectively. Screen shots attached below. ## Acceptance Criteria - [ ] Axe browser plugin doesn't report a heading nesting best practice warning on future runs - [ ] Current visual styles are maintained ## WCAG or Vendor Guidance (optional) * [Heading levels should only increase by one](https://dequeuniversity.com/rules/axe/4.0/heading-order) ## Screenshots or Trace Logs <!-- Drop any screenshots or error logs that might be useful for debugging --> ![Screen Shot 2020-11-19 at 2 22 53 PM](https://user-images.githubusercontent.com/934879/99720651-aacb7c00-2a73-11eb-9460-3649b9e5d32c.png) --- ![Screen Shot 2020-11-19 at 2 23 31 PM](https://user-images.githubusercontent.com/934879/99720682-b323b700-2a73-11eb-849b-f2b7171f2418.png)
non_test
defect app directory heading levels should increase by one enter an issue title using the format brief description of the problem edit buttons need aria label for context add another user link will not receive keyboard focus heading levels should increase by one error messages should be more specific blue button on blue background does not have sufficient contrast ratio feedback framework ❗️ must for if the feedback must be applied ⚠️ should if the feedback is best practice ✔️ consider for suggestions enhancements definition of done review and acknowledge feedback fix and or document decisions made accessibility specialist will close ticket after reviewing documented decisions validating fix point of contact vfs point of contact trevor user story or problem statement as an assistive tech user i want to hear headings read out in the correct nesting order details our app headings and sub headings inside the accordions need to be and headings respectively screen shots attached below acceptance criteria axe browser plugin doesn t report a heading nesting best practice warning on future runs current visual styles are maintained wcag or vendor guidance optional screenshots or trace logs
0
280,984
24,352,632,114
IssuesEvent
2022-10-03 02:42:31
ECP-WarpX/WarpX
https://api.github.com/repos/ECP-WarpX/WarpX
reopened
oneAPI 2022.2.0 Hangs in CI
bug component: tests install component: third party bug: affects latest release backend: dpc++ workaround
Since the update 1 week ago from `2022.1.0` to `2022.2.0`, most CI runs using either ICX (host) or DPC++ (device) compiles hang. It looks like this is from the linking part: https://github.com/ECP-WarpX/WarpX/pull/3421#issuecomment-1261311343 The same problem appears in the [AMReX](https://github.com/AMReX-Codes/amrex/) CI. Open this issue for triage and tracking. cc @rscohn2
2.0
oneAPI 2022.2.0 Hangs in CI - Since the update 1 week ago from `2022.1.0` to `2022.2.0`, most CI runs using either ICX (host) or DPC++ (device) compiles hang. It looks like this is from the linking part: https://github.com/ECP-WarpX/WarpX/pull/3421#issuecomment-1261311343 The same problem appears in the [AMReX](https://github.com/AMReX-Codes/amrex/) CI. Open this issue for triage and tracking. cc @rscohn2
test
oneapi hangs in ci since the update week ago from to most ci runs using either icx host or dpc device compiles hang it looks like this is from the linking part the same problem appears in the ci open this issue for triage and tracking cc
1
64,744
16,021,378,649
IssuesEvent
2021-04-21 00:14:04
jmuelbert/jmbde-QT
https://api.github.com/repos/jmuelbert/jmbde-QT
closed
Workflow: CD: RPM - openSUSE TW
build ci dependencies github_actions no-issue-activity
## Build the RPM is not really implemented Here is missing the dependencies.
1.0
Workflow: CD: RPM - openSUSE TW - ## Build the RPM is not really implemented Here is missing the dependencies.
non_test
workflow cd rpm opensuse tw build the rpm is not really implemented here is missing the dependencies
0
318,385
27,300,039,562
IssuesEvent
2023-02-24 00:37:24
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[REMOTO] [JAVA] [KAFKA] [AWS] [GIT] Pessoa Desenvolvedora Java Especialista na [INVILLIA]
HOME OFFICE JAVA SPRING SQL NOSQL AWS REMOTO JENKINS KAFKA GITFLOW TESTES UNITARIOS HELP WANTED ESPECIALISTA SPRING DATA SPRING BOOT Stale
<!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - Aproxime-se. A Invillia não apenas transformou a forma como as empresas mais revolucionárias do mundo criam e desenvolvem estratégias, negócios e produtos digitais. - Inovou também a maneira como pessoas apaixonadas por tecnologia, de qualquer lugar do planeta, podem interagir, evoluir, mais conectados do que nunca. - Para a Invillia, não importa onde você está. Se é um país grande. Ou uma cidade pequena. E sim a sua vontade. As suas ideias. O seu potencial. - O tamanho do seu talento. **Responsabilidades e atribuições:** - O profissional será responsável em prover soluções técnicas para novas features e dar o suporte necessário as features já existentes, afinal, nem tudo são flores. - Esperamos também que essa pessoa auxilie os outros membros do time em questões técnicas não esquecendo de fornecer a melhor solução para o negócio. - Algo que prezamos bastante é qualidade, isso inclui um código limpo e legível (clean code). - Também é desejável que o mesmo tenha um perfil intra-empreendedor, onde seus objetivos estejam alinhados com os objetivos da empresa, afinal, temos muito orgulho do que fazemos aqui! ## Local - Home Office ## Benefícios - Informações diretamente com o responsável pela vaga/recrutador. ## Requisitos **Obrigatórios:** - Experiência em desenvolvimento com Java; - Definição de Arquitetura exercendo o papel de Referência Técnica; - Experiência em desenvolvimento com Spring (Boot, Data, Cache, etc); - Conhecimentos em Java 8 (mínimo); - Conhecimento em Filas (Rabbit); - Conhecimento em Kafka; - Conhecimentos em AWS (SNS, SQS, S3); - Conhecimentos em Git e Git-Flow; - Experiência com bancos de dados SQL e NoSQL; - Desenvolvimento com foco em qualidade: testes unitários e Sonar(métricas); - Experiência em micro serviços e sistemas concorrentes; - Contínuos delivery (Jenkins); ## Contratação - a combinar ## Nossa empresa - A Invillia é uma empresa global que vem revolucionando a maneira como game-changers expandem o poder de inovar, implementar tecnologias de ponta e desenvolver novas estratégias, produtos e serviços digitais. - Nenhuma outra empresa no mundo atua como a Invillia. - E o que torna nosso Global Growth Framework tão único e poderoso? - Primeiro, dissolvemos os limites entre o físico e o virtual para ter em nosso time os melhores talentos do planeta. - Criamos infinitas práticas e metodologias para que que cada squad seja super customizado e engajado na cultura e desafios de cada cliente. - Adoramos usar ferramentas ágeis, métricas, inteligência de dados no dia-a-dia. Para que ideias e melhorias se multipliquem. - Mas acreditamos que é na educação contínua, na abordagem mais humana e colaborativa que a mágica acontece. - Novas oportunidades surgem. E a inovação nunca para. Infinite Digital Power. ## Como se candidatar - [Clique aqui para se candidatar](https://invillia.gupy.io/jobs/571873?jobBoardSource=gupy_public_page)
1.0
[REMOTO] [JAVA] [KAFKA] [AWS] [GIT] Pessoa Desenvolvedora Java Especialista na [INVILLIA] - <!-- ================================================== POR FAVOR, SÓ POSTE SE A VAGA FOR PARA SALVADOR E CIDADES VIZINHAS! Use: "Desenvolvedor Front-end" ao invés de "Front-End Developer" \o/ Exemplo: `[JAVASCRIPT] [MYSQL] [NODE.JS] Desenvolvedor Front-End na [NOME DA EMPRESA]` ================================================== --> ## Descrição da vaga - Aproxime-se. A Invillia não apenas transformou a forma como as empresas mais revolucionárias do mundo criam e desenvolvem estratégias, negócios e produtos digitais. - Inovou também a maneira como pessoas apaixonadas por tecnologia, de qualquer lugar do planeta, podem interagir, evoluir, mais conectados do que nunca. - Para a Invillia, não importa onde você está. Se é um país grande. Ou uma cidade pequena. E sim a sua vontade. As suas ideias. O seu potencial. - O tamanho do seu talento. **Responsabilidades e atribuições:** - O profissional será responsável em prover soluções técnicas para novas features e dar o suporte necessário as features já existentes, afinal, nem tudo são flores. - Esperamos também que essa pessoa auxilie os outros membros do time em questões técnicas não esquecendo de fornecer a melhor solução para o negócio. - Algo que prezamos bastante é qualidade, isso inclui um código limpo e legível (clean code). - Também é desejável que o mesmo tenha um perfil intra-empreendedor, onde seus objetivos estejam alinhados com os objetivos da empresa, afinal, temos muito orgulho do que fazemos aqui! ## Local - Home Office ## Benefícios - Informações diretamente com o responsável pela vaga/recrutador. ## Requisitos **Obrigatórios:** - Experiência em desenvolvimento com Java; - Definição de Arquitetura exercendo o papel de Referência Técnica; - Experiência em desenvolvimento com Spring (Boot, Data, Cache, etc); - Conhecimentos em Java 8 (mínimo); - Conhecimento em Filas (Rabbit); - Conhecimento em Kafka; - Conhecimentos em AWS (SNS, SQS, S3); - Conhecimentos em Git e Git-Flow; - Experiência com bancos de dados SQL e NoSQL; - Desenvolvimento com foco em qualidade: testes unitários e Sonar(métricas); - Experiência em micro serviços e sistemas concorrentes; - Contínuos delivery (Jenkins); ## Contratação - a combinar ## Nossa empresa - A Invillia é uma empresa global que vem revolucionando a maneira como game-changers expandem o poder de inovar, implementar tecnologias de ponta e desenvolver novas estratégias, produtos e serviços digitais. - Nenhuma outra empresa no mundo atua como a Invillia. - E o que torna nosso Global Growth Framework tão único e poderoso? - Primeiro, dissolvemos os limites entre o físico e o virtual para ter em nosso time os melhores talentos do planeta. - Criamos infinitas práticas e metodologias para que que cada squad seja super customizado e engajado na cultura e desafios de cada cliente. - Adoramos usar ferramentas ágeis, métricas, inteligência de dados no dia-a-dia. Para que ideias e melhorias se multipliquem. - Mas acreditamos que é na educação contínua, na abordagem mais humana e colaborativa que a mágica acontece. - Novas oportunidades surgem. E a inovação nunca para. Infinite Digital Power. ## Como se candidatar - [Clique aqui para se candidatar](https://invillia.gupy.io/jobs/571873?jobBoardSource=gupy_public_page)
test
pessoa desenvolvedora java especialista na por favor só poste se a vaga for para salvador e cidades vizinhas use desenvolvedor front end ao invés de front end developer o exemplo desenvolvedor front end na descrição da vaga aproxime se a invillia não apenas transformou a forma como as empresas mais revolucionárias do mundo criam e desenvolvem estratégias negócios e produtos digitais inovou também a maneira como pessoas apaixonadas por tecnologia de qualquer lugar do planeta podem interagir evoluir mais conectados do que nunca para a invillia não importa onde você está se é um país grande ou uma cidade pequena e sim a sua vontade as suas ideias o seu potencial o tamanho do seu talento responsabilidades e atribuições o profissional será responsável em prover soluções técnicas para novas features e dar o suporte necessário as features já existentes afinal nem tudo são flores esperamos também que essa pessoa auxilie os outros membros do time em questões técnicas não esquecendo de fornecer a melhor solução para o negócio algo que prezamos bastante é qualidade isso inclui um código limpo e legível clean code também é desejável que o mesmo tenha um perfil intra empreendedor onde seus objetivos estejam alinhados com os objetivos da empresa afinal temos muito orgulho do que fazemos aqui local home office benefícios informações diretamente com o responsável pela vaga recrutador requisitos obrigatórios experiência em desenvolvimento com java definição de arquitetura exercendo o papel de referência técnica experiência em desenvolvimento com spring boot data cache etc conhecimentos em java mínimo conhecimento em filas rabbit conhecimento em kafka conhecimentos em aws sns sqs conhecimentos em git e git flow experiência com bancos de dados sql e nosql desenvolvimento com foco em qualidade testes unitários e sonar métricas experiência em micro serviços e sistemas concorrentes contínuos delivery jenkins contratação a combinar nossa empresa a invillia é uma empresa global que vem revolucionando a maneira como game changers expandem o poder de inovar implementar tecnologias de ponta e desenvolver novas estratégias produtos e serviços digitais nenhuma outra empresa no mundo atua como a invillia e o que torna nosso global growth framework tão único e poderoso primeiro dissolvemos os limites entre o físico e o virtual para ter em nosso time os melhores talentos do planeta criamos infinitas práticas e metodologias para que que cada squad seja super customizado e engajado na cultura e desafios de cada cliente adoramos usar ferramentas ágeis métricas inteligência de dados no dia a dia para que ideias e melhorias se multipliquem mas acreditamos que é na educação contínua na abordagem mais humana e colaborativa que a mágica acontece novas oportunidades surgem e a inovação nunca para infinite digital power como se candidatar
1
158,802
24,899,512,456
IssuesEvent
2022-10-28 19:15:10
vegaprotocol/vegawallet-desktop
https://api.github.com/repos/vegaprotocol/vegawallet-desktop
closed
Revoke permissions
feature desktop-wallet backend ux-and-visual-design refine
Revoking permissions between wallet and hostname should automatically shutdown the connection between these two entities, if any.
1.0
Revoke permissions - Revoking permissions between wallet and hostname should automatically shutdown the connection between these two entities, if any.
non_test
revoke permissions revoking permissions between wallet and hostname should automatically shutdown the connection between these two entities if any
0
281,182
21,315,383,327
IssuesEvent
2022-04-16 07:15:19
putaojuice/pe
https://api.github.com/repos/putaojuice/pe
opened
Unclear use case for sort task in DG
severity.Low type.DocumentationBug
In DG the sort task use case step 1 of MSS states that sort the task by certain property, I feel like this could have been elaborated and stated clearer because in the UG there are so many properties covered **DG UC11** ![image.png](https://raw.githubusercontent.com/putaojuice/pe/main/files/014d1885-f4d3-43bd-9840-d6f90a326197.png) **UG** `sort` ![image.png](https://raw.githubusercontent.com/putaojuice/pe/main/files/cc541099-bb85-41d6-91e6-4b6da5f0dd20.png) <!--session: 1650088126549-fc759982-4493-4e69-bd46-1702e0a9f91f--> <!--Version: Web v3.4.2-->
1.0
Unclear use case for sort task in DG - In DG the sort task use case step 1 of MSS states that sort the task by certain property, I feel like this could have been elaborated and stated clearer because in the UG there are so many properties covered **DG UC11** ![image.png](https://raw.githubusercontent.com/putaojuice/pe/main/files/014d1885-f4d3-43bd-9840-d6f90a326197.png) **UG** `sort` ![image.png](https://raw.githubusercontent.com/putaojuice/pe/main/files/cc541099-bb85-41d6-91e6-4b6da5f0dd20.png) <!--session: 1650088126549-fc759982-4493-4e69-bd46-1702e0a9f91f--> <!--Version: Web v3.4.2-->
non_test
unclear use case for sort task in dg in dg the sort task use case step of mss states that sort the task by certain property i feel like this could have been elaborated and stated clearer because in the ug there are so many properties covered dg ug sort
0
252,178
21,561,161,074
IssuesEvent
2022-05-01 07:10:34
prestodb/presto
https://api.github.com/repos/prestodb/presto
closed
Flaky TestJdbcWarnings.testLongRunningStatement
tests stale
Build failed with ``` [ERROR] testLongRunningStatement(com.facebook.presto.jdbc.TestJdbcWarnings) Time elapsed: 0.172 s <<< FAILURE! java.lang.NullPointerException: throwable is null at java.util.Objects.requireNonNull(Objects.java:228) at com.facebook.presto.jdbc.TestJdbcWarnings$WarningEntry.<init>(TestJdbcWarnings.java:305) at com.facebook.presto.jdbc.TestJdbcWarnings.testLongRunningStatement(TestJdbcWarnings.java:150) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:104) at org.testng.internal.Invoker.invokeMethod(Invoker.java:645) at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:851) at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1177) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:129) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:112) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ``` The relevant part of the test is ``` while (statement.getWarnings() == null) { Thread.sleep(100); } SQLWarning warning = statement.getWarnings(); Set<WarningEntry> currentWarnings = new HashSet<>(); assertTrue(currentWarnings.add(new WarningEntry(warning))); ``` It seems like somehow the warnings are getting cleared or something between the two getWarnings() calls so that they are null at the second call.
1.0
Flaky TestJdbcWarnings.testLongRunningStatement - Build failed with ``` [ERROR] testLongRunningStatement(com.facebook.presto.jdbc.TestJdbcWarnings) Time elapsed: 0.172 s <<< FAILURE! java.lang.NullPointerException: throwable is null at java.util.Objects.requireNonNull(Objects.java:228) at com.facebook.presto.jdbc.TestJdbcWarnings$WarningEntry.<init>(TestJdbcWarnings.java:305) at com.facebook.presto.jdbc.TestJdbcWarnings.testLongRunningStatement(TestJdbcWarnings.java:150) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.testng.internal.MethodInvocationHelper.invokeMethod(MethodInvocationHelper.java:104) at org.testng.internal.Invoker.invokeMethod(Invoker.java:645) at org.testng.internal.Invoker.invokeTestMethod(Invoker.java:851) at org.testng.internal.Invoker.invokeTestMethods(Invoker.java:1177) at org.testng.internal.TestMethodWorker.invokeTestMethods(TestMethodWorker.java:129) at org.testng.internal.TestMethodWorker.run(TestMethodWorker.java:112) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) ``` The relevant part of the test is ``` while (statement.getWarnings() == null) { Thread.sleep(100); } SQLWarning warning = statement.getWarnings(); Set<WarningEntry> currentWarnings = new HashSet<>(); assertTrue(currentWarnings.add(new WarningEntry(warning))); ``` It seems like somehow the warnings are getting cleared or something between the two getWarnings() calls so that they are null at the second call.
test
flaky testjdbcwarnings testlongrunningstatement build failed with testlongrunningstatement com facebook presto jdbc testjdbcwarnings time elapsed s failure java lang nullpointerexception throwable is null at java util objects requirenonnull objects java at com facebook presto jdbc testjdbcwarnings warningentry testjdbcwarnings java at com facebook presto jdbc testjdbcwarnings testlongrunningstatement testjdbcwarnings java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org testng internal methodinvocationhelper invokemethod methodinvocationhelper java at org testng internal invoker invokemethod invoker java at org testng internal invoker invoketestmethod invoker java at org testng internal invoker invoketestmethods invoker java at org testng internal testmethodworker invoketestmethods testmethodworker java at org testng internal testmethodworker run testmethodworker java at java util concurrent threadpoolexecutor runworker threadpoolexecutor java at java util concurrent threadpoolexecutor worker run threadpoolexecutor java at java lang thread run thread java the relevant part of the test is while statement getwarnings null thread sleep sqlwarning warning statement getwarnings set currentwarnings new hashset asserttrue currentwarnings add new warningentry warning it seems like somehow the warnings are getting cleared or something between the two getwarnings calls so that they are null at the second call
1
168,250
13,067,467,057
IssuesEvent
2020-07-31 00:32:53
ayumi-cloud/oc-security-module
https://api.github.com/repos/ayumi-cloud/oc-security-module
closed
Add AppEngine detector to the Google firewall module
Add to Blacklist Add to Whitelist FINSIHED Firewall Priority: Medium Testing - Passed enhancement
### Enhancement idea - [x] Add `AppEngine` detector to the Google firewall module. Google App Engine is a Platform as a Service and cloud computing platform for developing and hosting web applications in Google-managed data centers. Applications are sandboxed and run across multiple servers.
1.0
Add AppEngine detector to the Google firewall module - ### Enhancement idea - [x] Add `AppEngine` detector to the Google firewall module. Google App Engine is a Platform as a Service and cloud computing platform for developing and hosting web applications in Google-managed data centers. Applications are sandboxed and run across multiple servers.
test
add appengine detector to the google firewall module enhancement idea add appengine detector to the google firewall module google app engine is a platform as a service and cloud computing platform for developing and hosting web applications in google managed data centers applications are sandboxed and run across multiple servers
1
8,373
8,272,528,201
IssuesEvent
2018-09-16 21:07:49
javaee/glassfish
https://api.github.com/repos/javaee/glassfish
closed
PostConstruct called twice for web service
Component: web_services ERR: Assignee Priority: Minor Type: Bug
Simple annotated web service with public constructor, private @PostConstruct and @PreDestroy methods, and a single web service method. For each call to the web service method, I see a new service object created and two calls to the PostConstruct method, e.g. INFO: WEB0671: Loading application [PostConstructIssue] at [/PostConstructIssue] INFO: PostConstructIssue was successfully deployed in 959 milliseconds. INFO: parsing WSDL... INFO: Generating code... INFO: Compiling code... INFO: Invoking wsimport with [http://localhost:8080/PostConstructIssue/MyWebServiceService?WSDL](http://localhost:8080/PostConstructIssue/MyWebServiceService?WSDL) INFO: wsimport successful INFO: webapp.postconstructissue.MyWebService@d4d7db constructor INFO: webapp.postconstructissue.MyWebService@d4d7db @PostConstruct init INFO: webapp.postconstructissue.MyWebService@d4d7db @PostConstruct init INFO: webapp.postconstructissue.MyWebService@c8d0e constructor INFO: webapp.postconstructissue.MyWebService@c8d0e @PostConstruct init INFO: webapp.postconstructissue.MyWebService@c8d0e @PostConstruct init Example Maven project attached. Aside from the double-call to the @PostConstruct method, I notice each webservice method call results in a new service object being created. This is different from GlassFish 2.1 behaviour - is it expected, and if so, is the old behaviour configurable? #### Environment GlassFish 3.1, Windows Vista #### Affected Versions [3.1]
1.0
PostConstruct called twice for web service - Simple annotated web service with public constructor, private @PostConstruct and @PreDestroy methods, and a single web service method. For each call to the web service method, I see a new service object created and two calls to the PostConstruct method, e.g. INFO: WEB0671: Loading application [PostConstructIssue] at [/PostConstructIssue] INFO: PostConstructIssue was successfully deployed in 959 milliseconds. INFO: parsing WSDL... INFO: Generating code... INFO: Compiling code... INFO: Invoking wsimport with [http://localhost:8080/PostConstructIssue/MyWebServiceService?WSDL](http://localhost:8080/PostConstructIssue/MyWebServiceService?WSDL) INFO: wsimport successful INFO: webapp.postconstructissue.MyWebService@d4d7db constructor INFO: webapp.postconstructissue.MyWebService@d4d7db @PostConstruct init INFO: webapp.postconstructissue.MyWebService@d4d7db @PostConstruct init INFO: webapp.postconstructissue.MyWebService@c8d0e constructor INFO: webapp.postconstructissue.MyWebService@c8d0e @PostConstruct init INFO: webapp.postconstructissue.MyWebService@c8d0e @PostConstruct init Example Maven project attached. Aside from the double-call to the @PostConstruct method, I notice each webservice method call results in a new service object being created. This is different from GlassFish 2.1 behaviour - is it expected, and if so, is the old behaviour configurable? #### Environment GlassFish 3.1, Windows Vista #### Affected Versions [3.1]
non_test
postconstruct called twice for web service simple annotated web service with public constructor private postconstruct and predestroy methods and a single web service method for each call to the web service method i see a new service object created and two calls to the postconstruct method e g info loading application at info postconstructissue was successfully deployed in milliseconds info parsing wsdl info generating code info compiling code info invoking wsimport with info wsimport successful info webapp postconstructissue mywebservice constructor info webapp postconstructissue mywebservice postconstruct init info webapp postconstructissue mywebservice postconstruct init info webapp postconstructissue mywebservice constructor info webapp postconstructissue mywebservice postconstruct init info webapp postconstructissue mywebservice postconstruct init example maven project attached aside from the double call to the postconstruct method i notice each webservice method call results in a new service object being created this is different from glassfish behaviour is it expected and if so is the old behaviour configurable environment glassfish windows vista affected versions
0
10,492
8,584,213,515
IssuesEvent
2018-11-13 22:02:43
nest/nest-simulator
https://api.github.com/repos/nest/nest-simulator
closed
Are you interested in `pip install pynest` ?
C: Infrastructure I: No breaking change P: In progess S: Low T: Discussion
After working on #844, I was wondering, if I could decouple NEST and PyNEST even further and had some interesting hacking hours. Now, I have a [prove of concept](https://github.com/tammoippen/nest-simulator/tree/pynest) to make NEST and PyNEST two independent projects, with NEST completely dependancy-free from PyNEST or Python installations and PyNEST installs with setup.py using `nest-config`. This gives: * No python / cython discovery and handling in CMakefiles, except for `docs` generation. * NEST does not need to link to the python lib, not even with `conngen` enabled. * `pip install nest` ! One could think about deploying PyNEST to http://pypi.python.org/ ? * Parallel installation of PyNEST for Python2 and Python3 with only one installation of NEST. * Possibly have PyNEST in a separate repository? * pytest (unrelated) On the other hand, this would require to install NEST and PyNEST separately, something like: ```sh cmake ... make make install pip install pynest/ ``` And two build chains have to be maintained! There are still some TODO's regarding documentation and testing, and maybe some questions regarding the design decisions. I would like to know, whether there is interest in this refactoring and if i should continue? Addressing the missing parts will require some effort on my side and I am unsure if I should invest. I am happy to do so, if there is consent. What changed in detail so far: * Remove most Python/Cython related stuff from CMakefiles * Extend `pynest/setup.py` to compile the cython extension, build and install PyNEST. `pip install pynest/` will install pynest to default python `site-packages` directories. * Put the python extension from the `topology` module also into PyNEST. Right now, it is not possible, to install nest without topology, hence the python extension will also always be installed. To reflect this in this changes, I moved the topology python files. * Remove `cythonize-pynest` option from CMakefiles: the setup.py excepts options to use rebuild python extensions. Crosscompiling still has to be tested. * Remove libneurosim / python parts from `neststartup.h/.cpp` and put them into a new header file in PyNEST: `pynestkernel.h` * move `pynestkernel.pyx/pxd` into the `nest` folder of pynest. * install python examples twice: once for nest / once for the pynest installation. * have `do_tests.sh.in` discover, whether there is a pynest installation and do tests, if they are found. (Use `pytest` instead of `nose` see #761. One could also think about putting that part into a separate PR; accept both `nose` or `pytest`; stick to `pytest` and also remove `do_tests.py.in` support altogether.) * there is also `extras/ConnPlotter/` python package ... I did not put this into pynest, because I am not sure, how both relate. Maybe this could also be a separate package / project / repository? * changed travis ci to reflect the install procedures. TravisCI works, but finds one python test failing ... I have to investigate: https://travis-ci.org/tammoippen/nest-simulator/builds/298469055 * The code is still not PR ready, has some debug stuff, and a ugly git history... TODOs: * Documentation. * Code cleanup. * Pytest fixing. * git history fixing. * cross compiling testing and documenting. see http://spscicomp.org/wordpress/wp-content/uploads/2014/05/BertrandBrelier_PythonBGQ.pdf slide 17ff) * unforeseeables ...
1.0
Are you interested in `pip install pynest` ? - After working on #844, I was wondering, if I could decouple NEST and PyNEST even further and had some interesting hacking hours. Now, I have a [prove of concept](https://github.com/tammoippen/nest-simulator/tree/pynest) to make NEST and PyNEST two independent projects, with NEST completely dependancy-free from PyNEST or Python installations and PyNEST installs with setup.py using `nest-config`. This gives: * No python / cython discovery and handling in CMakefiles, except for `docs` generation. * NEST does not need to link to the python lib, not even with `conngen` enabled. * `pip install nest` ! One could think about deploying PyNEST to http://pypi.python.org/ ? * Parallel installation of PyNEST for Python2 and Python3 with only one installation of NEST. * Possibly have PyNEST in a separate repository? * pytest (unrelated) On the other hand, this would require to install NEST and PyNEST separately, something like: ```sh cmake ... make make install pip install pynest/ ``` And two build chains have to be maintained! There are still some TODO's regarding documentation and testing, and maybe some questions regarding the design decisions. I would like to know, whether there is interest in this refactoring and if i should continue? Addressing the missing parts will require some effort on my side and I am unsure if I should invest. I am happy to do so, if there is consent. What changed in detail so far: * Remove most Python/Cython related stuff from CMakefiles * Extend `pynest/setup.py` to compile the cython extension, build and install PyNEST. `pip install pynest/` will install pynest to default python `site-packages` directories. * Put the python extension from the `topology` module also into PyNEST. Right now, it is not possible, to install nest without topology, hence the python extension will also always be installed. To reflect this in this changes, I moved the topology python files. * Remove `cythonize-pynest` option from CMakefiles: the setup.py excepts options to use rebuild python extensions. Crosscompiling still has to be tested. * Remove libneurosim / python parts from `neststartup.h/.cpp` and put them into a new header file in PyNEST: `pynestkernel.h` * move `pynestkernel.pyx/pxd` into the `nest` folder of pynest. * install python examples twice: once for nest / once for the pynest installation. * have `do_tests.sh.in` discover, whether there is a pynest installation and do tests, if they are found. (Use `pytest` instead of `nose` see #761. One could also think about putting that part into a separate PR; accept both `nose` or `pytest`; stick to `pytest` and also remove `do_tests.py.in` support altogether.) * there is also `extras/ConnPlotter/` python package ... I did not put this into pynest, because I am not sure, how both relate. Maybe this could also be a separate package / project / repository? * changed travis ci to reflect the install procedures. TravisCI works, but finds one python test failing ... I have to investigate: https://travis-ci.org/tammoippen/nest-simulator/builds/298469055 * The code is still not PR ready, has some debug stuff, and a ugly git history... TODOs: * Documentation. * Code cleanup. * Pytest fixing. * git history fixing. * cross compiling testing and documenting. see http://spscicomp.org/wordpress/wp-content/uploads/2014/05/BertrandBrelier_PythonBGQ.pdf slide 17ff) * unforeseeables ...
non_test
are you interested in pip install pynest after working on i was wondering if i could decouple nest and pynest even further and had some interesting hacking hours now i have a to make nest and pynest two independent projects with nest completely dependancy free from pynest or python installations and pynest installs with setup py using nest config this gives no python cython discovery and handling in cmakefiles except for docs generation nest does not need to link to the python lib not even with conngen enabled pip install nest one could think about deploying pynest to parallel installation of pynest for and with only one installation of nest possibly have pynest in a separate repository pytest unrelated on the other hand this would require to install nest and pynest separately something like sh cmake make make install pip install pynest and two build chains have to be maintained there are still some todo s regarding documentation and testing and maybe some questions regarding the design decisions i would like to know whether there is interest in this refactoring and if i should continue addressing the missing parts will require some effort on my side and i am unsure if i should invest i am happy to do so if there is consent what changed in detail so far remove most python cython related stuff from cmakefiles extend pynest setup py to compile the cython extension build and install pynest pip install pynest will install pynest to default python site packages directories put the python extension from the topology module also into pynest right now it is not possible to install nest without topology hence the python extension will also always be installed to reflect this in this changes i moved the topology python files remove cythonize pynest option from cmakefiles the setup py excepts options to use rebuild python extensions crosscompiling still has to be tested remove libneurosim python parts from neststartup h cpp and put them into a new header file in pynest pynestkernel h move pynestkernel pyx pxd into the nest folder of pynest install python examples twice once for nest once for the pynest installation have do tests sh in discover whether there is a pynest installation and do tests if they are found use pytest instead of nose see one could also think about putting that part into a separate pr accept both nose or pytest stick to pytest and also remove do tests py in support altogether there is also extras connplotter python package i did not put this into pynest because i am not sure how both relate maybe this could also be a separate package project repository changed travis ci to reflect the install procedures travisci works but finds one python test failing i have to investigate the code is still not pr ready has some debug stuff and a ugly git history todos documentation code cleanup pytest fixing git history fixing cross compiling testing and documenting see slide unforeseeables
0
126,423
17,030,062,020
IssuesEvent
2021-07-04 11:27:54
alan-turing-institute/sktime
https://api.github.com/repos/alan-turing-institute/sktime
reopened
Benchmarking design & implementation
API design implementing framework
## Evaluation - [ ] Extend evaluation API and functionality to single dataset case, currently only multiple dataset case is supported ## Performance metrics A few implementation notes: - Vectorised vs iterative computations - Callable classes vs classes with methods for computation - Use of jackknife by default for non point-wise metrics - Computation of standard error as decorator/mix-in - Have separate classes for point-wise metrics which can be wrapped by aggregation functions (e.g. mean) Also see https://github.com/JuliaML/LossFunctions.jl. ## Orchestration ### Should have - [ ] Allow orchestrator to be persisted to replicate benchmarking studies - [ ] add unit tests for `evaluator` methods - [ ] update all methods on evaluator to work on new internal data representation, also see https://www.statsmodels.org/stable/stats.html for some additional test implementations, e.g. the sign test, to improve readability, so that we can deprecate `_get_metrics_per_estimator_dataset` and `_get_metrics_per_estimator` methods - [ ] for saving results inside the `orchestrator` and for loading results in results classes use `_ResultsWrapper` to simply/unify interface, `_ResultsWrapper` needs to have slots for at least: y_true, y_pred, y_proba, index, fit_time, predict_time, strategy_name, dataset_name, cv_fold, train_or_test - [ ] No timing of fit and predict available, see https://docs.python.org/3/library/time.html#time.perf_counter, potentially have new `save_timings` and `load_timings` method - [ ] `orchestrator` cannot make probabilistic predictions, orchestrator tries to make probabilistic predictions using `predict_proba`, but (i) this will only works for some but not all classifiers and it won't work in regression, (ii) strategies currently don't even have a `predict_proba` (not even `TSCStrategy`), and (iii) current computation of `y_proba` fails if `y_pred` contains strings instead of integers which however is an accepted output format for classification I believe, add `predict_proba` to `TSCStrategy` - [ ] handling of probabilistic metrics in `evaluator` - [ ] no longer sure that saving results object as a master file is a good idea, as it may cause problems when multiple processes try to update it and because it needs to reflect the state of the directory somehow, maybe better to have a method on results object that allow to infer datasets, strategies and so on, something like a `register_results` method, instead of loading a fully specified dumped result object - [ ] separate `predict` method on `orchestrator` which loads and uses already fitted strategies - [ ] fix UEA results class ### Could have - [ ] allow for pre-defined cv splits in files - [ ] allow for pre-defined tasks in files - [ ] add `random_state` as input arg to orchestrator which is propagated to all strategies and cv - [ ] perhaps also useful to catch exceptions and skip over them in `orchestrator` instead of breaking execution? - [ ] currently only works for ts data input format, add other use cases - [ ] better user feedback, logging, keeping track of progress - [ ] many docstrings still missing or outdated - [ ] perhaps metrics shouldn't be wrapped in classes and the evaluator should take care of it internally, working with kwargs (e.g. `pointwise=True`) - [ ] handling of multiple metrics in `evaluator` - [ ] functionality for space filling parameter grids for large hyper-parameter search spaces (e.g. latin hypercube design), see this Python package: https://github.com/tirthajyoti/doepy - [ ] monitoring and comparison of memory usage of different estimators Related issues/PRs: #132
1.0
Benchmarking design & implementation - ## Evaluation - [ ] Extend evaluation API and functionality to single dataset case, currently only multiple dataset case is supported ## Performance metrics A few implementation notes: - Vectorised vs iterative computations - Callable classes vs classes with methods for computation - Use of jackknife by default for non point-wise metrics - Computation of standard error as decorator/mix-in - Have separate classes for point-wise metrics which can be wrapped by aggregation functions (e.g. mean) Also see https://github.com/JuliaML/LossFunctions.jl. ## Orchestration ### Should have - [ ] Allow orchestrator to be persisted to replicate benchmarking studies - [ ] add unit tests for `evaluator` methods - [ ] update all methods on evaluator to work on new internal data representation, also see https://www.statsmodels.org/stable/stats.html for some additional test implementations, e.g. the sign test, to improve readability, so that we can deprecate `_get_metrics_per_estimator_dataset` and `_get_metrics_per_estimator` methods - [ ] for saving results inside the `orchestrator` and for loading results in results classes use `_ResultsWrapper` to simply/unify interface, `_ResultsWrapper` needs to have slots for at least: y_true, y_pred, y_proba, index, fit_time, predict_time, strategy_name, dataset_name, cv_fold, train_or_test - [ ] No timing of fit and predict available, see https://docs.python.org/3/library/time.html#time.perf_counter, potentially have new `save_timings` and `load_timings` method - [ ] `orchestrator` cannot make probabilistic predictions, orchestrator tries to make probabilistic predictions using `predict_proba`, but (i) this will only works for some but not all classifiers and it won't work in regression, (ii) strategies currently don't even have a `predict_proba` (not even `TSCStrategy`), and (iii) current computation of `y_proba` fails if `y_pred` contains strings instead of integers which however is an accepted output format for classification I believe, add `predict_proba` to `TSCStrategy` - [ ] handling of probabilistic metrics in `evaluator` - [ ] no longer sure that saving results object as a master file is a good idea, as it may cause problems when multiple processes try to update it and because it needs to reflect the state of the directory somehow, maybe better to have a method on results object that allow to infer datasets, strategies and so on, something like a `register_results` method, instead of loading a fully specified dumped result object - [ ] separate `predict` method on `orchestrator` which loads and uses already fitted strategies - [ ] fix UEA results class ### Could have - [ ] allow for pre-defined cv splits in files - [ ] allow for pre-defined tasks in files - [ ] add `random_state` as input arg to orchestrator which is propagated to all strategies and cv - [ ] perhaps also useful to catch exceptions and skip over them in `orchestrator` instead of breaking execution? - [ ] currently only works for ts data input format, add other use cases - [ ] better user feedback, logging, keeping track of progress - [ ] many docstrings still missing or outdated - [ ] perhaps metrics shouldn't be wrapped in classes and the evaluator should take care of it internally, working with kwargs (e.g. `pointwise=True`) - [ ] handling of multiple metrics in `evaluator` - [ ] functionality for space filling parameter grids for large hyper-parameter search spaces (e.g. latin hypercube design), see this Python package: https://github.com/tirthajyoti/doepy - [ ] monitoring and comparison of memory usage of different estimators Related issues/PRs: #132
non_test
benchmarking design implementation evaluation extend evaluation api and functionality to single dataset case currently only multiple dataset case is supported performance metrics a few implementation notes vectorised vs iterative computations callable classes vs classes with methods for computation use of jackknife by default for non point wise metrics computation of standard error as decorator mix in have separate classes for point wise metrics which can be wrapped by aggregation functions e g mean also see orchestration should have allow orchestrator to be persisted to replicate benchmarking studies add unit tests for evaluator methods update all methods on evaluator to work on new internal data representation also see for some additional test implementations e g the sign test to improve readability so that we can deprecate get metrics per estimator dataset and get metrics per estimator methods for saving results inside the orchestrator and for loading results in results classes use resultswrapper to simply unify interface resultswrapper needs to have slots for at least y true y pred y proba index fit time predict time strategy name dataset name cv fold train or test no timing of fit and predict available see potentially have new save timings and load timings method orchestrator cannot make probabilistic predictions orchestrator tries to make probabilistic predictions using predict proba but i this will only works for some but not all classifiers and it won t work in regression ii strategies currently don t even have a predict proba not even tscstrategy and iii current computation of y proba fails if y pred contains strings instead of integers which however is an accepted output format for classification i believe add predict proba to tscstrategy handling of probabilistic metrics in evaluator no longer sure that saving results object as a master file is a good idea as it may cause problems when multiple processes try to update it and because it needs to reflect the state of the directory somehow maybe better to have a method on results object that allow to infer datasets strategies and so on something like a register results method instead of loading a fully specified dumped result object separate predict method on orchestrator which loads and uses already fitted strategies fix uea results class could have allow for pre defined cv splits in files allow for pre defined tasks in files add random state as input arg to orchestrator which is propagated to all strategies and cv perhaps also useful to catch exceptions and skip over them in orchestrator instead of breaking execution currently only works for ts data input format add other use cases better user feedback logging keeping track of progress many docstrings still missing or outdated perhaps metrics shouldn t be wrapped in classes and the evaluator should take care of it internally working with kwargs e g pointwise true handling of multiple metrics in evaluator functionality for space filling parameter grids for large hyper parameter search spaces e g latin hypercube design see this python package monitoring and comparison of memory usage of different estimators related issues prs
0
304,900
26,344,792,989
IssuesEvent
2023-01-10 20:57:58
jmarou/EventMapping
https://api.github.com/repos/jmarou/EventMapping
closed
Method using NLP: Words of interest using python library for NLP
A-feature B-Backend D-High C-development C-test
Develop method for extraction of words relating to location based on NLP Use spacy
1.0
Method using NLP: Words of interest using python library for NLP - Develop method for extraction of words relating to location based on NLP Use spacy
test
method using nlp words of interest using python library for nlp develop method for extraction of words relating to location based on nlp use spacy
1
298,752
9,200,871,202
IssuesEvent
2019-03-07 18:07:54
Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
https://api.github.com/repos/Warcraft-GoA-Development-Team/Warcraft-Guardians-of-Azeroth
closed
[LOCALIZATION] | INVADER_INFAMY
:beetle: bug - localization :scroll: :exclamation: priority high
**Mod Version** Master branch **Please explain your issue in as much detail as possible:** No loc for INVADER_INFAMY threat modifier **Upload screenshots of the problem localization:** <details> <summary>Click to expand</summary> ![ck2_3](https://user-images.githubusercontent.com/46576860/53856635-c95dbb00-3fe3-11e9-985d-42d0516a68e6.png) </details>
1.0
[LOCALIZATION] | INVADER_INFAMY - **Mod Version** Master branch **Please explain your issue in as much detail as possible:** No loc for INVADER_INFAMY threat modifier **Upload screenshots of the problem localization:** <details> <summary>Click to expand</summary> ![ck2_3](https://user-images.githubusercontent.com/46576860/53856635-c95dbb00-3fe3-11e9-985d-42d0516a68e6.png) </details>
non_test
invader infamy mod version master branch please explain your issue in as much detail as possible no loc for invader infamy threat modifier upload screenshots of the problem localization click to expand
0
328,086
28,100,777,339
IssuesEvent
2023-03-30 19:18:18
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: schemachange/mixed-versions-compat failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.schemachange/mixed-versions-compat [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=artifacts#/schemachange/mixed-versions-compat) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c): ``` test artifacts and logs in: /artifacts/schemachange/mixed-versions-compat/run_1 (mixed_version_decl_schemachange_compat.go:83).fetchCorpusToTmpDir: Missing validation corpus for release-22.2 (output in run_191732.317448944_n1_gsutil-cp-gscockroac: gsutil cp gs://cockroach-corpus/corpus-release-22.2/corpus /tmp/corpus3626686228/corpus returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_191732.322819992_n1_gsutil-cp-gscockroac.log: exit status 1) ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #100164 roachtest: schemachange/mixed-versions-compat failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-23.1 release-blocker] </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions-compat.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: schemachange/mixed-versions-compat failed - roachtest.schemachange/mixed-versions-compat [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/9329887?buildTab=artifacts#/schemachange/mixed-versions-compat) on master @ [1f8024bf14433ca169e5a8c3768c5d223dc5018c](https://github.com/cockroachdb/cockroach/commits/1f8024bf14433ca169e5a8c3768c5d223dc5018c): ``` test artifacts and logs in: /artifacts/schemachange/mixed-versions-compat/run_1 (mixed_version_decl_schemachange_compat.go:83).fetchCorpusToTmpDir: Missing validation corpus for release-22.2 (output in run_191732.317448944_n1_gsutil-cp-gscockroac: gsutil cp gs://cockroach-corpus/corpus-release-22.2/corpus /tmp/corpus3626686228/corpus returned: COMMAND_PROBLEM: ssh verbose log retained in ssh_191732.322819992_n1_gsutil-cp-gscockroac.log: exit status 1) ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> <details><summary>Same failure on other branches</summary> <p> - #100164 roachtest: schemachange/mixed-versions-compat failed [C-test-failure O-roachtest O-robot T-sql-schema branch-release-23.1 release-blocker] </p> </details> /cc @cockroachdb/sql-schema <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*schemachange/mixed-versions-compat.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest schemachange mixed versions compat failed roachtest schemachange mixed versions compat with on master test artifacts and logs in artifacts schemachange mixed versions compat run mixed version decl schemachange compat go fetchcorpustotmpdir missing validation corpus for release output in run gsutil cp gscockroac gsutil cp gs cockroach corpus corpus release corpus tmp corpus returned command problem ssh verbose log retained in ssh gsutil cp gscockroac log exit status parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see same failure on other branches roachtest schemachange mixed versions compat failed cc cockroachdb sql schema
1
374,968
26,141,364,963
IssuesEvent
2022-12-29 19:00:20
sukesh-ak/ESP32-TUX
https://api.github.com/repos/sukesh-ak/ESP32-TUX
closed
ESP-IDF compiler error - need docs improvements.
documentation question
I'm using Windows and also mac. There are huge gabs in the readme on what to do and it seems that it was written for a veteran of ESP32 developers. Right now I'm having trouble with the following steps: `idf.py set-target esp32s3` `idf.py build` `idf.py flash monitor` Which produce errors like the following: ``` ZacharysMBPM16:ESP32-TUX niteris$ idf.py flash monitor Executing action: flash Running cmake in directory /Users/niteris/dev/ESP32-TUX/build Executing "cmake -G Ninja -DPYTHON_DEPS_CHECKED=1 -DPYTHON=/Users/niteris/.espressif/python_env/idf5.1_py3.10_env/bin/python -DESP_PLATFORM=1 -DCCACHE_ENABLE=0 /Users/niteris/dev/ESP32-TUX"... -- IDF_TARGET not set, using default target: esp32 -- Found Git: /usr/bin/git (found version "2.37.1 (Apple Git-137.1)") -- Component directory /Users/niteris/dev/ESP32-TUX/components/ConfigBase does not contain a CMakeLists.txt file. No component will be added -- The C compiler identification is GNU 11.2.0 -- The CXX compiler identification is GNU 11.2.0 -- The ASM compiler identification is GNU -- Found assembler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-gcc -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-gcc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-g++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Building ESP-IDF components for target esp32 Processing 2 dependencies: [1/2] espressif/fmt (9.1.0) [2/2] idf (5.1.0) CMake Error at /Users/niteris/esp/esp-idf/tools/cmake/build.cmake:542 (message): ERROR: Some components (espressif/fmt) in the "managed_components" directory were modified on the disk since the last run of the CMake. Content of this directory is managed automatically. If you want to keep the changes, you can move the directory with the component to the "components" directory of your project. I.E. for "espressif__fmt" run: mv /Users/niteris/dev/ESP32-TUX/managed_components/espressif__fmt /Users/niteris/dev/ESP32-TUX/components/espressif__fmt Or, if you want to discard the changes remove the ".component_hash" file from the component's directory. I.E. for "espressif__fmt" run: rm /Users/niteris/dev/ESP32-TUX/managed_components/espressif__fmt/.component_hash Call Stack (most recent call first): /Users/niteris/esp/esp-idf/tools/cmake/project.cmake:447 (idf_build_process) CMakeLists.txt:28 (project) -- Configuring incomplete, errors occurred! See also "/Users/niteris/dev/ESP32-TUX/build/CMakeFiles/CMakeOutput.log". cmake failed with exit code 1, output of the command is in the /Users/niteris/dev/ESP32-TUX/build/log/idf_py_stderr_output_39505 and /Users/niteris/dev/ESP32-TUX/build/log/idf_py_stdout_output_39505 ```
1.0
ESP-IDF compiler error - need docs improvements. - I'm using Windows and also mac. There are huge gabs in the readme on what to do and it seems that it was written for a veteran of ESP32 developers. Right now I'm having trouble with the following steps: `idf.py set-target esp32s3` `idf.py build` `idf.py flash monitor` Which produce errors like the following: ``` ZacharysMBPM16:ESP32-TUX niteris$ idf.py flash monitor Executing action: flash Running cmake in directory /Users/niteris/dev/ESP32-TUX/build Executing "cmake -G Ninja -DPYTHON_DEPS_CHECKED=1 -DPYTHON=/Users/niteris/.espressif/python_env/idf5.1_py3.10_env/bin/python -DESP_PLATFORM=1 -DCCACHE_ENABLE=0 /Users/niteris/dev/ESP32-TUX"... -- IDF_TARGET not set, using default target: esp32 -- Found Git: /usr/bin/git (found version "2.37.1 (Apple Git-137.1)") -- Component directory /Users/niteris/dev/ESP32-TUX/components/ConfigBase does not contain a CMakeLists.txt file. No component will be added -- The C compiler identification is GNU 11.2.0 -- The CXX compiler identification is GNU 11.2.0 -- The ASM compiler identification is GNU -- Found assembler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-gcc -- Detecting C compiler ABI info -- Detecting C compiler ABI info - done -- Check for working C compiler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-gcc - skipped -- Detecting C compile features -- Detecting C compile features - done -- Detecting CXX compiler ABI info -- Detecting CXX compiler ABI info - done -- Check for working CXX compiler: /Users/niteris/.espressif/tools/xtensa-esp32-elf/esp-2022r1-11.2.0/xtensa-esp32-elf/bin/xtensa-esp32-elf-g++ - skipped -- Detecting CXX compile features -- Detecting CXX compile features - done -- Building ESP-IDF components for target esp32 Processing 2 dependencies: [1/2] espressif/fmt (9.1.0) [2/2] idf (5.1.0) CMake Error at /Users/niteris/esp/esp-idf/tools/cmake/build.cmake:542 (message): ERROR: Some components (espressif/fmt) in the "managed_components" directory were modified on the disk since the last run of the CMake. Content of this directory is managed automatically. If you want to keep the changes, you can move the directory with the component to the "components" directory of your project. I.E. for "espressif__fmt" run: mv /Users/niteris/dev/ESP32-TUX/managed_components/espressif__fmt /Users/niteris/dev/ESP32-TUX/components/espressif__fmt Or, if you want to discard the changes remove the ".component_hash" file from the component's directory. I.E. for "espressif__fmt" run: rm /Users/niteris/dev/ESP32-TUX/managed_components/espressif__fmt/.component_hash Call Stack (most recent call first): /Users/niteris/esp/esp-idf/tools/cmake/project.cmake:447 (idf_build_process) CMakeLists.txt:28 (project) -- Configuring incomplete, errors occurred! See also "/Users/niteris/dev/ESP32-TUX/build/CMakeFiles/CMakeOutput.log". cmake failed with exit code 1, output of the command is in the /Users/niteris/dev/ESP32-TUX/build/log/idf_py_stderr_output_39505 and /Users/niteris/dev/ESP32-TUX/build/log/idf_py_stdout_output_39505 ```
non_test
esp idf compiler error need docs improvements i m using windows and also mac there are huge gabs in the readme on what to do and it seems that it was written for a veteran of developers right now i m having trouble with the following steps idf py set target idf py build idf py flash monitor which produce errors like the following tux niteris idf py flash monitor executing action flash running cmake in directory users niteris dev tux build executing cmake g ninja dpython deps checked dpython users niteris espressif python env env bin python desp platform dccache enable users niteris dev tux idf target not set using default target found git usr bin git found version apple git component directory users niteris dev tux components configbase does not contain a cmakelists txt file no component will be added the c compiler identification is gnu the cxx compiler identification is gnu the asm compiler identification is gnu found assembler users niteris espressif tools xtensa elf esp xtensa elf bin xtensa elf gcc detecting c compiler abi info detecting c compiler abi info done check for working c compiler users niteris espressif tools xtensa elf esp xtensa elf bin xtensa elf gcc skipped detecting c compile features detecting c compile features done detecting cxx compiler abi info detecting cxx compiler abi info done check for working cxx compiler users niteris espressif tools xtensa elf esp xtensa elf bin xtensa elf g skipped detecting cxx compile features detecting cxx compile features done building esp idf components for target processing dependencies espressif fmt idf cmake error at users niteris esp esp idf tools cmake build cmake message error some components espressif fmt in the managed components directory were modified on the disk since the last run of the cmake content of this directory is managed automatically if you want to keep the changes you can move the directory with the component to the components directory of your project i e for espressif fmt run mv users niteris dev tux managed components espressif fmt users niteris dev tux components espressif fmt or if you want to discard the changes remove the component hash file from the component s directory i e for espressif fmt run rm users niteris dev tux managed components espressif fmt component hash call stack most recent call first users niteris esp esp idf tools cmake project cmake idf build process cmakelists txt project configuring incomplete errors occurred see also users niteris dev tux build cmakefiles cmakeoutput log cmake failed with exit code output of the command is in the users niteris dev tux build log idf py stderr output and users niteris dev tux build log idf py stdout output
0
10,764
7,302,159,490
IssuesEvent
2018-02-27 08:45:57
avocode/avocode
https://api.github.com/repos/avocode/avocode
closed
App is not loading on Windows 10
needsMoreInfo performance waitingForUser windows
<!-- Which version of avocode are you using? --> <!-- Please make sure you are using the latest version. --> <!-- You can always check and download the latest version from https://avocode.com/changelog.html. --> * **Version**: 2.14.8 <!-- Are you using the web or desktop version of Avocode? --> * **Web/Desktop**: Desktop <!-- What OS are you using? If you use the web version, which browser are you using? --> * **Platform**: Windows 10 <!-- Enter your issue details below this comment. --> App is not loading. It installs fine, I can see the entry in my Application Manager saying it's installed, I can see the process running in my Task Manager but nothing is showing on screen. The app is simply not loading.
True
App is not loading on Windows 10 - <!-- Which version of avocode are you using? --> <!-- Please make sure you are using the latest version. --> <!-- You can always check and download the latest version from https://avocode.com/changelog.html. --> * **Version**: 2.14.8 <!-- Are you using the web or desktop version of Avocode? --> * **Web/Desktop**: Desktop <!-- What OS are you using? If you use the web version, which browser are you using? --> * **Platform**: Windows 10 <!-- Enter your issue details below this comment. --> App is not loading. It installs fine, I can see the entry in my Application Manager saying it's installed, I can see the process running in my Task Manager but nothing is showing on screen. The app is simply not loading.
non_test
app is not loading on windows version web desktop desktop platform windows app is not loading it installs fine i can see the entry in my application manager saying it s installed i can see the process running in my task manager but nothing is showing on screen the app is simply not loading
0
349,397
31,798,711,265
IssuesEvent
2023-09-13 09:40:35
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
DISABLED test_nondeterministic_resize_quantized_cpu_quint4x2 (__main__.TestTorchDeviceTypeCPU)
module: tests triaged module: flaky-tests skipped oncall: pt2
Platforms: dynamo This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_nondeterministic_resize_quantized_cpu_quint4x2&suite=TestTorchDeviceTypeCPU) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/16740983982). Over the past 3 hours, it has been determined flaky in 6 workflow(s) with 6 failures and 6 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_nondeterministic_resize_quantized_cpu_quint4x2` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_torch.py`
2.0
DISABLED test_nondeterministic_resize_quantized_cpu_quint4x2 (__main__.TestTorchDeviceTypeCPU) - Platforms: dynamo This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_nondeterministic_resize_quantized_cpu_quint4x2&suite=TestTorchDeviceTypeCPU) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/16740983982). Over the past 3 hours, it has been determined flaky in 6 workflow(s) with 6 failures and 6 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_nondeterministic_resize_quantized_cpu_quint4x2` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_torch.py`
test
disabled test nondeterministic resize quantized cpu main testtorchdevicetypecpu platforms dynamo this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test nondeterministic resize quantized cpu there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test torch py
1
87,720
8,120,164,393
IssuesEvent
2018-08-16 01:00:41
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: jepsen/register/strobe-skews failed on master
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/d957838032a2dc162794efd23432f2672322f231 Parameters: Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=837166&tab=buildLog ``` test.go:494,jepsen.go:244,jepsen.go:288: /home/agent/work/.go/bin/roachprod run teamcity-837166-jepsen:6 -- bash -e -c "\ cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ ~/lein run test \ --tarball file://${PWD}/cockroach.tgz \ --username ${USER} \ --ssh-private-key ~/.ssh/id_rsa \ --os ubuntu \ --time-limit 300 \ --concurrency 30 \ --recovery-time 25 \ --test-count 1 \ -n 10.128.0.21 -n 10.128.0.5 -n 10.128.0.25 -n 10.128.0.23 -n 10.128.0.12 \ --test register --nemesis strobe-skews \ > invoke.log 2>&1 \ ": exit status 1 ```
1.0
roachtest: jepsen/register/strobe-skews failed on master - SHA: https://github.com/cockroachdb/cockroach/commits/d957838032a2dc162794efd23432f2672322f231 Parameters: Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=837166&tab=buildLog ``` test.go:494,jepsen.go:244,jepsen.go:288: /home/agent/work/.go/bin/roachprod run teamcity-837166-jepsen:6 -- bash -e -c "\ cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ ~/lein run test \ --tarball file://${PWD}/cockroach.tgz \ --username ${USER} \ --ssh-private-key ~/.ssh/id_rsa \ --os ubuntu \ --time-limit 300 \ --concurrency 30 \ --recovery-time 25 \ --test-count 1 \ -n 10.128.0.21 -n 10.128.0.5 -n 10.128.0.25 -n 10.128.0.23 -n 10.128.0.12 \ --test register --nemesis strobe-skews \ > invoke.log 2>&1 \ ": exit status 1 ```
test
roachtest jepsen register strobe skews failed on master sha parameters failed test test go jepsen go jepsen go home agent work go bin roachprod run teamcity jepsen bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test register nemesis strobe skews invoke log exit status
1
75,866
9,895,509,828
IssuesEvent
2019-06-26 08:00:47
google/draco
https://api.github.com/repos/google/draco
closed
Running the WebAssembly version of Draco in a web worker
documentation
Hey All, I am trying to run the WebAssembly version of Draco in a web worker. I have created the following worker code based on [this example](https://github.com/google/draco/blob/master/javascript/time_draco_decode.html): ```JavaScript importScripts("draco_wasm_wrapper.js"); var decoderModule = {}; var dracoDecoderType = {}; onmessage = function (e) { var compressedMeshBlob = e.data; var fileReader = new FileReader(); loadWebAssemblyDecoder(function decoderReady() { console.log("Decoder is ready."); fileReader.onload = function () { var arrayBuffer = fileReader.result; var geometryData = decodeMesh(arrayBuffer); postMessage(geometryData); geometryData = undefined; close(); }; fileReader.readAsArrayBuffer(compressedMeshBlob); }); }; function loadWebAssemblyDecoder(decoderReadyCallback) { dracoDecoderType['wasmBinaryFile'] = 'draco_decoder.wasm'; const xhr = new XMLHttpRequest(); xhr.open('GET', 'draco_decoder.wasm', true); xhr.responseType = 'arraybuffer'; xhr.onload = function () { dracoDecoderType['wasmBinary'] = xhr.response; decoderModule = DracoDecoderModule(dracoDecoderType); decoderReadyCallback(); }; xhr.send(null) } function decodeMesh(arrayBuffer) { const byteArray = new Uint8Array(arrayBuffer); const buffer = new decoderModule.DecoderBuffer(); buffer.Init(byteArray, byteArray.length); const decoder = new decoderModule.Decoder(); const geometryType = decoder.GetEncodedGeometryType(buffer); var dracoGeometry; if (geometryType === decoderModule.TRIANGULAR_MESH) { dracoGeometry = new decoderModule.Mesh(); } else { throw Error("Only triangular meshes are supported."); } decoderModule.destroy(decoder); decoderModule.destroy(buffer); return dracoGeometry; } ``` The `compressedMeshBlob` is a compressed Draco mesh which is loaded as a [Blob](https://developer.mozilla.org/en-US/docs/Web/API/Blob) by the main thread and passed to the web worker through `postMessage()`. I am using the [draco_decoder.wasm](https://github.com/google/draco/blob/master/javascript/draco_decoder.wasm) and [draco_wasm_wrapper.js](https://github.com/google/draco/blob/master/javascript/draco_wasm_wrapper.js) from the respository. I am getting the following error message: ![image](https://user-images.githubusercontent.com/4266299/39691950-80fddf54-51df-11e8-98dc-a12a1e4d39b1.png) I am not sure if I am doing something incorrectly or if it is fundamentally not possible to run the WebAssembly version of Draco in a web worker. Using the JavaScript decoder works fine. Thank you very much for your help in advance!
1.0
Running the WebAssembly version of Draco in a web worker - Hey All, I am trying to run the WebAssembly version of Draco in a web worker. I have created the following worker code based on [this example](https://github.com/google/draco/blob/master/javascript/time_draco_decode.html): ```JavaScript importScripts("draco_wasm_wrapper.js"); var decoderModule = {}; var dracoDecoderType = {}; onmessage = function (e) { var compressedMeshBlob = e.data; var fileReader = new FileReader(); loadWebAssemblyDecoder(function decoderReady() { console.log("Decoder is ready."); fileReader.onload = function () { var arrayBuffer = fileReader.result; var geometryData = decodeMesh(arrayBuffer); postMessage(geometryData); geometryData = undefined; close(); }; fileReader.readAsArrayBuffer(compressedMeshBlob); }); }; function loadWebAssemblyDecoder(decoderReadyCallback) { dracoDecoderType['wasmBinaryFile'] = 'draco_decoder.wasm'; const xhr = new XMLHttpRequest(); xhr.open('GET', 'draco_decoder.wasm', true); xhr.responseType = 'arraybuffer'; xhr.onload = function () { dracoDecoderType['wasmBinary'] = xhr.response; decoderModule = DracoDecoderModule(dracoDecoderType); decoderReadyCallback(); }; xhr.send(null) } function decodeMesh(arrayBuffer) { const byteArray = new Uint8Array(arrayBuffer); const buffer = new decoderModule.DecoderBuffer(); buffer.Init(byteArray, byteArray.length); const decoder = new decoderModule.Decoder(); const geometryType = decoder.GetEncodedGeometryType(buffer); var dracoGeometry; if (geometryType === decoderModule.TRIANGULAR_MESH) { dracoGeometry = new decoderModule.Mesh(); } else { throw Error("Only triangular meshes are supported."); } decoderModule.destroy(decoder); decoderModule.destroy(buffer); return dracoGeometry; } ``` The `compressedMeshBlob` is a compressed Draco mesh which is loaded as a [Blob](https://developer.mozilla.org/en-US/docs/Web/API/Blob) by the main thread and passed to the web worker through `postMessage()`. I am using the [draco_decoder.wasm](https://github.com/google/draco/blob/master/javascript/draco_decoder.wasm) and [draco_wasm_wrapper.js](https://github.com/google/draco/blob/master/javascript/draco_wasm_wrapper.js) from the respository. I am getting the following error message: ![image](https://user-images.githubusercontent.com/4266299/39691950-80fddf54-51df-11e8-98dc-a12a1e4d39b1.png) I am not sure if I am doing something incorrectly or if it is fundamentally not possible to run the WebAssembly version of Draco in a web worker. Using the JavaScript decoder works fine. Thank you very much for your help in advance!
non_test
running the webassembly version of draco in a web worker hey all i am trying to run the webassembly version of draco in a web worker i have created the following worker code based on javascript importscripts draco wasm wrapper js var decodermodule var dracodecodertype onmessage function e var compressedmeshblob e data var filereader new filereader loadwebassemblydecoder function decoderready console log decoder is ready filereader onload function var arraybuffer filereader result var geometrydata decodemesh arraybuffer postmessage geometrydata geometrydata undefined close filereader readasarraybuffer compressedmeshblob function loadwebassemblydecoder decoderreadycallback dracodecodertype draco decoder wasm const xhr new xmlhttprequest xhr open get draco decoder wasm true xhr responsetype arraybuffer xhr onload function dracodecodertype xhr response decodermodule dracodecodermodule dracodecodertype decoderreadycallback xhr send null function decodemesh arraybuffer const bytearray new arraybuffer const buffer new decodermodule decoderbuffer buffer init bytearray bytearray length const decoder new decodermodule decoder const geometrytype decoder getencodedgeometrytype buffer var dracogeometry if geometrytype decodermodule triangular mesh dracogeometry new decodermodule mesh else throw error only triangular meshes are supported decodermodule destroy decoder decodermodule destroy buffer return dracogeometry the compressedmeshblob is a compressed draco mesh which is loaded as a by the main thread and passed to the web worker through postmessage i am using the and from the respository i am getting the following error message i am not sure if i am doing something incorrectly or if it is fundamentally not possible to run the webassembly version of draco in a web worker using the javascript decoder works fine thank you very much for your help in advance
0
1,360
2,511,938,065
IssuesEvent
2015-01-14 12:46:40
transientskp/tkp
https://api.github.com/repos/transientskp/tkp
opened
Quality Control: Significant number of particular station type flagged
enhancement priority normal
Method: 1. Select all baselines between the imaged minimum and maximum baseline length (max_bl and min_bl given in the LOFAR_OBSERVATION table of the CASA image). The antenna id numbers are obtained from the LOFAR_ANTENNA table (i.e. find all pairs of antennas with a separation in the baseline range). For all the available baselines, identify the unique stations used and sort into core, remote and international. 2. Use the NVIS_USED and NVIS_TOTAL tables in the LOFAR_OBSERVATION table. Sum the maximum possible data available and sum the amount of data used for each type of station. 3. Calculate the percentage of flagged data ((1 - used/total)*100%) and flag the station as 'bad' if this exceeds the user defined maximum percentage allowed (same value as used in #3788). 4. Sum the number of 'bad' stations of a particular type and the total number of stations of a particular type. Use these to calculate the percentage of 'bad' baselines. Print the number of good and bad stations into the pipeline.log for each type of station. $type = (core remote international) 5. Reject image if the percentage of 'bad' stations in a particular category exceeds a user defined value (bad_station_percentage given in the quality_control.parset). Quality rejection reason: “Large amount of Measurement Set flagged. $value % of $type stations flagged.” original issue: https://support.astron.nl/lofar_issuetracker/issues/3884
1.0
Quality Control: Significant number of particular station type flagged - Method: 1. Select all baselines between the imaged minimum and maximum baseline length (max_bl and min_bl given in the LOFAR_OBSERVATION table of the CASA image). The antenna id numbers are obtained from the LOFAR_ANTENNA table (i.e. find all pairs of antennas with a separation in the baseline range). For all the available baselines, identify the unique stations used and sort into core, remote and international. 2. Use the NVIS_USED and NVIS_TOTAL tables in the LOFAR_OBSERVATION table. Sum the maximum possible data available and sum the amount of data used for each type of station. 3. Calculate the percentage of flagged data ((1 - used/total)*100%) and flag the station as 'bad' if this exceeds the user defined maximum percentage allowed (same value as used in #3788). 4. Sum the number of 'bad' stations of a particular type and the total number of stations of a particular type. Use these to calculate the percentage of 'bad' baselines. Print the number of good and bad stations into the pipeline.log for each type of station. $type = (core remote international) 5. Reject image if the percentage of 'bad' stations in a particular category exceeds a user defined value (bad_station_percentage given in the quality_control.parset). Quality rejection reason: “Large amount of Measurement Set flagged. $value % of $type stations flagged.” original issue: https://support.astron.nl/lofar_issuetracker/issues/3884
non_test
quality control significant number of particular station type flagged method select all baselines between the imaged minimum and maximum baseline length max bl and min bl given in the lofar observation table of the casa image the antenna id numbers are obtained from the lofar antenna table i e find all pairs of antennas with a separation in the baseline range for all the available baselines identify the unique stations used and sort into core remote and international use the nvis used and nvis total tables in the lofar observation table sum the maximum possible data available and sum the amount of data used for each type of station calculate the percentage of flagged data used total and flag the station as bad if this exceeds the user defined maximum percentage allowed same value as used in sum the number of bad stations of a particular type and the total number of stations of a particular type use these to calculate the percentage of bad baselines print the number of good and bad stations into the pipeline log for each type of station type core remote international reject image if the percentage of bad stations in a particular category exceeds a user defined value bad station percentage given in the quality control parset quality rejection reason “large amount of measurement set flagged value of type stations flagged ” original issue
0
91,280
26,346,600,782
IssuesEvent
2023-01-10 22:46:26
FashionFreedom/Seamly2D
https://api.github.com/repos/FashionFreedom/Seamly2D
closed
Workflows are referencing vulnerable actions
bug build
Hello, there! As part of the university research we are currently doing regarding the security of Github Actions, we noticed that one or many of the workflows that are part of this repository are referencing vulnerable versions of the third-party actions. As part of a disclosure process, we decided to open issues to notify GitHub Community. *Please note that there are could be some false positives in our methodology, thus not all of the open issues could be valid. If that is the case, please let us know, so that we can improve on our approach. You can contact me directly using an email: ikoishy [at] ncsu.edu* Thanks in advance 1. The workflow **build-release-assets.yml** is referencing action **ilammy/msvc-dev-cmd** using references **v1.4.1**. However this reference is missing the commit [74a501b](https://github.com/ilammy/msvc-dev-cmd/commits/74a501b) which may contain fix to the vulnerability. 1. The workflow **build-weekly-release.yml** is referencing action **ilammy/msvc-dev-cmd** using references **v1.4.1**. However this reference is missing the commit [74a501b](https://github.com/ilammy/msvc-dev-cmd/commits/74a501b) which may contain fix to the vulnerability. The vulnerability fix that is missing by actions' versions could be related to: (1) CVE fix (2) upgrade of vulnerable dependency (3) fix to secret leak and others. Please consider updating the reference to the action. *If you end up updating the reference, please let us know. We need the stats for the paper :-)*
1.0
Workflows are referencing vulnerable actions - Hello, there! As part of the university research we are currently doing regarding the security of Github Actions, we noticed that one or many of the workflows that are part of this repository are referencing vulnerable versions of the third-party actions. As part of a disclosure process, we decided to open issues to notify GitHub Community. *Please note that there are could be some false positives in our methodology, thus not all of the open issues could be valid. If that is the case, please let us know, so that we can improve on our approach. You can contact me directly using an email: ikoishy [at] ncsu.edu* Thanks in advance 1. The workflow **build-release-assets.yml** is referencing action **ilammy/msvc-dev-cmd** using references **v1.4.1**. However this reference is missing the commit [74a501b](https://github.com/ilammy/msvc-dev-cmd/commits/74a501b) which may contain fix to the vulnerability. 1. The workflow **build-weekly-release.yml** is referencing action **ilammy/msvc-dev-cmd** using references **v1.4.1**. However this reference is missing the commit [74a501b](https://github.com/ilammy/msvc-dev-cmd/commits/74a501b) which may contain fix to the vulnerability. The vulnerability fix that is missing by actions' versions could be related to: (1) CVE fix (2) upgrade of vulnerable dependency (3) fix to secret leak and others. Please consider updating the reference to the action. *If you end up updating the reference, please let us know. We need the stats for the paper :-)*
non_test
workflows are referencing vulnerable actions hello there as part of the university research we are currently doing regarding the security of github actions we noticed that one or many of the workflows that are part of this repository are referencing vulnerable versions of the third party actions as part of a disclosure process we decided to open issues to notify github community please note that there are could be some false positives in our methodology thus not all of the open issues could be valid if that is the case please let us know so that we can improve on our approach you can contact me directly using an email ikoishy ncsu edu thanks in advance the workflow build release assets yml is referencing action ilammy msvc dev cmd using references however this reference is missing the commit which may contain fix to the vulnerability the workflow build weekly release yml is referencing action ilammy msvc dev cmd using references however this reference is missing the commit which may contain fix to the vulnerability the vulnerability fix that is missing by actions versions could be related to cve fix upgrade of vulnerable dependency fix to secret leak and others please consider updating the reference to the action if you end up updating the reference please let us know we need the stats for the paper
0
693,451
23,775,902,484
IssuesEvent
2022-09-01 20:54:54
GoogleCloudPlatform/microservices-demo
https://api.github.com/repos/GoogleCloudPlatform/microservices-demo
closed
Update `skaffold` in `cloudbuild.yaml`
type: bug priority: p3
### Describe the bug * Our `skaffold.yaml` uses `skaffold` "Modules" — added in https://github.com/GoogleCloudPlatform/microservices-demo/pull/609. * But the our `cloudbuild.yaml` file in the root directory uses `skaffold` **v0.20.0** which doesn't support `skaffold` "Modules". https://github.com/GoogleCloudPlatform/microservices-demo/blob/8cd3ae52c93b2c5be5808c47bbfba1f8002ea13e/cloudbuild.yaml#L28 * We need to: * [ ] Update the version of `skaffold` being used to at least 1.27 (ideally, the latest) in the above `cloudbuild.yaml` file (in a new branch). * [ ] Test that updated `cloudbuild.yaml` file by creating a Google Cloud Build Trigger targeting the new branch. * [ ] Merge changes into `main`. * This issue was discovered in #752. ### To Reproduce * We would have to create a Google Cloud Build Trigger pointing to the `cloudbuild.yaml` file in `main`. ### Logs * I have not tested this breakage manually, but I assume we should see a message similar to [this](https://github.com/GoogleCloudPlatform/bank-of-anthos/pull/514#pullrequestreview-713779181) in the Google Cloud Build Trigger's logs: ``` WARN[0000] Your Skaffold version might be too old. Download the latest version (1.28.1) from: https://github.com/GoogleContainerTools/skaffold/releases/tag/v1.28.1 parsing skaffold config: error parsing skaffold configuration file: unknown skaffold config API version "skaffold/v2beta18". Set the config 'apiVersion' to a known value. Check https://skaffold.dev/docs/references/yaml/ for the list of valid API versions. Otherwise, check that your skaffold version is up-to-date. ``` ### Screenshots N/A ### Environment * Google Cloud Build ### Additional context N/A ### Exposure * This has been broken for a while now (`skaffold` modules was [added in Oct 2021](https://github.com/GoogleCloudPlatform/microservices-demo/pull/609)). * We don't have data on how many people are relying on our `cloudbuild.yaml` file (i.e., for learning/demo purposes) — I assume it's minimal.
1.0
Update `skaffold` in `cloudbuild.yaml` - ### Describe the bug * Our `skaffold.yaml` uses `skaffold` "Modules" — added in https://github.com/GoogleCloudPlatform/microservices-demo/pull/609. * But the our `cloudbuild.yaml` file in the root directory uses `skaffold` **v0.20.0** which doesn't support `skaffold` "Modules". https://github.com/GoogleCloudPlatform/microservices-demo/blob/8cd3ae52c93b2c5be5808c47bbfba1f8002ea13e/cloudbuild.yaml#L28 * We need to: * [ ] Update the version of `skaffold` being used to at least 1.27 (ideally, the latest) in the above `cloudbuild.yaml` file (in a new branch). * [ ] Test that updated `cloudbuild.yaml` file by creating a Google Cloud Build Trigger targeting the new branch. * [ ] Merge changes into `main`. * This issue was discovered in #752. ### To Reproduce * We would have to create a Google Cloud Build Trigger pointing to the `cloudbuild.yaml` file in `main`. ### Logs * I have not tested this breakage manually, but I assume we should see a message similar to [this](https://github.com/GoogleCloudPlatform/bank-of-anthos/pull/514#pullrequestreview-713779181) in the Google Cloud Build Trigger's logs: ``` WARN[0000] Your Skaffold version might be too old. Download the latest version (1.28.1) from: https://github.com/GoogleContainerTools/skaffold/releases/tag/v1.28.1 parsing skaffold config: error parsing skaffold configuration file: unknown skaffold config API version "skaffold/v2beta18". Set the config 'apiVersion' to a known value. Check https://skaffold.dev/docs/references/yaml/ for the list of valid API versions. Otherwise, check that your skaffold version is up-to-date. ``` ### Screenshots N/A ### Environment * Google Cloud Build ### Additional context N/A ### Exposure * This has been broken for a while now (`skaffold` modules was [added in Oct 2021](https://github.com/GoogleCloudPlatform/microservices-demo/pull/609)). * We don't have data on how many people are relying on our `cloudbuild.yaml` file (i.e., for learning/demo purposes) — I assume it's minimal.
non_test
update skaffold in cloudbuild yaml describe the bug our skaffold yaml uses skaffold modules — added in but the our cloudbuild yaml file in the root directory uses skaffold which doesn t support skaffold modules we need to update the version of skaffold being used to at least ideally the latest in the above cloudbuild yaml file in a new branch test that updated cloudbuild yaml file by creating a google cloud build trigger targeting the new branch merge changes into main this issue was discovered in to reproduce we would have to create a google cloud build trigger pointing to the cloudbuild yaml file in main logs i have not tested this breakage manually but i assume we should see a message similar to in the google cloud build trigger s logs warn your skaffold version might be too old download the latest version from parsing skaffold config error parsing skaffold configuration file unknown skaffold config api version skaffold set the config apiversion to a known value check for the list of valid api versions otherwise check that your skaffold version is up to date screenshots n a environment google cloud build additional context n a exposure this has been broken for a while now skaffold modules was we don t have data on how many people are relying on our cloudbuild yaml file i e for learning demo purposes — i assume it s minimal
0
684,745
23,428,876,485
IssuesEvent
2022-08-14 20:20:19
MartinXPN/profound.academy
https://api.github.com/repos/MartinXPN/profound.academy
opened
Re-evaluate is not scalable enough
bug enhancement priority/low
Re-evaluate dumps everything into a single transaction and becomes really not scalable. It can't process 200 submissions for instance. We should make it more scalable.
1.0
Re-evaluate is not scalable enough - Re-evaluate dumps everything into a single transaction and becomes really not scalable. It can't process 200 submissions for instance. We should make it more scalable.
non_test
re evaluate is not scalable enough re evaluate dumps everything into a single transaction and becomes really not scalable it can t process submissions for instance we should make it more scalable
0
52,153
27,402,195,518
IssuesEvent
2023-03-01 02:06:26
vmware-tanzu/velero
https://api.github.com/repos/vmware-tanzu/velero
closed
Velero backup duration using Kopia
Performance Needs investigation area/fs-backup
**Describe the problem/challenge you have** Noticed during testing, that once uploader=kopia is configured, cluster nodes are backup one by one instead of backup them in parallel. It increases duration of backup for Kopia, especially that node agents are run on each node. **Describe the solution you'd like** Consider to run Kopia backup in parallel on all nodes or have it configurable. **Anything else you would like to add:** Likely VCPU/Memory for velero deployment would increase, consider more tasks to run on node agents and velero deployment to manage status of agents. **Environment:** - Velero version (use `velero version`): 1.10.1 with Kopia uploaded - Kubernetes version (use `kubectl version`): AKS 1.24.6 - Kubernetes installer & version: 1.24.6 - Cloud provider or hardware configuration: Azure - OS (e.g. from `/etc/os-release`): Ubuntu 18 as part of AKS **Vote on this issue!** This is an invitation to the Velero community to vote on issues, you can see the project's [top voted issues listed here](https://github.com/vmware-tanzu/velero/issues?q=is%3Aissue+is%3Aopen+sort%3Areactions-%2B1-desc). Use the "reaction smiley face" up to the right of this comment to vote. - :+1: for "The project would be better with this feature added" - :-1: for "This feature will not enhance the project in a meaningful way"
True
Velero backup duration using Kopia - **Describe the problem/challenge you have** Noticed during testing, that once uploader=kopia is configured, cluster nodes are backup one by one instead of backup them in parallel. It increases duration of backup for Kopia, especially that node agents are run on each node. **Describe the solution you'd like** Consider to run Kopia backup in parallel on all nodes or have it configurable. **Anything else you would like to add:** Likely VCPU/Memory for velero deployment would increase, consider more tasks to run on node agents and velero deployment to manage status of agents. **Environment:** - Velero version (use `velero version`): 1.10.1 with Kopia uploaded - Kubernetes version (use `kubectl version`): AKS 1.24.6 - Kubernetes installer & version: 1.24.6 - Cloud provider or hardware configuration: Azure - OS (e.g. from `/etc/os-release`): Ubuntu 18 as part of AKS **Vote on this issue!** This is an invitation to the Velero community to vote on issues, you can see the project's [top voted issues listed here](https://github.com/vmware-tanzu/velero/issues?q=is%3Aissue+is%3Aopen+sort%3Areactions-%2B1-desc). Use the "reaction smiley face" up to the right of this comment to vote. - :+1: for "The project would be better with this feature added" - :-1: for "This feature will not enhance the project in a meaningful way"
non_test
velero backup duration using kopia describe the problem challenge you have noticed during testing that once uploader kopia is configured cluster nodes are backup one by one instead of backup them in parallel it increases duration of backup for kopia especially that node agents are run on each node describe the solution you d like consider to run kopia backup in parallel on all nodes or have it configurable anything else you would like to add likely vcpu memory for velero deployment would increase consider more tasks to run on node agents and velero deployment to manage status of agents environment velero version use velero version with kopia uploaded kubernetes version use kubectl version aks kubernetes installer version cloud provider or hardware configuration azure os e g from etc os release ubuntu as part of aks vote on this issue this is an invitation to the velero community to vote on issues you can see the project s use the reaction smiley face up to the right of this comment to vote for the project would be better with this feature added for this feature will not enhance the project in a meaningful way
0
295,563
25,484,535,942
IssuesEvent
2022-11-26 07:31:39
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
pkg/ccl/logictestccl/tests/3node-tenant/3node-tenant_test: TestTenantLogic_timetz failed
C-test-failure O-robot branch-release-22.2
pkg/ccl/logictestccl/tests/3node-tenant/3node-tenant_test.TestTenantLogic_timetz [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7698384?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7698384?buildTab=artifacts#/) on release-22.2 @ [fd133ba3af3954c93584a501cf68dfa3131e586c](https://github.com/cockroachdb/cockroach/commits/fd133ba3af3954c93584a501cf68dfa3131e586c): ``` Slow failing tests: TestTenantLogic_timetz - 1.41s Slow passing tests: TestTenantLogic_collatedstring - 160.43s TestTenantLogic_srfs - 133.88s TestTenantLogic_pg_builtins - 128.96s TestTenantLogic_fk - 84.38s TestTenantLogic_fk - 76.01s TestTenantLogic_alter_table - 70.13s TestTenantLogic_alter_table - 67.00s TestTenantLogic_alter_primary_key - 60.31s TestTenantLogic_alter_primary_key - 52.40s TestTenantLogic_create_as_non_metamorphic - 48.82s TestTenantLogic_cascade - 45.36s TestTenantLogic_enums - 43.74s TestTenantLogic_cascade - 43.16s TestTenantLogic_enums - 42.67s TestTenantLogic_drop_type - 40.63s TestTenantLogic_hash_sharded_index - 39.27s TestTenantLogic_new_schema_changer - 39.11s TestTenantLogic_new_schema_changer - 38.80s TestTenantLogic_hash_sharded_index - 38.06s TestTenantLogic_drop_type - 37.03s ``` <p>Parameters: <code>TAGS=bazel,gss</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantLogic_timetz.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
pkg/ccl/logictestccl/tests/3node-tenant/3node-tenant_test: TestTenantLogic_timetz failed - pkg/ccl/logictestccl/tests/3node-tenant/3node-tenant_test.TestTenantLogic_timetz [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7698384?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_StressBazel/7698384?buildTab=artifacts#/) on release-22.2 @ [fd133ba3af3954c93584a501cf68dfa3131e586c](https://github.com/cockroachdb/cockroach/commits/fd133ba3af3954c93584a501cf68dfa3131e586c): ``` Slow failing tests: TestTenantLogic_timetz - 1.41s Slow passing tests: TestTenantLogic_collatedstring - 160.43s TestTenantLogic_srfs - 133.88s TestTenantLogic_pg_builtins - 128.96s TestTenantLogic_fk - 84.38s TestTenantLogic_fk - 76.01s TestTenantLogic_alter_table - 70.13s TestTenantLogic_alter_table - 67.00s TestTenantLogic_alter_primary_key - 60.31s TestTenantLogic_alter_primary_key - 52.40s TestTenantLogic_create_as_non_metamorphic - 48.82s TestTenantLogic_cascade - 45.36s TestTenantLogic_enums - 43.74s TestTenantLogic_cascade - 43.16s TestTenantLogic_enums - 42.67s TestTenantLogic_drop_type - 40.63s TestTenantLogic_hash_sharded_index - 39.27s TestTenantLogic_new_schema_changer - 39.11s TestTenantLogic_new_schema_changer - 38.80s TestTenantLogic_hash_sharded_index - 38.06s TestTenantLogic_drop_type - 37.03s ``` <p>Parameters: <code>TAGS=bazel,gss</code> </p> <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) </p> </details> /cc @cockroachdb/sql-queries <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestTenantLogic_timetz.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
pkg ccl logictestccl tests tenant tenant test testtenantlogic timetz failed pkg ccl logictestccl tests tenant tenant test testtenantlogic timetz with on release slow failing tests testtenantlogic timetz slow passing tests testtenantlogic collatedstring testtenantlogic srfs testtenantlogic pg builtins testtenantlogic fk testtenantlogic fk testtenantlogic alter table testtenantlogic alter table testtenantlogic alter primary key testtenantlogic alter primary key testtenantlogic create as non metamorphic testtenantlogic cascade testtenantlogic enums testtenantlogic cascade testtenantlogic enums testtenantlogic drop type testtenantlogic hash sharded index testtenantlogic new schema changer testtenantlogic new schema changer testtenantlogic hash sharded index testtenantlogic drop type parameters tags bazel gss help see also cc cockroachdb sql queries
1
232,066
17,770,710,226
IssuesEvent
2021-08-30 13:21:35
juliuskoskela/corewar
https://api.github.com/repos/juliuskoskela/corewar
opened
create a README in project root
documentation
Add a README that would contain the following elements, for example: - short project description - how to compile and run each program - some example? - what kind of tests exist, and how to run them - links to documents that contain more information
1.0
create a README in project root - Add a README that would contain the following elements, for example: - short project description - how to compile and run each program - some example? - what kind of tests exist, and how to run them - links to documents that contain more information
non_test
create a readme in project root add a readme that would contain the following elements for example short project description how to compile and run each program some example what kind of tests exist and how to run them links to documents that contain more information
0
22,839
10,789,236,122
IssuesEvent
2019-11-05 11:27:12
zuhlke/ManageMyBudget
https://api.github.com/repos/zuhlke/ManageMyBudget
closed
Create VPC and security group to enable inbound connection to DocumentDB cluster
provisioning security terraform
Write a Terraform script to create VPC and security group for DocumentDB cluster to enable inbound connection
True
Create VPC and security group to enable inbound connection to DocumentDB cluster - Write a Terraform script to create VPC and security group for DocumentDB cluster to enable inbound connection
non_test
create vpc and security group to enable inbound connection to documentdb cluster write a terraform script to create vpc and security group for documentdb cluster to enable inbound connection
0
345,877
30,849,622,451
IssuesEvent
2023-08-02 15:48:25
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: knex failed
C-test-failure O-robot O-roachtest T-sql-foundations branch-release-22.2
roachtest.knex [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10044657?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10044657?buildTab=artifacts#/knex) on release-22.2 @ [97259ff3650b28abde99a8174d70016c9b5daa28](https://github.com/cockroachdb/cockroach/commits/97259ff3650b28abde99a8174d70016c9b5daa28): ``` test artifacts and logs in: /artifacts/knex/run_1 (knex.go:115).func1: COMMAND_PROBLEM: ssh verbose log retained in ssh_083745.354953804_n1_cd-mntdata1knex-DBco.log: exit status 1 ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-sessions <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*knex.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: knex failed - roachtest.knex [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10044657?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/10044657?buildTab=artifacts#/knex) on release-22.2 @ [97259ff3650b28abde99a8174d70016c9b5daa28](https://github.com/cockroachdb/cockroach/commits/97259ff3650b28abde99a8174d70016c9b5daa28): ``` test artifacts and logs in: /artifacts/knex/run_1 (knex.go:115).func1: COMMAND_PROBLEM: ssh verbose log retained in ssh_083745.354953804_n1_cd-mntdata1knex-DBco.log: exit status 1 ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/sql-sessions <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*knex.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest knex failed roachtest knex with on release test artifacts and logs in artifacts knex run knex go command problem ssh verbose log retained in ssh cd dbco log exit status parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest ssd help see see cc cockroachdb sql sessions
1
693,446
23,775,756,975
IssuesEvent
2022-09-01 20:45:43
kytos-ng/maintenance
https://api.github.com/repos/kytos-ng/maintenance
closed
rest endpoints don't include a version in the route
enhancement priority_low 2022.3 epic_maintenance_v1
Currently, this NApp isn't using versioned routes in the rest endpoints, most of the NApps we maintain such as `mef_eline`, `flow_manager`, `topology`, `storehouse`, `of_lldp` and so on uses a version like `v[1-9]+` in the route path which makes easier to provide certain guarantees for clients
1.0
rest endpoints don't include a version in the route - Currently, this NApp isn't using versioned routes in the rest endpoints, most of the NApps we maintain such as `mef_eline`, `flow_manager`, `topology`, `storehouse`, `of_lldp` and so on uses a version like `v[1-9]+` in the route path which makes easier to provide certain guarantees for clients
non_test
rest endpoints don t include a version in the route currently this napp isn t using versioned routes in the rest endpoints most of the napps we maintain such as mef eline flow manager topology storehouse of lldp and so on uses a version like v in the route path which makes easier to provide certain guarantees for clients
0
328,469
28,122,417,542
IssuesEvent
2023-03-31 15:04:20
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix ndarray.test_numpy_instance_copy__
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4508140756/jobs/7936571840" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4508140756/jobs/7936571840" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix ndarray.test_numpy_instance_copy__ - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4508140756/jobs/7936571840" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4508140756/jobs/7936571840" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="null" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
test
fix ndarray test numpy instance copy tensorflow img src torch img src numpy img src jax img src
1
9,575
3,059,000,962
IssuesEvent
2015-08-14 12:27:18
TypeCobolTeam/TypeCobol
https://api.github.com/repos/TypeCobolTeam/TypeCobol
closed
Unit Tests must be locale invariant
bug Tests
Whereas compiler should output message in the user's locale, we cannot allow unit tests to depend of their execution environment. In particular, current scanner tests compare numbers, separators and stuff against their string equivalents *in the french locale*. All tests should run in an invariant culture, so their results are predictable. Hint: add `Thread.CurrentThread.CurrentCulture = CultureInfo.InvariantCulture;` before any test runs.
1.0
Unit Tests must be locale invariant - Whereas compiler should output message in the user's locale, we cannot allow unit tests to depend of their execution environment. In particular, current scanner tests compare numbers, separators and stuff against their string equivalents *in the french locale*. All tests should run in an invariant culture, so their results are predictable. Hint: add `Thread.CurrentThread.CurrentCulture = CultureInfo.InvariantCulture;` before any test runs.
test
unit tests must be locale invariant whereas compiler should output message in the user s locale we cannot allow unit tests to depend of their execution environment in particular current scanner tests compare numbers separators and stuff against their string equivalents in the french locale all tests should run in an invariant culture so their results are predictable hint add thread currentthread currentculture cultureinfo invariantculture before any test runs
1
6,589
2,852,613,022
IssuesEvent
2015-06-01 14:31:00
NativeScript/NativeScript
https://api.github.com/repos/NativeScript/NativeScript
closed
Expose application level events as real events
2 - Open 3 - In Progress 4 - Ready For Test 5 - Done
Currently we are exposing the application level events with `onEventName` pattern. Consider providing real events as well: `addEventListener`/`on` in order to better handle all use cases. See [here](https://github.com/NativeScript/ios-runtime/issues/62#issuecomment-106268863) for more details.
1.0
Expose application level events as real events - Currently we are exposing the application level events with `onEventName` pattern. Consider providing real events as well: `addEventListener`/`on` in order to better handle all use cases. See [here](https://github.com/NativeScript/ios-runtime/issues/62#issuecomment-106268863) for more details.
test
expose application level events as real events currently we are exposing the application level events with oneventname pattern consider providing real events as well addeventlistener on in order to better handle all use cases see for more details
1
309,679
26,672,507,568
IssuesEvent
2023-01-26 11:33:55
wazuh/wazuh-qa
https://api.github.com/repos/wazuh/wazuh-qa
opened
Test unnecessary sanity check removed on wdb_syscollector_save2 function
team/qa type/manual-testing status/not-tracked role/qa-data-pirates subteam/qa-rainbow
| Target version | Related issue | Related PR/dev branch | |--------------------|--------------------|-----------------| | TBD | https://github.com/wazuh/wazuh/issues/10380 | https://github.com/wazuh/wazuh/pull/16008 | <!-- Important: No section may be left blank. If not, delete it directly (in principle only "Configurations" and "Considerations" could be left blank in case of not proceeding). --> ## Description <!-- Description that puts into context and shows the QA tester the changes that have been implemented and have to be tested. --> This development aims to fix a segmentation fault in Wazuh db syscollector in versions less than 4.2 ## Proposed test cases <!-- Indicate the minimum test cases proposed by the developer. --> - Upgrade to core-branch 1. Install wazuh agent 2. Enable the syscollector config 3. Check wazuh db does not crash in the manager 4. Check that there are no errors in the ossec.log - Fresh install core-branch 1. Install wazuh agent 2. Enable the syscollector config 3. Check wazuh db does not crash in the manager 4. Check that there are no errors in the ossec.log
1.0
Test unnecessary sanity check removed on wdb_syscollector_save2 function - | Target version | Related issue | Related PR/dev branch | |--------------------|--------------------|-----------------| | TBD | https://github.com/wazuh/wazuh/issues/10380 | https://github.com/wazuh/wazuh/pull/16008 | <!-- Important: No section may be left blank. If not, delete it directly (in principle only "Configurations" and "Considerations" could be left blank in case of not proceeding). --> ## Description <!-- Description that puts into context and shows the QA tester the changes that have been implemented and have to be tested. --> This development aims to fix a segmentation fault in Wazuh db syscollector in versions less than 4.2 ## Proposed test cases <!-- Indicate the minimum test cases proposed by the developer. --> - Upgrade to core-branch 1. Install wazuh agent 2. Enable the syscollector config 3. Check wazuh db does not crash in the manager 4. Check that there are no errors in the ossec.log - Fresh install core-branch 1. Install wazuh agent 2. Enable the syscollector config 3. Check wazuh db does not crash in the manager 4. Check that there are no errors in the ossec.log
test
test unnecessary sanity check removed on wdb syscollector function target version related issue related pr dev branch tbd description this development aims to fix a segmentation fault in wazuh db syscollector in versions less than proposed test cases upgrade to core branch install wazuh agent enable the syscollector config check wazuh db does not crash in the manager check that there are no errors in the ossec log fresh install core branch install wazuh agent enable the syscollector config check wazuh db does not crash in the manager check that there are no errors in the ossec log
1
95,824
27,628,379,722
IssuesEvent
2023-03-10 08:57:51
NVIDIA/spark-rapids-jni
https://api.github.com/repos/NVIDIA/spark-rapids-jni
closed
[FEA] Add CI for nightly or weekly fuzz testing for SparkRMM retry
build
**Is your feature request related to a problem? Please describe.** I added in a fuzz tester for Spark RMM retry https://github.com/NVIDIA/spark-rapids-jni/pull/977 It would be nice to run it with a random seed nightly or weekly to make sure that things are still working, and possibly find some errors that need to be fixed.
1.0
[FEA] Add CI for nightly or weekly fuzz testing for SparkRMM retry - **Is your feature request related to a problem? Please describe.** I added in a fuzz tester for Spark RMM retry https://github.com/NVIDIA/spark-rapids-jni/pull/977 It would be nice to run it with a random seed nightly or weekly to make sure that things are still working, and possibly find some errors that need to be fixed.
non_test
add ci for nightly or weekly fuzz testing for sparkrmm retry is your feature request related to a problem please describe i added in a fuzz tester for spark rmm retry it would be nice to run it with a random seed nightly or weekly to make sure that things are still working and possibly find some errors that need to be fixed
0
54,074
13,894,680,270
IssuesEvent
2020-10-19 14:57:50
jgeraigery/thingworx-gitbackup-extension
https://api.github.com/repos/jgeraigery/thingworx-gitbackup-extension
opened
CVE-2020-11111 (High) detected in jackson-databind-2.9.8.jar
security vulnerability
## CVE-2020-11111 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: thingworx-gitbackup-extension/twx-lib/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/thingworx-gitbackup-extension/commit/a274c5bf23bad29fe832613963ebac660dbc17bb">a274c5bf23bad29fe832613963ebac660dbc17bb</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.activemq.* (aka activemq-jms, activemq-core, activemq-pool, and activemq-pool-jms). <p>Publish Date: 2020-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11111>CVE-2020-11111</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11113">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11113</a></p> <p>Release Date: 2020-03-31</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0"}],"vulnerabilityIdentifier":"CVE-2020-11111","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.activemq.* (aka activemq-jms, activemq-core, activemq-pool, and activemq-pool-jms).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11111","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11111 (High) detected in jackson-databind-2.9.8.jar - ## CVE-2020-11111 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: thingworx-gitbackup-extension/twx-lib/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/jgeraigery/thingworx-gitbackup-extension/commit/a274c5bf23bad29fe832613963ebac660dbc17bb">a274c5bf23bad29fe832613963ebac660dbc17bb</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.activemq.* (aka activemq-jms, activemq-core, activemq-pool, and activemq-pool-jms). <p>Publish Date: 2020-03-31 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11111>CVE-2020-11111</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11113">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11113</a></p> <p>Release Date: 2020-03-31</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0</p> </p> </details> <p></p> *** :rescue_worker_helmet: Automatic Remediation is available for this issue <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4,2.10.0"}],"vulnerabilityIdentifier":"CVE-2020-11111","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.activemq.* (aka activemq-jms, activemq-core, activemq-pool, and activemq-pool-jms).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11111","cvss3Severity":"high","cvss3Score":"8.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in jackson databind jar cve high severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library thingworx gitbackup extension twx lib jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache activemq aka activemq jms activemq core activemq pool and activemq pool jms publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind rescue worker helmet automatic remediation is available for this issue isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache activemq aka activemq jms activemq core activemq pool and activemq pool jms vulnerabilityurl
0
47,704
5,908,625,104
IssuesEvent
2017-05-19 20:56:51
karmaradio/karma
https://api.github.com/repos/karmaradio/karma
closed
Email Verification
bug please-test T1d user-story
As a contractor/project manager, I want to verify my email when registering, So that my offers are more secure
1.0
Email Verification - As a contractor/project manager, I want to verify my email when registering, So that my offers are more secure
test
email verification as a contractor project manager i want to verify my email when registering so that my offers are more secure
1
217,640
16,857,866,969
IssuesEvent
2021-06-21 09:09:50
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
closed
PhoneHomeIntegrationTest.testMapLatenciesWithMapStore test fails on Windows Oracle JDK8
Source: Internal Team: Management Center Type: Test-Failure
**Failing test** com.hazelcast.internal.util.phonehome.PhoneHomeIntegrationTest.testMapLatenciesWithMapStore **Error message** ``` Expected 'mpgtlams' to be greater than or equal to 200, but was 199 ``` **Stacktrace** ``` java.lang.AssertionError: Expected 'mpgtlams' to be greater than or equal to 200, but was 199 at org.junit.Assert.fail(Assert.java:89) at org.junit.Assert.assertTrue(Assert.java:42) at com.hazelcast.test.HazelcastTestSupport.assertGreaterOrEquals(HazelcastTestSupport.java:1398) at com.hazelcast.internal.util.phonehome.PhoneHomeIntegrationTest.testMapLatenciesWithMapStore(PhoneHomeIntegrationTest.java:217) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) ``` **Jenkins build:** http://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-4.maintenance-Windows-OracleJDK8/311/
1.0
PhoneHomeIntegrationTest.testMapLatenciesWithMapStore test fails on Windows Oracle JDK8 - **Failing test** com.hazelcast.internal.util.phonehome.PhoneHomeIntegrationTest.testMapLatenciesWithMapStore **Error message** ``` Expected 'mpgtlams' to be greater than or equal to 200, but was 199 ``` **Stacktrace** ``` java.lang.AssertionError: Expected 'mpgtlams' to be greater than or equal to 200, but was 199 at org.junit.Assert.fail(Assert.java:89) at org.junit.Assert.assertTrue(Assert.java:42) at com.hazelcast.test.HazelcastTestSupport.assertGreaterOrEquals(HazelcastTestSupport.java:1398) at com.hazelcast.internal.util.phonehome.PhoneHomeIntegrationTest.testMapLatenciesWithMapStore(PhoneHomeIntegrationTest.java:217) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:59) at org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12) at org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:56) at org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:115) at com.hazelcast.test.FailOnTimeoutStatement$CallableStatement.call(FailOnTimeoutStatement.java:107) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.lang.Thread.run(Thread.java:748) ``` **Jenkins build:** http://jenkins.hazelcast.com/view/Official%20Builds/job/Hazelcast-4.maintenance-Windows-OracleJDK8/311/
test
phonehomeintegrationtest testmaplatencieswithmapstore test fails on windows oracle failing test com hazelcast internal util phonehome phonehomeintegrationtest testmaplatencieswithmapstore error message expected mpgtlams to be greater than or equal to but was stacktrace java lang assertionerror expected mpgtlams to be greater than or equal to but was at org junit assert fail assert java at org junit assert asserttrue assert java at com hazelcast test hazelcasttestsupport assertgreaterorequals hazelcasttestsupport java at com hazelcast internal util phonehome phonehomeintegrationtest testmaplatencieswithmapstore phonehomeintegrationtest java at sun reflect nativemethodaccessorimpl native method at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at org junit runners model frameworkmethod runreflectivecall frameworkmethod java at org junit internal runners model reflectivecallable run reflectivecallable java at org junit runners model frameworkmethod invokeexplosively frameworkmethod java at org junit internal runners statements invokemethod evaluate invokemethod java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at com hazelcast test failontimeoutstatement callablestatement call failontimeoutstatement java at java util concurrent futuretask run futuretask java at java lang thread run thread java jenkins build
1
167,321
13,019,751,408
IssuesEvent
2020-07-27 00:25:58
libra/libra
https://api.github.com/repos/libra/libra
closed
[cluster-test] Vault follow ups
cluster_test enhancement
Some things that are still missing from LSR/vault support - [x] We create instances for LSR/Vault, we don't add them to `Cluster`. We need to add vault_instances and lsr_instances fields and appropriate methods on the `Cluster`. This is required if we want people to use them in experiments - [x] Grouping API - we need some API on the cluster like `instances_for_group(ValidatorGroup)` - [x] Support in version compatibility experiment - need to leverage grouping api and update LSR tag in the compatibility experiment
1.0
[cluster-test] Vault follow ups - Some things that are still missing from LSR/vault support - [x] We create instances for LSR/Vault, we don't add them to `Cluster`. We need to add vault_instances and lsr_instances fields and appropriate methods on the `Cluster`. This is required if we want people to use them in experiments - [x] Grouping API - we need some API on the cluster like `instances_for_group(ValidatorGroup)` - [x] Support in version compatibility experiment - need to leverage grouping api and update LSR tag in the compatibility experiment
test
vault follow ups some things that are still missing from lsr vault support we create instances for lsr vault we don t add them to cluster we need to add vault instances and lsr instances fields and appropriate methods on the cluster this is required if we want people to use them in experiments grouping api we need some api on the cluster like instances for group validatorgroup support in version compatibility experiment need to leverage grouping api and update lsr tag in the compatibility experiment
1
113,005
9,608,167,727
IssuesEvent
2019-05-12 03:03:52
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: jepsen-batch3/multi-register/majority-ring failed
C-test-failure O-roachtest O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/45e15e05abff25e099ca59f4c5cb40a6cf695e6d Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=jepsen-batch3/multi-register/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1285294&tab=buildLog ``` The test failed on branch=master, cloud=gce: jepsen.go:260,jepsen.go:322,test.go:1251: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1285294-jepsen-batch3:6 -- bash -e -c "\ cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ ~/lein run test \ --tarball file://${PWD}/cockroach.tgz \ --username ${USER} \ --ssh-private-key ~/.ssh/id_rsa \ --os ubuntu \ --time-limit 300 \ --concurrency 30 \ --recovery-time 25 \ --test-count 1 \ -n 10.142.0.45 -n 10.142.0.4 -n 10.142.0.18 -n 10.142.0.25 -n 10.142.0.6 \ --test multi-register --nemesis majority-ring \ > invoke.log 2>&1 \ " returned: stderr: stdout: Error: exit status 255 : exit status 1 ```
2.0
roachtest: jepsen-batch3/multi-register/majority-ring failed - SHA: https://github.com/cockroachdb/cockroach/commits/45e15e05abff25e099ca59f4c5cb40a6cf695e6d Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=jepsen-batch3/multi-register/majority-ring PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1285294&tab=buildLog ``` The test failed on branch=master, cloud=gce: jepsen.go:260,jepsen.go:322,test.go:1251: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-1285294-jepsen-batch3:6 -- bash -e -c "\ cd /mnt/data1/jepsen/cockroachdb && set -eo pipefail && \ ~/lein run test \ --tarball file://${PWD}/cockroach.tgz \ --username ${USER} \ --ssh-private-key ~/.ssh/id_rsa \ --os ubuntu \ --time-limit 300 \ --concurrency 30 \ --recovery-time 25 \ --test-count 1 \ -n 10.142.0.45 -n 10.142.0.4 -n 10.142.0.18 -n 10.142.0.25 -n 10.142.0.6 \ --test multi-register --nemesis majority-ring \ > invoke.log 2>&1 \ " returned: stderr: stdout: Error: exit status 255 : exit status 1 ```
test
roachtest jepsen multi register majority ring failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests jepsen multi register majority ring pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on branch master cloud gce jepsen go jepsen go test go home agent work go src github com cockroachdb cockroach bin roachprod run teamcity jepsen bash e c cd mnt jepsen cockroachdb set eo pipefail lein run test tarball file pwd cockroach tgz username user ssh private key ssh id rsa os ubuntu time limit concurrency recovery time test count n n n n n test multi register nemesis majority ring invoke log returned stderr stdout error exit status exit status
1
25,194
4,148,880,376
IssuesEvent
2016-06-15 12:45:59
Gapminder/ddf-validation
https://api.github.com/repos/Gapminder/ddf-validation
closed
INVALID_DRILL_UP rule providing
effort1: medium (half-day) priority2: required status: done (tested) type: document type: enhancement type: feature type: rule
* **Rule name:** Should be filled if this issue depends on particular rule especially during new rule creation. This name should be a valid JS variable name and should be used for connecting with code. `INVALID_DRILL_UP` * **Rule test folder:** Should be filled if type of Request is `rule`. Needed for documentation auto generation. `test/fixtures/rules-cases/invalid-drill_up` * **Rule description:** Should be filled if type of Request is `rule`. Needed for documentation auto generation. An issue according to this rule will be fired when drillup in concept is defined and not valid: not a set of valid concepts * **Examples of correct data:** `ddf--concepts.csv` ``` concept,concept_type,domain,name,drill_up name,string,, geo,entity_domain,, region,entity_set,geo,Region country,entity_set,geo,Country,['geographic_regions', 'income_groups', 'landlocked', 'g77_and_oecd_countries', 'geographic_regions_in_4_colors', 'main_religion_2008'] capital,entity_set,geo,Capital pop,measure,geo,Population year,time,,year ``` * **Examples of incorrect data:** `ddf--concepts.csv` ``` concept,concept_type,domain,name,drill_up name,string,, geo,entity_domain,, region,entity_set,geo,Region country,entity_set,geo,Country,['foo', 'main_religion_2008'] capital,entity_set,geo,Capital,foo pop,measure,geo,Population year,time,,year ``` * **Scenarios** Should be filled if type of Request is `rule`. ``` when dataset is correct any issue should NOT be found for this rule ``` ``` when dataset is NOT correct issues in accordance with wrong concept drillups quantity should be detected for this rule output data for any issue should be expected ``` * **Output data format** Additional data that depends on particular issue type. Should be filled if type of Request is `rule`. wrong drillup value
1.0
INVALID_DRILL_UP rule providing - * **Rule name:** Should be filled if this issue depends on particular rule especially during new rule creation. This name should be a valid JS variable name and should be used for connecting with code. `INVALID_DRILL_UP` * **Rule test folder:** Should be filled if type of Request is `rule`. Needed for documentation auto generation. `test/fixtures/rules-cases/invalid-drill_up` * **Rule description:** Should be filled if type of Request is `rule`. Needed for documentation auto generation. An issue according to this rule will be fired when drillup in concept is defined and not valid: not a set of valid concepts * **Examples of correct data:** `ddf--concepts.csv` ``` concept,concept_type,domain,name,drill_up name,string,, geo,entity_domain,, region,entity_set,geo,Region country,entity_set,geo,Country,['geographic_regions', 'income_groups', 'landlocked', 'g77_and_oecd_countries', 'geographic_regions_in_4_colors', 'main_religion_2008'] capital,entity_set,geo,Capital pop,measure,geo,Population year,time,,year ``` * **Examples of incorrect data:** `ddf--concepts.csv` ``` concept,concept_type,domain,name,drill_up name,string,, geo,entity_domain,, region,entity_set,geo,Region country,entity_set,geo,Country,['foo', 'main_religion_2008'] capital,entity_set,geo,Capital,foo pop,measure,geo,Population year,time,,year ``` * **Scenarios** Should be filled if type of Request is `rule`. ``` when dataset is correct any issue should NOT be found for this rule ``` ``` when dataset is NOT correct issues in accordance with wrong concept drillups quantity should be detected for this rule output data for any issue should be expected ``` * **Output data format** Additional data that depends on particular issue type. Should be filled if type of Request is `rule`. wrong drillup value
test
invalid drill up rule providing rule name should be filled if this issue depends on particular rule especially during new rule creation this name should be a valid js variable name and should be used for connecting with code invalid drill up rule test folder should be filled if type of request is rule needed for documentation auto generation test fixtures rules cases invalid drill up rule description should be filled if type of request is rule needed for documentation auto generation an issue according to this rule will be fired when drillup in concept is defined and not valid not a set of valid concepts examples of correct data ddf concepts csv concept concept type domain name drill up name string geo entity domain region entity set geo region country entity set geo country capital entity set geo capital pop measure geo population year time year examples of incorrect data ddf concepts csv concept concept type domain name drill up name string geo entity domain region entity set geo region country entity set geo country capital entity set geo capital foo pop measure geo population year time year scenarios should be filled if type of request is rule when dataset is correct any issue should not be found for this rule when dataset is not correct issues in accordance with wrong concept drillups quantity should be detected for this rule output data for any issue should be expected output data format additional data that depends on particular issue type should be filled if type of request is rule wrong drillup value
1
300,385
25,962,668,139
IssuesEvent
2022-12-19 02:01:58
Ishiyuka/Kaizen
https://api.github.com/repos/Ishiyuka/Kaizen
opened
Rspecの実装
test
**model Spec バリデーションテスト** - [ ] user - [ ] team - [ ] issue - [ ] plan **System Spec ** - [ ] user - [ ] team, assign - [ ] issue - [ ] plan
1.0
Rspecの実装 - **model Spec バリデーションテスト** - [ ] user - [ ] team - [ ] issue - [ ] plan **System Spec ** - [ ] user - [ ] team, assign - [ ] issue - [ ] plan
test
rspecの実装 model spec バリデーションテスト user team issue plan system spec user team assign issue plan
1
331,126
24,292,676,415
IssuesEvent
2022-09-29 07:35:47
DHCJS/DHCJS
https://api.github.com/repos/DHCJS/DHCJS
closed
Heading
documentation
<h1 align="center">Hartato Cipta Jaya</h1> ![GitHub](https://user-images.githubusercontent.com/80773310/192715145-db470d3b-9ff9-4fc6-9be1-f38e71298683.png) <h2 align="center">About Me</h2> - 🍀Currently Learning HTML, CSS and JavaScript - 🍮Former FnB Staff & Part-Time 📈Indonesia Stock Trader - 🐑I Use [Medium](https://hartatociptajaya.medium.com/) To Blog My Journey, [Twitter](https://twitter.com/hartatocj) To Share My Achievements and Thoughts, and [LinkedIn](https://www.linkedin.com/in/hcj06/) To Connect. Feel Free To Connect With Me. <h2>Programming Language</h2> ![Li](https://user-images.githubusercontent.com/80773310/192737639-b309975f-e167-4887-8a18-aff16277bb2e.png)
1.0
Heading - <h1 align="center">Hartato Cipta Jaya</h1> ![GitHub](https://user-images.githubusercontent.com/80773310/192715145-db470d3b-9ff9-4fc6-9be1-f38e71298683.png) <h2 align="center">About Me</h2> - 🍀Currently Learning HTML, CSS and JavaScript - 🍮Former FnB Staff & Part-Time 📈Indonesia Stock Trader - 🐑I Use [Medium](https://hartatociptajaya.medium.com/) To Blog My Journey, [Twitter](https://twitter.com/hartatocj) To Share My Achievements and Thoughts, and [LinkedIn](https://www.linkedin.com/in/hcj06/) To Connect. Feel Free To Connect With Me. <h2>Programming Language</h2> ![Li](https://user-images.githubusercontent.com/80773310/192737639-b309975f-e167-4887-8a18-aff16277bb2e.png)
non_test
heading hartato cipta jaya about me 🍀currently learning html css and javascript 🍮former fnb staff part time 📈indonesia stock trader 🐑i use to blog my journey to share my achievements and thoughts and to connect feel free to connect with me programming language
0
98,458
12,325,454,671
IssuesEvent
2020-05-13 15:04:42
codingdavinci/relaunch2018
https://api.github.com/repos/codingdavinci/relaunch2018
opened
Abschnitt Zeitleiste- Fußzeile text align bottom
design
The Fußzeile text is aligning top when it should align bottom, as it does for Abschnitt Box/Box-Gruppen. Plx fix, ty!
1.0
Abschnitt Zeitleiste- Fußzeile text align bottom - The Fußzeile text is aligning top when it should align bottom, as it does for Abschnitt Box/Box-Gruppen. Plx fix, ty!
non_test
abschnitt zeitleiste fußzeile text align bottom the fußzeile text is aligning top when it should align bottom as it does for abschnitt box box gruppen plx fix ty
0
262,281
22,828,867,318
IssuesEvent
2022-07-12 11:05:23
clash-lang/clash-compiler
https://api.github.com/repos/clash-lang/clash-compiler
opened
Re-enable Verilator simulation in clash-testsuite
bug tests
I introduced a bug so that [verilator no longer runs simulation](https://github.com/clash-lang/clash-compiler/pull/2257#discussion_r916653627). We can also make some things more elegant, as Peter notes.
1.0
Re-enable Verilator simulation in clash-testsuite - I introduced a bug so that [verilator no longer runs simulation](https://github.com/clash-lang/clash-compiler/pull/2257#discussion_r916653627). We can also make some things more elegant, as Peter notes.
test
re enable verilator simulation in clash testsuite i introduced a bug so that we can also make some things more elegant as peter notes
1
171,679
14,342,861,395
IssuesEvent
2020-11-28 05:53:49
extratone/xyz
https://api.github.com/repos/extratone/xyz
opened
Create "The Unlicense, Dave Edition" Page
documentation
# The Unlicense, Dave Edition Considering my use of Git as a means to track revisions on my own writing, it's important to note that you should feel free to replace all instances of "software" in the statement below with "words," "writing," "expression," etc. While I appreciate tremendously the work and contribution of [Arlo Bendiken](https://ar.to/2010/01/set-your-code-free) in the form of The Unlicense, I would like to add that I see no need, personally, to make the refutation of Intellectual Property law regarding my own work into some profound ethical statement. As it stands, I have not to my knowledge been the victim of any sort of theft of any kind, and find the suggestion highly unlikely. If I *am* made aware of such a case, there is certainly a possibility that I may change my mind on this issue, so I suppose I should highlight that **THESE TERMS ARE SUBJECT TO CHANGE**, though I suspect I would be more flattered and/or amused than actually offended. I would also like to annotate that my use of this license should not be regarded as a suggestion that anyone else follow my example, or that I "believe in" taking such action regarding one's work, generally. While I did indeed take the time to set up [a repository](https://github.com/extratone/eoi) of John Perry Barlow's "[The Economy of Ideas](https://www.wired.com/1994/03/economy-ideas/)" in a bunch of different document formats, I do not necessarily agree with all of what he argued, especially in the context of the world 27 years later. If you *are* particularly interested in my opinion on the matter for whatever reason, here is what I have to say to you: The principled beginning of The Open Web was a great moment in history, but - like all historical principles - it is extremely important that we consider *context* and maintain an appropriate level of criticism when looking back on old manifestos written by old white guys. In general, try your best to *be reasonable*. *** This is free and unencumbered software released into the public domain. Anyone is free to copy, modify, publish, use, compile, sell, or distribute this software, either in source code form or as a compiled binary, for any purpose, commercial or non-commercial, and by any means. In jurisdictions that recognize copyright laws, the author or authors of this software dedicate any and all copyright interest in the software to the public domain. We make this dedication for the benefit of the public at large and to the detriment of our heirs and successors. We intend this dedication to be an overt act of relinquishment in perpetuity of all present and future rights to this software under copyright law. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. For more information, please refer to <https://unlicense.org>
1.0
Create "The Unlicense, Dave Edition" Page - # The Unlicense, Dave Edition Considering my use of Git as a means to track revisions on my own writing, it's important to note that you should feel free to replace all instances of "software" in the statement below with "words," "writing," "expression," etc. While I appreciate tremendously the work and contribution of [Arlo Bendiken](https://ar.to/2010/01/set-your-code-free) in the form of The Unlicense, I would like to add that I see no need, personally, to make the refutation of Intellectual Property law regarding my own work into some profound ethical statement. As it stands, I have not to my knowledge been the victim of any sort of theft of any kind, and find the suggestion highly unlikely. If I *am* made aware of such a case, there is certainly a possibility that I may change my mind on this issue, so I suppose I should highlight that **THESE TERMS ARE SUBJECT TO CHANGE**, though I suspect I would be more flattered and/or amused than actually offended. I would also like to annotate that my use of this license should not be regarded as a suggestion that anyone else follow my example, or that I "believe in" taking such action regarding one's work, generally. While I did indeed take the time to set up [a repository](https://github.com/extratone/eoi) of John Perry Barlow's "[The Economy of Ideas](https://www.wired.com/1994/03/economy-ideas/)" in a bunch of different document formats, I do not necessarily agree with all of what he argued, especially in the context of the world 27 years later. If you *are* particularly interested in my opinion on the matter for whatever reason, here is what I have to say to you: The principled beginning of The Open Web was a great moment in history, but - like all historical principles - it is extremely important that we consider *context* and maintain an appropriate level of criticism when looking back on old manifestos written by old white guys. In general, try your best to *be reasonable*. *** This is free and unencumbered software released into the public domain. Anyone is free to copy, modify, publish, use, compile, sell, or distribute this software, either in source code form or as a compiled binary, for any purpose, commercial or non-commercial, and by any means. In jurisdictions that recognize copyright laws, the author or authors of this software dedicate any and all copyright interest in the software to the public domain. We make this dedication for the benefit of the public at large and to the detriment of our heirs and successors. We intend this dedication to be an overt act of relinquishment in perpetuity of all present and future rights to this software under copyright law. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. For more information, please refer to <https://unlicense.org>
non_test
create the unlicense dave edition page the unlicense dave edition considering my use of git as a means to track revisions on my own writing it s important to note that you should feel free to replace all instances of software in the statement below with words writing expression etc while i appreciate tremendously the work and contribution of in the form of the unlicense i would like to add that i see no need personally to make the refutation of intellectual property law regarding my own work into some profound ethical statement as it stands i have not to my knowledge been the victim of any sort of theft of any kind and find the suggestion highly unlikely if i am made aware of such a case there is certainly a possibility that i may change my mind on this issue so i suppose i should highlight that these terms are subject to change though i suspect i would be more flattered and or amused than actually offended i would also like to annotate that my use of this license should not be regarded as a suggestion that anyone else follow my example or that i believe in taking such action regarding one s work generally while i did indeed take the time to set up of john perry barlow s in a bunch of different document formats i do not necessarily agree with all of what he argued especially in the context of the world years later if you are particularly interested in my opinion on the matter for whatever reason here is what i have to say to you the principled beginning of the open web was a great moment in history but like all historical principles it is extremely important that we consider context and maintain an appropriate level of criticism when looking back on old manifestos written by old white guys in general try your best to be reasonable this is free and unencumbered software released into the public domain anyone is free to copy modify publish use compile sell or distribute this software either in source code form or as a compiled binary for any purpose commercial or non commercial and by any means in jurisdictions that recognize copyright laws the author or authors of this software dedicate any and all copyright interest in the software to the public domain we make this dedication for the benefit of the public at large and to the detriment of our heirs and successors we intend this dedication to be an overt act of relinquishment in perpetuity of all present and future rights to this software under copyright law the software is provided as is without warranty of any kind express or implied including but not limited to the warranties of merchantability fitness for a particular purpose and noninfringement in no event shall the authors be liable for any claim damages or other liability whether in an action of contract tort or otherwise arising from out of or in connection with the software or the use or other dealings in the software for more information please refer to
0
150,426
11,960,386,275
IssuesEvent
2020-04-05 02:50:40
tensorflow/addons
https://api.github.com/repos/tensorflow/addons
closed
TimeStopping callback is untested.
bug callbacks help wanted test-cases
We have no way of knowing if it works or not. Some minimal tests would be greatly appreciated. See #964
1.0
TimeStopping callback is untested. - We have no way of knowing if it works or not. Some minimal tests would be greatly appreciated. See #964
test
timestopping callback is untested we have no way of knowing if it works or not some minimal tests would be greatly appreciated see
1
409,674
27,748,138,759
IssuesEvent
2023-03-15 18:31:13
LibertyDSNP/spec
https://api.github.com/repos/LibertyDSNP/spec
closed
DIP-227 User Data Operations
documentation enhancement blocked DSNP
Create a DIP that describes the new Replace User Data and Get User Data operations, their parameters, and outputs, as well as the User Data Changed Record.
1.0
DIP-227 User Data Operations - Create a DIP that describes the new Replace User Data and Get User Data operations, their parameters, and outputs, as well as the User Data Changed Record.
non_test
dip user data operations create a dip that describes the new replace user data and get user data operations their parameters and outputs as well as the user data changed record
0
293,177
25,275,304,640
IssuesEvent
2022-11-16 12:13:32
wazuh/wazuh
https://api.github.com/repos/wazuh/wazuh
closed
Release 4.3.10 - Revision 1 - Release Candidate RC1 - Footprint Metrics - ALL (2.5d)
team/cicd type/release release test/4.3.10
## Footprint metrics information | | | |---------------------------------| :--:| | **Main release candidate issue #** | #15372 | | **Main footprint metrics issue #** | #15348 | | **Version** | 4.3.10 | | **Release candidate #** | RC1 | | **Tag** | https://github.com/wazuh/wazuh/tree/4.3.10-rc1 | ## Stress test documentation ### Packages used - Repository: `packages-dev.wazuh.com` - Package path: `pre-release` - Package revision: `1` - **Jenkins build**: https://ci.wazuh.info/job/Test_stress/3655/ --- <details><summary>Manager</summary> + <details><summary>Plots</summary> ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_CPU.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Disk.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Disk_Read.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Disk_Written.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_FD.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_PSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Read_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_RSS_MAXMIN.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_RSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_SWAP.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_USS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_VMS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Write_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/Test_stress_B3655_manager_analysisd_state_EDPS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/Test_stress_B3655_manager_analysisd_state_Number_Events.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/Test_stress_B3655_manager_analysisd_state_Queues_state.png) </details> + <details><summary>Logs and configuration</summary> [ossec_Test_stress_B3655_manager_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/logs/ossec_Test_stress_B3655_manager_2022-11-14.zip) </details> + <details><summary>CSV</summary> [monitor-manager-Test_stress_B3655_manager-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/monitor-manager-Test_stress_B3655_manager-pre-release.csv) [Test_stress_B3655_manager_analysisd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/Test_stress_B3655_manager_analysisd_state.csv) [Test_stress_B3655_manager_remoted_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/Test_stress_B3655_manager_remoted_state.csv) </details> </details> <details><summary>Centos agent</summary> + <details><summary>Plots</summary> ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_CPU.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Disk.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Disk_Read.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Disk_Written.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_FD.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_PSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Read_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_RSS_MAXMIN.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_RSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_SWAP.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_USS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_VMS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Write_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/Test_stress_B3655_centos_agentd_state_AgentD_Number_of_events_buffered.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/Test_stress_B3655_centos_agentd_state_AgentD_Number_of_generated_events.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/Test_stress_B3655_centos_agentd_state_AgentD_Number_of_messages.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/Test_stress_B3655_centos_agentd_state_AgentD_Status.png) </details> + <details><summary>Logs and configuration</summary> [ossec_Test_stress_B3655_centos_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/logs/ossec_Test_stress_B3655_centos_2022-11-14.zip) </details> + <details><summary>CSV</summary> [monitor-agent-Test_stress_B3655_centos-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/data/monitor-agent-Test_stress_B3655_centos-pre-release.csv) [Test_stress_B3655_centos_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/data/Test_stress_B3655_centos_agentd_state.csv) </details> </details> <details><summary>Ubuntu agent</summary> + <details><summary>Plots</summary> ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_CPU.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Disk.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Disk_Read.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Disk_Written.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_FD.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_PSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Read_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_RSS_MAXMIN.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_RSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_SWAP.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_USS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_VMS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Write_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/Test_stress_B3655_ubuntu_agentd_state_AgentD_Number_of_events_buffered.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/Test_stress_B3655_ubuntu_agentd_state_AgentD_Number_of_generated_events.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/Test_stress_B3655_ubuntu_agentd_state_AgentD_Number_of_messages.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/Test_stress_B3655_ubuntu_agentd_state_AgentD_Status.png) </details> + <details><summary>Logs and configuration</summary> [ossec_Test_stress_B3655_ubuntu_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/logs/ossec_Test_stress_B3655_ubuntu_2022-11-14.zip) </details> + <details><summary>CSV</summary> [monitor-agent-Test_stress_B3655_ubuntu-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/data/monitor-agent-Test_stress_B3655_ubuntu-pre-release.csv) [Test_stress_B3655_ubuntu_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/data/Test_stress_B3655_ubuntu_agentd_state.csv) </details> </details> <details><summary>Windows agent</summary> + <details><summary>Plots</summary> ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_CPU.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Disk.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Disk_Read.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Disk_Written.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Handles.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Read_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_RSS_MAXMIN.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_RSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_USS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_VMS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Write_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/Test_stress_B3655_windows_agentd_state_AgentD_Number_of_events_buffered.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/Test_stress_B3655_windows_agentd_state_AgentD_Number_of_generated_events.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/Test_stress_B3655_windows_agentd_state_AgentD_Number_of_messages.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/Test_stress_B3655_windows_agentd_state_AgentD_Status.png) </details> + <details><summary>Logs and configuration</summary> [ossec_Test_stress_B3655_windows_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/logs/ossec_Test_stress_B3655_windows_2022-11-14.zip) </details> + <details><summary>CSV</summary> [monitor-winagent-Test_stress_B3655_windows-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/data/monitor-winagent-Test_stress_B3655_windows-pre-release.csv) [Test_stress_B3655_windows_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/data/Test_stress_B3655_windows_agentd_state.csv) </details> </details> <details><summary>macOS agent</summary> + <details><summary>Plots</summary> </details> + <details><summary>Logs and configuration</summary> </details> + <details><summary>CSV</summary> </details> </details> <details><summary>Solaris agent</summary> + <details><summary>Plots</summary> </details> + <details><summary>Logs and configuration</summary> </details> + <details><summary>CSV</summary> </details> </details>
1.0
Release 4.3.10 - Revision 1 - Release Candidate RC1 - Footprint Metrics - ALL (2.5d) - ## Footprint metrics information | | | |---------------------------------| :--:| | **Main release candidate issue #** | #15372 | | **Main footprint metrics issue #** | #15348 | | **Version** | 4.3.10 | | **Release candidate #** | RC1 | | **Tag** | https://github.com/wazuh/wazuh/tree/4.3.10-rc1 | ## Stress test documentation ### Packages used - Repository: `packages-dev.wazuh.com` - Package path: `pre-release` - Package revision: `1` - **Jenkins build**: https://ci.wazuh.info/job/Test_stress/3655/ --- <details><summary>Manager</summary> + <details><summary>Plots</summary> ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_CPU.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Disk.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Disk_Read.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Disk_Written.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_FD.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_PSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Read_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_RSS_MAXMIN.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_RSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_SWAP.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_USS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_VMS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/monitor-manager-Test_stress_B3655_manager-pre-release_Write_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/Test_stress_B3655_manager_analysisd_state_EDPS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/Test_stress_B3655_manager_analysisd_state_Number_Events.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/plots/Test_stress_B3655_manager_analysisd_state_Queues_state.png) </details> + <details><summary>Logs and configuration</summary> [ossec_Test_stress_B3655_manager_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/logs/ossec_Test_stress_B3655_manager_2022-11-14.zip) </details> + <details><summary>CSV</summary> [monitor-manager-Test_stress_B3655_manager-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/monitor-manager-Test_stress_B3655_manager-pre-release.csv) [Test_stress_B3655_manager_analysisd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/Test_stress_B3655_manager_analysisd_state.csv) [Test_stress_B3655_manager_remoted_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_manager_centos/data/Test_stress_B3655_manager_remoted_state.csv) </details> </details> <details><summary>Centos agent</summary> + <details><summary>Plots</summary> ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_CPU.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Disk.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Disk_Read.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Disk_Written.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_FD.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_PSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Read_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_RSS_MAXMIN.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_RSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_SWAP.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_USS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_VMS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/monitor-agent-Test_stress_B3655_centos-pre-release_Write_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/Test_stress_B3655_centos_agentd_state_AgentD_Number_of_events_buffered.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/Test_stress_B3655_centos_agentd_state_AgentD_Number_of_generated_events.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/Test_stress_B3655_centos_agentd_state_AgentD_Number_of_messages.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/plots/Test_stress_B3655_centos_agentd_state_AgentD_Status.png) </details> + <details><summary>Logs and configuration</summary> [ossec_Test_stress_B3655_centos_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/logs/ossec_Test_stress_B3655_centos_2022-11-14.zip) </details> + <details><summary>CSV</summary> [monitor-agent-Test_stress_B3655_centos-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/data/monitor-agent-Test_stress_B3655_centos-pre-release.csv) [Test_stress_B3655_centos_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_centos/data/Test_stress_B3655_centos_agentd_state.csv) </details> </details> <details><summary>Ubuntu agent</summary> + <details><summary>Plots</summary> ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_CPU.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Disk.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Disk_Read.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Disk_Written.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_FD.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_PSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Read_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_RSS_MAXMIN.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_RSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_SWAP.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_USS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_VMS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/monitor-agent-Test_stress_B3655_ubuntu-pre-release_Write_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/Test_stress_B3655_ubuntu_agentd_state_AgentD_Number_of_events_buffered.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/Test_stress_B3655_ubuntu_agentd_state_AgentD_Number_of_generated_events.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/Test_stress_B3655_ubuntu_agentd_state_AgentD_Number_of_messages.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/plots/Test_stress_B3655_ubuntu_agentd_state_AgentD_Status.png) </details> + <details><summary>Logs and configuration</summary> [ossec_Test_stress_B3655_ubuntu_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/logs/ossec_Test_stress_B3655_ubuntu_2022-11-14.zip) </details> + <details><summary>CSV</summary> [monitor-agent-Test_stress_B3655_ubuntu-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/data/monitor-agent-Test_stress_B3655_ubuntu-pre-release.csv) [Test_stress_B3655_ubuntu_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_ubuntu/data/Test_stress_B3655_ubuntu_agentd_state.csv) </details> </details> <details><summary>Windows agent</summary> + <details><summary>Plots</summary> ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_CPU.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Disk.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Disk_Read.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Disk_Written.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Handles.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Read_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_RSS_MAXMIN.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_RSS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_USS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_VMS.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/monitor-winagent-Test_stress_B3655_windows-pre-release_Write_Ops.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/Test_stress_B3655_windows_agentd_state_AgentD_Number_of_events_buffered.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/Test_stress_B3655_windows_agentd_state_AgentD_Number_of_generated_events.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/Test_stress_B3655_windows_agentd_state_AgentD_Number_of_messages.png) ![](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/plots/Test_stress_B3655_windows_agentd_state_AgentD_Status.png) </details> + <details><summary>Logs and configuration</summary> [ossec_Test_stress_B3655_windows_2022-11-14.zip](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/logs/ossec_Test_stress_B3655_windows_2022-11-14.zip) </details> + <details><summary>CSV</summary> [monitor-winagent-Test_stress_B3655_windows-pre-release.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/data/monitor-winagent-Test_stress_B3655_windows-pre-release.csv) [Test_stress_B3655_windows_agentd_state.csv](https://ci.wazuh.com/data/Test_stress/pre-release/4.3.10/B3655-3600m/B3655_agent_windows/data/Test_stress_B3655_windows_agentd_state.csv) </details> </details> <details><summary>macOS agent</summary> + <details><summary>Plots</summary> </details> + <details><summary>Logs and configuration</summary> </details> + <details><summary>CSV</summary> </details> </details> <details><summary>Solaris agent</summary> + <details><summary>Plots</summary> </details> + <details><summary>Logs and configuration</summary> </details> + <details><summary>CSV</summary> </details> </details>
test
release revision release candidate footprint metrics all footprint metrics information main release candidate issue main footprint metrics issue version release candidate tag stress test documentation packages used repository packages dev wazuh com package path pre release package revision jenkins build manager plots logs and configuration csv centos agent plots logs and configuration csv ubuntu agent plots logs and configuration csv windows agent plots logs and configuration csv macos agent plots logs and configuration csv solaris agent plots logs and configuration csv
1
214,277
24,056,925,381
IssuesEvent
2022-09-16 17:52:48
Agoric/agoric-sdk
https://api.github.com/repos/Agoric/agoric-sdk
opened
Await safety uncertain in cache/src/store.js
bug security
https://github.com/Agoric/agoric-sdk/blob/a1dedeae72908fda45afcb6038d76f8359adc8de/packages/cache/src/store.js#L82 The triage at https://github.com/Agoric/agoric-sdk/pull/6219 currently classifies this as // TODO FIXME This code should be refactored to make this // await checkably safe, or to remove it, or to record here // why it is actually safe. Git blame shows @turadg and @michaelfig as the ones who should probably investigate this, so I'm assigning to them. Feel free to reassign as appropriate of course.
True
Await safety uncertain in cache/src/store.js - https://github.com/Agoric/agoric-sdk/blob/a1dedeae72908fda45afcb6038d76f8359adc8de/packages/cache/src/store.js#L82 The triage at https://github.com/Agoric/agoric-sdk/pull/6219 currently classifies this as // TODO FIXME This code should be refactored to make this // await checkably safe, or to remove it, or to record here // why it is actually safe. Git blame shows @turadg and @michaelfig as the ones who should probably investigate this, so I'm assigning to them. Feel free to reassign as appropriate of course.
non_test
await safety uncertain in cache src store js the triage at currently classifies this as todo fixme this code should be refactored to make this await checkably safe or to remove it or to record here why it is actually safe git blame shows turadg and michaelfig as the ones who should probably investigate this so i m assigning to them feel free to reassign as appropriate of course
0
279,334
24,217,117,906
IssuesEvent
2022-09-26 07:48:33
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
[APM] Test Plan 8.5.0
Team:APM apm:test-plan-guide v8.5.0
# Test Plan 8.5.0 ## Phase 1 - Manual Testing: Monday 26th September - Pick an issue from the [curated list](https://github.com/elastic/kibana/issues?q=label%3Aapm%3Atest-plan-8.5.0+is%3Aclosed+-label%3Aapm%3Atest-plan-done) - After testing apply `apm:test-plan-done` label. If a problem is found a new issue should be opened and labelled `apm:test-plan-regression` and moved to "Ready" column. ### Run latest build candidate (BC) Using [apm-integration-testing](https://github.com/elastic/apm-integration-testing) it's trivial to start the stack for the latest BC: ``` ./scripts/compose.py start 8.5.0 --bc --all-opbeans ``` #### Check to see if issues are fixed before opening a bug Find the commit kibana was built with: ``` ./scripts/compose.py versions ... Kibana (image built: 2022-09-22 07:33:44 UTC): Version: 8.5.0-SNAPSHOT Branch: 8.5 Build SHA: - Build number: - ``` Using that Build SHA, visit https://github.com/elastic/kibana/compare/2aa076a33a2d67a72152240aa69ee1d26ff2485b...8.5 to see commits since the BC was built. ### Creating users locally ``` node ./x-pack/plugins/apm/scripts/create_apm_users.js --username elastic --password changeme --kibana-url http://localhost:5601 ``` Creates the following users (username / password): - `viewer` / `changeme` - `editor` / `changeme` ## Phase 2 - Improving Quality: Tuesday 27th September - Friday 30th September Primarily we need more e2e (cypress) tests and api tests. In addition to this we should replace the archives containing random opbeans data with synthbeans generated data. Below is a list of the important flows that should be covered in both a unit and an api test. ### Important areas & flows - Pending ### Migrate from opbeans archives to synthtrace scenarios We should replace the existing opbeans-based archives with data generated at runtime with synthtrace. ### Migrate API test to `apmApiClient` API tests should use the new `apmApiClient` instead of the legacy `supertest` clients. [Example on how to migrate.](https://github.com/elastic/kibana/issues/111308)
1.0
[APM] Test Plan 8.5.0 - # Test Plan 8.5.0 ## Phase 1 - Manual Testing: Monday 26th September - Pick an issue from the [curated list](https://github.com/elastic/kibana/issues?q=label%3Aapm%3Atest-plan-8.5.0+is%3Aclosed+-label%3Aapm%3Atest-plan-done) - After testing apply `apm:test-plan-done` label. If a problem is found a new issue should be opened and labelled `apm:test-plan-regression` and moved to "Ready" column. ### Run latest build candidate (BC) Using [apm-integration-testing](https://github.com/elastic/apm-integration-testing) it's trivial to start the stack for the latest BC: ``` ./scripts/compose.py start 8.5.0 --bc --all-opbeans ``` #### Check to see if issues are fixed before opening a bug Find the commit kibana was built with: ``` ./scripts/compose.py versions ... Kibana (image built: 2022-09-22 07:33:44 UTC): Version: 8.5.0-SNAPSHOT Branch: 8.5 Build SHA: - Build number: - ``` Using that Build SHA, visit https://github.com/elastic/kibana/compare/2aa076a33a2d67a72152240aa69ee1d26ff2485b...8.5 to see commits since the BC was built. ### Creating users locally ``` node ./x-pack/plugins/apm/scripts/create_apm_users.js --username elastic --password changeme --kibana-url http://localhost:5601 ``` Creates the following users (username / password): - `viewer` / `changeme` - `editor` / `changeme` ## Phase 2 - Improving Quality: Tuesday 27th September - Friday 30th September Primarily we need more e2e (cypress) tests and api tests. In addition to this we should replace the archives containing random opbeans data with synthbeans generated data. Below is a list of the important flows that should be covered in both a unit and an api test. ### Important areas & flows - Pending ### Migrate from opbeans archives to synthtrace scenarios We should replace the existing opbeans-based archives with data generated at runtime with synthtrace. ### Migrate API test to `apmApiClient` API tests should use the new `apmApiClient` instead of the legacy `supertest` clients. [Example on how to migrate.](https://github.com/elastic/kibana/issues/111308)
test
test plan test plan phase manual testing monday september pick an issue from the after testing apply apm test plan done label if a problem is found a new issue should be opened and labelled apm test plan regression and moved to ready column run latest build candidate bc using it s trivial to start the stack for the latest bc scripts compose py start bc all opbeans check to see if issues are fixed before opening a bug find the commit kibana was built with scripts compose py versions kibana image built utc version snapshot branch build sha build number using that build sha visit to see commits since the bc was built creating users locally node x pack plugins apm scripts create apm users js username elastic password changeme kibana url creates the following users username password viewer changeme editor changeme phase improving quality tuesday september friday september primarily we need more cypress tests and api tests in addition to this we should replace the archives containing random opbeans data with synthbeans generated data below is a list of the important flows that should be covered in both a unit and an api test important areas flows pending migrate from opbeans archives to synthtrace scenarios we should replace the existing opbeans based archives with data generated at runtime with synthtrace migrate api test to apmapiclient api tests should use the new apmapiclient instead of the legacy supertest clients
1
98,815
8,685,473,625
IssuesEvent
2018-12-03 07:52:45
humera987/FXLabs-Test-Automation
https://api.github.com/repos/humera987/FXLabs-Test-Automation
reopened
FX Testing 3 : ApiV1IssuesJobIdIdGetQueryParamPageInvalidDatatype
FX Testing 3
Project : FX Testing 3 Job : UAT Env : UAT Region : US_WEST Result : fail Status Code : 404 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=YzA4ZGZkNDItMjBkYS00NTk5LTkzMWItZDM2MDhmNWFhMjcw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 03 Dec 2018 07:45:48 GMT]} Endpoint : http://13.56.210.25/api/v1/api/v1/issues/job-id/buLiWKdz?page=n2rClY&status=buLiWKdz Request : Response : { "timestamp" : "2018-12-03T07:45:49.053+0000", "status" : 404, "error" : "Not Found", "message" : "No message available", "path" : "/api/v1/api/v1/issues/job-id/buLiWKdz" } Logs : Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed] --- FX Bot ---
1.0
FX Testing 3 : ApiV1IssuesJobIdIdGetQueryParamPageInvalidDatatype - Project : FX Testing 3 Job : UAT Env : UAT Region : US_WEST Result : fail Status Code : 404 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=YzA4ZGZkNDItMjBkYS00NTk5LTkzMWItZDM2MDhmNWFhMjcw; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 03 Dec 2018 07:45:48 GMT]} Endpoint : http://13.56.210.25/api/v1/api/v1/issues/job-id/buLiWKdz?page=n2rClY&status=buLiWKdz Request : Response : { "timestamp" : "2018-12-03T07:45:49.053+0000", "status" : 404, "error" : "Not Found", "message" : "No message available", "path" : "/api/v1/api/v1/issues/job-id/buLiWKdz" } Logs : Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed] --- FX Bot ---
test
fx testing project fx testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api issues job id buliwkdz logs assertion resolved to result assertion resolved to result fx bot
1
278,365
24,148,654,357
IssuesEvent
2022-09-21 21:22:19
unicode-org/icu4x
https://api.github.com/repos/unicode-org/icu4x
closed
Add benches for data provider constructors
T-docs-tests A-performance C-data-infra S-small
The data provider constructors are a big part of our startup cost. We should measure them and improve them if possible. I think it's fine for the tests to measure the testdata versions of the providers.
1.0
Add benches for data provider constructors - The data provider constructors are a big part of our startup cost. We should measure them and improve them if possible. I think it's fine for the tests to measure the testdata versions of the providers.
test
add benches for data provider constructors the data provider constructors are a big part of our startup cost we should measure them and improve them if possible i think it s fine for the tests to measure the testdata versions of the providers
1
63,345
14,656,703,165
IssuesEvent
2020-12-28 14:00:46
fu1771695yongxie/react-motion
https://api.github.com/repos/fu1771695yongxie/react-motion
opened
WS-2019-0032 (Medium) detected in js-yaml-3.7.0.tgz
security vulnerability
## WS-2019-0032 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary> <p>YAML 1.2 parser and serializer</p> <p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p> <p>Path to dependency file: react-motion/package.json</p> <p>Path to vulnerable library: react-motion/node_modules/js-yaml/package.json</p> <p> Dependency Hierarchy: - karma-coverage-1.1.2.tgz (Root Library) - istanbul-0.4.5.tgz - :x: **js-yaml-3.7.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/react-motion/commit/cc696c0e5f9cf48a30e0c7c74ea0d8811a5b1487">cc696c0e5f9cf48a30e0c7c74ea0d8811a5b1487</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service. <p>Publish Date: 2019-03-20 <p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p> <p>Release Date: 2019-03-20</p> <p>Fix Resolution: js-yaml - 3.13.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0032 (Medium) detected in js-yaml-3.7.0.tgz - ## WS-2019-0032 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>js-yaml-3.7.0.tgz</b></p></summary> <p>YAML 1.2 parser and serializer</p> <p>Library home page: <a href="https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz">https://registry.npmjs.org/js-yaml/-/js-yaml-3.7.0.tgz</a></p> <p>Path to dependency file: react-motion/package.json</p> <p>Path to vulnerable library: react-motion/node_modules/js-yaml/package.json</p> <p> Dependency Hierarchy: - karma-coverage-1.1.2.tgz (Root Library) - istanbul-0.4.5.tgz - :x: **js-yaml-3.7.0.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/fu1771695yongxie/react-motion/commit/cc696c0e5f9cf48a30e0c7c74ea0d8811a5b1487">cc696c0e5f9cf48a30e0c7c74ea0d8811a5b1487</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Versions js-yaml prior to 3.13.0 are vulnerable to Denial of Service. By parsing a carefully-crafted YAML file, the node process stalls and may exhaust system resources leading to a Denial of Service. <p>Publish Date: 2019-03-20 <p>URL: <a href=https://github.com/nodeca/js-yaml/commit/a567ef3c6e61eb319f0bfc2671d91061afb01235>WS-2019-0032</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/788/versions">https://www.npmjs.com/advisories/788/versions</a></p> <p>Release Date: 2019-03-20</p> <p>Fix Resolution: js-yaml - 3.13.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws medium detected in js yaml tgz ws medium severity vulnerability vulnerable library js yaml tgz yaml parser and serializer library home page a href path to dependency file react motion package json path to vulnerable library react motion node modules js yaml package json dependency hierarchy karma coverage tgz root library istanbul tgz x js yaml tgz vulnerable library found in head commit a href found in base branch master vulnerability details versions js yaml prior to are vulnerable to denial of service by parsing a carefully crafted yaml file the node process stalls and may exhaust system resources leading to a denial of service publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution js yaml step up your open source security game with whitesource
0
165,425
12,842,243,984
IssuesEvent
2020-07-08 01:29:27
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: django failed
C-test-failure O-roachtest O-robot branch-provisional_202007071743_v20.2.0-alpha.2 release-blocker
[(roachtest).django failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2071705&tab=buildLog) on [provisional_202007071743_v20.2.0-alpha.2@0b6e118bc1bcba4cfb4fc6c660153ec5be3989e8](https://github.com/cockroachdb/cockroach/commits/0b6e118bc1bcba4cfb4fc6c660153ec5be3989e8): ``` --- FAIL: introspection.tests.IntrospectionTests.test_table_names_with_views (unexpected) --- FAIL: queries.tests.ExcludeTest17600.test_only_orders_with_all_items_having_status_1 (unexpected) --- FAIL: backends.tests.FkConstraintsTests.test_disable_constraint_checks_manually (unexpected) --- FAIL: aggregation_regress.tests.AggregationTests.test_stddev (unexpected) --- FAIL: expressions.tests.FTimeDeltaTests.test_datetime_subtraction_microseconds (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_numeric_field_keep_null_status (unexpected) --- FAIL: schema.tests.SchemaTests.test_m2m_rename_field_in_target_model (unexpected) --- FAIL: schema.tests.SchemaTests.test_db_table (unexpected) --- FAIL: queries.tests.RelatedLookupTypeTests.test_values_queryset_lookup (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_order_with_respect_to (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_textual_field_keep_null_status (unexpected) --- FAIL: auth_tests.test_views.UUIDUserTests.test_admin_password_change (unexpected) --- FAIL: queries.tests.Ticket14056Tests.test_ticket_14056 (unexpected) --- FAIL: queries.tests.SubqueryTests.test_slice_subquery_and_query (unexpected) --- FAIL: many_to_one.tests.ManyToOneTests.test_fk_to_smallautofield (unexpected) --- FAIL: aggregation_regress.tests.AggregationTests.test_ticket_11293 (unexpected) --- FAIL: syndication_tests.tests.SyndicationFeedTest.test_latest_post_date (unexpected) --- FAIL: queries.test_explain.ExplainTests.test_unknown_format (unexpected) --- FAIL: db_functions.text.test_sha224.SHA224Tests.test_transform (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_not_unique_field_to_primary_key (unexpected) --- FAIL: admin_filters.tests.ListFiltersTests.test_simplelistfilter_with_queryset_based_lookups (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_field_pk_fk (unexpected) --- FAIL: syndication_tests.tests.SyndicationFeedTest.test_rss091_feed (unexpected) --- FAIL: queries.tests.NullInExcludeTest.test_col_not_in_list_containing_null (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_fk_non_fk (unexpected) --- FAIL: queries.test_bulk_update.BulkUpdateTests.test_inherited_fields (unexpected) --- FAIL: aggregation_regress.tests.AggregationTests.test_more_more (unexpected) --- FAIL: syndication_tests.tests.SyndicationFeedTest.test_rss2_feed (unexpected) --- FAIL: ordering.tests.OrderingTests.test_order_by_fk_attname (unexpected) --- FAIL: db_functions.math.test_power.PowerTests.test_integer (unexpected) --- FAIL: backends.tests.FkConstraintsTests.test_disable_constraint_checks_context_manager (unexpected) --- FAIL: queries.test_bulk_update.BulkUpdateNoteTests.test_multiple_fields (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_int_pk_to_int_unique (unexpected) --- FAIL: db_functions.text.test_sha384.SHA384Tests.test_basic (unexpected) --- FAIL: schema.tests.SchemaTests.test_add_field_temp_default (unexpected) --- FAIL: backends.tests.FkConstraintsTests.test_check_constraints (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_field_reloads_state_on_fk_with_to_field_target_changes (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_field_reloads_state_on_fk_with_to_field_target_type_change (unexpected) --- FAIL: introspection.tests.IntrospectionTests.test_get_table_description_types (unexpected) --- FAIL: schema.tests.SchemaTests.test_char_field_with_db_index_to_fk (unexpected) --- FAIL: db_functions.text.test_sha224.SHA224Tests.test_basic (unexpected) --- FAIL: timezones.tests.LegacyDatabaseTests.test_cursor_execute_accepts_naive_datetime (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_text_field_to_time_field (unexpected) --- FAIL: aggregation.test_filter_argument.FilteredAggregateTests.test_filtered_numerical_aggregates (unexpected) --- FAIL: introspection.tests.IntrospectionTests.test_get_table_description_col_lengths (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_text_field_to_datetime_field (unexpected) For a full summary look at the django artifacts An updated blocklist (djangoBlocklist20_2) is available in the artifacts' django log ``` <details><summary>More</summary><p> Artifacts: [/django](https://teamcity.cockroachdb.com/viewLog.html?buildId=2071705&tab=artifacts#/django) Related: - #51051 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #51049 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #50793 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006292135_v19.2.9](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006292135_v19.2.9) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #46800 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adjango.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: django failed - [(roachtest).django failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2071705&tab=buildLog) on [provisional_202007071743_v20.2.0-alpha.2@0b6e118bc1bcba4cfb4fc6c660153ec5be3989e8](https://github.com/cockroachdb/cockroach/commits/0b6e118bc1bcba4cfb4fc6c660153ec5be3989e8): ``` --- FAIL: introspection.tests.IntrospectionTests.test_table_names_with_views (unexpected) --- FAIL: queries.tests.ExcludeTest17600.test_only_orders_with_all_items_having_status_1 (unexpected) --- FAIL: backends.tests.FkConstraintsTests.test_disable_constraint_checks_manually (unexpected) --- FAIL: aggregation_regress.tests.AggregationTests.test_stddev (unexpected) --- FAIL: expressions.tests.FTimeDeltaTests.test_datetime_subtraction_microseconds (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_numeric_field_keep_null_status (unexpected) --- FAIL: schema.tests.SchemaTests.test_m2m_rename_field_in_target_model (unexpected) --- FAIL: schema.tests.SchemaTests.test_db_table (unexpected) --- FAIL: queries.tests.RelatedLookupTypeTests.test_values_queryset_lookup (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_order_with_respect_to (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_textual_field_keep_null_status (unexpected) --- FAIL: auth_tests.test_views.UUIDUserTests.test_admin_password_change (unexpected) --- FAIL: queries.tests.Ticket14056Tests.test_ticket_14056 (unexpected) --- FAIL: queries.tests.SubqueryTests.test_slice_subquery_and_query (unexpected) --- FAIL: many_to_one.tests.ManyToOneTests.test_fk_to_smallautofield (unexpected) --- FAIL: aggregation_regress.tests.AggregationTests.test_ticket_11293 (unexpected) --- FAIL: syndication_tests.tests.SyndicationFeedTest.test_latest_post_date (unexpected) --- FAIL: queries.test_explain.ExplainTests.test_unknown_format (unexpected) --- FAIL: db_functions.text.test_sha224.SHA224Tests.test_transform (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_not_unique_field_to_primary_key (unexpected) --- FAIL: admin_filters.tests.ListFiltersTests.test_simplelistfilter_with_queryset_based_lookups (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_field_pk_fk (unexpected) --- FAIL: syndication_tests.tests.SyndicationFeedTest.test_rss091_feed (unexpected) --- FAIL: queries.tests.NullInExcludeTest.test_col_not_in_list_containing_null (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_fk_non_fk (unexpected) --- FAIL: queries.test_bulk_update.BulkUpdateTests.test_inherited_fields (unexpected) --- FAIL: aggregation_regress.tests.AggregationTests.test_more_more (unexpected) --- FAIL: syndication_tests.tests.SyndicationFeedTest.test_rss2_feed (unexpected) --- FAIL: ordering.tests.OrderingTests.test_order_by_fk_attname (unexpected) --- FAIL: db_functions.math.test_power.PowerTests.test_integer (unexpected) --- FAIL: backends.tests.FkConstraintsTests.test_disable_constraint_checks_context_manager (unexpected) --- FAIL: queries.test_bulk_update.BulkUpdateNoteTests.test_multiple_fields (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_int_pk_to_int_unique (unexpected) --- FAIL: db_functions.text.test_sha384.SHA384Tests.test_basic (unexpected) --- FAIL: schema.tests.SchemaTests.test_add_field_temp_default (unexpected) --- FAIL: backends.tests.FkConstraintsTests.test_check_constraints (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_field_reloads_state_on_fk_with_to_field_target_changes (unexpected) --- FAIL: migrations.test_operations.OperationTests.test_alter_field_reloads_state_on_fk_with_to_field_target_type_change (unexpected) --- FAIL: introspection.tests.IntrospectionTests.test_get_table_description_types (unexpected) --- FAIL: schema.tests.SchemaTests.test_char_field_with_db_index_to_fk (unexpected) --- FAIL: db_functions.text.test_sha224.SHA224Tests.test_basic (unexpected) --- FAIL: timezones.tests.LegacyDatabaseTests.test_cursor_execute_accepts_naive_datetime (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_text_field_to_time_field (unexpected) --- FAIL: aggregation.test_filter_argument.FilteredAggregateTests.test_filtered_numerical_aggregates (unexpected) --- FAIL: introspection.tests.IntrospectionTests.test_get_table_description_col_lengths (unexpected) --- FAIL: schema.tests.SchemaTests.test_alter_text_field_to_datetime_field (unexpected) For a full summary look at the django artifacts An updated blocklist (djangoBlocklist20_2) is available in the artifacts' django log ``` <details><summary>More</summary><p> Artifacts: [/django](https://teamcity.cockroachdb.com/viewLog.html?buildId=2071705&tab=artifacts#/django) Related: - #51051 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #51049 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.2](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.2) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #50793 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-provisional_202006292135_v19.2.9](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-provisional_202006292135_v19.2.9) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) - #46800 roachtest: django failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-20.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-20.1) [release-blocker](https://api.github.com/repos/cockroachdb/cockroach/labels/release-blocker) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Adjango.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest django failed on fail introspection tests introspectiontests test table names with views unexpected fail queries tests test only orders with all items having status unexpected fail backends tests fkconstraintstests test disable constraint checks manually unexpected fail aggregation regress tests aggregationtests test stddev unexpected fail expressions tests ftimedeltatests test datetime subtraction microseconds unexpected fail schema tests schematests test alter numeric field keep null status unexpected fail schema tests schematests test rename field in target model unexpected fail schema tests schematests test db table unexpected fail queries tests relatedlookuptypetests test values queryset lookup unexpected fail schema tests schematests test alter unexpected fail migrations test operations operationtests test alter order with respect to unexpected fail schema tests schematests test alter textual field keep null status unexpected fail auth tests test views uuidusertests test admin password change unexpected fail queries tests test ticket unexpected fail queries tests subquerytests test slice subquery and query unexpected fail many to one tests manytoonetests test fk to smallautofield unexpected fail aggregation regress tests aggregationtests test ticket unexpected fail syndication tests tests syndicationfeedtest test latest post date unexpected fail queries test explain explaintests test unknown format unexpected fail db functions text test test transform unexpected fail schema tests schematests test alter not unique field to primary key unexpected fail admin filters tests listfilterstests test simplelistfilter with queryset based lookups unexpected fail migrations test operations operationtests test alter field pk fk unexpected fail syndication tests tests syndicationfeedtest test feed unexpected fail queries tests nullinexcludetest test col not in list containing null unexpected fail migrations test operations operationtests test alter fk non fk unexpected fail queries test bulk update bulkupdatetests test inherited fields unexpected fail aggregation regress tests aggregationtests test more more unexpected fail syndication tests tests syndicationfeedtest test feed unexpected fail ordering tests orderingtests test order by fk attname unexpected fail db functions math test power powertests test integer unexpected fail backends tests fkconstraintstests test disable constraint checks context manager unexpected fail queries test bulk update bulkupdatenotetests test multiple fields unexpected fail schema tests schematests test alter int pk to int unique unexpected fail db functions text test test basic unexpected fail schema tests schematests test add field temp default unexpected fail backends tests fkconstraintstests test check constraints unexpected fail migrations test operations operationtests test alter field reloads state on fk with to field target changes unexpected fail migrations test operations operationtests test alter field reloads state on fk with to field target type change unexpected fail introspection tests introspectiontests test get table description types unexpected fail schema tests schematests test char field with db index to fk unexpected fail db functions text test test basic unexpected fail timezones tests legacydatabasetests test cursor execute accepts naive datetime unexpected fail schema tests schematests test alter text field to time field unexpected fail aggregation test filter argument filteredaggregatetests test filtered numerical aggregates unexpected fail introspection tests introspectiontests test get table description col lengths unexpected fail schema tests schematests test alter text field to datetime field unexpected for a full summary look at the django artifacts an updated blocklist is available in the artifacts django log more artifacts related roachtest django failed roachtest django failed roachtest django failed roachtest django failed powered by
1
149,073
11,881,602,227
IssuesEvent
2020-03-27 13:00:33
emoncms/emoncms
https://api.github.com/repos/emoncms/emoncms
closed
Device Already Exists Error
testing required
https://github.com/emoncms/emoncms/blob/ccb3a41ed17879ff154dfcabf7fa58d010ef8dce/Modules/input/Views/input_view.js#L878 Getting these errors on a regular basis. Could be for any existing node. ![image](https://user-images.githubusercontent.com/1322749/70890566-45f2e080-1fdd-11ea-94ed-69197143d20d.png) It has also been reported on the community. https://community.openenergymonitor.org/t/constant-device-already-exists-error-alert-poup/12714 Whilst clearing REDIS cache does fix it, it does not solve why it is happening. The code has reached this point, because it does not think the device exists, yet, when it checks again, it does exist. Something not right......
1.0
Device Already Exists Error - https://github.com/emoncms/emoncms/blob/ccb3a41ed17879ff154dfcabf7fa58d010ef8dce/Modules/input/Views/input_view.js#L878 Getting these errors on a regular basis. Could be for any existing node. ![image](https://user-images.githubusercontent.com/1322749/70890566-45f2e080-1fdd-11ea-94ed-69197143d20d.png) It has also been reported on the community. https://community.openenergymonitor.org/t/constant-device-already-exists-error-alert-poup/12714 Whilst clearing REDIS cache does fix it, it does not solve why it is happening. The code has reached this point, because it does not think the device exists, yet, when it checks again, it does exist. Something not right......
test
device already exists error getting these errors on a regular basis could be for any existing node it has also been reported on the community whilst clearing redis cache does fix it it does not solve why it is happening the code has reached this point because it does not think the device exists yet when it checks again it does exist something not right
1
433,169
30,316,471,410
IssuesEvent
2023-07-10 15:54:13
supabase/supabase
https://api.github.com/repos/supabase/supabase
closed
Typo in Bangla README
documentation good first issue
# Improve documentation ## Link https://github.com/supabase/supabase/blob/master/i18n/README.bn.md ## Describe the problem ![image](https://github.com/supabase/supabase/assets/115442240/67030a5a-16be-4200-a800-75ee76e9570a) Just follow the last line. The word `রিপুর` will be exchanged with `রেপোর` ## Describe the improvement Change the spelling from `রিপুর` to `রেপোর` ## Additional context ![image](https://github.com/supabase/supabase/assets/115442240/a4e1a7fe-d3dd-41a1-8a64-0f445942daad)
1.0
Typo in Bangla README - # Improve documentation ## Link https://github.com/supabase/supabase/blob/master/i18n/README.bn.md ## Describe the problem ![image](https://github.com/supabase/supabase/assets/115442240/67030a5a-16be-4200-a800-75ee76e9570a) Just follow the last line. The word `রিপুর` will be exchanged with `রেপোর` ## Describe the improvement Change the spelling from `রিপুর` to `রেপোর` ## Additional context ![image](https://github.com/supabase/supabase/assets/115442240/a4e1a7fe-d3dd-41a1-8a64-0f445942daad)
non_test
typo in bangla readme improve documentation link describe the problem just follow the last line the word রিপুর will be exchanged with রেপোর describe the improvement change the spelling from রিপুর to রেপোর additional context
0
281,949
21,315,450,923
IssuesEvent
2022-04-16 07:30:23
kxshxsh/pe
https://api.github.com/repos/kxshxsh/pe
opened
Model component class diagram very complicated and labels tough to understand
severity.Medium type.DocumentationBug
![image.png](https://raw.githubusercontent.com/kxshxsh/pe/main/files/7aae2341-554a-4556-b326-7d5acf6ea42e.png) Extemely messy class diagram. All I see is curved arrows and the words "* filered" and "* all" floating everywhere with no clear indication of which association arrow they belong to or even what purpose they add. I don't know if * is for multiplicity or not, and it is why is it following "all" and "filtered" so closely? <!--session: 1650087226793-f266947a-1da7-4629-8f86-32a293c5b0a1--> <!--Version: Web v3.4.2-->
1.0
Model component class diagram very complicated and labels tough to understand - ![image.png](https://raw.githubusercontent.com/kxshxsh/pe/main/files/7aae2341-554a-4556-b326-7d5acf6ea42e.png) Extemely messy class diagram. All I see is curved arrows and the words "* filered" and "* all" floating everywhere with no clear indication of which association arrow they belong to or even what purpose they add. I don't know if * is for multiplicity or not, and it is why is it following "all" and "filtered" so closely? <!--session: 1650087226793-f266947a-1da7-4629-8f86-32a293c5b0a1--> <!--Version: Web v3.4.2-->
non_test
model component class diagram very complicated and labels tough to understand extemely messy class diagram all i see is curved arrows and the words filered and all floating everywhere with no clear indication of which association arrow they belong to or even what purpose they add i don t know if is for multiplicity or not and it is why is it following all and filtered so closely
0
124,425
10,311,233,361
IssuesEvent
2019-08-29 16:51:40
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Add windows node failed when rancher server url included port
[zube]: To Test area/windows kind/bug-qa team/cn
Rancher version: master-head Step to reproduce: 1. run rancher sever with the command: `docker run -p 8443:443 -p 8080:80 rancher/rancher:master-head` 2. enable windows 3. add windows node 4. then the agent container in windows got errors ``` INFO: https://52.197.xxx.xxx:8443 is accessible Copy-Item : The filename, directory name, or volume label syntax is incorrect : 'C:\etc\docker\certs.d\52.197.xxx.xxx:8443\ca.crt' At C:\etc\rancher\execute.ps1:299 char:5 + Copy-Item -Force -Path "$sslCertDir\serverca" -Destination "$dock ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : NotSpecified: (:) [Copy-Item], IOException + FullyQualifiedErrorId : System.IO.IOException,Microsoft.PowerShell.Commands.CopyItemCommand ```
1.0
Add windows node failed when rancher server url included port - Rancher version: master-head Step to reproduce: 1. run rancher sever with the command: `docker run -p 8443:443 -p 8080:80 rancher/rancher:master-head` 2. enable windows 3. add windows node 4. then the agent container in windows got errors ``` INFO: https://52.197.xxx.xxx:8443 is accessible Copy-Item : The filename, directory name, or volume label syntax is incorrect : 'C:\etc\docker\certs.d\52.197.xxx.xxx:8443\ca.crt' At C:\etc\rancher\execute.ps1:299 char:5 + Copy-Item -Force -Path "$sslCertDir\serverca" -Destination "$dock ... + ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ + CategoryInfo : NotSpecified: (:) [Copy-Item], IOException + FullyQualifiedErrorId : System.IO.IOException,Microsoft.PowerShell.Commands.CopyItemCommand ```
test
add windows node failed when rancher server url included port rancher version master head step to reproduce run rancher sever with the command docker run p p rancher rancher master head enable windows add windows node then the agent container in windows got errors info is accessible copy item the filename directory name or volume label syntax is incorrect c etc docker certs d xxx xxx ca crt at c etc rancher execute char copy item force path sslcertdir serverca destination dock categoryinfo notspecified ioexception fullyqualifiederrorid system io ioexception microsoft powershell commands copyitemcommand
1
402,510
27,371,859,849
IssuesEvent
2023-02-28 00:34:29
KA-devl/VueLovers-soen341project2023
https://api.github.com/repos/KA-devl/VueLovers-soen341project2023
closed
Employer Static Dashboard
documentation new feature Points: 2
As an employer, I should have a dashboard where I can see the applicants to my job postings
1.0
Employer Static Dashboard - As an employer, I should have a dashboard where I can see the applicants to my job postings
non_test
employer static dashboard as an employer i should have a dashboard where i can see the applicants to my job postings
0
274,022
23,803,865,772
IssuesEvent
2022-09-03 18:28:33
doctrine/dbal
https://api.github.com/repos/doctrine/dbal
closed
DBAL-1168: Schema's getMigrateFromSql always adds CREATE SCHEMA
Missing Tests Bug PostgreSQL
Jira issue originally created by user vbence: I originally posted this to Migrations; noticing that all the generated down() methods start with a "CREATE SCHEMA public" line. Inspecting the return from Schema#getMigrateFromSql it indeed contains the create statement.
1.0
DBAL-1168: Schema's getMigrateFromSql always adds CREATE SCHEMA - Jira issue originally created by user vbence: I originally posted this to Migrations; noticing that all the generated down() methods start with a "CREATE SCHEMA public" line. Inspecting the return from Schema#getMigrateFromSql it indeed contains the create statement.
test
dbal schema s getmigratefromsql always adds create schema jira issue originally created by user vbence i originally posted this to migrations noticing that all the generated down methods start with a create schema public line inspecting the return from schema getmigratefromsql it indeed contains the create statement
1
87,343
10,895,373,525
IssuesEvent
2019-11-19 10:31:33
betagouv/demarches-simplifiees.fr
https://api.github.com/repos/betagouv/demarches-simplifiees.fr
closed
ETQ Usager, je veux voir le logo de la démarche correctement sur ie11
a-communiquer contributions acceptées design usager
`hs #15254` Version correcte : ![dm-firefox](https://user-images.githubusercontent.com/907405/56216514-dac5c880-6061-11e9-8938-93907a65dfe7.png) Version IE11: ![dm-ie-11](https://user-images.githubusercontent.com/907405/56216553-eadda800-6061-11e9-9812-5f18385e48f5.png)
1.0
ETQ Usager, je veux voir le logo de la démarche correctement sur ie11 - `hs #15254` Version correcte : ![dm-firefox](https://user-images.githubusercontent.com/907405/56216514-dac5c880-6061-11e9-8938-93907a65dfe7.png) Version IE11: ![dm-ie-11](https://user-images.githubusercontent.com/907405/56216553-eadda800-6061-11e9-9812-5f18385e48f5.png)
non_test
etq usager je veux voir le logo de la démarche correctement sur hs version correcte version
0
201,628
15,216,117,944
IssuesEvent
2021-02-17 15:09:36
Princeton-CDH/geniza
https://api.github.com/repos/Princeton-CDH/geniza
reopened
As a global admin, I want a one-time import of the list of all languages and scripts, and their correlation, so that I can access, display and manage the information in the database.
🆕 enhancement 🗜️ awaiting testing
## testing notes - Confirm that the list of languages + scripts has been correctly imported from the ontology spreadsheet - Check that the edit history for a few records to confirm that there is a log entry showing the record was imported by script
1.0
As a global admin, I want a one-time import of the list of all languages and scripts, and their correlation, so that I can access, display and manage the information in the database. - ## testing notes - Confirm that the list of languages + scripts has been correctly imported from the ontology spreadsheet - Check that the edit history for a few records to confirm that there is a log entry showing the record was imported by script
test
as a global admin i want a one time import of the list of all languages and scripts and their correlation so that i can access display and manage the information in the database testing notes confirm that the list of languages scripts has been correctly imported from the ontology spreadsheet check that the edit history for a few records to confirm that there is a log entry showing the record was imported by script
1
246,994
20,954,966,112
IssuesEvent
2022-03-27 01:31:25
datafuselabs/databend
https://api.github.com/repos/datafuselabs/databend
closed
stateless-test `06_0002_show_create_database.sql` sometimes run slow
C-testing
In stateless-test, when running `06_0002_show_create_database.sql`, sometimes stuck. cc @drmingdrmer
1.0
stateless-test `06_0002_show_create_database.sql` sometimes run slow - In stateless-test, when running `06_0002_show_create_database.sql`, sometimes stuck. cc @drmingdrmer
test
stateless test show create database sql sometimes run slow in stateless test when running show create database sql sometimes stuck cc drmingdrmer
1
137,573
5,311,805,337
IssuesEvent
2017-02-13 06:07:34
BuiltBrokenModding/VoltzEngine
https://api.github.com/repos/BuiltBrokenModding/VoltzEngine
opened
Implement property system
enhancement High Priority
To improve the functionality of modifying common data we should switch over to using JSON files. These files will store information such as block name, localization key, hardness, resistance, and recipes. This way we do not need to worry about code implementation when handling each part. Instead, VE will read in the files and inject the data where it is needed. Additionally, this files can be used to register the block and item block without specifically the exacts for each. Example file: `{ "block": { "name":"block", "class":"BlockMeta", "Item":"ItemBlockMeta", "hardness":5, "resistance":5, "localization":"tile.#mod#.block" "recipes": { "1": { "type":"shaped", "output": { "item":"this", "meta":0, "nbt":"none" }, "recipe":"abc-abc-abc", "components": { "a":"apple", "b":"bow:0", "c": { "item":"stone", "meta":2 } } } } }`
1.0
Implement property system - To improve the functionality of modifying common data we should switch over to using JSON files. These files will store information such as block name, localization key, hardness, resistance, and recipes. This way we do not need to worry about code implementation when handling each part. Instead, VE will read in the files and inject the data where it is needed. Additionally, this files can be used to register the block and item block without specifically the exacts for each. Example file: `{ "block": { "name":"block", "class":"BlockMeta", "Item":"ItemBlockMeta", "hardness":5, "resistance":5, "localization":"tile.#mod#.block" "recipes": { "1": { "type":"shaped", "output": { "item":"this", "meta":0, "nbt":"none" }, "recipe":"abc-abc-abc", "components": { "a":"apple", "b":"bow:0", "c": { "item":"stone", "meta":2 } } } } }`
non_test
implement property system to improve the functionality of modifying common data we should switch over to using json files these files will store information such as block name localization key hardness resistance and recipes this way we do not need to worry about code implementation when handling each part instead ve will read in the files and inject the data where it is needed additionally this files can be used to register the block and item block without specifically the exacts for each example file block name block class blockmeta item itemblockmeta hardness resistance localization tile mod block recipes type shaped output item this meta nbt none recipe abc abc abc components a apple b bow c item stone meta
0
129,787
10,586,620,287
IssuesEvent
2019-10-08 20:08:40
eclipse/openj9
https://api.github.com/repos/eclipse/openj9
closed
Implement `TerminatingThreadLocal` as required by JTReg Test Failure: jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java
test excluded test failure
Failure link ------------ Rebuild Link: https://ci.adoptopenjdk.net/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&BUILD_LIST=openjdk&JenkinsFile=openjdk_x86-64_linux&CUSTOM_TARGET=jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java&TARGET=jdk_custom_0 - test category: openjdk - OS/architecture: all - version: 11, 13 - JVM: openj9 only Optional info ------------- - intermittent failure: No - regression or new test: Regression Failure output (captured from console output) --------------------------------------------- ``` 14:54:51 java.lang.AssertionError: Expected terminated values: [42] but got: [] 14:54:51 at TestTerminatingThreadLocal.ttlTest(TestTerminatingThreadLocal.java:86) 14:54:51 at TestTerminatingThreadLocal.ttlTestSet(TestTerminatingThreadLocal.java:48) 14:54:51 at TestTerminatingThreadLocal.main(TestTerminatingThreadLocal.java:41) 14:54:51 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 14:54:51 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 14:54:51 at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 14:54:51 at java.base/java.lang.reflect.Method.invoke(Method.java:566) 14:54:51 at com.sun.javatest.regtest.agent.MainActionHelper$AgentVMRunnable.run(MainActionHelper.java:298) 14:54:51 at java.base/java.lang.Thread.run(Thread.java:831) ``` After further investigation, it would seem that this method https://github.com/ibmruntimes/openj9-openjdk-jdk11/blob/8bf2a8166185765e9c6ddbffb66d1adbee67fd16/test/jdk/jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java#L66 is not being called by any of the test cases, so `terminatedValues` remains empty and causes the test to fail.
2.0
Implement `TerminatingThreadLocal` as required by JTReg Test Failure: jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java - Failure link ------------ Rebuild Link: https://ci.adoptopenjdk.net/job/Grinder/parambuild/?JDK_VERSION=11&JDK_IMPL=openj9&BUILD_LIST=openjdk&JenkinsFile=openjdk_x86-64_linux&CUSTOM_TARGET=jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java&TARGET=jdk_custom_0 - test category: openjdk - OS/architecture: all - version: 11, 13 - JVM: openj9 only Optional info ------------- - intermittent failure: No - regression or new test: Regression Failure output (captured from console output) --------------------------------------------- ``` 14:54:51 java.lang.AssertionError: Expected terminated values: [42] but got: [] 14:54:51 at TestTerminatingThreadLocal.ttlTest(TestTerminatingThreadLocal.java:86) 14:54:51 at TestTerminatingThreadLocal.ttlTestSet(TestTerminatingThreadLocal.java:48) 14:54:51 at TestTerminatingThreadLocal.main(TestTerminatingThreadLocal.java:41) 14:54:51 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke0(Native Method) 14:54:51 at java.base/jdk.internal.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) 14:54:51 at java.base/jdk.internal.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) 14:54:51 at java.base/java.lang.reflect.Method.invoke(Method.java:566) 14:54:51 at com.sun.javatest.regtest.agent.MainActionHelper$AgentVMRunnable.run(MainActionHelper.java:298) 14:54:51 at java.base/java.lang.Thread.run(Thread.java:831) ``` After further investigation, it would seem that this method https://github.com/ibmruntimes/openj9-openjdk-jdk11/blob/8bf2a8166185765e9c6ddbffb66d1adbee67fd16/test/jdk/jdk/internal/misc/TerminatingThreadLocal/TestTerminatingThreadLocal.java#L66 is not being called by any of the test cases, so `terminatedValues` remains empty and causes the test to fail.
test
implement terminatingthreadlocal as required by jtreg test failure jdk internal misc terminatingthreadlocal testterminatingthreadlocal java failure link rebuild link test category openjdk os architecture all version jvm only optional info intermittent failure no regression or new test regression failure output captured from console output java lang assertionerror expected terminated values but got at testterminatingthreadlocal ttltest testterminatingthreadlocal java at testterminatingthreadlocal ttltestset testterminatingthreadlocal java at testterminatingthreadlocal main testterminatingthreadlocal java at java base jdk internal reflect nativemethodaccessorimpl native method at java base jdk internal reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at java base jdk internal reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java base java lang reflect method invoke method java at com sun javatest regtest agent mainactionhelper agentvmrunnable run mainactionhelper java at java base java lang thread run thread java after further investigation it would seem that this method is not being called by any of the test cases so terminatedvalues remains empty and causes the test to fail
1
45,332
11,633,933,681
IssuesEvent
2020-02-28 09:21:44
feelpp/feelpp
https://api.github.com/repos/feelpp/feelpp
opened
Support clang-10
module:build type:feature
clang-10 is out and should be supported asap. here is the current issue ``` In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/quickstart/qs_hdg_laplacian.cpp:24: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feel.hpp:58: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/filters.hpp:30: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/savegmshmesh.hpp:32: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/exportergmsh.hpp:41: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/exporter.hpp:39: /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feeldiscr/timeset.hpp:369:39: error: no viable overloaded '=' M_complex[sanitize(name)] = __s; ~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~~ /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:378:7: note: candidate function not viable: no known conversion from 'const Feel::complex_type' (aka 'const complex<double>') to 'const std::pair<std::complex<double>, bool>' for 1st argument operator=(typename conditional< ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:389:7: note: candidate function not viable: no known conversion from 'const Feel::complex_type' (aka 'const complex<double>') to 'std::pair<std::complex<double>, bool>' for 1st argument operator=(typename conditional< ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:405:2: note: candidate template ignored: could not match 'pair' against 'complex' operator=(const pair<_U1, _U2>& __p) ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:416:2: note: candidate template ignored: could not match 'pair' against 'complex' operator=(pair<_U1, _U2>&& __p) ^ ```
1.0
Support clang-10 - clang-10 is out and should be supported asap. here is the current issue ``` In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/quickstart/qs_hdg_laplacian.cpp:24: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feel.hpp:58: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/filters.hpp:30: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/savegmshmesh.hpp:32: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/exportergmsh.hpp:41: In file included from /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feelfilters/exporter.hpp:39: /home/prudhomm/Devel/feelpp.clean/feelpp/feel/../feel/feeldiscr/timeset.hpp:369:39: error: no viable overloaded '=' M_complex[sanitize(name)] = __s; ~~~~~~~~~~~~~~~~~~~~~~~~~ ^ ~~~ /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:378:7: note: candidate function not viable: no known conversion from 'const Feel::complex_type' (aka 'const complex<double>') to 'const std::pair<std::complex<double>, bool>' for 1st argument operator=(typename conditional< ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:389:7: note: candidate function not viable: no known conversion from 'const Feel::complex_type' (aka 'const complex<double>') to 'std::pair<std::complex<double>, bool>' for 1st argument operator=(typename conditional< ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:405:2: note: candidate template ignored: could not match 'pair' against 'complex' operator=(const pair<_U1, _U2>& __p) ^ /usr/bin/../lib/gcc/x86_64-linux-gnu/9/../../../../include/c++/9/bits/stl_pair.h:416:2: note: candidate template ignored: could not match 'pair' against 'complex' operator=(pair<_U1, _U2>&& __p) ^ ```
non_test
support clang clang is out and should be supported asap here is the current issue in file included from home prudhomm devel feelpp clean feelpp quickstart qs hdg laplacian cpp in file included from home prudhomm devel feelpp clean feelpp feel feel feel hpp in file included from home prudhomm devel feelpp clean feelpp feel feel feelfilters filters hpp in file included from home prudhomm devel feelpp clean feelpp feel feel feelfilters savegmshmesh hpp in file included from home prudhomm devel feelpp clean feelpp feel feel feelfilters exportergmsh hpp in file included from home prudhomm devel feelpp clean feelpp feel feel feelfilters exporter hpp home prudhomm devel feelpp clean feelpp feel feel feeldiscr timeset hpp error no viable overloaded m complex s usr bin lib gcc linux gnu include c bits stl pair h note candidate function not viable no known conversion from const feel complex type aka const complex to const std pair bool for argument operator typename conditional usr bin lib gcc linux gnu include c bits stl pair h note candidate function not viable no known conversion from const feel complex type aka const complex to std pair bool for argument operator typename conditional usr bin lib gcc linux gnu include c bits stl pair h note candidate template ignored could not match pair against complex operator const pair p usr bin lib gcc linux gnu include c bits stl pair h note candidate template ignored could not match pair against complex operator pair p
0
215,349
16,666,971,423
IssuesEvent
2021-06-07 06:04:11
yyLeaves/DS-Project
https://api.github.com/repos/yyLeaves/DS-Project
opened
Tests and Debugs
tests
# Tests | Part | Tester1 | Tester2 | | --------- | ------- | ------- | | Event 1&2 | | | | Event 3 | L | | | Event 4 | | | | Event 5 | | | | Event 6 | | | | Event 0 | | | # TODO tests ## L - [ ] Own part & with GUI part - [ ] event 3 - [ ] - - [ ] - ## N - [ ] Own part & with GUI part - [ ] - - [ ] - - [ ] - ## T - [ ] Own part & with GUI part - [ ] - - [ ] - - [ ] - ## C - [ ] Own part & with GUI part - [ ] - - [ ] - - [ ] -
1.0
Tests and Debugs - # Tests | Part | Tester1 | Tester2 | | --------- | ------- | ------- | | Event 1&2 | | | | Event 3 | L | | | Event 4 | | | | Event 5 | | | | Event 6 | | | | Event 0 | | | # TODO tests ## L - [ ] Own part & with GUI part - [ ] event 3 - [ ] - - [ ] - ## N - [ ] Own part & with GUI part - [ ] - - [ ] - - [ ] - ## T - [ ] Own part & with GUI part - [ ] - - [ ] - - [ ] - ## C - [ ] Own part & with GUI part - [ ] - - [ ] - - [ ] -
test
tests and debugs tests part event event l event event event event todo tests l own part with gui part event n own part with gui part t own part with gui part c own part with gui part
1
57,797
24,230,869,393
IssuesEvent
2022-09-26 18:09:17
hashicorp/terraform-provider-aws
https://api.github.com/repos/hashicorp/terraform-provider-aws
closed
aws_acm_certificate - allow validation from a "validation domain"
enhancement service/acm
AWS terraform resource ```aws_acm_certificate``` Can this please be updated to allow a ```validation domain``` to be included when using email validation. The link below shows the AWS CLI documentation regarding a ```validation domain``` https://docs.aws.amazon.com/cli/latest/reference/acm/resend-validation-email.html Something like this in Terraform. This should allow the validation emails to be sent to @example.com rather than @my.example.com ``` resource "aws_acm_certificate" "aws_acm_certificate" { domain_name = "my.example.com" validation_method = "EMAIL" validation_domain = "example.com" } ````
1.0
aws_acm_certificate - allow validation from a "validation domain" - AWS terraform resource ```aws_acm_certificate``` Can this please be updated to allow a ```validation domain``` to be included when using email validation. The link below shows the AWS CLI documentation regarding a ```validation domain``` https://docs.aws.amazon.com/cli/latest/reference/acm/resend-validation-email.html Something like this in Terraform. This should allow the validation emails to be sent to @example.com rather than @my.example.com ``` resource "aws_acm_certificate" "aws_acm_certificate" { domain_name = "my.example.com" validation_method = "EMAIL" validation_domain = "example.com" } ````
non_test
aws acm certificate allow validation from a validation domain aws terraform resource aws acm certificate can this please be updated to allow a validation domain to be included when using email validation the link below shows the aws cli documentation regarding a validation domain something like this in terraform this should allow the validation emails to be sent to example com rather than my example com resource aws acm certificate aws acm certificate domain name my example com validation method email validation domain example com
0
325,461
27,879,929,864
IssuesEvent
2023-03-21 18:34:24
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
`brave_ads.enabled_last_profile` is `false` when only profile left has ads enabled
bug needs-investigation priority/P3 QA/Yes release-notes/exclude QA/Test-Plan-Specified feature/ads OS/Desktop
Follow up to: https://github.com/brave/brave-browser/issues/26864 This happens after deleting profile that ads were never enabled ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Profile1: Enabled ads 1. Profile2: Do not able ads 1. Run Brave 1. Delete Profile2 using Profile Manager (make sure you don't activate the window of Profile1) 1. Check `brave_ads.enabled_last_profile` in local state ## Actual result: <!--Please add screenshots if needed--> `brave_ads.enabled_last_profile` is `false` Note: the state is fixed when window with Profile1 is activated ## Expected result: `brave_ads.enabled_last_profile` is `true`, because only profile left has ads enabled ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easily reproduced ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.49.91 Chromium: 110.0.5481.77 (Official Build) beta (64-bit) -- | -- Revision | 65ed616c6e8ee3fe0ad64fe83796c020644d42af-refs/branch-heads/5481@{#839} OS | Ubuntu 18.04 LTS cc @tmancey @aseren @DJAndries
1.0
`brave_ads.enabled_last_profile` is `false` when only profile left has ads enabled - Follow up to: https://github.com/brave/brave-browser/issues/26864 This happens after deleting profile that ads were never enabled ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Profile1: Enabled ads 1. Profile2: Do not able ads 1. Run Brave 1. Delete Profile2 using Profile Manager (make sure you don't activate the window of Profile1) 1. Check `brave_ads.enabled_last_profile` in local state ## Actual result: <!--Please add screenshots if needed--> `brave_ads.enabled_last_profile` is `false` Note: the state is fixed when window with Profile1 is activated ## Expected result: `brave_ads.enabled_last_profile` is `true`, because only profile left has ads enabled ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easily reproduced ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> Brave | 1.49.91 Chromium: 110.0.5481.77 (Official Build) beta (64-bit) -- | -- Revision | 65ed616c6e8ee3fe0ad64fe83796c020644d42af-refs/branch-heads/5481@{#839} OS | Ubuntu 18.04 LTS cc @tmancey @aseren @DJAndries
test
brave ads enabled last profile is false when only profile left has ads enabled follow up to this happens after deleting profile that ads were never enabled steps to reproduce enabled ads do not able ads run brave delete using profile manager make sure you don t activate the window of check brave ads enabled last profile in local state actual result brave ads enabled last profile is false note the state is fixed when window with is activated expected result brave ads enabled last profile is true because only profile left has ads enabled reproduces how often easily reproduced brave version brave version info brave chromium official build beta bit revision refs branch heads os ubuntu lts cc tmancey aseren djandries
1
261,267
8,228,924,559
IssuesEvent
2018-09-07 07:39:57
mozilla/addons-frontend
https://api.github.com/repos/mozilla/addons-frontend
opened
Add smooth transitions to the AMInstallButton
component: installation contrib: welcome priority: p3
We should make the `AMInstallButton` look and feel slightly better by improving the transitions when the button switches from the "uninstalled" to "pending" to "installed" states.
1.0
Add smooth transitions to the AMInstallButton - We should make the `AMInstallButton` look and feel slightly better by improving the transitions when the button switches from the "uninstalled" to "pending" to "installed" states.
non_test
add smooth transitions to the aminstallbutton we should make the aminstallbutton look and feel slightly better by improving the transitions when the button switches from the uninstalled to pending to installed states
0
303,724
26,226,239,325
IssuesEvent
2023-01-04 18:58:24
rancher/dashboard
https://api.github.com/repos/rancher/dashboard
closed
[UI] - Search box - Select on Enter
area/navigation [zube]: To Test kind/enhancement good-first-issue
Rancher 2.6.4 - UI On the Search box ![image](https://user-images.githubusercontent.com/2582255/166195898-2e908e71-6cfc-44ff-a3e1-b9317af2dd2e.png) In this sample, when input "data", and press enter : Nothing happends It would be great to add "data" to the filter on Enter press, cause one item match exactly : ![image](https://user-images.githubusercontent.com/2582255/166196148-eccc11a4-1fe9-45c2-9794-d5a614df1614.png) And also clean the filter box to let user input another filter, and so on...
1.0
[UI] - Search box - Select on Enter - Rancher 2.6.4 - UI On the Search box ![image](https://user-images.githubusercontent.com/2582255/166195898-2e908e71-6cfc-44ff-a3e1-b9317af2dd2e.png) In this sample, when input "data", and press enter : Nothing happends It would be great to add "data" to the filter on Enter press, cause one item match exactly : ![image](https://user-images.githubusercontent.com/2582255/166196148-eccc11a4-1fe9-45c2-9794-d5a614df1614.png) And also clean the filter box to let user input another filter, and so on...
test
search box select on enter rancher ui on the search box in this sample when input data and press enter nothing happends it would be great to add data to the filter on enter press cause one item match exactly and also clean the filter box to let user input another filter and so on
1
12,188
3,257,019,449
IssuesEvent
2015-10-20 16:03:53
akvo/akvo-caddisfly
https://api.github.com/repos/akvo/akvo-caddisfly
closed
2.12 Create user interface for strip test
Strip test
The user interface for the strip test should include: 1) an easy way to select the type of test to be done 2) a way to visualise the steps that need to be taken, and where the user is in the current phase 3) realtime feedback on the quality of the image (illumination, saturation, focus) 4) a way to display the test the user is currently doing, as an extra check 5) after the image is captured, a way to display the results including the calibration scale 6) an optional visual 'walk through' that the user can do to see how the process works.
1.0
2.12 Create user interface for strip test - The user interface for the strip test should include: 1) an easy way to select the type of test to be done 2) a way to visualise the steps that need to be taken, and where the user is in the current phase 3) realtime feedback on the quality of the image (illumination, saturation, focus) 4) a way to display the test the user is currently doing, as an extra check 5) after the image is captured, a way to display the results including the calibration scale 6) an optional visual 'walk through' that the user can do to see how the process works.
test
create user interface for strip test the user interface for the strip test should include an easy way to select the type of test to be done a way to visualise the steps that need to be taken and where the user is in the current phase realtime feedback on the quality of the image illumination saturation focus a way to display the test the user is currently doing as an extra check after the image is captured a way to display the results including the calibration scale an optional visual walk through that the user can do to see how the process works
1
21,295
6,133,269,612
IssuesEvent
2017-06-25 12:20:18
TEAMMATES/teammates
https://api.github.com/repos/TEAMMATES/teammates
opened
Use href="javascript:;" to indicate anchor tags without link
a-CodeQuality d.Contributors t-HTML
```html <a href="#">Text</a> ``` The above is commonly used such that `Text` gains the appearance of a link while it does not actually link anywhere. However, this is the wrong method because `#` actually links to the top of the web page. The correct way to suppress such link is with `href="javascript:;"`. P.S. Of course, there might be cases where the intention _is_ to go to the top of the page. This needs to be examined case-by-case.
1.0
Use href="javascript:;" to indicate anchor tags without link - ```html <a href="#">Text</a> ``` The above is commonly used such that `Text` gains the appearance of a link while it does not actually link anywhere. However, this is the wrong method because `#` actually links to the top of the web page. The correct way to suppress such link is with `href="javascript:;"`. P.S. Of course, there might be cases where the intention _is_ to go to the top of the page. This needs to be examined case-by-case.
non_test
use href javascript to indicate anchor tags without link html text the above is commonly used such that text gains the appearance of a link while it does not actually link anywhere however this is the wrong method because actually links to the top of the web page the correct way to suppress such link is with href javascript p s of course there might be cases where the intention is to go to the top of the page this needs to be examined case by case
0
328,049
28,099,993,966
IssuesEvent
2023-03-30 18:41:03
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
closed
DISABLED test_variant_consistency_jit_expand_cuda_complex64 (__main__.TestJitCUDA)
triaged module: flaky-tests skipped module: unknown
Platforms: win, windows This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_variant_consistency_jit_expand_cuda_complex64&suite=TestJitCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/12024506828). Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_variant_consistency_jit_expand_cuda_complex64` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_ops_jit.py`
1.0
DISABLED test_variant_consistency_jit_expand_cuda_complex64 (__main__.TestJitCUDA) - Platforms: win, windows This test was disabled because it is failing in CI. See [recent examples](https://hud.pytorch.org/flakytest?name=test_variant_consistency_jit_expand_cuda_complex64&suite=TestJitCUDA) and the most recent trunk [workflow logs](https://github.com/pytorch/pytorch/runs/12024506828). Over the past 3 hours, it has been determined flaky in 3 workflow(s) with 3 failures and 3 successes. **Debugging instructions (after clicking on the recent samples link):** DO NOT ASSUME THINGS ARE OKAY IF THE CI IS GREEN. We now shield flaky tests from developers so CI will thus be green but it will be harder to parse the logs. To find relevant log snippets: 1. Click on the workflow logs linked above 2. Click on the Test step of the job so that it is expanded. Otherwise, the grepping will not work. 3. Grep for `test_variant_consistency_jit_expand_cuda_complex64` 4. There should be several instances run (as flaky tests are rerun in CI) from which you can study the logs. Test file path: `test_ops_jit.py`
test
disabled test variant consistency jit expand cuda main testjitcuda platforms win windows this test was disabled because it is failing in ci see and the most recent trunk over the past hours it has been determined flaky in workflow s with failures and successes debugging instructions after clicking on the recent samples link do not assume things are okay if the ci is green we now shield flaky tests from developers so ci will thus be green but it will be harder to parse the logs to find relevant log snippets click on the workflow logs linked above click on the test step of the job so that it is expanded otherwise the grepping will not work grep for test variant consistency jit expand cuda there should be several instances run as flaky tests are rerun in ci from which you can study the logs test file path test ops jit py
1
347,683
31,240,814,047
IssuesEvent
2023-08-20 21:02:31
MohistMC/Mohist
https://api.github.com/repos/MohistMC/Mohist
closed
[1.20.1] 在更新实体速度时的Minecraft世界计时崩溃
Wait Needs Testing 1.12.2
<!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.--> <!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.--> <!-- If you don't know where to upload your logs and crash reports, you can use these websites : --> <!-- https://gist.github.com (recommended) --> <!-- https://mclo.gs --> <!-- https://haste.mohistmc.com --> <!-- https://pastebin.com --> <!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT --> **Minecraft Version :** 1.20.1 **Mohist Version :** 357 **Operating System :** win10 **Concerned mod / plugin** : Plugins (60): Law, PlaceholderAPI, VoidWorld, CustomPlaceholder, Yum, AntiLoop, BBSToper, Ip2region, NBTAPI, MiniMOTD, LuckPerms, DreamCleaner, CrazyAuctionsPlus, PlayerChat, CMILib, ItemNBTEdit, LoneLibs, AlwaysKeepInventory, ProtocolLib, Notbuild, LiteSignIn, Snow-BetterRTP, MiaoBoard, Vault, WTF, SbPlaceholder, BKCommonLib, PlayerPoints, EconomicBalance, ItemCommand, PlayerTitle, DemonMarket, MyCommand, WorldEdit, PlayerKits, DecentHolograms, Multiverse-Core, AuthMe, SakuraPurchasePlugin, CMI, AllMusic, CMIEInjector, CoreProtect, MysqlEcoBridge, Towny, WorldBorder, PocketGames, Residence, ItemsAdder, LockettePro, BanItem, TownyGUI, PlayerGuild, PlayerTop, TrMenu, YeqiLib, DeluxeMenus, PlugManX (PlugMan), Mohist, Mohist 模组:[14:28:55 INFO]: Mod List: • mods_folder [前置][神秘主义][聪明大脑库]SmartBrainLib-fabric-1.20-1.11.jar : smartbrainlib (1.11) - 1 • mods_folder [前置][动态联合]CyclopsCore-1.20.1-1.18.8.jar : cyclopscore (1.18.8) - 1 • mods_folder [天空村庄] SkyVillages-1.0.2-1.20.1-forge-release.jar : skyvillages (1.0.2-1.20.1-forge) - 1 • mods_folder [前置][炼金化学]alchemylib-1.20.1-1.0.29.jar : alchemylib (1.0.29) - 1 • mods_folder [前置][合成扩展]Cucumber-1.20.1-7.0.0.jar : cucumber (7.0.0) - 1 • mods_folder [前置]geckolib-forge-1.20.1-4.2.1.jar : geckolib (4.2.1) - 1 • mods_folder [前置][铁魔法]player-animation-lib-forge-1.0.2-rc1+1.20.jar : playeranimator (1.0.2-rc1+1.20) - 1 • mods_folder [合成扩展]ExtendedCrafting-1.20.1-6.0.2.jar : extendedcrafting (6.0.2) - 1 • mods_folder [更多实用工具]scalar_utils-1.20.1-5.4.10.jar : scalar_utils (1.20.1-5.4.10) - 1 • mods_folder [网络发包修复]connectivity-1.20.1-4.4.jar : connectivity (1.20.1-4.4) - 1 • mods_folder [是,史蒂夫模型] yesstevemodel-1.20-release-1.1.5-hotfix-2.jar : yes_steve_model (1.1.5-hotfix-2) - 1 • mods_folder [前置][铁魔法]caelus-forge-3.1.0+1.20.jar : caelus (3.1.0+1.20) - 1 • mods_folder [更多结构]Structory_1.20.1_v1.3.2.jar : structory (1.3.2) - 1 • mods_folder [懒人厨房] cookingforblockheads-forge-1.20-16.0.0.jar : cookingforblockheads (16.0.0) - 1 • mods_folder [现代化修复] modernfix-forge-5.6.0+mc1.20.1.jar : modernfix (5.6.0+mc1.20.1) - 1 • mods_folder [JEI物品管理器]jei-1.20.1-forge-15.2.0.25.jar : jei (15.2.0.25) - 1 • mods_folder [服务端][前置]Placebo-1.20.1-8.3.2.jar : placebo (8.3.2) - 1 • mods_folder [自然罗盘/生物群系指南针] NaturesCompass-1.20.1-1.11.1-forge.jar : naturescompass (1.0.0) - 1 • mods_folder [大力能源]Powah-5.0.1.jar : powah (5.0.1) - 1 • JarInJar mixinextras-forge-0.2.0-beta.9.jar : mixinextras (0.2.0-beta.9) - 1 • mods_folder [前置][柠檬核心]lemon_core-1.20.1-1.0.21.jar : lemon_core (1.20.1-1.0.21) - 1 • mods_folder [魔法使]mahoutsukai-1.20.1-v1.34.53.jar : mahoutsukai (1.20.1-v1.34.53) - 1 • mods_folder [前置][懒人厨房]balm-forge-1.20-7.0.4.jar : balm (7.0.4) - 1 • mods_folder [合成表]CraftTweaker-forge-1.20.1-14.0.12.jar : crafttweaker (14.0.12) - 1 • mods_folder [更多磁盘] ExtraDisks-1.20.1-3.0.0.jar : extradisks (1.20.1-3.0.0) - 1 • mods_folder [原神仪器]genshinstrument-1.20.1-3.1.jar : genshinstrument (3.1) - 1 • mods_folder [前置]cloth-config-11.0.99-forge.jar : cloth_config (11.0.99) - 1 • minecraft forge-1.20.1-47.1.70-universal.jar : forge (47.1.70) - 2 • mods_folder [精致存储] refinedstorage-1.12.2.jar : refinedstorage (1.12.2) - 1 • mods_folder [服务端][机械动力扳手修复]windy-fixes-0.1.0.jar : windy_fixes (0.1.0) - 1 • mods_folder [更多箱子]ironchest-1.20.1-14.4.4.jar : ironchest (1.20.1-14.4.4) - 1 • mods_folder [炼金化学]alchemistry-1.20.1-2.3.3.jar : alchemistry (2.3.3) - 1 • mods_folder [动态联合][集成动力学]IntegratedCrafting-1.20.1-1.1.4.jar : integratedcrafting (1.1.4) - 1 • minecraft server-1.20.1-20230612.114412-srg.jar : minecraft (1.20.1) - 1 • mods_folder [农夫乐事] FarmersDelight-1.20.1-1.2.2a.jar : farmersdelight (1.20.1-1.2.2a) - 1 • mods_folder [农夫乐事][末地乐事] ends_delight-1.20.1-1.0.jar : ends_delight (1.0) - 1 • mods_folder [前置][集成动力]CommonCapabilities-1.20.1-2.9.0.jar : commoncapabilities (2.9.0) - 1 • mods_folder [前置][生物群系]TerraBlender-forge-1.20.1-3.0.0.169.jar : terrablender (3.0.0.169) - 1 • mods_folder [汉堡包]burgermod-2.9.4-1.20.1.jar : burgermod (2.9.4) - 1 • mods_folder [前置][妖怪之山通用库] mysterious_mountain_lib-1.2.7-1.20.jar : mysterious_mountain_lib (1.2.7-1.20) - 1 • mods_folder [更多生物群系]BiomesOPlenty-1.20.1-18.0.0.592.jar : biomesoplenty (18.0.0.592) - 1 • mods_folder [前置][创造核心]CreativeCore_FORGE_v2.10.31_mc1.20.1.jar : creativecore (2.10.31) - 1 • JarInJar spectrelib-forge-0.13.13+1.20.1.jar : spectrelib (0.13.13+1.20.1) - 1 • mods_folder [前置]supermartijn642corelib-1.1.9-forge-mc1.20.jar : supermartijn642corelib (1.1.9) - 1 • mods_folder [炼金术]theurgy-1.20.1-1.3.10.jar : theurgy (1.3.10) - 1 • mods_folder [植物魔法]Botania-1.20.1-441-FORGE-SNAPSHOT.jar : botania (1.20.1-441-FORGE-SNAPSHOT) - 1 • mods_folder [倒置火把] ceilingtorch-1.20-1.25.jar : ceilingtorch (1.25) - 1 • mods_folder [末影接口]EnderIO-1.20.1-6.0.14-alpha.jar : enderio (6.0.14-alpha) - 1 • mods_folder [服务端]spark-1.10.43-forge.jar : spark (1.10.43) - 1 • mods_folder [前置][饰品栏]curios-forge-5.2.0-beta.3+1.20.1.jar : curios (5.2.0-beta.3+1.20.1) - 1 • mods_folder [通用机械]Mekanism-1.20.1-10.3.9.homebaked-all.jar : mekanism (10.3.9) - 5 • mods_folder [赛🐎娘]umapyoi-0.2.15-1.20.jar : umapyoi (0.2.15-1.20) - 1 • mods_folder [管道]pipez-1.20.1-1.1.4.jar : pipez (1.20.1-1.1.4) - 1 • JarInJar flywheel-forge-1.20-0.6.9-4.jar : flywheel (0.6.9-4) - 1 • mods_folder [机械动力] create-1.20.1-0.5.1.d.jar : create (0.5.1.d) - 1 • mods_folder [前置][帕特抽手册]Patchouli-1.20.1-82-FORGE-SNAPSHOT.jar : patchouli (1.20.1-82-FORGE-SNAPSHOT) - 1 • mods_folder [动态联合]IntegratedDynamics-1.20.1-1.16.10.jar : integrateddynamics (1.16.10) - 2 • mods_folder [前置][化学]chemlib-1.20.1-2.0.17.jar : chemlib (2.0.17) - 1 • mods_folder [多态合成] polymorph-forge-0.49.0+1.20.1.jar : polymorph (0.49.0+1.20.1) - 1 • mods_folder [通量网络] FluxNetworks-1.20.1-7.2.0.14.jar : fluxnetworks (7.2.0.14) - 1 • mods_folder [铁魔法]irons_spellbooks-1.20.1-1.2.0.jar : irons_spellbooks (1.20.1-1.2.0) - 1 • mods_folder [前置][手册]modonomicon-1.20.1-forge-1.38.5.jar : modonomicon (1.38.5) - 1 • mods_folder [飞车奇匠] automobility-0.4.2+1.20.1-forge.jar : automobility (0.4.2+1.20.1-forge) - 1 • mods_folder [前置]architectury-9.0.8-forge.jar : architectury (9.0.8) - 1 • mods_folder [前置][葡园酒香]doapi-1.2.6-forge.jar : doapi (1.2.6) - 2 • mods_folder [葡园酒香] letsdo-vinery-forge-1.4.2.jar : vinery (1.4.2) - 1 • mods_folder [方块小镇]yuushya-1.20.1-forge-2.0.0.jar : yuushya (2.0.0) - 1 • mods_folder [农夫乐事][海洋乐事]oceansdelight-1.0.2-1.20.jar : oceansdelight (1.0.2-1.20) - 1 • mods_folder [铁氧体磁芯] ferritecore-6.0.0-forge.jar : ferritecore (6.0.0) - 1 • mods_folder [神秘主义]occultism-1.20.1-1.81.6.jar : occultism (1.81.6) - 1 • mods_folder [卡哇伊]KawaiiDishes_1.10.1-1.20.1.jar : kawaiidishes (1.10.1) - 1 • mods_folder [农夫乐事][玉米乐事] corn_delight-1.0.3-1.20.1.jar : corn_delight (1.0.3-1.20.1) - 1 • mods_folder [前置][循环]flib-1.20.1-0.0.9.jar : flib (0.0.9) - 1 • JarInJar l2library-2.4.14-slim.jar : l2library (2.4.14) - 1 • mods_folder [精致存储附属] refinedstorageaddons-0.10.0.jar : refinedstorageaddons (0.10.0) - 1 • mods_folder [农夫乐事][料理乐事]cuisinedelight-1.1.8.jar : cuisinedelight (1.1.8) - 1 • mods_folder [豆腐工艺]TofuCraftReload-1.20.1-5.4.0.0.jar : tofucraft (1.20.1-5.4.0.0) - 1 • mods_folder [循环]Cyclic-1.20.1-1.12.2.jar : cyclic (1.12.2) - 1 **Logs :** [crash-2023-08-20_14.07.08-server.txt](https://github.com/MohistMC/Mohist/files/12387390/crash-2023-08-20_14.07.08-server.txt) [debug.log](https://github.com/MohistMC/Mohist/files/12387391/debug.log) **Steps to Reproduce :** 1. 正常启动Minecraft服务器 2. 在游戏中进行常规操作 3.服务器在尝试设置某个实体的速度时崩溃 **Description of issue :** 在正常的服务器操作中,服务器因与"Exception ticking world"相关的错误而崩溃。主要的错误是"java.lang.IllegalArgumentException: x not finite",这个错误似乎是在尝试设置某个实体的速度时触发的。
1.0
[1.20.1] 在更新实体速度时的Minecraft世界计时崩溃 - <!-- ISSUE_TEMPLATE_1 -> IMPORTANT: DO NOT DELETE THIS LINE.--> <!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.--> <!-- If you don't know where to upload your logs and crash reports, you can use these websites : --> <!-- https://gist.github.com (recommended) --> <!-- https://mclo.gs --> <!-- https://haste.mohistmc.com --> <!-- https://pastebin.com --> <!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT --> **Minecraft Version :** 1.20.1 **Mohist Version :** 357 **Operating System :** win10 **Concerned mod / plugin** : Plugins (60): Law, PlaceholderAPI, VoidWorld, CustomPlaceholder, Yum, AntiLoop, BBSToper, Ip2region, NBTAPI, MiniMOTD, LuckPerms, DreamCleaner, CrazyAuctionsPlus, PlayerChat, CMILib, ItemNBTEdit, LoneLibs, AlwaysKeepInventory, ProtocolLib, Notbuild, LiteSignIn, Snow-BetterRTP, MiaoBoard, Vault, WTF, SbPlaceholder, BKCommonLib, PlayerPoints, EconomicBalance, ItemCommand, PlayerTitle, DemonMarket, MyCommand, WorldEdit, PlayerKits, DecentHolograms, Multiverse-Core, AuthMe, SakuraPurchasePlugin, CMI, AllMusic, CMIEInjector, CoreProtect, MysqlEcoBridge, Towny, WorldBorder, PocketGames, Residence, ItemsAdder, LockettePro, BanItem, TownyGUI, PlayerGuild, PlayerTop, TrMenu, YeqiLib, DeluxeMenus, PlugManX (PlugMan), Mohist, Mohist 模组:[14:28:55 INFO]: Mod List: • mods_folder [前置][神秘主义][聪明大脑库]SmartBrainLib-fabric-1.20-1.11.jar : smartbrainlib (1.11) - 1 • mods_folder [前置][动态联合]CyclopsCore-1.20.1-1.18.8.jar : cyclopscore (1.18.8) - 1 • mods_folder [天空村庄] SkyVillages-1.0.2-1.20.1-forge-release.jar : skyvillages (1.0.2-1.20.1-forge) - 1 • mods_folder [前置][炼金化学]alchemylib-1.20.1-1.0.29.jar : alchemylib (1.0.29) - 1 • mods_folder [前置][合成扩展]Cucumber-1.20.1-7.0.0.jar : cucumber (7.0.0) - 1 • mods_folder [前置]geckolib-forge-1.20.1-4.2.1.jar : geckolib (4.2.1) - 1 • mods_folder [前置][铁魔法]player-animation-lib-forge-1.0.2-rc1+1.20.jar : playeranimator (1.0.2-rc1+1.20) - 1 • mods_folder [合成扩展]ExtendedCrafting-1.20.1-6.0.2.jar : extendedcrafting (6.0.2) - 1 • mods_folder [更多实用工具]scalar_utils-1.20.1-5.4.10.jar : scalar_utils (1.20.1-5.4.10) - 1 • mods_folder [网络发包修复]connectivity-1.20.1-4.4.jar : connectivity (1.20.1-4.4) - 1 • mods_folder [是,史蒂夫模型] yesstevemodel-1.20-release-1.1.5-hotfix-2.jar : yes_steve_model (1.1.5-hotfix-2) - 1 • mods_folder [前置][铁魔法]caelus-forge-3.1.0+1.20.jar : caelus (3.1.0+1.20) - 1 • mods_folder [更多结构]Structory_1.20.1_v1.3.2.jar : structory (1.3.2) - 1 • mods_folder [懒人厨房] cookingforblockheads-forge-1.20-16.0.0.jar : cookingforblockheads (16.0.0) - 1 • mods_folder [现代化修复] modernfix-forge-5.6.0+mc1.20.1.jar : modernfix (5.6.0+mc1.20.1) - 1 • mods_folder [JEI物品管理器]jei-1.20.1-forge-15.2.0.25.jar : jei (15.2.0.25) - 1 • mods_folder [服务端][前置]Placebo-1.20.1-8.3.2.jar : placebo (8.3.2) - 1 • mods_folder [自然罗盘/生物群系指南针] NaturesCompass-1.20.1-1.11.1-forge.jar : naturescompass (1.0.0) - 1 • mods_folder [大力能源]Powah-5.0.1.jar : powah (5.0.1) - 1 • JarInJar mixinextras-forge-0.2.0-beta.9.jar : mixinextras (0.2.0-beta.9) - 1 • mods_folder [前置][柠檬核心]lemon_core-1.20.1-1.0.21.jar : lemon_core (1.20.1-1.0.21) - 1 • mods_folder [魔法使]mahoutsukai-1.20.1-v1.34.53.jar : mahoutsukai (1.20.1-v1.34.53) - 1 • mods_folder [前置][懒人厨房]balm-forge-1.20-7.0.4.jar : balm (7.0.4) - 1 • mods_folder [合成表]CraftTweaker-forge-1.20.1-14.0.12.jar : crafttweaker (14.0.12) - 1 • mods_folder [更多磁盘] ExtraDisks-1.20.1-3.0.0.jar : extradisks (1.20.1-3.0.0) - 1 • mods_folder [原神仪器]genshinstrument-1.20.1-3.1.jar : genshinstrument (3.1) - 1 • mods_folder [前置]cloth-config-11.0.99-forge.jar : cloth_config (11.0.99) - 1 • minecraft forge-1.20.1-47.1.70-universal.jar : forge (47.1.70) - 2 • mods_folder [精致存储] refinedstorage-1.12.2.jar : refinedstorage (1.12.2) - 1 • mods_folder [服务端][机械动力扳手修复]windy-fixes-0.1.0.jar : windy_fixes (0.1.0) - 1 • mods_folder [更多箱子]ironchest-1.20.1-14.4.4.jar : ironchest (1.20.1-14.4.4) - 1 • mods_folder [炼金化学]alchemistry-1.20.1-2.3.3.jar : alchemistry (2.3.3) - 1 • mods_folder [动态联合][集成动力学]IntegratedCrafting-1.20.1-1.1.4.jar : integratedcrafting (1.1.4) - 1 • minecraft server-1.20.1-20230612.114412-srg.jar : minecraft (1.20.1) - 1 • mods_folder [农夫乐事] FarmersDelight-1.20.1-1.2.2a.jar : farmersdelight (1.20.1-1.2.2a) - 1 • mods_folder [农夫乐事][末地乐事] ends_delight-1.20.1-1.0.jar : ends_delight (1.0) - 1 • mods_folder [前置][集成动力]CommonCapabilities-1.20.1-2.9.0.jar : commoncapabilities (2.9.0) - 1 • mods_folder [前置][生物群系]TerraBlender-forge-1.20.1-3.0.0.169.jar : terrablender (3.0.0.169) - 1 • mods_folder [汉堡包]burgermod-2.9.4-1.20.1.jar : burgermod (2.9.4) - 1 • mods_folder [前置][妖怪之山通用库] mysterious_mountain_lib-1.2.7-1.20.jar : mysterious_mountain_lib (1.2.7-1.20) - 1 • mods_folder [更多生物群系]BiomesOPlenty-1.20.1-18.0.0.592.jar : biomesoplenty (18.0.0.592) - 1 • mods_folder [前置][创造核心]CreativeCore_FORGE_v2.10.31_mc1.20.1.jar : creativecore (2.10.31) - 1 • JarInJar spectrelib-forge-0.13.13+1.20.1.jar : spectrelib (0.13.13+1.20.1) - 1 • mods_folder [前置]supermartijn642corelib-1.1.9-forge-mc1.20.jar : supermartijn642corelib (1.1.9) - 1 • mods_folder [炼金术]theurgy-1.20.1-1.3.10.jar : theurgy (1.3.10) - 1 • mods_folder [植物魔法]Botania-1.20.1-441-FORGE-SNAPSHOT.jar : botania (1.20.1-441-FORGE-SNAPSHOT) - 1 • mods_folder [倒置火把] ceilingtorch-1.20-1.25.jar : ceilingtorch (1.25) - 1 • mods_folder [末影接口]EnderIO-1.20.1-6.0.14-alpha.jar : enderio (6.0.14-alpha) - 1 • mods_folder [服务端]spark-1.10.43-forge.jar : spark (1.10.43) - 1 • mods_folder [前置][饰品栏]curios-forge-5.2.0-beta.3+1.20.1.jar : curios (5.2.0-beta.3+1.20.1) - 1 • mods_folder [通用机械]Mekanism-1.20.1-10.3.9.homebaked-all.jar : mekanism (10.3.9) - 5 • mods_folder [赛🐎娘]umapyoi-0.2.15-1.20.jar : umapyoi (0.2.15-1.20) - 1 • mods_folder [管道]pipez-1.20.1-1.1.4.jar : pipez (1.20.1-1.1.4) - 1 • JarInJar flywheel-forge-1.20-0.6.9-4.jar : flywheel (0.6.9-4) - 1 • mods_folder [机械动力] create-1.20.1-0.5.1.d.jar : create (0.5.1.d) - 1 • mods_folder [前置][帕特抽手册]Patchouli-1.20.1-82-FORGE-SNAPSHOT.jar : patchouli (1.20.1-82-FORGE-SNAPSHOT) - 1 • mods_folder [动态联合]IntegratedDynamics-1.20.1-1.16.10.jar : integrateddynamics (1.16.10) - 2 • mods_folder [前置][化学]chemlib-1.20.1-2.0.17.jar : chemlib (2.0.17) - 1 • mods_folder [多态合成] polymorph-forge-0.49.0+1.20.1.jar : polymorph (0.49.0+1.20.1) - 1 • mods_folder [通量网络] FluxNetworks-1.20.1-7.2.0.14.jar : fluxnetworks (7.2.0.14) - 1 • mods_folder [铁魔法]irons_spellbooks-1.20.1-1.2.0.jar : irons_spellbooks (1.20.1-1.2.0) - 1 • mods_folder [前置][手册]modonomicon-1.20.1-forge-1.38.5.jar : modonomicon (1.38.5) - 1 • mods_folder [飞车奇匠] automobility-0.4.2+1.20.1-forge.jar : automobility (0.4.2+1.20.1-forge) - 1 • mods_folder [前置]architectury-9.0.8-forge.jar : architectury (9.0.8) - 1 • mods_folder [前置][葡园酒香]doapi-1.2.6-forge.jar : doapi (1.2.6) - 2 • mods_folder [葡园酒香] letsdo-vinery-forge-1.4.2.jar : vinery (1.4.2) - 1 • mods_folder [方块小镇]yuushya-1.20.1-forge-2.0.0.jar : yuushya (2.0.0) - 1 • mods_folder [农夫乐事][海洋乐事]oceansdelight-1.0.2-1.20.jar : oceansdelight (1.0.2-1.20) - 1 • mods_folder [铁氧体磁芯] ferritecore-6.0.0-forge.jar : ferritecore (6.0.0) - 1 • mods_folder [神秘主义]occultism-1.20.1-1.81.6.jar : occultism (1.81.6) - 1 • mods_folder [卡哇伊]KawaiiDishes_1.10.1-1.20.1.jar : kawaiidishes (1.10.1) - 1 • mods_folder [农夫乐事][玉米乐事] corn_delight-1.0.3-1.20.1.jar : corn_delight (1.0.3-1.20.1) - 1 • mods_folder [前置][循环]flib-1.20.1-0.0.9.jar : flib (0.0.9) - 1 • JarInJar l2library-2.4.14-slim.jar : l2library (2.4.14) - 1 • mods_folder [精致存储附属] refinedstorageaddons-0.10.0.jar : refinedstorageaddons (0.10.0) - 1 • mods_folder [农夫乐事][料理乐事]cuisinedelight-1.1.8.jar : cuisinedelight (1.1.8) - 1 • mods_folder [豆腐工艺]TofuCraftReload-1.20.1-5.4.0.0.jar : tofucraft (1.20.1-5.4.0.0) - 1 • mods_folder [循环]Cyclic-1.20.1-1.12.2.jar : cyclic (1.12.2) - 1 **Logs :** [crash-2023-08-20_14.07.08-server.txt](https://github.com/MohistMC/Mohist/files/12387390/crash-2023-08-20_14.07.08-server.txt) [debug.log](https://github.com/MohistMC/Mohist/files/12387391/debug.log) **Steps to Reproduce :** 1. 正常启动Minecraft服务器 2. 在游戏中进行常规操作 3.服务器在尝试设置某个实体的速度时崩溃 **Description of issue :** 在正常的服务器操作中,服务器因与"Exception ticking world"相关的错误而崩溃。主要的错误是"java.lang.IllegalArgumentException: x not finite",这个错误似乎是在尝试设置某个实体的速度时触发的。
test
在更新实体速度时的minecraft世界计时崩溃 important do not delete this line minecraft version mohist version operating system concerned mod plugin plugins law placeholderapi voidworld customplaceholder yum antiloop bbstoper nbtapi minimotd luckperms dreamcleaner crazyauctionsplus playerchat cmilib itemnbtedit lonelibs alwayskeepinventory protocollib notbuild litesignin snow betterrtp miaoboard vault wtf sbplaceholder bkcommonlib playerpoints economicbalance itemcommand playertitle demonmarket mycommand worldedit playerkits decentholograms multiverse core authme sakurapurchaseplugin cmi allmusic cmieinjector coreprotect mysqlecobridge towny worldborder pocketgames residence itemsadder lockettepro banitem townygui playerguild playertop trmenu yeqilib deluxemenus plugmanx plugman mohist mohist 模组: mod list • mods folder smartbrainlib fabric jar smartbrainlib • mods folder cyclopscore jar cyclopscore • mods folder skyvillages forge release jar skyvillages forge • mods folder alchemylib jar alchemylib • mods folder cucumber jar cucumber • mods folder geckolib forge jar geckolib • mods folder player animation lib forge jar playeranimator • mods folder extendedcrafting jar extendedcrafting • mods folder scalar utils jar scalar utils • mods folder connectivity jar connectivity • mods folder yesstevemodel release hotfix jar yes steve model hotfix • mods folder caelus forge jar caelus • mods folder structory jar structory • mods folder cookingforblockheads forge jar cookingforblockheads • mods folder modernfix forge jar modernfix • mods folder jei forge jar jei • mods folder placebo jar placebo • mods folder naturescompass forge jar naturescompass • mods folder powah jar powah • jarinjar mixinextras forge beta jar mixinextras beta • mods folder lemon core jar lemon core • mods folder mahoutsukai jar mahoutsukai • mods folder balm forge jar balm • mods folder crafttweaker forge jar crafttweaker • mods folder extradisks jar extradisks • mods folder genshinstrument jar genshinstrument • mods folder cloth config forge jar cloth config • minecraft forge universal jar forge • mods folder refinedstorage jar refinedstorage • mods folder windy fixes jar windy fixes • mods folder ironchest jar ironchest • mods folder alchemistry jar alchemistry • mods folder integratedcrafting jar integratedcrafting • minecraft server srg jar minecraft • mods folder farmersdelight jar farmersdelight • mods folder ends delight jar ends delight • mods folder commoncapabilities jar commoncapabilities • mods folder terrablender forge jar terrablender • mods folder burgermod jar burgermod • mods folder mysterious mountain lib jar mysterious mountain lib • mods folder biomesoplenty jar biomesoplenty • mods folder creativecore forge jar creativecore • jarinjar spectrelib forge jar spectrelib • mods folder forge jar • mods folder theurgy jar theurgy • mods folder botania forge snapshot jar botania forge snapshot • mods folder ceilingtorch jar ceilingtorch • mods folder enderio alpha jar enderio alpha • mods folder spark forge jar spark • mods folder curios forge beta jar curios beta • mods folder mekanism homebaked all jar mekanism • mods folder umapyoi jar umapyoi • mods folder pipez jar pipez • jarinjar flywheel forge jar flywheel • mods folder create d jar create d • mods folder patchouli forge snapshot jar patchouli forge snapshot • mods folder integrateddynamics jar integrateddynamics • mods folder chemlib jar chemlib • mods folder polymorph forge jar polymorph • mods folder fluxnetworks jar fluxnetworks • mods folder irons spellbooks jar irons spellbooks • mods folder modonomicon forge jar modonomicon • mods folder automobility forge jar automobility forge • mods folder architectury forge jar architectury • mods folder doapi forge jar doapi • mods folder letsdo vinery forge jar vinery • mods folder yuushya forge jar yuushya • mods folder oceansdelight jar oceansdelight • mods folder ferritecore forge jar ferritecore • mods folder occultism jar occultism • mods folder kawaiidishes jar kawaiidishes • mods folder corn delight jar corn delight • mods folder flib jar flib • jarinjar slim jar • mods folder refinedstorageaddons jar refinedstorageaddons • mods folder cuisinedelight jar cuisinedelight • mods folder tofucraftreload jar tofucraft • mods folder cyclic jar cyclic logs steps to reproduce 正常启动minecraft服务器 在游戏中进行常规操作 服务器在尝试设置某个实体的速度时崩溃 description of issue 在正常的服务器操作中,服务器因与 exception ticking world 相关的错误而崩溃。主要的错误是 java lang illegalargumentexception x not finite ,这个错误似乎是在尝试设置某个实体的速度时触发的。
1
79,000
7,689,166,164
IssuesEvent
2018-05-17 11:52:31
NativeScript/nativescript-cli
https://api.github.com/repos/NativeScript/nativescript-cli
closed
Do not track local paths in Analytics
bug ready for test
### Did you verify this is a real problem by searching the [NativeScript Forum](http://forum.nativescript.org) and the [other open issues in this repo](https://github.com/NativeScript/nativescript/issues)? Yes ### Tell us about the problem Creating project from local template tracks the path in Google Analytics. The path may contain user information. ### Which platform(s) does your issue occur on? Both ### Please provide the following version numbers that your issue occurs with: - CLI: 4.0.1 - Cross-platform modules: Not applicable - Runtime(s): Not applicable - Plugin(s): Not applicable ### Please tell us how to recreate the issue in as much detail as possible. Use local template (dir or .tgz) and try the following: ``` $ tns usage-reporting enable $ tns create myApp --template <local path> --log trace ``` In the logs, you'll find out that the path is tracked in Google Analytics. ### Is there code involved? If so, please share the minimal amount of code needed to recreate the problem. No
1.0
Do not track local paths in Analytics - ### Did you verify this is a real problem by searching the [NativeScript Forum](http://forum.nativescript.org) and the [other open issues in this repo](https://github.com/NativeScript/nativescript/issues)? Yes ### Tell us about the problem Creating project from local template tracks the path in Google Analytics. The path may contain user information. ### Which platform(s) does your issue occur on? Both ### Please provide the following version numbers that your issue occurs with: - CLI: 4.0.1 - Cross-platform modules: Not applicable - Runtime(s): Not applicable - Plugin(s): Not applicable ### Please tell us how to recreate the issue in as much detail as possible. Use local template (dir or .tgz) and try the following: ``` $ tns usage-reporting enable $ tns create myApp --template <local path> --log trace ``` In the logs, you'll find out that the path is tracked in Google Analytics. ### Is there code involved? If so, please share the minimal amount of code needed to recreate the problem. No
test
do not track local paths in analytics did you verify this is a real problem by searching the and the yes tell us about the problem creating project from local template tracks the path in google analytics the path may contain user information which platform s does your issue occur on both please provide the following version numbers that your issue occurs with cli cross platform modules not applicable runtime s not applicable plugin s not applicable please tell us how to recreate the issue in as much detail as possible use local template dir or tgz and try the following tns usage reporting enable tns create myapp template log trace in the logs you ll find out that the path is tracked in google analytics is there code involved if so please share the minimal amount of code needed to recreate the problem no
1
349,050
31,769,876,918
IssuesEvent
2023-09-12 11:05:00
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] More multi_node.EsqlSpecIT failures
>test-failure Team:QL :Query Languages/ES|QL
### CI Link https://gradle-enterprise.elastic.co/s/kzhx4kcc6e3pw ### Repro line `./gradlew ':x-pack:plugin:esql:qa:server:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT" \ -Dtests.method="test {stats.ByStringAndLongWithAlias}" \ -Dtests.seed=9EEC253CF1741A06 \ -Dtests.locale=sv \ -Dtests.timezone=Africa/Bamako \ -Druntime.java=21` ### Does it reproduce? Didn't try ### Applicable branches main ### Failure history _No response_ ### Failure excerpt ``` 12:02:38 REPRODUCE WITH: ./gradlew ':x-pack:plugin:esql:qa:server:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT" -Dtests.method="test {stats.ByStringAndLong}" -Dtests.seed=9EEC253CF1741A06 -Dtests.locale=sv -Dtests.timezone=Africa/Bamako -Druntime.java=21 12:02:38 12:02:38 org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT > test {stats.ByStringAndLong} FAILED 12:02:38 org.junit.ComparisonFailure: expected:<[2]00000000> but was:<[3]00000000> 12:02:38 at __randomizedtesting.SeedInfo.seed([9EEC253CF1741A06:16B81AE65F8877FE]:0) 12:02:38 at org.junit.Assert.assertEquals(Assert.java:115) 12:02:38 at org.junit.Assert.assertEquals(Assert.java:144) 12:02:38 at org.elasticsearch.xpack.esql.CsvAssert.assertData(CsvAssert.java:208) 12:02:38 at org.elasticsearch.xpack.esql.qa.rest.EsqlSpecTestCase.doTest(EsqlSpecTestCase.java:103) 12:02:38 at org.elasticsearch.xpack.esql.qa.rest.EsqlSpecTestCase.test(EsqlSpecTestCase.java:84) 12:02:38 at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103) ```
1.0
[CI] More multi_node.EsqlSpecIT failures - ### CI Link https://gradle-enterprise.elastic.co/s/kzhx4kcc6e3pw ### Repro line `./gradlew ':x-pack:plugin:esql:qa:server:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT" \ -Dtests.method="test {stats.ByStringAndLongWithAlias}" \ -Dtests.seed=9EEC253CF1741A06 \ -Dtests.locale=sv \ -Dtests.timezone=Africa/Bamako \ -Druntime.java=21` ### Does it reproduce? Didn't try ### Applicable branches main ### Failure history _No response_ ### Failure excerpt ``` 12:02:38 REPRODUCE WITH: ./gradlew ':x-pack:plugin:esql:qa:server:multi-node:javaRestTest' --tests "org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT" -Dtests.method="test {stats.ByStringAndLong}" -Dtests.seed=9EEC253CF1741A06 -Dtests.locale=sv -Dtests.timezone=Africa/Bamako -Druntime.java=21 12:02:38 12:02:38 org.elasticsearch.xpack.esql.qa.multi_node.EsqlSpecIT > test {stats.ByStringAndLong} FAILED 12:02:38 org.junit.ComparisonFailure: expected:<[2]00000000> but was:<[3]00000000> 12:02:38 at __randomizedtesting.SeedInfo.seed([9EEC253CF1741A06:16B81AE65F8877FE]:0) 12:02:38 at org.junit.Assert.assertEquals(Assert.java:115) 12:02:38 at org.junit.Assert.assertEquals(Assert.java:144) 12:02:38 at org.elasticsearch.xpack.esql.CsvAssert.assertData(CsvAssert.java:208) 12:02:38 at org.elasticsearch.xpack.esql.qa.rest.EsqlSpecTestCase.doTest(EsqlSpecTestCase.java:103) 12:02:38 at org.elasticsearch.xpack.esql.qa.rest.EsqlSpecTestCase.test(EsqlSpecTestCase.java:84) 12:02:38 at java.base/jdk.internal.reflect.DirectMethodHandleAccessor.invoke(DirectMethodHandleAccessor.java:103) ```
test
more multi node esqlspecit failures ci link repro line gradlew x pack plugin esql qa server multi node javaresttest tests org elasticsearch xpack esql qa multi node esqlspecit dtests method test stats bystringandlongwithalias dtests seed dtests locale sv dtests timezone africa bamako druntime java does it reproduce didn t try applicable branches main failure history no response failure excerpt reproduce with gradlew x pack plugin esql qa server multi node javaresttest tests org elasticsearch xpack esql qa multi node esqlspecit dtests method test stats bystringandlong dtests seed dtests locale sv dtests timezone africa bamako druntime java org elasticsearch xpack esql qa multi node esqlspecit test stats bystringandlong failed org junit comparisonfailure expected but was at randomizedtesting seedinfo seed at org junit assert assertequals assert java at org junit assert assertequals assert java at org elasticsearch xpack esql csvassert assertdata csvassert java at org elasticsearch xpack esql qa rest esqlspectestcase dotest esqlspectestcase java at org elasticsearch xpack esql qa rest esqlspectestcase test esqlspectestcase java at java base jdk internal reflect directmethodhandleaccessor invoke directmethodhandleaccessor java
1
245,546
20,776,796,842
IssuesEvent
2022-03-16 11:14:49
cosmos/ibc-go
https://api.github.com/repos/cosmos/ibc-go
closed
Support validator set updates in testing package
testing
Currently the testing suite will fail on `UpdateClient` because the testing suite does not call `Endblock` and validator updates are not properly reflected in the updateclient headers. The chain simulation must call EndBlock, test headers must be allowed to have next vals different from vals, and validator updates need to correctly get simulated in the update headers
1.0
Support validator set updates in testing package - Currently the testing suite will fail on `UpdateClient` because the testing suite does not call `Endblock` and validator updates are not properly reflected in the updateclient headers. The chain simulation must call EndBlock, test headers must be allowed to have next vals different from vals, and validator updates need to correctly get simulated in the update headers
test
support validator set updates in testing package currently the testing suite will fail on updateclient because the testing suite does not call endblock and validator updates are not properly reflected in the updateclient headers the chain simulation must call endblock test headers must be allowed to have next vals different from vals and validator updates need to correctly get simulated in the update headers
1
215,922
24,197,962,877
IssuesEvent
2022-09-24 06:11:41
mkevenaar/OctoPrint-Slack
https://api.github.com/repos/mkevenaar/OctoPrint-Slack
closed
WS-2022-0249 (Medium) detected in OctoPrint-1.7.2-py2.py3-none-any.whl - autoclosed
security vulnerability
## WS-2022-0249 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>OctoPrint-1.7.2-py2.py3-none-any.whl</b></p></summary> <p>The snappy web interface for your 3D printer</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/85/f9/643f8de22bc50d11afee9560e0557859f0af7993c79cd4794c79bcd4d8c4/OctoPrint-1.7.2-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/85/f9/643f8de22bc50d11afee9560e0557859f0af7993c79cd4794c79bcd4d8c4/OctoPrint-1.7.2-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt,/tmp/ws-scm/OctoPrint-Slack</p> <p> Dependency Hierarchy: - :x: **OctoPrint-1.7.2-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An open-redirect vulnerability was discovered in In OctoPrint before 1.8.2. The redirect get variable in login page isn't properly checked. Currently, it check if url.scheme and url.netloc are empty using urllib, which does not handle all URL currently. An attacker could redirect a user to a malicious domain. <p>Publish Date: 2022-06-30 <p>URL: <a href=https://github.com/octoprint/octoprint/commit/dabdd40806e3409f134ab8e276f078e2a02d062b>WS-2022-0249</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/7d2e6212-7431-45c6-9f69-e5957b40fb36/">https://huntr.dev/bounties/7d2e6212-7431-45c6-9f69-e5957b40fb36/</a></p> <p>Release Date: 2022-06-30</p> <p>Fix Resolution: OctoPrint - 1.8.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2022-0249 (Medium) detected in OctoPrint-1.7.2-py2.py3-none-any.whl - autoclosed - ## WS-2022-0249 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>OctoPrint-1.7.2-py2.py3-none-any.whl</b></p></summary> <p>The snappy web interface for your 3D printer</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/85/f9/643f8de22bc50d11afee9560e0557859f0af7993c79cd4794c79bcd4d8c4/OctoPrint-1.7.2-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/85/f9/643f8de22bc50d11afee9560e0557859f0af7993c79cd4794c79bcd4d8c4/OctoPrint-1.7.2-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /requirements.txt</p> <p>Path to vulnerable library: /requirements.txt,/tmp/ws-scm/OctoPrint-Slack</p> <p> Dependency Hierarchy: - :x: **OctoPrint-1.7.2-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>develop</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An open-redirect vulnerability was discovered in In OctoPrint before 1.8.2. The redirect get variable in login page isn't properly checked. Currently, it check if url.scheme and url.netloc are empty using urllib, which does not handle all URL currently. An attacker could redirect a user to a malicious domain. <p>Publish Date: 2022-06-30 <p>URL: <a href=https://github.com/octoprint/octoprint/commit/dabdd40806e3409f134ab8e276f078e2a02d062b>WS-2022-0249</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/7d2e6212-7431-45c6-9f69-e5957b40fb36/">https://huntr.dev/bounties/7d2e6212-7431-45c6-9f69-e5957b40fb36/</a></p> <p>Release Date: 2022-06-30</p> <p>Fix Resolution: OctoPrint - 1.8.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws medium detected in octoprint none any whl autoclosed ws medium severity vulnerability vulnerable library octoprint none any whl the snappy web interface for your printer library home page a href path to dependency file requirements txt path to vulnerable library requirements txt tmp ws scm octoprint slack dependency hierarchy x octoprint none any whl vulnerable library found in base branch develop vulnerability details an open redirect vulnerability was discovered in in octoprint before the redirect get variable in login page isn t properly checked currently it check if url scheme and url netloc are empty using urllib which does not handle all url currently an attacker could redirect a user to a malicious domain publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution octoprint step up your open source security game with mend
0
1,438
2,757,003,911
IssuesEvent
2015-04-27 12:27:32
EricssonResearch/openwebrtc
https://api.github.com/repos/EricssonResearch/openwebrtc
closed
Versioning
All Platforms Build Issue enhancement help wanted Non-Build Issue
This issue is not to create a bikeshed about versioning, but I wanted to discuss some of the points of the versioning integration stuff in OpenWebRTC to make sure we have all bases covered. I plan to use [semantic versioning](http://semver.org) for OpenWebRTC releases. Given our internal iterations on the codebase, I think we will start at 0.3.0 (the current 0.1 is arbitrary.) We need to identify all points in the codebase that need to be changed when a version is set. We also need to make sure all the results are correctly versioned. What is needed here?
2.0
Versioning - This issue is not to create a bikeshed about versioning, but I wanted to discuss some of the points of the versioning integration stuff in OpenWebRTC to make sure we have all bases covered. I plan to use [semantic versioning](http://semver.org) for OpenWebRTC releases. Given our internal iterations on the codebase, I think we will start at 0.3.0 (the current 0.1 is arbitrary.) We need to identify all points in the codebase that need to be changed when a version is set. We also need to make sure all the results are correctly versioned. What is needed here?
non_test
versioning this issue is not to create a bikeshed about versioning but i wanted to discuss some of the points of the versioning integration stuff in openwebrtc to make sure we have all bases covered i plan to use for openwebrtc releases given our internal iterations on the codebase i think we will start at the current is arbitrary we need to identify all points in the codebase that need to be changed when a version is set we also need to make sure all the results are correctly versioned what is needed here
0
233,836
17,909,306,291
IssuesEvent
2021-09-09 01:27:40
ignitionrobotics/ign-gazebo
https://api.github.com/repos/ignitionrobotics/ign-gazebo
opened
Document the various ways of controlling a joint
documentation enhancement good first issue
<!-- If you're not sure on the specifics of the feature or would like a broader discussion, please consider posting a proposal to http://community.gazebosim.org/ instead.--> Ignition Gazebo offers many out-of-box controllers with different features and intended use cases, for example: * [JointController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointController.html) * [JointPositionController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointPositionController.html) * [JointTrajectoryController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointTrajectoryController.html) It can be difficult for users to choose what to use. So it would be nice to have a high-level joint control tutorial explaining the features and usage of each of them, and why a user would choose one over the other. ## Desired behavior <!-- Describe the current problem and the feature you want implemented.--> We should have a new tutorial in https://ignitionrobotics.org/api/gazebo/5.1/tutorials.html which summarizes all ways of controlling a joint. ## Alternatives considered <!-- Describe alternate solutions or features you've considered.--> Instead of documenting just joint controllers, we could document all ways of moving things, including links, models, etc. This reminds me of this great Gazebo classic tutorial: [Setting Velocity on Joints and Links](http://gazebosim.org/tutorials?tut=set_velocity&cat=) ## Implementation suggestion <!-- Provide a suggestion on how to implement this feature, which could help us expedite this implementation.--> Add a tutorial here: https://github.com/ignitionrobotics/ign-gazebo/tree/ign-gazebo5/tutorials
1.0
Document the various ways of controlling a joint - <!-- If you're not sure on the specifics of the feature or would like a broader discussion, please consider posting a proposal to http://community.gazebosim.org/ instead.--> Ignition Gazebo offers many out-of-box controllers with different features and intended use cases, for example: * [JointController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointController.html) * [JointPositionController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointPositionController.html) * [JointTrajectoryController](https://ignitionrobotics.org/api/gazebo/5.1/classignition_1_1gazebo_1_1systems_1_1JointTrajectoryController.html) It can be difficult for users to choose what to use. So it would be nice to have a high-level joint control tutorial explaining the features and usage of each of them, and why a user would choose one over the other. ## Desired behavior <!-- Describe the current problem and the feature you want implemented.--> We should have a new tutorial in https://ignitionrobotics.org/api/gazebo/5.1/tutorials.html which summarizes all ways of controlling a joint. ## Alternatives considered <!-- Describe alternate solutions or features you've considered.--> Instead of documenting just joint controllers, we could document all ways of moving things, including links, models, etc. This reminds me of this great Gazebo classic tutorial: [Setting Velocity on Joints and Links](http://gazebosim.org/tutorials?tut=set_velocity&cat=) ## Implementation suggestion <!-- Provide a suggestion on how to implement this feature, which could help us expedite this implementation.--> Add a tutorial here: https://github.com/ignitionrobotics/ign-gazebo/tree/ign-gazebo5/tutorials
non_test
document the various ways of controlling a joint if you re not sure on the specifics of the feature or would like a broader discussion please consider posting a proposal to instead ignition gazebo offers many out of box controllers with different features and intended use cases for example it can be difficult for users to choose what to use so it would be nice to have a high level joint control tutorial explaining the features and usage of each of them and why a user would choose one over the other desired behavior we should have a new tutorial in which summarizes all ways of controlling a joint alternatives considered instead of documenting just joint controllers we could document all ways of moving things including links models etc this reminds me of this great gazebo classic tutorial implementation suggestion provide a suggestion on how to implement this feature which could help us expedite this implementation add a tutorial here
0
231,307
18,759,406,504
IssuesEvent
2021-11-05 14:49:08
pytorch/pytorch
https://api.github.com/repos/pytorch/pytorch
opened
Rollup: ModuleInfo-based testing
module: nn module: testing
This issue tracks the progress of updating module testing from the old `ModuleTest` / `NewModuleTest` / `CriterionTest` structure to a more understandable `ModuleInfo` based framework modeled after `OpInfo`. - [X] Design `ModuleInfo` structure - [ ] Port module test coverage to use `ModuleInfo` - [ ] Test module forward with inputs compared to reference functions - [ ] Test module backward pass with gradcheck / gradgradcheck - [ ] Test with non-contiguous inputs - [ ] Test across devices / dtypes - [ ] Test across memory_formats - [ ] Test module instantiation with device / dtype kwargs - [ ] Test module printing (i.e. `str(m)` / `m.__repr__()`) - [ ] Test module pickling / unpickling - [ ] Test parameter casting and movement across devices (e.g. `m.cuda()`, `m.to()`, `m.half()`, etc.) - [ ] Test module in-place variants (i.e. `inplace=True`) vs. out-of-place variants - [ ] Test CPU / GPU parity - [ ] Test C++ / python parity - [ ] Other consumers of `ModuleTest`, etc. updated to use `ModuleInfo` - [ ] JIT tests use `ModuleInfo` - [ ] FX tests use `ModuleInfo` - [ ] ONNX tests use `ModuleInfo` - [ ] Full coverage for `ModuleInfo` entries across `torch.nn` - [ ] TODO: insert itemized list of modules here
1.0
Rollup: ModuleInfo-based testing - This issue tracks the progress of updating module testing from the old `ModuleTest` / `NewModuleTest` / `CriterionTest` structure to a more understandable `ModuleInfo` based framework modeled after `OpInfo`. - [X] Design `ModuleInfo` structure - [ ] Port module test coverage to use `ModuleInfo` - [ ] Test module forward with inputs compared to reference functions - [ ] Test module backward pass with gradcheck / gradgradcheck - [ ] Test with non-contiguous inputs - [ ] Test across devices / dtypes - [ ] Test across memory_formats - [ ] Test module instantiation with device / dtype kwargs - [ ] Test module printing (i.e. `str(m)` / `m.__repr__()`) - [ ] Test module pickling / unpickling - [ ] Test parameter casting and movement across devices (e.g. `m.cuda()`, `m.to()`, `m.half()`, etc.) - [ ] Test module in-place variants (i.e. `inplace=True`) vs. out-of-place variants - [ ] Test CPU / GPU parity - [ ] Test C++ / python parity - [ ] Other consumers of `ModuleTest`, etc. updated to use `ModuleInfo` - [ ] JIT tests use `ModuleInfo` - [ ] FX tests use `ModuleInfo` - [ ] ONNX tests use `ModuleInfo` - [ ] Full coverage for `ModuleInfo` entries across `torch.nn` - [ ] TODO: insert itemized list of modules here
test
rollup moduleinfo based testing this issue tracks the progress of updating module testing from the old moduletest newmoduletest criteriontest structure to a more understandable moduleinfo based framework modeled after opinfo design moduleinfo structure port module test coverage to use moduleinfo test module forward with inputs compared to reference functions test module backward pass with gradcheck gradgradcheck test with non contiguous inputs test across devices dtypes test across memory formats test module instantiation with device dtype kwargs test module printing i e str m m repr test module pickling unpickling test parameter casting and movement across devices e g m cuda m to m half etc test module in place variants i e inplace true vs out of place variants test cpu gpu parity test c python parity other consumers of moduletest etc updated to use moduleinfo jit tests use moduleinfo fx tests use moduleinfo onnx tests use moduleinfo full coverage for moduleinfo entries across torch nn todo insert itemized list of modules here
1
2,847
2,533,570,695
IssuesEvent
2015-01-24 01:00:07
golang/go
https://api.github.com/repos/golang/go
closed
cmd/gc: clang 3.3 complains about undefined behavior
accepted priority-later release-none repo-main
<pre>What steps will reproduce the problem? 1. Download clang 3.3 or later, <a href="http://llvm.org/releases/download.html#3.3">http://llvm.org/releases/download.html#3.3</a> 2. export CC=clang 3. apply CL <a href="https://golang.org/cl/10488044">https://golang.org/cl/10488044</a> 4. ./make.bash What is the expected output? What do you see instead? Expected, clean build Actual, quite a few warnings, some trivial, others not. <a href="https://gist.github.com/davecheney/5847237">https://gist.github.com/davecheney/5847237</a> Please use labels and text to provide additional information. See also the discussion here <a href="https://groups.google.com/forum/#">https://groups.google.com/forum/#</a>!topic/golang-dev/SlTKGOfHjtU</pre>
1.0
cmd/gc: clang 3.3 complains about undefined behavior - <pre>What steps will reproduce the problem? 1. Download clang 3.3 or later, <a href="http://llvm.org/releases/download.html#3.3">http://llvm.org/releases/download.html#3.3</a> 2. export CC=clang 3. apply CL <a href="https://golang.org/cl/10488044">https://golang.org/cl/10488044</a> 4. ./make.bash What is the expected output? What do you see instead? Expected, clean build Actual, quite a few warnings, some trivial, others not. <a href="https://gist.github.com/davecheney/5847237">https://gist.github.com/davecheney/5847237</a> Please use labels and text to provide additional information. See also the discussion here <a href="https://groups.google.com/forum/#">https://groups.google.com/forum/#</a>!topic/golang-dev/SlTKGOfHjtU</pre>
non_test
cmd gc clang complains about undefined behavior what steps will reproduce the problem download clang or later a href export cc clang apply cl a href make bash what is the expected output what do you see instead expected clean build actual quite a few warnings some trivial others not a href please use labels and text to provide additional information see also the discussion here a href
0
18,565
6,623,229,979
IssuesEvent
2017-09-22 05:54:53
commonality/generator-community
https://api.github.com/repos/commonality/generator-community
closed
feat(product-manifest): capture essential software configuration information
Priority: Medium Status: In Progress Type: Build Type: CI Type: Feature
💡 __TIP:__ Select the "Preview" Tab to help read these instructions. ## 1. Issue type - [x] __Feature:__ I'm requesting an enhancement. ## 2. User story summary > Describe what you want to accomplish and in what role/capacity, and why it's important to you. As a product maintainer, I need to read and write to a product manifest In order to manage the configuration of my product. ## 3. Acceptance criteria > ℹ️ Initial MVPs will serialize software configurations with Node.js product manifests, i.e., the `package.json` file. If a `package.json` exists, `generator-community` will only ever __*add*__ information: `generator-community` should never replace or overwrite existing data. > > 📡 Future releases should detect a software product's manifest by language and dependency manager, e.g., Ruby: Gemfile; Java (Gradle): build.gradle; Python: MANIFEST; etc. - [x] 1. A `package.json` is created if one doesn't exist already. - [x] 2. The manifest declares a canonical name for the software product. - [x] 3. The manifest provides a brief description of the product in order to communicate purpose and value. - [x] 4. The manifest provides the product's [semantic version](http://semver.org). - [x] 5. The manifest provides the primary product author's name, contact url, and contact email. - [x] 6. The manifest provides an issue tracking URL in case consumers need to report a defect. - [x] 7. The manifest lists all third-party production dependencies and their semantic versions. - [x] 8. The manifest lists all third-party development dependencies and their semantic versions. - [x] 9. The manifest provides the software product's homepage URL. - [x] 10. The manifest provides a list of keywords that can be used to help classify the software product. - [x] 11. The manifest provides the URI to the product's primary source code repository. - [x] 12. The manifest declares the product's open source license. ## 4. Reference material * [`generator-node's app` generator](https://github.com/yeoman/generator-node/blob/ca86c075176d029cf0174b31187b2e0a7b32eaa4/generators/app/index.js#L199) has good sample code that reads from and writes to a product's `package.json` file. * [NPM's package.json documentation](https://docs.npmjs.com/files/package.json) describes every property in a Node.js product manifest (the package.json file).
1.0
feat(product-manifest): capture essential software configuration information - 💡 __TIP:__ Select the "Preview" Tab to help read these instructions. ## 1. Issue type - [x] __Feature:__ I'm requesting an enhancement. ## 2. User story summary > Describe what you want to accomplish and in what role/capacity, and why it's important to you. As a product maintainer, I need to read and write to a product manifest In order to manage the configuration of my product. ## 3. Acceptance criteria > ℹ️ Initial MVPs will serialize software configurations with Node.js product manifests, i.e., the `package.json` file. If a `package.json` exists, `generator-community` will only ever __*add*__ information: `generator-community` should never replace or overwrite existing data. > > 📡 Future releases should detect a software product's manifest by language and dependency manager, e.g., Ruby: Gemfile; Java (Gradle): build.gradle; Python: MANIFEST; etc. - [x] 1. A `package.json` is created if one doesn't exist already. - [x] 2. The manifest declares a canonical name for the software product. - [x] 3. The manifest provides a brief description of the product in order to communicate purpose and value. - [x] 4. The manifest provides the product's [semantic version](http://semver.org). - [x] 5. The manifest provides the primary product author's name, contact url, and contact email. - [x] 6. The manifest provides an issue tracking URL in case consumers need to report a defect. - [x] 7. The manifest lists all third-party production dependencies and their semantic versions. - [x] 8. The manifest lists all third-party development dependencies and their semantic versions. - [x] 9. The manifest provides the software product's homepage URL. - [x] 10. The manifest provides a list of keywords that can be used to help classify the software product. - [x] 11. The manifest provides the URI to the product's primary source code repository. - [x] 12. The manifest declares the product's open source license. ## 4. Reference material * [`generator-node's app` generator](https://github.com/yeoman/generator-node/blob/ca86c075176d029cf0174b31187b2e0a7b32eaa4/generators/app/index.js#L199) has good sample code that reads from and writes to a product's `package.json` file. * [NPM's package.json documentation](https://docs.npmjs.com/files/package.json) describes every property in a Node.js product manifest (the package.json file).
non_test
feat product manifest capture essential software configuration information 💡 tip select the preview tab to help read these instructions issue type feature i m requesting an enhancement user story summary describe what you want to accomplish and in what role capacity and why it s important to you as a product maintainer i need to read and write to a product manifest in order to manage the configuration of my product acceptance criteria ℹ️ initial mvps will serialize software configurations with node js product manifests i e the package json file if a package json exists generator community will only ever add information generator community should never replace or overwrite existing data 📡 future releases should detect a software product s manifest by language and dependency manager e g ruby gemfile java gradle build gradle python manifest etc a package json is created if one doesn t exist already the manifest declares a canonical name for the software product the manifest provides a brief description of the product in order to communicate purpose and value the manifest provides the product s the manifest provides the primary product author s name contact url and contact email the manifest provides an issue tracking url in case consumers need to report a defect the manifest lists all third party production dependencies and their semantic versions the manifest lists all third party development dependencies and their semantic versions the manifest provides the software product s homepage url the manifest provides a list of keywords that can be used to help classify the software product the manifest provides the uri to the product s primary source code repository the manifest declares the product s open source license reference material has good sample code that reads from and writes to a product s package json file describes every property in a node js product manifest the package json file
0
123,802
10,291,644,529
IssuesEvent
2019-08-27 12:56:54
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed test: _size_out_of_range_direct=false
C-test-failure O-robot
The following tests appear to have failed on master (testrace): _size_out_of_range_direct=false You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+_size_out_of_range_direct=false). [#1451983](https://teamcity.cockroachdb.com/viewLog.html?buildId=1451983): ``` _size_out_of_range_direct=false --- FAIL: testrace/TestImportData/PGDUMP:_size_out_of_range_direct=false (0.000s) Test ended in panic. ------- Stdout: ------- I190823 22:18:23.891262 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "create_database", target: 120, info: {DatabaseName:d34 Statement:CREATE DATABASE d34 User:root} I190823 22:18:24.123243 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "drop_database", target: 120, info: {DatabaseName:d34 Statement:DROP DATABASE d34 User:root DroppedSchemaObjects:[]} _size_out_of_range_direct=false --- FAIL: testrace/TestImportData/PGCOPY:_size_out_of_range_direct=false (0.000s) Test ended in panic. ------- Stdout: ------- I190823 22:18:16.860213 187 storage/store.go:2593 [n1,s1,r59/1:/Table/86{-/1}] removing replica r57/1 I190823 22:18:16.887111 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "create_database", target: 107, info: {DatabaseName:d27 Statement:CREATE DATABASE d27 User:root} I190823 22:18:17.032604 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "drop_database", target: 107, info: {DatabaseName:d27 Statement:DROP DATABASE d27 User:root DroppedSchemaObjects:[]} ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: _size_out_of_range_direct=false - The following tests appear to have failed on master (testrace): _size_out_of_range_direct=false You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+_size_out_of_range_direct=false). [#1451983](https://teamcity.cockroachdb.com/viewLog.html?buildId=1451983): ``` _size_out_of_range_direct=false --- FAIL: testrace/TestImportData/PGDUMP:_size_out_of_range_direct=false (0.000s) Test ended in panic. ------- Stdout: ------- I190823 22:18:23.891262 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "create_database", target: 120, info: {DatabaseName:d34 Statement:CREATE DATABASE d34 User:root} I190823 22:18:24.123243 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "drop_database", target: 120, info: {DatabaseName:d34 Statement:DROP DATABASE d34 User:root DroppedSchemaObjects:[]} _size_out_of_range_direct=false --- FAIL: testrace/TestImportData/PGCOPY:_size_out_of_range_direct=false (0.000s) Test ended in panic. ------- Stdout: ------- I190823 22:18:16.860213 187 storage/store.go:2593 [n1,s1,r59/1:/Table/86{-/1}] removing replica r57/1 I190823 22:18:16.887111 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "create_database", target: 107, info: {DatabaseName:d27 Statement:CREATE DATABASE d27 User:root} I190823 22:18:17.032604 861 sql/event_log.go:130 [n1,client=127.0.0.1:46736,user=root] Event: "drop_database", target: 107, info: {DatabaseName:d27 Statement:DROP DATABASE d27 User:root DroppedSchemaObjects:[]} ``` Please assign, take a look and update the issue accordingly.
test
teamcity failed test size out of range direct false the following tests appear to have failed on master testrace size out of range direct false you may want to check size out of range direct false fail testrace testimportdata pgdump size out of range direct false test ended in panic stdout sql event log go event create database target info databasename statement create database user root sql event log go event drop database target info databasename statement drop database user root droppedschemaobjects size out of range direct false fail testrace testimportdata pgcopy size out of range direct false test ended in panic stdout storage store go removing replica sql event log go event create database target info databasename statement create database user root sql event log go event drop database target info databasename statement drop database user root droppedschemaobjects please assign take a look and update the issue accordingly
1
226,689
24,994,505,288
IssuesEvent
2022-11-02 22:10:00
RG4421/ampere-centos-kernel
https://api.github.com/repos/RG4421/ampere-centos-kernel
reopened
CVE-2019-15212 (Medium) detected in linux-yocto-devv5.3
security vulnerability
## CVE-2019-15212 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv5.3</b></p></summary> <p> <p>Linux Embedded Kernel - tracks the next mainline release</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.1.8. There is a double-free caused by a malicious USB device in the drivers/usb/misc/rio500.c driver. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15212>CVE-2019-15212</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212</a></p> <p>Release Date: 2019-09-03</p> <p>Fix Resolution: v5.2-rc3</p> </p> </details> <p></p>
True
CVE-2019-15212 (Medium) detected in linux-yocto-devv5.3 - ## CVE-2019-15212 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linux-yocto-devv5.3</b></p></summary> <p> <p>Linux Embedded Kernel - tracks the next mainline release</p> <p>Library home page: <a href=https://git.yoctoproject.org/git/linux-yocto-dev>https://git.yoctoproject.org/git/linux-yocto-dev</a></p> <p>Found in base branch: <b>amp-centos-8.0-kernel</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (3)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/drivers/usb/misc/rio500.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in the Linux kernel before 5.1.8. There is a double-free caused by a malicious USB device in the drivers/usb/misc/rio500.c driver. <p>Publish Date: 2019-08-19 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2019-15212>CVE-2019-15212</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-15212</a></p> <p>Release Date: 2019-09-03</p> <p>Fix Resolution: v5.2-rc3</p> </p> </details> <p></p>
non_test
cve medium detected in linux yocto cve medium severity vulnerability vulnerable library linux yocto linux embedded kernel tracks the next mainline release library home page a href found in base branch amp centos kernel vulnerable source files drivers usb misc c drivers usb misc c drivers usb misc c vulnerability details an issue was discovered in the linux kernel before there is a double free caused by a malicious usb device in the drivers usb misc c driver publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0