Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
5
112
repo_url
stringlengths
34
141
action
stringclasses
3 values
title
stringlengths
1
757
labels
stringlengths
4
664
body
stringlengths
3
261k
index
stringclasses
10 values
text_combine
stringlengths
96
261k
label
stringclasses
2 values
text
stringlengths
96
232k
binary_label
int64
0
1
734,637
25,356,805,879
IssuesEvent
2022-11-20 12:44:08
YunoHost/issues
https://api.github.com/repos/YunoHost/issues
closed
[idea] instead of remove domain, archive them instead
:star2: ux Priority: low
That would allow to "undo" mistakes and keep certificates in case of. Not sure it's really worth it.
1.0
[idea] instead of remove domain, archive them instead - That would allow to "undo" mistakes and keep certificates in case of. Not sure it's really worth it.
non_defect
instead of remove domain archive them instead that would allow to undo mistakes and keep certificates in case of not sure it s really worth it
0
79,784
29,048,982,505
IssuesEvent
2023-05-13 23:58:26
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
BOINC crashing
C: Manager P: Undetermined R: worksforme T: Defect E: to be determined Validate
Hello, so I run BOINC on Windows 10 and it sometimes throws errors or crashes. So my system is running 24/7 and it sometimes crashes and basically wastes the time where it could do useful work. Can you guys work a little on BOINC, improve UI, make it more user friendly and make not crash?
1.0
BOINC crashing - Hello, so I run BOINC on Windows 10 and it sometimes throws errors or crashes. So my system is running 24/7 and it sometimes crashes and basically wastes the time where it could do useful work. Can you guys work a little on BOINC, improve UI, make it more user friendly and make not crash?
defect
boinc crashing hello so i run boinc on windows and it sometimes throws errors or crashes so my system is running and it sometimes crashes and basically wastes the time where it could do useful work can you guys work a little on boinc improve ui make it more user friendly and make not crash
1
58,047
16,342,337,751
IssuesEvent
2021-05-13 00:04:35
darshan-hpc/darshan
https://api.github.com/repos/darshan-hpc/darshan
closed
concurrent I/O from threads gets counted twice in timing
defect wrapper libraries
In GitLab by @shanedsnyder on Sep 24, 2015, 16:25 If two threads (in the same MPI process) access the same file concurrently, then the cumulative time counters are incremented too far. We need to add a reference count to the run-time data structure to tell how many threads are accessing the same file at once. The time should not be incremented until the reference counter hits zero. This does not require a log format change.
1.0
concurrent I/O from threads gets counted twice in timing - In GitLab by @shanedsnyder on Sep 24, 2015, 16:25 If two threads (in the same MPI process) access the same file concurrently, then the cumulative time counters are incremented too far. We need to add a reference count to the run-time data structure to tell how many threads are accessing the same file at once. The time should not be incremented until the reference counter hits zero. This does not require a log format change.
defect
concurrent i o from threads gets counted twice in timing in gitlab by shanedsnyder on sep if two threads in the same mpi process access the same file concurrently then the cumulative time counters are incremented too far we need to add a reference count to the run time data structure to tell how many threads are accessing the same file at once the time should not be incremented until the reference counter hits zero this does not require a log format change
1
67,250
16,854,265,692
IssuesEvent
2021-06-21 02:53:40
rust-lang/docs.rs
https://api.github.com/repos/rust-lang/docs.rs
closed
Read-only filesystem error using cc during build
C-build-failure S-waiting-on-crate-author
**Crate name:** spirv_cross or crates depending on it (e.g. gfx-backend-dx12) **Build failure link:** https://docs.rs/crate/gfx-backend-dx12/0.5.9/builds/281262 **Additional details:** When running cc from spirv_cross we get the following message: > [INFO] [stderr] running: "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-g" "-fno-omit-frame-pointer" "-m64" "-Wall" "-Wextra" "-DSPIRV_CROSS_EXCEPTIONS_TO_ASSERTIONS" "-DSPIRV_CROSS_WRAPPER_NO_EXCEPTIONS" "-DSPIRV_CROSS_WRAPPER_GLSL" "-DSPIRV_CROSS_WRAPPER_HLSL" "-Fo/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o" "-c" "src/wrapper.cpp" [INFO] [stderr] cargo:warning=src/wrapper.cpp:2: warning: ignoring #pragma warning [-Wunknown-pragmas] [INFO] [stderr] cargo:warning= 2 | #pragma warning(disable : 4996 4101) [INFO] [stderr] cargo:warning= | [INFO] [stderr] cargo:warning=Assembler messages: [INFO] [stderr] cargo:warning=Fatal error: can't create wrapper.o: Read-only file system [INFO] [stderr] exit code: 1 [INFO] [stderr] [INFO] [stderr] --- stderr [INFO] [stderr] [INFO] [stderr] [INFO] [stderr] error occurred: Command "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-g" "-fno-omit-frame-pointer" "-m64" "-Wall" "-Wextra" "-DSPIRV_CROSS_EXCEPTIONS_TO_ASSERTIONS" "-DSPIRV_CROSS_WRAPPER_NO_EXCEPTIONS" "-DSPIRV_CROSS_WRAPPER_GLSL" "-DSPIRV_CROSS_WRAPPER_HLSL" "-Fo/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o" "-c" "src/wrapper.cpp" with args "c++" did not execute successfully (status code exit code: 1). [INFO] [stderr] [INFO] [stderr] [INFO] [stderr] warning: build failed, waiting for other jobs to finish... [INFO] [stderr] error: build failed I'm not sure how to resolve this so I was wondering if I could get some help. I thought the path (`/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o`) would be writable because it's inside of `out`. The spirv_cross build script is a pretty typical cc setup and should use `OUT_DIR` from the environment: https://github.com/grovesNL/spirv_cross/blob/master/spirv_cross/build.rs
1.0
Read-only filesystem error using cc during build - **Crate name:** spirv_cross or crates depending on it (e.g. gfx-backend-dx12) **Build failure link:** https://docs.rs/crate/gfx-backend-dx12/0.5.9/builds/281262 **Additional details:** When running cc from spirv_cross we get the following message: > [INFO] [stderr] running: "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-g" "-fno-omit-frame-pointer" "-m64" "-Wall" "-Wextra" "-DSPIRV_CROSS_EXCEPTIONS_TO_ASSERTIONS" "-DSPIRV_CROSS_WRAPPER_NO_EXCEPTIONS" "-DSPIRV_CROSS_WRAPPER_GLSL" "-DSPIRV_CROSS_WRAPPER_HLSL" "-Fo/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o" "-c" "src/wrapper.cpp" [INFO] [stderr] cargo:warning=src/wrapper.cpp:2: warning: ignoring #pragma warning [-Wunknown-pragmas] [INFO] [stderr] cargo:warning= 2 | #pragma warning(disable : 4996 4101) [INFO] [stderr] cargo:warning= | [INFO] [stderr] cargo:warning=Assembler messages: [INFO] [stderr] cargo:warning=Fatal error: can't create wrapper.o: Read-only file system [INFO] [stderr] exit code: 1 [INFO] [stderr] [INFO] [stderr] --- stderr [INFO] [stderr] [INFO] [stderr] [INFO] [stderr] error occurred: Command "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-g" "-fno-omit-frame-pointer" "-m64" "-Wall" "-Wextra" "-DSPIRV_CROSS_EXCEPTIONS_TO_ASSERTIONS" "-DSPIRV_CROSS_WRAPPER_NO_EXCEPTIONS" "-DSPIRV_CROSS_WRAPPER_GLSL" "-DSPIRV_CROSS_WRAPPER_HLSL" "-Fo/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o" "-c" "src/wrapper.cpp" with args "c++" did not execute successfully (status code exit code: 1). [INFO] [stderr] [INFO] [stderr] [INFO] [stderr] warning: build failed, waiting for other jobs to finish... [INFO] [stderr] error: build failed I'm not sure how to resolve this so I was wondering if I could get some help. I thought the path (`/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o`) would be writable because it's inside of `out`. The spirv_cross build script is a pretty typical cc setup and should use `OUT_DIR` from the environment: https://github.com/grovesNL/spirv_cross/blob/master/spirv_cross/build.rs
non_defect
read only filesystem error using cc during build crate name spirv cross or crates depending on it e g gfx backend build failure link additional details when running cc from spirv cross we get the following message running c ffunction sections fdata sections g fno omit frame pointer wall wextra dspirv cross exceptions to assertions dspirv cross wrapper no exceptions dspirv cross wrapper glsl dspirv cross wrapper hlsl fo opt rustwide target pc windows msvc debug build spirv cross out src wrapper o c src wrapper cpp cargo warning src wrapper cpp warning ignoring pragma warning cargo warning pragma warning disable cargo warning cargo warning assembler messages cargo warning fatal error can t create wrapper o read only file system exit code stderr error occurred command c ffunction sections fdata sections g fno omit frame pointer wall wextra dspirv cross exceptions to assertions dspirv cross wrapper no exceptions dspirv cross wrapper glsl dspirv cross wrapper hlsl fo opt rustwide target pc windows msvc debug build spirv cross out src wrapper o c src wrapper cpp with args c did not execute successfully status code exit code warning build failed waiting for other jobs to finish error build failed i m not sure how to resolve this so i was wondering if i could get some help i thought the path opt rustwide target pc windows msvc debug build spirv cross out src wrapper o would be writable because it s inside of out the spirv cross build script is a pretty typical cc setup and should use out dir from the environment
0
79,256
15,586,118,757
IssuesEvent
2021-03-18 01:12:55
wasimakh2/JPAGenratorRelease
https://api.github.com/repos/wasimakh2/JPAGenratorRelease
closed
CVE-2019-12814 (Medium) detected in jackson-databind-2.9.9.jar - autoclosed
security vulnerability
## CVE-2019-12814 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: _depth_0/JPAGenratorRelease/JPAGenrator-0.0.1-SNAPSHOT/WEB-INF/lib/jackson-databind-2.9.9.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/wasimakh2/JPAGenratorRelease/commit/2d85088b97dabdacb7e6cceaea3a329a2eb43f87">2d85088b97dabdacb7e6cceaea3a329a2eb43f87</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x through 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has JDOM 1.x or 2.x jar in the classpath, an attacker can send a specifically crafted JSON message that allows them to read arbitrary local files on the server. <p>Publish Date: 2019-06-19 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12814>CVE-2019-12814</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2341">https://github.com/FasterXML/jackson-databind/issues/2341</a></p> <p>Release Date: 2019-06-19</p> <p>Fix Resolution: 2.7.9.6, 2.8.11.4, 2.9.9.1, 2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-12814 (Medium) detected in jackson-databind-2.9.9.jar - autoclosed - ## CVE-2019-12814 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to vulnerable library: _depth_0/JPAGenratorRelease/JPAGenrator-0.0.1-SNAPSHOT/WEB-INF/lib/jackson-databind-2.9.9.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.9.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/wasimakh2/JPAGenratorRelease/commit/2d85088b97dabdacb7e6cceaea3a329a2eb43f87">2d85088b97dabdacb7e6cceaea3a329a2eb43f87</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x through 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has JDOM 1.x or 2.x jar in the classpath, an attacker can send a specifically crafted JSON message that allows them to read arbitrary local files on the server. <p>Publish Date: 2019-06-19 <p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12814>CVE-2019-12814</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2341">https://github.com/FasterXML/jackson-databind/issues/2341</a></p> <p>Release Date: 2019-06-19</p> <p>Fix Resolution: 2.7.9.6, 2.8.11.4, 2.9.9.1, 2.10.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve medium detected in jackson databind jar autoclosed cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library depth jpagenratorrelease jpagenrator snapshot web inf lib jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind x through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has jdom x or x jar in the classpath an attacker can send a specifically crafted json message that allows them to read arbitrary local files on the server publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
77,726
27,133,402,576
IssuesEvent
2023-02-16 11:25:31
vector-im/element-x-ios
https://api.github.com/repos/vector-im/element-x-ios
closed
Left Panel room previews are often stale
A-Room-List T-Defect TF:1
Running EIX on macOS \(as an iOS app on M1 macOS\), the room previews often show messages which are 3-4 events behind the actual most recent message in the room in question.
1.0
Left Panel room previews are often stale - Running EIX on macOS \(as an iOS app on M1 macOS\), the room previews often show messages which are 3-4 events behind the actual most recent message in the room in question.
defect
left panel room previews are often stale running eix on macos as an ios app on macos the room previews often show messages which are events behind the actual most recent message in the room in question
1
2,858
2,607,963,308
IssuesEvent
2015-02-26 00:41:12
chrsmithdemos/leveldb
https://api.github.com/repos/chrsmithdemos/leveldb
closed
Add OpenBSD support
auto-migrated Priority-Medium Type-Defect
``` Attached is a patch to allow leveldb to build (and pass regress tests) on OpenBSD. Tested succesfully on OpenBSD/amd64, 5.0. ``` ----- Original issue reported on code.google.com by `jasper.l...@gmail.com` on 18 Aug 2011 at 3:01 * Merged into: #59 Attachments: * [leveldb_openbsd.patch](https://storage.googleapis.com/google-code-attachments/leveldb/issue-31/comment-0/leveldb_openbsd.patch)
1.0
Add OpenBSD support - ``` Attached is a patch to allow leveldb to build (and pass regress tests) on OpenBSD. Tested succesfully on OpenBSD/amd64, 5.0. ``` ----- Original issue reported on code.google.com by `jasper.l...@gmail.com` on 18 Aug 2011 at 3:01 * Merged into: #59 Attachments: * [leveldb_openbsd.patch](https://storage.googleapis.com/google-code-attachments/leveldb/issue-31/comment-0/leveldb_openbsd.patch)
defect
add openbsd support attached is a patch to allow leveldb to build and pass regress tests on openbsd tested succesfully on openbsd original issue reported on code google com by jasper l gmail com on aug at merged into attachments
1
210,570
23,754,872,453
IssuesEvent
2022-09-01 01:23:59
LingalaShalini/sqlite-version-3.21.0_after_fix
https://api.github.com/repos/LingalaShalini/sqlite-version-3.21.0_after_fix
opened
CVE-2021-20223 (High) detected in sqliteversion-3.21.0
security vulnerability
## CVE-2021-20223 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sqliteversion-3.21.0</b></p></summary> <p> <p>Official Git mirror of the SQLite source tree</p> <p>Library home page: <a href=https://github.com/sqlite/sqlite.git>https://github.com/sqlite/sqlite.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/LingalaShalini/sqlite-version-3.21.0_after_fix/commit/c54261c167d3ed3bb3cdbcfc62a5337ec33bc1bd">c54261c167d3ed3bb3cdbcfc62a5337ec33bc1bd</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/ext/fts5/fts5_unicode2.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was found in fts5UnicodeTokenize() in ext/fts5/fts5_tokenize.c in Sqlite. A unicode61 tokenizer configured to treat unicode "control-characters" (class Cc), was treating embedded nul characters as tokens. The issue was fixed in sqlite-3.34.0 and later. <p>Publish Date: 2022-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20223>CVE-2021-20223</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.sqlite.org/forum/forumpost/09609d7e22">https://www.sqlite.org/forum/forumpost/09609d7e22</a></p> <p>Release Date: 2022-08-25</p> <p>Fix Resolution: version-3.34.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-20223 (High) detected in sqliteversion-3.21.0 - ## CVE-2021-20223 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sqliteversion-3.21.0</b></p></summary> <p> <p>Official Git mirror of the SQLite source tree</p> <p>Library home page: <a href=https://github.com/sqlite/sqlite.git>https://github.com/sqlite/sqlite.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/LingalaShalini/sqlite-version-3.21.0_after_fix/commit/c54261c167d3ed3bb3cdbcfc62a5337ec33bc1bd">c54261c167d3ed3bb3cdbcfc62a5337ec33bc1bd</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/ext/fts5/fts5_unicode2.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was found in fts5UnicodeTokenize() in ext/fts5/fts5_tokenize.c in Sqlite. A unicode61 tokenizer configured to treat unicode "control-characters" (class Cc), was treating embedded nul characters as tokens. The issue was fixed in sqlite-3.34.0 and later. <p>Publish Date: 2022-08-25 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20223>CVE-2021-20223</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.sqlite.org/forum/forumpost/09609d7e22">https://www.sqlite.org/forum/forumpost/09609d7e22</a></p> <p>Release Date: 2022-08-25</p> <p>Fix Resolution: version-3.34.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in sqliteversion cve high severity vulnerability vulnerable library sqliteversion official git mirror of the sqlite source tree library home page a href found in head commit a href found in base branch master vulnerable source files ext c vulnerability details an issue was found in in ext tokenize c in sqlite a tokenizer configured to treat unicode control characters class cc was treating embedded nul characters as tokens the issue was fixed in sqlite and later publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution version step up your open source security game with mend
0
20,072
3,295,088,182
IssuesEvent
2015-10-31 16:45:09
FreeRADIUS/freeradius-server
https://api.github.com/repos/FreeRADIUS/freeradius-server
closed
Fatal Python error: PyThreadState_Get: no current thread, cause by _locale.so
upstream defect v3.0.x v3.1.x
I found module rlm_python will crash when instantiating because of the _locale.so in python. ```bash $ ./radiusd -X -f Copyright (C) 1999-2015 The FreeRADIUS server project and contributors There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE You may redistribute copies of FreeRADIUS under the terms of the GNU General Public License For more information about these matters, see the file named COPYRIGHT Starting - reading configuration files ... .... # Instantiating module "attr_filter.accounting_response" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/attr_filter reading pairlist file /home/develop/app/freeraiuds-server/etc/raddb/mods-config/attr_filter/accounting_response # Instantiating module "mschap" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/mschap rlm_mschap (mschap): using internal authentication # Instantiating module "python" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/python Fatal Python error: PyThreadState_Get: no current thread core dumped) ``` but if i rename the _locale.so: ```bash $ ./radiusd -X -f Copyright (C) 1999-2015 The FreeRADIUS server project and contributors There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE You may redistribute copies of FreeRADIUS under the terms of the GNU General Public License For more information about these matters, see the file named COPYRIGHT Starting - reading configuration files ... .... cd $PYTHONHOME/lib/python2.7/lib-dynload mv _locale.so _locale.so.bak ``` freeradius-server will run. ```bash } } listen { type = "acct" ipv6addr = :: port = 0 limit { max_connections = 16 lifetime = 0 idle_timeout = 30 } } Listening on auth address 127.0.0.1 port 18120 bound to server inner-tunnel Listening on auth address * port 1812 bound to server default Listening on acct address * port 1813 bound to server default Listening on auth address :: port 1812 bound to server default Listening on acct address :: port 1813 bound to server default Opening new proxy socket 'proxy address * port 0' Listening on proxy address * port 34786 Ready to process requests ``` it's any idea for this ???
1.0
Fatal Python error: PyThreadState_Get: no current thread, cause by _locale.so - I found module rlm_python will crash when instantiating because of the _locale.so in python. ```bash $ ./radiusd -X -f Copyright (C) 1999-2015 The FreeRADIUS server project and contributors There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE You may redistribute copies of FreeRADIUS under the terms of the GNU General Public License For more information about these matters, see the file named COPYRIGHT Starting - reading configuration files ... .... # Instantiating module "attr_filter.accounting_response" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/attr_filter reading pairlist file /home/develop/app/freeraiuds-server/etc/raddb/mods-config/attr_filter/accounting_response # Instantiating module "mschap" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/mschap rlm_mschap (mschap): using internal authentication # Instantiating module "python" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/python Fatal Python error: PyThreadState_Get: no current thread core dumped) ``` but if i rename the _locale.so: ```bash $ ./radiusd -X -f Copyright (C) 1999-2015 The FreeRADIUS server project and contributors There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE You may redistribute copies of FreeRADIUS under the terms of the GNU General Public License For more information about these matters, see the file named COPYRIGHT Starting - reading configuration files ... .... cd $PYTHONHOME/lib/python2.7/lib-dynload mv _locale.so _locale.so.bak ``` freeradius-server will run. ```bash } } listen { type = "acct" ipv6addr = :: port = 0 limit { max_connections = 16 lifetime = 0 idle_timeout = 30 } } Listening on auth address 127.0.0.1 port 18120 bound to server inner-tunnel Listening on auth address * port 1812 bound to server default Listening on acct address * port 1813 bound to server default Listening on auth address :: port 1812 bound to server default Listening on acct address :: port 1813 bound to server default Opening new proxy socket 'proxy address * port 0' Listening on proxy address * port 34786 Ready to process requests ``` it's any idea for this ???
defect
fatal python error pythreadstate get no current thread cause by locale so i found module rlm python will crash when instantiating because of the locale so in python bash radiusd x f copyright c the freeradius server project and contributors there is no warranty not even for merchantability or fitness for a particular purpose you may redistribute copies of freeradius under the terms of the gnu general public license for more information about these matters see the file named copyright starting reading configuration files instantiating module attr filter accounting response from file home develop app freeraiuds server etc raddb mods enabled attr filter reading pairlist file home develop app freeraiuds server etc raddb mods config attr filter accounting response instantiating module mschap from file home develop app freeraiuds server etc raddb mods enabled mschap rlm mschap mschap using internal authentication instantiating module python from file home develop app freeraiuds server etc raddb mods enabled python fatal python error pythreadstate get no current thread core dumped but if i rename the locale so bash radiusd x f copyright c the freeradius server project and contributors there is no warranty not even for merchantability or fitness for a particular purpose you may redistribute copies of freeradius under the terms of the gnu general public license for more information about these matters see the file named copyright starting reading configuration files cd pythonhome lib lib dynload mv locale so locale so bak freeradius server will run bash listen type acct port limit max connections lifetime idle timeout listening on auth address port bound to server inner tunnel listening on auth address port bound to server default listening on acct address port bound to server default listening on auth address port bound to server default listening on acct address port bound to server default opening new proxy socket proxy address port listening on proxy address port ready to process requests it s any idea for this
1
85,545
7,976,421,370
IssuesEvent
2018-07-17 12:39:20
loot/skyrimse
https://api.github.com/repos/loot/skyrimse
closed
Wildcat and Ultimate Combat - Compatibility
Testing awaiting-feedback
Please remove UltimateCombat.esp from the "Load after" rules for Wildcat (added in 55e23755e2149c12dd76b4408ff0e0c2cb7b07aa). First of all, the Wildcat and Ultimate Combat pages give conflicting advice, and second, if using them both one needs to control their order manually based on patches (e.g. OMEGA MLU says to load Ultimate Combat after Wildcat, which is currently impossible due to the masterlist rule).
1.0
Wildcat and Ultimate Combat - Compatibility - Please remove UltimateCombat.esp from the "Load after" rules for Wildcat (added in 55e23755e2149c12dd76b4408ff0e0c2cb7b07aa). First of all, the Wildcat and Ultimate Combat pages give conflicting advice, and second, if using them both one needs to control their order manually based on patches (e.g. OMEGA MLU says to load Ultimate Combat after Wildcat, which is currently impossible due to the masterlist rule).
non_defect
wildcat and ultimate combat compatibility please remove ultimatecombat esp from the load after rules for wildcat added in first of all the wildcat and ultimate combat pages give conflicting advice and second if using them both one needs to control their order manually based on patches e g omega mlu says to load ultimate combat after wildcat which is currently impossible due to the masterlist rule
0
231,670
7,642,044,673
IssuesEvent
2018-05-08 07:54:53
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
opened
verdi calculation show should check whether the id references a calculation node
priority/nice to have topic/VerdiCommandLine type/bug
Currently it does not and will except on some of the properties: ``` Traceback (most recent call last): File "/home/aiida/.virtualenvs/aiida/bin/verdi", line 9, in <module> sys.exit(run()) File "/home/aiida/codes/aiida_core/aiida/cmdline/verdilib.py", line 1056, in run aiida.cmdline.verdilib.exec_from_cmdline(sys.argv) File "/home/aiida/codes/aiida_core/aiida/cmdline/verdilib.py", line 1041, in exec_from_cmdline CommandClass.run(*argv[command_position + 1:]) File "/home/aiida/codes/aiida_core/aiida/cmdline/baseclass.py", line 234, in run function_to_call(*args[1:]) File "/home/aiida/codes/aiida_core/aiida/cmdline/commands/calculation.py", line 315, in calculation_show print_node_info(calc) File "/home/aiida/codes/aiida_core/aiida/cmdline/utils/common.py", line 95, in print_node_info print_node_summary(node) File "/home/aiida/codes/aiida_core/aiida/cmdline/utils/common.py", line 60, in print_node_summary table.append(['process state', ProcessState(node.process_state)]) AttributeError: 'FolderData' object has no attribute 'process_state' ```
1.0
verdi calculation show should check whether the id references a calculation node - Currently it does not and will except on some of the properties: ``` Traceback (most recent call last): File "/home/aiida/.virtualenvs/aiida/bin/verdi", line 9, in <module> sys.exit(run()) File "/home/aiida/codes/aiida_core/aiida/cmdline/verdilib.py", line 1056, in run aiida.cmdline.verdilib.exec_from_cmdline(sys.argv) File "/home/aiida/codes/aiida_core/aiida/cmdline/verdilib.py", line 1041, in exec_from_cmdline CommandClass.run(*argv[command_position + 1:]) File "/home/aiida/codes/aiida_core/aiida/cmdline/baseclass.py", line 234, in run function_to_call(*args[1:]) File "/home/aiida/codes/aiida_core/aiida/cmdline/commands/calculation.py", line 315, in calculation_show print_node_info(calc) File "/home/aiida/codes/aiida_core/aiida/cmdline/utils/common.py", line 95, in print_node_info print_node_summary(node) File "/home/aiida/codes/aiida_core/aiida/cmdline/utils/common.py", line 60, in print_node_summary table.append(['process state', ProcessState(node.process_state)]) AttributeError: 'FolderData' object has no attribute 'process_state' ```
non_defect
verdi calculation show should check whether the id references a calculation node currently it does not and will except on some of the properties traceback most recent call last file home aiida virtualenvs aiida bin verdi line in sys exit run file home aiida codes aiida core aiida cmdline verdilib py line in run aiida cmdline verdilib exec from cmdline sys argv file home aiida codes aiida core aiida cmdline verdilib py line in exec from cmdline commandclass run argv file home aiida codes aiida core aiida cmdline baseclass py line in run function to call args file home aiida codes aiida core aiida cmdline commands calculation py line in calculation show print node info calc file home aiida codes aiida core aiida cmdline utils common py line in print node info print node summary node file home aiida codes aiida core aiida cmdline utils common py line in print node summary table append attributeerror folderdata object has no attribute process state
0
25,787
3,964,876,590
IssuesEvent
2016-05-03 04:21:11
uccser/cs-field-guide
https://api.github.com/repos/uccser/cs-field-guide
closed
Numbering questions in a box
bug help wanted website/design
In the data representation chapter we have "1. Represent each character in your name using UTF-32.", which is part of a numbered list. However, the rendered version doesn't number the list (same for the next spoiler box).
1.0
Numbering questions in a box - In the data representation chapter we have "1. Represent each character in your name using UTF-32.", which is part of a numbered list. However, the rendered version doesn't number the list (same for the next spoiler box).
non_defect
numbering questions in a box in the data representation chapter we have represent each character in your name using utf which is part of a numbered list however the rendered version doesn t number the list same for the next spoiler box
0
42,135
10,821,170,574
IssuesEvent
2019-11-08 18:01:33
vector-im/riot-web
https://api.github.com/repos/vector-im/riot-web
closed
Formatting bar has lost its rounded corners
bug defect p1 ui/ux 🔥 Fire 🔥
Latest develop / latest macOS / latest Chrome. Should be stylised the same as message actions. See GIF in https://github.com/vector-im/riot-web/issues/11202.
1.0
Formatting bar has lost its rounded corners - Latest develop / latest macOS / latest Chrome. Should be stylised the same as message actions. See GIF in https://github.com/vector-im/riot-web/issues/11202.
defect
formatting bar has lost its rounded corners latest develop latest macos latest chrome should be stylised the same as message actions see gif in
1
111,909
9,545,970,503
IssuesEvent
2019-05-01 18:34:35
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
teamcity: failed test: TestLint
C-test-failure O-robot
The following tests appear to have failed on master (lint): TestLint/TestVet: TestLint/TestVet/shadow, TestLint/TestVet, TestLint You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestLint). [#1270290](https://teamcity.cockroachdb.com/viewLog.html?buildId=1270290): ``` TestLint --- FAIL: lint/TestLint (274.410s) TestLint/TestVet --- FAIL: lint/TestLint: TestLint/TestVet (578.600s) ------- Stdout: ------- === PAUSE TestLint/TestVet TestLint/TestVet: TestLint/TestVet/shadow ...rc/runtime/malloc.go:913 +0x896 fp=0xc0086558c0 sp=0xc008655820 pc=0x40bb26 lint_test.go:1352: runtime.growslice(0xc73e40, 0xc017fda000, 0xf00, 0xf00, 0xf01, 0xc015d8cc88, 0x0, 0x0) lint_test.go:1352: /usr/local/go/src/runtime/slice.go:204 +0x145 fp=0xc008655928 sp=0xc0086558c0 pc=0x43f655 lint_test.go:1352: cmd/internal/obj.Addrel(0xc001383b90, 0xc008655a60) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/data.go:186 +0xf9 fp=0xc008655980 sp=0xc008655928 pc=0x4cf8f9 lint_test.go:1352: cmd/internal/obj/x86.(*AsmBuf).doasm(0xc008655d78, 0xc0002ea000, 0xc001383b90, 0xc015d8cdd0) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/x86/asm6.go:4677 +0xeb8 fp=0xc008655c58 sp=0xc008655980 pc=0x5c5748 lint_test.go:1352: cmd/internal/obj/x86.(*AsmBuf).asmins(0xc008655d78, 0xc0002ea000, 0xc001383b90, 0xc015d8cdd0) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/x86/asm6.go:5369 +0x4a6 fp=0xc008655d00 sp=0xc008655c58 pc=0x5cc986 lint_test.go:1352: cmd/internal/obj/x86.span6(0xc0002ea000, 0xc001383b90, 0xc00102a630) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/x86/asm6.go:1999 +0x3b2 fp=0xc008655e20 sp=0xc008655d00 pc=0x5be752 lint_test.go:1352: cmd/internal/obj.Flushplist(0xc0002ea000, 0xc008655f08, 0xc00102a630, 0x7ffd5d9278cb, 0x2f) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/plist.go:107 +0x6ab fp=0xc008655ee0 sp=0xc008655e20 pc=0x4d916b lint_test.go:1352: cmd/compile/internal/gc.(*Progs).Flush(0xc008581f10) lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/gsubr.go:93 +0xbc fp=0xc008655f30 sp=0xc008655ee0 pc=0xabd6bc lint_test.go:1352: cmd/compile/internal/gc.compileSSA(0xc00127d340, 0x3) lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/pgen.go:283 +0x2b9 fp=0xc008655f98 sp=0xc008655f30 pc=0xb02ed9 lint_test.go:1352: cmd/compile/internal/gc.compileFunctions.func2(0xc005966540, 0xc005975db0, 0x3) lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/pgen.go:323 +0x49 fp=0xc008655fc8 sp=0xc008655f98 pc=0xbb7f79 lint_test.go:1352: runtime.goexit() lint_test.go:1352: /usr/local/go/src/runtime/asm_amd64.s:1333 +0x1 fp=0xc008655fd0 sp=0xc008655fc8 pc=0x457da1 lint_test.go:1352: created by cmd/compile/internal/gc.compileFunctions lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/pgen.go:321 +0x11a lint_test.go:1352: goroutine 1 [semacquire]: lint_test.go:1352: sync.runtime_Semacquire(0xc005975db8) lint_test.go:1352: /usr/local/go/src/runtime/sema.go:56 +0x39 lint_test.go:1352: sync.(*WaitGroup).Wait(0xc005975db0) lint_test.go:1352: /usr/local/go/src/sync/waitgroup.go:130 +0x64 lint_test.go:1352: cmd/compile/internal/gc.compileFunctions() lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/pgen.go:333 +0x1c3 lint_test.go:1352: cmd/compile/internal/gc.Main(0xcc51f8) lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/main.go:651 +0x276d lint_test.go:1352: main.main() lint_test.go:1352: /usr/local/go/src/cmd/compile/main.go:51 +0x96 ``` Please assign, take a look and update the issue accordingly.
1.0
teamcity: failed test: TestLint - The following tests appear to have failed on master (lint): TestLint/TestVet: TestLint/TestVet/shadow, TestLint/TestVet, TestLint You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestLint). [#1270290](https://teamcity.cockroachdb.com/viewLog.html?buildId=1270290): ``` TestLint --- FAIL: lint/TestLint (274.410s) TestLint/TestVet --- FAIL: lint/TestLint: TestLint/TestVet (578.600s) ------- Stdout: ------- === PAUSE TestLint/TestVet TestLint/TestVet: TestLint/TestVet/shadow ...rc/runtime/malloc.go:913 +0x896 fp=0xc0086558c0 sp=0xc008655820 pc=0x40bb26 lint_test.go:1352: runtime.growslice(0xc73e40, 0xc017fda000, 0xf00, 0xf00, 0xf01, 0xc015d8cc88, 0x0, 0x0) lint_test.go:1352: /usr/local/go/src/runtime/slice.go:204 +0x145 fp=0xc008655928 sp=0xc0086558c0 pc=0x43f655 lint_test.go:1352: cmd/internal/obj.Addrel(0xc001383b90, 0xc008655a60) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/data.go:186 +0xf9 fp=0xc008655980 sp=0xc008655928 pc=0x4cf8f9 lint_test.go:1352: cmd/internal/obj/x86.(*AsmBuf).doasm(0xc008655d78, 0xc0002ea000, 0xc001383b90, 0xc015d8cdd0) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/x86/asm6.go:4677 +0xeb8 fp=0xc008655c58 sp=0xc008655980 pc=0x5c5748 lint_test.go:1352: cmd/internal/obj/x86.(*AsmBuf).asmins(0xc008655d78, 0xc0002ea000, 0xc001383b90, 0xc015d8cdd0) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/x86/asm6.go:5369 +0x4a6 fp=0xc008655d00 sp=0xc008655c58 pc=0x5cc986 lint_test.go:1352: cmd/internal/obj/x86.span6(0xc0002ea000, 0xc001383b90, 0xc00102a630) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/x86/asm6.go:1999 +0x3b2 fp=0xc008655e20 sp=0xc008655d00 pc=0x5be752 lint_test.go:1352: cmd/internal/obj.Flushplist(0xc0002ea000, 0xc008655f08, 0xc00102a630, 0x7ffd5d9278cb, 0x2f) lint_test.go:1352: /usr/local/go/src/cmd/internal/obj/plist.go:107 +0x6ab fp=0xc008655ee0 sp=0xc008655e20 pc=0x4d916b lint_test.go:1352: cmd/compile/internal/gc.(*Progs).Flush(0xc008581f10) lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/gsubr.go:93 +0xbc fp=0xc008655f30 sp=0xc008655ee0 pc=0xabd6bc lint_test.go:1352: cmd/compile/internal/gc.compileSSA(0xc00127d340, 0x3) lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/pgen.go:283 +0x2b9 fp=0xc008655f98 sp=0xc008655f30 pc=0xb02ed9 lint_test.go:1352: cmd/compile/internal/gc.compileFunctions.func2(0xc005966540, 0xc005975db0, 0x3) lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/pgen.go:323 +0x49 fp=0xc008655fc8 sp=0xc008655f98 pc=0xbb7f79 lint_test.go:1352: runtime.goexit() lint_test.go:1352: /usr/local/go/src/runtime/asm_amd64.s:1333 +0x1 fp=0xc008655fd0 sp=0xc008655fc8 pc=0x457da1 lint_test.go:1352: created by cmd/compile/internal/gc.compileFunctions lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/pgen.go:321 +0x11a lint_test.go:1352: goroutine 1 [semacquire]: lint_test.go:1352: sync.runtime_Semacquire(0xc005975db8) lint_test.go:1352: /usr/local/go/src/runtime/sema.go:56 +0x39 lint_test.go:1352: sync.(*WaitGroup).Wait(0xc005975db0) lint_test.go:1352: /usr/local/go/src/sync/waitgroup.go:130 +0x64 lint_test.go:1352: cmd/compile/internal/gc.compileFunctions() lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/pgen.go:333 +0x1c3 lint_test.go:1352: cmd/compile/internal/gc.Main(0xcc51f8) lint_test.go:1352: /usr/local/go/src/cmd/compile/internal/gc/main.go:651 +0x276d lint_test.go:1352: main.main() lint_test.go:1352: /usr/local/go/src/cmd/compile/main.go:51 +0x96 ``` Please assign, take a look and update the issue accordingly.
non_defect
teamcity failed test testlint the following tests appear to have failed on master lint testlint testvet testlint testvet shadow testlint testvet testlint you may want to check testlint fail lint testlint testlint testvet fail lint testlint testlint testvet stdout pause testlint testvet testlint testvet testlint testvet shadow rc runtime malloc go fp sp pc lint test go runtime growslice lint test go usr local go src runtime slice go fp sp pc lint test go cmd internal obj addrel lint test go usr local go src cmd internal obj data go fp sp pc lint test go cmd internal obj asmbuf doasm lint test go usr local go src cmd internal obj go fp sp pc lint test go cmd internal obj asmbuf asmins lint test go usr local go src cmd internal obj go fp sp pc lint test go cmd internal obj lint test go usr local go src cmd internal obj go fp sp pc lint test go cmd internal obj flushplist lint test go usr local go src cmd internal obj plist go fp sp pc lint test go cmd compile internal gc progs flush lint test go usr local go src cmd compile internal gc gsubr go fp sp pc lint test go cmd compile internal gc compilessa lint test go usr local go src cmd compile internal gc pgen go fp sp pc lint test go cmd compile internal gc compilefunctions lint test go usr local go src cmd compile internal gc pgen go fp sp pc lint test go runtime goexit lint test go usr local go src runtime asm s fp sp pc lint test go created by cmd compile internal gc compilefunctions lint test go usr local go src cmd compile internal gc pgen go lint test go goroutine lint test go sync runtime semacquire lint test go usr local go src runtime sema go lint test go sync waitgroup wait lint test go usr local go src sync waitgroup go lint test go cmd compile internal gc compilefunctions lint test go usr local go src cmd compile internal gc pgen go lint test go cmd compile internal gc main lint test go usr local go src cmd compile internal gc main go lint test go main main lint test go usr local go src cmd compile main go please assign take a look and update the issue accordingly
0
40,058
9,815,339,958
IssuesEvent
2019-06-13 12:26:46
roc-project/roc
https://api.github.com/repos/roc-project/roc
closed
OpenFEC LDPC incorrect decoding
defect
Can be reproduced with "full_repair_payload_sizes" test. Reproducible only on LDPC and on some payload sizes, e.g. 69. The last byte of the decoded packet is incorrect.
1.0
OpenFEC LDPC incorrect decoding - Can be reproduced with "full_repair_payload_sizes" test. Reproducible only on LDPC and on some payload sizes, e.g. 69. The last byte of the decoded packet is incorrect.
defect
openfec ldpc incorrect decoding can be reproduced with full repair payload sizes test reproducible only on ldpc and on some payload sizes e g the last byte of the decoded packet is incorrect
1
72,531
24,167,087,429
IssuesEvent
2022-09-22 15:51:05
idaholab/moose
https://api.github.com/repos/idaholab/moose
opened
AdvancedExtrudedMeshGenerator Documentation Update to Cover Triple-Indexed Parameter
T: defect P: normal
## Bug Description `AdvancedExtrudedMeshGenerator` was recently updated to use a triple-indexed input parameter. But the documentation has not been updated accordingly. ## Steps to Reproduce The documentation page of `AdvancedExtrudedMeshGenerator` still uses stacked double-indexed input parameter. ## Impact The documentation is confusing due to the inconsistence.
1.0
AdvancedExtrudedMeshGenerator Documentation Update to Cover Triple-Indexed Parameter - ## Bug Description `AdvancedExtrudedMeshGenerator` was recently updated to use a triple-indexed input parameter. But the documentation has not been updated accordingly. ## Steps to Reproduce The documentation page of `AdvancedExtrudedMeshGenerator` still uses stacked double-indexed input parameter. ## Impact The documentation is confusing due to the inconsistence.
defect
advancedextrudedmeshgenerator documentation update to cover triple indexed parameter bug description advancedextrudedmeshgenerator was recently updated to use a triple indexed input parameter but the documentation has not been updated accordingly steps to reproduce the documentation page of advancedextrudedmeshgenerator still uses stacked double indexed input parameter impact the documentation is confusing due to the inconsistence
1
62,618
17,092,699,146
IssuesEvent
2021-07-08 19:50:24
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
DatePicker: TimeOnly 12:04 a.m. (midnight plus 4 minutes) is not handled correctly
defect
(Apologies in advance, I am new to reporting defects. I kinda just looked at what others were doing to fill in the format below.) **Describe the defect** Using the time-only 12-hour datePicker causes 12:04 a.m. to become 12:04 p.m. The input box correctly displays "12:04 AM", but when opening the time picker, it gets changed to 12:04 PM" and updates the input box with the incorrect value. GIF below: ![PrimefacesDefect](https://user-images.githubusercontent.com/38224657/101688276-ca820e80-3a39-11eb-9dde-a22135f2a66e.gif) **Environment** PF Version: 8.0.5 JSF + version: Mojarra 2.3.14 Tomcat 8.5 Affected browsers: Chrome (at least) **To Reproduce** Steps to reproduce the behavior: 1. XHTML: ```<?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"> <html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en" lang="en" xmlns:p="http://primefaces.org/ui" xmlns:h="http://java.sun.com/jsf/html" xmlns:f="http://java.sun.com/jsf/core"> <h:head> <title>Title</title> </h:head> <body> <p:datePicker id="datePickerDemo" hourFormat="12" value="#{rTest.dateExample}" timeOnly="true"/> </body> </html> ``` 2. Backing Bean (note: same behavior with LocalTime and Date) ```import java.time.LocalTime; import javax.enterprise.context.RequestScoped; import javax.inject.Named; @RequestScoped @Named("rTest") public class RequestTest { private LocalTime dateExample = LocalTime.of(0, 4); public LocalTime getDateExample() { return dateExample; } public void setDateExample(LocalTime dateExample) { this.dateExample = dateExample; } } ``` **Commentary** The issue appears to be in [this code](https://github.com/primefaces/primefaces/blob/6c67dadd07ab2b7887f43f14da82d73f30427659/src/main/resources/META-INF/resources/primefaces/datepicker/0-datepicker.js#L752) snippet. Passing "12:04" and "AM" into parseTime causes the AM to get dropped and hour: 12, minute: 4, second: null to be returned. I'm not much of a JavaScript developer, but maybe the hour needs to get set to zero? Thanks!
1.0
DatePicker: TimeOnly 12:04 a.m. (midnight plus 4 minutes) is not handled correctly - (Apologies in advance, I am new to reporting defects. I kinda just looked at what others were doing to fill in the format below.) **Describe the defect** Using the time-only 12-hour datePicker causes 12:04 a.m. to become 12:04 p.m. The input box correctly displays "12:04 AM", but when opening the time picker, it gets changed to 12:04 PM" and updates the input box with the incorrect value. GIF below: ![PrimefacesDefect](https://user-images.githubusercontent.com/38224657/101688276-ca820e80-3a39-11eb-9dde-a22135f2a66e.gif) **Environment** PF Version: 8.0.5 JSF + version: Mojarra 2.3.14 Tomcat 8.5 Affected browsers: Chrome (at least) **To Reproduce** Steps to reproduce the behavior: 1. XHTML: ```<?xml version="1.0" encoding="UTF-8"?> <!DOCTYPE html PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN" "http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd"> <html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en" lang="en" xmlns:p="http://primefaces.org/ui" xmlns:h="http://java.sun.com/jsf/html" xmlns:f="http://java.sun.com/jsf/core"> <h:head> <title>Title</title> </h:head> <body> <p:datePicker id="datePickerDemo" hourFormat="12" value="#{rTest.dateExample}" timeOnly="true"/> </body> </html> ``` 2. Backing Bean (note: same behavior with LocalTime and Date) ```import java.time.LocalTime; import javax.enterprise.context.RequestScoped; import javax.inject.Named; @RequestScoped @Named("rTest") public class RequestTest { private LocalTime dateExample = LocalTime.of(0, 4); public LocalTime getDateExample() { return dateExample; } public void setDateExample(LocalTime dateExample) { this.dateExample = dateExample; } } ``` **Commentary** The issue appears to be in [this code](https://github.com/primefaces/primefaces/blob/6c67dadd07ab2b7887f43f14da82d73f30427659/src/main/resources/META-INF/resources/primefaces/datepicker/0-datepicker.js#L752) snippet. Passing "12:04" and "AM" into parseTime causes the AM to get dropped and hour: 12, minute: 4, second: null to be returned. I'm not much of a JavaScript developer, but maybe the hour needs to get set to zero? Thanks!
defect
datepicker timeonly a m midnight plus minutes is not handled correctly apologies in advance i am new to reporting defects i kinda just looked at what others were doing to fill in the format below describe the defect using the time only hour datepicker causes a m to become p m the input box correctly displays am but when opening the time picker it gets changed to pm and updates the input box with the incorrect value gif below environment pf version jsf version mojarra tomcat affected browsers chrome at least to reproduce steps to reproduce the behavior xhtml doctype html public dtd xhtml transitional en html xmlns xml lang en lang en xmlns p xmlns h xmlns f title backing bean note same behavior with localtime and date import java time localtime import javax enterprise context requestscoped import javax inject named requestscoped named rtest public class requesttest private localtime dateexample localtime of public localtime getdateexample return dateexample public void setdateexample localtime dateexample this dateexample dateexample commentary the issue appears to be in snippet passing and am into parsetime causes the am to get dropped and hour minute second null to be returned i m not much of a javascript developer but maybe the hour needs to get set to zero thanks
1
308,141
23,234,190,018
IssuesEvent
2022-08-03 10:14:52
open-metadata/OpenMetadata
https://api.github.com/repos/open-metadata/OpenMetadata
opened
Documentation or Tutorial Requests!!
documentation
Dear developers, first of all I am very interested in this project, but I have some problems when I perform local deployment, I will be very grateful if the developers can reply to me.The tutorial of the reference documentation of this open source project is behind the code. If possible, the documentation can be updated in time. At present, I can start the service locally, but the ingestion module cannot connect to airflow through rest-api. The ingestion module uses the official image provided. Thank you very much.
1.0
Documentation or Tutorial Requests!! - Dear developers, first of all I am very interested in this project, but I have some problems when I perform local deployment, I will be very grateful if the developers can reply to me.The tutorial of the reference documentation of this open source project is behind the code. If possible, the documentation can be updated in time. At present, I can start the service locally, but the ingestion module cannot connect to airflow through rest-api. The ingestion module uses the official image provided. Thank you very much.
non_defect
documentation or tutorial requests!! dear developers first of all i am very interested in this project but i have some problems when i perform local deployment i will be very grateful if the developers can reply to me the tutorial of the reference documentation of this open source project is behind the code if possible the documentation can be updated in time at present i can start the service locally but the ingestion module cannot connect to airflow through rest api the ingestion module uses the official image provided thank you very much
0
214
2,519,740,251
IssuesEvent
2015-01-18 09:01:26
mbunkus/mtx-trac-import-test
https://api.github.com/repos/mbunkus/mtx-trac-import-test
closed
No mux aac and RV9
C: mkvmerge P: normal R: invalid T: defect
**Reported by slickbitch1 on 21 Dec 2003 13:25 UTC** Windows XP, AutoRV9, OagMachine aac 5.1 After mkvmrege says opened " ".mkv for writing, it crashes and returns error code 1073741676
1.0
No mux aac and RV9 - **Reported by slickbitch1 on 21 Dec 2003 13:25 UTC** Windows XP, AutoRV9, OagMachine aac 5.1 After mkvmrege says opened " ".mkv for writing, it crashes and returns error code 1073741676
defect
no mux aac and reported by on dec utc windows xp oagmachine aac after mkvmrege says opened mkv for writing it crashes and returns error code
1
57,795
16,065,123,067
IssuesEvent
2021-04-23 17:49:36
department-of-veterans-affairs/va.gov-cms
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
closed
Menu item should not have a default parent for any node type.
Defect Needs refining Product Support Team
**Describe the defect** **To Reproduce** As Victor.A.MCtest 1. Go to /node/add/event and enable menu item 2. note that the parent is set to Altoona **Expected behavior** There should be no default parent. Also, grep the config/sync folder for `parent: '`, you'll find this is set on on a number of `node.type.*.yml` **ACs** - [ ] Menu items for all content types should always be specified manually.
1.0
Menu item should not have a default parent for any node type. - **Describe the defect** **To Reproduce** As Victor.A.MCtest 1. Go to /node/add/event and enable menu item 2. note that the parent is set to Altoona **Expected behavior** There should be no default parent. Also, grep the config/sync folder for `parent: '`, you'll find this is set on on a number of `node.type.*.yml` **ACs** - [ ] Menu items for all content types should always be specified manually.
defect
menu item should not have a default parent for any node type describe the defect to reproduce as victor a mctest go to node add event and enable menu item note that the parent is set to altoona expected behavior there should be no default parent also grep the config sync folder for parent you ll find this is set on on a number of node type yml acs menu items for all content types should always be specified manually
1
801,911
28,506,655,274
IssuesEvent
2023-04-18 22:12:30
googleapis/python-aiplatform
https://api.github.com/repos/googleapis/python-aiplatform
closed
tests.system.aiplatform.test_matching_engine_index.TestMatchingEngine: test_create_get_list_matching_engine_index failed
type: bug priority: p1 flakybot: issue api: vertex-ai
Note: #1932 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: d75322c24d1b802c77493550fa08bfffdf3ec8fb buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f382ab13-7b18-4cf4-becd-148dae7c3389), [Sponge](http://sponge2/f382ab13-7b18-4cf4-becd-148dae7c3389) status: failed <details><summary>Test output</summary><br><pre>args = (name: "projects/580378083368/locations/us-central1/indexes/2091385465239240704" ,) kwargs = {'metadata': [('x-goog-request-params', 'name=projects/580378083368/locations/us-central1/indexes/2091385465239240704'), ('x-goog-api-client', 'model-builder/1.24.0 gl-python/3.8.13 grpc/1.47.0 gax/1.32.0 gapic/1.24.0')]} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:67: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f6f51674df0> request = name: "projects/580378083368/locations/us-central1/indexes/2091385465239240704" timeout = None metadata = [('x-goog-request-params', 'name=projects/580378083368/locations/us-central1/indexes/2091385465239240704'), ('x-goog-api-client', 'model-builder/1.24.0 gl-python/3.8.13 grpc/1.47.0 gax/1.32.0 gapic/1.24.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:946: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7f6f516880d0> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f6f51588bc0> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _MultiThreadedRendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _InactiveRpcError(state) E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: E status = StatusCode.FAILED_PRECONDITION E details = "The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136." E debug_error_string = "{"created":"@1681797321.728722218","description":"Error received from peer ipv4:142.250.107.95:443","file":"src/core/lib/surface/call.cc","file_line":966,"grpc_message":"The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.","grpc_status":9}" E > .nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:849: _InactiveRpcError The above exception was the direct cause of the following exception: self = <tests.system.aiplatform.test_matching_engine_index.TestMatchingEngine object at 0x7f6f54dfd130> shared_state = {'index': <google.cloud.aiplatform.matching_engine.matching_engine_index.MatchingEngineIndex object at 0x7f6f52beafd0>...ect at 0x7f6f5168f130> resource name: projects/580378083368/locations/us-central1/indexEndpoints/3989652708175904768]} def test_create_get_list_matching_engine_index(self, shared_state): aiplatform.init( project=e2e_base._PROJECT, location=e2e_base._LOCATION, ) # Create an index index = aiplatform.MatchingEngineIndex.create_tree_ah_index( display_name=_TEST_INDEX_DISPLAY_NAME, contents_delta_uri=_TEST_CONTENTS_DELTA_URI, dimensions=_TEST_INDEX_CONFIG_DIMENSIONS, approximate_neighbors_count=_TEST_INDEX_APPROXIMATE_NEIGHBORS_COUNT, distance_measure_type=_TEST_INDEX_DISTANCE_MEASURE_TYPE, leaf_node_embedding_count=_TEST_LEAF_NODE_EMBEDDING_COUNT, leaf_nodes_to_search_percent=_TEST_LEAF_NODES_TO_SEARCH_PERCENT, description=_TEST_INDEX_DESCRIPTION, labels=_TEST_LABELS, ) shared_state["resources"] = [index] shared_state["index"] = index shared_state["index_name"] = index.resource_name # Verify that the retrieved index is the same get_index = aiplatform.MatchingEngineIndex(index_name=index.resource_name) assert index.resource_name == get_index.resource_name # Create index and check that it is listed list_indexes = aiplatform.MatchingEngineIndex.list() assert get_index.resource_name in [ index.resource_name for index in list_indexes ] # Update the index metadata updated_index = get_index.update_metadata( display_name=_TEST_DISPLAY_NAME_UPDATE, description=_TEST_DESCRIPTION_UPDATE, labels=_TEST_LABELS_UPDATE, ) assert updated_index.name == get_index.name # TODO: Reinstate assertions once b/220005272 is fixed. # assert updated_index.display_name == _TEST_DISPLAY_NAME_UPDATE # assert updated_index.description == _TEST_DESCRIPTION_UPDATE # assert updated_index.labels == _TEST_LABELS_UPDATE # Update the index embeddings updated_index = get_index.update_embeddings( contents_delta_uri=_TEST_CONTENTS_DELTA_URI_UPDATE, is_complete_overwrite=_TEST_IS_COMPLETE_OVERWRITE, ) assert updated_index.name == get_index.name # Create endpoint and check that it is listed my_index_endpoint = aiplatform.MatchingEngineIndexEndpoint.create( display_name=_TEST_INDEX_ENDPOINT_DISPLAY_NAME, description=_TEST_INDEX_ENDPOINT_DESCRIPTION, network=e2e_base._VPC_NETWORK_URI, labels=_TEST_LABELS, ) assert my_index_endpoint.resource_name in [ index_endpoint.resource_name for index_endpoint in aiplatform.MatchingEngineIndexEndpoint.list() ] assert my_index_endpoint.labels == _TEST_LABELS assert my_index_endpoint.display_name == _TEST_INDEX_ENDPOINT_DISPLAY_NAME assert my_index_endpoint.description == _TEST_INDEX_ENDPOINT_DESCRIPTION # Create endpoint and check that it is listed public_index_endpoint = aiplatform.MatchingEngineIndexEndpoint.create( display_name=_TEST_PUBLIC_INDEX_ENDPOINT_DISPLAY_NAME, description=_TEST_PUBLIC_INDEX_ENDPOINT_DESCRIPTION, public_endpoint_enabled=True, labels=_TEST_LABELS, ) assert public_index_endpoint.resource_name in [ index_endpoint.resource_name for index_endpoint in aiplatform.MatchingEngineIndexEndpoint.list() ] assert public_index_endpoint.labels == _TEST_LABELS assert ( public_index_endpoint.display_name == _TEST_PUBLIC_INDEX_ENDPOINT_DISPLAY_NAME ) assert ( public_index_endpoint.description == _TEST_PUBLIC_INDEX_ENDPOINT_DESCRIPTION ) shared_state["resources"].append(my_index_endpoint) # Deploy endpoint my_index_endpoint = my_index_endpoint.deploy_index( index=index, deployed_index_id=_TEST_DEPLOYED_INDEX_ID, display_name=_TEST_DEPLOYED_INDEX_DISPLAY_NAME, ) # Deploy public endpoint public_index_endpoint = public_index_endpoint.deploy_index( index=index, deployed_index_id=_TEST_DEPLOYED_INDEX_ID_PUBLIC, display_name=_TEST_DEPLOYED_INDEX_DISPLAY_NAME_PUBLIC, min_replica_count=_TEST_MIN_REPLICA_COUNT_UPDATED, max_replica_count=_TEST_MAX_REPLICA_COUNT_UPDATED, ) # Update endpoint updated_index_endpoint = my_index_endpoint.update( display_name=_TEST_DISPLAY_NAME_UPDATE, description=_TEST_DESCRIPTION_UPDATE, labels=_TEST_LABELS_UPDATE, ) assert updated_index_endpoint.labels == _TEST_LABELS_UPDATE assert updated_index_endpoint.display_name == _TEST_DISPLAY_NAME_UPDATE assert updated_index_endpoint.description == _TEST_DESCRIPTION_UPDATE # Mutate deployed index my_index_endpoint.mutate_deployed_index( deployed_index_id=_TEST_DEPLOYED_INDEX_ID, min_replica_count=_TEST_MIN_REPLICA_COUNT_UPDATED, max_replica_count=_TEST_MAX_REPLICA_COUNT_UPDATED, ) # deployed index on private endpoint. deployed_index = my_index_endpoint.deployed_indexes[0] assert deployed_index.id == _TEST_DEPLOYED_INDEX_ID assert deployed_index.index == index.resource_name assert ( deployed_index.automatic_resources.min_replica_count == _TEST_MIN_REPLICA_COUNT_UPDATED ) assert ( deployed_index.automatic_resources.max_replica_count == _TEST_MAX_REPLICA_COUNT_UPDATED ) # deployed index on public endpoint. deployed_index_public = public_index_endpoint.deployed_indexes[0] assert deployed_index_public.id == _TEST_DEPLOYED_INDEX_ID_PUBLIC assert deployed_index_public.index == index.resource_name assert ( deployed_index_public.automatic_resources.min_replica_count == _TEST_MIN_REPLICA_COUNT_UPDATED ) assert ( deployed_index_public.automatic_resources.max_replica_count == _TEST_MAX_REPLICA_COUNT_UPDATED ) # TODO: Test `my_index_endpoint.match` request. This requires running this test in a VPC. # results = my_index_endpoint.match( # deployed_index_id=_TEST_DEPLOYED_INDEX_ID, queries=[_TEST_MATCH_QUERY] # ) # assert results[0][0].id == 870 # TODO: Test `my_index_endpoint.match` with filter. # This requires uploading a new content of the Matching Engine Index to Cloud Storage. # results = my_index_endpoint.match( # deployed_index_id=_TEST_DEPLOYED_INDEX_ID, # queries=[_TEST_MATCH_QUERY], # num_neighbors=1, # filter=_TEST_FILTER, # ) # assert results[0][0].id == 9999 # Undeploy index my_index_endpoint = my_index_endpoint.undeploy_index( deployed_index_id=deployed_index.id ) # Delete index and check that it is no longer listed > index.delete() tests/system/aiplatform/test_matching_engine_index.py:356: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/aiplatform/base.py:814: in wrapper return method(*args, **kwargs) google/cloud/aiplatform/base.py:1269: in delete lro = getattr(self.api_client, self._delete_method)(name=self.resource_name) google/cloud/aiplatform_v1/services/index_service/client.py:1071: in delete_index response = rpc( .nox/system-3-8/lib/python3.8/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_InactiveRpcError of RPC that terminated with: status = StatusCode.FAILED_PRECONDITION details = "The Index "projec... IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.","grpc_status":9}" > > ??? E google.api_core.exceptions.FailedPrecondition: 400 The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136. <string>:3: FailedPrecondition</pre></details>
1.0
tests.system.aiplatform.test_matching_engine_index.TestMatchingEngine: test_create_get_list_matching_engine_index failed - Note: #1932 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: d75322c24d1b802c77493550fa08bfffdf3ec8fb buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f382ab13-7b18-4cf4-becd-148dae7c3389), [Sponge](http://sponge2/f382ab13-7b18-4cf4-becd-148dae7c3389) status: failed <details><summary>Test output</summary><br><pre>args = (name: "projects/580378083368/locations/us-central1/indexes/2091385465239240704" ,) kwargs = {'metadata': [('x-goog-request-params', 'name=projects/580378083368/locations/us-central1/indexes/2091385465239240704'), ('x-goog-api-client', 'model-builder/1.24.0 gl-python/3.8.13 grpc/1.47.0 gax/1.32.0 gapic/1.24.0')]} @six.wraps(callable_) def error_remapped_callable(*args, **kwargs): try: > return callable_(*args, **kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:67: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f6f51674df0> request = name: "projects/580378083368/locations/us-central1/indexes/2091385465239240704" timeout = None metadata = [('x-goog-request-params', 'name=projects/580378083368/locations/us-central1/indexes/2091385465239240704'), ('x-goog-api-client', 'model-builder/1.24.0 gl-python/3.8.13 grpc/1.47.0 gax/1.32.0 gapic/1.24.0')] credentials = None, wait_for_ready = None, compression = None def __call__(self, request, timeout=None, metadata=None, credentials=None, wait_for_ready=None, compression=None): state, call, = self._blocking(request, timeout, metadata, credentials, wait_for_ready, compression) > return _end_unary_response_blocking(state, call, False, None) .nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:946: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ state = <grpc._channel._RPCState object at 0x7f6f516880d0> call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f6f51588bc0> with_call = False, deadline = None def _end_unary_response_blocking(state, call, with_call, deadline): if state.code is grpc.StatusCode.OK: if with_call: rendezvous = _MultiThreadedRendezvous(state, call, None, deadline) return state.response, rendezvous else: return state.response else: > raise _InactiveRpcError(state) E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: E status = StatusCode.FAILED_PRECONDITION E details = "The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136." E debug_error_string = "{"created":"@1681797321.728722218","description":"Error received from peer ipv4:142.250.107.95:443","file":"src/core/lib/surface/call.cc","file_line":966,"grpc_message":"The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.","grpc_status":9}" E > .nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:849: _InactiveRpcError The above exception was the direct cause of the following exception: self = <tests.system.aiplatform.test_matching_engine_index.TestMatchingEngine object at 0x7f6f54dfd130> shared_state = {'index': <google.cloud.aiplatform.matching_engine.matching_engine_index.MatchingEngineIndex object at 0x7f6f52beafd0>...ect at 0x7f6f5168f130> resource name: projects/580378083368/locations/us-central1/indexEndpoints/3989652708175904768]} def test_create_get_list_matching_engine_index(self, shared_state): aiplatform.init( project=e2e_base._PROJECT, location=e2e_base._LOCATION, ) # Create an index index = aiplatform.MatchingEngineIndex.create_tree_ah_index( display_name=_TEST_INDEX_DISPLAY_NAME, contents_delta_uri=_TEST_CONTENTS_DELTA_URI, dimensions=_TEST_INDEX_CONFIG_DIMENSIONS, approximate_neighbors_count=_TEST_INDEX_APPROXIMATE_NEIGHBORS_COUNT, distance_measure_type=_TEST_INDEX_DISTANCE_MEASURE_TYPE, leaf_node_embedding_count=_TEST_LEAF_NODE_EMBEDDING_COUNT, leaf_nodes_to_search_percent=_TEST_LEAF_NODES_TO_SEARCH_PERCENT, description=_TEST_INDEX_DESCRIPTION, labels=_TEST_LABELS, ) shared_state["resources"] = [index] shared_state["index"] = index shared_state["index_name"] = index.resource_name # Verify that the retrieved index is the same get_index = aiplatform.MatchingEngineIndex(index_name=index.resource_name) assert index.resource_name == get_index.resource_name # Create index and check that it is listed list_indexes = aiplatform.MatchingEngineIndex.list() assert get_index.resource_name in [ index.resource_name for index in list_indexes ] # Update the index metadata updated_index = get_index.update_metadata( display_name=_TEST_DISPLAY_NAME_UPDATE, description=_TEST_DESCRIPTION_UPDATE, labels=_TEST_LABELS_UPDATE, ) assert updated_index.name == get_index.name # TODO: Reinstate assertions once b/220005272 is fixed. # assert updated_index.display_name == _TEST_DISPLAY_NAME_UPDATE # assert updated_index.description == _TEST_DESCRIPTION_UPDATE # assert updated_index.labels == _TEST_LABELS_UPDATE # Update the index embeddings updated_index = get_index.update_embeddings( contents_delta_uri=_TEST_CONTENTS_DELTA_URI_UPDATE, is_complete_overwrite=_TEST_IS_COMPLETE_OVERWRITE, ) assert updated_index.name == get_index.name # Create endpoint and check that it is listed my_index_endpoint = aiplatform.MatchingEngineIndexEndpoint.create( display_name=_TEST_INDEX_ENDPOINT_DISPLAY_NAME, description=_TEST_INDEX_ENDPOINT_DESCRIPTION, network=e2e_base._VPC_NETWORK_URI, labels=_TEST_LABELS, ) assert my_index_endpoint.resource_name in [ index_endpoint.resource_name for index_endpoint in aiplatform.MatchingEngineIndexEndpoint.list() ] assert my_index_endpoint.labels == _TEST_LABELS assert my_index_endpoint.display_name == _TEST_INDEX_ENDPOINT_DISPLAY_NAME assert my_index_endpoint.description == _TEST_INDEX_ENDPOINT_DESCRIPTION # Create endpoint and check that it is listed public_index_endpoint = aiplatform.MatchingEngineIndexEndpoint.create( display_name=_TEST_PUBLIC_INDEX_ENDPOINT_DISPLAY_NAME, description=_TEST_PUBLIC_INDEX_ENDPOINT_DESCRIPTION, public_endpoint_enabled=True, labels=_TEST_LABELS, ) assert public_index_endpoint.resource_name in [ index_endpoint.resource_name for index_endpoint in aiplatform.MatchingEngineIndexEndpoint.list() ] assert public_index_endpoint.labels == _TEST_LABELS assert ( public_index_endpoint.display_name == _TEST_PUBLIC_INDEX_ENDPOINT_DISPLAY_NAME ) assert ( public_index_endpoint.description == _TEST_PUBLIC_INDEX_ENDPOINT_DESCRIPTION ) shared_state["resources"].append(my_index_endpoint) # Deploy endpoint my_index_endpoint = my_index_endpoint.deploy_index( index=index, deployed_index_id=_TEST_DEPLOYED_INDEX_ID, display_name=_TEST_DEPLOYED_INDEX_DISPLAY_NAME, ) # Deploy public endpoint public_index_endpoint = public_index_endpoint.deploy_index( index=index, deployed_index_id=_TEST_DEPLOYED_INDEX_ID_PUBLIC, display_name=_TEST_DEPLOYED_INDEX_DISPLAY_NAME_PUBLIC, min_replica_count=_TEST_MIN_REPLICA_COUNT_UPDATED, max_replica_count=_TEST_MAX_REPLICA_COUNT_UPDATED, ) # Update endpoint updated_index_endpoint = my_index_endpoint.update( display_name=_TEST_DISPLAY_NAME_UPDATE, description=_TEST_DESCRIPTION_UPDATE, labels=_TEST_LABELS_UPDATE, ) assert updated_index_endpoint.labels == _TEST_LABELS_UPDATE assert updated_index_endpoint.display_name == _TEST_DISPLAY_NAME_UPDATE assert updated_index_endpoint.description == _TEST_DESCRIPTION_UPDATE # Mutate deployed index my_index_endpoint.mutate_deployed_index( deployed_index_id=_TEST_DEPLOYED_INDEX_ID, min_replica_count=_TEST_MIN_REPLICA_COUNT_UPDATED, max_replica_count=_TEST_MAX_REPLICA_COUNT_UPDATED, ) # deployed index on private endpoint. deployed_index = my_index_endpoint.deployed_indexes[0] assert deployed_index.id == _TEST_DEPLOYED_INDEX_ID assert deployed_index.index == index.resource_name assert ( deployed_index.automatic_resources.min_replica_count == _TEST_MIN_REPLICA_COUNT_UPDATED ) assert ( deployed_index.automatic_resources.max_replica_count == _TEST_MAX_REPLICA_COUNT_UPDATED ) # deployed index on public endpoint. deployed_index_public = public_index_endpoint.deployed_indexes[0] assert deployed_index_public.id == _TEST_DEPLOYED_INDEX_ID_PUBLIC assert deployed_index_public.index == index.resource_name assert ( deployed_index_public.automatic_resources.min_replica_count == _TEST_MIN_REPLICA_COUNT_UPDATED ) assert ( deployed_index_public.automatic_resources.max_replica_count == _TEST_MAX_REPLICA_COUNT_UPDATED ) # TODO: Test `my_index_endpoint.match` request. This requires running this test in a VPC. # results = my_index_endpoint.match( # deployed_index_id=_TEST_DEPLOYED_INDEX_ID, queries=[_TEST_MATCH_QUERY] # ) # assert results[0][0].id == 870 # TODO: Test `my_index_endpoint.match` with filter. # This requires uploading a new content of the Matching Engine Index to Cloud Storage. # results = my_index_endpoint.match( # deployed_index_id=_TEST_DEPLOYED_INDEX_ID, # queries=[_TEST_MATCH_QUERY], # num_neighbors=1, # filter=_TEST_FILTER, # ) # assert results[0][0].id == 9999 # Undeploy index my_index_endpoint = my_index_endpoint.undeploy_index( deployed_index_id=deployed_index.id ) # Delete index and check that it is no longer listed > index.delete() tests/system/aiplatform/test_matching_engine_index.py:356: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ google/cloud/aiplatform/base.py:814: in wrapper return method(*args, **kwargs) google/cloud/aiplatform/base.py:1269: in delete lro = getattr(self.api_client, self._delete_method)(name=self.resource_name) google/cloud/aiplatform_v1/services/index_service/client.py:1071: in delete_index response = rpc( .nox/system-3-8/lib/python3.8/site-packages/google/api_core/gapic_v1/method.py:145: in __call__ return wrapped_func(*args, **kwargs) .nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ value = None from_value = <_InactiveRpcError of RPC that terminated with: status = StatusCode.FAILED_PRECONDITION details = "The Index "projec... IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.","grpc_status":9}" > > ??? E google.api_core.exceptions.FailedPrecondition: 400 The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136. <string>:3: FailedPrecondition</pre></details>
non_defect
tests system aiplatform test matching engine index testmatchingengine test create get list matching engine index failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output args name projects locations us indexes kwargs metadata six wraps callable def error remapped callable args kwargs try return callable args kwargs nox system lib site packages google api core grpc helpers py self request name projects locations us indexes timeout none metadata credentials none wait for ready none compression none def call self request timeout none metadata none credentials none wait for ready none compression none state call self blocking request timeout metadata credentials wait for ready compression return end unary response blocking state call false none nox system lib site packages grpc channel py state call with call false deadline none def end unary response blocking state call with call deadline if state code is grpc statuscode ok if with call rendezvous multithreadedrendezvous state call none deadline return state response rendezvous else return state response else raise inactiverpcerror state e grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with e status statuscode failed precondition e details the index projects locations us indexes is deployed or being deployed at the following indexendpoint s projects locations us indexendpoints e debug error string created description error received from peer file src core lib surface call cc file line grpc message the index projects locations us indexes is deployed or being deployed at the following indexendpoint s projects locations us indexendpoints grpc status e nox system lib site packages grpc channel py inactiverpcerror the above exception was the direct cause of the following exception self shared state index ect at resource name projects locations us indexendpoints def test create get list matching engine index self shared state aiplatform init project base project location base location create an index index aiplatform matchingengineindex create tree ah index display name test index display name contents delta uri test contents delta uri dimensions test index config dimensions approximate neighbors count test index approximate neighbors count distance measure type test index distance measure type leaf node embedding count test leaf node embedding count leaf nodes to search percent test leaf nodes to search percent description test index description labels test labels shared state shared state index shared state index resource name verify that the retrieved index is the same get index aiplatform matchingengineindex index name index resource name assert index resource name get index resource name create index and check that it is listed list indexes aiplatform matchingengineindex list assert get index resource name in index resource name for index in list indexes update the index metadata updated index get index update metadata display name test display name update description test description update labels test labels update assert updated index name get index name todo reinstate assertions once b is fixed assert updated index display name test display name update assert updated index description test description update assert updated index labels test labels update update the index embeddings updated index get index update embeddings contents delta uri test contents delta uri update is complete overwrite test is complete overwrite assert updated index name get index name create endpoint and check that it is listed my index endpoint aiplatform matchingengineindexendpoint create display name test index endpoint display name description test index endpoint description network base vpc network uri labels test labels assert my index endpoint resource name in index endpoint resource name for index endpoint in aiplatform matchingengineindexendpoint list assert my index endpoint labels test labels assert my index endpoint display name test index endpoint display name assert my index endpoint description test index endpoint description create endpoint and check that it is listed public index endpoint aiplatform matchingengineindexendpoint create display name test public index endpoint display name description test public index endpoint description public endpoint enabled true labels test labels assert public index endpoint resource name in index endpoint resource name for index endpoint in aiplatform matchingengineindexendpoint list assert public index endpoint labels test labels assert public index endpoint display name test public index endpoint display name assert public index endpoint description test public index endpoint description shared state append my index endpoint deploy endpoint my index endpoint my index endpoint deploy index index index deployed index id test deployed index id display name test deployed index display name deploy public endpoint public index endpoint public index endpoint deploy index index index deployed index id test deployed index id public display name test deployed index display name public min replica count test min replica count updated max replica count test max replica count updated update endpoint updated index endpoint my index endpoint update display name test display name update description test description update labels test labels update assert updated index endpoint labels test labels update assert updated index endpoint display name test display name update assert updated index endpoint description test description update mutate deployed index my index endpoint mutate deployed index deployed index id test deployed index id min replica count test min replica count updated max replica count test max replica count updated deployed index on private endpoint deployed index my index endpoint deployed indexes assert deployed index id test deployed index id assert deployed index index index resource name assert deployed index automatic resources min replica count test min replica count updated assert deployed index automatic resources max replica count test max replica count updated deployed index on public endpoint deployed index public public index endpoint deployed indexes assert deployed index public id test deployed index id public assert deployed index public index index resource name assert deployed index public automatic resources min replica count test min replica count updated assert deployed index public automatic resources max replica count test max replica count updated todo test my index endpoint match request this requires running this test in a vpc results my index endpoint match deployed index id test deployed index id queries assert results id todo test my index endpoint match with filter this requires uploading a new content of the matching engine index to cloud storage results my index endpoint match deployed index id test deployed index id queries num neighbors filter test filter assert results id undeploy index my index endpoint my index endpoint undeploy index deployed index id deployed index id delete index and check that it is no longer listed index delete tests system aiplatform test matching engine index py google cloud aiplatform base py in wrapper return method args kwargs google cloud aiplatform base py in delete lro getattr self api client self delete method name self resource name google cloud aiplatform services index service client py in delete index response rpc nox system lib site packages google api core gapic method py in call return wrapped func args kwargs nox system lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value inactiverpcerror of rpc that terminated with status statuscode failed precondition details the index projec indexendpoint s projects locations us indexendpoints grpc status e google api core exceptions failedprecondition the index projects locations us indexes is deployed or being deployed at the following indexendpoint s projects locations us indexendpoints failedprecondition
0
233,607
7,700,728,397
IssuesEvent
2018-05-20 06:00:43
python/mypy
https://api.github.com/repos/python/mypy
closed
"SupportsInt" doesn't work with Union types
bug false-positive priority-1-normal topic-union-types
The following shouldn't be an error: `error: Argument 1 to "int" has incompatible type "Union[unicode, int]"; expected "SupportsInt"`
1.0
"SupportsInt" doesn't work with Union types - The following shouldn't be an error: `error: Argument 1 to "int" has incompatible type "Union[unicode, int]"; expected "SupportsInt"`
non_defect
supportsint doesn t work with union types the following shouldn t be an error error argument to int has incompatible type union expected supportsint
0
142,373
19,090,476,180
IssuesEvent
2021-11-29 11:30:31
sultanabubaker/NuGet_Project_SDK_NonSDK
https://api.github.com/repos/sultanabubaker/NuGet_Project_SDK_NonSDK
opened
CVE-2019-1428 (High) detected in multiple libraries
security vulnerability
## CVE-2019-1428 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ChakraCore-1.8.1.0.dll</b>, <b>microsoft.chakracore.1.8.1.nupkg</b>, <b>ChakraCore-1.8.1.0.dll</b>, <b>ChakraCore-1.8.1.0.dll</b></p></summary> <p> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/net5.0/runtimes/win8-arm/native/ChakraCore.dll,/SDK/bin/Debug/netcoreapp3.1/runtimes/win8-arm/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>microsoft.chakracore.1.8.1.nupkg</b></p></summary> <p>ChakraCore is the core part of the Chakra Javascript engine that powers Microsoft Edge.</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to dependency file: NuGet_Project_SDK_NonSDK/SDK/SDK.csproj</p> <p>Path to vulnerable library: crosoft.chakracore/1.8.1/microsoft.chakracore.1.8.1.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.chakracore.1.8.1.nupkg** (Vulnerable Library) </details> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x64/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x64/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x86/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x86/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/NuGet_Project_SDK_NonSDK/commit/2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927">2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in Microsoft Edge (HTML-based), aka 'Scripting Engine Memory Corruption Vulnerability'. This CVE ID is unique from CVE-2019-1426, CVE-2019-1427, CVE-2019-1429. <p>Publish Date: 2019-11-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1428>CVE-2019-1428</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/Microsoft/ChakraCore/wiki/Roadmap">https://github.com/Microsoft/ChakraCore/wiki/Roadmap</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: Microsoft.ChakraCore - 1.11.15</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true},{"packageType":"Nuget","packageName":"Microsoft.ChakraCore","packageVersion":"1.8.1","packageFilePaths":["/SDK/SDK.csproj"],"isTransitiveDependency":false,"dependencyTree":"Microsoft.ChakraCore:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":false},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-1428","vulnerabilityDetails":"A remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in Microsoft Edge (HTML-based), aka \u0027Scripting Engine Memory Corruption Vulnerability\u0027. This CVE ID is unique from CVE-2019-1426, CVE-2019-1427, CVE-2019-1429.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1428","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2019-1428 (High) detected in multiple libraries - ## CVE-2019-1428 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ChakraCore-1.8.1.0.dll</b>, <b>microsoft.chakracore.1.8.1.nupkg</b>, <b>ChakraCore-1.8.1.0.dll</b>, <b>ChakraCore-1.8.1.0.dll</b></p></summary> <p> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/net5.0/runtimes/win8-arm/native/ChakraCore.dll,/SDK/bin/Debug/netcoreapp3.1/runtimes/win8-arm/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>microsoft.chakracore.1.8.1.nupkg</b></p></summary> <p>ChakraCore is the core part of the Chakra Javascript engine that powers Microsoft Edge.</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to dependency file: NuGet_Project_SDK_NonSDK/SDK/SDK.csproj</p> <p>Path to vulnerable library: crosoft.chakracore/1.8.1/microsoft.chakracore.1.8.1.nupkg</p> <p> Dependency Hierarchy: - :x: **microsoft.chakracore.1.8.1.nupkg** (Vulnerable Library) </details> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x64/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x64/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary> <p>Microsoft ? Chakra Core</p> <p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p> <p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x86/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x86/native/ChakraCore.dll</p> <p> Dependency Hierarchy: - :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/NuGet_Project_SDK_NonSDK/commit/2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927">2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in Microsoft Edge (HTML-based), aka 'Scripting Engine Memory Corruption Vulnerability'. This CVE ID is unique from CVE-2019-1426, CVE-2019-1427, CVE-2019-1429. <p>Publish Date: 2019-11-12 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1428>CVE-2019-1428</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/Microsoft/ChakraCore/wiki/Roadmap">https://github.com/Microsoft/ChakraCore/wiki/Roadmap</a></p> <p>Release Date: 2020-08-24</p> <p>Fix Resolution: Microsoft.ChakraCore - 1.11.15</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true},{"packageType":"Nuget","packageName":"Microsoft.ChakraCore","packageVersion":"1.8.1","packageFilePaths":["/SDK/SDK.csproj"],"isTransitiveDependency":false,"dependencyTree":"Microsoft.ChakraCore:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":false},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-1428","vulnerabilityDetails":"A remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in Microsoft Edge (HTML-based), aka \u0027Scripting Engine Memory Corruption Vulnerability\u0027. This CVE ID is unique from CVE-2019-1426, CVE-2019-1427, CVE-2019-1429.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1428","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries chakracore dll microsoft chakracore nupkg chakracore dll chakracore dll chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes arm native chakracore dll sdk bin debug runtimes arm native chakracore dll dependency hierarchy x chakracore dll vulnerable library microsoft chakracore nupkg chakracore is the core part of the chakra javascript engine that powers microsoft edge library home page a href path to dependency file nuget project sdk nonsdk sdk sdk csproj path to vulnerable library crosoft chakracore microsoft chakracore nupkg dependency hierarchy x microsoft chakracore nupkg vulnerable library chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes native chakracore dll sdk bin debug runtimes native chakracore dll dependency hierarchy x chakracore dll vulnerable library chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes native chakracore dll sdk bin debug runtimes native chakracore dll dependency hierarchy x chakracore dll vulnerable library found in head commit a href found in base branch master vulnerability details a remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in microsoft edge html based aka scripting engine memory corruption vulnerability this cve id is unique from cve cve cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft chakracore isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion microsoft chakracore isbinary true packagetype nuget packagename microsoft chakracore packageversion packagefilepaths istransitivedependency false dependencytree microsoft chakracore isminimumfixversionavailable true minimumfixversion microsoft chakracore isbinary false packagetype net packagename chakracore packageversion packagefilepaths istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion microsoft chakracore isbinary true packagetype net packagename chakracore packageversion packagefilepaths istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion microsoft chakracore isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails a remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in microsoft edge html based aka engine memory corruption vulnerability this cve id is unique from cve cve cve vulnerabilityurl
0
161,713
12,559,414,919
IssuesEvent
2020-06-07 18:53:01
GTNewHorizons/GT-New-Horizons-Modpack
https://api.github.com/repos/GTNewHorizons/GT-New-Horizons-Modpack
closed
Crash with 512V Creative Energy Buffer
FixedInDev GT++ Issue need to be tested
#### Which modpack version are you using? 2.0.8.3 # #### What did you try to do, and what did you expect to happen? Change the output of the Buffer with a screwdriver, to cycle back to ULV from MAX. # #### What happened instead? (Attach screenshots if needed) It crashed. # #### What do you suggest instead/what changes do you propose? Fix the crash. Do you need a log? It seems easy to reproduce so I'm not sure you needed one.
1.0
Crash with 512V Creative Energy Buffer - #### Which modpack version are you using? 2.0.8.3 # #### What did you try to do, and what did you expect to happen? Change the output of the Buffer with a screwdriver, to cycle back to ULV from MAX. # #### What happened instead? (Attach screenshots if needed) It crashed. # #### What do you suggest instead/what changes do you propose? Fix the crash. Do you need a log? It seems easy to reproduce so I'm not sure you needed one.
non_defect
crash with creative energy buffer which modpack version are you using what did you try to do and what did you expect to happen change the output of the buffer with a screwdriver to cycle back to ulv from max what happened instead attach screenshots if needed it crashed what do you suggest instead what changes do you propose fix the crash do you need a log it seems easy to reproduce so i m not sure you needed one
0
193,061
6,877,817,527
IssuesEvent
2017-11-20 09:36:07
OpenNebula/one
https://api.github.com/repos/OpenNebula/one
opened
euca-describe-instances returns UnsupportedOperation
Priority: Normal Status: Pending Tracker: Bug
--- Author Name: **Valerio Schiavoni** (Valerio Schiavoni) Original Redmine Issue: 4244, https://dev.opennebula.org/issues/4244 Original Date: 2015-12-14 --- We get this error in scone-server.log @ Mon Dec 14 14:08:04 2015 [E]: Unexpected error executing code for particular method, detected by Xmlrpc-c method registry code. Method did not fail; rather, it did not complete at all. 25637787-byte supposed UTF-8 string is not valid UTF-8. Unrecognized UTF-8 initial byte value 0x93 @
1.0
euca-describe-instances returns UnsupportedOperation - --- Author Name: **Valerio Schiavoni** (Valerio Schiavoni) Original Redmine Issue: 4244, https://dev.opennebula.org/issues/4244 Original Date: 2015-12-14 --- We get this error in scone-server.log @ Mon Dec 14 14:08:04 2015 [E]: Unexpected error executing code for particular method, detected by Xmlrpc-c method registry code. Method did not fail; rather, it did not complete at all. 25637787-byte supposed UTF-8 string is not valid UTF-8. Unrecognized UTF-8 initial byte value 0x93 @
non_defect
euca describe instances returns unsupportedoperation author name valerio schiavoni valerio schiavoni original redmine issue original date we get this error in scone server log mon dec unexpected error executing code for particular method detected by xmlrpc c method registry code method did not fail rather it did not complete at all byte supposed utf string is not valid utf unrecognized utf initial byte value
0
44,960
12,493,538,777
IssuesEvent
2020-06-01 09:27:44
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
User Code Deployment it’s not capable of resolving the package private classes and interfaces
Team: Core Type: Critical Type: Defect
User Code Deployment it’s not capable of resolving the package private classes and interfaces. We found this while developing the Map Store Support for the Hazelcast Cloud. When some class inside the Hikari Connection Pool references package a private stuff - it fails with the IllegalAccessError on the attempt to load it: ``` { "errorMessage": "class com.zaxxer.hikari.pool.PoolBase$NopMetricsTrackerDelegate cannot access its superinterface com.zaxxer.hikari.pool.PoolBase$IMetricsTrackerDelegate", "errorType": "java.lang.IllegalAccessError", "stackTrace": [ "java.lang.ClassLoader.defineClass1(Native Method)", "java.lang.ClassLoader.defineClass(ClassLoader.java:763)", "java.lang.ClassLoader.defineClass(ClassLoader.java:642)", "com.hazelcast.internal.usercodedeployment.impl.ClassSource.define(ClassSource.java:50)", "com.hazelcast.internal.usercodedeployment.impl.ClassLocator.tryToGetClassFromRemote(ClassLocator.java:163)", "com.hazelcast.internal.usercodedeployment.impl.ClassLocator.handleClassNotFoundException(ClassLocator.java:95)", ``` **Now, this fully blocks the end-to-end scenario working at our side.** **Expected behavior** The expected behavior is that the uploaded JAR, with all its dependencies, will be correctly loaded and the Map Store instance will be successfully created and initialized. **To Reproduce** See the branch with reproducer - https://github.com/hazelcast/hazelcast-cloud-code-samples/tree/hazelcast.code-deployment.issue. Steps to reproduce right here. 1. Clone the repo. 2. Checkout branch `hazelcast.code-deployment.issue` 3. cd `mapstore` 4. Open README.md **Additional context** Hazelcast ver. 3.12.6 In our case, we share the uploaded JARs through the lite members. When we inspected the source code, we noticed that for every loaded class it creates a separate class loader. So we suspect this can be causing this issue.
1.0
User Code Deployment it’s not capable of resolving the package private classes and interfaces - User Code Deployment it’s not capable of resolving the package private classes and interfaces. We found this while developing the Map Store Support for the Hazelcast Cloud. When some class inside the Hikari Connection Pool references package a private stuff - it fails with the IllegalAccessError on the attempt to load it: ``` { "errorMessage": "class com.zaxxer.hikari.pool.PoolBase$NopMetricsTrackerDelegate cannot access its superinterface com.zaxxer.hikari.pool.PoolBase$IMetricsTrackerDelegate", "errorType": "java.lang.IllegalAccessError", "stackTrace": [ "java.lang.ClassLoader.defineClass1(Native Method)", "java.lang.ClassLoader.defineClass(ClassLoader.java:763)", "java.lang.ClassLoader.defineClass(ClassLoader.java:642)", "com.hazelcast.internal.usercodedeployment.impl.ClassSource.define(ClassSource.java:50)", "com.hazelcast.internal.usercodedeployment.impl.ClassLocator.tryToGetClassFromRemote(ClassLocator.java:163)", "com.hazelcast.internal.usercodedeployment.impl.ClassLocator.handleClassNotFoundException(ClassLocator.java:95)", ``` **Now, this fully blocks the end-to-end scenario working at our side.** **Expected behavior** The expected behavior is that the uploaded JAR, with all its dependencies, will be correctly loaded and the Map Store instance will be successfully created and initialized. **To Reproduce** See the branch with reproducer - https://github.com/hazelcast/hazelcast-cloud-code-samples/tree/hazelcast.code-deployment.issue. Steps to reproduce right here. 1. Clone the repo. 2. Checkout branch `hazelcast.code-deployment.issue` 3. cd `mapstore` 4. Open README.md **Additional context** Hazelcast ver. 3.12.6 In our case, we share the uploaded JARs through the lite members. When we inspected the source code, we noticed that for every loaded class it creates a separate class loader. So we suspect this can be causing this issue.
defect
user code deployment it’s not capable of resolving the package private classes and interfaces user code deployment it’s not capable of resolving the package private classes and interfaces we found this while developing the map store support for the hazelcast cloud when some class inside the hikari connection pool references package a private stuff it fails with the illegalaccesserror on the attempt to load it errormessage class com zaxxer hikari pool poolbase nopmetricstrackerdelegate cannot access its superinterface com zaxxer hikari pool poolbase imetricstrackerdelegate errortype java lang illegalaccesserror stacktrace java lang classloader native method java lang classloader defineclass classloader java java lang classloader defineclass classloader java com hazelcast internal usercodedeployment impl classsource define classsource java com hazelcast internal usercodedeployment impl classlocator trytogetclassfromremote classlocator java com hazelcast internal usercodedeployment impl classlocator handleclassnotfoundexception classlocator java now this fully blocks the end to end scenario working at our side expected behavior the expected behavior is that the uploaded jar with all its dependencies will be correctly loaded and the map store instance will be successfully created and initialized to reproduce see the branch with reproducer steps to reproduce right here clone the repo checkout branch hazelcast code deployment issue cd mapstore open readme md additional context hazelcast ver in our case we share the uploaded jars through the lite members when we inspected the source code we noticed that for every loaded class it creates a separate class loader so we suspect this can be causing this issue
1
244,228
18,751,236,553
IssuesEvent
2021-11-05 02:24:32
AY2122S1-CS2103T-T13-2/tp
https://api.github.com/repos/AY2122S1-CS2103T-T13-2/tp
closed
[PE-D] No "uncompleted status"
documentation
Assignments only have "pending" status, not "uncompleted" status. ![image.png](https://raw.githubusercontent.com/Th-429B/ped/main/files/e16d2203-9f59-4de9-880a-a902d79dbb80.png) <!--session: 1635501358553-926ea394-739d-48cb-b26f-eaa548f33f4d--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.DocumentationBug` original: Th-429B/ped#10
1.0
[PE-D] No "uncompleted status" - Assignments only have "pending" status, not "uncompleted" status. ![image.png](https://raw.githubusercontent.com/Th-429B/ped/main/files/e16d2203-9f59-4de9-880a-a902d79dbb80.png) <!--session: 1635501358553-926ea394-739d-48cb-b26f-eaa548f33f4d--> <!--Version: Web v3.4.1--> ------------- Labels: `severity.Low` `type.DocumentationBug` original: Th-429B/ped#10
non_defect
no uncompleted status assignments only have pending status not uncompleted status labels severity low type documentationbug original th ped
0
75,839
26,090,377,439
IssuesEvent
2022-12-26 10:27:47
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: "xl" not returned if success = False for scipy.optimize.shgo
defect scipy.optimize
### Describe your issue. The lists `xl` and `funl` are not returned if the optimization did not return with `success = True`. It would seem that these lists should be populated and returned for access to the user. The code below is taken from the documentation with `n` changed to a small number such that the issue is reproduced. ### Reproducing Code Example ```python import scipy.optimize as sopt def eggholder(x): return (-(x[1] + 47.0) * np.sin(np.sqrt(abs(x[0]/2.0 + (x[1] + 47.0)))) - x[0] * np.sin(np.sqrt(abs(x[0] - (x[1] + 47.0)))) ) bounds = [(-512, 512), (-512, 512)] result = sopt.shgo(eggholder, bounds, n=2, sampling_method='sobol') ``` ### Error message ```shell result Out[27]: fun: -25.460337185286313 message: 'Failed to find a feasible minimizer point. Lowest sampling point = -25.460337185286313' nfev: 2 nit: 2 nlfev: 0 nlhev: 0 nljev: 0 success: False x: array([0., 0.]) ``` ### SciPy/NumPy/Python version information 1.9.3
1.0
BUG: "xl" not returned if success = False for scipy.optimize.shgo - ### Describe your issue. The lists `xl` and `funl` are not returned if the optimization did not return with `success = True`. It would seem that these lists should be populated and returned for access to the user. The code below is taken from the documentation with `n` changed to a small number such that the issue is reproduced. ### Reproducing Code Example ```python import scipy.optimize as sopt def eggholder(x): return (-(x[1] + 47.0) * np.sin(np.sqrt(abs(x[0]/2.0 + (x[1] + 47.0)))) - x[0] * np.sin(np.sqrt(abs(x[0] - (x[1] + 47.0)))) ) bounds = [(-512, 512), (-512, 512)] result = sopt.shgo(eggholder, bounds, n=2, sampling_method='sobol') ``` ### Error message ```shell result Out[27]: fun: -25.460337185286313 message: 'Failed to find a feasible minimizer point. Lowest sampling point = -25.460337185286313' nfev: 2 nit: 2 nlfev: 0 nlhev: 0 nljev: 0 success: False x: array([0., 0.]) ``` ### SciPy/NumPy/Python version information 1.9.3
defect
bug xl not returned if success false for scipy optimize shgo describe your issue the lists xl and funl are not returned if the optimization did not return with success true it would seem that these lists should be populated and returned for access to the user the code below is taken from the documentation with n changed to a small number such that the issue is reproduced reproducing code example python import scipy optimize as sopt def eggholder x return x np sin np sqrt abs x x x np sin np sqrt abs x x bounds result sopt shgo eggholder bounds n sampling method sobol error message shell result out fun message failed to find a feasible minimizer point lowest sampling point nfev nit nlfev nlhev nljev success false x array scipy numpy python version information
1
297,223
25,710,885,929
IssuesEvent
2022-12-07 06:28:16
ByteOtter/OtterDen
https://api.github.com/repos/ByteOtter/OtterDen
closed
[BUG:]QA/QE: Broken import statement in environment.py
bug testsuite behave/cucumber
# Found an issue? Please let us know! when trying to launch any testscenario, behave throws the error "module test_utils not found" despite the directory being where it should be. This directory holds all necessary utility functions selenium needs to set up correctly so not importing it borks the testsuite. ### Please tick the applicable box(es) regarding the kind of bug you found: - [x] Broken feature - [ ] Login/Logout system - [ ] Database broken - [ ] Post-related problem - [ ] Unusable route - [ ] UI Bug - [ ] Wrong documentation - [ ] Broken dependency ### Steps to reproduce 1. try to run behave testsuite/features/core/init.py 2. get greeted by error message ### Links This issue is connected to Issue [#53 ](https://github.com/ByteOtter/OtterDen/issues/53) ### Additional context Stacktrace: ``` Exception ModuleNotFoundError: No module named 'test_utils' Traceback (most recent call last): File "/home/chris/Dev/PrivateProjects/LogBlog/env/bin/behave", line 8, in <module> sys.exit(main()) File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/__main__.py", line 183, in main return run_behave(config) File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/__main__.py", line 127, in run_behave failed = runner.run() File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 804, in run return self.run_with_paths() File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 808, in run_with_paths self.load_hooks() File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 784, in load_hooks exec_file(hooks_path, self.hooks) File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner_util.py", line 386, in exec_file exec(code, globals_, locals_) File "testsuite/features/environment.py", line 10, in <module> from test_utils.utils_web import get_browser ModuleNotFoundError: No module named 'test_utils' ```
1.0
[BUG:]QA/QE: Broken import statement in environment.py - # Found an issue? Please let us know! when trying to launch any testscenario, behave throws the error "module test_utils not found" despite the directory being where it should be. This directory holds all necessary utility functions selenium needs to set up correctly so not importing it borks the testsuite. ### Please tick the applicable box(es) regarding the kind of bug you found: - [x] Broken feature - [ ] Login/Logout system - [ ] Database broken - [ ] Post-related problem - [ ] Unusable route - [ ] UI Bug - [ ] Wrong documentation - [ ] Broken dependency ### Steps to reproduce 1. try to run behave testsuite/features/core/init.py 2. get greeted by error message ### Links This issue is connected to Issue [#53 ](https://github.com/ByteOtter/OtterDen/issues/53) ### Additional context Stacktrace: ``` Exception ModuleNotFoundError: No module named 'test_utils' Traceback (most recent call last): File "/home/chris/Dev/PrivateProjects/LogBlog/env/bin/behave", line 8, in <module> sys.exit(main()) File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/__main__.py", line 183, in main return run_behave(config) File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/__main__.py", line 127, in run_behave failed = runner.run() File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 804, in run return self.run_with_paths() File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 808, in run_with_paths self.load_hooks() File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 784, in load_hooks exec_file(hooks_path, self.hooks) File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner_util.py", line 386, in exec_file exec(code, globals_, locals_) File "testsuite/features/environment.py", line 10, in <module> from test_utils.utils_web import get_browser ModuleNotFoundError: No module named 'test_utils' ```
non_defect
qa qe broken import statement in environment py found an issue please let us know when trying to launch any testscenario behave throws the error module test utils not found despite the directory being where it should be this directory holds all necessary utility functions selenium needs to set up correctly so not importing it borks the testsuite please tick the applicable box es regarding the kind of bug you found broken feature login logout system database broken post related problem unusable route ui bug wrong documentation broken dependency steps to reproduce try to run behave testsuite features core init py get greeted by error message links this issue is connected to issue additional context stacktrace exception modulenotfounderror no module named test utils traceback most recent call last file home chris dev privateprojects logblog env bin behave line in sys exit main file home chris dev privateprojects logblog env site packages behave main py line in main return run behave config file home chris dev privateprojects logblog env site packages behave main py line in run behave failed runner run file home chris dev privateprojects logblog env site packages behave runner py line in run return self run with paths file home chris dev privateprojects logblog env site packages behave runner py line in run with paths self load hooks file home chris dev privateprojects logblog env site packages behave runner py line in load hooks exec file hooks path self hooks file home chris dev privateprojects logblog env site packages behave runner util py line in exec file exec code globals locals file testsuite features environment py line in from test utils utils web import get browser modulenotfounderror no module named test utils
0
66,818
20,682,119,457
IssuesEvent
2022-03-10 14:47:55
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
opened
Spotlight search doesn't find unnamed rooms
T-Defect
### Steps to reproduce 1. I wanted to find my 3-way DM with Amandine and Rick 2. I searched for `Rick` (didn't show up; too many rooms). I searched for `Amandine` (didn't show up; too many rooms). I searched for `amandine rick` (nothing showed up at all) 3. I eventually searched for `Rick` and then manually expanded the results list and went through one by one until i eventually found it ### Outcome #### What did you expect? Searching for `amandine rick` should return nameless rooms where the room name has been synthesised to be 'Amandine and rick'. #### What happened instead? Nothing showed up. Turns out that searching for `amandine and rick` found it, fwiw. ### Operating system macOS ### Browser information chrome ### URL for webapp nightly ### Application version nightly ### Homeserver matrix.org ### Will you send logs? No
1.0
Spotlight search doesn't find unnamed rooms - ### Steps to reproduce 1. I wanted to find my 3-way DM with Amandine and Rick 2. I searched for `Rick` (didn't show up; too many rooms). I searched for `Amandine` (didn't show up; too many rooms). I searched for `amandine rick` (nothing showed up at all) 3. I eventually searched for `Rick` and then manually expanded the results list and went through one by one until i eventually found it ### Outcome #### What did you expect? Searching for `amandine rick` should return nameless rooms where the room name has been synthesised to be 'Amandine and rick'. #### What happened instead? Nothing showed up. Turns out that searching for `amandine and rick` found it, fwiw. ### Operating system macOS ### Browser information chrome ### URL for webapp nightly ### Application version nightly ### Homeserver matrix.org ### Will you send logs? No
defect
spotlight search doesn t find unnamed rooms steps to reproduce i wanted to find my way dm with amandine and rick i searched for rick didn t show up too many rooms i searched for amandine didn t show up too many rooms i searched for amandine rick nothing showed up at all i eventually searched for rick and then manually expanded the results list and went through one by one until i eventually found it outcome what did you expect searching for amandine rick should return nameless rooms where the room name has been synthesised to be amandine and rick what happened instead nothing showed up turns out that searching for amandine and rick found it fwiw operating system macos browser information chrome url for webapp nightly application version nightly homeserver matrix org will you send logs no
1
79,429
28,240,719,951
IssuesEvent
2023-04-06 06:54:41
line/armeria
https://api.github.com/repos/line/armeria
closed
Handle exceptions thrown by Servlet
defect
> Hi folks, we recently encountered an issue where Armeria (we are on Armeria 1.16.2) won’t finish collecting the response information for a RequestLog when the corresponding HttpResponse is closed without any data written. As a result, RequestLogAccess.whenComplete() never completes after closing the HttpResponse. A simple reproduction can be found [here](https://gist.github.com/linwuz-db/2c7adae5a546c87760c69a89ddccc70e). So far we find this issue is exposed in [JettyService](https://github.com/line/armeria/blob/main/jetty9/src/main/java/com/linecorp/armeria/server/jetty/JettyService.java#L323) and [TomcatService](https://github.com/line/armeria/blob/main/tomcat9/src/main/java/com/linecorp/armeria/server/tomcat/TomcatService.java#L427) when the exception is thrown outside the “servlet handling” code path. We have a couple of questions: Is it expected behavior that the RequestLog never completes when neither data nor failure is being set on the response (just close)? This leads us to the second question: Is it a legit use of the API to close a HttpResponse without writing anything or “failing” it. Slack thread: https://line-armeria.slack.com/archives/C1NGPBUH2/p1679553378276729 If `JettyService` get an exception while handling a request, it closes the response silently without propagating the exception. https://github.com/line/armeria/blob/49054ad7f22a7f2d64faaeb694674b0df09ad165/jetty9/src/main/java/com/linecorp/armeria/server/jetty/JettyService.java#L322-L324 https://github.com/line/armeria/blob/49054ad7f22a7f2d64faaeb694674b0df09ad165/tomcat9/src/main/java/com/linecorp/armeria/server/tomcat/TomcatService.java#L424-L428 These can make the failed response publish no `HttpObject`s. https://github.com/line/armeria/blob/cef7b9c359b383e2b703f13c135f7800e980e422/core/src/main/java/com/linecorp/armeria/server/HttpResponseSubscriber.java#L307-L317 As a result, 1) the exception is not propagated to `RequestLog.responseCause()` and 2) the `RequestLog.whenComplete()` isn't completed because of the bug in `HttpResponseSubscriber`.
1.0
Handle exceptions thrown by Servlet - > Hi folks, we recently encountered an issue where Armeria (we are on Armeria 1.16.2) won’t finish collecting the response information for a RequestLog when the corresponding HttpResponse is closed without any data written. As a result, RequestLogAccess.whenComplete() never completes after closing the HttpResponse. A simple reproduction can be found [here](https://gist.github.com/linwuz-db/2c7adae5a546c87760c69a89ddccc70e). So far we find this issue is exposed in [JettyService](https://github.com/line/armeria/blob/main/jetty9/src/main/java/com/linecorp/armeria/server/jetty/JettyService.java#L323) and [TomcatService](https://github.com/line/armeria/blob/main/tomcat9/src/main/java/com/linecorp/armeria/server/tomcat/TomcatService.java#L427) when the exception is thrown outside the “servlet handling” code path. We have a couple of questions: Is it expected behavior that the RequestLog never completes when neither data nor failure is being set on the response (just close)? This leads us to the second question: Is it a legit use of the API to close a HttpResponse without writing anything or “failing” it. Slack thread: https://line-armeria.slack.com/archives/C1NGPBUH2/p1679553378276729 If `JettyService` get an exception while handling a request, it closes the response silently without propagating the exception. https://github.com/line/armeria/blob/49054ad7f22a7f2d64faaeb694674b0df09ad165/jetty9/src/main/java/com/linecorp/armeria/server/jetty/JettyService.java#L322-L324 https://github.com/line/armeria/blob/49054ad7f22a7f2d64faaeb694674b0df09ad165/tomcat9/src/main/java/com/linecorp/armeria/server/tomcat/TomcatService.java#L424-L428 These can make the failed response publish no `HttpObject`s. https://github.com/line/armeria/blob/cef7b9c359b383e2b703f13c135f7800e980e422/core/src/main/java/com/linecorp/armeria/server/HttpResponseSubscriber.java#L307-L317 As a result, 1) the exception is not propagated to `RequestLog.responseCause()` and 2) the `RequestLog.whenComplete()` isn't completed because of the bug in `HttpResponseSubscriber`.
defect
handle exceptions thrown by servlet hi folks we recently encountered an issue where armeria we are on armeria won’t finish collecting the response information for a requestlog when the corresponding httpresponse is closed without any data written as a result requestlogaccess whencomplete never completes after closing the httpresponse a simple reproduction can be found so far we find this issue is exposed in and when the exception is thrown outside the “servlet handling” code path we have a couple of questions is it expected behavior that the requestlog never completes when neither data nor failure is being set on the response just close this leads us to the second question is it a legit use of the api to close a httpresponse without writing anything or “failing” it slack thread if jettyservice get an exception while handling a request it closes the response silently without propagating the exception these can make the failed response publish no httpobject s as a result the exception is not propagated to requestlog responsecause and the requestlog whencomplete isn t completed because of the bug in httpresponsesubscriber
1
24,912
5,110,408,740
IssuesEvent
2017-01-06 00:09:40
emberjs/ember.js
https://api.github.com/repos/emberjs/ember.js
closed
[DOC] Ember.Component#layout is not documented as public
Documentation
Inspired by: https://github.com/ember-cli/ember-cli.github.io/issues/100 Currently the [default addon component blueprint](https://github.com/emberjs/ember.js/blob/v2.10.0-beta.2/blueprints/component/index.js#L59-L68) sets up the component template via the `layout` property. There is no evidence in the [API docs](http://emberjs.com/api/classes/Ember.Component.html) that it's public, or any mention or use of it in [the appropriate section of the guides](https://guides.emberjs.com/v2.9.0/components/defining-a-component/), leading to confusion as to whether it's ok to use. Add in [the deprecated `Ember.View` concept of the same name](http://emberjs.com/api/classes/Ember.View.html#toc_layouts), and it's understandable why people new to the framework are tripping up here.
1.0
[DOC] Ember.Component#layout is not documented as public - Inspired by: https://github.com/ember-cli/ember-cli.github.io/issues/100 Currently the [default addon component blueprint](https://github.com/emberjs/ember.js/blob/v2.10.0-beta.2/blueprints/component/index.js#L59-L68) sets up the component template via the `layout` property. There is no evidence in the [API docs](http://emberjs.com/api/classes/Ember.Component.html) that it's public, or any mention or use of it in [the appropriate section of the guides](https://guides.emberjs.com/v2.9.0/components/defining-a-component/), leading to confusion as to whether it's ok to use. Add in [the deprecated `Ember.View` concept of the same name](http://emberjs.com/api/classes/Ember.View.html#toc_layouts), and it's understandable why people new to the framework are tripping up here.
non_defect
ember component layout is not documented as public inspired by currently the sets up the component template via the layout property there is no evidence in the that it s public or any mention or use of it in leading to confusion as to whether it s ok to use add in and it s understandable why people new to the framework are tripping up here
0
34,543
7,453,494,658
IssuesEvent
2018-03-29 12:11:55
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Javascripti bugi seoste laadimisel
P: highest R: fixed T: defect
**Reported by jaak tamre on 16 Aug 2013 13:33 UTC** '''Object''' KÜ detailvaade '''Description''' Kui kirjeldusüksuse detailvaates avada sakk aktid, kohad või märksõnad, siis pärast uue seose lisamist või kustutamist läheb osa javascripti funktsionaalsust lehel katki. XhrLoader ei tööta enam ja akkordioni sinised ribad ei lähe kinni/lahti. Samas kui vajutada "näita kõiki" siis töötab. Äkki mingi live eventide probleem? '''Todo''' Uurida probleemi ja lahendada
1.0
Javascripti bugi seoste laadimisel - **Reported by jaak tamre on 16 Aug 2013 13:33 UTC** '''Object''' KÜ detailvaade '''Description''' Kui kirjeldusüksuse detailvaates avada sakk aktid, kohad või märksõnad, siis pärast uue seose lisamist või kustutamist läheb osa javascripti funktsionaalsust lehel katki. XhrLoader ei tööta enam ja akkordioni sinised ribad ei lähe kinni/lahti. Samas kui vajutada "näita kõiki" siis töötab. Äkki mingi live eventide probleem? '''Todo''' Uurida probleemi ja lahendada
defect
javascripti bugi seoste laadimisel reported by jaak tamre on aug utc object kü detailvaade description kui kirjeldusüksuse detailvaates avada sakk aktid kohad või märksõnad siis pärast uue seose lisamist või kustutamist läheb osa javascripti funktsionaalsust lehel katki xhrloader ei tööta enam ja akkordioni sinised ribad ei lähe kinni lahti samas kui vajutada näita kõiki siis töötab äkki mingi live eventide probleem todo uurida probleemi ja lahendada
1
124,445
16,603,131,670
IssuesEvent
2021-06-01 22:36:01
microsoft/TypeScript
https://api.github.com/repos/microsoft/TypeScript
closed
Function.prototype.bind typings do not work properly when supplied a function with variadic arguments.
Design Limitation
# Bug Report If you pass a function that might behave differently when supplied different arguments, the resulting type does not properly accommodate all call patterns. I ran into this while trying to figure out how TypeScript handled situations similar to RxJS's `bindNodeCallback` and `bindCallback`, which suffers from similar problems. ### 🔎 Search Terms - `bind` - `bind "Arguments for the rest parameter"` ### 🕗 Version & Regression Information versions 3.3.* - 4.1.3 - This is the behavior in every version I tried, and I reviewed the FAQ for entries about (I have no idea where the FAQ is, reporting anyway, because this is friction) ### ⏯ Playground Link [Playground link with relevant code](https://www.typescriptlang.org/play?#code/JYOwLgpgTgZghgYwgAgILIN4FgBQz-JwBcyAzmFKAOYDcuAvrqJLIigEKa4HIBGJ5SiFoMm4aPCTIAwlzwEEAitTo5GOZhLbIAInJ4ATJUJFrcuBAHsQ5QiXQBeTHeQByOK+T1VVm2D4knE4YAW68nt4W1raKMsjByLGuCBE+0f5GuvHOma4GqeY4APRFyAAyEGCupGRwAJ7IYAAWEKQocFAozSjWADYNAG5wvcAGyADu9TVglrglicO9yJYArlDIAAaQ5BvIMCsgCGDA1kS4+4fH1o2tYAAUxGgANKHsAJT2NMjzAIznB0cTiAbuQHvYXrFpB9kOwvvMAEz-S5AkH3R6oF78GEvTI6aHSOGlADMhXmACJgABbAAOvTJyBanSRgOu2zR4L4AHkoNJAsgAD4yHEAfhIePsAphktk2Hk+GAMGQd2ApFQDzeb30PAICqVKvYd143Khmtl2vNyF1ytIOjuBg1WotFs6YDWwJCSTZqW+xO4ToI9GQEF6bUd-p4Lrdzkeri9XkJyD+cv96idqYtVpV0kNxodZvDyEjUHdoVjt29CL95vT+BrjSaUEs42QIAgzYAolBG1A7q5xo3hAter1XG9VOpmVdgSq1aBqSswCQ4CA6tC5wvLTV0Pn8EXpyB5-4AGRHtweS37w-IE+NOrUiCWRXrsAAbXcrgAuvEHE5qsphK446FBcLLTqQBrPkuK5rgeG4qlKO6FpUUbPtep6uOEF4XleN5gHeD5PrBr4YZ+36-oI1CAaIOCTiiWbKkRUGriQqHwTKVZIa6xbYRuN7JJ4oA8cep54fej5CW+KRfj+5H-lQVFmDROAgVOm62pBhDQSxRGblkiF7kJaFuPkWGobh+Hic+b75NJP5uBRAFAUpUR+MgAAe2RsgAdLwoAGHcIArMOLxwJiY6FL4tjrE4bl3OFSmRZYvQQF5vSWFQdxQGOQA) ### 💻 Code ```ts interface A { a: string; } interface B { b: string; } interface C { c: string; } interface D { d: string; } const a: A = { a: 'a' }; const b: B = { b: 'b' }; const c: C = { c: 'c' }; const d: D = { d: 'd' }; // Let's say these are the only valid ways to // call our `test` function: function test(a: A, b: B): A; // 1 function test(a: A, c: C): B; // 2 function test(a: A, b: B, d: D): C; // 3 // "impl" here function test(a: A, bOrC: B | C, d?: D): A | B | C { if (isA(a)) { if (isB(bOrC)) { if (isD(d)) { return { c: 'test' }; // 3 } else { return { a: 'test' }; // 1 } } if (isC(bOrC)) { return { b: 'test' }; // 2 } } throw new Error('wrong call'); } function isA(input: any): input is A { return input && 'a' in input && typeof input['a'] === 'string'; } function isB(input: any): input is B { return input && 'b' in input && typeof input['b'] === 'string'; } function isC(input: any): input is C { return input && 'c' in input && typeof input['c'] === 'string'; } function isD(input: any): input is D { return input && 'd' in input && typeof input['d'] === 'string'; } const x = test.bind(null, a, b); const r = x(); console.log(r); ``` ### 🙁 Actual behavior ![image](https://user-images.githubusercontent.com/1540597/103491491-ef279700-4de9-11eb-98c5-e9acfc597e2d.png) I got a compilation error for a perfectly valid call of my bound function. ### 🙂 Expected behavior The return type (the type of `r`) should be `C`, and the code should compile successfully.
1.0
Function.prototype.bind typings do not work properly when supplied a function with variadic arguments. - # Bug Report If you pass a function that might behave differently when supplied different arguments, the resulting type does not properly accommodate all call patterns. I ran into this while trying to figure out how TypeScript handled situations similar to RxJS's `bindNodeCallback` and `bindCallback`, which suffers from similar problems. ### 🔎 Search Terms - `bind` - `bind "Arguments for the rest parameter"` ### 🕗 Version & Regression Information versions 3.3.* - 4.1.3 - This is the behavior in every version I tried, and I reviewed the FAQ for entries about (I have no idea where the FAQ is, reporting anyway, because this is friction) ### ⏯ Playground Link [Playground link with relevant code](https://www.typescriptlang.org/play?#code/JYOwLgpgTgZghgYwgAgILIN4FgBQz-JwBcyAzmFKAOYDcuAvrqJLIigEKa4HIBGJ5SiFoMm4aPCTIAwlzwEEAitTo5GOZhLbIAInJ4ATJUJFrcuBAHsQ5QiXQBeTHeQByOK+T1VVm2D4knE4YAW68nt4W1raKMsjByLGuCBE+0f5GuvHOma4GqeY4APRFyAAyEGCupGRwAJ7IYAAWEKQocFAozSjWADYNAG5wvcAGyADu9TVglrglicO9yJYArlDIAAaQ5BvIMCsgCGDA1kS4+4fH1o2tYAAUxGgANKHsAJT2NMjzAIznB0cTiAbuQHvYXrFpB9kOwvvMAEz-S5AkH3R6oF78GEvTI6aHSOGlADMhXmACJgABbAAOvTJyBanSRgOu2zR4L4AHkoNJAsgAD4yHEAfhIePsAphktk2Hk+GAMGQd2ApFQDzeb30PAICqVKvYd143Khmtl2vNyF1ytIOjuBg1WotFs6YDWwJCSTZqW+xO4ToI9GQEF6bUd-p4Lrdzkeri9XkJyD+cv96idqYtVpV0kNxodZvDyEjUHdoVjt29CL95vT+BrjSaUEs42QIAgzYAolBG1A7q5xo3hAter1XG9VOpmVdgSq1aBqSswCQ4CA6tC5wvLTV0Pn8EXpyB5-4AGRHtweS37w-IE+NOrUiCWRXrsAAbXcrgAuvEHE5qsphK446FBcLLTqQBrPkuK5rgeG4qlKO6FpUUbPtep6uOEF4XleN5gHeD5PrBr4YZ+36-oI1CAaIOCTiiWbKkRUGriQqHwTKVZIa6xbYRuN7JJ4oA8cep54fej5CW+KRfj+5H-lQVFmDROAgVOm62pBhDQSxRGblkiF7kJaFuPkWGobh+Hic+b75NJP5uBRAFAUpUR+MgAAe2RsgAdLwoAGHcIArMOLxwJiY6FL4tjrE4bl3OFSmRZYvQQF5vSWFQdxQGOQA) ### 💻 Code ```ts interface A { a: string; } interface B { b: string; } interface C { c: string; } interface D { d: string; } const a: A = { a: 'a' }; const b: B = { b: 'b' }; const c: C = { c: 'c' }; const d: D = { d: 'd' }; // Let's say these are the only valid ways to // call our `test` function: function test(a: A, b: B): A; // 1 function test(a: A, c: C): B; // 2 function test(a: A, b: B, d: D): C; // 3 // "impl" here function test(a: A, bOrC: B | C, d?: D): A | B | C { if (isA(a)) { if (isB(bOrC)) { if (isD(d)) { return { c: 'test' }; // 3 } else { return { a: 'test' }; // 1 } } if (isC(bOrC)) { return { b: 'test' }; // 2 } } throw new Error('wrong call'); } function isA(input: any): input is A { return input && 'a' in input && typeof input['a'] === 'string'; } function isB(input: any): input is B { return input && 'b' in input && typeof input['b'] === 'string'; } function isC(input: any): input is C { return input && 'c' in input && typeof input['c'] === 'string'; } function isD(input: any): input is D { return input && 'd' in input && typeof input['d'] === 'string'; } const x = test.bind(null, a, b); const r = x(); console.log(r); ``` ### 🙁 Actual behavior ![image](https://user-images.githubusercontent.com/1540597/103491491-ef279700-4de9-11eb-98c5-e9acfc597e2d.png) I got a compilation error for a perfectly valid call of my bound function. ### 🙂 Expected behavior The return type (the type of `r`) should be `C`, and the code should compile successfully.
non_defect
function prototype bind typings do not work properly when supplied a function with variadic arguments bug report if you pass a function that might behave differently when supplied different arguments the resulting type does not properly accommodate all call patterns i ran into this while trying to figure out how typescript handled situations similar to rxjs s bindnodecallback and bindcallback which suffers from similar problems 🔎 search terms bind bind arguments for the rest parameter 🕗 version regression information versions this is the behavior in every version i tried and i reviewed the faq for entries about i have no idea where the faq is reporting anyway because this is friction ⏯ playground link 💻 code ts interface a a string interface b b string interface c c string interface d d string const a a a a const b b b b const c c c c const d d d d let s say these are the only valid ways to call our test function function test a a b b a function test a a c c b function test a a b b d d c impl here function test a a borc b c d d a b c if isa a if isb borc if isd d return c test else return a test if isc borc return b test throw new error wrong call function isa input any input is a return input a in input typeof input string function isb input any input is b return input b in input typeof input string function isc input any input is c return input c in input typeof input string function isd input any input is d return input d in input typeof input string const x test bind null a b const r x console log r 🙁 actual behavior i got a compilation error for a perfectly valid call of my bound function 🙂 expected behavior the return type the type of r should be c and the code should compile successfully
0
578,306
17,146,593,764
IssuesEvent
2021-07-13 15:11:43
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
healthcare.api-client.v1.fhir.fhir_resources_test: test_delete_patient failed
api: healthcare flakybot: flaky flakybot: issue priority: p1 samples type: bug
This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: b99df8d36109e4fe3e397bfd2cbacac06960340c buildURL: [Build Status](https://source.cloud.google.com/results/invocations/325e6648-2604-4a2d-a611-53de8a0492aa), [Sponge](http://sponge2/325e6648-2604-4a2d-a611-53de8a0492aa) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/healthcare/api-client/v1/fhir/fhir_resources_test.py", line 160, in test_patient clean_up() File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/backoff/_sync.py", line 94, in retry ret = target(*args, **kwargs) File "/workspace/healthcare/api-client/v1/fhir/fhir_resources_test.py", line 147, in clean_up patient_resource_id, File "/workspace/healthcare/api-client/v1/fhir/fhir_resources.py", line 299, in delete_resource response = session.delete(resource_path) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/requests/sessions.py", line 624, in delete return self.request('DELETE', url, **kwargs) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/auth/transport/requests.py", line 478, in request self.credentials.before_request(auth_request, method, url, request_headers) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/auth/credentials.py", line 133, in before_request self.refresh(request) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/service_account.py", line 377, in refresh request, self._token_uri, assertion File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 193, in jwt_grant response_data = _token_endpoint_request(request, token_uri, body) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 165, in _token_endpoint_request _handle_error_response(response_data) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 60, in _handle_error_response raise exceptions.RefreshError(error_details, response_data) google.auth.exceptions.RefreshError: ('invalid_grant: Invalid JWT Signature.', {'error': 'invalid_grant', 'error_description': 'Invalid JWT Signature.'})</pre></details>
1.0
healthcare.api-client.v1.fhir.fhir_resources_test: test_delete_patient failed - This test failed! To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot). If I'm commenting on this issue too often, add the `flakybot: quiet` label and I will stop commenting. --- commit: b99df8d36109e4fe3e397bfd2cbacac06960340c buildURL: [Build Status](https://source.cloud.google.com/results/invocations/325e6648-2604-4a2d-a611-53de8a0492aa), [Sponge](http://sponge2/325e6648-2604-4a2d-a611-53de8a0492aa) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/healthcare/api-client/v1/fhir/fhir_resources_test.py", line 160, in test_patient clean_up() File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/backoff/_sync.py", line 94, in retry ret = target(*args, **kwargs) File "/workspace/healthcare/api-client/v1/fhir/fhir_resources_test.py", line 147, in clean_up patient_resource_id, File "/workspace/healthcare/api-client/v1/fhir/fhir_resources.py", line 299, in delete_resource response = session.delete(resource_path) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/requests/sessions.py", line 624, in delete return self.request('DELETE', url, **kwargs) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/auth/transport/requests.py", line 478, in request self.credentials.before_request(auth_request, method, url, request_headers) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/auth/credentials.py", line 133, in before_request self.refresh(request) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/service_account.py", line 377, in refresh request, self._token_uri, assertion File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 193, in jwt_grant response_data = _token_endpoint_request(request, token_uri, body) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 165, in _token_endpoint_request _handle_error_response(response_data) File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 60, in _handle_error_response raise exceptions.RefreshError(error_details, response_data) google.auth.exceptions.RefreshError: ('invalid_grant: Invalid JWT Signature.', {'error': 'invalid_grant', 'error_description': 'Invalid JWT Signature.'})</pre></details>
non_defect
healthcare api client fhir fhir resources test test delete patient failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace healthcare api client fhir fhir resources test py line in test patient clean up file workspace healthcare api client fhir nox py lib site packages backoff sync py line in retry ret target args kwargs file workspace healthcare api client fhir fhir resources test py line in clean up patient resource id file workspace healthcare api client fhir fhir resources py line in delete resource response session delete resource path file workspace healthcare api client fhir nox py lib site packages requests sessions py line in delete return self request delete url kwargs file workspace healthcare api client fhir nox py lib site packages google auth transport requests py line in request self credentials before request auth request method url request headers file workspace healthcare api client fhir nox py lib site packages google auth credentials py line in before request self refresh request file workspace healthcare api client fhir nox py lib site packages google service account py line in refresh request self token uri assertion file workspace healthcare api client fhir nox py lib site packages google client py line in jwt grant response data token endpoint request request token uri body file workspace healthcare api client fhir nox py lib site packages google client py line in token endpoint request handle error response response data file workspace healthcare api client fhir nox py lib site packages google client py line in handle error response raise exceptions refresherror error details response data google auth exceptions refresherror invalid grant invalid jwt signature error invalid grant error description invalid jwt signature
0
68,644
21,775,146,357
IssuesEvent
2022-05-13 13:08:30
hazelcast/hazelcast
https://api.github.com/repos/hazelcast/hazelcast
opened
DataPersistenceConfig and TieredStoreConfig are missing in dynamic AddMapConfig
Type: Defect Source: Internal Module: Config
**Describe the bug** `DataPersistenceConfig` and `TieredStoreConfig` submap configurations are not transmitted while dynamically adding map config. See: https://github.com/hazelcast/hazelcast/blob/master/hazelcast/src/main/java/com/hazelcast/client/impl/protocol/codec/DynamicConfigAddMapConfigCodec.java So, currently these subconfig elements cannot be configured while dynamically adding the map config. To fix it: Update the client protocol of [DynamicConfig#addMapConfig](https://github.com/hazelcast/hazelcast-client-protocol/blob/master/protocol-definitions/DynamicConfig.yaml#L698-L887) to include these mentioned subConfigs. Also, the configuration update checklist wiki page, https://github.com/hazelcast/hazelcast/wiki/Checklist-for-Hazelcast-config-changes, doesn't include the necessary information about this. We need to update this page to cover this.
1.0
DataPersistenceConfig and TieredStoreConfig are missing in dynamic AddMapConfig - **Describe the bug** `DataPersistenceConfig` and `TieredStoreConfig` submap configurations are not transmitted while dynamically adding map config. See: https://github.com/hazelcast/hazelcast/blob/master/hazelcast/src/main/java/com/hazelcast/client/impl/protocol/codec/DynamicConfigAddMapConfigCodec.java So, currently these subconfig elements cannot be configured while dynamically adding the map config. To fix it: Update the client protocol of [DynamicConfig#addMapConfig](https://github.com/hazelcast/hazelcast-client-protocol/blob/master/protocol-definitions/DynamicConfig.yaml#L698-L887) to include these mentioned subConfigs. Also, the configuration update checklist wiki page, https://github.com/hazelcast/hazelcast/wiki/Checklist-for-Hazelcast-config-changes, doesn't include the necessary information about this. We need to update this page to cover this.
defect
datapersistenceconfig and tieredstoreconfig are missing in dynamic addmapconfig describe the bug datapersistenceconfig and tieredstoreconfig submap configurations are not transmitted while dynamically adding map config see so currently these subconfig elements cannot be configured while dynamically adding the map config to fix it update the client protocol of to include these mentioned subconfigs also the configuration update checklist wiki page doesn t include the necessary information about this we need to update this page to cover this
1
284,632
24,611,544,958
IssuesEvent
2022-10-14 22:13:13
gitpod-io/gitpod
https://api.github.com/repos/gitpod-io/gitpod
closed
Performance test PVC with a single node saturated with workspaces
aspect: testing aspect: performance
## Is your feature request related to a problem? Please describe We should test how a node behaves when it is full of workspaces using PVCs, where there is disk activity in the workspaces. ## Describe the behaviour you'd like 1. Start a regular workspace in the cluster using PVC on a single node. 2. Cordon the other two nodes 3. Run loadgen, once to fill half of the node, ~9 workspaces 4. Begin to stop the first loadgen run & start second loadgen run with ~20 workspaces (to fill the initial node and trigger scale-up of a new one) ### Questions - What IOPS and bandwidth do we achieve on the initial node? How is it different from a normal cluster? - How does the regular workspace that was initially started respond as we go through the various phases described above? ## Additional context We're not sure if the IO limiter is needed to cover these disks or not, and whether there will excessive CPU usage as a result of using PVCs for `/workspace`.
1.0
Performance test PVC with a single node saturated with workspaces - ## Is your feature request related to a problem? Please describe We should test how a node behaves when it is full of workspaces using PVCs, where there is disk activity in the workspaces. ## Describe the behaviour you'd like 1. Start a regular workspace in the cluster using PVC on a single node. 2. Cordon the other two nodes 3. Run loadgen, once to fill half of the node, ~9 workspaces 4. Begin to stop the first loadgen run & start second loadgen run with ~20 workspaces (to fill the initial node and trigger scale-up of a new one) ### Questions - What IOPS and bandwidth do we achieve on the initial node? How is it different from a normal cluster? - How does the regular workspace that was initially started respond as we go through the various phases described above? ## Additional context We're not sure if the IO limiter is needed to cover these disks or not, and whether there will excessive CPU usage as a result of using PVCs for `/workspace`.
non_defect
performance test pvc with a single node saturated with workspaces is your feature request related to a problem please describe we should test how a node behaves when it is full of workspaces using pvcs where there is disk activity in the workspaces describe the behaviour you d like start a regular workspace in the cluster using pvc on a single node cordon the other two nodes run loadgen once to fill half of the node workspaces begin to stop the first loadgen run start second loadgen run with workspaces to fill the initial node and trigger scale up of a new one questions what iops and bandwidth do we achieve on the initial node how is it different from a normal cluster how does the regular workspace that was initially started respond as we go through the various phases described above additional context we re not sure if the io limiter is needed to cover these disks or not and whether there will excessive cpu usage as a result of using pvcs for workspace
0
95,026
19,655,695,542
IssuesEvent
2022-01-10 12:16:28
sourcegraph/sourcegraph
https://api.github.com/repos/sourcegraph/sourcegraph
opened
Don't display the definition as a reference when doing "find references" from the definition
team/code-intelligence
#### Feature request description <!-- A description of what feature you would like. --> When doing "Find references" from the definition location, it would be nice if the definition was excluded from the list of results. #### Is your feature request related to a problem? If so, please describe. <!-- A description of what the problem is. Ex. I'm always frustrated when [...] --> Currently, the "references" panel gets loaded when a symbol has only one reference. It would be nice if the navigation went straight to the single reference instead of opening the references panel with two results (and one result is the definition location where you triggered references) #### Describe alternatives you've considered. <!-- A description of any alternative solutions or features you've considered. --> Manually click on the reference result inside the references panel. #### Additional context <!-- Add any other context or other information you'd like to provide. -->
1.0
Don't display the definition as a reference when doing "find references" from the definition - #### Feature request description <!-- A description of what feature you would like. --> When doing "Find references" from the definition location, it would be nice if the definition was excluded from the list of results. #### Is your feature request related to a problem? If so, please describe. <!-- A description of what the problem is. Ex. I'm always frustrated when [...] --> Currently, the "references" panel gets loaded when a symbol has only one reference. It would be nice if the navigation went straight to the single reference instead of opening the references panel with two results (and one result is the definition location where you triggered references) #### Describe alternatives you've considered. <!-- A description of any alternative solutions or features you've considered. --> Manually click on the reference result inside the references panel. #### Additional context <!-- Add any other context or other information you'd like to provide. -->
non_defect
don t display the definition as a reference when doing find references from the definition feature request description when doing find references from the definition location it would be nice if the definition was excluded from the list of results is your feature request related to a problem if so please describe currently the references panel gets loaded when a symbol has only one reference it would be nice if the navigation went straight to the single reference instead of opening the references panel with two results and one result is the definition location where you triggered references describe alternatives you ve considered manually click on the reference result inside the references panel additional context
0
73,782
24,797,605,441
IssuesEvent
2022-10-24 18:41:31
fecgov/fecfile-web-app
https://api.github.com/repos/fecgov/fecfile-web-app
opened
Defect: System to itemize JF Transfer regardless of amount
defect
Implementation of #486 was broken by subsequent development. This ticket is to re-code and re-test this functionality. Note that itemization logic is _not_ determined by aggregation group. ### Business Reason ### Per FECA regulations, all Form 3X transactions on Line 12 must be itemized on Schedule A, regardless of amount. This includes JF Transfer transactions. ### Acceptance Criteria ### **Given** one of the following transactions: - Joint Fundraising Transfer - PAC JF Transfer Memo **Then** the transaction will be itemized upon saving **and** appear on it's own line item on the *.fec file
1.0
Defect: System to itemize JF Transfer regardless of amount - Implementation of #486 was broken by subsequent development. This ticket is to re-code and re-test this functionality. Note that itemization logic is _not_ determined by aggregation group. ### Business Reason ### Per FECA regulations, all Form 3X transactions on Line 12 must be itemized on Schedule A, regardless of amount. This includes JF Transfer transactions. ### Acceptance Criteria ### **Given** one of the following transactions: - Joint Fundraising Transfer - PAC JF Transfer Memo **Then** the transaction will be itemized upon saving **and** appear on it's own line item on the *.fec file
defect
defect system to itemize jf transfer regardless of amount implementation of was broken by subsequent development this ticket is to re code and re test this functionality note that itemization logic is not determined by aggregation group business reason per feca regulations all form transactions on line must be itemized on schedule a regardless of amount this includes jf transfer transactions acceptance criteria given one of the following transactions joint fundraising transfer pac jf transfer memo then the transaction will be itemized upon saving and appear on it s own line item on the fec file
1
33,423
12,216,531,738
IssuesEvent
2020-05-01 15:20:11
robertjfinn/hadoop
https://api.github.com/repos/robertjfinn/hadoop
opened
CVE-2019-11834 (High) detected in cJSON-v1.7.8
security vulnerability
## CVE-2019-11834 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cJSONv1.7.8</b></p></summary> <p> <p>Ultralightweight JSON parser in ANSI C</p> <p>Library home page: <a href=https://github.com/DaveGamble/cJSON.git>https://github.com/DaveGamble/cJSON.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/robertjfinn/hadoop/commit/876b3d37847317582197087627081de9f19f88d9">876b3d37847317582197087627081de9f19f88d9</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (2)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/utils/cJSON/cJSON.c - /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/utils/cJSON/cJSON.h </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> cJSON before 1.7.11 allows out-of-bounds access, related to \x00 in a string literal. <p>Publish Date: 2019-05-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11834>CVE-2019-11834</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11835">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11835</a></p> <p>Release Date: 2019-05-09</p> <p>Fix Resolution: v1.7.11</p> </p> </details> <p></p>
True
CVE-2019-11834 (High) detected in cJSON-v1.7.8 - ## CVE-2019-11834 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cJSONv1.7.8</b></p></summary> <p> <p>Ultralightweight JSON parser in ANSI C</p> <p>Library home page: <a href=https://github.com/DaveGamble/cJSON.git>https://github.com/DaveGamble/cJSON.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/robertjfinn/hadoop/commit/876b3d37847317582197087627081de9f19f88d9">876b3d37847317582197087627081de9f19f88d9</a></p> </p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (2)</summary> <p></p> <p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p> <p> - /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/utils/cJSON/cJSON.c - /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/utils/cJSON/cJSON.h </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> cJSON before 1.7.11 allows out-of-bounds access, related to \x00 in a string literal. <p>Publish Date: 2019-05-09 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11834>CVE-2019-11834</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11835">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11835</a></p> <p>Release Date: 2019-05-09</p> <p>Fix Resolution: v1.7.11</p> </p> </details> <p></p>
non_defect
cve high detected in cjson cve high severity vulnerability vulnerable library ultralightweight json parser in ansi c library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server nodemanager src main native container executor impl utils cjson cjson c hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server nodemanager src main native container executor impl utils cjson cjson h vulnerability details cjson before allows out of bounds access related to in a string literal publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
0
280,963
21,315,298,955
IssuesEvent
2022-04-16 06:56:59
Shivang211/pe
https://api.github.com/repos/Shivang211/pe
opened
No visuals for commands
severity.Low type.DocumentationBug
![Screenshot 2022-04-16 at 2.56.06 PM.png](https://raw.githubusercontent.com/Shivang211/pe/main/files/a0641701-e72d-46c7-b326-5cd9f2191b9b.png) There are a lot of commands and no visuals provided for any of them, which makes reading and understanding the UG very difficult. <!--session: 1650088058079-eb80a396-2b55-442e-a1e8-c84d767ca243--> <!--Version: Web v3.4.2-->
1.0
No visuals for commands - ![Screenshot 2022-04-16 at 2.56.06 PM.png](https://raw.githubusercontent.com/Shivang211/pe/main/files/a0641701-e72d-46c7-b326-5cd9f2191b9b.png) There are a lot of commands and no visuals provided for any of them, which makes reading and understanding the UG very difficult. <!--session: 1650088058079-eb80a396-2b55-442e-a1e8-c84d767ca243--> <!--Version: Web v3.4.2-->
non_defect
no visuals for commands there are a lot of commands and no visuals provided for any of them which makes reading and understanding the ug very difficult
0
337,867
30,268,262,312
IssuesEvent
2023-07-07 13:31:01
apache/camel-quarkus
https://api.github.com/repos/apache/camel-quarkus
closed
Nitrite native integration test could not serialize [[3941604762136]NO₂]
bug native test area/testing area/native area/nitrite
### Bug description For some reason this has recently popped up on the CI builds: ``` 2023-07-06T15:51:28.1555920Z [INFO] ------------------------------------------------------- 2023-07-06T15:51:28.1556320Z [INFO] T E S T S 2023-07-06T15:51:28.1556760Z [INFO] ------------------------------------------------------- 2023-07-06T15:51:29.0215396Z [INFO] Running org.apache.camel.quarkus.component.nitrite.it.NitriteIT 2023-07-06T15:51:33.0323722Z ====================================================================================== 2023-07-06T15:51:33.0324350Z native.image.path was not set, making a guess for the correct path of native image 2023-07-06T15:51:33.0325453Z guessed path: /home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/camel-quarkus-integration-test-nitrite-3.0.0-SNAPSHOT-runner 2023-07-06T15:51:33.0326449Z ====================================================================================== 2023-07-06T15:51:33.0336065Z Executing "/home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/camel-quarkus-integration-test-nitrite-3.0.0-SNAPSHOT-runner -Dquarkus.http.port=8081 -Dquarkus.http.ssl-port=8444 -Dtest.url=http://localhost:8081 -Dquarkus.log.file.path=/home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/quarkus.log -Dquarkus.log.file.enable=true -Dquarkus.log.category."io.quarkus".level=INFO -Dcamel.quarkus.nitrite.test.db.file=/tmp/NitriteTestResource-db-file-17819878958287804087" 2023-07-06T15:51:33.1319606Z 2023-07-06 15:51:33,050 INFO [org.apa.cam.qua.cor.CamelBootstrapRecorder] (main) Bootstrap runtime: org.apache.camel.quarkus.main.CamelMainRuntime 2023-07-06T15:51:33.1321026Z 2023-07-06 15:51:33,051 INFO [org.apa.cam.mai.MainSupport] (main) Apache Camel (Main) 4.0.0-RC1 is starting 2023-07-06T15:51:33.1321867Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Apache Camel 4.0.0-RC1 (camel-1) is starting 2023-07-06T15:51:33.1322916Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Routes startup (started:0) 2023-07-06T15:51:33.1323847Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Apache Camel 4.0.0-RC1 (camel-1) started in 0ms (build:0ms init:0ms start:0ms) 2023-07-06T15:51:33.1325085Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) camel-quarkus-integration-test-nitrite 3.0.0-SNAPSHOT native (powered by Quarkus 3.2.0.Final) started in 0.042s. Listening on: http://0.0.0.0:8081 2023-07-06T15:51:33.1325889Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) Profile prod activated. 2023-07-06T15:51:33.1326938Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) Installed features: [camel-core, camel-nitrite, cdi, resteasy, resteasy-jackson, smallrye-context-propagation, vertx] 2023-07-06T15:51:43.1229021Z ##[error]Exception in thread "MVStore background writer nio:/tmp/NitriteTestResource-db-file-17819878958287804087" java.lang.IllegalStateException: java.lang.IllegalArgumentException: Could not serialize [[3941604762136]NO₂] [1.4.200/0] [1.4.200/3] 2023-07-06T15:51:43.1524187Z at org.h2.mvstore.DataUtils.newIllegalStateException(DataUtils.java:950) 2023-07-06T15:51:43.2171401Z at org.h2.mvstore.MVStore.store(MVStore.java:1338) 2023-07-06T15:51:43.2172037Z at org.h2.mvstore.MVStore.store(MVStore.java:1312) 2023-07-06T15:51:43.2172518Z at org.h2.mvstore.MVStore.tryCommit(MVStore.java:1272) 2023-07-06T15:51:43.2173014Z at org.h2.mvstore.MVStore.writeInBackground(MVStore.java:2780) 2023-07-06T15:51:43.2173539Z at org.h2.mvstore.MVStore$BackgroundWriterThread.run(MVStore.java:3290) 2023-07-06T15:51:43.2174239Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.thread.PlatformThreads.threadStartRoutine(PlatformThreads.java:807) 2023-07-06T15:51:43.2175172Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.posix.thread.PosixPlatformThreads.pthreadStartRoutine(PosixPlatformThreads.java:210) 2023-07-06T15:51:43.2176353Z Caused by: java.lang.IllegalArgumentException: Could not serialize [[3941604762136]NO₂] [1.4.200/0] 2023-07-06T15:51:43.2177316Z at org.h2.mvstore.DataUtils.newIllegalArgumentException(DataUtils.java:924) 2023-07-06T15:51:43.2177939Z at org.h2.mvstore.type.ObjectDataType.serialize(ObjectDataType.java:359) 2023-07-06T15:51:43.2178655Z at org.h2.mvstore.type.ObjectDataType$SerializedObjectType.write(ObjectDataType.java:1527) 2023-07-06T15:51:43.2179593Z at org.h2.mvstore.type.ObjectDataType.write(ObjectDataType.java:127) 2023-07-06T15:51:43.2180134Z at org.h2.mvstore.type.ObjectDataType.write(ObjectDataType.java:121) 2023-07-06T15:51:43.2180627Z at org.h2.mvstore.Page$Leaf.writeValues(Page.java:1516) 2023-07-06T15:51:43.2181033Z at org.h2.mvstore.Page.write(Page.java:670) 2023-07-06T15:51:43.2181483Z at org.h2.mvstore.Page$Leaf.writeUnsavedRecursive(Page.java:1525) 2023-07-06T15:51:43.2181950Z at org.h2.mvstore.MVStore.storeNow(MVStore.java:1427) 2023-07-06T15:51:43.2182597Z at org.h2.mvstore.MVStore.store(MVStore.java:1334) 2023-07-06T15:51:43.2182939Z ... 6 more 2023-07-06T15:51:43.2184215Z Caused by: com.oracle.svm.core.jdk.UnsupportedFeatureError: SerializationConstructorAccessor class not found for declaringClass: java.util.concurrent.ConcurrentSkipListSet (targetConstructorClass: java.util.AbstractSet). Usually adding java.util.concurrent.ConcurrentSkipListSet to serialization-config.json fixes the problem. 2023-07-06T15:51:43.2185371Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.util.VMError.unsupportedFeature(VMError.java:92) 2023-07-06T15:51:43.2186351Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.reflect.serialize.SerializationSupport.getSerializationConstructorAccessor(SerializationSupport.java:143) 2023-07-06T15:51:43.2187439Z at java.base@17.0.7/jdk.internal.reflect.MethodAccessorGenerator.generateSerializationConstructor(MethodAccessorGenerator.java:61) 2023-07-06T15:51:43.2188304Z at java.base@17.0.7/jdk.internal.reflect.ReflectionFactory.generateConstructor(ReflectionFactory.java:463) 2023-07-06T15:51:43.2189111Z at java.base@17.0.7/jdk.internal.reflect.ReflectionFactory.newConstructorForSerialization(ReflectionFactory.java:455) 2023-07-06T15:51:43.2189887Z at java.base@17.0.7/java.io.ObjectStreamClass.getSerializableConstructor(ObjectStreamClass.java:1443) 2023-07-06T15:51:43.2190498Z at java.base@17.0.7/java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:412) 2023-07-06T15:51:43.2191003Z at java.base@17.0.7/java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:384) 2023-07-06T15:51:43.2191593Z at java.base@17.0.7/java.security.AccessController.executePrivileged(AccessController.java:171) 2023-07-06T15:51:43.2192233Z at java.base@17.0.7/java.security.AccessController.doPrivileged(AccessController.java:318) 2023-07-06T15:51:43.2192802Z at java.base@17.0.7/java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:384) 2023-07-06T15:51:43.2193385Z at java.base@17.0.7/java.io.ObjectStreamClass$Caches$1.computeValue(ObjectStreamClass.java:110) 2023-07-06T15:51:43.2194017Z at java.base@17.0.7/java.io.ObjectStreamClass$Caches$1.computeValue(ObjectStreamClass.java:107) 2023-07-06T15:51:43.2195550Z at java.base@17.0.7/java.io.ClassCache$1.computeValue(ClassCache.java:73) 2023-07-06T15:51:43.2195961Z at java.base@17.0.7/java.io.ClassCache$1.computeValue(ClassCache.java:70) 2023-07-06T15:51:43.2196389Z at java.base@17.0.7/java.lang.ClassValue.get(JavaLangSubstitutions.java:681) 2023-07-06T15:51:43.2196790Z at java.base@17.0.7/java.io.ClassCache.get(ClassCache.java:84) 2023-07-06T15:51:43.2197216Z at java.base@17.0.7/java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:363) 2023-07-06T15:51:43.2197720Z at java.base@17.0.7/java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1137) 2023-07-06T15:51:43.2198233Z at java.base@17.0.7/java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:350) 2023-07-06T15:51:43.2198983Z at org.h2.mvstore.type.ObjectDataType.serialize(ObjectDataType.java:356) 2023-07-06T15:51:43.2199343Z ... 14 more 2023-07-06T15:51:43.2200696Z 2023-07-06 15:51:43,064 ERROR [io.qua.ver.htt.run.QuarkusErrorHandler] (executor-thread-1) HTTP Request to /nitrite/repositoryClass?mappable=false failed, error id: 36f95039-4468-4295-a769-49c146d7d407-1: org.jboss.resteasy.spi.UnhandledException: org.apache.camel.CamelExecutionException: Exception occurred during execution on the exchange: Exchange[] 2023-07-06T15:51:43.2201712Z at org.jboss.resteasy.core.ExceptionHandler.handleApplicationException(ExceptionHandler.java:107) 2023-07-06T15:51:43.2202363Z at org.jboss.resteasy.core.ExceptionHandler.handleException(ExceptionHandler.java:344) 2023-07-06T15:51:43.2203006Z at org.jboss.resteasy.core.SynchronousDispatcher.writeException(SynchronousDispatcher.java:205) 2023-07-06T15:51:43.2203641Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:452) 2023-07-06T15:51:43.2204243Z at org.jboss.resteasy.core.SynchronousDispatcher.lambda$invoke$4(SynchronousDispatcher.java:240) 2023-07-06T15:51:43.2205009Z at org.jboss.resteasy.core.SynchronousDispatcher.lambda$preprocess$0(SynchronousDispatcher.java:154) 2023-07-06T15:51:43.2205810Z at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:321) 2023-07-06T15:51:43.2206598Z at org.jboss.resteasy.core.SynchronousDispatcher.preprocess(SynchronousDispatcher.java:157) 2023-07-06T15:51:43.2207202Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:229) 2023-07-06T15:51:43.2207839Z at io.quarkus.resteasy.runtime.standalone.RequestDispatcher.service(RequestDispatcher.java:82) 2023-07-06T15:51:43.2208538Z at io.quarkus.resteasy.runtime.standalone.VertxRequestHandler.dispatch(VertxRequestHandler.java:147) 2023-07-06T15:51:43.2209222Z at io.quarkus.resteasy.runtime.standalone.VertxRequestHandler$1.run(VertxRequestHandler.java:93) 2023-07-06T15:51:43.2209813Z at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:576) 2023-07-06T15:51:43.2210364Z at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2513) 2023-07-06T15:51:43.2210919Z at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1538) 2023-07-06T15:51:43.2211427Z at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29) 2023-07-06T15:51:43.2211987Z at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29) 2023-07-06T15:51:43.2212601Z at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) 2023-07-06T15:51:43.2213074Z at java.base@17.0.7/java.lang.Thread.run(Thread.java:833) 2023-07-06T15:51:43.2213640Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.thread.PlatformThreads.threadStartRoutine(PlatformThreads.java:807) 2023-07-06T15:51:43.2214493Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.posix.thread.PosixPlatformThreads.pthreadStartRoutine(PosixPlatformThreads.java:210) 2023-07-06T15:51:43.2215255Z Caused by: org.apache.camel.CamelExecutionException: Exception occurred during execution on the exchange: Exchange[] 2023-07-06T15:51:43.2215862Z at org.apache.camel.support.ExchangeHelper.extractResultBody(ExchangeHelper.java:653) 2023-07-06T15:51:43.2216545Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:368) 2023-07-06T15:51:43.2217320Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:333) 2023-07-06T15:51:43.2218127Z at org.apache.camel.quarkus.component.nitrite.it.NitriteResource.postRepositoryClass(NitriteResource.java:87) 2023-07-06T15:51:43.2218916Z at org.apache.camel.quarkus.component.nitrite.it.NitriteResource_ClientProxy.postRepositoryClass(Unknown Source) 2023-07-06T15:51:43.2219474Z at java.base@17.0.7/java.lang.reflect.Method.invoke(Method.java:568) 2023-07-06T15:51:43.2220237Z at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:154) 2023-07-06T15:51:43.2220834Z at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:118) 2023-07-06T15:51:43.2221593Z at org.jboss.resteasy.core.ResourceMethodInvoker.internalInvokeOnTarget(ResourceMethodInvoker.java:560) 2023-07-06T15:51:43.2222359Z at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTargetAfterFilter(ResourceMethodInvoker.java:452) 2023-07-06T15:51:43.2223073Z at org.jboss.resteasy.core.ResourceMethodInvoker.lambda$invokeOnTarget$2(ResourceMethodInvoker.java:413) 2023-07-06T15:51:43.2223876Z at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:321) 2023-07-06T15:51:43.2224673Z at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTarget(ResourceMethodInvoker.java:415) 2023-07-06T15:51:43.2225303Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:378) 2023-07-06T15:51:43.2225895Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:356) 2023-07-06T15:51:43.2226579Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:70) 2023-07-06T15:51:43.2227163Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:429) 2023-07-06T15:51:43.2227556Z ... 17 more 2023-07-06T15:51:43.2227921Z Caused by: org.dizitart.no2.exceptions.NitriteIOException: NO2.2004: store is closed 2023-07-06T15:51:43.2228524Z at org.dizitart.no2.internals.DefaultNitriteCollection.checkOpened(DefaultNitriteCollection.java:394) 2023-07-06T15:51:43.2229214Z at org.dizitart.no2.internals.DefaultNitriteCollection.update(DefaultNitriteCollection.java:320) 2023-07-06T15:51:43.2229861Z at org.dizitart.no2.objects.DefaultObjectRepository.update(DefaultObjectRepository.java:141) 2023-07-06T15:51:43.2230489Z at org.dizitart.no2.objects.DefaultObjectRepository.update(DefaultObjectRepository.java:126) 2023-07-06T15:51:43.2231180Z at org.apache.camel.component.nitrite.operation.common.UpsertOperation.execute(UpsertOperation.java:41) 2023-07-06T15:51:43.2231880Z at org.apache.camel.component.nitrite.NitriteProducer.process(NitriteProducer.java:78) 2023-07-06T15:51:43.2232637Z at org.apache.camel.support.AsyncProcessorConverterHelper$ProcessorToAsyncProcessorBridge.process(AsyncProcessorConverterHelper.java:66) 2023-07-06T15:51:43.2233453Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:214) 2023-07-06T15:51:43.2234191Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor$1.process(SharedCamelInternalProcessor.java:111) 2023-07-06T15:51:43.2234982Z at org.apache.camel.impl.engine.DefaultAsyncProcessorAwaitManager.process(DefaultAsyncProcessorAwaitManager.java:81) 2023-07-06T15:51:43.2235782Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:108) 2023-07-06T15:51:43.2236467Z at org.apache.camel.support.cache.DefaultProducerCache.send(DefaultProducerCache.java:180) 2023-07-06T15:51:43.2237116Z at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:176) 2023-07-06T15:51:43.2237769Z at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:172) 2023-07-06T15:51:43.2238490Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:360) 2023-07-06T15:51:43.2239137Z ... 32 more ```
2.0
Nitrite native integration test could not serialize [[3941604762136]NO₂] - ### Bug description For some reason this has recently popped up on the CI builds: ``` 2023-07-06T15:51:28.1555920Z [INFO] ------------------------------------------------------- 2023-07-06T15:51:28.1556320Z [INFO] T E S T S 2023-07-06T15:51:28.1556760Z [INFO] ------------------------------------------------------- 2023-07-06T15:51:29.0215396Z [INFO] Running org.apache.camel.quarkus.component.nitrite.it.NitriteIT 2023-07-06T15:51:33.0323722Z ====================================================================================== 2023-07-06T15:51:33.0324350Z native.image.path was not set, making a guess for the correct path of native image 2023-07-06T15:51:33.0325453Z guessed path: /home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/camel-quarkus-integration-test-nitrite-3.0.0-SNAPSHOT-runner 2023-07-06T15:51:33.0326449Z ====================================================================================== 2023-07-06T15:51:33.0336065Z Executing "/home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/camel-quarkus-integration-test-nitrite-3.0.0-SNAPSHOT-runner -Dquarkus.http.port=8081 -Dquarkus.http.ssl-port=8444 -Dtest.url=http://localhost:8081 -Dquarkus.log.file.path=/home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/quarkus.log -Dquarkus.log.file.enable=true -Dquarkus.log.category."io.quarkus".level=INFO -Dcamel.quarkus.nitrite.test.db.file=/tmp/NitriteTestResource-db-file-17819878958287804087" 2023-07-06T15:51:33.1319606Z 2023-07-06 15:51:33,050 INFO [org.apa.cam.qua.cor.CamelBootstrapRecorder] (main) Bootstrap runtime: org.apache.camel.quarkus.main.CamelMainRuntime 2023-07-06T15:51:33.1321026Z 2023-07-06 15:51:33,051 INFO [org.apa.cam.mai.MainSupport] (main) Apache Camel (Main) 4.0.0-RC1 is starting 2023-07-06T15:51:33.1321867Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Apache Camel 4.0.0-RC1 (camel-1) is starting 2023-07-06T15:51:33.1322916Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Routes startup (started:0) 2023-07-06T15:51:33.1323847Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Apache Camel 4.0.0-RC1 (camel-1) started in 0ms (build:0ms init:0ms start:0ms) 2023-07-06T15:51:33.1325085Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) camel-quarkus-integration-test-nitrite 3.0.0-SNAPSHOT native (powered by Quarkus 3.2.0.Final) started in 0.042s. Listening on: http://0.0.0.0:8081 2023-07-06T15:51:33.1325889Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) Profile prod activated. 2023-07-06T15:51:33.1326938Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) Installed features: [camel-core, camel-nitrite, cdi, resteasy, resteasy-jackson, smallrye-context-propagation, vertx] 2023-07-06T15:51:43.1229021Z ##[error]Exception in thread "MVStore background writer nio:/tmp/NitriteTestResource-db-file-17819878958287804087" java.lang.IllegalStateException: java.lang.IllegalArgumentException: Could not serialize [[3941604762136]NO₂] [1.4.200/0] [1.4.200/3] 2023-07-06T15:51:43.1524187Z at org.h2.mvstore.DataUtils.newIllegalStateException(DataUtils.java:950) 2023-07-06T15:51:43.2171401Z at org.h2.mvstore.MVStore.store(MVStore.java:1338) 2023-07-06T15:51:43.2172037Z at org.h2.mvstore.MVStore.store(MVStore.java:1312) 2023-07-06T15:51:43.2172518Z at org.h2.mvstore.MVStore.tryCommit(MVStore.java:1272) 2023-07-06T15:51:43.2173014Z at org.h2.mvstore.MVStore.writeInBackground(MVStore.java:2780) 2023-07-06T15:51:43.2173539Z at org.h2.mvstore.MVStore$BackgroundWriterThread.run(MVStore.java:3290) 2023-07-06T15:51:43.2174239Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.thread.PlatformThreads.threadStartRoutine(PlatformThreads.java:807) 2023-07-06T15:51:43.2175172Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.posix.thread.PosixPlatformThreads.pthreadStartRoutine(PosixPlatformThreads.java:210) 2023-07-06T15:51:43.2176353Z Caused by: java.lang.IllegalArgumentException: Could not serialize [[3941604762136]NO₂] [1.4.200/0] 2023-07-06T15:51:43.2177316Z at org.h2.mvstore.DataUtils.newIllegalArgumentException(DataUtils.java:924) 2023-07-06T15:51:43.2177939Z at org.h2.mvstore.type.ObjectDataType.serialize(ObjectDataType.java:359) 2023-07-06T15:51:43.2178655Z at org.h2.mvstore.type.ObjectDataType$SerializedObjectType.write(ObjectDataType.java:1527) 2023-07-06T15:51:43.2179593Z at org.h2.mvstore.type.ObjectDataType.write(ObjectDataType.java:127) 2023-07-06T15:51:43.2180134Z at org.h2.mvstore.type.ObjectDataType.write(ObjectDataType.java:121) 2023-07-06T15:51:43.2180627Z at org.h2.mvstore.Page$Leaf.writeValues(Page.java:1516) 2023-07-06T15:51:43.2181033Z at org.h2.mvstore.Page.write(Page.java:670) 2023-07-06T15:51:43.2181483Z at org.h2.mvstore.Page$Leaf.writeUnsavedRecursive(Page.java:1525) 2023-07-06T15:51:43.2181950Z at org.h2.mvstore.MVStore.storeNow(MVStore.java:1427) 2023-07-06T15:51:43.2182597Z at org.h2.mvstore.MVStore.store(MVStore.java:1334) 2023-07-06T15:51:43.2182939Z ... 6 more 2023-07-06T15:51:43.2184215Z Caused by: com.oracle.svm.core.jdk.UnsupportedFeatureError: SerializationConstructorAccessor class not found for declaringClass: java.util.concurrent.ConcurrentSkipListSet (targetConstructorClass: java.util.AbstractSet). Usually adding java.util.concurrent.ConcurrentSkipListSet to serialization-config.json fixes the problem. 2023-07-06T15:51:43.2185371Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.util.VMError.unsupportedFeature(VMError.java:92) 2023-07-06T15:51:43.2186351Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.reflect.serialize.SerializationSupport.getSerializationConstructorAccessor(SerializationSupport.java:143) 2023-07-06T15:51:43.2187439Z at java.base@17.0.7/jdk.internal.reflect.MethodAccessorGenerator.generateSerializationConstructor(MethodAccessorGenerator.java:61) 2023-07-06T15:51:43.2188304Z at java.base@17.0.7/jdk.internal.reflect.ReflectionFactory.generateConstructor(ReflectionFactory.java:463) 2023-07-06T15:51:43.2189111Z at java.base@17.0.7/jdk.internal.reflect.ReflectionFactory.newConstructorForSerialization(ReflectionFactory.java:455) 2023-07-06T15:51:43.2189887Z at java.base@17.0.7/java.io.ObjectStreamClass.getSerializableConstructor(ObjectStreamClass.java:1443) 2023-07-06T15:51:43.2190498Z at java.base@17.0.7/java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:412) 2023-07-06T15:51:43.2191003Z at java.base@17.0.7/java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:384) 2023-07-06T15:51:43.2191593Z at java.base@17.0.7/java.security.AccessController.executePrivileged(AccessController.java:171) 2023-07-06T15:51:43.2192233Z at java.base@17.0.7/java.security.AccessController.doPrivileged(AccessController.java:318) 2023-07-06T15:51:43.2192802Z at java.base@17.0.7/java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:384) 2023-07-06T15:51:43.2193385Z at java.base@17.0.7/java.io.ObjectStreamClass$Caches$1.computeValue(ObjectStreamClass.java:110) 2023-07-06T15:51:43.2194017Z at java.base@17.0.7/java.io.ObjectStreamClass$Caches$1.computeValue(ObjectStreamClass.java:107) 2023-07-06T15:51:43.2195550Z at java.base@17.0.7/java.io.ClassCache$1.computeValue(ClassCache.java:73) 2023-07-06T15:51:43.2195961Z at java.base@17.0.7/java.io.ClassCache$1.computeValue(ClassCache.java:70) 2023-07-06T15:51:43.2196389Z at java.base@17.0.7/java.lang.ClassValue.get(JavaLangSubstitutions.java:681) 2023-07-06T15:51:43.2196790Z at java.base@17.0.7/java.io.ClassCache.get(ClassCache.java:84) 2023-07-06T15:51:43.2197216Z at java.base@17.0.7/java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:363) 2023-07-06T15:51:43.2197720Z at java.base@17.0.7/java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1137) 2023-07-06T15:51:43.2198233Z at java.base@17.0.7/java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:350) 2023-07-06T15:51:43.2198983Z at org.h2.mvstore.type.ObjectDataType.serialize(ObjectDataType.java:356) 2023-07-06T15:51:43.2199343Z ... 14 more 2023-07-06T15:51:43.2200696Z 2023-07-06 15:51:43,064 ERROR [io.qua.ver.htt.run.QuarkusErrorHandler] (executor-thread-1) HTTP Request to /nitrite/repositoryClass?mappable=false failed, error id: 36f95039-4468-4295-a769-49c146d7d407-1: org.jboss.resteasy.spi.UnhandledException: org.apache.camel.CamelExecutionException: Exception occurred during execution on the exchange: Exchange[] 2023-07-06T15:51:43.2201712Z at org.jboss.resteasy.core.ExceptionHandler.handleApplicationException(ExceptionHandler.java:107) 2023-07-06T15:51:43.2202363Z at org.jboss.resteasy.core.ExceptionHandler.handleException(ExceptionHandler.java:344) 2023-07-06T15:51:43.2203006Z at org.jboss.resteasy.core.SynchronousDispatcher.writeException(SynchronousDispatcher.java:205) 2023-07-06T15:51:43.2203641Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:452) 2023-07-06T15:51:43.2204243Z at org.jboss.resteasy.core.SynchronousDispatcher.lambda$invoke$4(SynchronousDispatcher.java:240) 2023-07-06T15:51:43.2205009Z at org.jboss.resteasy.core.SynchronousDispatcher.lambda$preprocess$0(SynchronousDispatcher.java:154) 2023-07-06T15:51:43.2205810Z at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:321) 2023-07-06T15:51:43.2206598Z at org.jboss.resteasy.core.SynchronousDispatcher.preprocess(SynchronousDispatcher.java:157) 2023-07-06T15:51:43.2207202Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:229) 2023-07-06T15:51:43.2207839Z at io.quarkus.resteasy.runtime.standalone.RequestDispatcher.service(RequestDispatcher.java:82) 2023-07-06T15:51:43.2208538Z at io.quarkus.resteasy.runtime.standalone.VertxRequestHandler.dispatch(VertxRequestHandler.java:147) 2023-07-06T15:51:43.2209222Z at io.quarkus.resteasy.runtime.standalone.VertxRequestHandler$1.run(VertxRequestHandler.java:93) 2023-07-06T15:51:43.2209813Z at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:576) 2023-07-06T15:51:43.2210364Z at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2513) 2023-07-06T15:51:43.2210919Z at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1538) 2023-07-06T15:51:43.2211427Z at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29) 2023-07-06T15:51:43.2211987Z at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29) 2023-07-06T15:51:43.2212601Z at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30) 2023-07-06T15:51:43.2213074Z at java.base@17.0.7/java.lang.Thread.run(Thread.java:833) 2023-07-06T15:51:43.2213640Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.thread.PlatformThreads.threadStartRoutine(PlatformThreads.java:807) 2023-07-06T15:51:43.2214493Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.posix.thread.PosixPlatformThreads.pthreadStartRoutine(PosixPlatformThreads.java:210) 2023-07-06T15:51:43.2215255Z Caused by: org.apache.camel.CamelExecutionException: Exception occurred during execution on the exchange: Exchange[] 2023-07-06T15:51:43.2215862Z at org.apache.camel.support.ExchangeHelper.extractResultBody(ExchangeHelper.java:653) 2023-07-06T15:51:43.2216545Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:368) 2023-07-06T15:51:43.2217320Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:333) 2023-07-06T15:51:43.2218127Z at org.apache.camel.quarkus.component.nitrite.it.NitriteResource.postRepositoryClass(NitriteResource.java:87) 2023-07-06T15:51:43.2218916Z at org.apache.camel.quarkus.component.nitrite.it.NitriteResource_ClientProxy.postRepositoryClass(Unknown Source) 2023-07-06T15:51:43.2219474Z at java.base@17.0.7/java.lang.reflect.Method.invoke(Method.java:568) 2023-07-06T15:51:43.2220237Z at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:154) 2023-07-06T15:51:43.2220834Z at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:118) 2023-07-06T15:51:43.2221593Z at org.jboss.resteasy.core.ResourceMethodInvoker.internalInvokeOnTarget(ResourceMethodInvoker.java:560) 2023-07-06T15:51:43.2222359Z at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTargetAfterFilter(ResourceMethodInvoker.java:452) 2023-07-06T15:51:43.2223073Z at org.jboss.resteasy.core.ResourceMethodInvoker.lambda$invokeOnTarget$2(ResourceMethodInvoker.java:413) 2023-07-06T15:51:43.2223876Z at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:321) 2023-07-06T15:51:43.2224673Z at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTarget(ResourceMethodInvoker.java:415) 2023-07-06T15:51:43.2225303Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:378) 2023-07-06T15:51:43.2225895Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:356) 2023-07-06T15:51:43.2226579Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:70) 2023-07-06T15:51:43.2227163Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:429) 2023-07-06T15:51:43.2227556Z ... 17 more 2023-07-06T15:51:43.2227921Z Caused by: org.dizitart.no2.exceptions.NitriteIOException: NO2.2004: store is closed 2023-07-06T15:51:43.2228524Z at org.dizitart.no2.internals.DefaultNitriteCollection.checkOpened(DefaultNitriteCollection.java:394) 2023-07-06T15:51:43.2229214Z at org.dizitart.no2.internals.DefaultNitriteCollection.update(DefaultNitriteCollection.java:320) 2023-07-06T15:51:43.2229861Z at org.dizitart.no2.objects.DefaultObjectRepository.update(DefaultObjectRepository.java:141) 2023-07-06T15:51:43.2230489Z at org.dizitart.no2.objects.DefaultObjectRepository.update(DefaultObjectRepository.java:126) 2023-07-06T15:51:43.2231180Z at org.apache.camel.component.nitrite.operation.common.UpsertOperation.execute(UpsertOperation.java:41) 2023-07-06T15:51:43.2231880Z at org.apache.camel.component.nitrite.NitriteProducer.process(NitriteProducer.java:78) 2023-07-06T15:51:43.2232637Z at org.apache.camel.support.AsyncProcessorConverterHelper$ProcessorToAsyncProcessorBridge.process(AsyncProcessorConverterHelper.java:66) 2023-07-06T15:51:43.2233453Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:214) 2023-07-06T15:51:43.2234191Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor$1.process(SharedCamelInternalProcessor.java:111) 2023-07-06T15:51:43.2234982Z at org.apache.camel.impl.engine.DefaultAsyncProcessorAwaitManager.process(DefaultAsyncProcessorAwaitManager.java:81) 2023-07-06T15:51:43.2235782Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:108) 2023-07-06T15:51:43.2236467Z at org.apache.camel.support.cache.DefaultProducerCache.send(DefaultProducerCache.java:180) 2023-07-06T15:51:43.2237116Z at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:176) 2023-07-06T15:51:43.2237769Z at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:172) 2023-07-06T15:51:43.2238490Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:360) 2023-07-06T15:51:43.2239137Z ... 32 more ```
non_defect
nitrite native integration test could not serialize no₂ bug description for some reason this has recently popped up on the ci builds t e s t s running org apache camel quarkus component nitrite it nitriteit native image path was not set making a guess for the correct path of native image guessed path home runner work camel quarkus camel quarkus integration tests nitrite target camel quarkus integration test nitrite snapshot runner executing home runner work camel quarkus camel quarkus integration tests nitrite target camel quarkus integration test nitrite snapshot runner dquarkus http port dquarkus http ssl port dtest url dquarkus log file path home runner work camel quarkus camel quarkus integration tests nitrite target quarkus log dquarkus log file enable true dquarkus log category io quarkus level info dcamel quarkus nitrite test db file tmp nitritetestresource db file info main bootstrap runtime org apache camel quarkus main camelmainruntime info main apache camel main is starting info main apache camel camel is starting info main routes startup started info main apache camel camel started in build init start info main camel quarkus integration test nitrite snapshot native powered by quarkus final started in listening on info main profile prod activated info main installed features exception in thread mvstore background writer nio tmp nitritetestresource db file java lang illegalstateexception java lang illegalargumentexception could not serialize no₂ at org mvstore datautils newillegalstateexception datautils java at org mvstore mvstore store mvstore java at org mvstore mvstore store mvstore java at org mvstore mvstore trycommit mvstore java at org mvstore mvstore writeinbackground mvstore java at org mvstore mvstore backgroundwriterthread run mvstore java at org graalvm nativeimage builder com oracle svm core thread platformthreads threadstartroutine platformthreads java at org graalvm nativeimage builder com oracle svm core posix thread posixplatformthreads pthreadstartroutine posixplatformthreads java caused by java lang illegalargumentexception could not serialize no₂ at org mvstore datautils newillegalargumentexception datautils java at org mvstore type objectdatatype serialize objectdatatype java at org mvstore type objectdatatype serializedobjecttype write objectdatatype java at org mvstore type objectdatatype write objectdatatype java at org mvstore type objectdatatype write objectdatatype java at org mvstore page leaf writevalues page java at org mvstore page write page java at org mvstore page leaf writeunsavedrecursive page java at org mvstore mvstore storenow mvstore java at org mvstore mvstore store mvstore java more caused by com oracle svm core jdk unsupportedfeatureerror serializationconstructoraccessor class not found for declaringclass java util concurrent concurrentskiplistset targetconstructorclass java util abstractset usually adding java util concurrent concurrentskiplistset to serialization config json fixes the problem at org graalvm nativeimage builder com oracle svm core util vmerror unsupportedfeature vmerror java at org graalvm nativeimage builder com oracle svm core reflect serialize serializationsupport getserializationconstructoraccessor serializationsupport java at java base jdk internal reflect methodaccessorgenerator generateserializationconstructor methodaccessorgenerator java at java base jdk internal reflect reflectionfactory generateconstructor reflectionfactory java at java base jdk internal reflect reflectionfactory newconstructorforserialization reflectionfactory java at java base java io objectstreamclass getserializableconstructor objectstreamclass java at java base java io objectstreamclass run objectstreamclass java at java base java io objectstreamclass run objectstreamclass java at java base java security accesscontroller executeprivileged accesscontroller java at java base java security accesscontroller doprivileged accesscontroller java at java base java io objectstreamclass objectstreamclass java at java base java io objectstreamclass caches computevalue objectstreamclass java at java base java io objectstreamclass caches computevalue objectstreamclass java at java base java io classcache computevalue classcache java at java base java io classcache computevalue classcache java at java base java lang classvalue get javalangsubstitutions java at java base java io classcache get classcache java at java base java io objectstreamclass lookup objectstreamclass java at java base java io objectoutputstream objectoutputstream java at java base java io objectoutputstream writeobject objectoutputstream java at org mvstore type objectdatatype serialize objectdatatype java more error executor thread http request to nitrite repositoryclass mappable false failed error id org jboss resteasy spi unhandledexception org apache camel camelexecutionexception exception occurred during execution on the exchange exchange at org jboss resteasy core exceptionhandler handleapplicationexception exceptionhandler java at org jboss resteasy core exceptionhandler handleexception exceptionhandler java at org jboss resteasy core synchronousdispatcher writeexception synchronousdispatcher java at org jboss resteasy core synchronousdispatcher invoke synchronousdispatcher java at org jboss resteasy core synchronousdispatcher lambda invoke synchronousdispatcher java at org jboss resteasy core synchronousdispatcher lambda preprocess synchronousdispatcher java at org jboss resteasy core interception jaxrs prematchcontainerrequestcontext filter prematchcontainerrequestcontext java at org jboss resteasy core synchronousdispatcher preprocess synchronousdispatcher java at org jboss resteasy core synchronousdispatcher invoke synchronousdispatcher java at io quarkus resteasy runtime standalone requestdispatcher service requestdispatcher java at io quarkus resteasy runtime standalone vertxrequesthandler dispatch vertxrequesthandler java at io quarkus resteasy runtime standalone vertxrequesthandler run vertxrequesthandler java at io quarkus vertx core runtime vertxcorerecorder runwith vertxcorerecorder java at org jboss threads enhancedqueueexecutor task run enhancedqueueexecutor java at org jboss threads enhancedqueueexecutor threadbody run enhancedqueueexecutor java at org jboss threads delegatingrunnable run delegatingrunnable java at org jboss threads threadlocalresettingrunnable run threadlocalresettingrunnable java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java at org graalvm nativeimage builder com oracle svm core thread platformthreads threadstartroutine platformthreads java at org graalvm nativeimage builder com oracle svm core posix thread posixplatformthreads pthreadstartroutine posixplatformthreads java caused by org apache camel camelexecutionexception exception occurred during execution on the exchange exchange at org apache camel support exchangehelper extractresultbody exchangehelper java at org apache camel impl engine defaultfluentproducertemplate request defaultfluentproducertemplate java at org apache camel impl engine defaultfluentproducertemplate request defaultfluentproducertemplate java at org apache camel quarkus component nitrite it nitriteresource postrepositoryclass nitriteresource java at org apache camel quarkus component nitrite it nitriteresource clientproxy postrepositoryclass unknown source at java base java lang reflect method invoke method java at org jboss resteasy core methodinjectorimpl invoke methodinjectorimpl java at org jboss resteasy core methodinjectorimpl invoke methodinjectorimpl java at org jboss resteasy core resourcemethodinvoker internalinvokeontarget resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker invokeontargetafterfilter resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker lambda invokeontarget resourcemethodinvoker java at org jboss resteasy core interception jaxrs prematchcontainerrequestcontext filter prematchcontainerrequestcontext java at org jboss resteasy core resourcemethodinvoker invokeontarget resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker invoke resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker invoke resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker invoke resourcemethodinvoker java at org jboss resteasy core synchronousdispatcher invoke synchronousdispatcher java more caused by org dizitart exceptions nitriteioexception store is closed at org dizitart internals defaultnitritecollection checkopened defaultnitritecollection java at org dizitart internals defaultnitritecollection update defaultnitritecollection java at org dizitart objects defaultobjectrepository update defaultobjectrepository java at org dizitart objects defaultobjectrepository update defaultobjectrepository java at org apache camel component nitrite operation common upsertoperation execute upsertoperation java at org apache camel component nitrite nitriteproducer process nitriteproducer java at org apache camel support asyncprocessorconverterhelper processortoasyncprocessorbridge process asyncprocessorconverterhelper java at org apache camel impl engine sharedcamelinternalprocessor process sharedcamelinternalprocessor java at org apache camel impl engine sharedcamelinternalprocessor process sharedcamelinternalprocessor java at org apache camel impl engine defaultasyncprocessorawaitmanager process defaultasyncprocessorawaitmanager java at org apache camel impl engine sharedcamelinternalprocessor process sharedcamelinternalprocessor java at org apache camel support cache defaultproducercache send defaultproducercache java at org apache camel impl engine defaultproducertemplate send defaultproducertemplate java at org apache camel impl engine defaultproducertemplate send defaultproducertemplate java at org apache camel impl engine defaultfluentproducertemplate request defaultfluentproducertemplate java more
0
5,095
26,007,472,462
IssuesEvent
2022-12-20 20:56:11
aws/aws-sam-cli
https://api.github.com/repos/aws/aws-sam-cli
closed
[Feature Request] Cognito Offline Testing
type/feature area/local/start-api stage/pm-review maintainer/need-followup
Just curious whether SAM will support cognito offline testing? Pasting in some links to the same feature request in various other forums: - https://github.com/localstack/localstack/issues/26 - https://stackoverflow.com/questions/41633800/use-cognito-locally-offline - https://stackoverflow.com/questions/51719765/invoke-lambda-function-without-authorization-on-local-serverless-offline-environ
True
[Feature Request] Cognito Offline Testing - Just curious whether SAM will support cognito offline testing? Pasting in some links to the same feature request in various other forums: - https://github.com/localstack/localstack/issues/26 - https://stackoverflow.com/questions/41633800/use-cognito-locally-offline - https://stackoverflow.com/questions/51719765/invoke-lambda-function-without-authorization-on-local-serverless-offline-environ
non_defect
cognito offline testing just curious whether sam will support cognito offline testing pasting in some links to the same feature request in various other forums
0
10,291
2,622,141,104
IssuesEvent
2015-03-04 00:02:03
byzhang/spserver
https://api.github.com/repos/byzhang/spserver
closed
关于超时,是怎么回事?
auto-migrated Priority-Medium Type-Defect
``` 0.9.5的版本。 SP_IocpEventHelper :: timeoutNext 有以下代码: DWORD ret = ( first->tv_sec - curr.tv_sec ) * 1000 + ( first->tv_usec - curr.tv_usec ) / 1000; if( ret < 0 ) ret = 0; return ret; ret 永远都会是>=0 的。 ``` Original issue reported on code.google.com by `xChenJun` on 17 Jul 2010 at 4:04
1.0
关于超时,是怎么回事? - ``` 0.9.5的版本。 SP_IocpEventHelper :: timeoutNext 有以下代码: DWORD ret = ( first->tv_sec - curr.tv_sec ) * 1000 + ( first->tv_usec - curr.tv_usec ) / 1000; if( ret < 0 ) ret = 0; return ret; ret 永远都会是>=0 的。 ``` Original issue reported on code.google.com by `xChenJun` on 17 Jul 2010 at 4:04
defect
关于超时,是怎么回事? 。 sp iocpeventhelper timeoutnext 有以下代码: dword ret first tv sec curr tv sec first tv usec curr tv usec if ret ret return ret ret 永远都会是 的。 original issue reported on code google com by xchenjun on jul at
1
33,840
7,267,597,152
IssuesEvent
2018-02-20 06:10:42
colour-science/colour
https://api.github.com/repos/colour-science/colour
closed
Python 3 Syntax Error in "docs/conf.py" file.
Defect Distribution Documentation Minor
flake8 testing of https://github.com/colour-science/colour on Python 3.6.4 $ __flake8 . --count --select=E901,E999,F821,F822,F823 --show-source --statistics__ ``` ./docs/conf.py:219:0: E999 SyntaxError: (unicode error) 'unicodeescape' codec can't decode bytes in position 9-10: truncated \uXXXX escape """, ^ ./utilities/unicode_to_ascii.py:9:1: F821 undefined name 'reload' reload(sys) ^ 1 E999 SyntaxError: (unicode error) 'unicodeescape' codec can't decode bytes in position 9-10: truncated \uXXXX escape 1 F821 undefined name 'reload' 2 ```
1.0
Python 3 Syntax Error in "docs/conf.py" file. - flake8 testing of https://github.com/colour-science/colour on Python 3.6.4 $ __flake8 . --count --select=E901,E999,F821,F822,F823 --show-source --statistics__ ``` ./docs/conf.py:219:0: E999 SyntaxError: (unicode error) 'unicodeescape' codec can't decode bytes in position 9-10: truncated \uXXXX escape """, ^ ./utilities/unicode_to_ascii.py:9:1: F821 undefined name 'reload' reload(sys) ^ 1 E999 SyntaxError: (unicode error) 'unicodeescape' codec can't decode bytes in position 9-10: truncated \uXXXX escape 1 F821 undefined name 'reload' 2 ```
defect
python syntax error in docs conf py file testing of on python count select show source statistics docs conf py syntaxerror unicode error unicodeescape codec can t decode bytes in position truncated uxxxx escape utilities unicode to ascii py undefined name reload reload sys syntaxerror unicode error unicodeescape codec can t decode bytes in position truncated uxxxx escape undefined name reload
1
471,417
13,566,326,802
IssuesEvent
2020-09-18 13:07:59
willowtreeapps/vocable-ios
https://api.github.com/repos/willowtreeapps/vocable-ios
closed
Able to save blank string to my sayings
bug priority - low v1.3.3
On the keyboard screen, the "my sayings" button should be disabled when no text is present Steps to reproduce: 1. navigate to the keyboard screen 2. press the star (my sayings) button expected: - button is disabled actual: - blank string is saved to my sayings
1.0
Able to save blank string to my sayings - On the keyboard screen, the "my sayings" button should be disabled when no text is present Steps to reproduce: 1. navigate to the keyboard screen 2. press the star (my sayings) button expected: - button is disabled actual: - blank string is saved to my sayings
non_defect
able to save blank string to my sayings on the keyboard screen the my sayings button should be disabled when no text is present steps to reproduce navigate to the keyboard screen press the star my sayings button expected button is disabled actual blank string is saved to my sayings
0
79,461
28,289,299,638
IssuesEvent
2023-04-09 01:53:08
HewlettPackard/yoda-demo
https://api.github.com/repos/HewlettPackard/yoda-demo
closed
Issue editing text on Mondays.
S3 - Medium T1 - Defect no-issue-activity
<p>Lorem ipsum nam nulla duis aenean facilisis venenatis viverra eleifend, sollicitudin sodales nisi primis sollicitudin mi ad.</p> <p>Purus convallis lorem eget hac curabitur fusce duis himenaeos pulvinar maecenas, feugiat tempor fusce netus enim lacinia nisi et nisi.</p> <p>At nisi elit vivamus feugiat orci non, tellus elit sagittis erat ac, sodales elit mauris aliquet fringilla.</p> <p>Pharetra vitae habitant tristique scelerisque id mollis odio tempus, venenatis fusce arcu dictumst massa curabitur dolor.</p> > estimate 5
1.0
Issue editing text on Mondays. - <p>Lorem ipsum nam nulla duis aenean facilisis venenatis viverra eleifend, sollicitudin sodales nisi primis sollicitudin mi ad.</p> <p>Purus convallis lorem eget hac curabitur fusce duis himenaeos pulvinar maecenas, feugiat tempor fusce netus enim lacinia nisi et nisi.</p> <p>At nisi elit vivamus feugiat orci non, tellus elit sagittis erat ac, sodales elit mauris aliquet fringilla.</p> <p>Pharetra vitae habitant tristique scelerisque id mollis odio tempus, venenatis fusce arcu dictumst massa curabitur dolor.</p> > estimate 5
defect
issue editing text on mondays lorem ipsum nam nulla duis aenean facilisis venenatis viverra eleifend sollicitudin sodales nisi primis sollicitudin mi ad purus convallis lorem eget hac curabitur fusce duis himenaeos pulvinar maecenas feugiat tempor fusce netus enim lacinia nisi et nisi at nisi elit vivamus feugiat orci non tellus elit sagittis erat ac sodales elit mauris aliquet fringilla pharetra vitae habitant tristique scelerisque id mollis odio tempus venenatis fusce arcu dictumst massa curabitur dolor estimate
1
53,786
13,262,298,332
IssuesEvent
2020-08-20 21:29:44
icecube-trac/tix4
https://api.github.com/repos/icecube-trac/tix4
closed
Assertion "`bin_edges[q+1]-bin_edges[q] > 0' failed." triggered in MillipedeDOMCacheMap.cxx:247 (Trac #2092)
Migrated from Trac combo reconstruction defect
Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion: /home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed. Unfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2092">https://code.icecube.wisc.edu/projects/icecube/ticket/2092</a>, reported by claudio.kopperand owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "_ts": "1550067323910946", "description": "Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion:\n\n/home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed.\n\nUnfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered?", "reporter": "claudio.kopper", "cc": "austin.schneider", "resolution": "insufficient resources", "time": "2017-09-15T16:30:44", "component": "combo reconstruction", "summary": "Assertion \"`bin_edges[q+1]-bin_edges[q] > 0' failed.\" triggered in MillipedeDOMCacheMap.cxx:247", "priority": "normal", "keywords": "", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
1.0
Assertion "`bin_edges[q+1]-bin_edges[q] > 0' failed." triggered in MillipedeDOMCacheMap.cxx:247 (Trac #2092) - Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion: /home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed. Unfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered? <details> <summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2092">https://code.icecube.wisc.edu/projects/icecube/ticket/2092</a>, reported by claudio.kopperand owned by jvansanten</em></summary> <p> ```json { "status": "closed", "changetime": "2019-02-13T14:15:23", "_ts": "1550067323910946", "description": "Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion:\n\n/home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed.\n\nUnfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered?", "reporter": "claudio.kopper", "cc": "austin.schneider", "resolution": "insufficient resources", "time": "2017-09-15T16:30:44", "component": "combo reconstruction", "summary": "Assertion \"`bin_edges[q+1]-bin_edges[q] > 0' failed.\" triggered in MillipedeDOMCacheMap.cxx:247", "priority": "normal", "keywords": "", "milestone": "", "owner": "jvansanten", "type": "defect" } ``` </p> </details>
defect
assertion bin edges bin edges failed triggered in millipededomcachemap cxx trac austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion home aschneider combo source millipede private millipede millipededomcachemap cxx int rebin bayesian blocks int const double const double bool double double double double assertion bin edges bin edges failed unfortunately we do not have a minimal example right now i guess distribute needs better reporting or we would need to add logging but maybe it is obvious why this assertion could be triggered migrated from json status closed changetime ts description austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion n n home aschneider combo source millipede private millipede millipededomcachemap cxx int rebin bayesian blocks int const double const double bool double double double double assertion bin edges bin edges failed n nunfortunately we do not have a minimal example right now i guess distribute needs better reporting or we would need to add logging but maybe it is obvious why this assertion could be triggered reporter claudio kopper cc austin schneider resolution insufficient resources time component combo reconstruction summary assertion bin edges bin edges failed triggered in millipededomcachemap cxx priority normal keywords milestone owner jvansanten type defect
1
69,622
15,021,855,863
IssuesEvent
2021-02-01 16:16:44
fabric8-services/fabric8-wit
https://api.github.com/repos/fabric8-services/fabric8-wit
closed
It should be possible to make PostgreSQL database connection secure through TLS
:snake: security database
PostgreSQL has TLS support. It should be possible to make PostgreSQL database connection secure through TLS. **The default configuration can continue to be disabled**. BTW, TLS can be used with RDS also: http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_PostgreSQL.html#PostgreSQL.Concepts.General.SSL
True
It should be possible to make PostgreSQL database connection secure through TLS - PostgreSQL has TLS support. It should be possible to make PostgreSQL database connection secure through TLS. **The default configuration can continue to be disabled**. BTW, TLS can be used with RDS also: http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_PostgreSQL.html#PostgreSQL.Concepts.General.SSL
non_defect
it should be possible to make postgresql database connection secure through tls postgresql has tls support it should be possible to make postgresql database connection secure through tls the default configuration can continue to be disabled btw tls can be used with rds also
0
49,117
13,185,241,592
IssuesEvent
2020-08-12 21:00:25
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
opened
[sim-services] (Trac #781)
Incomplete Migration Migrated from Trac combo simulation defect
<details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/781 , reported by olivas and owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:03", "description": "Make sure the sanity checkers are solid, all have tests, and the testing scripts are enabled.", "reporter": "olivas", "cc": "", "resolution": "fixed", "_ts": "1458335643235016", "component": "combo simulation", "summary": "[sim-services]", "priority": "critical", "keywords": "", "time": "2014-10-11T19:02:30", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
1.0
[sim-services] (Trac #781) - <details> <summary><em>Migrated from https://code.icecube.wisc.edu/ticket/781 , reported by olivas and owned by olivas</em></summary> <p> ```json { "status": "closed", "changetime": "2016-03-18T21:14:03", "description": "Make sure the sanity checkers are solid, all have tests, and the testing scripts are enabled.", "reporter": "olivas", "cc": "", "resolution": "fixed", "_ts": "1458335643235016", "component": "combo simulation", "summary": "[sim-services]", "priority": "critical", "keywords": "", "time": "2014-10-11T19:02:30", "milestone": "", "owner": "olivas", "type": "defect" } ``` </p> </details>
defect
trac migrated from reported by olivas and owned by olivas json status closed changetime description make sure the sanity checkers are solid all have tests and the testing scripts are enabled reporter olivas cc resolution fixed ts component combo simulation summary priority critical keywords time milestone owner olivas type defect
1
139,764
18,853,807,585
IssuesEvent
2021-11-12 01:46:30
LalithK90/aroComputers
https://api.github.com/repos/LalithK90/aroComputers
opened
CVE-2021-22118 (High) detected in spring-web-5.2.3.RELEASE.jar
security vulnerability
## CVE-2021-22118 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.2.3.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: aroComputers/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-web/5.2.3.RELEASE/dd386a02e40b915ab400a3bf9f586d2dc4c0852c/spring-web-5.2.3.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.4.RELEASE.jar (Root Library) - spring-boot-starter-json-2.2.4.RELEASE.jar - :x: **spring-web-5.2.3.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 5.2.x prior to 5.2.15 and versions 5.3.x prior to 5.3.7, a WebFlux application is vulnerable to a privilege escalation: by (re)creating the temporary storage directory, a locally authenticated malicious user can read or modify files that have been uploaded to the WebFlux application, or overwrite arbitrary files with multipart request data. <p>Publish Date: 2021-05-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22118>CVE-2021-22118</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22118">https://tanzu.vmware.com/security/cve-2021-22118</a></p> <p>Release Date: 2021-05-27</p> <p>Fix Resolution: org.springframework:spring-web:5.2.15,5.3.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-22118 (High) detected in spring-web-5.2.3.RELEASE.jar - ## CVE-2021-22118 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.2.3.RELEASE.jar</b></p></summary> <p>Spring Web</p> <p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p> <p>Path to dependency file: aroComputers/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-web/5.2.3.RELEASE/dd386a02e40b915ab400a3bf9f586d2dc4c0852c/spring-web-5.2.3.RELEASE.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.2.4.RELEASE.jar (Root Library) - spring-boot-starter-json-2.2.4.RELEASE.jar - :x: **spring-web-5.2.3.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 5.2.x prior to 5.2.15 and versions 5.3.x prior to 5.3.7, a WebFlux application is vulnerable to a privilege escalation: by (re)creating the temporary storage directory, a locally authenticated malicious user can read or modify files that have been uploaded to the WebFlux application, or overwrite arbitrary files with multipart request data. <p>Publish Date: 2021-05-27 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22118>CVE-2021-22118</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22118">https://tanzu.vmware.com/security/cve-2021-22118</a></p> <p>Release Date: 2021-05-27</p> <p>Fix Resolution: org.springframework:spring-web:5.2.15,5.3.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in spring web release jar cve high severity vulnerability vulnerable library spring web release jar spring web library home page a href path to dependency file arocomputers build gradle path to vulnerable library home wss scanner gradle caches modules files org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x spring web release jar vulnerable library found in base branch master vulnerability details in spring framework versions x prior to and versions x prior to a webflux application is vulnerable to a privilege escalation by re creating the temporary storage directory a locally authenticated malicious user can read or modify files that have been uploaded to the webflux application or overwrite arbitrary files with multipart request data publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring web step up your open source security game with whitesource
0
63,066
17,365,970,475
IssuesEvent
2021-07-30 07:19:44
vector-im/element-web
https://api.github.com/repos/vector-im/element-web
closed
no UI when switching room whilst setting up a voice/video call
A-VoIP P1 S-Tolerable T-Defect
If you set up a voice/video call and whilst it's ringing switch to a different room, there is zero UI to tell you what's going on. We should at least show the 'ongoing call' UI at the top of LeftPanel
1.0
no UI when switching room whilst setting up a voice/video call - If you set up a voice/video call and whilst it's ringing switch to a different room, there is zero UI to tell you what's going on. We should at least show the 'ongoing call' UI at the top of LeftPanel
defect
no ui when switching room whilst setting up a voice video call if you set up a voice video call and whilst it s ringing switch to a different room there is zero ui to tell you what s going on we should at least show the ongoing call ui at the top of leftpanel
1
56,190
23,720,740,872
IssuesEvent
2022-08-30 15:09:04
hashicorp/terraform-provider-azurerm
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
closed
Error: no schema available for azurerm_traffic_manager_endpoint.secondary_endpoint[0] while reading state;
question service/traffic-manager
### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note <!--- Please keep this note for the community ---> * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version 1.2.8 ### AzureRM Provider Version 3.19.1 ### Affected Resource(s)/Data Source(s) traffic_manager_endpoint ### Terraform Configuration Files ```hcl resource "azurerm_traffic_manager_azure_endpoint" "primary_endpoint" { count = var.env == "prod" || var.env == "uat" ? 1 : 0 name = "Primary" profile_id = azurerm_traffic_manager_profile.gsr_spoke_tm[0].id weight = 1 priority = 1 target_resource_id = data.azurerm_public_ip.gsr_primary_appgw_uksouth_ip[0].id } resource "azurerm_traffic_manager_azure_endpoint" "secondary_endpoint" { count = var.env == "prod" ? 1 : 0 name = "Secondary" profile_id = azurerm_traffic_manager_profile.gsr_spoke_tm[0].id weight = 1 priority = 2 target_resource_id = data.azurerm_public_ip.gsr_secondary_appgw_ukwest_ip[0].id } ``` ### Debug Output/Panic Output ```shell Error: no schema available for azurerm_traffic_manager_endpoint.secondary_endpoint[0] while reading state; this is a bug in Terraform and should be reported ``` ### Expected Behaviour _No response_ ### Actual Behaviour _No response_ ### Steps to Reproduce _No response_ ### Important Factoids _No response_ ### References _No response_
1.0
Error: no schema available for azurerm_traffic_manager_endpoint.secondary_endpoint[0] while reading state; - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Community Note <!--- Please keep this note for the community ---> * Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request * Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request * If you are interested in working on this issue or have submitted a pull request, please leave a comment <!--- Thank you for keeping this note for the community ---> ### Terraform Version 1.2.8 ### AzureRM Provider Version 3.19.1 ### Affected Resource(s)/Data Source(s) traffic_manager_endpoint ### Terraform Configuration Files ```hcl resource "azurerm_traffic_manager_azure_endpoint" "primary_endpoint" { count = var.env == "prod" || var.env == "uat" ? 1 : 0 name = "Primary" profile_id = azurerm_traffic_manager_profile.gsr_spoke_tm[0].id weight = 1 priority = 1 target_resource_id = data.azurerm_public_ip.gsr_primary_appgw_uksouth_ip[0].id } resource "azurerm_traffic_manager_azure_endpoint" "secondary_endpoint" { count = var.env == "prod" ? 1 : 0 name = "Secondary" profile_id = azurerm_traffic_manager_profile.gsr_spoke_tm[0].id weight = 1 priority = 2 target_resource_id = data.azurerm_public_ip.gsr_secondary_appgw_ukwest_ip[0].id } ``` ### Debug Output/Panic Output ```shell Error: no schema available for azurerm_traffic_manager_endpoint.secondary_endpoint[0] while reading state; this is a bug in Terraform and should be reported ``` ### Expected Behaviour _No response_ ### Actual Behaviour _No response_ ### Steps to Reproduce _No response_ ### Important Factoids _No response_ ### References _No response_
non_defect
error no schema available for azurerm traffic manager endpoint secondary endpoint while reading state is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s traffic manager endpoint terraform configuration files hcl resource azurerm traffic manager azure endpoint primary endpoint count var env prod var env uat name primary profile id azurerm traffic manager profile gsr spoke tm id weight priority target resource id data azurerm public ip gsr primary appgw uksouth ip id resource azurerm traffic manager azure endpoint secondary endpoint count var env prod name secondary profile id azurerm traffic manager profile gsr spoke tm id weight priority target resource id data azurerm public ip gsr secondary appgw ukwest ip id debug output panic output shell error no schema available for azurerm traffic manager endpoint secondary endpoint while reading state this is a bug in terraform and should be reported expected behaviour no response actual behaviour no response steps to reproduce no response important factoids no response references no response
0
18,320
24,438,849,687
IssuesEvent
2022-10-06 13:22:41
Ultimate-Hosts-Blacklist/whitelist
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
opened
[FALSE-POSITIVE?]
whitelisting process
**Domains or links** js.pusher.com **More Information** How did you discover your web site or domain was listed here? 2. Reported by another user. **Have you requested removal from other sources?** No **Additional context** Add any other context about the problem here. :exclamation: We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
1.0
[FALSE-POSITIVE?] - **Domains or links** js.pusher.com **More Information** How did you discover your web site or domain was listed here? 2. Reported by another user. **Have you requested removal from other sources?** No **Additional context** Add any other context about the problem here. :exclamation: We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
non_defect
domains or links js pusher com more information how did you discover your web site or domain was listed here reported by another user have you requested removal from other sources no additional context add any other context about the problem here exclamation we understand being listed on a list like this can be frustrating and embarrassing for many web site owners the first step is to remain calm the second step is to rest assured one of our maintainers will address your issue as soon as possible please make sure you have provided as much information as possible to help speed up the process
0
56,435
15,091,694,926
IssuesEvent
2021-02-06 16:31:56
dkfurrow/moneydance-investment-reports
https://api.github.com/repos/dkfurrow/moneydance-investment-reports
closed
Test initial issue report
Priority-Medium Type-Defect auto-migrated
``` Test initial issue report ``` Original issue reported on code.google.com by `dkfur...@gmail.com` on 23 Dec 2010 at 4:22
1.0
Test initial issue report - ``` Test initial issue report ``` Original issue reported on code.google.com by `dkfur...@gmail.com` on 23 Dec 2010 at 4:22
defect
test initial issue report test initial issue report original issue reported on code google com by dkfur gmail com on dec at
1
6,242
2,610,223,969
IssuesEvent
2015-02-26 19:10:59
chrsmith/somefinders
https://api.github.com/repos/chrsmith/somefinders
opened
kh 1171 инструкция на русском
auto-migrated Priority-Medium Type-Defect
``` '''Антонин Евсеев''' Привет всем не подскажите где можно найти .kh 1171 инструкция на русском. где то видел уже '''Боеслав Овчинников''' Вот держи линк http://bit.ly/16T1P32 '''Витольд Козлов''' Спасибо вроде то но просит телефон вводить '''Анастасий Калинин''' Не это не влияет на баланс '''Аверкий Пахомов''' Неа все ок у меня ничего не списало Информация о файле: kh 1171 инструкция на русском Загружен: В этом месяце Скачан раз: 425 Рейтинг: 272 Средняя скорость скачивания: 693 Похожих файлов: 37 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 8:29
1.0
kh 1171 инструкция на русском - ``` '''Антонин Евсеев''' Привет всем не подскажите где можно найти .kh 1171 инструкция на русском. где то видел уже '''Боеслав Овчинников''' Вот держи линк http://bit.ly/16T1P32 '''Витольд Козлов''' Спасибо вроде то но просит телефон вводить '''Анастасий Калинин''' Не это не влияет на баланс '''Аверкий Пахомов''' Неа все ок у меня ничего не списало Информация о файле: kh 1171 инструкция на русском Загружен: В этом месяце Скачан раз: 425 Рейтинг: 272 Средняя скорость скачивания: 693 Похожих файлов: 37 ``` ----- Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 8:29
defect
kh инструкция на русском антонин евсеев привет всем не подскажите где можно найти kh инструкция на русском где то видел уже боеслав овчинников вот держи линк витольд козлов спасибо вроде то но просит телефон вводить анастасий калинин не это не влияет на баланс аверкий пахомов неа все ок у меня ничего не списало информация о файле kh инструкция на русском загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
1
32,647
6,884,884,161
IssuesEvent
2017-11-21 14:31:29
tpfinal-pp1/tp-final
https://api.github.com/repos/tpfinal-pp1/tp-final
reopened
Contrato: Error en fecha de ingreso
bug Defecto medio Liberado por desarrollo
No deberia pasar (a mi criterio) que la fecha de celebracion sea despues de la fecha de ingreso ![image.png](https://images.zenhubusercontent.com/59c0646db0222d5de47835ea/7e044a8f-915c-44e4-b776-b9ab9c9de09c)
1.0
Contrato: Error en fecha de ingreso - No deberia pasar (a mi criterio) que la fecha de celebracion sea despues de la fecha de ingreso ![image.png](https://images.zenhubusercontent.com/59c0646db0222d5de47835ea/7e044a8f-915c-44e4-b776-b9ab9c9de09c)
defect
contrato error en fecha de ingreso no deberia pasar a mi criterio que la fecha de celebracion sea despues de la fecha de ingreso
1
18,283
3,040,177,737
IssuesEvent
2015-08-07 14:04:11
googlei18n/noto-fonts
https://api.github.com/repos/googlei18n/noto-fonts
opened
Imperial Aramaic: Tip of 100 and 10,000 characters should point right
Script-ImperialAramaic Type-Defect
moved from https://github.com/googlei18n/noto-alpha/issues/191 Imported from [Google Code issue #191](https://code.google.com/p/noto-alpha/issues/detail?id=191) created by [roozbeh@google.com](https://code.google.com/u/roozbeh@google.com/) on 2014-04-02T22:12:45.000Z: ---- The tip of ONE HUNDRED (U+1085D) and TEN THOUSAND (U+1085F) characters currently point to the left side. They should point to the right. Compare with Figures 13 and 14 of the Unicode proposal for Imperial Aramaic, where attestation of the numbers are given: http://www.unicode.org/L2/L2007/07197r2-n3273r2-aramaic.pdf Also, compare with the Unicode chart.
1.0
Imperial Aramaic: Tip of 100 and 10,000 characters should point right - moved from https://github.com/googlei18n/noto-alpha/issues/191 Imported from [Google Code issue #191](https://code.google.com/p/noto-alpha/issues/detail?id=191) created by [roozbeh@google.com](https://code.google.com/u/roozbeh@google.com/) on 2014-04-02T22:12:45.000Z: ---- The tip of ONE HUNDRED (U+1085D) and TEN THOUSAND (U+1085F) characters currently point to the left side. They should point to the right. Compare with Figures 13 and 14 of the Unicode proposal for Imperial Aramaic, where attestation of the numbers are given: http://www.unicode.org/L2/L2007/07197r2-n3273r2-aramaic.pdf Also, compare with the Unicode chart.
defect
imperial aramaic tip of and characters should point right moved from imported from created by on the tip of one hundred u and ten thousand u characters currently point to the left side they should point to the right compare with figures and of the unicode proposal for imperial aramaic where attestation of the numbers are given also compare with the unicode chart
1
72,903
24,351,475,056
IssuesEvent
2022-10-03 00:45:50
scipy/scipy
https://api.github.com/repos/scipy/scipy
closed
BUG: 1.9.0rc1: `OptimizeResult` not populated when `optimize.milp` runs into `time_limit` even though a feasible solution was found.
defect scipy.optimize
### Describe your issue. When `optimize.milp` terminates due to the time limit being reached, the `OptimizeResult` object is not populated, even though one or more feasible solutions have been found. Example HiGHS log below shows that a feasible solution has been found (see solution status) when the 1 second time limit is reached. [HiGHS log.txt](https://github.com/scipy/scipy/files/9039809/HiGHS.log.txt) but the corresponding `OptimizeResult` object (see "Error message" section) is more or less empty and the only thing that is propagated correctly is the `message`, saying that the time limit was reached. ### Reproducing Code Example ```python import numpy as np from random import randint from scipy.optimize import milp, LinearConstraint, Bounds A = np.zeros((100, 100)) for c in range(100): for v in range(100): A[c, v] = randint(0, 5) b_lb = [-1*np.inf for c in range(100)] b_ub = [25 for c in range(100)] constraints = LinearConstraint(A, b_lb, b_ub) variable_lb = [0 for v in range(100)] variable_ub = [1 for v in range(100)] variable_bounds = Bounds(variable_lb, variable_ub) integrality = [1 for v in range(100)] c_vector = 100 * [-1] res = milp( c_vector, integrality=integrality, bounds=variable_bounds, constraints=constraints, options={"time_limit": 1, "disp": True} ) ``` ### Error message ```shell {'fun': None, 'message': 'Time limit reached. (HiGHS Status 13: model_status is Time limit ' "reached; primal_status is b'At upper bound')", 'mip_dual_bound': None, 'mip_gap': None, 'mip_node_count': None, 'status': 1, 'success': False, 'x': None} ``` ### SciPy/NumPy/Python version information 1.9.0rc1 1.23.0 sys.version_info(major=3, minor=9, micro=13, releaselevel='final', serial=0)
1.0
BUG: 1.9.0rc1: `OptimizeResult` not populated when `optimize.milp` runs into `time_limit` even though a feasible solution was found. - ### Describe your issue. When `optimize.milp` terminates due to the time limit being reached, the `OptimizeResult` object is not populated, even though one or more feasible solutions have been found. Example HiGHS log below shows that a feasible solution has been found (see solution status) when the 1 second time limit is reached. [HiGHS log.txt](https://github.com/scipy/scipy/files/9039809/HiGHS.log.txt) but the corresponding `OptimizeResult` object (see "Error message" section) is more or less empty and the only thing that is propagated correctly is the `message`, saying that the time limit was reached. ### Reproducing Code Example ```python import numpy as np from random import randint from scipy.optimize import milp, LinearConstraint, Bounds A = np.zeros((100, 100)) for c in range(100): for v in range(100): A[c, v] = randint(0, 5) b_lb = [-1*np.inf for c in range(100)] b_ub = [25 for c in range(100)] constraints = LinearConstraint(A, b_lb, b_ub) variable_lb = [0 for v in range(100)] variable_ub = [1 for v in range(100)] variable_bounds = Bounds(variable_lb, variable_ub) integrality = [1 for v in range(100)] c_vector = 100 * [-1] res = milp( c_vector, integrality=integrality, bounds=variable_bounds, constraints=constraints, options={"time_limit": 1, "disp": True} ) ``` ### Error message ```shell {'fun': None, 'message': 'Time limit reached. (HiGHS Status 13: model_status is Time limit ' "reached; primal_status is b'At upper bound')", 'mip_dual_bound': None, 'mip_gap': None, 'mip_node_count': None, 'status': 1, 'success': False, 'x': None} ``` ### SciPy/NumPy/Python version information 1.9.0rc1 1.23.0 sys.version_info(major=3, minor=9, micro=13, releaselevel='final', serial=0)
defect
bug optimizeresult not populated when optimize milp runs into time limit even though a feasible solution was found describe your issue when optimize milp terminates due to the time limit being reached the optimizeresult object is not populated even though one or more feasible solutions have been found example highs log below shows that a feasible solution has been found see solution status when the second time limit is reached but the corresponding optimizeresult object see error message section is more or less empty and the only thing that is propagated correctly is the message saying that the time limit was reached reproducing code example python import numpy as np from random import randint from scipy optimize import milp linearconstraint bounds a np zeros for c in range for v in range a randint b lb b ub constraints linearconstraint a b lb b ub variable lb variable ub variable bounds bounds variable lb variable ub integrality c vector res milp c vector integrality integrality bounds variable bounds constraints constraints options time limit disp true error message shell fun none message time limit reached highs status model status is time limit reached primal status is b at upper bound mip dual bound none mip gap none mip node count none status success false x none scipy numpy python version information sys version info major minor micro releaselevel final serial
1
8,951
2,612,934,070
IssuesEvent
2015-02-27 17:35:53
skelterjohn/go-wiki
https://api.github.com/repos/skelterjohn/go-wiki
closed
Broken url to video Practical_Go_Programming
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. On page http://code.google.com/p/go-wiki/wiki/GoTalks#Practical_Go_Programming url to video and slides is broken 2. URL to video at this moment is http://www.youtube.com/watch?v=2-pPAvqylu but should be: http://www.youtube.com/watch?v=2-pPAvqyluI ``` Original issue reported on code.google.com by `Dey...@gmail.com` on 15 Dec 2012 at 7:53
1.0
Broken url to video Practical_Go_Programming - ``` What steps will reproduce the problem? 1. On page http://code.google.com/p/go-wiki/wiki/GoTalks#Practical_Go_Programming url to video and slides is broken 2. URL to video at this moment is http://www.youtube.com/watch?v=2-pPAvqylu but should be: http://www.youtube.com/watch?v=2-pPAvqyluI ``` Original issue reported on code.google.com by `Dey...@gmail.com` on 15 Dec 2012 at 7:53
defect
broken url to video practical go programming what steps will reproduce the problem on page url to video and slides is broken url to video at this moment is but should be original issue reported on code google com by dey gmail com on dec at
1
57,580
15,866,618,468
IssuesEvent
2021-04-08 15:56:38
NREL/EnergyPlus
https://api.github.com/repos/NREL/EnergyPlus
closed
Airflow curve input defect in AirloopHVAC:UnitarySystem when MS DX Heating coils operate at Speed = 1
Defect
Issue overview -------------- PR #8501 fixed the issue #8335 to pass correct mass flow rate to calculate flow fraction for DX cooling coil performance curves at Speed = 1. However, the same problem occurs for a MS DX heating coil at speed=1, when a UnitarySystem is used as a parent. The input values vary with node mass flow rate as show below. ![Pinture3](https://user-images.githubusercontent.com/5970656/109869610-1ad2cc00-7c37-11eb-9053-b030d3591364.png) It is expected to get a constant input value. ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [X] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
1.0
Airflow curve input defect in AirloopHVAC:UnitarySystem when MS DX Heating coils operate at Speed = 1 - Issue overview -------------- PR #8501 fixed the issue #8335 to pass correct mass flow rate to calculate flow fraction for DX cooling coil performance curves at Speed = 1. However, the same problem occurs for a MS DX heating coil at speed=1, when a UnitarySystem is used as a parent. The input values vary with node mass flow rate as show below. ![Pinture3](https://user-images.githubusercontent.com/5970656/109869610-1ad2cc00-7c37-11eb-9053-b030d3591364.png) It is expected to get a constant input value. ### Details Some additional details for this issue (if relevant): - Platform (Operating system, version) - Version of EnergyPlus (if using an intermediate build, include SHA) - Unmethours link or helpdesk ticket number ### Checklist Add to this list or remove from it as applicable. This is a simple templated set of guidelines. - [X] Defect file added (list location of defect file here) - [ ] Ticket added to Pivotal for defect (development team task) - [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
defect
airflow curve input defect in airloophvac unitarysystem when ms dx heating coils operate at speed issue overview pr fixed the issue to pass correct mass flow rate to calculate flow fraction for dx cooling coil performance curves at speed however the same problem occurs for a ms dx heating coil at speed when a unitarysystem is used as a parent the input values vary with node mass flow rate as show below it is expected to get a constant input value details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
1
19,710
3,248,226,238
IssuesEvent
2015-10-17 04:13:45
jimradford/superputty
https://api.github.com/repos/jimradford/superputty
closed
Automatic logon with -pw password
auto-migrated Priority-Medium Type-Defect
``` See the attached screenshot for more details! sp version: 1.3.0.11 / Windows 7 Pro. SP1 ``` Original issue reported on code.google.com by `strugare...@gmail.com` on 23 Nov 2012 at 2:19 Attachments: * [sperror.png](https://storage.googleapis.com/google-code-attachments/superputty/issue-258/comment-0/sperror.png)
1.0
Automatic logon with -pw password - ``` See the attached screenshot for more details! sp version: 1.3.0.11 / Windows 7 Pro. SP1 ``` Original issue reported on code.google.com by `strugare...@gmail.com` on 23 Nov 2012 at 2:19 Attachments: * [sperror.png](https://storage.googleapis.com/google-code-attachments/superputty/issue-258/comment-0/sperror.png)
defect
automatic logon with pw password see the attached screenshot for more details sp version windows pro original issue reported on code google com by strugare gmail com on nov at attachments
1
39,620
9,565,486,443
IssuesEvent
2019-05-05 12:55:58
jmc2obj/j-mc-2-obj
https://api.github.com/repos/jmc2obj/j-mc-2-obj
closed
r302 Launching for OS X
Priority-Medium Type-Defect auto-migrated
``` What steps will reproduce the problem? 1. Have Apple OS X Mountain Lion 10.8.5 with supplement update 2. Open the .jar file by double clicking or right click - Jar Launcher 3. Open the .jar file by following steps from Issue 78 in terminal. What is the expected output? What do you see instead? A conf file to be generated and the application GUI to display, but instead an error report stating the jar file could not be launched. What version of the product are you using? On what operating system? jMc2Obj-dev_r302.jar on OS X 10.8.5 w/supplemental update Please describe your error as detailed as possible. Error Pop-up (Double Click): The Java JAR file "jMc2Obj-dev_r302.jar" could not be launched. Check the Console for possible error messages. Error in Terminal: new-host-6:Desktop Michael$ java -jar jMc2Obj-dev_r302.jar Exception in thread "main" java.lang.UnsupportedClassVersionError: org/jmc/Main : Unsupported major.minor version 51.0 at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClassCond(ClassLoader.java:637) at java.lang.ClassLoader.defineClass(ClassLoader.java:621) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141) at java.net.URLClassLoader.defineClass(URLClassLoader.java:283) at java.net.URLClassLoader.access$000(URLClassLoader.java:58) at java.net.URLClassLoader$1.run(URLClassLoader.java:197) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:190) at java.lang.ClassLoader.loadClass(ClassLoader.java:306) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) at java.lang.ClassLoader.loadClass(ClassLoader.java:247) Please provide any additional information below. Checked Console and found no errors, checked terminal and received the above list. Also, r301.jar was working 100% beforehand (no resizing for GUI to update, etc), with the exception of missing some blocks in MC 1.7.x. Initial test was done through the built-in GUI updater, and then direct download from the site. I've also read on you GUI Redesign Reddit thread that there have been problems with the GUI engine on OS X. I'm not sure if this is one of those issues. I hope this helps! I also apologize for all the issues I've given you guys. ``` Original issue reported on code.google.com by `rochefor...@gmail.com` on 7 Jan 2014 at 11:43 Attachments: - [Screen Shot 2014-01-07 at 6.41.15 PM.png](https://storage.googleapis.com/google-code-attachments/j-mc-2-obj/issue-81/comment-0/Screen Shot 2014-01-07 at 6.41.15 PM.png)
1.0
r302 Launching for OS X - ``` What steps will reproduce the problem? 1. Have Apple OS X Mountain Lion 10.8.5 with supplement update 2. Open the .jar file by double clicking or right click - Jar Launcher 3. Open the .jar file by following steps from Issue 78 in terminal. What is the expected output? What do you see instead? A conf file to be generated and the application GUI to display, but instead an error report stating the jar file could not be launched. What version of the product are you using? On what operating system? jMc2Obj-dev_r302.jar on OS X 10.8.5 w/supplemental update Please describe your error as detailed as possible. Error Pop-up (Double Click): The Java JAR file "jMc2Obj-dev_r302.jar" could not be launched. Check the Console for possible error messages. Error in Terminal: new-host-6:Desktop Michael$ java -jar jMc2Obj-dev_r302.jar Exception in thread "main" java.lang.UnsupportedClassVersionError: org/jmc/Main : Unsupported major.minor version 51.0 at java.lang.ClassLoader.defineClass1(Native Method) at java.lang.ClassLoader.defineClassCond(ClassLoader.java:637) at java.lang.ClassLoader.defineClass(ClassLoader.java:621) at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141) at java.net.URLClassLoader.defineClass(URLClassLoader.java:283) at java.net.URLClassLoader.access$000(URLClassLoader.java:58) at java.net.URLClassLoader$1.run(URLClassLoader.java:197) at java.security.AccessController.doPrivileged(Native Method) at java.net.URLClassLoader.findClass(URLClassLoader.java:190) at java.lang.ClassLoader.loadClass(ClassLoader.java:306) at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301) at java.lang.ClassLoader.loadClass(ClassLoader.java:247) Please provide any additional information below. Checked Console and found no errors, checked terminal and received the above list. Also, r301.jar was working 100% beforehand (no resizing for GUI to update, etc), with the exception of missing some blocks in MC 1.7.x. Initial test was done through the built-in GUI updater, and then direct download from the site. I've also read on you GUI Redesign Reddit thread that there have been problems with the GUI engine on OS X. I'm not sure if this is one of those issues. I hope this helps! I also apologize for all the issues I've given you guys. ``` Original issue reported on code.google.com by `rochefor...@gmail.com` on 7 Jan 2014 at 11:43 Attachments: - [Screen Shot 2014-01-07 at 6.41.15 PM.png](https://storage.googleapis.com/google-code-attachments/j-mc-2-obj/issue-81/comment-0/Screen Shot 2014-01-07 at 6.41.15 PM.png)
defect
launching for os x what steps will reproduce the problem have apple os x mountain lion with supplement update open the jar file by double clicking or right click jar launcher open the jar file by following steps from issue in terminal what is the expected output what do you see instead a conf file to be generated and the application gui to display but instead an error report stating the jar file could not be launched what version of the product are you using on what operating system dev jar on os x w supplemental update please describe your error as detailed as possible error pop up double click the java jar file dev jar could not be launched check the console for possible error messages error in terminal new host desktop michael java jar dev jar exception in thread main java lang unsupportedclassversionerror org jmc main unsupported major minor version at java lang classloader native method at java lang classloader defineclasscond classloader java at java lang classloader defineclass classloader java at java security secureclassloader defineclass secureclassloader java at java net urlclassloader defineclass urlclassloader java at java net urlclassloader access urlclassloader java at java net urlclassloader run urlclassloader java at java security accesscontroller doprivileged native method at java net urlclassloader findclass urlclassloader java at java lang classloader loadclass classloader java at sun misc launcher appclassloader loadclass launcher java at java lang classloader loadclass classloader java please provide any additional information below checked console and found no errors checked terminal and received the above list also jar was working beforehand no resizing for gui to update etc with the exception of missing some blocks in mc x initial test was done through the built in gui updater and then direct download from the site i ve also read on you gui redesign reddit thread that there have been problems with the gui engine on os x i m not sure if this is one of those issues i hope this helps i also apologize for all the issues i ve given you guys original issue reported on code google com by rochefor gmail com on jan at attachments shot at pm png
1
41,286
10,354,590,144
IssuesEvent
2019-09-05 14:02:16
vector-im/riot-web
https://api.github.com/repos/vector-im/riot-web
closed
Terms account data is meant to be additive, but currently sets only new URLs
bug defect phase:1 privacy privacy-sprint
We're meant to keep adding newly agreed URLs to account data, but a bug means we're only storing the ones you just agreed to this time.
1.0
Terms account data is meant to be additive, but currently sets only new URLs - We're meant to keep adding newly agreed URLs to account data, but a bug means we're only storing the ones you just agreed to this time.
defect
terms account data is meant to be additive but currently sets only new urls we re meant to keep adding newly agreed urls to account data but a bug means we re only storing the ones you just agreed to this time
1
23,793
3,851,867,809
IssuesEvent
2016-04-06 05:28:58
GPF/imame4all
https://api.github.com/repos/GPF/imame4all
closed
Button issues with Samurai Showdown 4 or maybe all NEO GEO games?
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. Load Samurai Showdown 4 2. Enable 4-buttons or All buttons 3. Test all buttons. What is the expected output? What do you see instead? Samurai Showdown uses 4 buttons. A,B,C,D for LP, MP, HP, K. In MAME4Droid, button L,R is not used. A is HP, B is LP, X is MP. Y opens up NEO GEO GENIE. The button for Kick is missing. Also, the Virtual button for Select is the coin button. But the onscreen coin button is not working. What version of the product are you using? On what operating system? 1.5 Android. Please provide any additional information below. I wonder about NEO GEO GENIE whether you can disable it. Maybe it is conflicting with the last button for Neo Geo. esp since the portrait mode of MAME4Droid only displays 4 buttons on my Galaxy S2. I have only managed to test Samurai Showdown 4 since that was the only game I found and the only one I really wanted to play. ``` Original issue reported on code.google.com by `alloutfo...@gmail.com` on 20 Dec 2011 at 8:16
1.0
Button issues with Samurai Showdown 4 or maybe all NEO GEO games? - ``` What steps will reproduce the problem? 1. Load Samurai Showdown 4 2. Enable 4-buttons or All buttons 3. Test all buttons. What is the expected output? What do you see instead? Samurai Showdown uses 4 buttons. A,B,C,D for LP, MP, HP, K. In MAME4Droid, button L,R is not used. A is HP, B is LP, X is MP. Y opens up NEO GEO GENIE. The button for Kick is missing. Also, the Virtual button for Select is the coin button. But the onscreen coin button is not working. What version of the product are you using? On what operating system? 1.5 Android. Please provide any additional information below. I wonder about NEO GEO GENIE whether you can disable it. Maybe it is conflicting with the last button for Neo Geo. esp since the portrait mode of MAME4Droid only displays 4 buttons on my Galaxy S2. I have only managed to test Samurai Showdown 4 since that was the only game I found and the only one I really wanted to play. ``` Original issue reported on code.google.com by `alloutfo...@gmail.com` on 20 Dec 2011 at 8:16
defect
button issues with samurai showdown or maybe all neo geo games what steps will reproduce the problem load samurai showdown enable buttons or all buttons test all buttons what is the expected output what do you see instead samurai showdown uses buttons a b c d for lp mp hp k in button l r is not used a is hp b is lp x is mp y opens up neo geo genie the button for kick is missing also the virtual button for select is the coin button but the onscreen coin button is not working what version of the product are you using on what operating system android please provide any additional information below i wonder about neo geo genie whether you can disable it maybe it is conflicting with the last button for neo geo esp since the portrait mode of only displays buttons on my galaxy i have only managed to test samurai showdown since that was the only game i found and the only one i really wanted to play original issue reported on code google com by alloutfo gmail com on dec at
1
136,820
11,088,758,520
IssuesEvent
2019-12-14 13:40:08
BiancoRoyal/node-red-contrib-modbus
https://api.github.com/repos/BiancoRoyal/node-red-contrib-modbus
closed
Read Node stopped working
fixed pull request welcome testing unexpected behaviour
<!-- IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MIGHT CLOSE YOUR ISSUE WITHOUT INVESTIGATING PLEASE, DELETE NOT NEEDED PARTS FROM THAT ISSUE! --> **I'm submitting a ...** (check one with "x") ## Problem - ( ) bug report => please search github for a similar issue or PR and test Modbus problems with the [modbus-serial](https://github.com/yaacov/node-modbus-serial) package before submitting - (X) unexpected behavior => please visit the Wiki before **Current behavior** <!-- Describe how the bug or problem manifests. --> Modbus Read node stopped working **Expected behavior** <!-- Describe what the behavior would be without the bug or problem. --> Gets readings from modes devices ### Your Environment **Please tell us about your environment:** <!-- system, Modbus server or client type, package manager, equipment, vendor info --> RPi 3+ with Hassos 2.12. Hassio 0.102.3. Node-red add-on 5.0.5 SDM120 and SDM230 energy meters connected to Pi via USB. ``` Add-on version: 5.0.5 You are running the latest version of this add-on. System: HassOS 2.12 (armv7 / raspberrypi3) Home Assistant version: 0.102.3 Supervisor version: 192 ----------------------------------------------------------- Please, share the above information when looking for help or support in, e.g., GitHub, forums or the Discord chat. ----------------------------------------------------------- [cont-init.d] 00-banner.sh: exited 0. [cont-init.d] 01-log-level.sh: executing... [cont-init.d] 01-log-level.sh: exited 0. [cont-init.d] nginx.sh: executing... [cont-init.d] nginx.sh: exited 0. [cont-init.d] node-red.sh: executing... patching file nodes/ui_base.html Hunk #1 succeeded at 1243 (offset 712 lines). audited 200 packages in 6.917s found 0 vulnerabilities [cont-init.d] node-red.sh: exited 0. [cont-init.d] user.sh: executing... [cont-init.d] user.sh: exited 0. [cont-init.d] done. [services.d] starting services [services.d] done. [10:09:46] INFO: Starting Node-RED... > hassio-addon-node-red@ start /opt > node $NODE_OPTIONS node_modules/node-red/red.js "--settings" "/etc/node-red/config.js" 7 Dec 10:09:51 - [info] Welcome to Node-RED =================== 7 Dec 10:09:51 - [info] Node-RED version: v1.0.3 7 Dec 10:09:51 - [info] Node.js version: v10.16.3 7 Dec 10:09:51 - [info] Linux 4.14.98-v7 arm LE 7 Dec 10:09:53 - [info] Loading palette nodes 7 Dec 10:10:04 - [info] Dashboard version 2.14.0 started at /endpoint/ui 7 Dec 10:10:08 - [info] Settings file : /etc/node-red/config.js 7 Dec 10:10:08 - [info] Context store : 'default' [module=memory] 7 Dec 10:10:08 - [info] User directory : /config/node-red/ 7 Dec 10:10:08 - [warn] Projects disabled : editorTheme.projects.enabled=false 7 Dec 10:10:08 - [info] Flows file : /config/node-red/flows.json 7 Dec 10:10:08 - [info] Starting flows 7 Dec 10:10:09 - [info] Started flows 7 Dec 10:10:09 - [info] Server now running at http://127.0.0.1:46836/ [10:10:09] INFO: Starting NGinx... nginx: [alert] detected a LuaJIT version which is not OpenResty's; many optimizations will be disabled and performance will be compromised (see https://github.com/openresty/luajit2 for OpenResty's LuaJIT or, even better, consider using the OpenResty releases from https://openresty.org/en/download.html) 7 Dec 10:10:11 - [info] [mqtt-broker:c28a2dad.22663] Connected to broker: mqtt://192.168.1.200:1883 7 Dec 10:10:13 - [info] [server:Home Assistant] WebSocket Connected to http://hassio/homeassistant 7 Dec 10:10:21 - [error] [modbus-read:SDM 230 Voltage] Timed out 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Voltage] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Current] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Power] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Imp] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Exp] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Voltage] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Current] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Power] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Energy Sol] FSM Reset On State [object Object] 7 Dec 10:10:21 - [error] [modbus-read:SDM 120 Voltage] Timed out 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Voltage] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Current] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Power] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Imp] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Exp] FSM Reset On State [object Object] ``` * ( ) I am using just the Node-RED package and got [modbus-serial](https://github.com/yaacov/node-modbus-serial) package installed with it * ( ) I have installed [modbus-serial](https://github.com/yaacov/node-modbus-serial) package global from source (using git clone) * ( ) I have installed [modbus-serial](https://github.com/yaacov/node-modbus-serial) as a global package (using npm i -g) Yes, modbus-serial is the name but it does all, TCP and Serial. The node-modbus package is just for the Modbus Server node. * Device : - ( ) macOS: v - ( ) Linux: v - ( ) Windows: v - ( ) Raspbian: v - (X) Other: HASSOS on RPi3 - Name: - Version: * Description of the connecting Modbus system: - Name : - Version: - Vendor : - Link : https:// * **node-red-contrib-modbus version:** `` <!-- Check which is the hash of the last commit from node-red-contrib-modbus that you have locally --> * **Node:** (min. 6.x) `node --version` = 5.1.0
1.0
Read Node stopped working - <!-- IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MIGHT CLOSE YOUR ISSUE WITHOUT INVESTIGATING PLEASE, DELETE NOT NEEDED PARTS FROM THAT ISSUE! --> **I'm submitting a ...** (check one with "x") ## Problem - ( ) bug report => please search github for a similar issue or PR and test Modbus problems with the [modbus-serial](https://github.com/yaacov/node-modbus-serial) package before submitting - (X) unexpected behavior => please visit the Wiki before **Current behavior** <!-- Describe how the bug or problem manifests. --> Modbus Read node stopped working **Expected behavior** <!-- Describe what the behavior would be without the bug or problem. --> Gets readings from modes devices ### Your Environment **Please tell us about your environment:** <!-- system, Modbus server or client type, package manager, equipment, vendor info --> RPi 3+ with Hassos 2.12. Hassio 0.102.3. Node-red add-on 5.0.5 SDM120 and SDM230 energy meters connected to Pi via USB. ``` Add-on version: 5.0.5 You are running the latest version of this add-on. System: HassOS 2.12 (armv7 / raspberrypi3) Home Assistant version: 0.102.3 Supervisor version: 192 ----------------------------------------------------------- Please, share the above information when looking for help or support in, e.g., GitHub, forums or the Discord chat. ----------------------------------------------------------- [cont-init.d] 00-banner.sh: exited 0. [cont-init.d] 01-log-level.sh: executing... [cont-init.d] 01-log-level.sh: exited 0. [cont-init.d] nginx.sh: executing... [cont-init.d] nginx.sh: exited 0. [cont-init.d] node-red.sh: executing... patching file nodes/ui_base.html Hunk #1 succeeded at 1243 (offset 712 lines). audited 200 packages in 6.917s found 0 vulnerabilities [cont-init.d] node-red.sh: exited 0. [cont-init.d] user.sh: executing... [cont-init.d] user.sh: exited 0. [cont-init.d] done. [services.d] starting services [services.d] done. [10:09:46] INFO: Starting Node-RED... > hassio-addon-node-red@ start /opt > node $NODE_OPTIONS node_modules/node-red/red.js "--settings" "/etc/node-red/config.js" 7 Dec 10:09:51 - [info] Welcome to Node-RED =================== 7 Dec 10:09:51 - [info] Node-RED version: v1.0.3 7 Dec 10:09:51 - [info] Node.js version: v10.16.3 7 Dec 10:09:51 - [info] Linux 4.14.98-v7 arm LE 7 Dec 10:09:53 - [info] Loading palette nodes 7 Dec 10:10:04 - [info] Dashboard version 2.14.0 started at /endpoint/ui 7 Dec 10:10:08 - [info] Settings file : /etc/node-red/config.js 7 Dec 10:10:08 - [info] Context store : 'default' [module=memory] 7 Dec 10:10:08 - [info] User directory : /config/node-red/ 7 Dec 10:10:08 - [warn] Projects disabled : editorTheme.projects.enabled=false 7 Dec 10:10:08 - [info] Flows file : /config/node-red/flows.json 7 Dec 10:10:08 - [info] Starting flows 7 Dec 10:10:09 - [info] Started flows 7 Dec 10:10:09 - [info] Server now running at http://127.0.0.1:46836/ [10:10:09] INFO: Starting NGinx... nginx: [alert] detected a LuaJIT version which is not OpenResty's; many optimizations will be disabled and performance will be compromised (see https://github.com/openresty/luajit2 for OpenResty's LuaJIT or, even better, consider using the OpenResty releases from https://openresty.org/en/download.html) 7 Dec 10:10:11 - [info] [mqtt-broker:c28a2dad.22663] Connected to broker: mqtt://192.168.1.200:1883 7 Dec 10:10:13 - [info] [server:Home Assistant] WebSocket Connected to http://hassio/homeassistant 7 Dec 10:10:21 - [error] [modbus-read:SDM 230 Voltage] Timed out 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Voltage] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Current] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Power] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Imp] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Exp] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Voltage] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Current] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Power] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Energy Sol] FSM Reset On State [object Object] 7 Dec 10:10:21 - [error] [modbus-read:SDM 120 Voltage] Timed out 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Voltage] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Current] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Power] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Imp] FSM Reset On State [object Object] 7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Exp] FSM Reset On State [object Object] ``` * ( ) I am using just the Node-RED package and got [modbus-serial](https://github.com/yaacov/node-modbus-serial) package installed with it * ( ) I have installed [modbus-serial](https://github.com/yaacov/node-modbus-serial) package global from source (using git clone) * ( ) I have installed [modbus-serial](https://github.com/yaacov/node-modbus-serial) as a global package (using npm i -g) Yes, modbus-serial is the name but it does all, TCP and Serial. The node-modbus package is just for the Modbus Server node. * Device : - ( ) macOS: v - ( ) Linux: v - ( ) Windows: v - ( ) Raspbian: v - (X) Other: HASSOS on RPi3 - Name: - Version: * Description of the connecting Modbus system: - Name : - Version: - Vendor : - Link : https:// * **node-red-contrib-modbus version:** `` <!-- Check which is the hash of the last commit from node-red-contrib-modbus that you have locally --> * **Node:** (min. 6.x) `node --version` = 5.1.0
non_defect
read node stopped working if you don t fill out the following information we might close your issue without investigating please delete not needed parts from that issue i m submitting a check one with x problem bug report please search github for a similar issue or pr and test modbus problems with the package before submitting x unexpected behavior please visit the wiki before current behavior modbus read node stopped working expected behavior gets readings from modes devices your environment please tell us about your environment rpi with hassos hassio node red add on and energy meters connected to pi via usb add on version you are running the latest version of this add on system hassos home assistant version supervisor version please share the above information when looking for help or support in e g github forums or the discord chat banner sh exited log level sh executing log level sh exited nginx sh executing nginx sh exited node red sh executing patching file nodes ui base html hunk succeeded at offset lines audited packages in found vulnerabilities node red sh exited user sh executing user sh exited done starting services done info starting node red hassio addon node red start opt node node options node modules node red red js settings etc node red config js dec welcome to node red dec node red version dec node js version dec linux arm le dec loading palette nodes dec dashboard version started at endpoint ui dec settings file etc node red config js dec context store default dec user directory config node red dec projects disabled editortheme projects enabled false dec flows file config node red flows json dec starting flows dec started flows dec server now running at info starting nginx nginx detected a luajit version which is not openresty s many optimizations will be disabled and performance will be compromised see for openresty s luajit or even better consider using the openresty releases from dec connected to broker mqtt dec websocket connected to dec timed out dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec timed out dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state i am using just the node red package and got package installed with it i have installed package global from source using git clone i have installed as a global package using npm i g yes modbus serial is the name but it does all tcp and serial the node modbus package is just for the modbus server node device macos v linux v windows v raspbian v x other hassos on name version description of the connecting modbus system name version vendor link https node red contrib modbus version node min x node version
0
110,589
13,917,947,586
IssuesEvent
2020-10-21 06:55:24
vaadin/vaadin-avatar
https://api.github.com/repos/vaadin/vaadin-avatar
closed
The plus sign in "+5" when avatars are overflowing is not completely with a number larger than 9
needs design
Consider the following code ``` <vaadin-avatar-group id="avatarGroup" items={{items}} max-items-visible="3"> </vaadin-avatar-group> ``` Where `items` contains more than 13 items. At this point the overflow number is partly obscured. ![image](https://user-images.githubusercontent.com/29859656/96098319-a8a55900-0eda-11eb-87a1-23b3d0ae8b8b.png) I think this boils down to whether 10+ avatars in one group is realistic. Maybe the overflow circle could be shown on top (on the z axis)?
1.0
The plus sign in "+5" when avatars are overflowing is not completely with a number larger than 9 - Consider the following code ``` <vaadin-avatar-group id="avatarGroup" items={{items}} max-items-visible="3"> </vaadin-avatar-group> ``` Where `items` contains more than 13 items. At this point the overflow number is partly obscured. ![image](https://user-images.githubusercontent.com/29859656/96098319-a8a55900-0eda-11eb-87a1-23b3d0ae8b8b.png) I think this boils down to whether 10+ avatars in one group is realistic. Maybe the overflow circle could be shown on top (on the z axis)?
non_defect
the plus sign in when avatars are overflowing is not completely with a number larger than consider the following code where items contains more than items at this point the overflow number is partly obscured i think this boils down to whether avatars in one group is realistic maybe the overflow circle could be shown on top on the z axis
0
219,537
17,099,256,254
IssuesEvent
2021-07-09 08:54:09
tracim/tracim
https://api.github.com/repos/tracim/tracim
closed
Feat: Create TagList component
frontend manually tested not in changelog
## Feature description and goals We add tags to contents (#4681), which requires a new component to render the tag list. This component will be rendered in the apps file and html-document, in a dedicated tab above the other ones. The goal of this feature is to let the user choose tags in the available tag list, or to create new ones in a form. The selected tags will be displayed in the left part of the app, on top of the "download" icon. This component should look like MemberList The chosen tags will be marked as "checked", and the other ones will be "unchecked". The checked tags will appear on top of the unchecked ones. When a new tag is created in the app, it will be marked as "checked" and it will be rendered in the list. This tag will be available in every content in its space, and will be displayed as an "unchecked" tag. To do : - [x] Create TagList component - [x] Create NewTagForm component - [x] Create Tag component - [x] Render selected tags - [x] Add translations - [x] Link with backend endpoints <!-- Explain why we want this feature and describe it. --> ## Prerequisites Backend endpoint to GET the tag list: #4685 ## Mockups ![TAGS](https://user-images.githubusercontent.com/50776142/120205581-65f67e00-c22a-11eb-9e4b-4d3f21fe5ba5.png) <!-- Optional - list the issues that must be solved or what needs to be done before handling this issue. --> <!-- ## Required sections, if relevant ## --> <!-- - To be discussed before development --> <!-- - Interface --> <!-- - Translations --> <!-- - Workaround --> <!-- - Extra information -->
1.0
Feat: Create TagList component - ## Feature description and goals We add tags to contents (#4681), which requires a new component to render the tag list. This component will be rendered in the apps file and html-document, in a dedicated tab above the other ones. The goal of this feature is to let the user choose tags in the available tag list, or to create new ones in a form. The selected tags will be displayed in the left part of the app, on top of the "download" icon. This component should look like MemberList The chosen tags will be marked as "checked", and the other ones will be "unchecked". The checked tags will appear on top of the unchecked ones. When a new tag is created in the app, it will be marked as "checked" and it will be rendered in the list. This tag will be available in every content in its space, and will be displayed as an "unchecked" tag. To do : - [x] Create TagList component - [x] Create NewTagForm component - [x] Create Tag component - [x] Render selected tags - [x] Add translations - [x] Link with backend endpoints <!-- Explain why we want this feature and describe it. --> ## Prerequisites Backend endpoint to GET the tag list: #4685 ## Mockups ![TAGS](https://user-images.githubusercontent.com/50776142/120205581-65f67e00-c22a-11eb-9e4b-4d3f21fe5ba5.png) <!-- Optional - list the issues that must be solved or what needs to be done before handling this issue. --> <!-- ## Required sections, if relevant ## --> <!-- - To be discussed before development --> <!-- - Interface --> <!-- - Translations --> <!-- - Workaround --> <!-- - Extra information -->
non_defect
feat create taglist component feature description and goals we add tags to contents which requires a new component to render the tag list this component will be rendered in the apps file and html document in a dedicated tab above the other ones the goal of this feature is to let the user choose tags in the available tag list or to create new ones in a form the selected tags will be displayed in the left part of the app on top of the download icon this component should look like memberlist the chosen tags will be marked as checked and the other ones will be unchecked the checked tags will appear on top of the unchecked ones when a new tag is created in the app it will be marked as checked and it will be rendered in the list this tag will be available in every content in its space and will be displayed as an unchecked tag to do create taglist component create newtagform component create tag component render selected tags add translations link with backend endpoints prerequisites backend endpoint to get the tag list mockups
0
44,525
12,227,249,219
IssuesEvent
2020-05-03 14:31:36
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
SelectBooleanCheckbox: Disabled selectBooleanCheckbox with itemLabel defined
defect
## 1) Environment - PrimeFaces version: 7.0 - Does it work on the newest released PrimeFaces version? Version?: No - Does it work on the newest sources in GitHub? (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source) - Application server + version: tomee 7.1.1 - Affected browsers: Firefox, Chrome, ... ## 2) Expected behavior If SelectBooleanCheckbox is disabled and an ItemLabel is defined, the label should also be disabled. (as in the 2. field of screenshot) ## 3) Actual behavior The label looks like activated and has also the "click-finger" cursor (as the 3. field in the screenshot) ![SelectBooleanCheckbox](https://user-images.githubusercontent.com/37800416/80712398-3d192c00-8af2-11ea-92ab-8eeb55785d5f.png) ## 4) Steps to reproduce .. ## 5) Sample XHTML ```xml <p:panelGrid columns="2"> <p:selectBooleanCheckbox id="cb1" value="true" disabled="false" itemLabel="1. disabled='false'"> </p:selectBooleanCheckbox> <p:selectBooleanCheckbox id="cb2" value="true" styleClass="ui-state-disabled" disabled="false" itemLabel="2. disabled='false' styleClass='ui-state-disabled'"> </p:selectBooleanCheckbox> <p:selectBooleanCheckbox id="cb3" value="true" disabled="true" itemLabel="3. disabled='true'"> </p:selectBooleanCheckbox> <p:selectBooleanCheckbox id="cb4" value="true" styleClass="ui-state-disabled" disabled="true" itemLabel="4. disabled='true' styleClass='ui-state-disabled'"> </p:selectBooleanCheckbox> </p:panelGrid> ``` ## 6) Sample bean ..
1.0
SelectBooleanCheckbox: Disabled selectBooleanCheckbox with itemLabel defined - ## 1) Environment - PrimeFaces version: 7.0 - Does it work on the newest released PrimeFaces version? Version?: No - Does it work on the newest sources in GitHub? (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source) - Application server + version: tomee 7.1.1 - Affected browsers: Firefox, Chrome, ... ## 2) Expected behavior If SelectBooleanCheckbox is disabled and an ItemLabel is defined, the label should also be disabled. (as in the 2. field of screenshot) ## 3) Actual behavior The label looks like activated and has also the "click-finger" cursor (as the 3. field in the screenshot) ![SelectBooleanCheckbox](https://user-images.githubusercontent.com/37800416/80712398-3d192c00-8af2-11ea-92ab-8eeb55785d5f.png) ## 4) Steps to reproduce .. ## 5) Sample XHTML ```xml <p:panelGrid columns="2"> <p:selectBooleanCheckbox id="cb1" value="true" disabled="false" itemLabel="1. disabled='false'"> </p:selectBooleanCheckbox> <p:selectBooleanCheckbox id="cb2" value="true" styleClass="ui-state-disabled" disabled="false" itemLabel="2. disabled='false' styleClass='ui-state-disabled'"> </p:selectBooleanCheckbox> <p:selectBooleanCheckbox id="cb3" value="true" disabled="true" itemLabel="3. disabled='true'"> </p:selectBooleanCheckbox> <p:selectBooleanCheckbox id="cb4" value="true" styleClass="ui-state-disabled" disabled="true" itemLabel="4. disabled='true' styleClass='ui-state-disabled'"> </p:selectBooleanCheckbox> </p:panelGrid> ``` ## 6) Sample bean ..
defect
selectbooleancheckbox disabled selectbooleancheckbox with itemlabel defined environment primefaces version does it work on the newest released primefaces version version no does it work on the newest sources in github build by source application server version tomee affected browsers firefox chrome expected behavior if selectbooleancheckbox is disabled and an itemlabel is defined the label should also be disabled as in the field of screenshot actual behavior the label looks like activated and has also the click finger cursor as the field in the screenshot steps to reproduce sample xhtml xml p selectbooleancheckbox id value true disabled false itemlabel disabled false p selectbooleancheckbox id value true styleclass ui state disabled disabled false itemlabel disabled false styleclass ui state disabled p selectbooleancheckbox id value true disabled true itemlabel disabled true p selectbooleancheckbox id value true styleclass ui state disabled disabled true itemlabel disabled true styleclass ui state disabled sample bean
1
45,435
12,799,854,314
IssuesEvent
2020-07-02 16:02:57
snowplow/snowplow-android-tracker
https://api.github.com/repos/snowplow/snowplow-android-tracker
closed
Fix importing of kotlin on gradle
priority:medium status:completed type:defect
This project is written 100% in Java, however the SDK ships with a dependency on [the Kotlin stdlib](https://github.com/snowplow/snowplow-android-tracker/blob/master/snowplow-tracker/build.gradle#L85) and [Kotlin Android extensions](https://github.com/snowplow/snowplow-android-tracker/blob/master/snowplow-tracker/build.gradle#L7). Kotlin was added in [this PR](https://github.com/snowplow/snowplow-android-tracker/pull/358), but seems unrelated? Also as an aside it would be great if this library added nullability annotations to make Kotlin interoperability nicer! I can open up a separate issue for this if you'd prefer.
1.0
Fix importing of kotlin on gradle - This project is written 100% in Java, however the SDK ships with a dependency on [the Kotlin stdlib](https://github.com/snowplow/snowplow-android-tracker/blob/master/snowplow-tracker/build.gradle#L85) and [Kotlin Android extensions](https://github.com/snowplow/snowplow-android-tracker/blob/master/snowplow-tracker/build.gradle#L7). Kotlin was added in [this PR](https://github.com/snowplow/snowplow-android-tracker/pull/358), but seems unrelated? Also as an aside it would be great if this library added nullability annotations to make Kotlin interoperability nicer! I can open up a separate issue for this if you'd prefer.
defect
fix importing of kotlin on gradle this project is written in java however the sdk ships with a dependency on and kotlin was added in but seems unrelated also as an aside it would be great if this library added nullability annotations to make kotlin interoperability nicer i can open up a separate issue for this if you d prefer
1
4,103
2,610,087,444
IssuesEvent
2015-02-26 18:26:33
chrsmith/dsdsdaadf
https://api.github.com/repos/chrsmith/dsdsdaadf
opened
深圳痤疮哪里祛比较好
auto-migrated Priority-Medium Type-Defect
``` 深圳痤疮哪里祛比较好【深圳韩方科颜全国热线400-869-1818,24 小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:20
1.0
深圳痤疮哪里祛比较好 - ``` 深圳痤疮哪里祛比较好【深圳韩方科颜全国热线400-869-1818,24 小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 ``` ----- Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:20
defect
深圳痤疮哪里祛比较好 深圳痤疮哪里祛比较好【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 original issue reported on code google com by szft com on may at
1
795,706
28,083,324,494
IssuesEvent
2023-03-30 08:07:25
frequenz-floss/frequenz-sdk-python
https://api.github.com/repos/frequenz-floss/frequenz-sdk-python
opened
Update SDK dependency of channels to the new release 0.14
part:❓ priority:❓ type:enhancement
### What's needed? Interfaces changed and need to be adjusted ### Proposed solution _No response_ ### Use cases _No response_ ### Alternatives and workarounds _No response_ ### Additional context _No response_
1.0
Update SDK dependency of channels to the new release 0.14 - ### What's needed? Interfaces changed and need to be adjusted ### Proposed solution _No response_ ### Use cases _No response_ ### Alternatives and workarounds _No response_ ### Additional context _No response_
non_defect
update sdk dependency of channels to the new release what s needed interfaces changed and need to be adjusted proposed solution no response use cases no response alternatives and workarounds no response additional context no response
0
7,091
2,597,979,510
IssuesEvent
2015-02-22 01:06:43
JukkaL/mypy
https://api.github.com/repos/JukkaL/mypy
closed
For loop parse error with single-item tuple lvalue
bug priority
This is valid Python but mypy doesn't accept it (`Parse error before in`): ```python for x, in [(1,), (2,), (3,)]: print(x) ```
1.0
For loop parse error with single-item tuple lvalue - This is valid Python but mypy doesn't accept it (`Parse error before in`): ```python for x, in [(1,), (2,), (3,)]: print(x) ```
non_defect
for loop parse error with single item tuple lvalue this is valid python but mypy doesn t accept it parse error before in python for x in print x
0
13,406
2,755,583,219
IssuesEvent
2015-04-26 19:40:11
jakejustus/pct
https://api.github.com/repos/jakejustus/pct
closed
PCTCompileExt doesn't set resultProperty
auto-migrated Type-Defect wontfix
``` When running PCTCompileExt with failOnError="false" resultProperty="result", the result property is not set at the end of the PCTCompileExt task. With PCTCompile it works fine. ``` Original issue reported on code.google.com by `tif...@gmail.com` on 3 Oct 2013 at 10:21
1.0
PCTCompileExt doesn't set resultProperty - ``` When running PCTCompileExt with failOnError="false" resultProperty="result", the result property is not set at the end of the PCTCompileExt task. With PCTCompile it works fine. ``` Original issue reported on code.google.com by `tif...@gmail.com` on 3 Oct 2013 at 10:21
defect
pctcompileext doesn t set resultproperty when running pctcompileext with failonerror false resultproperty result the result property is not set at the end of the pctcompileext task with pctcompile it works fine original issue reported on code google com by tif gmail com on oct at
1
47,077
13,056,028,043
IssuesEvent
2020-07-30 03:26:08
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
[PROPOSAL] fatal in decay (Trac #2397)
Incomplete Migration Migrated from Trac combo simulation defect
Migrated from https://code.icecube.wisc.edu/ticket/2397 ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "description": "I found this fatal error running PROPOSAL via the script [https://code.icecube.wisc.edu/projects/icecube/browser/IceCube/meta-projects/combo/trunk/simprod-scripts/resources/scripts/clsim.py]:\n\n\n{{{\nFATAL (PROPOSAL): Kaellen function is negative. Cannot caluclate momentum (Decay\nChannel.cxx:88 in static double PROPOSAL::DecayChannel::Momentum(double, double,\n double))\nERROR (I3Module): propagator_propagator: Exception thrown (I3Module.cxx:123 in v\noid I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/data/user/grenzi/combo-py3/src/simprod-scripts/resources/scripts/clsim.\npy\", line 188, in <module>\n main()\n File \"/data/user/grenzi/combo-py3/src/simprod-scripts/resources/scripts/clsim.\npy\", line 184, in main\n usegslrng=params['usegslrng'])\n File \"/data/user/grenzi/combo-py3/build/lib/icecube/simprod/util/simprodtray.py\", line 150, in RunI3Tray\n _execute(tray, executionmaxcount)\n File \"/data/user/grenzi/combo-py3/build/lib/icecube/simprod/util/simprodtray.py\", line 53, in _execute\n tray.Execute()\n File \"/data/user/grenzi/combo-py3/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\nRuntimeError: Kaellen function is negative. Cannot caluclate momentum (in static double PROPOSAL::DecayChannel::Momentum(double, double, double))\n}}}\n\nThe error seems to come from here at line 88 [https://code.icecube.wisc.edu/projects/icecube/browser/IceCube/meta-projects/combo/trunk/PROPOSAL/private/PROPOSAL/decay/DecayChannel.cxx]\n\n\n{{{\n79 double DecayChannel::Momentum(double m1, double m2, double m3)\n80\t{\n81\t double kaellen = (m1 - m2 - m3) * (m1 + m2 + m3) * (m1 - m2 + m3) * (m1 + m2 - m3);\n82\t\n83\t if (kaellen > 0.0)\n84\t {\n85\t return std::sqrt(kaellen) / (2.0 * m1);\n86\t } else\n87\t {\n88\t log_fatal(\"Kaellen function is negative. Cannot caluclate momentum\");\n89\t return 0.0;\n90\t }\n91\t}\n}}}\n\n\nThis error prevents the propagation process for all of the following frames. I think the right thing would be just skipping the single wrong one.\n", "reporter": "grenzi", "cc": "", "resolution": "fixed", "_ts": "1593001902142004", "component": "combo simulation", "summary": "[PROPOSAL] fatal in decay", "priority": "major", "keywords": "", "time": "2020-01-08T09:47:52", "milestone": "Autumnal Equinox 2020", "owner": "jsoedingrekso", "type": "defect" } ```
1.0
[PROPOSAL] fatal in decay (Trac #2397) - Migrated from https://code.icecube.wisc.edu/ticket/2397 ```json { "status": "closed", "changetime": "2020-06-24T12:31:42", "description": "I found this fatal error running PROPOSAL via the script [https://code.icecube.wisc.edu/projects/icecube/browser/IceCube/meta-projects/combo/trunk/simprod-scripts/resources/scripts/clsim.py]:\n\n\n{{{\nFATAL (PROPOSAL): Kaellen function is negative. Cannot caluclate momentum (Decay\nChannel.cxx:88 in static double PROPOSAL::DecayChannel::Momentum(double, double,\n double))\nERROR (I3Module): propagator_propagator: Exception thrown (I3Module.cxx:123 in v\noid I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/data/user/grenzi/combo-py3/src/simprod-scripts/resources/scripts/clsim.\npy\", line 188, in <module>\n main()\n File \"/data/user/grenzi/combo-py3/src/simprod-scripts/resources/scripts/clsim.\npy\", line 184, in main\n usegslrng=params['usegslrng'])\n File \"/data/user/grenzi/combo-py3/build/lib/icecube/simprod/util/simprodtray.py\", line 150, in RunI3Tray\n _execute(tray, executionmaxcount)\n File \"/data/user/grenzi/combo-py3/build/lib/icecube/simprod/util/simprodtray.py\", line 53, in _execute\n tray.Execute()\n File \"/data/user/grenzi/combo-py3/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\nRuntimeError: Kaellen function is negative. Cannot caluclate momentum (in static double PROPOSAL::DecayChannel::Momentum(double, double, double))\n}}}\n\nThe error seems to come from here at line 88 [https://code.icecube.wisc.edu/projects/icecube/browser/IceCube/meta-projects/combo/trunk/PROPOSAL/private/PROPOSAL/decay/DecayChannel.cxx]\n\n\n{{{\n79 double DecayChannel::Momentum(double m1, double m2, double m3)\n80\t{\n81\t double kaellen = (m1 - m2 - m3) * (m1 + m2 + m3) * (m1 - m2 + m3) * (m1 + m2 - m3);\n82\t\n83\t if (kaellen > 0.0)\n84\t {\n85\t return std::sqrt(kaellen) / (2.0 * m1);\n86\t } else\n87\t {\n88\t log_fatal(\"Kaellen function is negative. Cannot caluclate momentum\");\n89\t return 0.0;\n90\t }\n91\t}\n}}}\n\n\nThis error prevents the propagation process for all of the following frames. I think the right thing would be just skipping the single wrong one.\n", "reporter": "grenzi", "cc": "", "resolution": "fixed", "_ts": "1593001902142004", "component": "combo simulation", "summary": "[PROPOSAL] fatal in decay", "priority": "major", "keywords": "", "time": "2020-01-08T09:47:52", "milestone": "Autumnal Equinox 2020", "owner": "jsoedingrekso", "type": "defect" } ```
defect
fatal in decay trac migrated from json status closed changetime description i found this fatal error running proposal via the script n n n nfatal proposal kaellen function is negative cannot caluclate momentum decay nchannel cxx in static double proposal decaychannel momentum double double n double nerror propagator propagator exception thrown cxx in v noid do void ntraceback most recent call last n file data user grenzi combo src simprod scripts resources scripts clsim npy line in n main n file data user grenzi combo src simprod scripts resources scripts clsim npy line in main n usegslrng params n file data user grenzi combo build lib icecube simprod util simprodtray py line in n execute tray executionmaxcount n file data user grenzi combo build lib icecube simprod util simprodtray py line in execute n tray execute n file data user grenzi combo build lib py line in execute n super self execute nruntimeerror kaellen function is negative cannot caluclate momentum in static double proposal decaychannel momentum double double double n n nthe error seems to come from here at line n n n double decaychannel momentum double double double t t double kaellen t t if kaellen t t return std sqrt kaellen t else t t log fatal kaellen function is negative cannot caluclate momentum t return t t n n n nthis error prevents the propagation process for all of the following frames i think the right thing would be just skipping the single wrong one n reporter grenzi cc resolution fixed ts component combo simulation summary fatal in decay priority major keywords time milestone autumnal equinox owner jsoedingrekso type defect
1
154,353
24,281,391,238
IssuesEvent
2022-09-28 17:44:15
microsoft/pylance-release
https://api.github.com/repos/microsoft/pylance-release
closed
Python language server no longer recommends auto imports from local files (still works for site-packages in venv)
by design
Type: <b>Bug</b> Unsure, it worked 3 days ago, not it doesn't, didn't change any workspace settings what so ever. Extension version: 2022.14.0 VS Code version: Code 1.71.2 (Universal) (74b1f979648cc44d385a2286793c226e611f59e7, 2022-09-14T21:05:37.721Z) OS version: Darwin x64 21.6.0 Modes: Sandboxed: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i9-9880H CPU @ 2.30GHz (16 x 2300)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off| |Load (avg)|2, 2, 2| |Memory (System)|32.00GB (6.37GB free)| |Process Argv|--crash-reporter-id 0731fb54-df68-48c8-9bc7-cb94d92db4d8| |Screen Reader|no| |VM|0%| </details> <!-- generated by issue reporter -->
1.0
Python language server no longer recommends auto imports from local files (still works for site-packages in venv) - Type: <b>Bug</b> Unsure, it worked 3 days ago, not it doesn't, didn't change any workspace settings what so ever. Extension version: 2022.14.0 VS Code version: Code 1.71.2 (Universal) (74b1f979648cc44d385a2286793c226e611f59e7, 2022-09-14T21:05:37.721Z) OS version: Darwin x64 21.6.0 Modes: Sandboxed: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|Intel(R) Core(TM) i9-9880H CPU @ 2.30GHz (16 x 2300)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off| |Load (avg)|2, 2, 2| |Memory (System)|32.00GB (6.37GB free)| |Process Argv|--crash-reporter-id 0731fb54-df68-48c8-9bc7-cb94d92db4d8| |Screen Reader|no| |VM|0%| </details> <!-- generated by issue reporter -->
non_defect
python language server no longer recommends auto imports from local files still works for site packages in venv type bug unsure it worked days ago not it doesn t didn t change any workspace settings what so ever extension version vs code version code universal os version darwin modes sandboxed no system info item value cpus intel r core tm cpu x gpu status canvas enabled canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing enabled metal disabled off multiple raster threads enabled on opengl enabled on rasterization enabled raw draw disabled off ok skia renderer enabled on video decode enabled video encode enabled vulkan disabled off webgl enabled enabled webgpu disabled off load avg memory system free process argv crash reporter id screen reader no vm
0
78,426
27,516,800,032
IssuesEvent
2023-03-06 12:32:22
primefaces/primefaces
https://api.github.com/repos/primefaces/primefaces
closed
Tree: pfAjaxComplete and pfAjaxStart not working when filter a tree
:lady_beetle: defect :bangbang: needs-triage
### Describe the bug When a Tree is filtered, an Ajax call is performed (see browser console), but event handlers for `pfAjaxStart` `pfAjaxComplete` don't react. ### Reproducer [primefaces-ajax-trigger-missing.zip](https://github.com/primefaces/primefaces/files/10897430/primefaces-ajax-trigger-missing.zip) ```xhtml <p:tree value="#{testView.root}" var="node" filterBy="#{node}"> <p:treeNode> <h:outputText value="#{node}" /> </p:treeNode> </p:tree> <script> $(document).on('pfAjaxStart',()=>console.info('Test')) </script> ``` Also reproducable in PrimeFaces Showcase (type `$(document).on('pfAjaxStart',()=>console.info('Test'))` in the browser console and look for messages. Compare [Tree](https://www.primefaces.org/showcase/ui/data/tree/filter.xhtml) with [Data Table](https://www.primefaces.org/showcase/ui/data/datatable/filter.xhtml). ### Expected behavior `pfAjaxStart` and `pfAjaxComplete` should be called. ### PrimeFaces edition Community ### PrimeFaces version 12.0.0 ### Theme _No response_ ### JSF implementation Mojarra ### JSF version 2.2 ### Java version 17 ### Browser(s) _No response_
1.0
Tree: pfAjaxComplete and pfAjaxStart not working when filter a tree - ### Describe the bug When a Tree is filtered, an Ajax call is performed (see browser console), but event handlers for `pfAjaxStart` `pfAjaxComplete` don't react. ### Reproducer [primefaces-ajax-trigger-missing.zip](https://github.com/primefaces/primefaces/files/10897430/primefaces-ajax-trigger-missing.zip) ```xhtml <p:tree value="#{testView.root}" var="node" filterBy="#{node}"> <p:treeNode> <h:outputText value="#{node}" /> </p:treeNode> </p:tree> <script> $(document).on('pfAjaxStart',()=>console.info('Test')) </script> ``` Also reproducable in PrimeFaces Showcase (type `$(document).on('pfAjaxStart',()=>console.info('Test'))` in the browser console and look for messages. Compare [Tree](https://www.primefaces.org/showcase/ui/data/tree/filter.xhtml) with [Data Table](https://www.primefaces.org/showcase/ui/data/datatable/filter.xhtml). ### Expected behavior `pfAjaxStart` and `pfAjaxComplete` should be called. ### PrimeFaces edition Community ### PrimeFaces version 12.0.0 ### Theme _No response_ ### JSF implementation Mojarra ### JSF version 2.2 ### Java version 17 ### Browser(s) _No response_
defect
tree pfajaxcomplete and pfajaxstart not working when filter a tree describe the bug when a tree is filtered an ajax call is performed see browser console but event handlers for pfajaxstart pfajaxcomplete don t react reproducer xhtml document on pfajaxstart console info test also reproducable in primefaces showcase type document on pfajaxstart console info test in the browser console and look for messages compare with expected behavior pfajaxstart and pfajaxcomplete should be called primefaces edition community primefaces version theme no response jsf implementation mojarra jsf version java version browser s no response
1
4,879
2,610,159,473
IssuesEvent
2015-02-26 18:50:36
chrsmith/republic-at-war
https://api.github.com/repos/chrsmith/republic-at-war
closed
Map Issue
auto-migrated Priority-Medium Type-Defect
``` Naboo... Some gungans trapped in trees.. possible passability issue ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 12:35
1.0
Map Issue - ``` Naboo... Some gungans trapped in trees.. possible passability issue ``` ----- Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 12:35
defect
map issue naboo some gungans trapped in trees possible passability issue original issue reported on code google com by gmail com on jan at
1
64,365
18,506,943,317
IssuesEvent
2021-10-19 19:51:00
scipy/scipy
https://api.github.com/repos/scipy/scipy
opened
BUG: NumPy's `random` module should not be in the `scipy` namespace.
defect
### Describe your issue. Currently, `scipy.random` is an alias for `numpy.random`: ``` In [12]: import numpy as np In [13]: import scipy In [14]: scipy.__version__ Out[14]: '1.8.0.dev0+1859.14341ea' In [15]: scipy.random Out[15]: <module 'numpy.random' from '/home/warren/mc39scipy/lib/python3.9/site-packages/numpy/random/__init__.py'> In [16]: scipy.random is np.random Out[16]: True ``` We have already deprecated other names in `scipy` that come from `numpy`, e.g. ``` In [17]: scipy.array([1, 2, 3]) <ipython-input-17-773c09c10413>:1: DeprecationWarning: scipy.array is deprecated and will be removed in SciPy 2.0.0, use numpy.array instead scipy.array([1, 2, 3]) Out[17]: array([1, 2, 3]) ``` We should to the same for `random`, so a call such as the following generates a deprecation warning: ``` In [18]: scipy.random.randint(10, size=3) Out[18]: array([8, 5, 5]) ``` ### Reproducing Code Example ```python See above. ``` ### Error message ```shell See above. ``` ### SciPy/NumPy/Python version information 1.8.0.dev0+1859.14341ea 1.21.2 sys.version_info(major=3, minor=9, micro=5, releaselevel='final', serial=0)
1.0
BUG: NumPy's `random` module should not be in the `scipy` namespace. - ### Describe your issue. Currently, `scipy.random` is an alias for `numpy.random`: ``` In [12]: import numpy as np In [13]: import scipy In [14]: scipy.__version__ Out[14]: '1.8.0.dev0+1859.14341ea' In [15]: scipy.random Out[15]: <module 'numpy.random' from '/home/warren/mc39scipy/lib/python3.9/site-packages/numpy/random/__init__.py'> In [16]: scipy.random is np.random Out[16]: True ``` We have already deprecated other names in `scipy` that come from `numpy`, e.g. ``` In [17]: scipy.array([1, 2, 3]) <ipython-input-17-773c09c10413>:1: DeprecationWarning: scipy.array is deprecated and will be removed in SciPy 2.0.0, use numpy.array instead scipy.array([1, 2, 3]) Out[17]: array([1, 2, 3]) ``` We should to the same for `random`, so a call such as the following generates a deprecation warning: ``` In [18]: scipy.random.randint(10, size=3) Out[18]: array([8, 5, 5]) ``` ### Reproducing Code Example ```python See above. ``` ### Error message ```shell See above. ``` ### SciPy/NumPy/Python version information 1.8.0.dev0+1859.14341ea 1.21.2 sys.version_info(major=3, minor=9, micro=5, releaselevel='final', serial=0)
defect
bug numpy s random module should not be in the scipy namespace describe your issue currently scipy random is an alias for numpy random in import numpy as np in import scipy in scipy version out in scipy random out in scipy random is np random out true we have already deprecated other names in scipy that come from numpy e g in scipy array deprecationwarning scipy array is deprecated and will be removed in scipy use numpy array instead scipy array out array we should to the same for random so a call such as the following generates a deprecation warning in scipy random randint size out array reproducing code example python see above error message shell see above scipy numpy python version information sys version info major minor micro releaselevel final serial
1
67,535
20,980,637,225
IssuesEvent
2022-03-28 19:32:59
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
closed
CFDI 4.0 - exportación y régimen. Enviar clave y descripción al PDF
resolved defect
Ahora mismo solo envía la clave, y se requiere la descripción también. En general para todos los datos que son claves, enviar al PDF la descripción.
1.0
CFDI 4.0 - exportación y régimen. Enviar clave y descripción al PDF - Ahora mismo solo envía la clave, y se requiere la descripción también. En general para todos los datos que son claves, enviar al PDF la descripción.
defect
cfdi exportación y régimen enviar clave y descripción al pdf ahora mismo solo envía la clave y se requiere la descripción también en general para todos los datos que son claves enviar al pdf la descripción
1
29,017
5,477,086,453
IssuesEvent
2017-03-12 03:45:37
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
postgresql: NUMERIC field in database appears as float when debug($entity)
Defect
This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.4.3 / Postgresql 9.4 ### What you did 1- defined a field as type NUMERIC(1) in PostgreSQL database 2- did a find() over the table ### What happened when I debug() over the query the field appears as float: ``` ... 'sgn' => (float) -1, ... ``` ### What you expected to happen should appear as integer
1.0
postgresql: NUMERIC field in database appears as float when debug($entity) - This is a (multiple allowed): * [x] bug * [ ] enhancement * [ ] feature-discussion (RFC) * CakePHP Version: 3.4.3 / Postgresql 9.4 ### What you did 1- defined a field as type NUMERIC(1) in PostgreSQL database 2- did a find() over the table ### What happened when I debug() over the query the field appears as float: ``` ... 'sgn' => (float) -1, ... ``` ### What you expected to happen should appear as integer
defect
postgresql numeric field in database appears as float when debug entity this is a multiple allowed bug enhancement feature discussion rfc cakephp version postgresql what you did defined a field as type numeric in postgresql database did a find over the table what happened when i debug over the query the field appears as float sgn float what you expected to happen should appear as integer
1
44,035
11,914,518,904
IssuesEvent
2020-03-31 13:43:18
SasView/sasview
https://api.github.com/repos/SasView/sasview
closed
When loading plugin model with a spurious unicode character plugin editor crashes
CCX: GUI defect
This a follow up to #1426. There is currently no handling for the a spurious unicode character in plugin model editor (Edit custom model), so it crashes when one tries to edit it.
1.0
When loading plugin model with a spurious unicode character plugin editor crashes - This a follow up to #1426. There is currently no handling for the a spurious unicode character in plugin model editor (Edit custom model), so it crashes when one tries to edit it.
defect
when loading plugin model with a spurious unicode character plugin editor crashes this a follow up to there is currently no handling for the a spurious unicode character in plugin model editor edit custom model so it crashes when one tries to edit it
1
234,143
7,717,198,671
IssuesEvent
2018-05-23 13:07:09
unfoldingWord-dev/translationCore
https://api.github.com/repos/unfoldingWord-dev/translationCore
closed
Again able to put GL words in wrong order
Kind/Bug Priority/Medium QA/Pass
beta.92 - same in rc2 [fr_tit_text_ulb.tstudio.zip](https://github.com/unfoldingWord-dev/translationCore/files/1381075/fr_tit_text_ulb.tstudio.zip) 1. ![image](https://user-images.githubusercontent.com/8171759/31524739-393e161c-af81-11e7-9d06-e10ac51d6b40.png) Move the French words into the target in this order: sévèrement, témoignage, reprimende, Ce 2. ![image](https://user-images.githubusercontent.com/8171759/31524885-1b1bcdb8-af82-11e7-83b5-6644e0fdf12a.png) Move the French words into the target in this order: saine, foi, une
1.0
Again able to put GL words in wrong order - beta.92 - same in rc2 [fr_tit_text_ulb.tstudio.zip](https://github.com/unfoldingWord-dev/translationCore/files/1381075/fr_tit_text_ulb.tstudio.zip) 1. ![image](https://user-images.githubusercontent.com/8171759/31524739-393e161c-af81-11e7-9d06-e10ac51d6b40.png) Move the French words into the target in this order: sévèrement, témoignage, reprimende, Ce 2. ![image](https://user-images.githubusercontent.com/8171759/31524885-1b1bcdb8-af82-11e7-83b5-6644e0fdf12a.png) Move the French words into the target in this order: saine, foi, une
non_defect
again able to put gl words in wrong order beta same in move the french words into the target in this order sévèrement témoignage reprimende ce move the french words into the target in this order saine foi une
0
108,214
16,762,817,311
IssuesEvent
2021-06-14 03:16:33
gms-ws-sandbox/nibrs-pr-test
https://api.github.com/repos/gms-ws-sandbox/nibrs-pr-test
opened
CVE-2020-11620 (High) detected in multiple libraries
security vulnerability
## CVE-2020-11620 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.6.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.8.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.8.0.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-summary-report-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.5.RELEASE.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-validate-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: nibrs-pr-test/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.10.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-route/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs-pr-test/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs-pr-test/commit/860cc22f54e17594e32e303f0716fb065202fff5">860cc22f54e17594e32e303f0716fb065202fff5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly). <p>Publish Date: 2020-04-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620>CVE-2020-11620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620</a></p> <p>Release Date: 2020-04-07</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-route/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-validation/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11620","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11620 (High) detected in multiple libraries - ## CVE-2020-11620 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.6.jar</b></p></summary> <p> <details><summary><b>jackson-databind-2.8.0.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.8.0.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.8.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-summary-report-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library) - spring-boot-starter-json-2.1.5.RELEASE.jar - :x: **jackson-databind-2.9.8.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.5.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-validate-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - :x: **jackson-databind-2.9.5.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.8.10.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-fbi-service/pom.xml</p> <p>Path to vulnerable library: nibrs-pr-test/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.8.10.jar** (Vulnerable Library) </details> <details><summary><b>jackson-databind-2.9.6.jar</b></p></summary> <p>General data-binding functionality for Jackson: works on core streaming API</p> <p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p> <p>Path to dependency file: nibrs-pr-test/tools/nibrs-route/pom.xml</p> <p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs-pr-test/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p> <p> Dependency Hierarchy: - :x: **jackson-databind-2.9.6.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs-pr-test/commit/860cc22f54e17594e32e303f0716fb065202fff5">860cc22f54e17594e32e303f0716fb065202fff5</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly). <p>Publish Date: 2020-04-07 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620>CVE-2020-11620</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620</a></p> <p>Release Date: 2020-04-07</p> <p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-route/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-validation/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11620","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
non_defect
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs summary report common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs validate common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs fbi service pom xml path to vulnerable library nibrs pr test tools nibrs fbi service target nibrs fbi service web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs route pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar nibrs pr test web nibrs web target nibrs web web inf lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons jelly impl embedded aka commons jelly publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter json release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons jelly impl embedded aka commons jelly vulnerabilityurl
0
1,154
2,598,004,861
IssuesEvent
2015-02-22 01:41:55
chrsmith/bwapi
https://api.github.com/repos/chrsmith/bwapi
opened
BWAPI 4 example projects still reference BWTA
auto-migrated Milestone-Release Priority-Critical Type-Defect
``` What steps will reproduce the problem? 1. Download the BWAPI 4.0.0 and install it as a full installation from https://code.google.com/p/bwapi/downloads/list including the additional gateways. 2. Open the example.sln in a compiler like Microsoft Visual Studio. 3. Attempt to build ExampleAIModule in release mode or debug mode. What is the expected output? What do you see instead? Expected output in release mode: 1>------ Build started: Project: ExampleAIModule, Configuration: Release Win32 ------ 1> Dll.cpp 1> ExampleAIModule.cpp 1>LINK : warning LNK4075: ignoring '/INCREMENTAL' due to '/LTCG' specification 1> Creating library C:\Program Files (x86)\BWAPI\Release\ExampleAIModule.lib and object C:\Program Files (x86)\BWAPI\Release\ExampleAIModule.exp 1> Generating code 1> Finished generating code 1> ExampleAIModule.vcxproj -> C:\Program Files (x86)\BWAPI\Release\ExampleAIModule.dll ========== Build: 1 succeeded, 0 failed, 0 up-to-date, 0 skipped ========== Instead it stops after: 1>LINK : fatal error LNK1104: cannot open file '../lib/BWTAd.lib' What version of the product are you using? On what operating system? Using BWAPI 4.0.0 beta on Windows 7 64 bit. Please provide any additional information below. If you remove the BWTA.lib from properties->linker->Input->Additional Dependencies you will also get the errors of other libraries like CGAL.lib or tinyxml.lib being missing. Other than BWAPI.lib, all other libs from previous version are missing. Workaround is to copy the BWAPI 3.7.4 libs into BWAPI 4.0.0 and unlink the currently not working BWTA.lib and any others. ``` ----- Original issue reported on code.google.com by `georgetw...@gmail.com` on 24 Jun 2013 at 12:55
1.0
BWAPI 4 example projects still reference BWTA - ``` What steps will reproduce the problem? 1. Download the BWAPI 4.0.0 and install it as a full installation from https://code.google.com/p/bwapi/downloads/list including the additional gateways. 2. Open the example.sln in a compiler like Microsoft Visual Studio. 3. Attempt to build ExampleAIModule in release mode or debug mode. What is the expected output? What do you see instead? Expected output in release mode: 1>------ Build started: Project: ExampleAIModule, Configuration: Release Win32 ------ 1> Dll.cpp 1> ExampleAIModule.cpp 1>LINK : warning LNK4075: ignoring '/INCREMENTAL' due to '/LTCG' specification 1> Creating library C:\Program Files (x86)\BWAPI\Release\ExampleAIModule.lib and object C:\Program Files (x86)\BWAPI\Release\ExampleAIModule.exp 1> Generating code 1> Finished generating code 1> ExampleAIModule.vcxproj -> C:\Program Files (x86)\BWAPI\Release\ExampleAIModule.dll ========== Build: 1 succeeded, 0 failed, 0 up-to-date, 0 skipped ========== Instead it stops after: 1>LINK : fatal error LNK1104: cannot open file '../lib/BWTAd.lib' What version of the product are you using? On what operating system? Using BWAPI 4.0.0 beta on Windows 7 64 bit. Please provide any additional information below. If you remove the BWTA.lib from properties->linker->Input->Additional Dependencies you will also get the errors of other libraries like CGAL.lib or tinyxml.lib being missing. Other than BWAPI.lib, all other libs from previous version are missing. Workaround is to copy the BWAPI 3.7.4 libs into BWAPI 4.0.0 and unlink the currently not working BWTA.lib and any others. ``` ----- Original issue reported on code.google.com by `georgetw...@gmail.com` on 24 Jun 2013 at 12:55
defect
bwapi example projects still reference bwta what steps will reproduce the problem download the bwapi and install it as a full installation from including the additional gateways open the example sln in a compiler like microsoft visual studio attempt to build exampleaimodule in release mode or debug mode what is the expected output what do you see instead expected output in release mode build started project exampleaimodule configuration release dll cpp exampleaimodule cpp link warning ignoring incremental due to ltcg specification creating library c program files bwapi release exampleaimodule lib and object c program files bwapi release exampleaimodule exp generating code finished generating code exampleaimodule vcxproj c program files bwapi release exampleaimodule dll build succeeded failed up to date skipped instead it stops after link fatal error cannot open file lib bwtad lib what version of the product are you using on what operating system using bwapi beta on windows bit please provide any additional information below if you remove the bwta lib from properties linker input additional dependencies you will also get the errors of other libraries like cgal lib or tinyxml lib being missing other than bwapi lib all other libs from previous version are missing workaround is to copy the bwapi libs into bwapi and unlink the currently not working bwta lib and any others original issue reported on code google com by georgetw gmail com on jun at
1
350,323
31,879,341,758
IssuesEvent
2023-09-16 07:10:58
depromeet/street-drop-server
https://api.github.com/repos/depromeet/street-drop-server
opened
Add Test Case to Geo Controller and Service
✅test
## DESCRIPTION - Add Test Case to Geo Controller and Service
1.0
Add Test Case to Geo Controller and Service - ## DESCRIPTION - Add Test Case to Geo Controller and Service
non_defect
add test case to geo controller and service description add test case to geo controller and service
0
80,605
30,386,748,880
IssuesEvent
2023-07-13 01:52:42
SeleniumHQ/selenium
https://api.github.com/repos/SeleniumHQ/selenium
opened
[🐛 Bug]: Grid closeed not wrigh, when click close button just close the remot desk page,but not disconnect the vnc.
I-defect needs-triaging
### What happened? when click close button, just close the remot desk page,but not disconnect the vnc. ### How can we reproduce the issue? ```shell 1.connect remot desk window over grid session page. 2.open chrome develped tools,watch the network tab. 3.you can see new request over grid. 4.click close button. 5.you can see always request over grid vnc. ``` ### Relevant log output ```shell blank ``` ### Operating System linux ### Selenium version 4.8.3 ### What are the browser(s) and version(s) where you see this issue? chrome 111 ### What are the browser driver(s) and version(s) where you see this issue? use grid docker ### Are you using Selenium Grid? 4.8.3
1.0
[🐛 Bug]: Grid closeed not wrigh, when click close button just close the remot desk page,but not disconnect the vnc. - ### What happened? when click close button, just close the remot desk page,but not disconnect the vnc. ### How can we reproduce the issue? ```shell 1.connect remot desk window over grid session page. 2.open chrome develped tools,watch the network tab. 3.you can see new request over grid. 4.click close button. 5.you can see always request over grid vnc. ``` ### Relevant log output ```shell blank ``` ### Operating System linux ### Selenium version 4.8.3 ### What are the browser(s) and version(s) where you see this issue? chrome 111 ### What are the browser driver(s) and version(s) where you see this issue? use grid docker ### Are you using Selenium Grid? 4.8.3
defect
grid closeed not wrigh when click close button just close the remot desk page but not disconnect the vnc what happened when click close button, just close the remot desk page but not disconnect the vnc how can we reproduce the issue shell connect remot desk window over grid session page open chrome develped tools watch the network tab you can see new request over grid click close button you can see always request over grid vnc relevant log output shell blank operating system linux selenium version what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue use grid docker are you using selenium grid
1
63,902
7,751,514,491
IssuesEvent
2018-05-30 17:18:22
researchstudio-sat/webofneeds
https://api.github.com/repos/researchstudio-sat/webofneeds
closed
Draft to Redux-State
UX User Story design decision enhancement suggestion
The draft used to be in the redux-state, it would be nice to have it there again. This would avoid data-loss if people leave the create-post view for a sec. Also it would enable later caching systems that allow to finish editing a draft later (e.g. by storing them in localstorage).
1.0
Draft to Redux-State - The draft used to be in the redux-state, it would be nice to have it there again. This would avoid data-loss if people leave the create-post view for a sec. Also it would enable later caching systems that allow to finish editing a draft later (e.g. by storing them in localstorage).
non_defect
draft to redux state the draft used to be in the redux state it would be nice to have it there again this would avoid data loss if people leave the create post view for a sec also it would enable later caching systems that allow to finish editing a draft later e g by storing them in localstorage
0
73,127
24,469,482,213
IssuesEvent
2022-10-07 18:16:36
BOINC/boinc
https://api.github.com/repos/BOINC/boinc
closed
Work Fetch - Leaves part of a GPU unused, when it should instead fetch work
C: Client - Work Fetch Policy P: Major R: wontfix T: Defect
**Reported by JacobKlein on 25 Mar 43272124 22:13 UTC** If a project's GPU apps are setup to use only part of the GPU (ie: app_config.xml), then when the last remaining task(s) for that project are running and not utilizing the full GPU, work fetch should fetch more, but doesn't. This issue was confirmed with both 7.0.60, as well as on 4/8/2013 in the simulator (which has several unreleased work fetch changes). It would seem that the prerequisites to reproducing the bug are: - use an app_config.xml file (to set an app to use part of a GPU, so multiple tasks could run at the same time on the same device). - use a small buffer setting I'm not certain if GPU Exclusions are necessary to create the issue, but I believe that using GPU Exclusions makes this problem worse. As a workaround, I had to increase my buffer settings way above what I would normally expect. It feels like, in addition to work fetch not realizing a portion of the GPU is idle, it might also not be realizing that the tasks run 2-at-a-time. Details, including examples in a simulation, are in the email below: --- From: jacob_w_klein@msn.com[davea@ssl.berkeley.edu[[BR]([BR]]To:)]Subject: RE: job scheduling[Mon, 8 Apr 2013 09:51:16 -0400[[BR]([BR]]Date:)][ Thank you. I really appreciate you looking at these issues, and I'll try to verify they work.[[BR]([BR]])]Your WCG project sounds interesting; maybe they're going to support Android?[wish we had a Windows Phone platform, I'd love to test on it.[[BR]([BR]]I)][you remember Ed (Beyond) reporting a GPU Exclusion Work Fetch issue?[[BR]([BR]]Do)]I might have found examples of what he was trying to explain...[noticing an issue, both on my computer (7.0.60's work fetch algorithm), as well as the simulator (new work fetch algorithm).[[BR]([BR]][[BR]]I'm)]If a GPU is only partially-loaded (ie: 0.5 GPU) by the last remaining task(s) for a project that has GPU-Exclusions,[get into a scenario where GPUs are left part-idle, and work fetch won't fetch more.[[BR]([BR]]We)][task scheduler (correctly) schedules the workload, which is scheduled in a way where a GPU is left part-idle,[[BR]([BR]]The)]But work fetch thinks we have plenty of work, and sees no fully idle instances, so it doesn't ask for any.[are some examples where that occurred, even with our work fetch changes:[[BR]([BR]][[BR]]Here)][[[BR]([BR]]http://boinc.berkeley.edu/dev/sim_web.php?action=show_simulation&scen=86&sim=26)]2 days 17:03:00[days 14:33:00[[BR]([BR]]3)]6 days 06:13:00[days 16:43:00[[BR]([BR]]8)]9 days 16:07:00[fix might involve evaluating the project's GPU apps to see if it has any that use partial GPU[[BR]([BR]][[BR]]The)]... or maybe checking to see that all of its GPU apps use <= amount of currently idle GPU (to ensure we don't keep asking/getting work we cannot immediately use)[sounds to me like the fix for this one might be tricky instead of straight-forward, though I'm not sure.[[BR]([BR]][[BR]]It)]Do you plan on tackling this soon (fixed in short term), or should I create a ticket (fixed eventually, maybe months/years)?[[BR]][[BR]]Regards,[[BR]]Jacob Migrated-From: http://boinc.berkeley.edu/trac/ticket/1239
1.0
Work Fetch - Leaves part of a GPU unused, when it should instead fetch work - **Reported by JacobKlein on 25 Mar 43272124 22:13 UTC** If a project's GPU apps are setup to use only part of the GPU (ie: app_config.xml), then when the last remaining task(s) for that project are running and not utilizing the full GPU, work fetch should fetch more, but doesn't. This issue was confirmed with both 7.0.60, as well as on 4/8/2013 in the simulator (which has several unreleased work fetch changes). It would seem that the prerequisites to reproducing the bug are: - use an app_config.xml file (to set an app to use part of a GPU, so multiple tasks could run at the same time on the same device). - use a small buffer setting I'm not certain if GPU Exclusions are necessary to create the issue, but I believe that using GPU Exclusions makes this problem worse. As a workaround, I had to increase my buffer settings way above what I would normally expect. It feels like, in addition to work fetch not realizing a portion of the GPU is idle, it might also not be realizing that the tasks run 2-at-a-time. Details, including examples in a simulation, are in the email below: --- From: jacob_w_klein@msn.com[davea@ssl.berkeley.edu[[BR]([BR]]To:)]Subject: RE: job scheduling[Mon, 8 Apr 2013 09:51:16 -0400[[BR]([BR]]Date:)][ Thank you. I really appreciate you looking at these issues, and I'll try to verify they work.[[BR]([BR]])]Your WCG project sounds interesting; maybe they're going to support Android?[wish we had a Windows Phone platform, I'd love to test on it.[[BR]([BR]]I)][you remember Ed (Beyond) reporting a GPU Exclusion Work Fetch issue?[[BR]([BR]]Do)]I might have found examples of what he was trying to explain...[noticing an issue, both on my computer (7.0.60's work fetch algorithm), as well as the simulator (new work fetch algorithm).[[BR]([BR]][[BR]]I'm)]If a GPU is only partially-loaded (ie: 0.5 GPU) by the last remaining task(s) for a project that has GPU-Exclusions,[get into a scenario where GPUs are left part-idle, and work fetch won't fetch more.[[BR]([BR]]We)][task scheduler (correctly) schedules the workload, which is scheduled in a way where a GPU is left part-idle,[[BR]([BR]]The)]But work fetch thinks we have plenty of work, and sees no fully idle instances, so it doesn't ask for any.[are some examples where that occurred, even with our work fetch changes:[[BR]([BR]][[BR]]Here)][[[BR]([BR]]http://boinc.berkeley.edu/dev/sim_web.php?action=show_simulation&scen=86&sim=26)]2 days 17:03:00[days 14:33:00[[BR]([BR]]3)]6 days 06:13:00[days 16:43:00[[BR]([BR]]8)]9 days 16:07:00[fix might involve evaluating the project's GPU apps to see if it has any that use partial GPU[[BR]([BR]][[BR]]The)]... or maybe checking to see that all of its GPU apps use <= amount of currently idle GPU (to ensure we don't keep asking/getting work we cannot immediately use)[sounds to me like the fix for this one might be tricky instead of straight-forward, though I'm not sure.[[BR]([BR]][[BR]]It)]Do you plan on tackling this soon (fixed in short term), or should I create a ticket (fixed eventually, maybe months/years)?[[BR]][[BR]]Regards,[[BR]]Jacob Migrated-From: http://boinc.berkeley.edu/trac/ticket/1239
defect
work fetch leaves part of a gpu unused when it should instead fetch work reported by jacobklein on mar utc if a project s gpu apps are setup to use only part of the gpu ie app config xml then when the last remaining task s for that project are running and not utilizing the full gpu work fetch should fetch more but doesn t this issue was confirmed with both as well as on in the simulator which has several unreleased work fetch changes it would seem that the prerequisites to reproducing the bug are use an app config xml file to set an app to use part of a gpu so multiple tasks could run at the same time on the same device use a small buffer setting i m not certain if gpu exclusions are necessary to create the issue but i believe that using gpu exclusions makes this problem worse as a workaround i had to increase my buffer settings way above what i would normally expect it feels like in addition to work fetch not realizing a portion of the gpu is idle it might also not be realizing that the tasks run at a time details including examples in a simulation are in the email below from jacob w klein msn com to subject re job scheduling date your wcg project sounds interesting maybe they re going to support android i do i might have found examples of what he was trying to explain i m if a gpu is only partially loaded ie gpu by the last remaining task s for a project that has gpu exclusions we the but work fetch thinks we have plenty of work and sees no fully idle instances so it doesn t ask for any here days days days the or maybe checking to see that all of its gpu apps use amount of currently idle gpu to ensure we don t keep asking getting work we cannot immediately use it do you plan on tackling this soon fixed in short term or should i create a ticket fixed eventually maybe months years regards jacob migrated from
1
212,720
16,494,949,770
IssuesEvent
2021-05-25 09:19:37
jeongukjae/tfds-korean
https://api.github.com/repos/jeongukjae/tfds-korean
closed
Add CONTRIBUTING.md
before-release documentation
- [ ] 프로젝트에서 사용하는 언어에 대한 설명. 사용법/데이터셋 설명은 가능하면 영어로 적되, 이슈/PR 소통은 한국어로 하는게 좋지 않을까? - [ ] 데이터셋 추가하는 법 - [ ] 이슈/PR/Discussion 간단한 설명 - [ ] 추가로 같이 관리하고 싶은 분들에 대한 설명 - [ ] 데이터셋 라이선스에 대한 문제에 대한 설명
1.0
Add CONTRIBUTING.md - - [ ] 프로젝트에서 사용하는 언어에 대한 설명. 사용법/데이터셋 설명은 가능하면 영어로 적되, 이슈/PR 소통은 한국어로 하는게 좋지 않을까? - [ ] 데이터셋 추가하는 법 - [ ] 이슈/PR/Discussion 간단한 설명 - [ ] 추가로 같이 관리하고 싶은 분들에 대한 설명 - [ ] 데이터셋 라이선스에 대한 문제에 대한 설명
non_defect
add contributing md 프로젝트에서 사용하는 언어에 대한 설명 사용법 데이터셋 설명은 가능하면 영어로 적되 이슈 pr 소통은 한국어로 하는게 좋지 않을까 데이터셋 추가하는 법 이슈 pr discussion 간단한 설명 추가로 같이 관리하고 싶은 분들에 대한 설명 데이터셋 라이선스에 대한 문제에 대한 설명
0
45,792
13,055,749,812
IssuesEvent
2020-07-30 02:37:26
icecube-trac/tix2
https://api.github.com/repos/icecube-trac/tix2
opened
cmake header-only projects. (Trac #62)
Incomplete Migration Migrated from Trac cmake defect
Migrated from https://code.icecube.wisc.edu/ticket/62 ```json { "status": "closed", "changetime": "2007-11-09T22:34:28", "description": "header-only and executable-only projects aren't supported.", "reporter": "troy", "cc": "", "resolution": "duplicate", "_ts": "1194647668000000", "component": "cmake", "summary": "cmake header-only projects.", "priority": "normal", "keywords": "", "time": "2007-06-12T17:56:22", "milestone": "", "owner": "troy", "type": "defect" } ```
1.0
cmake header-only projects. (Trac #62) - Migrated from https://code.icecube.wisc.edu/ticket/62 ```json { "status": "closed", "changetime": "2007-11-09T22:34:28", "description": "header-only and executable-only projects aren't supported.", "reporter": "troy", "cc": "", "resolution": "duplicate", "_ts": "1194647668000000", "component": "cmake", "summary": "cmake header-only projects.", "priority": "normal", "keywords": "", "time": "2007-06-12T17:56:22", "milestone": "", "owner": "troy", "type": "defect" } ```
defect
cmake header only projects trac migrated from json status closed changetime description header only and executable only projects aren t supported reporter troy cc resolution duplicate ts component cmake summary cmake header only projects priority normal keywords time milestone owner troy type defect
1
731,796
25,231,336,992
IssuesEvent
2022-11-14 20:07:05
PrefectHQ/prefect
https://api.github.com/repos/PrefectHQ/prefect
closed
Ensure link styling gets consistently applied to all clickable links
ui priority:high
### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I refreshed the page and this issue still occurred. - [X] I checked if this issue was specific to the browser I was using by testing with a different browser. ### Bug summary <img width="755" alt="image" src="https://user-images.githubusercontent.com/40272060/198748508-d7844d45-d03c-404b-9982-fc26f94c9312.png"> We have added link styling to work queues to make it more clear they are clickable links. We should add the same to deployment and flow run links. In addition, let's clean up the Flow Run and Deployment wells so that they have a consistent order and group the "primary meta" details together. [Here's the figma design.](https://www.figma.com/file/WQwcuqBAxug8JqLigWaztu/Designs?node-id=1600%3A19270) Notice that date formatting is more consistent as well. <img width="1307" alt="Screen Shot 2022-11-01 at 11 05 37 AM" src="https://user-images.githubusercontent.com/6776415/199266121-bb787d15-c162-4f23-b71f-a906b4fa3ee2.png"> ### Reproduction See image above ### Error See image above ### Browers - [X] Chrome - [ ] Firefox - [ ] Safari - [ ] Edge ### Prefect version _No response_ ### Additional context _No response_
1.0
Ensure link styling gets consistently applied to all clickable links - ### First check - [X] I added a descriptive title to this issue. - [X] I used the GitHub search to find a similar issue and didn't find it. - [X] I refreshed the page and this issue still occurred. - [X] I checked if this issue was specific to the browser I was using by testing with a different browser. ### Bug summary <img width="755" alt="image" src="https://user-images.githubusercontent.com/40272060/198748508-d7844d45-d03c-404b-9982-fc26f94c9312.png"> We have added link styling to work queues to make it more clear they are clickable links. We should add the same to deployment and flow run links. In addition, let's clean up the Flow Run and Deployment wells so that they have a consistent order and group the "primary meta" details together. [Here's the figma design.](https://www.figma.com/file/WQwcuqBAxug8JqLigWaztu/Designs?node-id=1600%3A19270) Notice that date formatting is more consistent as well. <img width="1307" alt="Screen Shot 2022-11-01 at 11 05 37 AM" src="https://user-images.githubusercontent.com/6776415/199266121-bb787d15-c162-4f23-b71f-a906b4fa3ee2.png"> ### Reproduction See image above ### Error See image above ### Browers - [X] Chrome - [ ] Firefox - [ ] Safari - [ ] Edge ### Prefect version _No response_ ### Additional context _No response_
non_defect
ensure link styling gets consistently applied to all clickable links first check i added a descriptive title to this issue i used the github search to find a similar issue and didn t find it i refreshed the page and this issue still occurred i checked if this issue was specific to the browser i was using by testing with a different browser bug summary img width alt image src we have added link styling to work queues to make it more clear they are clickable links we should add the same to deployment and flow run links in addition let s clean up the flow run and deployment wells so that they have a consistent order and group the primary meta details together notice that date formatting is more consistent as well img width alt screen shot at am src reproduction see image above error see image above browers chrome firefox safari edge prefect version no response additional context no response
0
41,738
10,583,959,405
IssuesEvent
2019-10-08 14:37:46
zfsonlinux/zfs
https://api.github.com/repos/zfsonlinux/zfs
closed
After a week of running array, issuing zpool scrub causes system hang
Type: Defect
<!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Dell R620 | Sandy Bridge --- | --- Distribution Name | Gentoo Distribution Version | Rolling Linux Kernel | 4.15.16 Architecture | x86_64 ZFS Version | 0.7.9-r0-gentoo SPL Version | 0.7.9-r0-gentoo <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing My array, which is made up of 10x5TB disks attached to an LSI 9300 SAS controller running RAIDz2, and is currently about 18.2TB full out of ~36TB of usable storage, is seeing a system hang after zpool scrub runs. However, I can run zpool scrub on my pool after a fresh reboot, and the scrub runs to completion with no issues (and finds no problems). But if I have the scrub run out of cron once a week, as it has been running for about 2 years now, it will cause the system to become unresponsive. If I run the scrub manually after about a week of running the system, the same behavior occurs. My system is a Dell R620 running Gentoo. It has 72GB of ECC RAM. I have checked SMART data on all the disks, and run other health checks against the RAM, and nothing has indicated a hardware issue. This started occurring after upgrading ZFS to 0.7.x at some point. I honestly don't know where the cutoff happened, since I wrote off the the strange crashes as anomalies until I noticed the pattern. ### Describe how to reproduce the problem My system just has to run for about a week doing its normal workloads (a couple of VMs, serving data to my Plex server, etc.), and then kick off a zpool scrub on the pool. This can be done via cron or interactively. Either way, same issue. ### Include any warning/errors/backtraces from the system logs I'm in the process of getting a serial console hooked up to capture this. Because my cron job is currently scheduled to run at 2am on Sunday's, I forget to get this configured until it's too late. I'm hoping someone has also seen this (Google'ing around did show some similar issues, but nothing specific). <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
1.0
After a week of running array, issuing zpool scrub causes system hang - <!-- Thank you for reporting an issue. *IMPORTANT* - Please search our issue tracker *before* making a new issue. If you cannot find a similar issue, then create a new issue. https://github.com/zfsonlinux/zfs/issues *IMPORTANT* - This issue tracker is for *bugs* and *issues* only. Please search the wiki and the mailing list archives before asking questions on the mailing list. https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists Please fill in as much of the template as possible. --> ### System information <!-- add version after "|" character --> Dell R620 | Sandy Bridge --- | --- Distribution Name | Gentoo Distribution Version | Rolling Linux Kernel | 4.15.16 Architecture | x86_64 ZFS Version | 0.7.9-r0-gentoo SPL Version | 0.7.9-r0-gentoo <!-- Commands to find ZFS/SPL versions: modinfo zfs | grep -iw version modinfo spl | grep -iw version --> ### Describe the problem you're observing My array, which is made up of 10x5TB disks attached to an LSI 9300 SAS controller running RAIDz2, and is currently about 18.2TB full out of ~36TB of usable storage, is seeing a system hang after zpool scrub runs. However, I can run zpool scrub on my pool after a fresh reboot, and the scrub runs to completion with no issues (and finds no problems). But if I have the scrub run out of cron once a week, as it has been running for about 2 years now, it will cause the system to become unresponsive. If I run the scrub manually after about a week of running the system, the same behavior occurs. My system is a Dell R620 running Gentoo. It has 72GB of ECC RAM. I have checked SMART data on all the disks, and run other health checks against the RAM, and nothing has indicated a hardware issue. This started occurring after upgrading ZFS to 0.7.x at some point. I honestly don't know where the cutoff happened, since I wrote off the the strange crashes as anomalies until I noticed the pattern. ### Describe how to reproduce the problem My system just has to run for about a week doing its normal workloads (a couple of VMs, serving data to my Plex server, etc.), and then kick off a zpool scrub on the pool. This can be done via cron or interactively. Either way, same issue. ### Include any warning/errors/backtraces from the system logs I'm in the process of getting a serial console hooked up to capture this. Because my cron job is currently scheduled to run at 2am on Sunday's, I forget to get this configured until it's too late. I'm hoping someone has also seen this (Google'ing around did show some similar issues, but nothing specific). <!-- *IMPORTANT* - Please mark logs and text output from terminal commands or else Github will not display them correctly. An example is provided below. Example: ``` this is an example how log text should be marked (wrap it with ```) ``` -->
defect
after a week of running array issuing zpool scrub causes system hang thank you for reporting an issue important please search our issue tracker before making a new issue if you cannot find a similar issue then create a new issue important this issue tracker is for bugs and issues only please search the wiki and the mailing list archives before asking questions on the mailing list please fill in as much of the template as possible system information dell sandy bridge distribution name gentoo distribution version rolling linux kernel architecture zfs version gentoo spl version gentoo commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing my array which is made up of disks attached to an lsi sas controller running and is currently about full out of of usable storage is seeing a system hang after zpool scrub runs however i can run zpool scrub on my pool after a fresh reboot and the scrub runs to completion with no issues and finds no problems but if i have the scrub run out of cron once a week as it has been running for about years now it will cause the system to become unresponsive if i run the scrub manually after about a week of running the system the same behavior occurs my system is a dell running gentoo it has of ecc ram i have checked smart data on all the disks and run other health checks against the ram and nothing has indicated a hardware issue this started occurring after upgrading zfs to x at some point i honestly don t know where the cutoff happened since i wrote off the the strange crashes as anomalies until i noticed the pattern describe how to reproduce the problem my system just has to run for about a week doing its normal workloads a couple of vms serving data to my plex server etc and then kick off a zpool scrub on the pool this can be done via cron or interactively either way same issue include any warning errors backtraces from the system logs i m in the process of getting a serial console hooked up to capture this because my cron job is currently scheduled to run at on sunday s i forget to get this configured until it s too late i m hoping someone has also seen this google ing around did show some similar issues but nothing specific important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
1
19,109
3,142,923,502
IssuesEvent
2015-09-14 00:08:22
CompEvol/beast2
https://api.github.com/repos/CompEvol/beast2
closed
Automatic detection of package directories gives false positives
defect MEDIUM priority
The method `AddOnManager.getBeastDirectories()` returns a list of directories that "may contain packages". All jar files within the lib/ subdirectory of these directories are loaded by `AddOnManager.loadExternalJars()`. If a jar file contains a class that is already in the class path, that jar file will not be loaded. One of the criteria for a directory to be included in the list returned by `getBeastDirectories()` is that the directory is a subdirectory of the current directory (or whatever System.getProperty("user.dir") returns on your system) and that it contains a "/lib" or "/templates" directory. I contend that this is not a strict enough criterion, as /lib in particular is a very common directory name on *nix systems. As the "user.dir" subdirectories are included before the official package directories (eg $HOME/.beast/2.1/*), jars detected here may prevent the installed package jars containing at least one identical class from being loaded. This could cause beast to behave in surprising ways if, for instance, an installed package used a different version of a library than the one picked up from a directory such as "user.dir"/BLAH/lib. An alternative might be to look for (and ideally parse) "version.xml". The related question of how to handle real packages using different versions of the same library would still exist though.
1.0
Automatic detection of package directories gives false positives - The method `AddOnManager.getBeastDirectories()` returns a list of directories that "may contain packages". All jar files within the lib/ subdirectory of these directories are loaded by `AddOnManager.loadExternalJars()`. If a jar file contains a class that is already in the class path, that jar file will not be loaded. One of the criteria for a directory to be included in the list returned by `getBeastDirectories()` is that the directory is a subdirectory of the current directory (or whatever System.getProperty("user.dir") returns on your system) and that it contains a "/lib" or "/templates" directory. I contend that this is not a strict enough criterion, as /lib in particular is a very common directory name on *nix systems. As the "user.dir" subdirectories are included before the official package directories (eg $HOME/.beast/2.1/*), jars detected here may prevent the installed package jars containing at least one identical class from being loaded. This could cause beast to behave in surprising ways if, for instance, an installed package used a different version of a library than the one picked up from a directory such as "user.dir"/BLAH/lib. An alternative might be to look for (and ideally parse) "version.xml". The related question of how to handle real packages using different versions of the same library would still exist though.
defect
automatic detection of package directories gives false positives the method addonmanager getbeastdirectories returns a list of directories that may contain packages all jar files within the lib subdirectory of these directories are loaded by addonmanager loadexternaljars if a jar file contains a class that is already in the class path that jar file will not be loaded one of the criteria for a directory to be included in the list returned by getbeastdirectories is that the directory is a subdirectory of the current directory or whatever system getproperty user dir returns on your system and that it contains a lib or templates directory i contend that this is not a strict enough criterion as lib in particular is a very common directory name on nix systems as the user dir subdirectories are included before the official package directories eg home beast jars detected here may prevent the installed package jars containing at least one identical class from being loaded this could cause beast to behave in surprising ways if for instance an installed package used a different version of a library than the one picked up from a directory such as user dir blah lib an alternative might be to look for and ideally parse version xml the related question of how to handle real packages using different versions of the same library would still exist though
1
74,152
24,978,172,181
IssuesEvent
2022-11-02 09:35:39
matrix-org/matrix-public-archive
https://api.github.com/repos/matrix-org/matrix-public-archive
closed
Don't change the URL as you scroll around
T-Defect A-archive-room-view
Even if we had https://github.com/matrix-org/matrix-public-archive/issues/72 to page-load the room with a stable URL, it still probably doesn't make sense to change the URL around as you scroll. The content on the page should be consistent with the URL. > ## UX problems > > ### One person's link will not look like the other once they start scrolling > > Because we change the URL as you scroll around. Someone may link a page expecting all the same messages to be there but it will actually only be the current day they're scrolled to plus whatever is above it. > > *-- https://github.com/matrix-org/matrix-public-archive/pull/71* It will probably also affect how search engines index the site attributing messages with a certain URL that when visiting will be missing. --- I think it does make sense to still change the calendar and text around to make it easy to jump to another day from where you're currently viewing in the surrounding context.
1.0
Don't change the URL as you scroll around - Even if we had https://github.com/matrix-org/matrix-public-archive/issues/72 to page-load the room with a stable URL, it still probably doesn't make sense to change the URL around as you scroll. The content on the page should be consistent with the URL. > ## UX problems > > ### One person's link will not look like the other once they start scrolling > > Because we change the URL as you scroll around. Someone may link a page expecting all the same messages to be there but it will actually only be the current day they're scrolled to plus whatever is above it. > > *-- https://github.com/matrix-org/matrix-public-archive/pull/71* It will probably also affect how search engines index the site attributing messages with a certain URL that when visiting will be missing. --- I think it does make sense to still change the calendar and text around to make it easy to jump to another day from where you're currently viewing in the surrounding context.
defect
don t change the url as you scroll around even if we had to page load the room with a stable url it still probably doesn t make sense to change the url around as you scroll the content on the page should be consistent with the url ux problems one person s link will not look like the other once they start scrolling because we change the url as you scroll around someone may link a page expecting all the same messages to be there but it will actually only be the current day they re scrolled to plus whatever is above it it will probably also affect how search engines index the site attributing messages with a certain url that when visiting will be missing i think it does make sense to still change the calendar and text around to make it easy to jump to another day from where you re currently viewing in the surrounding context
1
27,267
4,956,984,106
IssuesEvent
2016-12-02 01:53:33
cakephp/cakephp
https://api.github.com/repos/cakephp/cakephp
closed
ExistsIn Rule (since 3.3) causing SQL generation issue
Defect ORM validation
- [x] bug - [ ] enhancement - [ ] feature-discussion (RFC) - CakePHP Version: 3.3.0 - Platform and Target: Apache 2.4.18,PHP 5.6,MYSQL 5.5.51 ### What you did table1 has 'hasOne' relationship with table2 ``` php $this->hasOne('table2', [ 'className' => 'table2', 'foreignKey' => false, 'joinType' => 'inner', 'conditions' => ['table1.target_id = table2.id'], ]); ``` ``` php $rules->add($rules->existsIn(['target_id'], 'table2')); ``` ### Expected Behavior cake 3.2.14 before when inserting a value in table1 cake was checking in table2 if target_id existing without including the condition of the relation ``` sql SELECT 1 AS `existing` FROM `table2` `table2` WHERE `table2`.`id` = 1 LIMIT 1 ``` ### Actual Behavior cake 3.3.0 now cake add the relation to the query and broke the query because in this query table1 is unknown ``` sql SELECT 1 AS `existing` FROM `table2` `table2` WHERE (`table2`.`id` = 1 AND table1.target_id = table2.id) LIMIT 1 ```
1.0
ExistsIn Rule (since 3.3) causing SQL generation issue - - [x] bug - [ ] enhancement - [ ] feature-discussion (RFC) - CakePHP Version: 3.3.0 - Platform and Target: Apache 2.4.18,PHP 5.6,MYSQL 5.5.51 ### What you did table1 has 'hasOne' relationship with table2 ``` php $this->hasOne('table2', [ 'className' => 'table2', 'foreignKey' => false, 'joinType' => 'inner', 'conditions' => ['table1.target_id = table2.id'], ]); ``` ``` php $rules->add($rules->existsIn(['target_id'], 'table2')); ``` ### Expected Behavior cake 3.2.14 before when inserting a value in table1 cake was checking in table2 if target_id existing without including the condition of the relation ``` sql SELECT 1 AS `existing` FROM `table2` `table2` WHERE `table2`.`id` = 1 LIMIT 1 ``` ### Actual Behavior cake 3.3.0 now cake add the relation to the query and broke the query because in this query table1 is unknown ``` sql SELECT 1 AS `existing` FROM `table2` `table2` WHERE (`table2`.`id` = 1 AND table1.target_id = table2.id) LIMIT 1 ```
defect
existsin rule since causing sql generation issue bug enhancement feature discussion rfc cakephp version platform and target apache php mysql what you did has hasone relationship with php this hasone classname foreignkey false jointype inner conditions php rules add rules existsin expected behavior cake before when inserting a value in cake was checking in if target id existing without including the condition of the relation sql select as existing from where id limit actual behavior cake now cake add the relation to the query and broke the query because in this query is unknown sql select as existing from where id and target id id limit
1
252,565
27,245,824,211
IssuesEvent
2023-02-22 01:51:41
idmarinas/lotgd-modules
https://api.github.com/repos/idmarinas/lotgd-modules
closed
CVE-2021-35065 (High) detected in glob-parent-5.1.2.tgz - autoclosed
security vulnerability
## CVE-2021-35065 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - del-6.0.0.tgz (Root Library) - globby-11.0.4.tgz - fast-glob-3.2.7.tgz - :x: **glob-parent-5.1.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package glob-parent from 6.0.0 and before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-35065>CVE-2021-35065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution: glob-parent - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-35065 (High) detected in glob-parent-5.1.2.tgz - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p></summary> <p>Extract the non-magic parent path from a glob string.</p> <p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/glob-parent/package.json</p> <p> Dependency Hierarchy: - del-6.0.0.tgz (Root Library) - globby-11.0.4.tgz - fast-glob-3.2.7.tgz - :x: **glob-parent-5.1.2.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package glob-parent from 6.0.0 and before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS) <p>Publish Date: 2021-06-22 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-35065>CVE-2021-35065</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p> <p>Release Date: 2021-06-22</p> <p>Fix Resolution: glob-parent - 6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_defect
cve high detected in glob parent tgz autoclosed cve high severity vulnerability vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy del tgz root library globby tgz fast glob tgz x glob parent tgz vulnerable library found in base branch master vulnerability details the package glob parent from and before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with mend
0
6,516
14,707,438,935
IssuesEvent
2021-01-04 21:37:19
MicrosoftDocs/architecture-center
https://api.github.com/repos/MicrosoftDocs/architecture-center
closed
Change title to Reference Architecture or Create a new article for it!
architecture-center/svc cxp doc-idea example-scenario/subsvc triaged
A number of my customers are looking for reference architecture and this article is the closest to it (https://docs.microsoft.com/en-us/azure/architecture/example-scenario/wvd/windows-virtual-desktop ) Can the title of this article be changed to reference arch. or have a new article written for it? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: f26be54d-0f67-5098-7839-98f5bb064188 * Version Independent ID: 187fa3ce-ade9-1b07-2af3-c874ec3d0aac * Content: [Windows Virtual Desktop for the enterprise - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/wvd/windows-virtual-desktop) * Content Source: [docs/example-scenario/wvd/windows-virtual-desktop.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/wvd/windows-virtual-desktop.yml) * Service: **architecture-center** * Sub-service: **example-scenario** * GitHub Login: @doodlemania2 * Microsoft Alias: **pnp**
1.0
Change title to Reference Architecture or Create a new article for it! - A number of my customers are looking for reference architecture and this article is the closest to it (https://docs.microsoft.com/en-us/azure/architecture/example-scenario/wvd/windows-virtual-desktop ) Can the title of this article be changed to reference arch. or have a new article written for it? --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: f26be54d-0f67-5098-7839-98f5bb064188 * Version Independent ID: 187fa3ce-ade9-1b07-2af3-c874ec3d0aac * Content: [Windows Virtual Desktop for the enterprise - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/wvd/windows-virtual-desktop) * Content Source: [docs/example-scenario/wvd/windows-virtual-desktop.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/wvd/windows-virtual-desktop.yml) * Service: **architecture-center** * Sub-service: **example-scenario** * GitHub Login: @doodlemania2 * Microsoft Alias: **pnp**
non_defect
change title to reference architecture or create a new article for it a number of my customers are looking for reference architecture and this article is the closest to it can the title of this article be changed to reference arch or have a new article written for it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service example scenario github login microsoft alias pnp
0
79,371
28,136,764,348
IssuesEvent
2023-04-01 13:21:56
dkfans/keeperfx
https://api.github.com/repos/dkfans/keeperfx
closed
Heap crash in log even though heap is gone
Type-Defect Priority-Low
Start any map and see this in the log: `=== Crash ===Sync: GraphicsHeap Size 50331648` Most likely caused by 8443f4094f9accc6725ce39e9a6b9062472c5ceb
1.0
Heap crash in log even though heap is gone - Start any map and see this in the log: `=== Crash ===Sync: GraphicsHeap Size 50331648` Most likely caused by 8443f4094f9accc6725ce39e9a6b9062472c5ceb
defect
heap crash in log even though heap is gone start any map and see this in the log crash sync graphicsheap size most likely caused by
1
307,187
23,188,305,257
IssuesEvent
2022-08-01 10:19:29
aitos-io/BoAT-X-Framework
https://api.github.com/repos/aitos-io/BoAT-X-Framework
closed
Some functions of utility lack annotation descriptions
documentation good first issue
As the title says: <img width="626" alt="image" src="https://user-images.githubusercontent.com/44628575/178891105-ce13857f-c26b-44e7-8609-abdda89474d0.png">
1.0
Some functions of utility lack annotation descriptions - As the title says: <img width="626" alt="image" src="https://user-images.githubusercontent.com/44628575/178891105-ce13857f-c26b-44e7-8609-abdda89474d0.png">
non_defect
some functions of utility lack annotation descriptions as the title says: img width alt image src
0