Unnamed: 0
int64 0
832k
| id
float64 2.49B
32.1B
| type
stringclasses 1
value | created_at
stringlengths 19
19
| repo
stringlengths 5
112
| repo_url
stringlengths 34
141
| action
stringclasses 3
values | title
stringlengths 1
757
| labels
stringlengths 4
664
| body
stringlengths 3
261k
| index
stringclasses 10
values | text_combine
stringlengths 96
261k
| label
stringclasses 2
values | text
stringlengths 96
232k
| binary_label
int64 0
1
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
734,637
| 25,356,805,879
|
IssuesEvent
|
2022-11-20 12:44:08
|
YunoHost/issues
|
https://api.github.com/repos/YunoHost/issues
|
closed
|
[idea] instead of remove domain, archive them instead
|
:star2: ux Priority: low
|
That would allow to "undo" mistakes and keep certificates in case of.
Not sure it's really worth it.
|
1.0
|
[idea] instead of remove domain, archive them instead - That would allow to "undo" mistakes and keep certificates in case of.
Not sure it's really worth it.
|
non_defect
|
instead of remove domain archive them instead that would allow to undo mistakes and keep certificates in case of not sure it s really worth it
| 0
|
79,784
| 29,048,982,505
|
IssuesEvent
|
2023-05-13 23:58:26
|
BOINC/boinc
|
https://api.github.com/repos/BOINC/boinc
|
closed
|
BOINC crashing
|
C: Manager P: Undetermined R: worksforme T: Defect E: to be determined Validate
|
Hello,
so I run BOINC on Windows 10 and it sometimes throws errors or crashes.
So my system is running 24/7 and it sometimes crashes and basically wastes the time where it could do useful work.
Can you guys work a little on BOINC, improve UI, make it more user friendly and make not crash?
|
1.0
|
BOINC crashing - Hello,
so I run BOINC on Windows 10 and it sometimes throws errors or crashes.
So my system is running 24/7 and it sometimes crashes and basically wastes the time where it could do useful work.
Can you guys work a little on BOINC, improve UI, make it more user friendly and make not crash?
|
defect
|
boinc crashing hello so i run boinc on windows and it sometimes throws errors or crashes so my system is running and it sometimes crashes and basically wastes the time where it could do useful work can you guys work a little on boinc improve ui make it more user friendly and make not crash
| 1
|
58,047
| 16,342,337,751
|
IssuesEvent
|
2021-05-13 00:04:35
|
darshan-hpc/darshan
|
https://api.github.com/repos/darshan-hpc/darshan
|
closed
|
concurrent I/O from threads gets counted twice in timing
|
defect wrapper libraries
|
In GitLab by @shanedsnyder on Sep 24, 2015, 16:25
If two threads (in the same MPI process) access the same file concurrently, then the cumulative time counters are incremented too far.
We need to add a reference count to the run-time data structure to tell how many threads are accessing the same file at once. The time should not be incremented until the reference counter hits zero.
This does not require a log format change.
|
1.0
|
concurrent I/O from threads gets counted twice in timing - In GitLab by @shanedsnyder on Sep 24, 2015, 16:25
If two threads (in the same MPI process) access the same file concurrently, then the cumulative time counters are incremented too far.
We need to add a reference count to the run-time data structure to tell how many threads are accessing the same file at once. The time should not be incremented until the reference counter hits zero.
This does not require a log format change.
|
defect
|
concurrent i o from threads gets counted twice in timing in gitlab by shanedsnyder on sep if two threads in the same mpi process access the same file concurrently then the cumulative time counters are incremented too far we need to add a reference count to the run time data structure to tell how many threads are accessing the same file at once the time should not be incremented until the reference counter hits zero this does not require a log format change
| 1
|
67,250
| 16,854,265,692
|
IssuesEvent
|
2021-06-21 02:53:40
|
rust-lang/docs.rs
|
https://api.github.com/repos/rust-lang/docs.rs
|
closed
|
Read-only filesystem error using cc during build
|
C-build-failure S-waiting-on-crate-author
|
**Crate name:** spirv_cross or crates depending on it (e.g. gfx-backend-dx12)
**Build failure link:** https://docs.rs/crate/gfx-backend-dx12/0.5.9/builds/281262
**Additional details:** When running cc from spirv_cross we get the following message:
> [INFO] [stderr] running: "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-g" "-fno-omit-frame-pointer" "-m64" "-Wall" "-Wextra" "-DSPIRV_CROSS_EXCEPTIONS_TO_ASSERTIONS" "-DSPIRV_CROSS_WRAPPER_NO_EXCEPTIONS" "-DSPIRV_CROSS_WRAPPER_GLSL" "-DSPIRV_CROSS_WRAPPER_HLSL" "-Fo/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o" "-c" "src/wrapper.cpp"
[INFO] [stderr] cargo:warning=src/wrapper.cpp:2: warning: ignoring #pragma warning [-Wunknown-pragmas]
[INFO] [stderr] cargo:warning= 2 | #pragma warning(disable : 4996 4101)
[INFO] [stderr] cargo:warning= |
[INFO] [stderr] cargo:warning=Assembler messages:
[INFO] [stderr] cargo:warning=Fatal error: can't create wrapper.o: Read-only file system
[INFO] [stderr] exit code: 1
[INFO] [stderr]
[INFO] [stderr] --- stderr
[INFO] [stderr]
[INFO] [stderr]
[INFO] [stderr] error occurred: Command "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-g" "-fno-omit-frame-pointer" "-m64" "-Wall" "-Wextra" "-DSPIRV_CROSS_EXCEPTIONS_TO_ASSERTIONS" "-DSPIRV_CROSS_WRAPPER_NO_EXCEPTIONS" "-DSPIRV_CROSS_WRAPPER_GLSL" "-DSPIRV_CROSS_WRAPPER_HLSL" "-Fo/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o" "-c" "src/wrapper.cpp" with args "c++" did not execute successfully (status code exit code: 1).
[INFO] [stderr]
[INFO] [stderr]
[INFO] [stderr] warning: build failed, waiting for other jobs to finish...
[INFO] [stderr] error: build failed
I'm not sure how to resolve this so I was wondering if I could get some help. I thought the path (`/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o`) would be writable because it's inside of `out`.
The spirv_cross build script is a pretty typical cc setup and should use `OUT_DIR` from the environment: https://github.com/grovesNL/spirv_cross/blob/master/spirv_cross/build.rs
|
1.0
|
Read-only filesystem error using cc during build - **Crate name:** spirv_cross or crates depending on it (e.g. gfx-backend-dx12)
**Build failure link:** https://docs.rs/crate/gfx-backend-dx12/0.5.9/builds/281262
**Additional details:** When running cc from spirv_cross we get the following message:
> [INFO] [stderr] running: "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-g" "-fno-omit-frame-pointer" "-m64" "-Wall" "-Wextra" "-DSPIRV_CROSS_EXCEPTIONS_TO_ASSERTIONS" "-DSPIRV_CROSS_WRAPPER_NO_EXCEPTIONS" "-DSPIRV_CROSS_WRAPPER_GLSL" "-DSPIRV_CROSS_WRAPPER_HLSL" "-Fo/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o" "-c" "src/wrapper.cpp"
[INFO] [stderr] cargo:warning=src/wrapper.cpp:2: warning: ignoring #pragma warning [-Wunknown-pragmas]
[INFO] [stderr] cargo:warning= 2 | #pragma warning(disable : 4996 4101)
[INFO] [stderr] cargo:warning= |
[INFO] [stderr] cargo:warning=Assembler messages:
[INFO] [stderr] cargo:warning=Fatal error: can't create wrapper.o: Read-only file system
[INFO] [stderr] exit code: 1
[INFO] [stderr]
[INFO] [stderr] --- stderr
[INFO] [stderr]
[INFO] [stderr]
[INFO] [stderr] error occurred: Command "c++" "-O0" "-ffunction-sections" "-fdata-sections" "-g" "-fno-omit-frame-pointer" "-m64" "-Wall" "-Wextra" "-DSPIRV_CROSS_EXCEPTIONS_TO_ASSERTIONS" "-DSPIRV_CROSS_WRAPPER_NO_EXCEPTIONS" "-DSPIRV_CROSS_WRAPPER_GLSL" "-DSPIRV_CROSS_WRAPPER_HLSL" "-Fo/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o" "-c" "src/wrapper.cpp" with args "c++" did not execute successfully (status code exit code: 1).
[INFO] [stderr]
[INFO] [stderr]
[INFO] [stderr] warning: build failed, waiting for other jobs to finish...
[INFO] [stderr] error: build failed
I'm not sure how to resolve this so I was wondering if I could get some help. I thought the path (`/opt/rustwide/target/x86_64-pc-windows-msvc/debug/build/spirv_cross-f7ca93fa2885997b/out/src/wrapper.o`) would be writable because it's inside of `out`.
The spirv_cross build script is a pretty typical cc setup and should use `OUT_DIR` from the environment: https://github.com/grovesNL/spirv_cross/blob/master/spirv_cross/build.rs
|
non_defect
|
read only filesystem error using cc during build crate name spirv cross or crates depending on it e g gfx backend build failure link additional details when running cc from spirv cross we get the following message running c ffunction sections fdata sections g fno omit frame pointer wall wextra dspirv cross exceptions to assertions dspirv cross wrapper no exceptions dspirv cross wrapper glsl dspirv cross wrapper hlsl fo opt rustwide target pc windows msvc debug build spirv cross out src wrapper o c src wrapper cpp cargo warning src wrapper cpp warning ignoring pragma warning cargo warning pragma warning disable cargo warning cargo warning assembler messages cargo warning fatal error can t create wrapper o read only file system exit code stderr error occurred command c ffunction sections fdata sections g fno omit frame pointer wall wextra dspirv cross exceptions to assertions dspirv cross wrapper no exceptions dspirv cross wrapper glsl dspirv cross wrapper hlsl fo opt rustwide target pc windows msvc debug build spirv cross out src wrapper o c src wrapper cpp with args c did not execute successfully status code exit code warning build failed waiting for other jobs to finish error build failed i m not sure how to resolve this so i was wondering if i could get some help i thought the path opt rustwide target pc windows msvc debug build spirv cross out src wrapper o would be writable because it s inside of out the spirv cross build script is a pretty typical cc setup and should use out dir from the environment
| 0
|
79,256
| 15,586,118,757
|
IssuesEvent
|
2021-03-18 01:12:55
|
wasimakh2/JPAGenratorRelease
|
https://api.github.com/repos/wasimakh2/JPAGenratorRelease
|
closed
|
CVE-2019-12814 (Medium) detected in jackson-databind-2.9.9.jar - autoclosed
|
security vulnerability
|
## CVE-2019-12814 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: _depth_0/JPAGenratorRelease/JPAGenrator-0.0.1-SNAPSHOT/WEB-INF/lib/jackson-databind-2.9.9.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/wasimakh2/JPAGenratorRelease/commit/2d85088b97dabdacb7e6cceaea3a329a2eb43f87">2d85088b97dabdacb7e6cceaea3a329a2eb43f87</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x through 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has JDOM 1.x or 2.x jar in the classpath, an attacker can send a specifically crafted JSON message that allows them to read arbitrary local files on the server.
<p>Publish Date: 2019-06-19
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12814>CVE-2019-12814</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2341">https://github.com/FasterXML/jackson-databind/issues/2341</a></p>
<p>Release Date: 2019-06-19</p>
<p>Fix Resolution: 2.7.9.6, 2.8.11.4, 2.9.9.1, 2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2019-12814 (Medium) detected in jackson-databind-2.9.9.jar - autoclosed - ## CVE-2019-12814 - Medium Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jackson-databind-2.9.9.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to vulnerable library: _depth_0/JPAGenratorRelease/JPAGenrator-0.0.1-SNAPSHOT/WEB-INF/lib/jackson-databind-2.9.9.jar,/root/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.9/jackson-databind-2.9.9.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.9.jar** (Vulnerable Library)
<p>Found in HEAD commit: <a href="https://github.com/wasimakh2/JPAGenratorRelease/commit/2d85088b97dabdacb7e6cceaea3a329a2eb43f87">2d85088b97dabdacb7e6cceaea3a329a2eb43f87</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A Polymorphic Typing issue was discovered in FasterXML jackson-databind 2.x through 2.9.9. When Default Typing is enabled (either globally or for a specific property) for an externally exposed JSON endpoint and the service has JDOM 1.x or 2.x jar in the classpath, an attacker can send a specifically crafted JSON message that allows them to read arbitrary local files on the server.
<p>Publish Date: 2019-06-19
<p>URL: <a href=https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-12814>CVE-2019-12814</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.9</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: None
- Availability Impact: None
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/FasterXML/jackson-databind/issues/2341">https://github.com/FasterXML/jackson-databind/issues/2341</a></p>
<p>Release Date: 2019-06-19</p>
<p>Fix Resolution: 2.7.9.6, 2.8.11.4, 2.9.9.1, 2.10.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve medium detected in jackson databind jar autoclosed cve medium severity vulnerability vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to vulnerable library depth jpagenratorrelease jpagenrator snapshot web inf lib jackson databind jar root repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href vulnerability details a polymorphic typing issue was discovered in fasterxml jackson databind x through when default typing is enabled either globally or for a specific property for an externally exposed json endpoint and the service has jdom x or x jar in the classpath an attacker can send a specifically crafted json message that allows them to read arbitrary local files on the server publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
| 0
|
77,726
| 27,133,402,576
|
IssuesEvent
|
2023-02-16 11:25:31
|
vector-im/element-x-ios
|
https://api.github.com/repos/vector-im/element-x-ios
|
closed
|
Left Panel room previews are often stale
|
A-Room-List T-Defect TF:1
|
Running EIX on macOS \(as an iOS app on M1 macOS\), the room previews often show messages which are 3-4 events behind the actual most recent message in the room in question.
|
1.0
|
Left Panel room previews are often stale - Running EIX on macOS \(as an iOS app on M1 macOS\), the room previews often show messages which are 3-4 events behind the actual most recent message in the room in question.
|
defect
|
left panel room previews are often stale running eix on macos as an ios app on macos the room previews often show messages which are events behind the actual most recent message in the room in question
| 1
|
2,858
| 2,607,963,308
|
IssuesEvent
|
2015-02-26 00:41:12
|
chrsmithdemos/leveldb
|
https://api.github.com/repos/chrsmithdemos/leveldb
|
closed
|
Add OpenBSD support
|
auto-migrated Priority-Medium Type-Defect
|
```
Attached is a patch to allow leveldb to build (and pass regress tests) on
OpenBSD. Tested succesfully on OpenBSD/amd64, 5.0.
```
-----
Original issue reported on code.google.com by `jasper.l...@gmail.com` on 18 Aug 2011 at 3:01
* Merged into: #59
Attachments:
* [leveldb_openbsd.patch](https://storage.googleapis.com/google-code-attachments/leveldb/issue-31/comment-0/leveldb_openbsd.patch)
|
1.0
|
Add OpenBSD support - ```
Attached is a patch to allow leveldb to build (and pass regress tests) on
OpenBSD. Tested succesfully on OpenBSD/amd64, 5.0.
```
-----
Original issue reported on code.google.com by `jasper.l...@gmail.com` on 18 Aug 2011 at 3:01
* Merged into: #59
Attachments:
* [leveldb_openbsd.patch](https://storage.googleapis.com/google-code-attachments/leveldb/issue-31/comment-0/leveldb_openbsd.patch)
|
defect
|
add openbsd support attached is a patch to allow leveldb to build and pass regress tests on openbsd tested succesfully on openbsd original issue reported on code google com by jasper l gmail com on aug at merged into attachments
| 1
|
210,570
| 23,754,872,453
|
IssuesEvent
|
2022-09-01 01:23:59
|
LingalaShalini/sqlite-version-3.21.0_after_fix
|
https://api.github.com/repos/LingalaShalini/sqlite-version-3.21.0_after_fix
|
opened
|
CVE-2021-20223 (High) detected in sqliteversion-3.21.0
|
security vulnerability
|
## CVE-2021-20223 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sqliteversion-3.21.0</b></p></summary>
<p>
<p>Official Git mirror of the SQLite source tree</p>
<p>Library home page: <a href=https://github.com/sqlite/sqlite.git>https://github.com/sqlite/sqlite.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/LingalaShalini/sqlite-version-3.21.0_after_fix/commit/c54261c167d3ed3bb3cdbcfc62a5337ec33bc1bd">c54261c167d3ed3bb3cdbcfc62a5337ec33bc1bd</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/ext/fts5/fts5_unicode2.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was found in fts5UnicodeTokenize() in ext/fts5/fts5_tokenize.c in Sqlite. A unicode61 tokenizer configured to treat unicode "control-characters" (class Cc), was treating embedded nul characters as tokens. The issue was fixed in sqlite-3.34.0 and later.
<p>Publish Date: 2022-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20223>CVE-2021-20223</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.sqlite.org/forum/forumpost/09609d7e22">https://www.sqlite.org/forum/forumpost/09609d7e22</a></p>
<p>Release Date: 2022-08-25</p>
<p>Fix Resolution: version-3.34.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-20223 (High) detected in sqliteversion-3.21.0 - ## CVE-2021-20223 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>sqliteversion-3.21.0</b></p></summary>
<p>
<p>Official Git mirror of the SQLite source tree</p>
<p>Library home page: <a href=https://github.com/sqlite/sqlite.git>https://github.com/sqlite/sqlite.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/LingalaShalini/sqlite-version-3.21.0_after_fix/commit/c54261c167d3ed3bb3cdbcfc62a5337ec33bc1bd">c54261c167d3ed3bb3cdbcfc62a5337ec33bc1bd</a></p>
<p>Found in base branch: <b>master</b></p></p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary>
<p></p>
<p>
<img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/ext/fts5/fts5_unicode2.c</b>
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
An issue was found in fts5UnicodeTokenize() in ext/fts5/fts5_tokenize.c in Sqlite. A unicode61 tokenizer configured to treat unicode "control-characters" (class Cc), was treating embedded nul characters as tokens. The issue was fixed in sqlite-3.34.0 and later.
<p>Publish Date: 2022-08-25
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-20223>CVE-2021-20223</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://www.sqlite.org/forum/forumpost/09609d7e22">https://www.sqlite.org/forum/forumpost/09609d7e22</a></p>
<p>Release Date: 2022-08-25</p>
<p>Fix Resolution: version-3.34.0</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in sqliteversion cve high severity vulnerability vulnerable library sqliteversion official git mirror of the sqlite source tree library home page a href found in head commit a href found in base branch master vulnerable source files ext c vulnerability details an issue was found in in ext tokenize c in sqlite a tokenizer configured to treat unicode control characters class cc was treating embedded nul characters as tokens the issue was fixed in sqlite and later publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution version step up your open source security game with mend
| 0
|
20,072
| 3,295,088,182
|
IssuesEvent
|
2015-10-31 16:45:09
|
FreeRADIUS/freeradius-server
|
https://api.github.com/repos/FreeRADIUS/freeradius-server
|
closed
|
Fatal Python error: PyThreadState_Get: no current thread, cause by _locale.so
|
upstream defect v3.0.x v3.1.x
|
I found module rlm_python will crash when instantiating because of the _locale.so in python.
```bash
$ ./radiusd -X -f
Copyright (C) 1999-2015 The FreeRADIUS server project and contributors
There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A
PARTICULAR PURPOSE
You may redistribute copies of FreeRADIUS under the terms of the
GNU General Public License
For more information about these matters, see the file named COPYRIGHT
Starting - reading configuration files ...
....
# Instantiating module "attr_filter.accounting_response" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/attr_filter
reading pairlist file /home/develop/app/freeraiuds-server/etc/raddb/mods-config/attr_filter/accounting_response
# Instantiating module "mschap" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/mschap
rlm_mschap (mschap): using internal authentication
# Instantiating module "python" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/python
Fatal Python error: PyThreadState_Get: no current thread
core dumped)
```
but if i rename the _locale.so:
```bash
$ ./radiusd -X -f
Copyright (C) 1999-2015 The FreeRADIUS server project and contributors
There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A
PARTICULAR PURPOSE
You may redistribute copies of FreeRADIUS under the terms of the
GNU General Public License
For more information about these matters, see the file named COPYRIGHT
Starting - reading configuration files ...
....
cd $PYTHONHOME/lib/python2.7/lib-dynload
mv _locale.so _locale.so.bak
```
freeradius-server will run.
```bash
}
}
listen {
type = "acct"
ipv6addr = ::
port = 0
limit {
max_connections = 16
lifetime = 0
idle_timeout = 30
}
}
Listening on auth address 127.0.0.1 port 18120 bound to server inner-tunnel
Listening on auth address * port 1812 bound to server default
Listening on acct address * port 1813 bound to server default
Listening on auth address :: port 1812 bound to server default
Listening on acct address :: port 1813 bound to server default
Opening new proxy socket 'proxy address * port 0'
Listening on proxy address * port 34786
Ready to process requests
```
it's any idea for this ???
|
1.0
|
Fatal Python error: PyThreadState_Get: no current thread, cause by _locale.so - I found module rlm_python will crash when instantiating because of the _locale.so in python.
```bash
$ ./radiusd -X -f
Copyright (C) 1999-2015 The FreeRADIUS server project and contributors
There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A
PARTICULAR PURPOSE
You may redistribute copies of FreeRADIUS under the terms of the
GNU General Public License
For more information about these matters, see the file named COPYRIGHT
Starting - reading configuration files ...
....
# Instantiating module "attr_filter.accounting_response" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/attr_filter
reading pairlist file /home/develop/app/freeraiuds-server/etc/raddb/mods-config/attr_filter/accounting_response
# Instantiating module "mschap" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/mschap
rlm_mschap (mschap): using internal authentication
# Instantiating module "python" from file /home/develop/app/freeraiuds-server/etc/raddb/mods-enabled/python
Fatal Python error: PyThreadState_Get: no current thread
core dumped)
```
but if i rename the _locale.so:
```bash
$ ./radiusd -X -f
Copyright (C) 1999-2015 The FreeRADIUS server project and contributors
There is NO warranty; not even for MERCHANTABILITY or FITNESS FOR A
PARTICULAR PURPOSE
You may redistribute copies of FreeRADIUS under the terms of the
GNU General Public License
For more information about these matters, see the file named COPYRIGHT
Starting - reading configuration files ...
....
cd $PYTHONHOME/lib/python2.7/lib-dynload
mv _locale.so _locale.so.bak
```
freeradius-server will run.
```bash
}
}
listen {
type = "acct"
ipv6addr = ::
port = 0
limit {
max_connections = 16
lifetime = 0
idle_timeout = 30
}
}
Listening on auth address 127.0.0.1 port 18120 bound to server inner-tunnel
Listening on auth address * port 1812 bound to server default
Listening on acct address * port 1813 bound to server default
Listening on auth address :: port 1812 bound to server default
Listening on acct address :: port 1813 bound to server default
Opening new proxy socket 'proxy address * port 0'
Listening on proxy address * port 34786
Ready to process requests
```
it's any idea for this ???
|
defect
|
fatal python error pythreadstate get no current thread cause by locale so i found module rlm python will crash when instantiating because of the locale so in python bash radiusd x f copyright c the freeradius server project and contributors there is no warranty not even for merchantability or fitness for a particular purpose you may redistribute copies of freeradius under the terms of the gnu general public license for more information about these matters see the file named copyright starting reading configuration files instantiating module attr filter accounting response from file home develop app freeraiuds server etc raddb mods enabled attr filter reading pairlist file home develop app freeraiuds server etc raddb mods config attr filter accounting response instantiating module mschap from file home develop app freeraiuds server etc raddb mods enabled mschap rlm mschap mschap using internal authentication instantiating module python from file home develop app freeraiuds server etc raddb mods enabled python fatal python error pythreadstate get no current thread core dumped but if i rename the locale so bash radiusd x f copyright c the freeradius server project and contributors there is no warranty not even for merchantability or fitness for a particular purpose you may redistribute copies of freeradius under the terms of the gnu general public license for more information about these matters see the file named copyright starting reading configuration files cd pythonhome lib lib dynload mv locale so locale so bak freeradius server will run bash listen type acct port limit max connections lifetime idle timeout listening on auth address port bound to server inner tunnel listening on auth address port bound to server default listening on acct address port bound to server default listening on auth address port bound to server default listening on acct address port bound to server default opening new proxy socket proxy address port listening on proxy address port ready to process requests it s any idea for this
| 1
|
85,545
| 7,976,421,370
|
IssuesEvent
|
2018-07-17 12:39:20
|
loot/skyrimse
|
https://api.github.com/repos/loot/skyrimse
|
closed
|
Wildcat and Ultimate Combat - Compatibility
|
Testing awaiting-feedback
|
Please remove UltimateCombat.esp from the "Load after" rules for Wildcat (added in 55e23755e2149c12dd76b4408ff0e0c2cb7b07aa). First of all, the Wildcat and Ultimate Combat pages give conflicting advice, and second, if using them both one needs to control their order manually based on patches (e.g. OMEGA MLU says to load Ultimate Combat after Wildcat, which is currently impossible due to the masterlist rule).
|
1.0
|
Wildcat and Ultimate Combat - Compatibility - Please remove UltimateCombat.esp from the "Load after" rules for Wildcat (added in 55e23755e2149c12dd76b4408ff0e0c2cb7b07aa). First of all, the Wildcat and Ultimate Combat pages give conflicting advice, and second, if using them both one needs to control their order manually based on patches (e.g. OMEGA MLU says to load Ultimate Combat after Wildcat, which is currently impossible due to the masterlist rule).
|
non_defect
|
wildcat and ultimate combat compatibility please remove ultimatecombat esp from the load after rules for wildcat added in first of all the wildcat and ultimate combat pages give conflicting advice and second if using them both one needs to control their order manually based on patches e g omega mlu says to load ultimate combat after wildcat which is currently impossible due to the masterlist rule
| 0
|
231,670
| 7,642,044,673
|
IssuesEvent
|
2018-05-08 07:54:53
|
aiidateam/aiida_core
|
https://api.github.com/repos/aiidateam/aiida_core
|
opened
|
verdi calculation show should check whether the id references a calculation node
|
priority/nice to have topic/VerdiCommandLine type/bug
|
Currently it does not and will except on some of the properties:
```
Traceback (most recent call last):
File "/home/aiida/.virtualenvs/aiida/bin/verdi", line 9, in <module>
sys.exit(run())
File "/home/aiida/codes/aiida_core/aiida/cmdline/verdilib.py", line 1056, in run
aiida.cmdline.verdilib.exec_from_cmdline(sys.argv)
File "/home/aiida/codes/aiida_core/aiida/cmdline/verdilib.py", line 1041, in exec_from_cmdline
CommandClass.run(*argv[command_position + 1:])
File "/home/aiida/codes/aiida_core/aiida/cmdline/baseclass.py", line 234, in run
function_to_call(*args[1:])
File "/home/aiida/codes/aiida_core/aiida/cmdline/commands/calculation.py", line 315, in calculation_show
print_node_info(calc)
File "/home/aiida/codes/aiida_core/aiida/cmdline/utils/common.py", line 95, in print_node_info
print_node_summary(node)
File "/home/aiida/codes/aiida_core/aiida/cmdline/utils/common.py", line 60, in print_node_summary
table.append(['process state', ProcessState(node.process_state)])
AttributeError: 'FolderData' object has no attribute 'process_state'
```
|
1.0
|
verdi calculation show should check whether the id references a calculation node - Currently it does not and will except on some of the properties:
```
Traceback (most recent call last):
File "/home/aiida/.virtualenvs/aiida/bin/verdi", line 9, in <module>
sys.exit(run())
File "/home/aiida/codes/aiida_core/aiida/cmdline/verdilib.py", line 1056, in run
aiida.cmdline.verdilib.exec_from_cmdline(sys.argv)
File "/home/aiida/codes/aiida_core/aiida/cmdline/verdilib.py", line 1041, in exec_from_cmdline
CommandClass.run(*argv[command_position + 1:])
File "/home/aiida/codes/aiida_core/aiida/cmdline/baseclass.py", line 234, in run
function_to_call(*args[1:])
File "/home/aiida/codes/aiida_core/aiida/cmdline/commands/calculation.py", line 315, in calculation_show
print_node_info(calc)
File "/home/aiida/codes/aiida_core/aiida/cmdline/utils/common.py", line 95, in print_node_info
print_node_summary(node)
File "/home/aiida/codes/aiida_core/aiida/cmdline/utils/common.py", line 60, in print_node_summary
table.append(['process state', ProcessState(node.process_state)])
AttributeError: 'FolderData' object has no attribute 'process_state'
```
|
non_defect
|
verdi calculation show should check whether the id references a calculation node currently it does not and will except on some of the properties traceback most recent call last file home aiida virtualenvs aiida bin verdi line in sys exit run file home aiida codes aiida core aiida cmdline verdilib py line in run aiida cmdline verdilib exec from cmdline sys argv file home aiida codes aiida core aiida cmdline verdilib py line in exec from cmdline commandclass run argv file home aiida codes aiida core aiida cmdline baseclass py line in run function to call args file home aiida codes aiida core aiida cmdline commands calculation py line in calculation show print node info calc file home aiida codes aiida core aiida cmdline utils common py line in print node info print node summary node file home aiida codes aiida core aiida cmdline utils common py line in print node summary table append attributeerror folderdata object has no attribute process state
| 0
|
25,787
| 3,964,876,590
|
IssuesEvent
|
2016-05-03 04:21:11
|
uccser/cs-field-guide
|
https://api.github.com/repos/uccser/cs-field-guide
|
closed
|
Numbering questions in a box
|
bug help wanted website/design
|
In the data representation chapter we have "1. Represent each character in your name using UTF-32.", which is part of a numbered list. However, the rendered version doesn't number the list (same for the next spoiler box).
|
1.0
|
Numbering questions in a box - In the data representation chapter we have "1. Represent each character in your name using UTF-32.", which is part of a numbered list. However, the rendered version doesn't number the list (same for the next spoiler box).
|
non_defect
|
numbering questions in a box in the data representation chapter we have represent each character in your name using utf which is part of a numbered list however the rendered version doesn t number the list same for the next spoiler box
| 0
|
42,135
| 10,821,170,574
|
IssuesEvent
|
2019-11-08 18:01:33
|
vector-im/riot-web
|
https://api.github.com/repos/vector-im/riot-web
|
closed
|
Formatting bar has lost its rounded corners
|
bug defect p1 ui/ux 🔥 Fire 🔥
|
Latest develop / latest macOS / latest Chrome. Should be stylised the same as message actions. See GIF in https://github.com/vector-im/riot-web/issues/11202.
|
1.0
|
Formatting bar has lost its rounded corners - Latest develop / latest macOS / latest Chrome. Should be stylised the same as message actions. See GIF in https://github.com/vector-im/riot-web/issues/11202.
|
defect
|
formatting bar has lost its rounded corners latest develop latest macos latest chrome should be stylised the same as message actions see gif in
| 1
|
111,909
| 9,545,970,503
|
IssuesEvent
|
2019-05-01 18:34:35
|
cockroachdb/cockroach
|
https://api.github.com/repos/cockroachdb/cockroach
|
closed
|
teamcity: failed test: TestLint
|
C-test-failure O-robot
|
The following tests appear to have failed on master (lint): TestLint/TestVet: TestLint/TestVet/shadow, TestLint/TestVet, TestLint
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestLint).
[#1270290](https://teamcity.cockroachdb.com/viewLog.html?buildId=1270290):
```
TestLint
--- FAIL: lint/TestLint (274.410s)
TestLint/TestVet
--- FAIL: lint/TestLint: TestLint/TestVet (578.600s)
------- Stdout: -------
=== PAUSE TestLint/TestVet
TestLint/TestVet: TestLint/TestVet/shadow
...rc/runtime/malloc.go:913 +0x896 fp=0xc0086558c0 sp=0xc008655820 pc=0x40bb26
lint_test.go:1352:
runtime.growslice(0xc73e40, 0xc017fda000, 0xf00, 0xf00, 0xf01, 0xc015d8cc88, 0x0, 0x0)
lint_test.go:1352:
/usr/local/go/src/runtime/slice.go:204 +0x145 fp=0xc008655928 sp=0xc0086558c0 pc=0x43f655
lint_test.go:1352:
cmd/internal/obj.Addrel(0xc001383b90, 0xc008655a60)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/data.go:186 +0xf9 fp=0xc008655980 sp=0xc008655928 pc=0x4cf8f9
lint_test.go:1352:
cmd/internal/obj/x86.(*AsmBuf).doasm(0xc008655d78, 0xc0002ea000, 0xc001383b90, 0xc015d8cdd0)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/x86/asm6.go:4677 +0xeb8 fp=0xc008655c58 sp=0xc008655980 pc=0x5c5748
lint_test.go:1352:
cmd/internal/obj/x86.(*AsmBuf).asmins(0xc008655d78, 0xc0002ea000, 0xc001383b90, 0xc015d8cdd0)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/x86/asm6.go:5369 +0x4a6 fp=0xc008655d00 sp=0xc008655c58 pc=0x5cc986
lint_test.go:1352:
cmd/internal/obj/x86.span6(0xc0002ea000, 0xc001383b90, 0xc00102a630)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/x86/asm6.go:1999 +0x3b2 fp=0xc008655e20 sp=0xc008655d00 pc=0x5be752
lint_test.go:1352:
cmd/internal/obj.Flushplist(0xc0002ea000, 0xc008655f08, 0xc00102a630, 0x7ffd5d9278cb, 0x2f)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/plist.go:107 +0x6ab fp=0xc008655ee0 sp=0xc008655e20 pc=0x4d916b
lint_test.go:1352:
cmd/compile/internal/gc.(*Progs).Flush(0xc008581f10)
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/gsubr.go:93 +0xbc fp=0xc008655f30 sp=0xc008655ee0 pc=0xabd6bc
lint_test.go:1352:
cmd/compile/internal/gc.compileSSA(0xc00127d340, 0x3)
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/pgen.go:283 +0x2b9 fp=0xc008655f98 sp=0xc008655f30 pc=0xb02ed9
lint_test.go:1352:
cmd/compile/internal/gc.compileFunctions.func2(0xc005966540, 0xc005975db0, 0x3)
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/pgen.go:323 +0x49 fp=0xc008655fc8 sp=0xc008655f98 pc=0xbb7f79
lint_test.go:1352:
runtime.goexit()
lint_test.go:1352:
/usr/local/go/src/runtime/asm_amd64.s:1333 +0x1 fp=0xc008655fd0 sp=0xc008655fc8 pc=0x457da1
lint_test.go:1352:
created by cmd/compile/internal/gc.compileFunctions
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/pgen.go:321 +0x11a
lint_test.go:1352:
goroutine 1 [semacquire]:
lint_test.go:1352:
sync.runtime_Semacquire(0xc005975db8)
lint_test.go:1352:
/usr/local/go/src/runtime/sema.go:56 +0x39
lint_test.go:1352:
sync.(*WaitGroup).Wait(0xc005975db0)
lint_test.go:1352:
/usr/local/go/src/sync/waitgroup.go:130 +0x64
lint_test.go:1352:
cmd/compile/internal/gc.compileFunctions()
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/pgen.go:333 +0x1c3
lint_test.go:1352:
cmd/compile/internal/gc.Main(0xcc51f8)
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/main.go:651 +0x276d
lint_test.go:1352:
main.main()
lint_test.go:1352:
/usr/local/go/src/cmd/compile/main.go:51 +0x96
```
Please assign, take a look and update the issue accordingly.
|
1.0
|
teamcity: failed test: TestLint - The following tests appear to have failed on master (lint): TestLint/TestVet: TestLint/TestVet/shadow, TestLint/TestVet, TestLint
You may want to check [for open issues](https://github.com/cockroachdb/cockroach/issues?q=is%3Aissue+is%3Aopen+TestLint).
[#1270290](https://teamcity.cockroachdb.com/viewLog.html?buildId=1270290):
```
TestLint
--- FAIL: lint/TestLint (274.410s)
TestLint/TestVet
--- FAIL: lint/TestLint: TestLint/TestVet (578.600s)
------- Stdout: -------
=== PAUSE TestLint/TestVet
TestLint/TestVet: TestLint/TestVet/shadow
...rc/runtime/malloc.go:913 +0x896 fp=0xc0086558c0 sp=0xc008655820 pc=0x40bb26
lint_test.go:1352:
runtime.growslice(0xc73e40, 0xc017fda000, 0xf00, 0xf00, 0xf01, 0xc015d8cc88, 0x0, 0x0)
lint_test.go:1352:
/usr/local/go/src/runtime/slice.go:204 +0x145 fp=0xc008655928 sp=0xc0086558c0 pc=0x43f655
lint_test.go:1352:
cmd/internal/obj.Addrel(0xc001383b90, 0xc008655a60)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/data.go:186 +0xf9 fp=0xc008655980 sp=0xc008655928 pc=0x4cf8f9
lint_test.go:1352:
cmd/internal/obj/x86.(*AsmBuf).doasm(0xc008655d78, 0xc0002ea000, 0xc001383b90, 0xc015d8cdd0)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/x86/asm6.go:4677 +0xeb8 fp=0xc008655c58 sp=0xc008655980 pc=0x5c5748
lint_test.go:1352:
cmd/internal/obj/x86.(*AsmBuf).asmins(0xc008655d78, 0xc0002ea000, 0xc001383b90, 0xc015d8cdd0)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/x86/asm6.go:5369 +0x4a6 fp=0xc008655d00 sp=0xc008655c58 pc=0x5cc986
lint_test.go:1352:
cmd/internal/obj/x86.span6(0xc0002ea000, 0xc001383b90, 0xc00102a630)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/x86/asm6.go:1999 +0x3b2 fp=0xc008655e20 sp=0xc008655d00 pc=0x5be752
lint_test.go:1352:
cmd/internal/obj.Flushplist(0xc0002ea000, 0xc008655f08, 0xc00102a630, 0x7ffd5d9278cb, 0x2f)
lint_test.go:1352:
/usr/local/go/src/cmd/internal/obj/plist.go:107 +0x6ab fp=0xc008655ee0 sp=0xc008655e20 pc=0x4d916b
lint_test.go:1352:
cmd/compile/internal/gc.(*Progs).Flush(0xc008581f10)
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/gsubr.go:93 +0xbc fp=0xc008655f30 sp=0xc008655ee0 pc=0xabd6bc
lint_test.go:1352:
cmd/compile/internal/gc.compileSSA(0xc00127d340, 0x3)
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/pgen.go:283 +0x2b9 fp=0xc008655f98 sp=0xc008655f30 pc=0xb02ed9
lint_test.go:1352:
cmd/compile/internal/gc.compileFunctions.func2(0xc005966540, 0xc005975db0, 0x3)
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/pgen.go:323 +0x49 fp=0xc008655fc8 sp=0xc008655f98 pc=0xbb7f79
lint_test.go:1352:
runtime.goexit()
lint_test.go:1352:
/usr/local/go/src/runtime/asm_amd64.s:1333 +0x1 fp=0xc008655fd0 sp=0xc008655fc8 pc=0x457da1
lint_test.go:1352:
created by cmd/compile/internal/gc.compileFunctions
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/pgen.go:321 +0x11a
lint_test.go:1352:
goroutine 1 [semacquire]:
lint_test.go:1352:
sync.runtime_Semacquire(0xc005975db8)
lint_test.go:1352:
/usr/local/go/src/runtime/sema.go:56 +0x39
lint_test.go:1352:
sync.(*WaitGroup).Wait(0xc005975db0)
lint_test.go:1352:
/usr/local/go/src/sync/waitgroup.go:130 +0x64
lint_test.go:1352:
cmd/compile/internal/gc.compileFunctions()
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/pgen.go:333 +0x1c3
lint_test.go:1352:
cmd/compile/internal/gc.Main(0xcc51f8)
lint_test.go:1352:
/usr/local/go/src/cmd/compile/internal/gc/main.go:651 +0x276d
lint_test.go:1352:
main.main()
lint_test.go:1352:
/usr/local/go/src/cmd/compile/main.go:51 +0x96
```
Please assign, take a look and update the issue accordingly.
|
non_defect
|
teamcity failed test testlint the following tests appear to have failed on master lint testlint testvet testlint testvet shadow testlint testvet testlint you may want to check testlint fail lint testlint testlint testvet fail lint testlint testlint testvet stdout pause testlint testvet testlint testvet testlint testvet shadow rc runtime malloc go fp sp pc lint test go runtime growslice lint test go usr local go src runtime slice go fp sp pc lint test go cmd internal obj addrel lint test go usr local go src cmd internal obj data go fp sp pc lint test go cmd internal obj asmbuf doasm lint test go usr local go src cmd internal obj go fp sp pc lint test go cmd internal obj asmbuf asmins lint test go usr local go src cmd internal obj go fp sp pc lint test go cmd internal obj lint test go usr local go src cmd internal obj go fp sp pc lint test go cmd internal obj flushplist lint test go usr local go src cmd internal obj plist go fp sp pc lint test go cmd compile internal gc progs flush lint test go usr local go src cmd compile internal gc gsubr go fp sp pc lint test go cmd compile internal gc compilessa lint test go usr local go src cmd compile internal gc pgen go fp sp pc lint test go cmd compile internal gc compilefunctions lint test go usr local go src cmd compile internal gc pgen go fp sp pc lint test go runtime goexit lint test go usr local go src runtime asm s fp sp pc lint test go created by cmd compile internal gc compilefunctions lint test go usr local go src cmd compile internal gc pgen go lint test go goroutine lint test go sync runtime semacquire lint test go usr local go src runtime sema go lint test go sync waitgroup wait lint test go usr local go src sync waitgroup go lint test go cmd compile internal gc compilefunctions lint test go usr local go src cmd compile internal gc pgen go lint test go cmd compile internal gc main lint test go usr local go src cmd compile internal gc main go lint test go main main lint test go usr local go src cmd compile main go please assign take a look and update the issue accordingly
| 0
|
40,058
| 9,815,339,958
|
IssuesEvent
|
2019-06-13 12:26:46
|
roc-project/roc
|
https://api.github.com/repos/roc-project/roc
|
closed
|
OpenFEC LDPC incorrect decoding
|
defect
|
Can be reproduced with "full_repair_payload_sizes" test. Reproducible only on LDPC and on some payload sizes, e.g. 69.
The last byte of the decoded packet is incorrect.
|
1.0
|
OpenFEC LDPC incorrect decoding - Can be reproduced with "full_repair_payload_sizes" test. Reproducible only on LDPC and on some payload sizes, e.g. 69.
The last byte of the decoded packet is incorrect.
|
defect
|
openfec ldpc incorrect decoding can be reproduced with full repair payload sizes test reproducible only on ldpc and on some payload sizes e g the last byte of the decoded packet is incorrect
| 1
|
72,531
| 24,167,087,429
|
IssuesEvent
|
2022-09-22 15:51:05
|
idaholab/moose
|
https://api.github.com/repos/idaholab/moose
|
opened
|
AdvancedExtrudedMeshGenerator Documentation Update to Cover Triple-Indexed Parameter
|
T: defect P: normal
|
## Bug Description
`AdvancedExtrudedMeshGenerator` was recently updated to use a triple-indexed input parameter. But the documentation has not been updated accordingly.
## Steps to Reproduce
The documentation page of `AdvancedExtrudedMeshGenerator` still uses stacked double-indexed input parameter.
## Impact
The documentation is confusing due to the inconsistence.
|
1.0
|
AdvancedExtrudedMeshGenerator Documentation Update to Cover Triple-Indexed Parameter - ## Bug Description
`AdvancedExtrudedMeshGenerator` was recently updated to use a triple-indexed input parameter. But the documentation has not been updated accordingly.
## Steps to Reproduce
The documentation page of `AdvancedExtrudedMeshGenerator` still uses stacked double-indexed input parameter.
## Impact
The documentation is confusing due to the inconsistence.
|
defect
|
advancedextrudedmeshgenerator documentation update to cover triple indexed parameter bug description advancedextrudedmeshgenerator was recently updated to use a triple indexed input parameter but the documentation has not been updated accordingly steps to reproduce the documentation page of advancedextrudedmeshgenerator still uses stacked double indexed input parameter impact the documentation is confusing due to the inconsistence
| 1
|
62,618
| 17,092,699,146
|
IssuesEvent
|
2021-07-08 19:50:24
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
DatePicker: TimeOnly 12:04 a.m. (midnight plus 4 minutes) is not handled correctly
|
defect
|
(Apologies in advance, I am new to reporting defects. I kinda just looked at what others were doing to fill in the format below.)
**Describe the defect**
Using the time-only 12-hour datePicker causes 12:04 a.m. to become 12:04 p.m. The input box correctly displays "12:04 AM", but when opening the time picker, it gets changed to 12:04 PM" and updates the input box with the incorrect value.
GIF below:

**Environment**
PF Version: 8.0.5
JSF + version: Mojarra 2.3.14
Tomcat 8.5
Affected browsers: Chrome (at least)
**To Reproduce**
Steps to reproduce the behavior:
1. XHTML:
```<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE html
PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN"
"http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
<html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en" lang="en" xmlns:p="http://primefaces.org/ui" xmlns:h="http://java.sun.com/jsf/html"
xmlns:f="http://java.sun.com/jsf/core">
<h:head>
<title>Title</title>
</h:head>
<body>
<p:datePicker id="datePickerDemo" hourFormat="12" value="#{rTest.dateExample}" timeOnly="true"/>
</body>
</html>
```
2. Backing Bean (note: same behavior with LocalTime and Date)
```import java.time.LocalTime;
import javax.enterprise.context.RequestScoped;
import javax.inject.Named;
@RequestScoped
@Named("rTest")
public class RequestTest {
private LocalTime dateExample = LocalTime.of(0, 4);
public LocalTime getDateExample() {
return dateExample;
}
public void setDateExample(LocalTime dateExample) {
this.dateExample = dateExample;
}
}
```
**Commentary**
The issue appears to be in [this code](https://github.com/primefaces/primefaces/blob/6c67dadd07ab2b7887f43f14da82d73f30427659/src/main/resources/META-INF/resources/primefaces/datepicker/0-datepicker.js#L752) snippet. Passing "12:04" and "AM" into parseTime causes the AM to get dropped and hour: 12, minute: 4, second: null to be returned. I'm not much of a JavaScript developer, but maybe the hour needs to get set to zero?
Thanks!
|
1.0
|
DatePicker: TimeOnly 12:04 a.m. (midnight plus 4 minutes) is not handled correctly - (Apologies in advance, I am new to reporting defects. I kinda just looked at what others were doing to fill in the format below.)
**Describe the defect**
Using the time-only 12-hour datePicker causes 12:04 a.m. to become 12:04 p.m. The input box correctly displays "12:04 AM", but when opening the time picker, it gets changed to 12:04 PM" and updates the input box with the incorrect value.
GIF below:

**Environment**
PF Version: 8.0.5
JSF + version: Mojarra 2.3.14
Tomcat 8.5
Affected browsers: Chrome (at least)
**To Reproduce**
Steps to reproduce the behavior:
1. XHTML:
```<?xml version="1.0" encoding="UTF-8"?>
<!DOCTYPE html
PUBLIC "-//W3C//DTD XHTML 1.0 Transitional//EN"
"http://www.w3.org/TR/xhtml1/DTD/xhtml1-transitional.dtd">
<html xmlns="http://www.w3.org/1999/xhtml" xml:lang="en" lang="en" xmlns:p="http://primefaces.org/ui" xmlns:h="http://java.sun.com/jsf/html"
xmlns:f="http://java.sun.com/jsf/core">
<h:head>
<title>Title</title>
</h:head>
<body>
<p:datePicker id="datePickerDemo" hourFormat="12" value="#{rTest.dateExample}" timeOnly="true"/>
</body>
</html>
```
2. Backing Bean (note: same behavior with LocalTime and Date)
```import java.time.LocalTime;
import javax.enterprise.context.RequestScoped;
import javax.inject.Named;
@RequestScoped
@Named("rTest")
public class RequestTest {
private LocalTime dateExample = LocalTime.of(0, 4);
public LocalTime getDateExample() {
return dateExample;
}
public void setDateExample(LocalTime dateExample) {
this.dateExample = dateExample;
}
}
```
**Commentary**
The issue appears to be in [this code](https://github.com/primefaces/primefaces/blob/6c67dadd07ab2b7887f43f14da82d73f30427659/src/main/resources/META-INF/resources/primefaces/datepicker/0-datepicker.js#L752) snippet. Passing "12:04" and "AM" into parseTime causes the AM to get dropped and hour: 12, minute: 4, second: null to be returned. I'm not much of a JavaScript developer, but maybe the hour needs to get set to zero?
Thanks!
|
defect
|
datepicker timeonly a m midnight plus minutes is not handled correctly apologies in advance i am new to reporting defects i kinda just looked at what others were doing to fill in the format below describe the defect using the time only hour datepicker causes a m to become p m the input box correctly displays am but when opening the time picker it gets changed to pm and updates the input box with the incorrect value gif below environment pf version jsf version mojarra tomcat affected browsers chrome at least to reproduce steps to reproduce the behavior xhtml doctype html public dtd xhtml transitional en html xmlns xml lang en lang en xmlns p xmlns h xmlns f title backing bean note same behavior with localtime and date import java time localtime import javax enterprise context requestscoped import javax inject named requestscoped named rtest public class requesttest private localtime dateexample localtime of public localtime getdateexample return dateexample public void setdateexample localtime dateexample this dateexample dateexample commentary the issue appears to be in snippet passing and am into parsetime causes the am to get dropped and hour minute second null to be returned i m not much of a javascript developer but maybe the hour needs to get set to zero thanks
| 1
|
308,141
| 23,234,190,018
|
IssuesEvent
|
2022-08-03 10:14:52
|
open-metadata/OpenMetadata
|
https://api.github.com/repos/open-metadata/OpenMetadata
|
opened
|
Documentation or Tutorial Requests!!
|
documentation
|
Dear developers, first of all I am very interested in this project, but I have some problems when I perform local deployment, I will be very grateful if the developers can reply to me.The tutorial of the reference documentation of this open source project is behind the code. If possible, the documentation can be updated in time. At present, I can start the service locally, but the ingestion module cannot connect to airflow through rest-api. The ingestion module uses the official image provided. Thank you very much.
|
1.0
|
Documentation or Tutorial Requests!! - Dear developers, first of all I am very interested in this project, but I have some problems when I perform local deployment, I will be very grateful if the developers can reply to me.The tutorial of the reference documentation of this open source project is behind the code. If possible, the documentation can be updated in time. At present, I can start the service locally, but the ingestion module cannot connect to airflow through rest-api. The ingestion module uses the official image provided. Thank you very much.
|
non_defect
|
documentation or tutorial requests!! dear developers first of all i am very interested in this project but i have some problems when i perform local deployment i will be very grateful if the developers can reply to me the tutorial of the reference documentation of this open source project is behind the code if possible the documentation can be updated in time at present i can start the service locally but the ingestion module cannot connect to airflow through rest api the ingestion module uses the official image provided thank you very much
| 0
|
214
| 2,519,740,251
|
IssuesEvent
|
2015-01-18 09:01:26
|
mbunkus/mtx-trac-import-test
|
https://api.github.com/repos/mbunkus/mtx-trac-import-test
|
closed
|
No mux aac and RV9
|
C: mkvmerge P: normal R: invalid T: defect
|
**Reported by slickbitch1 on 21 Dec 2003 13:25 UTC**
Windows XP, AutoRV9, OagMachine aac 5.1
After mkvmrege says opened " ".mkv for writing, it crashes and returns
error code 1073741676
|
1.0
|
No mux aac and RV9 - **Reported by slickbitch1 on 21 Dec 2003 13:25 UTC**
Windows XP, AutoRV9, OagMachine aac 5.1
After mkvmrege says opened " ".mkv for writing, it crashes and returns
error code 1073741676
|
defect
|
no mux aac and reported by on dec utc windows xp oagmachine aac after mkvmrege says opened mkv for writing it crashes and returns error code
| 1
|
57,795
| 16,065,123,067
|
IssuesEvent
|
2021-04-23 17:49:36
|
department-of-veterans-affairs/va.gov-cms
|
https://api.github.com/repos/department-of-veterans-affairs/va.gov-cms
|
closed
|
Menu item should not have a default parent for any node type.
|
Defect Needs refining Product Support Team
|
**Describe the defect**
**To Reproduce**
As Victor.A.MCtest
1. Go to /node/add/event and enable menu item
2. note that the parent is set to Altoona
**Expected behavior**
There should be no default parent.
Also, grep the config/sync folder for `parent: '`, you'll find this is set on on a number of `node.type.*.yml`
**ACs**
- [ ] Menu items for all content types should always be specified manually.
|
1.0
|
Menu item should not have a default parent for any node type. - **Describe the defect**
**To Reproduce**
As Victor.A.MCtest
1. Go to /node/add/event and enable menu item
2. note that the parent is set to Altoona
**Expected behavior**
There should be no default parent.
Also, grep the config/sync folder for `parent: '`, you'll find this is set on on a number of `node.type.*.yml`
**ACs**
- [ ] Menu items for all content types should always be specified manually.
|
defect
|
menu item should not have a default parent for any node type describe the defect to reproduce as victor a mctest go to node add event and enable menu item note that the parent is set to altoona expected behavior there should be no default parent also grep the config sync folder for parent you ll find this is set on on a number of node type yml acs menu items for all content types should always be specified manually
| 1
|
801,911
| 28,506,655,274
|
IssuesEvent
|
2023-04-18 22:12:30
|
googleapis/python-aiplatform
|
https://api.github.com/repos/googleapis/python-aiplatform
|
closed
|
tests.system.aiplatform.test_matching_engine_index.TestMatchingEngine: test_create_get_list_matching_engine_index failed
|
type: bug priority: p1 flakybot: issue api: vertex-ai
|
Note: #1932 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: d75322c24d1b802c77493550fa08bfffdf3ec8fb
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f382ab13-7b18-4cf4-becd-148dae7c3389), [Sponge](http://sponge2/f382ab13-7b18-4cf4-becd-148dae7c3389)
status: failed
<details><summary>Test output</summary><br><pre>args = (name: "projects/580378083368/locations/us-central1/indexes/2091385465239240704"
,)
kwargs = {'metadata': [('x-goog-request-params', 'name=projects/580378083368/locations/us-central1/indexes/2091385465239240704'), ('x-goog-api-client', 'model-builder/1.24.0 gl-python/3.8.13 grpc/1.47.0 gax/1.32.0 gapic/1.24.0')]}
@six.wraps(callable_)
def error_remapped_callable(*args, **kwargs):
try:
> return callable_(*args, **kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:67:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f6f51674df0>
request = name: "projects/580378083368/locations/us-central1/indexes/2091385465239240704"
timeout = None
metadata = [('x-goog-request-params', 'name=projects/580378083368/locations/us-central1/indexes/2091385465239240704'), ('x-goog-api-client', 'model-builder/1.24.0 gl-python/3.8.13 grpc/1.47.0 gax/1.32.0 gapic/1.24.0')]
credentials = None, wait_for_ready = None, compression = None
def __call__(self,
request,
timeout=None,
metadata=None,
credentials=None,
wait_for_ready=None,
compression=None):
state, call, = self._blocking(request, timeout, metadata, credentials,
wait_for_ready, compression)
> return _end_unary_response_blocking(state, call, False, None)
.nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:946:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
state = <grpc._channel._RPCState object at 0x7f6f516880d0>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f6f51588bc0>
with_call = False, deadline = None
def _end_unary_response_blocking(state, call, with_call, deadline):
if state.code is grpc.StatusCode.OK:
if with_call:
rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
return state.response, rendezvous
else:
return state.response
else:
> raise _InactiveRpcError(state)
E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E status = StatusCode.FAILED_PRECONDITION
E details = "The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136."
E debug_error_string = "{"created":"@1681797321.728722218","description":"Error received from peer ipv4:142.250.107.95:443","file":"src/core/lib/surface/call.cc","file_line":966,"grpc_message":"The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.","grpc_status":9}"
E >
.nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:849: _InactiveRpcError
The above exception was the direct cause of the following exception:
self = <tests.system.aiplatform.test_matching_engine_index.TestMatchingEngine object at 0x7f6f54dfd130>
shared_state = {'index': <google.cloud.aiplatform.matching_engine.matching_engine_index.MatchingEngineIndex object at 0x7f6f52beafd0>...ect at 0x7f6f5168f130>
resource name: projects/580378083368/locations/us-central1/indexEndpoints/3989652708175904768]}
def test_create_get_list_matching_engine_index(self, shared_state):
aiplatform.init(
project=e2e_base._PROJECT,
location=e2e_base._LOCATION,
)
# Create an index
index = aiplatform.MatchingEngineIndex.create_tree_ah_index(
display_name=_TEST_INDEX_DISPLAY_NAME,
contents_delta_uri=_TEST_CONTENTS_DELTA_URI,
dimensions=_TEST_INDEX_CONFIG_DIMENSIONS,
approximate_neighbors_count=_TEST_INDEX_APPROXIMATE_NEIGHBORS_COUNT,
distance_measure_type=_TEST_INDEX_DISTANCE_MEASURE_TYPE,
leaf_node_embedding_count=_TEST_LEAF_NODE_EMBEDDING_COUNT,
leaf_nodes_to_search_percent=_TEST_LEAF_NODES_TO_SEARCH_PERCENT,
description=_TEST_INDEX_DESCRIPTION,
labels=_TEST_LABELS,
)
shared_state["resources"] = [index]
shared_state["index"] = index
shared_state["index_name"] = index.resource_name
# Verify that the retrieved index is the same
get_index = aiplatform.MatchingEngineIndex(index_name=index.resource_name)
assert index.resource_name == get_index.resource_name
# Create index and check that it is listed
list_indexes = aiplatform.MatchingEngineIndex.list()
assert get_index.resource_name in [
index.resource_name for index in list_indexes
]
# Update the index metadata
updated_index = get_index.update_metadata(
display_name=_TEST_DISPLAY_NAME_UPDATE,
description=_TEST_DESCRIPTION_UPDATE,
labels=_TEST_LABELS_UPDATE,
)
assert updated_index.name == get_index.name
# TODO: Reinstate assertions once b/220005272 is fixed.
# assert updated_index.display_name == _TEST_DISPLAY_NAME_UPDATE
# assert updated_index.description == _TEST_DESCRIPTION_UPDATE
# assert updated_index.labels == _TEST_LABELS_UPDATE
# Update the index embeddings
updated_index = get_index.update_embeddings(
contents_delta_uri=_TEST_CONTENTS_DELTA_URI_UPDATE,
is_complete_overwrite=_TEST_IS_COMPLETE_OVERWRITE,
)
assert updated_index.name == get_index.name
# Create endpoint and check that it is listed
my_index_endpoint = aiplatform.MatchingEngineIndexEndpoint.create(
display_name=_TEST_INDEX_ENDPOINT_DISPLAY_NAME,
description=_TEST_INDEX_ENDPOINT_DESCRIPTION,
network=e2e_base._VPC_NETWORK_URI,
labels=_TEST_LABELS,
)
assert my_index_endpoint.resource_name in [
index_endpoint.resource_name
for index_endpoint in aiplatform.MatchingEngineIndexEndpoint.list()
]
assert my_index_endpoint.labels == _TEST_LABELS
assert my_index_endpoint.display_name == _TEST_INDEX_ENDPOINT_DISPLAY_NAME
assert my_index_endpoint.description == _TEST_INDEX_ENDPOINT_DESCRIPTION
# Create endpoint and check that it is listed
public_index_endpoint = aiplatform.MatchingEngineIndexEndpoint.create(
display_name=_TEST_PUBLIC_INDEX_ENDPOINT_DISPLAY_NAME,
description=_TEST_PUBLIC_INDEX_ENDPOINT_DESCRIPTION,
public_endpoint_enabled=True,
labels=_TEST_LABELS,
)
assert public_index_endpoint.resource_name in [
index_endpoint.resource_name
for index_endpoint in aiplatform.MatchingEngineIndexEndpoint.list()
]
assert public_index_endpoint.labels == _TEST_LABELS
assert (
public_index_endpoint.display_name
== _TEST_PUBLIC_INDEX_ENDPOINT_DISPLAY_NAME
)
assert (
public_index_endpoint.description == _TEST_PUBLIC_INDEX_ENDPOINT_DESCRIPTION
)
shared_state["resources"].append(my_index_endpoint)
# Deploy endpoint
my_index_endpoint = my_index_endpoint.deploy_index(
index=index,
deployed_index_id=_TEST_DEPLOYED_INDEX_ID,
display_name=_TEST_DEPLOYED_INDEX_DISPLAY_NAME,
)
# Deploy public endpoint
public_index_endpoint = public_index_endpoint.deploy_index(
index=index,
deployed_index_id=_TEST_DEPLOYED_INDEX_ID_PUBLIC,
display_name=_TEST_DEPLOYED_INDEX_DISPLAY_NAME_PUBLIC,
min_replica_count=_TEST_MIN_REPLICA_COUNT_UPDATED,
max_replica_count=_TEST_MAX_REPLICA_COUNT_UPDATED,
)
# Update endpoint
updated_index_endpoint = my_index_endpoint.update(
display_name=_TEST_DISPLAY_NAME_UPDATE,
description=_TEST_DESCRIPTION_UPDATE,
labels=_TEST_LABELS_UPDATE,
)
assert updated_index_endpoint.labels == _TEST_LABELS_UPDATE
assert updated_index_endpoint.display_name == _TEST_DISPLAY_NAME_UPDATE
assert updated_index_endpoint.description == _TEST_DESCRIPTION_UPDATE
# Mutate deployed index
my_index_endpoint.mutate_deployed_index(
deployed_index_id=_TEST_DEPLOYED_INDEX_ID,
min_replica_count=_TEST_MIN_REPLICA_COUNT_UPDATED,
max_replica_count=_TEST_MAX_REPLICA_COUNT_UPDATED,
)
# deployed index on private endpoint.
deployed_index = my_index_endpoint.deployed_indexes[0]
assert deployed_index.id == _TEST_DEPLOYED_INDEX_ID
assert deployed_index.index == index.resource_name
assert (
deployed_index.automatic_resources.min_replica_count
== _TEST_MIN_REPLICA_COUNT_UPDATED
)
assert (
deployed_index.automatic_resources.max_replica_count
== _TEST_MAX_REPLICA_COUNT_UPDATED
)
# deployed index on public endpoint.
deployed_index_public = public_index_endpoint.deployed_indexes[0]
assert deployed_index_public.id == _TEST_DEPLOYED_INDEX_ID_PUBLIC
assert deployed_index_public.index == index.resource_name
assert (
deployed_index_public.automatic_resources.min_replica_count
== _TEST_MIN_REPLICA_COUNT_UPDATED
)
assert (
deployed_index_public.automatic_resources.max_replica_count
== _TEST_MAX_REPLICA_COUNT_UPDATED
)
# TODO: Test `my_index_endpoint.match` request. This requires running this test in a VPC.
# results = my_index_endpoint.match(
# deployed_index_id=_TEST_DEPLOYED_INDEX_ID, queries=[_TEST_MATCH_QUERY]
# )
# assert results[0][0].id == 870
# TODO: Test `my_index_endpoint.match` with filter.
# This requires uploading a new content of the Matching Engine Index to Cloud Storage.
# results = my_index_endpoint.match(
# deployed_index_id=_TEST_DEPLOYED_INDEX_ID,
# queries=[_TEST_MATCH_QUERY],
# num_neighbors=1,
# filter=_TEST_FILTER,
# )
# assert results[0][0].id == 9999
# Undeploy index
my_index_endpoint = my_index_endpoint.undeploy_index(
deployed_index_id=deployed_index.id
)
# Delete index and check that it is no longer listed
> index.delete()
tests/system/aiplatform/test_matching_engine_index.py:356:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/aiplatform/base.py:814: in wrapper
return method(*args, **kwargs)
google/cloud/aiplatform/base.py:1269: in delete
lro = getattr(self.api_client, self._delete_method)(name=self.resource_name)
google/cloud/aiplatform_v1/services/index_service/client.py:1071: in delete_index
response = rpc(
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/gapic_v1/method.py:145: in __call__
return wrapped_func(*args, **kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = None
from_value = <_InactiveRpcError of RPC that terminated with:
status = StatusCode.FAILED_PRECONDITION
details = "The Index "projec... IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.","grpc_status":9}"
>
> ???
E google.api_core.exceptions.FailedPrecondition: 400 The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.
<string>:3: FailedPrecondition</pre></details>
|
1.0
|
tests.system.aiplatform.test_matching_engine_index.TestMatchingEngine: test_create_get_list_matching_engine_index failed - Note: #1932 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky.
----
commit: d75322c24d1b802c77493550fa08bfffdf3ec8fb
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/f382ab13-7b18-4cf4-becd-148dae7c3389), [Sponge](http://sponge2/f382ab13-7b18-4cf4-becd-148dae7c3389)
status: failed
<details><summary>Test output</summary><br><pre>args = (name: "projects/580378083368/locations/us-central1/indexes/2091385465239240704"
,)
kwargs = {'metadata': [('x-goog-request-params', 'name=projects/580378083368/locations/us-central1/indexes/2091385465239240704'), ('x-goog-api-client', 'model-builder/1.24.0 gl-python/3.8.13 grpc/1.47.0 gax/1.32.0 gapic/1.24.0')]}
@six.wraps(callable_)
def error_remapped_callable(*args, **kwargs):
try:
> return callable_(*args, **kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:67:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
self = <grpc._channel._UnaryUnaryMultiCallable object at 0x7f6f51674df0>
request = name: "projects/580378083368/locations/us-central1/indexes/2091385465239240704"
timeout = None
metadata = [('x-goog-request-params', 'name=projects/580378083368/locations/us-central1/indexes/2091385465239240704'), ('x-goog-api-client', 'model-builder/1.24.0 gl-python/3.8.13 grpc/1.47.0 gax/1.32.0 gapic/1.24.0')]
credentials = None, wait_for_ready = None, compression = None
def __call__(self,
request,
timeout=None,
metadata=None,
credentials=None,
wait_for_ready=None,
compression=None):
state, call, = self._blocking(request, timeout, metadata, credentials,
wait_for_ready, compression)
> return _end_unary_response_blocking(state, call, False, None)
.nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:946:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
state = <grpc._channel._RPCState object at 0x7f6f516880d0>
call = <grpc._cython.cygrpc.SegregatedCall object at 0x7f6f51588bc0>
with_call = False, deadline = None
def _end_unary_response_blocking(state, call, with_call, deadline):
if state.code is grpc.StatusCode.OK:
if with_call:
rendezvous = _MultiThreadedRendezvous(state, call, None, deadline)
return state.response, rendezvous
else:
return state.response
else:
> raise _InactiveRpcError(state)
E grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with:
E status = StatusCode.FAILED_PRECONDITION
E details = "The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136."
E debug_error_string = "{"created":"@1681797321.728722218","description":"Error received from peer ipv4:142.250.107.95:443","file":"src/core/lib/surface/call.cc","file_line":966,"grpc_message":"The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.","grpc_status":9}"
E >
.nox/system-3-8/lib/python3.8/site-packages/grpc/_channel.py:849: _InactiveRpcError
The above exception was the direct cause of the following exception:
self = <tests.system.aiplatform.test_matching_engine_index.TestMatchingEngine object at 0x7f6f54dfd130>
shared_state = {'index': <google.cloud.aiplatform.matching_engine.matching_engine_index.MatchingEngineIndex object at 0x7f6f52beafd0>...ect at 0x7f6f5168f130>
resource name: projects/580378083368/locations/us-central1/indexEndpoints/3989652708175904768]}
def test_create_get_list_matching_engine_index(self, shared_state):
aiplatform.init(
project=e2e_base._PROJECT,
location=e2e_base._LOCATION,
)
# Create an index
index = aiplatform.MatchingEngineIndex.create_tree_ah_index(
display_name=_TEST_INDEX_DISPLAY_NAME,
contents_delta_uri=_TEST_CONTENTS_DELTA_URI,
dimensions=_TEST_INDEX_CONFIG_DIMENSIONS,
approximate_neighbors_count=_TEST_INDEX_APPROXIMATE_NEIGHBORS_COUNT,
distance_measure_type=_TEST_INDEX_DISTANCE_MEASURE_TYPE,
leaf_node_embedding_count=_TEST_LEAF_NODE_EMBEDDING_COUNT,
leaf_nodes_to_search_percent=_TEST_LEAF_NODES_TO_SEARCH_PERCENT,
description=_TEST_INDEX_DESCRIPTION,
labels=_TEST_LABELS,
)
shared_state["resources"] = [index]
shared_state["index"] = index
shared_state["index_name"] = index.resource_name
# Verify that the retrieved index is the same
get_index = aiplatform.MatchingEngineIndex(index_name=index.resource_name)
assert index.resource_name == get_index.resource_name
# Create index and check that it is listed
list_indexes = aiplatform.MatchingEngineIndex.list()
assert get_index.resource_name in [
index.resource_name for index in list_indexes
]
# Update the index metadata
updated_index = get_index.update_metadata(
display_name=_TEST_DISPLAY_NAME_UPDATE,
description=_TEST_DESCRIPTION_UPDATE,
labels=_TEST_LABELS_UPDATE,
)
assert updated_index.name == get_index.name
# TODO: Reinstate assertions once b/220005272 is fixed.
# assert updated_index.display_name == _TEST_DISPLAY_NAME_UPDATE
# assert updated_index.description == _TEST_DESCRIPTION_UPDATE
# assert updated_index.labels == _TEST_LABELS_UPDATE
# Update the index embeddings
updated_index = get_index.update_embeddings(
contents_delta_uri=_TEST_CONTENTS_DELTA_URI_UPDATE,
is_complete_overwrite=_TEST_IS_COMPLETE_OVERWRITE,
)
assert updated_index.name == get_index.name
# Create endpoint and check that it is listed
my_index_endpoint = aiplatform.MatchingEngineIndexEndpoint.create(
display_name=_TEST_INDEX_ENDPOINT_DISPLAY_NAME,
description=_TEST_INDEX_ENDPOINT_DESCRIPTION,
network=e2e_base._VPC_NETWORK_URI,
labels=_TEST_LABELS,
)
assert my_index_endpoint.resource_name in [
index_endpoint.resource_name
for index_endpoint in aiplatform.MatchingEngineIndexEndpoint.list()
]
assert my_index_endpoint.labels == _TEST_LABELS
assert my_index_endpoint.display_name == _TEST_INDEX_ENDPOINT_DISPLAY_NAME
assert my_index_endpoint.description == _TEST_INDEX_ENDPOINT_DESCRIPTION
# Create endpoint and check that it is listed
public_index_endpoint = aiplatform.MatchingEngineIndexEndpoint.create(
display_name=_TEST_PUBLIC_INDEX_ENDPOINT_DISPLAY_NAME,
description=_TEST_PUBLIC_INDEX_ENDPOINT_DESCRIPTION,
public_endpoint_enabled=True,
labels=_TEST_LABELS,
)
assert public_index_endpoint.resource_name in [
index_endpoint.resource_name
for index_endpoint in aiplatform.MatchingEngineIndexEndpoint.list()
]
assert public_index_endpoint.labels == _TEST_LABELS
assert (
public_index_endpoint.display_name
== _TEST_PUBLIC_INDEX_ENDPOINT_DISPLAY_NAME
)
assert (
public_index_endpoint.description == _TEST_PUBLIC_INDEX_ENDPOINT_DESCRIPTION
)
shared_state["resources"].append(my_index_endpoint)
# Deploy endpoint
my_index_endpoint = my_index_endpoint.deploy_index(
index=index,
deployed_index_id=_TEST_DEPLOYED_INDEX_ID,
display_name=_TEST_DEPLOYED_INDEX_DISPLAY_NAME,
)
# Deploy public endpoint
public_index_endpoint = public_index_endpoint.deploy_index(
index=index,
deployed_index_id=_TEST_DEPLOYED_INDEX_ID_PUBLIC,
display_name=_TEST_DEPLOYED_INDEX_DISPLAY_NAME_PUBLIC,
min_replica_count=_TEST_MIN_REPLICA_COUNT_UPDATED,
max_replica_count=_TEST_MAX_REPLICA_COUNT_UPDATED,
)
# Update endpoint
updated_index_endpoint = my_index_endpoint.update(
display_name=_TEST_DISPLAY_NAME_UPDATE,
description=_TEST_DESCRIPTION_UPDATE,
labels=_TEST_LABELS_UPDATE,
)
assert updated_index_endpoint.labels == _TEST_LABELS_UPDATE
assert updated_index_endpoint.display_name == _TEST_DISPLAY_NAME_UPDATE
assert updated_index_endpoint.description == _TEST_DESCRIPTION_UPDATE
# Mutate deployed index
my_index_endpoint.mutate_deployed_index(
deployed_index_id=_TEST_DEPLOYED_INDEX_ID,
min_replica_count=_TEST_MIN_REPLICA_COUNT_UPDATED,
max_replica_count=_TEST_MAX_REPLICA_COUNT_UPDATED,
)
# deployed index on private endpoint.
deployed_index = my_index_endpoint.deployed_indexes[0]
assert deployed_index.id == _TEST_DEPLOYED_INDEX_ID
assert deployed_index.index == index.resource_name
assert (
deployed_index.automatic_resources.min_replica_count
== _TEST_MIN_REPLICA_COUNT_UPDATED
)
assert (
deployed_index.automatic_resources.max_replica_count
== _TEST_MAX_REPLICA_COUNT_UPDATED
)
# deployed index on public endpoint.
deployed_index_public = public_index_endpoint.deployed_indexes[0]
assert deployed_index_public.id == _TEST_DEPLOYED_INDEX_ID_PUBLIC
assert deployed_index_public.index == index.resource_name
assert (
deployed_index_public.automatic_resources.min_replica_count
== _TEST_MIN_REPLICA_COUNT_UPDATED
)
assert (
deployed_index_public.automatic_resources.max_replica_count
== _TEST_MAX_REPLICA_COUNT_UPDATED
)
# TODO: Test `my_index_endpoint.match` request. This requires running this test in a VPC.
# results = my_index_endpoint.match(
# deployed_index_id=_TEST_DEPLOYED_INDEX_ID, queries=[_TEST_MATCH_QUERY]
# )
# assert results[0][0].id == 870
# TODO: Test `my_index_endpoint.match` with filter.
# This requires uploading a new content of the Matching Engine Index to Cloud Storage.
# results = my_index_endpoint.match(
# deployed_index_id=_TEST_DEPLOYED_INDEX_ID,
# queries=[_TEST_MATCH_QUERY],
# num_neighbors=1,
# filter=_TEST_FILTER,
# )
# assert results[0][0].id == 9999
# Undeploy index
my_index_endpoint = my_index_endpoint.undeploy_index(
deployed_index_id=deployed_index.id
)
# Delete index and check that it is no longer listed
> index.delete()
tests/system/aiplatform/test_matching_engine_index.py:356:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
google/cloud/aiplatform/base.py:814: in wrapper
return method(*args, **kwargs)
google/cloud/aiplatform/base.py:1269: in delete
lro = getattr(self.api_client, self._delete_method)(name=self.resource_name)
google/cloud/aiplatform_v1/services/index_service/client.py:1071: in delete_index
response = rpc(
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/gapic_v1/method.py:145: in __call__
return wrapped_func(*args, **kwargs)
.nox/system-3-8/lib/python3.8/site-packages/google/api_core/grpc_helpers.py:69: in error_remapped_callable
six.raise_from(exceptions.from_grpc_error(exc), exc)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
value = None
from_value = <_InactiveRpcError of RPC that terminated with:
status = StatusCode.FAILED_PRECONDITION
details = "The Index "projec... IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.","grpc_status":9}"
>
> ???
E google.api_core.exceptions.FailedPrecondition: 400 The Index "projects/580378083368/locations/us-central1/indexes/2091385465239240704" is deployed or being deployed at the following IndexEndpoint(s): projects/580378083368/locations/us-central1/indexEndpoints/1288337356683739136.
<string>:3: FailedPrecondition</pre></details>
|
non_defect
|
tests system aiplatform test matching engine index testmatchingengine test create get list matching engine index failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output args name projects locations us indexes kwargs metadata six wraps callable def error remapped callable args kwargs try return callable args kwargs nox system lib site packages google api core grpc helpers py self request name projects locations us indexes timeout none metadata credentials none wait for ready none compression none def call self request timeout none metadata none credentials none wait for ready none compression none state call self blocking request timeout metadata credentials wait for ready compression return end unary response blocking state call false none nox system lib site packages grpc channel py state call with call false deadline none def end unary response blocking state call with call deadline if state code is grpc statuscode ok if with call rendezvous multithreadedrendezvous state call none deadline return state response rendezvous else return state response else raise inactiverpcerror state e grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with e status statuscode failed precondition e details the index projects locations us indexes is deployed or being deployed at the following indexendpoint s projects locations us indexendpoints e debug error string created description error received from peer file src core lib surface call cc file line grpc message the index projects locations us indexes is deployed or being deployed at the following indexendpoint s projects locations us indexendpoints grpc status e nox system lib site packages grpc channel py inactiverpcerror the above exception was the direct cause of the following exception self shared state index ect at resource name projects locations us indexendpoints def test create get list matching engine index self shared state aiplatform init project base project location base location create an index index aiplatform matchingengineindex create tree ah index display name test index display name contents delta uri test contents delta uri dimensions test index config dimensions approximate neighbors count test index approximate neighbors count distance measure type test index distance measure type leaf node embedding count test leaf node embedding count leaf nodes to search percent test leaf nodes to search percent description test index description labels test labels shared state shared state index shared state index resource name verify that the retrieved index is the same get index aiplatform matchingengineindex index name index resource name assert index resource name get index resource name create index and check that it is listed list indexes aiplatform matchingengineindex list assert get index resource name in index resource name for index in list indexes update the index metadata updated index get index update metadata display name test display name update description test description update labels test labels update assert updated index name get index name todo reinstate assertions once b is fixed assert updated index display name test display name update assert updated index description test description update assert updated index labels test labels update update the index embeddings updated index get index update embeddings contents delta uri test contents delta uri update is complete overwrite test is complete overwrite assert updated index name get index name create endpoint and check that it is listed my index endpoint aiplatform matchingengineindexendpoint create display name test index endpoint display name description test index endpoint description network base vpc network uri labels test labels assert my index endpoint resource name in index endpoint resource name for index endpoint in aiplatform matchingengineindexendpoint list assert my index endpoint labels test labels assert my index endpoint display name test index endpoint display name assert my index endpoint description test index endpoint description create endpoint and check that it is listed public index endpoint aiplatform matchingengineindexendpoint create display name test public index endpoint display name description test public index endpoint description public endpoint enabled true labels test labels assert public index endpoint resource name in index endpoint resource name for index endpoint in aiplatform matchingengineindexendpoint list assert public index endpoint labels test labels assert public index endpoint display name test public index endpoint display name assert public index endpoint description test public index endpoint description shared state append my index endpoint deploy endpoint my index endpoint my index endpoint deploy index index index deployed index id test deployed index id display name test deployed index display name deploy public endpoint public index endpoint public index endpoint deploy index index index deployed index id test deployed index id public display name test deployed index display name public min replica count test min replica count updated max replica count test max replica count updated update endpoint updated index endpoint my index endpoint update display name test display name update description test description update labels test labels update assert updated index endpoint labels test labels update assert updated index endpoint display name test display name update assert updated index endpoint description test description update mutate deployed index my index endpoint mutate deployed index deployed index id test deployed index id min replica count test min replica count updated max replica count test max replica count updated deployed index on private endpoint deployed index my index endpoint deployed indexes assert deployed index id test deployed index id assert deployed index index index resource name assert deployed index automatic resources min replica count test min replica count updated assert deployed index automatic resources max replica count test max replica count updated deployed index on public endpoint deployed index public public index endpoint deployed indexes assert deployed index public id test deployed index id public assert deployed index public index index resource name assert deployed index public automatic resources min replica count test min replica count updated assert deployed index public automatic resources max replica count test max replica count updated todo test my index endpoint match request this requires running this test in a vpc results my index endpoint match deployed index id test deployed index id queries assert results id todo test my index endpoint match with filter this requires uploading a new content of the matching engine index to cloud storage results my index endpoint match deployed index id test deployed index id queries num neighbors filter test filter assert results id undeploy index my index endpoint my index endpoint undeploy index deployed index id deployed index id delete index and check that it is no longer listed index delete tests system aiplatform test matching engine index py google cloud aiplatform base py in wrapper return method args kwargs google cloud aiplatform base py in delete lro getattr self api client self delete method name self resource name google cloud aiplatform services index service client py in delete index response rpc nox system lib site packages google api core gapic method py in call return wrapped func args kwargs nox system lib site packages google api core grpc helpers py in error remapped callable six raise from exceptions from grpc error exc exc value none from value inactiverpcerror of rpc that terminated with status statuscode failed precondition details the index projec indexendpoint s projects locations us indexendpoints grpc status e google api core exceptions failedprecondition the index projects locations us indexes is deployed or being deployed at the following indexendpoint s projects locations us indexendpoints failedprecondition
| 0
|
233,607
| 7,700,728,397
|
IssuesEvent
|
2018-05-20 06:00:43
|
python/mypy
|
https://api.github.com/repos/python/mypy
|
closed
|
"SupportsInt" doesn't work with Union types
|
bug false-positive priority-1-normal topic-union-types
|
The following shouldn't be an error:
`error: Argument 1 to "int" has incompatible type "Union[unicode, int]"; expected "SupportsInt"`
|
1.0
|
"SupportsInt" doesn't work with Union types - The following shouldn't be an error:
`error: Argument 1 to "int" has incompatible type "Union[unicode, int]"; expected "SupportsInt"`
|
non_defect
|
supportsint doesn t work with union types the following shouldn t be an error error argument to int has incompatible type union expected supportsint
| 0
|
142,373
| 19,090,476,180
|
IssuesEvent
|
2021-11-29 11:30:31
|
sultanabubaker/NuGet_Project_SDK_NonSDK
|
https://api.github.com/repos/sultanabubaker/NuGet_Project_SDK_NonSDK
|
opened
|
CVE-2019-1428 (High) detected in multiple libraries
|
security vulnerability
|
## CVE-2019-1428 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ChakraCore-1.8.1.0.dll</b>, <b>microsoft.chakracore.1.8.1.nupkg</b>, <b>ChakraCore-1.8.1.0.dll</b>, <b>ChakraCore-1.8.1.0.dll</b></p></summary>
<p>
<details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary>
<p>Microsoft ? Chakra Core</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p>
<p>Path to vulnerable library: /SDK/bin/Debug/net5.0/runtimes/win8-arm/native/ChakraCore.dll,/SDK/bin/Debug/netcoreapp3.1/runtimes/win8-arm/native/ChakraCore.dll</p>
<p>
Dependency Hierarchy:
- :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library)
</details>
<details><summary><b>microsoft.chakracore.1.8.1.nupkg</b></p></summary>
<p>ChakraCore is the core part of the Chakra Javascript engine that powers Microsoft Edge.</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p>
<p>Path to dependency file: NuGet_Project_SDK_NonSDK/SDK/SDK.csproj</p>
<p>Path to vulnerable library: crosoft.chakracore/1.8.1/microsoft.chakracore.1.8.1.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **microsoft.chakracore.1.8.1.nupkg** (Vulnerable Library)
</details>
<details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary>
<p>Microsoft ? Chakra Core</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p>
<p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x64/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x64/native/ChakraCore.dll</p>
<p>
Dependency Hierarchy:
- :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library)
</details>
<details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary>
<p>Microsoft ? Chakra Core</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p>
<p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x86/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x86/native/ChakraCore.dll</p>
<p>
Dependency Hierarchy:
- :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/NuGet_Project_SDK_NonSDK/commit/2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927">2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in Microsoft Edge (HTML-based), aka 'Scripting Engine Memory Corruption Vulnerability'. This CVE ID is unique from CVE-2019-1426, CVE-2019-1427, CVE-2019-1429.
<p>Publish Date: 2019-11-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1428>CVE-2019-1428</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/Microsoft/ChakraCore/wiki/Roadmap">https://github.com/Microsoft/ChakraCore/wiki/Roadmap</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: Microsoft.ChakraCore - 1.11.15</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true},{"packageType":"Nuget","packageName":"Microsoft.ChakraCore","packageVersion":"1.8.1","packageFilePaths":["/SDK/SDK.csproj"],"isTransitiveDependency":false,"dependencyTree":"Microsoft.ChakraCore:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":false},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-1428","vulnerabilityDetails":"A remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in Microsoft Edge (HTML-based), aka \u0027Scripting Engine Memory Corruption Vulnerability\u0027. This CVE ID is unique from CVE-2019-1426, CVE-2019-1427, CVE-2019-1429.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1428","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2019-1428 (High) detected in multiple libraries - ## CVE-2019-1428 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ChakraCore-1.8.1.0.dll</b>, <b>microsoft.chakracore.1.8.1.nupkg</b>, <b>ChakraCore-1.8.1.0.dll</b>, <b>ChakraCore-1.8.1.0.dll</b></p></summary>
<p>
<details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary>
<p>Microsoft ? Chakra Core</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p>
<p>Path to vulnerable library: /SDK/bin/Debug/net5.0/runtimes/win8-arm/native/ChakraCore.dll,/SDK/bin/Debug/netcoreapp3.1/runtimes/win8-arm/native/ChakraCore.dll</p>
<p>
Dependency Hierarchy:
- :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library)
</details>
<details><summary><b>microsoft.chakracore.1.8.1.nupkg</b></p></summary>
<p>ChakraCore is the core part of the Chakra Javascript engine that powers Microsoft Edge.</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p>
<p>Path to dependency file: NuGet_Project_SDK_NonSDK/SDK/SDK.csproj</p>
<p>Path to vulnerable library: crosoft.chakracore/1.8.1/microsoft.chakracore.1.8.1.nupkg</p>
<p>
Dependency Hierarchy:
- :x: **microsoft.chakracore.1.8.1.nupkg** (Vulnerable Library)
</details>
<details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary>
<p>Microsoft ? Chakra Core</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p>
<p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x64/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x64/native/ChakraCore.dll</p>
<p>
Dependency Hierarchy:
- :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library)
</details>
<details><summary><b>ChakraCore-1.8.1.0.dll</b></p></summary>
<p>Microsoft ? Chakra Core</p>
<p>Library home page: <a href="https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg">https://api.nuget.org/packages/microsoft.chakracore.1.8.1.nupkg</a></p>
<p>Path to vulnerable library: /SDK/bin/Debug/netcoreapp3.1/runtimes/win7-x86/native/ChakraCore.dll,/SDK/bin/Debug/net5.0/runtimes/win7-x86/native/ChakraCore.dll</p>
<p>
Dependency Hierarchy:
- :x: **ChakraCore-1.8.1.0.dll** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/sultanabubaker/NuGet_Project_SDK_NonSDK/commit/2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927">2cdcbe42d2efe636b5e9b1d4c29c9da6e2c9b927</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
A remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in Microsoft Edge (HTML-based), aka 'Scripting Engine Memory Corruption Vulnerability'. This CVE ID is unique from CVE-2019-1426, CVE-2019-1427, CVE-2019-1429.
<p>Publish Date: 2019-11-12
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1428>CVE-2019-1428</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: Required
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/Microsoft/ChakraCore/wiki/Roadmap">https://github.com/Microsoft/ChakraCore/wiki/Roadmap</a></p>
<p>Release Date: 2020-08-24</p>
<p>Fix Resolution: Microsoft.ChakraCore - 1.11.15</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true},{"packageType":"Nuget","packageName":"Microsoft.ChakraCore","packageVersion":"1.8.1","packageFilePaths":["/SDK/SDK.csproj"],"isTransitiveDependency":false,"dependencyTree":"Microsoft.ChakraCore:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":false},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true},{"packageType":".NET","packageName":"ChakraCore","packageVersion":"1.8.1.0","packageFilePaths":[],"isTransitiveDependency":false,"dependencyTree":"ChakraCore:1.8.1.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"Microsoft.ChakraCore - 1.11.15","isBinary":true}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2019-1428","vulnerabilityDetails":"A remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in Microsoft Edge (HTML-based), aka \u0027Scripting Engine Memory Corruption Vulnerability\u0027. This CVE ID is unique from CVE-2019-1426, CVE-2019-1427, CVE-2019-1429.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-1428","cvss3Severity":"high","cvss3Score":"7.5","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"Required","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries chakracore dll microsoft chakracore nupkg chakracore dll chakracore dll chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes arm native chakracore dll sdk bin debug runtimes arm native chakracore dll dependency hierarchy x chakracore dll vulnerable library microsoft chakracore nupkg chakracore is the core part of the chakra javascript engine that powers microsoft edge library home page a href path to dependency file nuget project sdk nonsdk sdk sdk csproj path to vulnerable library crosoft chakracore microsoft chakracore nupkg dependency hierarchy x microsoft chakracore nupkg vulnerable library chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes native chakracore dll sdk bin debug runtimes native chakracore dll dependency hierarchy x chakracore dll vulnerable library chakracore dll microsoft chakra core library home page a href path to vulnerable library sdk bin debug runtimes native chakracore dll sdk bin debug runtimes native chakracore dll dependency hierarchy x chakracore dll vulnerable library found in head commit a href found in base branch master vulnerability details a remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in microsoft edge html based aka scripting engine memory corruption vulnerability this cve id is unique from cve cve cve publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution microsoft chakracore isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion microsoft chakracore isbinary true packagetype nuget packagename microsoft chakracore packageversion packagefilepaths istransitivedependency false dependencytree microsoft chakracore isminimumfixversionavailable true minimumfixversion microsoft chakracore isbinary false packagetype net packagename chakracore packageversion packagefilepaths istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion microsoft chakracore isbinary true packagetype net packagename chakracore packageversion packagefilepaths istransitivedependency false dependencytree chakracore isminimumfixversionavailable true minimumfixversion microsoft chakracore isbinary true basebranches vulnerabilityidentifier cve vulnerabilitydetails a remote code execution vulnerability exists in the way that the scripting engine handles objects in memory in microsoft edge html based aka engine memory corruption vulnerability this cve id is unique from cve cve cve vulnerabilityurl
| 0
|
161,713
| 12,559,414,919
|
IssuesEvent
|
2020-06-07 18:53:01
|
GTNewHorizons/GT-New-Horizons-Modpack
|
https://api.github.com/repos/GTNewHorizons/GT-New-Horizons-Modpack
|
closed
|
Crash with 512V Creative Energy Buffer
|
FixedInDev GT++ Issue need to be tested
|
#### Which modpack version are you using?
2.0.8.3
#
#### What did you try to do, and what did you expect to happen?
Change the output of the Buffer with a screwdriver, to cycle back to ULV from MAX.
#
#### What happened instead? (Attach screenshots if needed)
It crashed.
#
#### What do you suggest instead/what changes do you propose?
Fix the crash. Do you need a log? It seems easy to reproduce so I'm not sure you needed one.
|
1.0
|
Crash with 512V Creative Energy Buffer - #### Which modpack version are you using?
2.0.8.3
#
#### What did you try to do, and what did you expect to happen?
Change the output of the Buffer with a screwdriver, to cycle back to ULV from MAX.
#
#### What happened instead? (Attach screenshots if needed)
It crashed.
#
#### What do you suggest instead/what changes do you propose?
Fix the crash. Do you need a log? It seems easy to reproduce so I'm not sure you needed one.
|
non_defect
|
crash with creative energy buffer which modpack version are you using what did you try to do and what did you expect to happen change the output of the buffer with a screwdriver to cycle back to ulv from max what happened instead attach screenshots if needed it crashed what do you suggest instead what changes do you propose fix the crash do you need a log it seems easy to reproduce so i m not sure you needed one
| 0
|
193,061
| 6,877,817,527
|
IssuesEvent
|
2017-11-20 09:36:07
|
OpenNebula/one
|
https://api.github.com/repos/OpenNebula/one
|
opened
|
euca-describe-instances returns UnsupportedOperation
|
Priority: Normal Status: Pending Tracker: Bug
|
---
Author Name: **Valerio Schiavoni** (Valerio Schiavoni)
Original Redmine Issue: 4244, https://dev.opennebula.org/issues/4244
Original Date: 2015-12-14
---
We get this error in scone-server.log
@
Mon Dec 14 14:08:04 2015 [E]: Unexpected error executing code for particular method, detected by Xmlrpc-c method registry code. Method did not fail; rather, it did not complete at all. 25637787-byte supposed UTF-8 string is not valid UTF-8. Unrecognized UTF-8 initial byte value 0x93
@
|
1.0
|
euca-describe-instances returns UnsupportedOperation - ---
Author Name: **Valerio Schiavoni** (Valerio Schiavoni)
Original Redmine Issue: 4244, https://dev.opennebula.org/issues/4244
Original Date: 2015-12-14
---
We get this error in scone-server.log
@
Mon Dec 14 14:08:04 2015 [E]: Unexpected error executing code for particular method, detected by Xmlrpc-c method registry code. Method did not fail; rather, it did not complete at all. 25637787-byte supposed UTF-8 string is not valid UTF-8. Unrecognized UTF-8 initial byte value 0x93
@
|
non_defect
|
euca describe instances returns unsupportedoperation author name valerio schiavoni valerio schiavoni original redmine issue original date we get this error in scone server log mon dec unexpected error executing code for particular method detected by xmlrpc c method registry code method did not fail rather it did not complete at all byte supposed utf string is not valid utf unrecognized utf initial byte value
| 0
|
44,960
| 12,493,538,777
|
IssuesEvent
|
2020-06-01 09:27:44
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
opened
|
User Code Deployment it’s not capable of resolving the package private classes and interfaces
|
Team: Core Type: Critical Type: Defect
|
User Code Deployment it’s not capable of resolving the package private classes and interfaces. We found this while developing the Map Store Support for the Hazelcast Cloud. When some class inside the Hikari Connection Pool references package a private stuff - it fails with the IllegalAccessError on the attempt to load it:
```
{
"errorMessage": "class com.zaxxer.hikari.pool.PoolBase$NopMetricsTrackerDelegate cannot access its superinterface com.zaxxer.hikari.pool.PoolBase$IMetricsTrackerDelegate",
"errorType": "java.lang.IllegalAccessError",
"stackTrace": [
"java.lang.ClassLoader.defineClass1(Native Method)",
"java.lang.ClassLoader.defineClass(ClassLoader.java:763)",
"java.lang.ClassLoader.defineClass(ClassLoader.java:642)",
"com.hazelcast.internal.usercodedeployment.impl.ClassSource.define(ClassSource.java:50)",
"com.hazelcast.internal.usercodedeployment.impl.ClassLocator.tryToGetClassFromRemote(ClassLocator.java:163)",
"com.hazelcast.internal.usercodedeployment.impl.ClassLocator.handleClassNotFoundException(ClassLocator.java:95)",
```
**Now, this fully blocks the end-to-end scenario working at our side.**
**Expected behavior**
The expected behavior is that the uploaded JAR, with all its dependencies, will be correctly loaded and the Map Store instance will be successfully created and initialized.
**To Reproduce**
See the branch with reproducer - https://github.com/hazelcast/hazelcast-cloud-code-samples/tree/hazelcast.code-deployment.issue. Steps to reproduce right here.
1. Clone the repo.
2. Checkout branch `hazelcast.code-deployment.issue`
3. cd `mapstore`
4. Open README.md
**Additional context**
Hazelcast ver. 3.12.6
In our case, we share the uploaded JARs through the lite members.
When we inspected the source code, we noticed that for every loaded class it creates a separate class loader. So we suspect this can be causing this issue.
|
1.0
|
User Code Deployment it’s not capable of resolving the package private classes and interfaces - User Code Deployment it’s not capable of resolving the package private classes and interfaces. We found this while developing the Map Store Support for the Hazelcast Cloud. When some class inside the Hikari Connection Pool references package a private stuff - it fails with the IllegalAccessError on the attempt to load it:
```
{
"errorMessage": "class com.zaxxer.hikari.pool.PoolBase$NopMetricsTrackerDelegate cannot access its superinterface com.zaxxer.hikari.pool.PoolBase$IMetricsTrackerDelegate",
"errorType": "java.lang.IllegalAccessError",
"stackTrace": [
"java.lang.ClassLoader.defineClass1(Native Method)",
"java.lang.ClassLoader.defineClass(ClassLoader.java:763)",
"java.lang.ClassLoader.defineClass(ClassLoader.java:642)",
"com.hazelcast.internal.usercodedeployment.impl.ClassSource.define(ClassSource.java:50)",
"com.hazelcast.internal.usercodedeployment.impl.ClassLocator.tryToGetClassFromRemote(ClassLocator.java:163)",
"com.hazelcast.internal.usercodedeployment.impl.ClassLocator.handleClassNotFoundException(ClassLocator.java:95)",
```
**Now, this fully blocks the end-to-end scenario working at our side.**
**Expected behavior**
The expected behavior is that the uploaded JAR, with all its dependencies, will be correctly loaded and the Map Store instance will be successfully created and initialized.
**To Reproduce**
See the branch with reproducer - https://github.com/hazelcast/hazelcast-cloud-code-samples/tree/hazelcast.code-deployment.issue. Steps to reproduce right here.
1. Clone the repo.
2. Checkout branch `hazelcast.code-deployment.issue`
3. cd `mapstore`
4. Open README.md
**Additional context**
Hazelcast ver. 3.12.6
In our case, we share the uploaded JARs through the lite members.
When we inspected the source code, we noticed that for every loaded class it creates a separate class loader. So we suspect this can be causing this issue.
|
defect
|
user code deployment it’s not capable of resolving the package private classes and interfaces user code deployment it’s not capable of resolving the package private classes and interfaces we found this while developing the map store support for the hazelcast cloud when some class inside the hikari connection pool references package a private stuff it fails with the illegalaccesserror on the attempt to load it errormessage class com zaxxer hikari pool poolbase nopmetricstrackerdelegate cannot access its superinterface com zaxxer hikari pool poolbase imetricstrackerdelegate errortype java lang illegalaccesserror stacktrace java lang classloader native method java lang classloader defineclass classloader java java lang classloader defineclass classloader java com hazelcast internal usercodedeployment impl classsource define classsource java com hazelcast internal usercodedeployment impl classlocator trytogetclassfromremote classlocator java com hazelcast internal usercodedeployment impl classlocator handleclassnotfoundexception classlocator java now this fully blocks the end to end scenario working at our side expected behavior the expected behavior is that the uploaded jar with all its dependencies will be correctly loaded and the map store instance will be successfully created and initialized to reproduce see the branch with reproducer steps to reproduce right here clone the repo checkout branch hazelcast code deployment issue cd mapstore open readme md additional context hazelcast ver in our case we share the uploaded jars through the lite members when we inspected the source code we noticed that for every loaded class it creates a separate class loader so we suspect this can be causing this issue
| 1
|
244,228
| 18,751,236,553
|
IssuesEvent
|
2021-11-05 02:24:32
|
AY2122S1-CS2103T-T13-2/tp
|
https://api.github.com/repos/AY2122S1-CS2103T-T13-2/tp
|
closed
|
[PE-D] No "uncompleted status"
|
documentation
|
Assignments only have "pending" status, not "uncompleted" status.

<!--session: 1635501358553-926ea394-739d-48cb-b26f-eaa548f33f4d-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: Th-429B/ped#10
|
1.0
|
[PE-D] No "uncompleted status" - Assignments only have "pending" status, not "uncompleted" status.

<!--session: 1635501358553-926ea394-739d-48cb-b26f-eaa548f33f4d-->
<!--Version: Web v3.4.1-->
-------------
Labels: `severity.Low` `type.DocumentationBug`
original: Th-429B/ped#10
|
non_defect
|
no uncompleted status assignments only have pending status not uncompleted status labels severity low type documentationbug original th ped
| 0
|
75,839
| 26,090,377,439
|
IssuesEvent
|
2022-12-26 10:27:47
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
BUG: "xl" not returned if success = False for scipy.optimize.shgo
|
defect scipy.optimize
|
### Describe your issue.
The lists `xl` and `funl` are not returned if the optimization did not return with `success = True`. It would seem that these lists should be populated and returned for access to the user. The code below is taken from the documentation with `n` changed to a small number such that the issue is reproduced.
### Reproducing Code Example
```python
import scipy.optimize as sopt
def eggholder(x):
return (-(x[1] + 47.0)
* np.sin(np.sqrt(abs(x[0]/2.0 + (x[1] + 47.0))))
- x[0] * np.sin(np.sqrt(abs(x[0] - (x[1] + 47.0))))
)
bounds = [(-512, 512), (-512, 512)]
result = sopt.shgo(eggholder, bounds, n=2, sampling_method='sobol')
```
### Error message
```shell
result
Out[27]:
fun: -25.460337185286313
message: 'Failed to find a feasible minimizer point. Lowest sampling point = -25.460337185286313'
nfev: 2
nit: 2
nlfev: 0
nlhev: 0
nljev: 0
success: False
x: array([0., 0.])
```
### SciPy/NumPy/Python version information
1.9.3
|
1.0
|
BUG: "xl" not returned if success = False for scipy.optimize.shgo - ### Describe your issue.
The lists `xl` and `funl` are not returned if the optimization did not return with `success = True`. It would seem that these lists should be populated and returned for access to the user. The code below is taken from the documentation with `n` changed to a small number such that the issue is reproduced.
### Reproducing Code Example
```python
import scipy.optimize as sopt
def eggholder(x):
return (-(x[1] + 47.0)
* np.sin(np.sqrt(abs(x[0]/2.0 + (x[1] + 47.0))))
- x[0] * np.sin(np.sqrt(abs(x[0] - (x[1] + 47.0))))
)
bounds = [(-512, 512), (-512, 512)]
result = sopt.shgo(eggholder, bounds, n=2, sampling_method='sobol')
```
### Error message
```shell
result
Out[27]:
fun: -25.460337185286313
message: 'Failed to find a feasible minimizer point. Lowest sampling point = -25.460337185286313'
nfev: 2
nit: 2
nlfev: 0
nlhev: 0
nljev: 0
success: False
x: array([0., 0.])
```
### SciPy/NumPy/Python version information
1.9.3
|
defect
|
bug xl not returned if success false for scipy optimize shgo describe your issue the lists xl and funl are not returned if the optimization did not return with success true it would seem that these lists should be populated and returned for access to the user the code below is taken from the documentation with n changed to a small number such that the issue is reproduced reproducing code example python import scipy optimize as sopt def eggholder x return x np sin np sqrt abs x x x np sin np sqrt abs x x bounds result sopt shgo eggholder bounds n sampling method sobol error message shell result out fun message failed to find a feasible minimizer point lowest sampling point nfev nit nlfev nlhev nljev success false x array scipy numpy python version information
| 1
|
297,223
| 25,710,885,929
|
IssuesEvent
|
2022-12-07 06:28:16
|
ByteOtter/OtterDen
|
https://api.github.com/repos/ByteOtter/OtterDen
|
closed
|
[BUG:]QA/QE: Broken import statement in environment.py
|
bug testsuite behave/cucumber
|
# Found an issue? Please let us know!
when trying to launch any testscenario, behave throws the error "module test_utils not found" despite the directory being where it should be. This directory holds all necessary utility functions selenium needs to set up correctly so not importing it borks the testsuite.
### Please tick the applicable box(es) regarding the kind of bug you found:
- [x] Broken feature
- [ ] Login/Logout system
- [ ] Database broken
- [ ] Post-related problem
- [ ] Unusable route
- [ ] UI Bug
- [ ] Wrong documentation
- [ ] Broken dependency
### Steps to reproduce
1. try to run behave testsuite/features/core/init.py
2. get greeted by error message
### Links
This issue is connected to Issue [#53 ](https://github.com/ByteOtter/OtterDen/issues/53)
### Additional context
Stacktrace:
```
Exception ModuleNotFoundError: No module named 'test_utils'
Traceback (most recent call last):
File "/home/chris/Dev/PrivateProjects/LogBlog/env/bin/behave", line 8, in <module>
sys.exit(main())
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/__main__.py", line 183, in main
return run_behave(config)
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/__main__.py", line 127, in run_behave
failed = runner.run()
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 804, in run
return self.run_with_paths()
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 808, in run_with_paths
self.load_hooks()
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 784, in load_hooks
exec_file(hooks_path, self.hooks)
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner_util.py", line 386, in exec_file
exec(code, globals_, locals_)
File "testsuite/features/environment.py", line 10, in <module>
from test_utils.utils_web import get_browser
ModuleNotFoundError: No module named 'test_utils'
```
|
1.0
|
[BUG:]QA/QE: Broken import statement in environment.py - # Found an issue? Please let us know!
when trying to launch any testscenario, behave throws the error "module test_utils not found" despite the directory being where it should be. This directory holds all necessary utility functions selenium needs to set up correctly so not importing it borks the testsuite.
### Please tick the applicable box(es) regarding the kind of bug you found:
- [x] Broken feature
- [ ] Login/Logout system
- [ ] Database broken
- [ ] Post-related problem
- [ ] Unusable route
- [ ] UI Bug
- [ ] Wrong documentation
- [ ] Broken dependency
### Steps to reproduce
1. try to run behave testsuite/features/core/init.py
2. get greeted by error message
### Links
This issue is connected to Issue [#53 ](https://github.com/ByteOtter/OtterDen/issues/53)
### Additional context
Stacktrace:
```
Exception ModuleNotFoundError: No module named 'test_utils'
Traceback (most recent call last):
File "/home/chris/Dev/PrivateProjects/LogBlog/env/bin/behave", line 8, in <module>
sys.exit(main())
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/__main__.py", line 183, in main
return run_behave(config)
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/__main__.py", line 127, in run_behave
failed = runner.run()
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 804, in run
return self.run_with_paths()
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 808, in run_with_paths
self.load_hooks()
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner.py", line 784, in load_hooks
exec_file(hooks_path, self.hooks)
File "/home/chris/Dev/PrivateProjects/LogBlog/env/lib64/python3.10/site-packages/behave/runner_util.py", line 386, in exec_file
exec(code, globals_, locals_)
File "testsuite/features/environment.py", line 10, in <module>
from test_utils.utils_web import get_browser
ModuleNotFoundError: No module named 'test_utils'
```
|
non_defect
|
qa qe broken import statement in environment py found an issue please let us know when trying to launch any testscenario behave throws the error module test utils not found despite the directory being where it should be this directory holds all necessary utility functions selenium needs to set up correctly so not importing it borks the testsuite please tick the applicable box es regarding the kind of bug you found broken feature login logout system database broken post related problem unusable route ui bug wrong documentation broken dependency steps to reproduce try to run behave testsuite features core init py get greeted by error message links this issue is connected to issue additional context stacktrace exception modulenotfounderror no module named test utils traceback most recent call last file home chris dev privateprojects logblog env bin behave line in sys exit main file home chris dev privateprojects logblog env site packages behave main py line in main return run behave config file home chris dev privateprojects logblog env site packages behave main py line in run behave failed runner run file home chris dev privateprojects logblog env site packages behave runner py line in run return self run with paths file home chris dev privateprojects logblog env site packages behave runner py line in run with paths self load hooks file home chris dev privateprojects logblog env site packages behave runner py line in load hooks exec file hooks path self hooks file home chris dev privateprojects logblog env site packages behave runner util py line in exec file exec code globals locals file testsuite features environment py line in from test utils utils web import get browser modulenotfounderror no module named test utils
| 0
|
66,818
| 20,682,119,457
|
IssuesEvent
|
2022-03-10 14:47:55
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
opened
|
Spotlight search doesn't find unnamed rooms
|
T-Defect
|
### Steps to reproduce
1. I wanted to find my 3-way DM with Amandine and Rick
2. I searched for `Rick` (didn't show up; too many rooms). I searched for `Amandine` (didn't show up; too many rooms). I searched for `amandine rick` (nothing showed up at all)
3. I eventually searched for `Rick` and then manually expanded the results list and went through one by one until i eventually found it
### Outcome
#### What did you expect?
Searching for `amandine rick` should return nameless rooms where the room name has been synthesised to be 'Amandine and rick'.
#### What happened instead?
Nothing showed up.
Turns out that searching for `amandine and rick` found it, fwiw.
### Operating system
macOS
### Browser information
chrome
### URL for webapp
nightly
### Application version
nightly
### Homeserver
matrix.org
### Will you send logs?
No
|
1.0
|
Spotlight search doesn't find unnamed rooms - ### Steps to reproduce
1. I wanted to find my 3-way DM with Amandine and Rick
2. I searched for `Rick` (didn't show up; too many rooms). I searched for `Amandine` (didn't show up; too many rooms). I searched for `amandine rick` (nothing showed up at all)
3. I eventually searched for `Rick` and then manually expanded the results list and went through one by one until i eventually found it
### Outcome
#### What did you expect?
Searching for `amandine rick` should return nameless rooms where the room name has been synthesised to be 'Amandine and rick'.
#### What happened instead?
Nothing showed up.
Turns out that searching for `amandine and rick` found it, fwiw.
### Operating system
macOS
### Browser information
chrome
### URL for webapp
nightly
### Application version
nightly
### Homeserver
matrix.org
### Will you send logs?
No
|
defect
|
spotlight search doesn t find unnamed rooms steps to reproduce i wanted to find my way dm with amandine and rick i searched for rick didn t show up too many rooms i searched for amandine didn t show up too many rooms i searched for amandine rick nothing showed up at all i eventually searched for rick and then manually expanded the results list and went through one by one until i eventually found it outcome what did you expect searching for amandine rick should return nameless rooms where the room name has been synthesised to be amandine and rick what happened instead nothing showed up turns out that searching for amandine and rick found it fwiw operating system macos browser information chrome url for webapp nightly application version nightly homeserver matrix org will you send logs no
| 1
|
79,429
| 28,240,719,951
|
IssuesEvent
|
2023-04-06 06:54:41
|
line/armeria
|
https://api.github.com/repos/line/armeria
|
closed
|
Handle exceptions thrown by Servlet
|
defect
|
> Hi folks, we recently encountered an issue where Armeria (we are on Armeria 1.16.2) won’t finish collecting the response information for a RequestLog when the corresponding HttpResponse is closed without any data written. As a result, RequestLogAccess.whenComplete() never completes after closing the HttpResponse. A simple reproduction can be found [here](https://gist.github.com/linwuz-db/2c7adae5a546c87760c69a89ddccc70e). So far we find this issue is exposed in [JettyService](https://github.com/line/armeria/blob/main/jetty9/src/main/java/com/linecorp/armeria/server/jetty/JettyService.java#L323) and [TomcatService](https://github.com/line/armeria/blob/main/tomcat9/src/main/java/com/linecorp/armeria/server/tomcat/TomcatService.java#L427) when the exception is thrown outside the “servlet handling” code path. We have a couple of questions:
Is it expected behavior that the RequestLog never completes when neither data nor failure is being set on the response (just close)? This leads us to the second question:
Is it a legit use of the API to close a HttpResponse without writing anything or “failing” it.
Slack thread: https://line-armeria.slack.com/archives/C1NGPBUH2/p1679553378276729
If `JettyService` get an exception while handling a request, it closes the response silently without propagating the exception.
https://github.com/line/armeria/blob/49054ad7f22a7f2d64faaeb694674b0df09ad165/jetty9/src/main/java/com/linecorp/armeria/server/jetty/JettyService.java#L322-L324
https://github.com/line/armeria/blob/49054ad7f22a7f2d64faaeb694674b0df09ad165/tomcat9/src/main/java/com/linecorp/armeria/server/tomcat/TomcatService.java#L424-L428
These can make the failed response publish no `HttpObject`s.
https://github.com/line/armeria/blob/cef7b9c359b383e2b703f13c135f7800e980e422/core/src/main/java/com/linecorp/armeria/server/HttpResponseSubscriber.java#L307-L317
As a result, 1) the exception is not propagated to `RequestLog.responseCause()` and 2) the `RequestLog.whenComplete()` isn't completed because of the bug in `HttpResponseSubscriber`.
|
1.0
|
Handle exceptions thrown by Servlet - > Hi folks, we recently encountered an issue where Armeria (we are on Armeria 1.16.2) won’t finish collecting the response information for a RequestLog when the corresponding HttpResponse is closed without any data written. As a result, RequestLogAccess.whenComplete() never completes after closing the HttpResponse. A simple reproduction can be found [here](https://gist.github.com/linwuz-db/2c7adae5a546c87760c69a89ddccc70e). So far we find this issue is exposed in [JettyService](https://github.com/line/armeria/blob/main/jetty9/src/main/java/com/linecorp/armeria/server/jetty/JettyService.java#L323) and [TomcatService](https://github.com/line/armeria/blob/main/tomcat9/src/main/java/com/linecorp/armeria/server/tomcat/TomcatService.java#L427) when the exception is thrown outside the “servlet handling” code path. We have a couple of questions:
Is it expected behavior that the RequestLog never completes when neither data nor failure is being set on the response (just close)? This leads us to the second question:
Is it a legit use of the API to close a HttpResponse without writing anything or “failing” it.
Slack thread: https://line-armeria.slack.com/archives/C1NGPBUH2/p1679553378276729
If `JettyService` get an exception while handling a request, it closes the response silently without propagating the exception.
https://github.com/line/armeria/blob/49054ad7f22a7f2d64faaeb694674b0df09ad165/jetty9/src/main/java/com/linecorp/armeria/server/jetty/JettyService.java#L322-L324
https://github.com/line/armeria/blob/49054ad7f22a7f2d64faaeb694674b0df09ad165/tomcat9/src/main/java/com/linecorp/armeria/server/tomcat/TomcatService.java#L424-L428
These can make the failed response publish no `HttpObject`s.
https://github.com/line/armeria/blob/cef7b9c359b383e2b703f13c135f7800e980e422/core/src/main/java/com/linecorp/armeria/server/HttpResponseSubscriber.java#L307-L317
As a result, 1) the exception is not propagated to `RequestLog.responseCause()` and 2) the `RequestLog.whenComplete()` isn't completed because of the bug in `HttpResponseSubscriber`.
|
defect
|
handle exceptions thrown by servlet hi folks we recently encountered an issue where armeria we are on armeria won’t finish collecting the response information for a requestlog when the corresponding httpresponse is closed without any data written as a result requestlogaccess whencomplete never completes after closing the httpresponse a simple reproduction can be found so far we find this issue is exposed in and when the exception is thrown outside the “servlet handling” code path we have a couple of questions is it expected behavior that the requestlog never completes when neither data nor failure is being set on the response just close this leads us to the second question is it a legit use of the api to close a httpresponse without writing anything or “failing” it slack thread if jettyservice get an exception while handling a request it closes the response silently without propagating the exception these can make the failed response publish no httpobject s as a result the exception is not propagated to requestlog responsecause and the requestlog whencomplete isn t completed because of the bug in httpresponsesubscriber
| 1
|
24,912
| 5,110,408,740
|
IssuesEvent
|
2017-01-06 00:09:40
|
emberjs/ember.js
|
https://api.github.com/repos/emberjs/ember.js
|
closed
|
[DOC] Ember.Component#layout is not documented as public
|
Documentation
|
Inspired by: https://github.com/ember-cli/ember-cli.github.io/issues/100
Currently the [default addon component blueprint](https://github.com/emberjs/ember.js/blob/v2.10.0-beta.2/blueprints/component/index.js#L59-L68) sets up the component template via the `layout` property.
There is no evidence in the [API docs](http://emberjs.com/api/classes/Ember.Component.html) that it's public, or any mention or use of it in [the appropriate section of the guides](https://guides.emberjs.com/v2.9.0/components/defining-a-component/), leading to confusion as to whether it's ok to use. Add in [the deprecated `Ember.View` concept of the same name](http://emberjs.com/api/classes/Ember.View.html#toc_layouts), and it's understandable why people new to the framework are tripping up here.
|
1.0
|
[DOC] Ember.Component#layout is not documented as public - Inspired by: https://github.com/ember-cli/ember-cli.github.io/issues/100
Currently the [default addon component blueprint](https://github.com/emberjs/ember.js/blob/v2.10.0-beta.2/blueprints/component/index.js#L59-L68) sets up the component template via the `layout` property.
There is no evidence in the [API docs](http://emberjs.com/api/classes/Ember.Component.html) that it's public, or any mention or use of it in [the appropriate section of the guides](https://guides.emberjs.com/v2.9.0/components/defining-a-component/), leading to confusion as to whether it's ok to use. Add in [the deprecated `Ember.View` concept of the same name](http://emberjs.com/api/classes/Ember.View.html#toc_layouts), and it's understandable why people new to the framework are tripping up here.
|
non_defect
|
ember component layout is not documented as public inspired by currently the sets up the component template via the layout property there is no evidence in the that it s public or any mention or use of it in leading to confusion as to whether it s ok to use add in and it s understandable why people new to the framework are tripping up here
| 0
|
34,543
| 7,453,494,658
|
IssuesEvent
|
2018-03-29 12:11:55
|
kerdokullamae/test_koik_issued
|
https://api.github.com/repos/kerdokullamae/test_koik_issued
|
closed
|
Javascripti bugi seoste laadimisel
|
P: highest R: fixed T: defect
|
**Reported by jaak tamre on 16 Aug 2013 13:33 UTC**
'''Object'''
KÜ detailvaade
'''Description'''
Kui kirjeldusüksuse detailvaates avada sakk aktid, kohad või märksõnad, siis pärast uue seose lisamist või kustutamist läheb osa javascripti funktsionaalsust lehel katki. XhrLoader ei tööta enam ja akkordioni sinised ribad ei lähe kinni/lahti. Samas kui vajutada "näita kõiki" siis töötab. Äkki mingi live eventide probleem?
'''Todo'''
Uurida probleemi ja lahendada
|
1.0
|
Javascripti bugi seoste laadimisel - **Reported by jaak tamre on 16 Aug 2013 13:33 UTC**
'''Object'''
KÜ detailvaade
'''Description'''
Kui kirjeldusüksuse detailvaates avada sakk aktid, kohad või märksõnad, siis pärast uue seose lisamist või kustutamist läheb osa javascripti funktsionaalsust lehel katki. XhrLoader ei tööta enam ja akkordioni sinised ribad ei lähe kinni/lahti. Samas kui vajutada "näita kõiki" siis töötab. Äkki mingi live eventide probleem?
'''Todo'''
Uurida probleemi ja lahendada
|
defect
|
javascripti bugi seoste laadimisel reported by jaak tamre on aug utc object kü detailvaade description kui kirjeldusüksuse detailvaates avada sakk aktid kohad või märksõnad siis pärast uue seose lisamist või kustutamist läheb osa javascripti funktsionaalsust lehel katki xhrloader ei tööta enam ja akkordioni sinised ribad ei lähe kinni lahti samas kui vajutada näita kõiki siis töötab äkki mingi live eventide probleem todo uurida probleemi ja lahendada
| 1
|
124,445
| 16,603,131,670
|
IssuesEvent
|
2021-06-01 22:36:01
|
microsoft/TypeScript
|
https://api.github.com/repos/microsoft/TypeScript
|
closed
|
Function.prototype.bind typings do not work properly when supplied a function with variadic arguments.
|
Design Limitation
|
# Bug Report
If you pass a function that might behave differently when supplied different arguments, the resulting type does not properly accommodate all call patterns.
I ran into this while trying to figure out how TypeScript handled situations similar to RxJS's `bindNodeCallback` and `bindCallback`, which suffers from similar problems.
### 🔎 Search Terms
- `bind`
- `bind "Arguments for the rest parameter"`
### 🕗 Version & Regression Information
versions 3.3.* - 4.1.3
- This is the behavior in every version I tried, and I reviewed the FAQ for entries about (I have no idea where the FAQ is, reporting anyway, because this is friction)
### ⏯ Playground Link
[Playground link with relevant code](https://www.typescriptlang.org/play?#code/JYOwLgpgTgZghgYwgAgILIN4FgBQz-JwBcyAzmFKAOYDcuAvrqJLIigEKa4HIBGJ5SiFoMm4aPCTIAwlzwEEAitTo5GOZhLbIAInJ4ATJUJFrcuBAHsQ5QiXQBeTHeQByOK+T1VVm2D4knE4YAW68nt4W1raKMsjByLGuCBE+0f5GuvHOma4GqeY4APRFyAAyEGCupGRwAJ7IYAAWEKQocFAozSjWADYNAG5wvcAGyADu9TVglrglicO9yJYArlDIAAaQ5BvIMCsgCGDA1kS4+4fH1o2tYAAUxGgANKHsAJT2NMjzAIznB0cTiAbuQHvYXrFpB9kOwvvMAEz-S5AkH3R6oF78GEvTI6aHSOGlADMhXmACJgABbAAOvTJyBanSRgOu2zR4L4AHkoNJAsgAD4yHEAfhIePsAphktk2Hk+GAMGQd2ApFQDzeb30PAICqVKvYd143Khmtl2vNyF1ytIOjuBg1WotFs6YDWwJCSTZqW+xO4ToI9GQEF6bUd-p4Lrdzkeri9XkJyD+cv96idqYtVpV0kNxodZvDyEjUHdoVjt29CL95vT+BrjSaUEs42QIAgzYAolBG1A7q5xo3hAter1XG9VOpmVdgSq1aBqSswCQ4CA6tC5wvLTV0Pn8EXpyB5-4AGRHtweS37w-IE+NOrUiCWRXrsAAbXcrgAuvEHE5qsphK446FBcLLTqQBrPkuK5rgeG4qlKO6FpUUbPtep6uOEF4XleN5gHeD5PrBr4YZ+36-oI1CAaIOCTiiWbKkRUGriQqHwTKVZIa6xbYRuN7JJ4oA8cep54fej5CW+KRfj+5H-lQVFmDROAgVOm62pBhDQSxRGblkiF7kJaFuPkWGobh+Hic+b75NJP5uBRAFAUpUR+MgAAe2RsgAdLwoAGHcIArMOLxwJiY6FL4tjrE4bl3OFSmRZYvQQF5vSWFQdxQGOQA)
### 💻 Code
```ts
interface A {
a: string;
}
interface B {
b: string;
}
interface C {
c: string;
}
interface D {
d: string;
}
const a: A = { a: 'a' };
const b: B = { b: 'b' };
const c: C = { c: 'c' };
const d: D = { d: 'd' };
// Let's say these are the only valid ways to
// call our `test` function:
function test(a: A, b: B): A; // 1
function test(a: A, c: C): B; // 2
function test(a: A, b: B, d: D): C; // 3
// "impl" here
function test(a: A, bOrC: B | C, d?: D): A | B | C {
if (isA(a)) {
if (isB(bOrC)) {
if (isD(d)) {
return { c: 'test' }; // 3
} else {
return { a: 'test' }; // 1
}
}
if (isC(bOrC)) {
return { b: 'test' }; // 2
}
}
throw new Error('wrong call');
}
function isA(input: any): input is A {
return input && 'a' in input && typeof input['a'] === 'string';
}
function isB(input: any): input is B {
return input && 'b' in input && typeof input['b'] === 'string';
}
function isC(input: any): input is C {
return input && 'c' in input && typeof input['c'] === 'string';
}
function isD(input: any): input is D {
return input && 'd' in input && typeof input['d'] === 'string';
}
const x = test.bind(null, a, b);
const r = x();
console.log(r);
```
### 🙁 Actual behavior

I got a compilation error for a perfectly valid call of my bound function.
### 🙂 Expected behavior
The return type (the type of `r`) should be `C`, and the code should compile successfully.
|
1.0
|
Function.prototype.bind typings do not work properly when supplied a function with variadic arguments. - # Bug Report
If you pass a function that might behave differently when supplied different arguments, the resulting type does not properly accommodate all call patterns.
I ran into this while trying to figure out how TypeScript handled situations similar to RxJS's `bindNodeCallback` and `bindCallback`, which suffers from similar problems.
### 🔎 Search Terms
- `bind`
- `bind "Arguments for the rest parameter"`
### 🕗 Version & Regression Information
versions 3.3.* - 4.1.3
- This is the behavior in every version I tried, and I reviewed the FAQ for entries about (I have no idea where the FAQ is, reporting anyway, because this is friction)
### ⏯ Playground Link
[Playground link with relevant code](https://www.typescriptlang.org/play?#code/JYOwLgpgTgZghgYwgAgILIN4FgBQz-JwBcyAzmFKAOYDcuAvrqJLIigEKa4HIBGJ5SiFoMm4aPCTIAwlzwEEAitTo5GOZhLbIAInJ4ATJUJFrcuBAHsQ5QiXQBeTHeQByOK+T1VVm2D4knE4YAW68nt4W1raKMsjByLGuCBE+0f5GuvHOma4GqeY4APRFyAAyEGCupGRwAJ7IYAAWEKQocFAozSjWADYNAG5wvcAGyADu9TVglrglicO9yJYArlDIAAaQ5BvIMCsgCGDA1kS4+4fH1o2tYAAUxGgANKHsAJT2NMjzAIznB0cTiAbuQHvYXrFpB9kOwvvMAEz-S5AkH3R6oF78GEvTI6aHSOGlADMhXmACJgABbAAOvTJyBanSRgOu2zR4L4AHkoNJAsgAD4yHEAfhIePsAphktk2Hk+GAMGQd2ApFQDzeb30PAICqVKvYd143Khmtl2vNyF1ytIOjuBg1WotFs6YDWwJCSTZqW+xO4ToI9GQEF6bUd-p4Lrdzkeri9XkJyD+cv96idqYtVpV0kNxodZvDyEjUHdoVjt29CL95vT+BrjSaUEs42QIAgzYAolBG1A7q5xo3hAter1XG9VOpmVdgSq1aBqSswCQ4CA6tC5wvLTV0Pn8EXpyB5-4AGRHtweS37w-IE+NOrUiCWRXrsAAbXcrgAuvEHE5qsphK446FBcLLTqQBrPkuK5rgeG4qlKO6FpUUbPtep6uOEF4XleN5gHeD5PrBr4YZ+36-oI1CAaIOCTiiWbKkRUGriQqHwTKVZIa6xbYRuN7JJ4oA8cep54fej5CW+KRfj+5H-lQVFmDROAgVOm62pBhDQSxRGblkiF7kJaFuPkWGobh+Hic+b75NJP5uBRAFAUpUR+MgAAe2RsgAdLwoAGHcIArMOLxwJiY6FL4tjrE4bl3OFSmRZYvQQF5vSWFQdxQGOQA)
### 💻 Code
```ts
interface A {
a: string;
}
interface B {
b: string;
}
interface C {
c: string;
}
interface D {
d: string;
}
const a: A = { a: 'a' };
const b: B = { b: 'b' };
const c: C = { c: 'c' };
const d: D = { d: 'd' };
// Let's say these are the only valid ways to
// call our `test` function:
function test(a: A, b: B): A; // 1
function test(a: A, c: C): B; // 2
function test(a: A, b: B, d: D): C; // 3
// "impl" here
function test(a: A, bOrC: B | C, d?: D): A | B | C {
if (isA(a)) {
if (isB(bOrC)) {
if (isD(d)) {
return { c: 'test' }; // 3
} else {
return { a: 'test' }; // 1
}
}
if (isC(bOrC)) {
return { b: 'test' }; // 2
}
}
throw new Error('wrong call');
}
function isA(input: any): input is A {
return input && 'a' in input && typeof input['a'] === 'string';
}
function isB(input: any): input is B {
return input && 'b' in input && typeof input['b'] === 'string';
}
function isC(input: any): input is C {
return input && 'c' in input && typeof input['c'] === 'string';
}
function isD(input: any): input is D {
return input && 'd' in input && typeof input['d'] === 'string';
}
const x = test.bind(null, a, b);
const r = x();
console.log(r);
```
### 🙁 Actual behavior

I got a compilation error for a perfectly valid call of my bound function.
### 🙂 Expected behavior
The return type (the type of `r`) should be `C`, and the code should compile successfully.
|
non_defect
|
function prototype bind typings do not work properly when supplied a function with variadic arguments bug report if you pass a function that might behave differently when supplied different arguments the resulting type does not properly accommodate all call patterns i ran into this while trying to figure out how typescript handled situations similar to rxjs s bindnodecallback and bindcallback which suffers from similar problems 🔎 search terms bind bind arguments for the rest parameter 🕗 version regression information versions this is the behavior in every version i tried and i reviewed the faq for entries about i have no idea where the faq is reporting anyway because this is friction ⏯ playground link 💻 code ts interface a a string interface b b string interface c c string interface d d string const a a a a const b b b b const c c c c const d d d d let s say these are the only valid ways to call our test function function test a a b b a function test a a c c b function test a a b b d d c impl here function test a a borc b c d d a b c if isa a if isb borc if isd d return c test else return a test if isc borc return b test throw new error wrong call function isa input any input is a return input a in input typeof input string function isb input any input is b return input b in input typeof input string function isc input any input is c return input c in input typeof input string function isd input any input is d return input d in input typeof input string const x test bind null a b const r x console log r 🙁 actual behavior i got a compilation error for a perfectly valid call of my bound function 🙂 expected behavior the return type the type of r should be c and the code should compile successfully
| 0
|
578,306
| 17,146,593,764
|
IssuesEvent
|
2021-07-13 15:11:43
|
GoogleCloudPlatform/python-docs-samples
|
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
|
closed
|
healthcare.api-client.v1.fhir.fhir_resources_test: test_delete_patient failed
|
api: healthcare flakybot: flaky flakybot: issue priority: p1 samples type: bug
|
This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: b99df8d36109e4fe3e397bfd2cbacac06960340c
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/325e6648-2604-4a2d-a611-53de8a0492aa), [Sponge](http://sponge2/325e6648-2604-4a2d-a611-53de8a0492aa)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/healthcare/api-client/v1/fhir/fhir_resources_test.py", line 160, in test_patient
clean_up()
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/backoff/_sync.py", line 94, in retry
ret = target(*args, **kwargs)
File "/workspace/healthcare/api-client/v1/fhir/fhir_resources_test.py", line 147, in clean_up
patient_resource_id,
File "/workspace/healthcare/api-client/v1/fhir/fhir_resources.py", line 299, in delete_resource
response = session.delete(resource_path)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/requests/sessions.py", line 624, in delete
return self.request('DELETE', url, **kwargs)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/auth/transport/requests.py", line 478, in request
self.credentials.before_request(auth_request, method, url, request_headers)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/auth/credentials.py", line 133, in before_request
self.refresh(request)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/service_account.py", line 377, in refresh
request, self._token_uri, assertion
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 193, in jwt_grant
response_data = _token_endpoint_request(request, token_uri, body)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 165, in _token_endpoint_request
_handle_error_response(response_data)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 60, in _handle_error_response
raise exceptions.RefreshError(error_details, response_data)
google.auth.exceptions.RefreshError: ('invalid_grant: Invalid JWT Signature.', {'error': 'invalid_grant', 'error_description': 'Invalid JWT Signature.'})</pre></details>
|
1.0
|
healthcare.api-client.v1.fhir.fhir_resources_test: test_delete_patient failed - This test failed!
To configure my behavior, see [the Flaky Bot documentation](https://github.com/googleapis/repo-automation-bots/tree/master/packages/flakybot).
If I'm commenting on this issue too often, add the `flakybot: quiet` label and
I will stop commenting.
---
commit: b99df8d36109e4fe3e397bfd2cbacac06960340c
buildURL: [Build Status](https://source.cloud.google.com/results/invocations/325e6648-2604-4a2d-a611-53de8a0492aa), [Sponge](http://sponge2/325e6648-2604-4a2d-a611-53de8a0492aa)
status: failed
<details><summary>Test output</summary><br><pre>Traceback (most recent call last):
File "/workspace/healthcare/api-client/v1/fhir/fhir_resources_test.py", line 160, in test_patient
clean_up()
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/backoff/_sync.py", line 94, in retry
ret = target(*args, **kwargs)
File "/workspace/healthcare/api-client/v1/fhir/fhir_resources_test.py", line 147, in clean_up
patient_resource_id,
File "/workspace/healthcare/api-client/v1/fhir/fhir_resources.py", line 299, in delete_resource
response = session.delete(resource_path)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/requests/sessions.py", line 624, in delete
return self.request('DELETE', url, **kwargs)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/auth/transport/requests.py", line 478, in request
self.credentials.before_request(auth_request, method, url, request_headers)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/auth/credentials.py", line 133, in before_request
self.refresh(request)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/service_account.py", line 377, in refresh
request, self._token_uri, assertion
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 193, in jwt_grant
response_data = _token_endpoint_request(request, token_uri, body)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 165, in _token_endpoint_request
_handle_error_response(response_data)
File "/workspace/healthcare/api-client/v1/fhir/.nox/py-3-6/lib/python3.6/site-packages/google/oauth2/_client.py", line 60, in _handle_error_response
raise exceptions.RefreshError(error_details, response_data)
google.auth.exceptions.RefreshError: ('invalid_grant: Invalid JWT Signature.', {'error': 'invalid_grant', 'error_description': 'Invalid JWT Signature.'})</pre></details>
|
non_defect
|
healthcare api client fhir fhir resources test test delete patient failed this test failed to configure my behavior see if i m commenting on this issue too often add the flakybot quiet label and i will stop commenting commit buildurl status failed test output traceback most recent call last file workspace healthcare api client fhir fhir resources test py line in test patient clean up file workspace healthcare api client fhir nox py lib site packages backoff sync py line in retry ret target args kwargs file workspace healthcare api client fhir fhir resources test py line in clean up patient resource id file workspace healthcare api client fhir fhir resources py line in delete resource response session delete resource path file workspace healthcare api client fhir nox py lib site packages requests sessions py line in delete return self request delete url kwargs file workspace healthcare api client fhir nox py lib site packages google auth transport requests py line in request self credentials before request auth request method url request headers file workspace healthcare api client fhir nox py lib site packages google auth credentials py line in before request self refresh request file workspace healthcare api client fhir nox py lib site packages google service account py line in refresh request self token uri assertion file workspace healthcare api client fhir nox py lib site packages google client py line in jwt grant response data token endpoint request request token uri body file workspace healthcare api client fhir nox py lib site packages google client py line in token endpoint request handle error response response data file workspace healthcare api client fhir nox py lib site packages google client py line in handle error response raise exceptions refresherror error details response data google auth exceptions refresherror invalid grant invalid jwt signature error invalid grant error description invalid jwt signature
| 0
|
68,644
| 21,775,146,357
|
IssuesEvent
|
2022-05-13 13:08:30
|
hazelcast/hazelcast
|
https://api.github.com/repos/hazelcast/hazelcast
|
opened
|
DataPersistenceConfig and TieredStoreConfig are missing in dynamic AddMapConfig
|
Type: Defect Source: Internal Module: Config
|
**Describe the bug**
`DataPersistenceConfig` and `TieredStoreConfig` submap configurations are
not transmitted while dynamically adding map config.
See: https://github.com/hazelcast/hazelcast/blob/master/hazelcast/src/main/java/com/hazelcast/client/impl/protocol/codec/DynamicConfigAddMapConfigCodec.java
So, currently these subconfig elements cannot be configured while dynamically
adding the map config.
To fix it:
Update the client protocol of [DynamicConfig#addMapConfig](https://github.com/hazelcast/hazelcast-client-protocol/blob/master/protocol-definitions/DynamicConfig.yaml#L698-L887) to include these mentioned subConfigs.
Also, the configuration update checklist wiki page, https://github.com/hazelcast/hazelcast/wiki/Checklist-for-Hazelcast-config-changes, doesn't include the necessary information about this. We need to update this page to cover this.
|
1.0
|
DataPersistenceConfig and TieredStoreConfig are missing in dynamic AddMapConfig - **Describe the bug**
`DataPersistenceConfig` and `TieredStoreConfig` submap configurations are
not transmitted while dynamically adding map config.
See: https://github.com/hazelcast/hazelcast/blob/master/hazelcast/src/main/java/com/hazelcast/client/impl/protocol/codec/DynamicConfigAddMapConfigCodec.java
So, currently these subconfig elements cannot be configured while dynamically
adding the map config.
To fix it:
Update the client protocol of [DynamicConfig#addMapConfig](https://github.com/hazelcast/hazelcast-client-protocol/blob/master/protocol-definitions/DynamicConfig.yaml#L698-L887) to include these mentioned subConfigs.
Also, the configuration update checklist wiki page, https://github.com/hazelcast/hazelcast/wiki/Checklist-for-Hazelcast-config-changes, doesn't include the necessary information about this. We need to update this page to cover this.
|
defect
|
datapersistenceconfig and tieredstoreconfig are missing in dynamic addmapconfig describe the bug datapersistenceconfig and tieredstoreconfig submap configurations are not transmitted while dynamically adding map config see so currently these subconfig elements cannot be configured while dynamically adding the map config to fix it update the client protocol of to include these mentioned subconfigs also the configuration update checklist wiki page doesn t include the necessary information about this we need to update this page to cover this
| 1
|
284,632
| 24,611,544,958
|
IssuesEvent
|
2022-10-14 22:13:13
|
gitpod-io/gitpod
|
https://api.github.com/repos/gitpod-io/gitpod
|
closed
|
Performance test PVC with a single node saturated with workspaces
|
aspect: testing aspect: performance
|
## Is your feature request related to a problem? Please describe
We should test how a node behaves when it is full of workspaces using PVCs, where there is disk activity in the workspaces.
## Describe the behaviour you'd like
1. Start a regular workspace in the cluster using PVC on a single node.
2. Cordon the other two nodes
3. Run loadgen, once to fill half of the node, ~9 workspaces
4. Begin to stop the first loadgen run & start second loadgen run with ~20 workspaces (to fill the initial node and trigger scale-up of a new one)
### Questions
- What IOPS and bandwidth do we achieve on the initial node? How is it different from a normal cluster?
- How does the regular workspace that was initially started respond as we go through the various phases described above?
## Additional context
We're not sure if the IO limiter is needed to cover these disks or not, and whether there will excessive CPU usage as a result of using PVCs for `/workspace`.
|
1.0
|
Performance test PVC with a single node saturated with workspaces - ## Is your feature request related to a problem? Please describe
We should test how a node behaves when it is full of workspaces using PVCs, where there is disk activity in the workspaces.
## Describe the behaviour you'd like
1. Start a regular workspace in the cluster using PVC on a single node.
2. Cordon the other two nodes
3. Run loadgen, once to fill half of the node, ~9 workspaces
4. Begin to stop the first loadgen run & start second loadgen run with ~20 workspaces (to fill the initial node and trigger scale-up of a new one)
### Questions
- What IOPS and bandwidth do we achieve on the initial node? How is it different from a normal cluster?
- How does the regular workspace that was initially started respond as we go through the various phases described above?
## Additional context
We're not sure if the IO limiter is needed to cover these disks or not, and whether there will excessive CPU usage as a result of using PVCs for `/workspace`.
|
non_defect
|
performance test pvc with a single node saturated with workspaces is your feature request related to a problem please describe we should test how a node behaves when it is full of workspaces using pvcs where there is disk activity in the workspaces describe the behaviour you d like start a regular workspace in the cluster using pvc on a single node cordon the other two nodes run loadgen once to fill half of the node workspaces begin to stop the first loadgen run start second loadgen run with workspaces to fill the initial node and trigger scale up of a new one questions what iops and bandwidth do we achieve on the initial node how is it different from a normal cluster how does the regular workspace that was initially started respond as we go through the various phases described above additional context we re not sure if the io limiter is needed to cover these disks or not and whether there will excessive cpu usage as a result of using pvcs for workspace
| 0
|
95,026
| 19,655,695,542
|
IssuesEvent
|
2022-01-10 12:16:28
|
sourcegraph/sourcegraph
|
https://api.github.com/repos/sourcegraph/sourcegraph
|
opened
|
Don't display the definition as a reference when doing "find references" from the definition
|
team/code-intelligence
|
#### Feature request description
<!-- A description of what feature you would like. -->
When doing "Find references" from the definition location, it would be nice if the definition was excluded from the list of results.
#### Is your feature request related to a problem? If so, please describe.
<!-- A description of what the problem is. Ex. I'm always frustrated when [...] -->
Currently, the "references" panel gets loaded when a symbol has only one reference. It would be nice if the navigation went straight to the single reference instead of opening the references panel with two results (and one result is the definition location where you triggered references)
#### Describe alternatives you've considered.
<!-- A description of any alternative solutions or features you've considered. -->
Manually click on the reference result inside the references panel.
#### Additional context
<!-- Add any other context or other information you'd like to provide. -->
|
1.0
|
Don't display the definition as a reference when doing "find references" from the definition - #### Feature request description
<!-- A description of what feature you would like. -->
When doing "Find references" from the definition location, it would be nice if the definition was excluded from the list of results.
#### Is your feature request related to a problem? If so, please describe.
<!-- A description of what the problem is. Ex. I'm always frustrated when [...] -->
Currently, the "references" panel gets loaded when a symbol has only one reference. It would be nice if the navigation went straight to the single reference instead of opening the references panel with two results (and one result is the definition location where you triggered references)
#### Describe alternatives you've considered.
<!-- A description of any alternative solutions or features you've considered. -->
Manually click on the reference result inside the references panel.
#### Additional context
<!-- Add any other context or other information you'd like to provide. -->
|
non_defect
|
don t display the definition as a reference when doing find references from the definition feature request description when doing find references from the definition location it would be nice if the definition was excluded from the list of results is your feature request related to a problem if so please describe currently the references panel gets loaded when a symbol has only one reference it would be nice if the navigation went straight to the single reference instead of opening the references panel with two results and one result is the definition location where you triggered references describe alternatives you ve considered manually click on the reference result inside the references panel additional context
| 0
|
73,782
| 24,797,605,441
|
IssuesEvent
|
2022-10-24 18:41:31
|
fecgov/fecfile-web-app
|
https://api.github.com/repos/fecgov/fecfile-web-app
|
opened
|
Defect: System to itemize JF Transfer regardless of amount
|
defect
|
Implementation of #486 was broken by subsequent development. This ticket is to re-code and re-test this functionality. Note that itemization logic is _not_ determined by aggregation group.
### Business Reason ###
Per FECA regulations, all Form 3X transactions on Line 12 must be itemized on Schedule A, regardless of amount. This includes JF Transfer transactions.
### Acceptance Criteria ###
**Given** one of the following transactions:
- Joint Fundraising Transfer
- PAC JF Transfer Memo
**Then** the transaction will be itemized upon saving
**and** appear on it's own line item on the *.fec file
|
1.0
|
Defect: System to itemize JF Transfer regardless of amount - Implementation of #486 was broken by subsequent development. This ticket is to re-code and re-test this functionality. Note that itemization logic is _not_ determined by aggregation group.
### Business Reason ###
Per FECA regulations, all Form 3X transactions on Line 12 must be itemized on Schedule A, regardless of amount. This includes JF Transfer transactions.
### Acceptance Criteria ###
**Given** one of the following transactions:
- Joint Fundraising Transfer
- PAC JF Transfer Memo
**Then** the transaction will be itemized upon saving
**and** appear on it's own line item on the *.fec file
|
defect
|
defect system to itemize jf transfer regardless of amount implementation of was broken by subsequent development this ticket is to re code and re test this functionality note that itemization logic is not determined by aggregation group business reason per feca regulations all form transactions on line must be itemized on schedule a regardless of amount this includes jf transfer transactions acceptance criteria given one of the following transactions joint fundraising transfer pac jf transfer memo then the transaction will be itemized upon saving and appear on it s own line item on the fec file
| 1
|
33,423
| 12,216,531,738
|
IssuesEvent
|
2020-05-01 15:20:11
|
robertjfinn/hadoop
|
https://api.github.com/repos/robertjfinn/hadoop
|
opened
|
CVE-2019-11834 (High) detected in cJSON-v1.7.8
|
security vulnerability
|
## CVE-2019-11834 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cJSONv1.7.8</b></p></summary>
<p>
<p>Ultralightweight JSON parser in ANSI C</p>
<p>Library home page: <a href=https://github.com/DaveGamble/cJSON.git>https://github.com/DaveGamble/cJSON.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/robertjfinn/hadoop/commit/876b3d37847317582197087627081de9f19f88d9">876b3d37847317582197087627081de9f19f88d9</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (2)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/utils/cJSON/cJSON.c
- /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/utils/cJSON/cJSON.h
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
cJSON before 1.7.11 allows out-of-bounds access, related to \x00 in a string literal.
<p>Publish Date: 2019-05-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11834>CVE-2019-11834</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11835">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11835</a></p>
<p>Release Date: 2019-05-09</p>
<p>Fix Resolution: v1.7.11</p>
</p>
</details>
<p></p>
|
True
|
CVE-2019-11834 (High) detected in cJSON-v1.7.8 - ## CVE-2019-11834 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cJSONv1.7.8</b></p></summary>
<p>
<p>Ultralightweight JSON parser in ANSI C</p>
<p>Library home page: <a href=https://github.com/DaveGamble/cJSON.git>https://github.com/DaveGamble/cJSON.git</a></p>
<p>Found in HEAD commit: <a href="https://github.com/robertjfinn/hadoop/commit/876b3d37847317582197087627081de9f19f88d9">876b3d37847317582197087627081de9f19f88d9</a></p>
</p>
</details>
</p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Library Source Files (2)</summary>
<p></p>
<p> * The source files were matched to this source library based on a best effort match. Source libraries are selected from a list of probable public libraries.</p>
<p>
- /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/utils/cJSON/cJSON.c
- /hadoop/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/main/native/container-executor/impl/utils/cJSON/cJSON.h
</p>
</details>
<p></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
cJSON before 1.7.11 allows out-of-bounds access, related to \x00 in a string literal.
<p>Publish Date: 2019-05-09
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11834>CVE-2019-11834</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11835">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11835</a></p>
<p>Release Date: 2019-05-09</p>
<p>Fix Resolution: v1.7.11</p>
</p>
</details>
<p></p>
|
non_defect
|
cve high detected in cjson cve high severity vulnerability vulnerable library ultralightweight json parser in ansi c library home page a href found in head commit a href library source files the source files were matched to this source library based on a best effort match source libraries are selected from a list of probable public libraries hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server nodemanager src main native container executor impl utils cjson cjson c hadoop hadoop yarn project hadoop yarn hadoop yarn server hadoop yarn server nodemanager src main native container executor impl utils cjson cjson h vulnerability details cjson before allows out of bounds access related to in a string literal publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution
| 0
|
280,963
| 21,315,298,955
|
IssuesEvent
|
2022-04-16 06:56:59
|
Shivang211/pe
|
https://api.github.com/repos/Shivang211/pe
|
opened
|
No visuals for commands
|
severity.Low type.DocumentationBug
|

There are a lot of commands and no visuals provided for any of them, which makes reading and understanding the UG very difficult.
<!--session: 1650088058079-eb80a396-2b55-442e-a1e8-c84d767ca243-->
<!--Version: Web v3.4.2-->
|
1.0
|
No visuals for commands - 
There are a lot of commands and no visuals provided for any of them, which makes reading and understanding the UG very difficult.
<!--session: 1650088058079-eb80a396-2b55-442e-a1e8-c84d767ca243-->
<!--Version: Web v3.4.2-->
|
non_defect
|
no visuals for commands there are a lot of commands and no visuals provided for any of them which makes reading and understanding the ug very difficult
| 0
|
337,867
| 30,268,262,312
|
IssuesEvent
|
2023-07-07 13:31:01
|
apache/camel-quarkus
|
https://api.github.com/repos/apache/camel-quarkus
|
closed
|
Nitrite native integration test could not serialize [[3941604762136]NO₂]
|
bug native test area/testing area/native area/nitrite
|
### Bug description
For some reason this has recently popped up on the CI builds:
```
2023-07-06T15:51:28.1555920Z [INFO] -------------------------------------------------------
2023-07-06T15:51:28.1556320Z [INFO] T E S T S
2023-07-06T15:51:28.1556760Z [INFO] -------------------------------------------------------
2023-07-06T15:51:29.0215396Z [INFO] Running org.apache.camel.quarkus.component.nitrite.it.NitriteIT
2023-07-06T15:51:33.0323722Z ======================================================================================
2023-07-06T15:51:33.0324350Z native.image.path was not set, making a guess for the correct path of native image
2023-07-06T15:51:33.0325453Z guessed path: /home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/camel-quarkus-integration-test-nitrite-3.0.0-SNAPSHOT-runner
2023-07-06T15:51:33.0326449Z ======================================================================================
2023-07-06T15:51:33.0336065Z Executing "/home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/camel-quarkus-integration-test-nitrite-3.0.0-SNAPSHOT-runner -Dquarkus.http.port=8081 -Dquarkus.http.ssl-port=8444 -Dtest.url=http://localhost:8081 -Dquarkus.log.file.path=/home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/quarkus.log -Dquarkus.log.file.enable=true -Dquarkus.log.category."io.quarkus".level=INFO -Dcamel.quarkus.nitrite.test.db.file=/tmp/NitriteTestResource-db-file-17819878958287804087"
2023-07-06T15:51:33.1319606Z 2023-07-06 15:51:33,050 INFO [org.apa.cam.qua.cor.CamelBootstrapRecorder] (main) Bootstrap runtime: org.apache.camel.quarkus.main.CamelMainRuntime
2023-07-06T15:51:33.1321026Z 2023-07-06 15:51:33,051 INFO [org.apa.cam.mai.MainSupport] (main) Apache Camel (Main) 4.0.0-RC1 is starting
2023-07-06T15:51:33.1321867Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Apache Camel 4.0.0-RC1 (camel-1) is starting
2023-07-06T15:51:33.1322916Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Routes startup (started:0)
2023-07-06T15:51:33.1323847Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Apache Camel 4.0.0-RC1 (camel-1) started in 0ms (build:0ms init:0ms start:0ms)
2023-07-06T15:51:33.1325085Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) camel-quarkus-integration-test-nitrite 3.0.0-SNAPSHOT native (powered by Quarkus 3.2.0.Final) started in 0.042s. Listening on: http://0.0.0.0:8081
2023-07-06T15:51:33.1325889Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) Profile prod activated.
2023-07-06T15:51:33.1326938Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) Installed features: [camel-core, camel-nitrite, cdi, resteasy, resteasy-jackson, smallrye-context-propagation, vertx]
2023-07-06T15:51:43.1229021Z ##[error]Exception in thread "MVStore background writer nio:/tmp/NitriteTestResource-db-file-17819878958287804087" java.lang.IllegalStateException: java.lang.IllegalArgumentException: Could not serialize [[3941604762136]NO₂] [1.4.200/0] [1.4.200/3]
2023-07-06T15:51:43.1524187Z at org.h2.mvstore.DataUtils.newIllegalStateException(DataUtils.java:950)
2023-07-06T15:51:43.2171401Z at org.h2.mvstore.MVStore.store(MVStore.java:1338)
2023-07-06T15:51:43.2172037Z at org.h2.mvstore.MVStore.store(MVStore.java:1312)
2023-07-06T15:51:43.2172518Z at org.h2.mvstore.MVStore.tryCommit(MVStore.java:1272)
2023-07-06T15:51:43.2173014Z at org.h2.mvstore.MVStore.writeInBackground(MVStore.java:2780)
2023-07-06T15:51:43.2173539Z at org.h2.mvstore.MVStore$BackgroundWriterThread.run(MVStore.java:3290)
2023-07-06T15:51:43.2174239Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.thread.PlatformThreads.threadStartRoutine(PlatformThreads.java:807)
2023-07-06T15:51:43.2175172Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.posix.thread.PosixPlatformThreads.pthreadStartRoutine(PosixPlatformThreads.java:210)
2023-07-06T15:51:43.2176353Z Caused by: java.lang.IllegalArgumentException: Could not serialize [[3941604762136]NO₂] [1.4.200/0]
2023-07-06T15:51:43.2177316Z at org.h2.mvstore.DataUtils.newIllegalArgumentException(DataUtils.java:924)
2023-07-06T15:51:43.2177939Z at org.h2.mvstore.type.ObjectDataType.serialize(ObjectDataType.java:359)
2023-07-06T15:51:43.2178655Z at org.h2.mvstore.type.ObjectDataType$SerializedObjectType.write(ObjectDataType.java:1527)
2023-07-06T15:51:43.2179593Z at org.h2.mvstore.type.ObjectDataType.write(ObjectDataType.java:127)
2023-07-06T15:51:43.2180134Z at org.h2.mvstore.type.ObjectDataType.write(ObjectDataType.java:121)
2023-07-06T15:51:43.2180627Z at org.h2.mvstore.Page$Leaf.writeValues(Page.java:1516)
2023-07-06T15:51:43.2181033Z at org.h2.mvstore.Page.write(Page.java:670)
2023-07-06T15:51:43.2181483Z at org.h2.mvstore.Page$Leaf.writeUnsavedRecursive(Page.java:1525)
2023-07-06T15:51:43.2181950Z at org.h2.mvstore.MVStore.storeNow(MVStore.java:1427)
2023-07-06T15:51:43.2182597Z at org.h2.mvstore.MVStore.store(MVStore.java:1334)
2023-07-06T15:51:43.2182939Z ... 6 more
2023-07-06T15:51:43.2184215Z Caused by: com.oracle.svm.core.jdk.UnsupportedFeatureError: SerializationConstructorAccessor class not found for declaringClass: java.util.concurrent.ConcurrentSkipListSet (targetConstructorClass: java.util.AbstractSet). Usually adding java.util.concurrent.ConcurrentSkipListSet to serialization-config.json fixes the problem.
2023-07-06T15:51:43.2185371Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.util.VMError.unsupportedFeature(VMError.java:92)
2023-07-06T15:51:43.2186351Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.reflect.serialize.SerializationSupport.getSerializationConstructorAccessor(SerializationSupport.java:143)
2023-07-06T15:51:43.2187439Z at java.base@17.0.7/jdk.internal.reflect.MethodAccessorGenerator.generateSerializationConstructor(MethodAccessorGenerator.java:61)
2023-07-06T15:51:43.2188304Z at java.base@17.0.7/jdk.internal.reflect.ReflectionFactory.generateConstructor(ReflectionFactory.java:463)
2023-07-06T15:51:43.2189111Z at java.base@17.0.7/jdk.internal.reflect.ReflectionFactory.newConstructorForSerialization(ReflectionFactory.java:455)
2023-07-06T15:51:43.2189887Z at java.base@17.0.7/java.io.ObjectStreamClass.getSerializableConstructor(ObjectStreamClass.java:1443)
2023-07-06T15:51:43.2190498Z at java.base@17.0.7/java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:412)
2023-07-06T15:51:43.2191003Z at java.base@17.0.7/java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:384)
2023-07-06T15:51:43.2191593Z at java.base@17.0.7/java.security.AccessController.executePrivileged(AccessController.java:171)
2023-07-06T15:51:43.2192233Z at java.base@17.0.7/java.security.AccessController.doPrivileged(AccessController.java:318)
2023-07-06T15:51:43.2192802Z at java.base@17.0.7/java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:384)
2023-07-06T15:51:43.2193385Z at java.base@17.0.7/java.io.ObjectStreamClass$Caches$1.computeValue(ObjectStreamClass.java:110)
2023-07-06T15:51:43.2194017Z at java.base@17.0.7/java.io.ObjectStreamClass$Caches$1.computeValue(ObjectStreamClass.java:107)
2023-07-06T15:51:43.2195550Z at java.base@17.0.7/java.io.ClassCache$1.computeValue(ClassCache.java:73)
2023-07-06T15:51:43.2195961Z at java.base@17.0.7/java.io.ClassCache$1.computeValue(ClassCache.java:70)
2023-07-06T15:51:43.2196389Z at java.base@17.0.7/java.lang.ClassValue.get(JavaLangSubstitutions.java:681)
2023-07-06T15:51:43.2196790Z at java.base@17.0.7/java.io.ClassCache.get(ClassCache.java:84)
2023-07-06T15:51:43.2197216Z at java.base@17.0.7/java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:363)
2023-07-06T15:51:43.2197720Z at java.base@17.0.7/java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1137)
2023-07-06T15:51:43.2198233Z at java.base@17.0.7/java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:350)
2023-07-06T15:51:43.2198983Z at org.h2.mvstore.type.ObjectDataType.serialize(ObjectDataType.java:356)
2023-07-06T15:51:43.2199343Z ... 14 more
2023-07-06T15:51:43.2200696Z 2023-07-06 15:51:43,064 ERROR [io.qua.ver.htt.run.QuarkusErrorHandler] (executor-thread-1) HTTP Request to /nitrite/repositoryClass?mappable=false failed, error id: 36f95039-4468-4295-a769-49c146d7d407-1: org.jboss.resteasy.spi.UnhandledException: org.apache.camel.CamelExecutionException: Exception occurred during execution on the exchange: Exchange[]
2023-07-06T15:51:43.2201712Z at org.jboss.resteasy.core.ExceptionHandler.handleApplicationException(ExceptionHandler.java:107)
2023-07-06T15:51:43.2202363Z at org.jboss.resteasy.core.ExceptionHandler.handleException(ExceptionHandler.java:344)
2023-07-06T15:51:43.2203006Z at org.jboss.resteasy.core.SynchronousDispatcher.writeException(SynchronousDispatcher.java:205)
2023-07-06T15:51:43.2203641Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:452)
2023-07-06T15:51:43.2204243Z at org.jboss.resteasy.core.SynchronousDispatcher.lambda$invoke$4(SynchronousDispatcher.java:240)
2023-07-06T15:51:43.2205009Z at org.jboss.resteasy.core.SynchronousDispatcher.lambda$preprocess$0(SynchronousDispatcher.java:154)
2023-07-06T15:51:43.2205810Z at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:321)
2023-07-06T15:51:43.2206598Z at org.jboss.resteasy.core.SynchronousDispatcher.preprocess(SynchronousDispatcher.java:157)
2023-07-06T15:51:43.2207202Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:229)
2023-07-06T15:51:43.2207839Z at io.quarkus.resteasy.runtime.standalone.RequestDispatcher.service(RequestDispatcher.java:82)
2023-07-06T15:51:43.2208538Z at io.quarkus.resteasy.runtime.standalone.VertxRequestHandler.dispatch(VertxRequestHandler.java:147)
2023-07-06T15:51:43.2209222Z at io.quarkus.resteasy.runtime.standalone.VertxRequestHandler$1.run(VertxRequestHandler.java:93)
2023-07-06T15:51:43.2209813Z at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:576)
2023-07-06T15:51:43.2210364Z at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2513)
2023-07-06T15:51:43.2210919Z at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1538)
2023-07-06T15:51:43.2211427Z at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29)
2023-07-06T15:51:43.2211987Z at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29)
2023-07-06T15:51:43.2212601Z at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
2023-07-06T15:51:43.2213074Z at java.base@17.0.7/java.lang.Thread.run(Thread.java:833)
2023-07-06T15:51:43.2213640Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.thread.PlatformThreads.threadStartRoutine(PlatformThreads.java:807)
2023-07-06T15:51:43.2214493Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.posix.thread.PosixPlatformThreads.pthreadStartRoutine(PosixPlatformThreads.java:210)
2023-07-06T15:51:43.2215255Z Caused by: org.apache.camel.CamelExecutionException: Exception occurred during execution on the exchange: Exchange[]
2023-07-06T15:51:43.2215862Z at org.apache.camel.support.ExchangeHelper.extractResultBody(ExchangeHelper.java:653)
2023-07-06T15:51:43.2216545Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:368)
2023-07-06T15:51:43.2217320Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:333)
2023-07-06T15:51:43.2218127Z at org.apache.camel.quarkus.component.nitrite.it.NitriteResource.postRepositoryClass(NitriteResource.java:87)
2023-07-06T15:51:43.2218916Z at org.apache.camel.quarkus.component.nitrite.it.NitriteResource_ClientProxy.postRepositoryClass(Unknown Source)
2023-07-06T15:51:43.2219474Z at java.base@17.0.7/java.lang.reflect.Method.invoke(Method.java:568)
2023-07-06T15:51:43.2220237Z at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:154)
2023-07-06T15:51:43.2220834Z at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:118)
2023-07-06T15:51:43.2221593Z at org.jboss.resteasy.core.ResourceMethodInvoker.internalInvokeOnTarget(ResourceMethodInvoker.java:560)
2023-07-06T15:51:43.2222359Z at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTargetAfterFilter(ResourceMethodInvoker.java:452)
2023-07-06T15:51:43.2223073Z at org.jboss.resteasy.core.ResourceMethodInvoker.lambda$invokeOnTarget$2(ResourceMethodInvoker.java:413)
2023-07-06T15:51:43.2223876Z at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:321)
2023-07-06T15:51:43.2224673Z at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTarget(ResourceMethodInvoker.java:415)
2023-07-06T15:51:43.2225303Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:378)
2023-07-06T15:51:43.2225895Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:356)
2023-07-06T15:51:43.2226579Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:70)
2023-07-06T15:51:43.2227163Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:429)
2023-07-06T15:51:43.2227556Z ... 17 more
2023-07-06T15:51:43.2227921Z Caused by: org.dizitart.no2.exceptions.NitriteIOException: NO2.2004: store is closed
2023-07-06T15:51:43.2228524Z at org.dizitart.no2.internals.DefaultNitriteCollection.checkOpened(DefaultNitriteCollection.java:394)
2023-07-06T15:51:43.2229214Z at org.dizitart.no2.internals.DefaultNitriteCollection.update(DefaultNitriteCollection.java:320)
2023-07-06T15:51:43.2229861Z at org.dizitart.no2.objects.DefaultObjectRepository.update(DefaultObjectRepository.java:141)
2023-07-06T15:51:43.2230489Z at org.dizitart.no2.objects.DefaultObjectRepository.update(DefaultObjectRepository.java:126)
2023-07-06T15:51:43.2231180Z at org.apache.camel.component.nitrite.operation.common.UpsertOperation.execute(UpsertOperation.java:41)
2023-07-06T15:51:43.2231880Z at org.apache.camel.component.nitrite.NitriteProducer.process(NitriteProducer.java:78)
2023-07-06T15:51:43.2232637Z at org.apache.camel.support.AsyncProcessorConverterHelper$ProcessorToAsyncProcessorBridge.process(AsyncProcessorConverterHelper.java:66)
2023-07-06T15:51:43.2233453Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:214)
2023-07-06T15:51:43.2234191Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor$1.process(SharedCamelInternalProcessor.java:111)
2023-07-06T15:51:43.2234982Z at org.apache.camel.impl.engine.DefaultAsyncProcessorAwaitManager.process(DefaultAsyncProcessorAwaitManager.java:81)
2023-07-06T15:51:43.2235782Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:108)
2023-07-06T15:51:43.2236467Z at org.apache.camel.support.cache.DefaultProducerCache.send(DefaultProducerCache.java:180)
2023-07-06T15:51:43.2237116Z at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:176)
2023-07-06T15:51:43.2237769Z at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:172)
2023-07-06T15:51:43.2238490Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:360)
2023-07-06T15:51:43.2239137Z ... 32 more
```
|
2.0
|
Nitrite native integration test could not serialize [[3941604762136]NO₂] - ### Bug description
For some reason this has recently popped up on the CI builds:
```
2023-07-06T15:51:28.1555920Z [INFO] -------------------------------------------------------
2023-07-06T15:51:28.1556320Z [INFO] T E S T S
2023-07-06T15:51:28.1556760Z [INFO] -------------------------------------------------------
2023-07-06T15:51:29.0215396Z [INFO] Running org.apache.camel.quarkus.component.nitrite.it.NitriteIT
2023-07-06T15:51:33.0323722Z ======================================================================================
2023-07-06T15:51:33.0324350Z native.image.path was not set, making a guess for the correct path of native image
2023-07-06T15:51:33.0325453Z guessed path: /home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/camel-quarkus-integration-test-nitrite-3.0.0-SNAPSHOT-runner
2023-07-06T15:51:33.0326449Z ======================================================================================
2023-07-06T15:51:33.0336065Z Executing "/home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/camel-quarkus-integration-test-nitrite-3.0.0-SNAPSHOT-runner -Dquarkus.http.port=8081 -Dquarkus.http.ssl-port=8444 -Dtest.url=http://localhost:8081 -Dquarkus.log.file.path=/home/runner/work/camel-quarkus/camel-quarkus/integration-tests/nitrite/target/quarkus.log -Dquarkus.log.file.enable=true -Dquarkus.log.category."io.quarkus".level=INFO -Dcamel.quarkus.nitrite.test.db.file=/tmp/NitriteTestResource-db-file-17819878958287804087"
2023-07-06T15:51:33.1319606Z 2023-07-06 15:51:33,050 INFO [org.apa.cam.qua.cor.CamelBootstrapRecorder] (main) Bootstrap runtime: org.apache.camel.quarkus.main.CamelMainRuntime
2023-07-06T15:51:33.1321026Z 2023-07-06 15:51:33,051 INFO [org.apa.cam.mai.MainSupport] (main) Apache Camel (Main) 4.0.0-RC1 is starting
2023-07-06T15:51:33.1321867Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Apache Camel 4.0.0-RC1 (camel-1) is starting
2023-07-06T15:51:33.1322916Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Routes startup (started:0)
2023-07-06T15:51:33.1323847Z 2023-07-06 15:51:33,060 INFO [org.apa.cam.imp.eng.AbstractCamelContext] (main) Apache Camel 4.0.0-RC1 (camel-1) started in 0ms (build:0ms init:0ms start:0ms)
2023-07-06T15:51:33.1325085Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) camel-quarkus-integration-test-nitrite 3.0.0-SNAPSHOT native (powered by Quarkus 3.2.0.Final) started in 0.042s. Listening on: http://0.0.0.0:8081
2023-07-06T15:51:33.1325889Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) Profile prod activated.
2023-07-06T15:51:33.1326938Z 2023-07-06 15:51:33,064 INFO [io.quarkus] (main) Installed features: [camel-core, camel-nitrite, cdi, resteasy, resteasy-jackson, smallrye-context-propagation, vertx]
2023-07-06T15:51:43.1229021Z ##[error]Exception in thread "MVStore background writer nio:/tmp/NitriteTestResource-db-file-17819878958287804087" java.lang.IllegalStateException: java.lang.IllegalArgumentException: Could not serialize [[3941604762136]NO₂] [1.4.200/0] [1.4.200/3]
2023-07-06T15:51:43.1524187Z at org.h2.mvstore.DataUtils.newIllegalStateException(DataUtils.java:950)
2023-07-06T15:51:43.2171401Z at org.h2.mvstore.MVStore.store(MVStore.java:1338)
2023-07-06T15:51:43.2172037Z at org.h2.mvstore.MVStore.store(MVStore.java:1312)
2023-07-06T15:51:43.2172518Z at org.h2.mvstore.MVStore.tryCommit(MVStore.java:1272)
2023-07-06T15:51:43.2173014Z at org.h2.mvstore.MVStore.writeInBackground(MVStore.java:2780)
2023-07-06T15:51:43.2173539Z at org.h2.mvstore.MVStore$BackgroundWriterThread.run(MVStore.java:3290)
2023-07-06T15:51:43.2174239Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.thread.PlatformThreads.threadStartRoutine(PlatformThreads.java:807)
2023-07-06T15:51:43.2175172Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.posix.thread.PosixPlatformThreads.pthreadStartRoutine(PosixPlatformThreads.java:210)
2023-07-06T15:51:43.2176353Z Caused by: java.lang.IllegalArgumentException: Could not serialize [[3941604762136]NO₂] [1.4.200/0]
2023-07-06T15:51:43.2177316Z at org.h2.mvstore.DataUtils.newIllegalArgumentException(DataUtils.java:924)
2023-07-06T15:51:43.2177939Z at org.h2.mvstore.type.ObjectDataType.serialize(ObjectDataType.java:359)
2023-07-06T15:51:43.2178655Z at org.h2.mvstore.type.ObjectDataType$SerializedObjectType.write(ObjectDataType.java:1527)
2023-07-06T15:51:43.2179593Z at org.h2.mvstore.type.ObjectDataType.write(ObjectDataType.java:127)
2023-07-06T15:51:43.2180134Z at org.h2.mvstore.type.ObjectDataType.write(ObjectDataType.java:121)
2023-07-06T15:51:43.2180627Z at org.h2.mvstore.Page$Leaf.writeValues(Page.java:1516)
2023-07-06T15:51:43.2181033Z at org.h2.mvstore.Page.write(Page.java:670)
2023-07-06T15:51:43.2181483Z at org.h2.mvstore.Page$Leaf.writeUnsavedRecursive(Page.java:1525)
2023-07-06T15:51:43.2181950Z at org.h2.mvstore.MVStore.storeNow(MVStore.java:1427)
2023-07-06T15:51:43.2182597Z at org.h2.mvstore.MVStore.store(MVStore.java:1334)
2023-07-06T15:51:43.2182939Z ... 6 more
2023-07-06T15:51:43.2184215Z Caused by: com.oracle.svm.core.jdk.UnsupportedFeatureError: SerializationConstructorAccessor class not found for declaringClass: java.util.concurrent.ConcurrentSkipListSet (targetConstructorClass: java.util.AbstractSet). Usually adding java.util.concurrent.ConcurrentSkipListSet to serialization-config.json fixes the problem.
2023-07-06T15:51:43.2185371Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.util.VMError.unsupportedFeature(VMError.java:92)
2023-07-06T15:51:43.2186351Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.reflect.serialize.SerializationSupport.getSerializationConstructorAccessor(SerializationSupport.java:143)
2023-07-06T15:51:43.2187439Z at java.base@17.0.7/jdk.internal.reflect.MethodAccessorGenerator.generateSerializationConstructor(MethodAccessorGenerator.java:61)
2023-07-06T15:51:43.2188304Z at java.base@17.0.7/jdk.internal.reflect.ReflectionFactory.generateConstructor(ReflectionFactory.java:463)
2023-07-06T15:51:43.2189111Z at java.base@17.0.7/jdk.internal.reflect.ReflectionFactory.newConstructorForSerialization(ReflectionFactory.java:455)
2023-07-06T15:51:43.2189887Z at java.base@17.0.7/java.io.ObjectStreamClass.getSerializableConstructor(ObjectStreamClass.java:1443)
2023-07-06T15:51:43.2190498Z at java.base@17.0.7/java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:412)
2023-07-06T15:51:43.2191003Z at java.base@17.0.7/java.io.ObjectStreamClass$2.run(ObjectStreamClass.java:384)
2023-07-06T15:51:43.2191593Z at java.base@17.0.7/java.security.AccessController.executePrivileged(AccessController.java:171)
2023-07-06T15:51:43.2192233Z at java.base@17.0.7/java.security.AccessController.doPrivileged(AccessController.java:318)
2023-07-06T15:51:43.2192802Z at java.base@17.0.7/java.io.ObjectStreamClass.<init>(ObjectStreamClass.java:384)
2023-07-06T15:51:43.2193385Z at java.base@17.0.7/java.io.ObjectStreamClass$Caches$1.computeValue(ObjectStreamClass.java:110)
2023-07-06T15:51:43.2194017Z at java.base@17.0.7/java.io.ObjectStreamClass$Caches$1.computeValue(ObjectStreamClass.java:107)
2023-07-06T15:51:43.2195550Z at java.base@17.0.7/java.io.ClassCache$1.computeValue(ClassCache.java:73)
2023-07-06T15:51:43.2195961Z at java.base@17.0.7/java.io.ClassCache$1.computeValue(ClassCache.java:70)
2023-07-06T15:51:43.2196389Z at java.base@17.0.7/java.lang.ClassValue.get(JavaLangSubstitutions.java:681)
2023-07-06T15:51:43.2196790Z at java.base@17.0.7/java.io.ClassCache.get(ClassCache.java:84)
2023-07-06T15:51:43.2197216Z at java.base@17.0.7/java.io.ObjectStreamClass.lookup(ObjectStreamClass.java:363)
2023-07-06T15:51:43.2197720Z at java.base@17.0.7/java.io.ObjectOutputStream.writeObject0(ObjectOutputStream.java:1137)
2023-07-06T15:51:43.2198233Z at java.base@17.0.7/java.io.ObjectOutputStream.writeObject(ObjectOutputStream.java:350)
2023-07-06T15:51:43.2198983Z at org.h2.mvstore.type.ObjectDataType.serialize(ObjectDataType.java:356)
2023-07-06T15:51:43.2199343Z ... 14 more
2023-07-06T15:51:43.2200696Z 2023-07-06 15:51:43,064 ERROR [io.qua.ver.htt.run.QuarkusErrorHandler] (executor-thread-1) HTTP Request to /nitrite/repositoryClass?mappable=false failed, error id: 36f95039-4468-4295-a769-49c146d7d407-1: org.jboss.resteasy.spi.UnhandledException: org.apache.camel.CamelExecutionException: Exception occurred during execution on the exchange: Exchange[]
2023-07-06T15:51:43.2201712Z at org.jboss.resteasy.core.ExceptionHandler.handleApplicationException(ExceptionHandler.java:107)
2023-07-06T15:51:43.2202363Z at org.jboss.resteasy.core.ExceptionHandler.handleException(ExceptionHandler.java:344)
2023-07-06T15:51:43.2203006Z at org.jboss.resteasy.core.SynchronousDispatcher.writeException(SynchronousDispatcher.java:205)
2023-07-06T15:51:43.2203641Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:452)
2023-07-06T15:51:43.2204243Z at org.jboss.resteasy.core.SynchronousDispatcher.lambda$invoke$4(SynchronousDispatcher.java:240)
2023-07-06T15:51:43.2205009Z at org.jboss.resteasy.core.SynchronousDispatcher.lambda$preprocess$0(SynchronousDispatcher.java:154)
2023-07-06T15:51:43.2205810Z at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:321)
2023-07-06T15:51:43.2206598Z at org.jboss.resteasy.core.SynchronousDispatcher.preprocess(SynchronousDispatcher.java:157)
2023-07-06T15:51:43.2207202Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:229)
2023-07-06T15:51:43.2207839Z at io.quarkus.resteasy.runtime.standalone.RequestDispatcher.service(RequestDispatcher.java:82)
2023-07-06T15:51:43.2208538Z at io.quarkus.resteasy.runtime.standalone.VertxRequestHandler.dispatch(VertxRequestHandler.java:147)
2023-07-06T15:51:43.2209222Z at io.quarkus.resteasy.runtime.standalone.VertxRequestHandler$1.run(VertxRequestHandler.java:93)
2023-07-06T15:51:43.2209813Z at io.quarkus.vertx.core.runtime.VertxCoreRecorder$14.runWith(VertxCoreRecorder.java:576)
2023-07-06T15:51:43.2210364Z at org.jboss.threads.EnhancedQueueExecutor$Task.run(EnhancedQueueExecutor.java:2513)
2023-07-06T15:51:43.2210919Z at org.jboss.threads.EnhancedQueueExecutor$ThreadBody.run(EnhancedQueueExecutor.java:1538)
2023-07-06T15:51:43.2211427Z at org.jboss.threads.DelegatingRunnable.run(DelegatingRunnable.java:29)
2023-07-06T15:51:43.2211987Z at org.jboss.threads.ThreadLocalResettingRunnable.run(ThreadLocalResettingRunnable.java:29)
2023-07-06T15:51:43.2212601Z at io.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
2023-07-06T15:51:43.2213074Z at java.base@17.0.7/java.lang.Thread.run(Thread.java:833)
2023-07-06T15:51:43.2213640Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.thread.PlatformThreads.threadStartRoutine(PlatformThreads.java:807)
2023-07-06T15:51:43.2214493Z at org.graalvm.nativeimage.builder/com.oracle.svm.core.posix.thread.PosixPlatformThreads.pthreadStartRoutine(PosixPlatformThreads.java:210)
2023-07-06T15:51:43.2215255Z Caused by: org.apache.camel.CamelExecutionException: Exception occurred during execution on the exchange: Exchange[]
2023-07-06T15:51:43.2215862Z at org.apache.camel.support.ExchangeHelper.extractResultBody(ExchangeHelper.java:653)
2023-07-06T15:51:43.2216545Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:368)
2023-07-06T15:51:43.2217320Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:333)
2023-07-06T15:51:43.2218127Z at org.apache.camel.quarkus.component.nitrite.it.NitriteResource.postRepositoryClass(NitriteResource.java:87)
2023-07-06T15:51:43.2218916Z at org.apache.camel.quarkus.component.nitrite.it.NitriteResource_ClientProxy.postRepositoryClass(Unknown Source)
2023-07-06T15:51:43.2219474Z at java.base@17.0.7/java.lang.reflect.Method.invoke(Method.java:568)
2023-07-06T15:51:43.2220237Z at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:154)
2023-07-06T15:51:43.2220834Z at org.jboss.resteasy.core.MethodInjectorImpl.invoke(MethodInjectorImpl.java:118)
2023-07-06T15:51:43.2221593Z at org.jboss.resteasy.core.ResourceMethodInvoker.internalInvokeOnTarget(ResourceMethodInvoker.java:560)
2023-07-06T15:51:43.2222359Z at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTargetAfterFilter(ResourceMethodInvoker.java:452)
2023-07-06T15:51:43.2223073Z at org.jboss.resteasy.core.ResourceMethodInvoker.lambda$invokeOnTarget$2(ResourceMethodInvoker.java:413)
2023-07-06T15:51:43.2223876Z at org.jboss.resteasy.core.interception.jaxrs.PreMatchContainerRequestContext.filter(PreMatchContainerRequestContext.java:321)
2023-07-06T15:51:43.2224673Z at org.jboss.resteasy.core.ResourceMethodInvoker.invokeOnTarget(ResourceMethodInvoker.java:415)
2023-07-06T15:51:43.2225303Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:378)
2023-07-06T15:51:43.2225895Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:356)
2023-07-06T15:51:43.2226579Z at org.jboss.resteasy.core.ResourceMethodInvoker.invoke(ResourceMethodInvoker.java:70)
2023-07-06T15:51:43.2227163Z at org.jboss.resteasy.core.SynchronousDispatcher.invoke(SynchronousDispatcher.java:429)
2023-07-06T15:51:43.2227556Z ... 17 more
2023-07-06T15:51:43.2227921Z Caused by: org.dizitart.no2.exceptions.NitriteIOException: NO2.2004: store is closed
2023-07-06T15:51:43.2228524Z at org.dizitart.no2.internals.DefaultNitriteCollection.checkOpened(DefaultNitriteCollection.java:394)
2023-07-06T15:51:43.2229214Z at org.dizitart.no2.internals.DefaultNitriteCollection.update(DefaultNitriteCollection.java:320)
2023-07-06T15:51:43.2229861Z at org.dizitart.no2.objects.DefaultObjectRepository.update(DefaultObjectRepository.java:141)
2023-07-06T15:51:43.2230489Z at org.dizitart.no2.objects.DefaultObjectRepository.update(DefaultObjectRepository.java:126)
2023-07-06T15:51:43.2231180Z at org.apache.camel.component.nitrite.operation.common.UpsertOperation.execute(UpsertOperation.java:41)
2023-07-06T15:51:43.2231880Z at org.apache.camel.component.nitrite.NitriteProducer.process(NitriteProducer.java:78)
2023-07-06T15:51:43.2232637Z at org.apache.camel.support.AsyncProcessorConverterHelper$ProcessorToAsyncProcessorBridge.process(AsyncProcessorConverterHelper.java:66)
2023-07-06T15:51:43.2233453Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:214)
2023-07-06T15:51:43.2234191Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor$1.process(SharedCamelInternalProcessor.java:111)
2023-07-06T15:51:43.2234982Z at org.apache.camel.impl.engine.DefaultAsyncProcessorAwaitManager.process(DefaultAsyncProcessorAwaitManager.java:81)
2023-07-06T15:51:43.2235782Z at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:108)
2023-07-06T15:51:43.2236467Z at org.apache.camel.support.cache.DefaultProducerCache.send(DefaultProducerCache.java:180)
2023-07-06T15:51:43.2237116Z at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:176)
2023-07-06T15:51:43.2237769Z at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:172)
2023-07-06T15:51:43.2238490Z at org.apache.camel.impl.engine.DefaultFluentProducerTemplate.request(DefaultFluentProducerTemplate.java:360)
2023-07-06T15:51:43.2239137Z ... 32 more
```
|
non_defect
|
nitrite native integration test could not serialize no₂ bug description for some reason this has recently popped up on the ci builds t e s t s running org apache camel quarkus component nitrite it nitriteit native image path was not set making a guess for the correct path of native image guessed path home runner work camel quarkus camel quarkus integration tests nitrite target camel quarkus integration test nitrite snapshot runner executing home runner work camel quarkus camel quarkus integration tests nitrite target camel quarkus integration test nitrite snapshot runner dquarkus http port dquarkus http ssl port dtest url dquarkus log file path home runner work camel quarkus camel quarkus integration tests nitrite target quarkus log dquarkus log file enable true dquarkus log category io quarkus level info dcamel quarkus nitrite test db file tmp nitritetestresource db file info main bootstrap runtime org apache camel quarkus main camelmainruntime info main apache camel main is starting info main apache camel camel is starting info main routes startup started info main apache camel camel started in build init start info main camel quarkus integration test nitrite snapshot native powered by quarkus final started in listening on info main profile prod activated info main installed features exception in thread mvstore background writer nio tmp nitritetestresource db file java lang illegalstateexception java lang illegalargumentexception could not serialize no₂ at org mvstore datautils newillegalstateexception datautils java at org mvstore mvstore store mvstore java at org mvstore mvstore store mvstore java at org mvstore mvstore trycommit mvstore java at org mvstore mvstore writeinbackground mvstore java at org mvstore mvstore backgroundwriterthread run mvstore java at org graalvm nativeimage builder com oracle svm core thread platformthreads threadstartroutine platformthreads java at org graalvm nativeimage builder com oracle svm core posix thread posixplatformthreads pthreadstartroutine posixplatformthreads java caused by java lang illegalargumentexception could not serialize no₂ at org mvstore datautils newillegalargumentexception datautils java at org mvstore type objectdatatype serialize objectdatatype java at org mvstore type objectdatatype serializedobjecttype write objectdatatype java at org mvstore type objectdatatype write objectdatatype java at org mvstore type objectdatatype write objectdatatype java at org mvstore page leaf writevalues page java at org mvstore page write page java at org mvstore page leaf writeunsavedrecursive page java at org mvstore mvstore storenow mvstore java at org mvstore mvstore store mvstore java more caused by com oracle svm core jdk unsupportedfeatureerror serializationconstructoraccessor class not found for declaringclass java util concurrent concurrentskiplistset targetconstructorclass java util abstractset usually adding java util concurrent concurrentskiplistset to serialization config json fixes the problem at org graalvm nativeimage builder com oracle svm core util vmerror unsupportedfeature vmerror java at org graalvm nativeimage builder com oracle svm core reflect serialize serializationsupport getserializationconstructoraccessor serializationsupport java at java base jdk internal reflect methodaccessorgenerator generateserializationconstructor methodaccessorgenerator java at java base jdk internal reflect reflectionfactory generateconstructor reflectionfactory java at java base jdk internal reflect reflectionfactory newconstructorforserialization reflectionfactory java at java base java io objectstreamclass getserializableconstructor objectstreamclass java at java base java io objectstreamclass run objectstreamclass java at java base java io objectstreamclass run objectstreamclass java at java base java security accesscontroller executeprivileged accesscontroller java at java base java security accesscontroller doprivileged accesscontroller java at java base java io objectstreamclass objectstreamclass java at java base java io objectstreamclass caches computevalue objectstreamclass java at java base java io objectstreamclass caches computevalue objectstreamclass java at java base java io classcache computevalue classcache java at java base java io classcache computevalue classcache java at java base java lang classvalue get javalangsubstitutions java at java base java io classcache get classcache java at java base java io objectstreamclass lookup objectstreamclass java at java base java io objectoutputstream objectoutputstream java at java base java io objectoutputstream writeobject objectoutputstream java at org mvstore type objectdatatype serialize objectdatatype java more error executor thread http request to nitrite repositoryclass mappable false failed error id org jboss resteasy spi unhandledexception org apache camel camelexecutionexception exception occurred during execution on the exchange exchange at org jboss resteasy core exceptionhandler handleapplicationexception exceptionhandler java at org jboss resteasy core exceptionhandler handleexception exceptionhandler java at org jboss resteasy core synchronousdispatcher writeexception synchronousdispatcher java at org jboss resteasy core synchronousdispatcher invoke synchronousdispatcher java at org jboss resteasy core synchronousdispatcher lambda invoke synchronousdispatcher java at org jboss resteasy core synchronousdispatcher lambda preprocess synchronousdispatcher java at org jboss resteasy core interception jaxrs prematchcontainerrequestcontext filter prematchcontainerrequestcontext java at org jboss resteasy core synchronousdispatcher preprocess synchronousdispatcher java at org jboss resteasy core synchronousdispatcher invoke synchronousdispatcher java at io quarkus resteasy runtime standalone requestdispatcher service requestdispatcher java at io quarkus resteasy runtime standalone vertxrequesthandler dispatch vertxrequesthandler java at io quarkus resteasy runtime standalone vertxrequesthandler run vertxrequesthandler java at io quarkus vertx core runtime vertxcorerecorder runwith vertxcorerecorder java at org jboss threads enhancedqueueexecutor task run enhancedqueueexecutor java at org jboss threads enhancedqueueexecutor threadbody run enhancedqueueexecutor java at org jboss threads delegatingrunnable run delegatingrunnable java at org jboss threads threadlocalresettingrunnable run threadlocalresettingrunnable java at io netty util concurrent fastthreadlocalrunnable run fastthreadlocalrunnable java at java base java lang thread run thread java at org graalvm nativeimage builder com oracle svm core thread platformthreads threadstartroutine platformthreads java at org graalvm nativeimage builder com oracle svm core posix thread posixplatformthreads pthreadstartroutine posixplatformthreads java caused by org apache camel camelexecutionexception exception occurred during execution on the exchange exchange at org apache camel support exchangehelper extractresultbody exchangehelper java at org apache camel impl engine defaultfluentproducertemplate request defaultfluentproducertemplate java at org apache camel impl engine defaultfluentproducertemplate request defaultfluentproducertemplate java at org apache camel quarkus component nitrite it nitriteresource postrepositoryclass nitriteresource java at org apache camel quarkus component nitrite it nitriteresource clientproxy postrepositoryclass unknown source at java base java lang reflect method invoke method java at org jboss resteasy core methodinjectorimpl invoke methodinjectorimpl java at org jboss resteasy core methodinjectorimpl invoke methodinjectorimpl java at org jboss resteasy core resourcemethodinvoker internalinvokeontarget resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker invokeontargetafterfilter resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker lambda invokeontarget resourcemethodinvoker java at org jboss resteasy core interception jaxrs prematchcontainerrequestcontext filter prematchcontainerrequestcontext java at org jboss resteasy core resourcemethodinvoker invokeontarget resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker invoke resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker invoke resourcemethodinvoker java at org jboss resteasy core resourcemethodinvoker invoke resourcemethodinvoker java at org jboss resteasy core synchronousdispatcher invoke synchronousdispatcher java more caused by org dizitart exceptions nitriteioexception store is closed at org dizitart internals defaultnitritecollection checkopened defaultnitritecollection java at org dizitart internals defaultnitritecollection update defaultnitritecollection java at org dizitart objects defaultobjectrepository update defaultobjectrepository java at org dizitart objects defaultobjectrepository update defaultobjectrepository java at org apache camel component nitrite operation common upsertoperation execute upsertoperation java at org apache camel component nitrite nitriteproducer process nitriteproducer java at org apache camel support asyncprocessorconverterhelper processortoasyncprocessorbridge process asyncprocessorconverterhelper java at org apache camel impl engine sharedcamelinternalprocessor process sharedcamelinternalprocessor java at org apache camel impl engine sharedcamelinternalprocessor process sharedcamelinternalprocessor java at org apache camel impl engine defaultasyncprocessorawaitmanager process defaultasyncprocessorawaitmanager java at org apache camel impl engine sharedcamelinternalprocessor process sharedcamelinternalprocessor java at org apache camel support cache defaultproducercache send defaultproducercache java at org apache camel impl engine defaultproducertemplate send defaultproducertemplate java at org apache camel impl engine defaultproducertemplate send defaultproducertemplate java at org apache camel impl engine defaultfluentproducertemplate request defaultfluentproducertemplate java more
| 0
|
5,095
| 26,007,472,462
|
IssuesEvent
|
2022-12-20 20:56:11
|
aws/aws-sam-cli
|
https://api.github.com/repos/aws/aws-sam-cli
|
closed
|
[Feature Request] Cognito Offline Testing
|
type/feature area/local/start-api stage/pm-review maintainer/need-followup
|
Just curious whether SAM will support cognito offline testing? Pasting in some links to the same feature request in various other forums:
- https://github.com/localstack/localstack/issues/26
- https://stackoverflow.com/questions/41633800/use-cognito-locally-offline
- https://stackoverflow.com/questions/51719765/invoke-lambda-function-without-authorization-on-local-serverless-offline-environ
|
True
|
[Feature Request] Cognito Offline Testing - Just curious whether SAM will support cognito offline testing? Pasting in some links to the same feature request in various other forums:
- https://github.com/localstack/localstack/issues/26
- https://stackoverflow.com/questions/41633800/use-cognito-locally-offline
- https://stackoverflow.com/questions/51719765/invoke-lambda-function-without-authorization-on-local-serverless-offline-environ
|
non_defect
|
cognito offline testing just curious whether sam will support cognito offline testing pasting in some links to the same feature request in various other forums
| 0
|
10,291
| 2,622,141,104
|
IssuesEvent
|
2015-03-04 00:02:03
|
byzhang/spserver
|
https://api.github.com/repos/byzhang/spserver
|
closed
|
关于超时,是怎么回事?
|
auto-migrated Priority-Medium Type-Defect
|
```
0.9.5的版本。
SP_IocpEventHelper :: timeoutNext 有以下代码:
DWORD ret = ( first->tv_sec - curr.tv_sec ) * 1000
+ ( first->tv_usec - curr.tv_usec ) / 1000;
if( ret < 0 ) ret = 0;
return ret;
ret 永远都会是>=0 的。
```
Original issue reported on code.google.com by `xChenJun` on 17 Jul 2010 at 4:04
|
1.0
|
关于超时,是怎么回事? - ```
0.9.5的版本。
SP_IocpEventHelper :: timeoutNext 有以下代码:
DWORD ret = ( first->tv_sec - curr.tv_sec ) * 1000
+ ( first->tv_usec - curr.tv_usec ) / 1000;
if( ret < 0 ) ret = 0;
return ret;
ret 永远都会是>=0 的。
```
Original issue reported on code.google.com by `xChenJun` on 17 Jul 2010 at 4:04
|
defect
|
关于超时,是怎么回事? 。 sp iocpeventhelper timeoutnext 有以下代码: dword ret first tv sec curr tv sec first tv usec curr tv usec if ret ret return ret ret 永远都会是 的。 original issue reported on code google com by xchenjun on jul at
| 1
|
33,840
| 7,267,597,152
|
IssuesEvent
|
2018-02-20 06:10:42
|
colour-science/colour
|
https://api.github.com/repos/colour-science/colour
|
closed
|
Python 3 Syntax Error in "docs/conf.py" file.
|
Defect Distribution Documentation Minor
|
flake8 testing of https://github.com/colour-science/colour on Python 3.6.4
$ __flake8 . --count --select=E901,E999,F821,F822,F823 --show-source --statistics__
```
./docs/conf.py:219:0: E999 SyntaxError: (unicode error) 'unicodeescape' codec can't decode bytes in position 9-10: truncated \uXXXX escape
""",
^
./utilities/unicode_to_ascii.py:9:1: F821 undefined name 'reload'
reload(sys)
^
1 E999 SyntaxError: (unicode error) 'unicodeescape' codec can't decode bytes in position 9-10: truncated \uXXXX escape
1 F821 undefined name 'reload'
2
```
|
1.0
|
Python 3 Syntax Error in "docs/conf.py" file. - flake8 testing of https://github.com/colour-science/colour on Python 3.6.4
$ __flake8 . --count --select=E901,E999,F821,F822,F823 --show-source --statistics__
```
./docs/conf.py:219:0: E999 SyntaxError: (unicode error) 'unicodeescape' codec can't decode bytes in position 9-10: truncated \uXXXX escape
""",
^
./utilities/unicode_to_ascii.py:9:1: F821 undefined name 'reload'
reload(sys)
^
1 E999 SyntaxError: (unicode error) 'unicodeescape' codec can't decode bytes in position 9-10: truncated \uXXXX escape
1 F821 undefined name 'reload'
2
```
|
defect
|
python syntax error in docs conf py file testing of on python count select show source statistics docs conf py syntaxerror unicode error unicodeescape codec can t decode bytes in position truncated uxxxx escape utilities unicode to ascii py undefined name reload reload sys syntaxerror unicode error unicodeescape codec can t decode bytes in position truncated uxxxx escape undefined name reload
| 1
|
471,417
| 13,566,326,802
|
IssuesEvent
|
2020-09-18 13:07:59
|
willowtreeapps/vocable-ios
|
https://api.github.com/repos/willowtreeapps/vocable-ios
|
closed
|
Able to save blank string to my sayings
|
bug priority - low v1.3.3
|
On the keyboard screen, the "my sayings" button should be disabled when no text is present
Steps to reproduce:
1. navigate to the keyboard screen
2. press the star (my sayings) button
expected:
- button is disabled
actual:
- blank string is saved to my sayings
|
1.0
|
Able to save blank string to my sayings - On the keyboard screen, the "my sayings" button should be disabled when no text is present
Steps to reproduce:
1. navigate to the keyboard screen
2. press the star (my sayings) button
expected:
- button is disabled
actual:
- blank string is saved to my sayings
|
non_defect
|
able to save blank string to my sayings on the keyboard screen the my sayings button should be disabled when no text is present steps to reproduce navigate to the keyboard screen press the star my sayings button expected button is disabled actual blank string is saved to my sayings
| 0
|
79,461
| 28,289,299,638
|
IssuesEvent
|
2023-04-09 01:53:08
|
HewlettPackard/yoda-demo
|
https://api.github.com/repos/HewlettPackard/yoda-demo
|
closed
|
Issue editing text on Mondays.
|
S3 - Medium T1 - Defect no-issue-activity
|
<p>Lorem ipsum nam nulla duis aenean facilisis venenatis viverra eleifend, sollicitudin sodales nisi primis sollicitudin mi ad.</p>
<p>Purus convallis lorem eget hac curabitur fusce duis himenaeos pulvinar maecenas, feugiat tempor fusce netus enim lacinia nisi et nisi.</p>
<p>At nisi elit vivamus feugiat orci non, tellus elit sagittis erat ac, sodales elit mauris aliquet fringilla.</p>
<p>Pharetra vitae habitant tristique scelerisque id mollis odio tempus, venenatis fusce arcu dictumst massa curabitur dolor.</p>
> estimate 5
|
1.0
|
Issue editing text on Mondays. - <p>Lorem ipsum nam nulla duis aenean facilisis venenatis viverra eleifend, sollicitudin sodales nisi primis sollicitudin mi ad.</p>
<p>Purus convallis lorem eget hac curabitur fusce duis himenaeos pulvinar maecenas, feugiat tempor fusce netus enim lacinia nisi et nisi.</p>
<p>At nisi elit vivamus feugiat orci non, tellus elit sagittis erat ac, sodales elit mauris aliquet fringilla.</p>
<p>Pharetra vitae habitant tristique scelerisque id mollis odio tempus, venenatis fusce arcu dictumst massa curabitur dolor.</p>
> estimate 5
|
defect
|
issue editing text on mondays lorem ipsum nam nulla duis aenean facilisis venenatis viverra eleifend sollicitudin sodales nisi primis sollicitudin mi ad purus convallis lorem eget hac curabitur fusce duis himenaeos pulvinar maecenas feugiat tempor fusce netus enim lacinia nisi et nisi at nisi elit vivamus feugiat orci non tellus elit sagittis erat ac sodales elit mauris aliquet fringilla pharetra vitae habitant tristique scelerisque id mollis odio tempus venenatis fusce arcu dictumst massa curabitur dolor estimate
| 1
|
53,786
| 13,262,298,332
|
IssuesEvent
|
2020-08-20 21:29:44
|
icecube-trac/tix4
|
https://api.github.com/repos/icecube-trac/tix4
|
closed
|
Assertion "`bin_edges[q+1]-bin_edges[q] > 0' failed." triggered in MillipedeDOMCacheMap.cxx:247 (Trac #2092)
|
Migrated from Trac combo reconstruction defect
|
Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion:
/home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed.
Unfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2092">https://code.icecube.wisc.edu/projects/icecube/ticket/2092</a>, reported by claudio.kopperand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:23",
"_ts": "1550067323910946",
"description": "Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion:\n\n/home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed.\n\nUnfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered?",
"reporter": "claudio.kopper",
"cc": "austin.schneider",
"resolution": "insufficient resources",
"time": "2017-09-15T16:30:44",
"component": "combo reconstruction",
"summary": "Assertion \"`bin_edges[q+1]-bin_edges[q] > 0' failed.\" triggered in MillipedeDOMCacheMap.cxx:247",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
Assertion "`bin_edges[q+1]-bin_edges[q] > 0' failed." triggered in MillipedeDOMCacheMap.cxx:247 (Trac #2092) - Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion:
/home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed.
Unfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered?
<details>
<summary><em>Migrated from <a href="https://code.icecube.wisc.edu/projects/icecube/ticket/2092">https://code.icecube.wisc.edu/projects/icecube/ticket/2092</a>, reported by claudio.kopperand owned by jvansanten</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2019-02-13T14:15:23",
"_ts": "1550067323910946",
"description": "Austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion:\n\n/home/aschneider/combo/source/millipede/private/millipede/MillipedeDOMCacheMap.cxx:247: int rebin_bayesian_blocks(int, const double *, const double *, bool *, double *, double *, double, double): Assertion `bin_edges[q+1]-bin_edges[q] > 0' failed.\n\nUnfortunately we do not have a minimal example right now (I guess `distribute` needs better reporting or we would need to add logging..), but maybe it is obvious why this assertion could be triggered?",
"reporter": "claudio.kopper",
"cc": "austin.schneider",
"resolution": "insufficient resources",
"time": "2017-09-15T16:30:44",
"component": "combo reconstruction",
"summary": "Assertion \"`bin_edges[q+1]-bin_edges[q] > 0' failed.\" triggered in MillipedeDOMCacheMap.cxx:247",
"priority": "normal",
"keywords": "",
"milestone": "",
"owner": "jvansanten",
"type": "defect"
}
```
</p>
</details>
|
defect
|
assertion bin edges bin edges failed triggered in millipededomcachemap cxx trac austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion home aschneider combo source millipede private millipede millipededomcachemap cxx int rebin bayesian blocks int const double const double bool double double double double assertion bin edges bin edges failed unfortunately we do not have a minimal example right now i guess distribute needs better reporting or we would need to add logging but maybe it is obvious why this assertion could be triggered migrated from json status closed changetime ts description austin was running a recent trunk of millipede for a distributed likelihood scan and encountered this assertion n n home aschneider combo source millipede private millipede millipededomcachemap cxx int rebin bayesian blocks int const double const double bool double double double double assertion bin edges bin edges failed n nunfortunately we do not have a minimal example right now i guess distribute needs better reporting or we would need to add logging but maybe it is obvious why this assertion could be triggered reporter claudio kopper cc austin schneider resolution insufficient resources time component combo reconstruction summary assertion bin edges bin edges failed triggered in millipededomcachemap cxx priority normal keywords milestone owner jvansanten type defect
| 1
|
69,622
| 15,021,855,863
|
IssuesEvent
|
2021-02-01 16:16:44
|
fabric8-services/fabric8-wit
|
https://api.github.com/repos/fabric8-services/fabric8-wit
|
closed
|
It should be possible to make PostgreSQL database connection secure through TLS
|
:snake: security database
|
PostgreSQL has TLS support. It should be possible to make PostgreSQL database connection secure through TLS. **The default configuration can continue to be disabled**.
BTW, TLS can be used with RDS also:
http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_PostgreSQL.html#PostgreSQL.Concepts.General.SSL
|
True
|
It should be possible to make PostgreSQL database connection secure through TLS - PostgreSQL has TLS support. It should be possible to make PostgreSQL database connection secure through TLS. **The default configuration can continue to be disabled**.
BTW, TLS can be used with RDS also:
http://docs.aws.amazon.com/AmazonRDS/latest/UserGuide/CHAP_PostgreSQL.html#PostgreSQL.Concepts.General.SSL
|
non_defect
|
it should be possible to make postgresql database connection secure through tls postgresql has tls support it should be possible to make postgresql database connection secure through tls the default configuration can continue to be disabled btw tls can be used with rds also
| 0
|
49,117
| 13,185,241,592
|
IssuesEvent
|
2020-08-12 21:00:25
|
icecube-trac/tix3
|
https://api.github.com/repos/icecube-trac/tix3
|
opened
|
[sim-services] (Trac #781)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
<details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/781
, reported by olivas and owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:03",
"description": "Make sure the sanity checkers are solid, all have tests, and the testing scripts are enabled.",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1458335643235016",
"component": "combo simulation",
"summary": "[sim-services]",
"priority": "critical",
"keywords": "",
"time": "2014-10-11T19:02:30",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
1.0
|
[sim-services] (Trac #781) - <details>
<summary><em>Migrated from https://code.icecube.wisc.edu/ticket/781
, reported by olivas and owned by olivas</em></summary>
<p>
```json
{
"status": "closed",
"changetime": "2016-03-18T21:14:03",
"description": "Make sure the sanity checkers are solid, all have tests, and the testing scripts are enabled.",
"reporter": "olivas",
"cc": "",
"resolution": "fixed",
"_ts": "1458335643235016",
"component": "combo simulation",
"summary": "[sim-services]",
"priority": "critical",
"keywords": "",
"time": "2014-10-11T19:02:30",
"milestone": "",
"owner": "olivas",
"type": "defect"
}
```
</p>
</details>
|
defect
|
trac migrated from reported by olivas and owned by olivas json status closed changetime description make sure the sanity checkers are solid all have tests and the testing scripts are enabled reporter olivas cc resolution fixed ts component combo simulation summary priority critical keywords time milestone owner olivas type defect
| 1
|
139,764
| 18,853,807,585
|
IssuesEvent
|
2021-11-12 01:46:30
|
LalithK90/aroComputers
|
https://api.github.com/repos/LalithK90/aroComputers
|
opened
|
CVE-2021-22118 (High) detected in spring-web-5.2.3.RELEASE.jar
|
security vulnerability
|
## CVE-2021-22118 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.2.3.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: aroComputers/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-web/5.2.3.RELEASE/dd386a02e40b915ab400a3bf9f586d2dc4c0852c/spring-web-5.2.3.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.4.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.2.4.RELEASE.jar
- :x: **spring-web-5.2.3.RELEASE.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework, versions 5.2.x prior to 5.2.15 and versions 5.3.x prior to 5.3.7, a WebFlux application is vulnerable to a privilege escalation: by (re)creating the temporary storage directory, a locally authenticated malicious user can read or modify files that have been uploaded to the WebFlux application, or overwrite arbitrary files with multipart request data.
<p>Publish Date: 2021-05-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22118>CVE-2021-22118</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22118">https://tanzu.vmware.com/security/cve-2021-22118</a></p>
<p>Release Date: 2021-05-27</p>
<p>Fix Resolution: org.springframework:spring-web:5.2.15,5.3.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-22118 (High) detected in spring-web-5.2.3.RELEASE.jar - ## CVE-2021-22118 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-web-5.2.3.RELEASE.jar</b></p></summary>
<p>Spring Web</p>
<p>Library home page: <a href="https://github.com/spring-projects/spring-framework">https://github.com/spring-projects/spring-framework</a></p>
<p>Path to dependency file: aroComputers/build.gradle</p>
<p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.springframework/spring-web/5.2.3.RELEASE/dd386a02e40b915ab400a3bf9f586d2dc4c0852c/spring-web-5.2.3.RELEASE.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.2.4.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.2.4.RELEASE.jar
- :x: **spring-web-5.2.3.RELEASE.jar** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
In Spring Framework, versions 5.2.x prior to 5.2.15 and versions 5.3.x prior to 5.3.7, a WebFlux application is vulnerable to a privilege escalation: by (re)creating the temporary storage directory, a locally authenticated malicious user can read or modify files that have been uploaded to the WebFlux application, or overwrite arbitrary files with multipart request data.
<p>Publish Date: 2021-05-27
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-22118>CVE-2021-22118</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Local
- Attack Complexity: Low
- Privileges Required: Low
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://tanzu.vmware.com/security/cve-2021-22118">https://tanzu.vmware.com/security/cve-2021-22118</a></p>
<p>Release Date: 2021-05-27</p>
<p>Fix Resolution: org.springframework:spring-web:5.2.15,5.3.7</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in spring web release jar cve high severity vulnerability vulnerable library spring web release jar spring web library home page a href path to dependency file arocomputers build gradle path to vulnerable library home wss scanner gradle caches modules files org springframework spring web release spring web release jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x spring web release jar vulnerable library found in base branch master vulnerability details in spring framework versions x prior to and versions x prior to a webflux application is vulnerable to a privilege escalation by re creating the temporary storage directory a locally authenticated malicious user can read or modify files that have been uploaded to the webflux application or overwrite arbitrary files with multipart request data publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring web step up your open source security game with whitesource
| 0
|
63,066
| 17,365,970,475
|
IssuesEvent
|
2021-07-30 07:19:44
|
vector-im/element-web
|
https://api.github.com/repos/vector-im/element-web
|
closed
|
no UI when switching room whilst setting up a voice/video call
|
A-VoIP P1 S-Tolerable T-Defect
|
If you set up a voice/video call and whilst it's ringing switch to a different room, there is zero UI to tell you what's going on. We should at least show the 'ongoing call' UI at the top of LeftPanel
|
1.0
|
no UI when switching room whilst setting up a voice/video call - If you set up a voice/video call and whilst it's ringing switch to a different room, there is zero UI to tell you what's going on. We should at least show the 'ongoing call' UI at the top of LeftPanel
|
defect
|
no ui when switching room whilst setting up a voice video call if you set up a voice video call and whilst it s ringing switch to a different room there is zero ui to tell you what s going on we should at least show the ongoing call ui at the top of leftpanel
| 1
|
56,190
| 23,720,740,872
|
IssuesEvent
|
2022-08-30 15:09:04
|
hashicorp/terraform-provider-azurerm
|
https://api.github.com/repos/hashicorp/terraform-provider-azurerm
|
closed
|
Error: no schema available for azurerm_traffic_manager_endpoint.secondary_endpoint[0] while reading state;
|
question service/traffic-manager
|
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.2.8
### AzureRM Provider Version
3.19.1
### Affected Resource(s)/Data Source(s)
traffic_manager_endpoint
### Terraform Configuration Files
```hcl
resource "azurerm_traffic_manager_azure_endpoint" "primary_endpoint" {
count = var.env == "prod" || var.env == "uat" ? 1 : 0
name = "Primary"
profile_id = azurerm_traffic_manager_profile.gsr_spoke_tm[0].id
weight = 1
priority = 1
target_resource_id = data.azurerm_public_ip.gsr_primary_appgw_uksouth_ip[0].id
}
resource "azurerm_traffic_manager_azure_endpoint" "secondary_endpoint" {
count = var.env == "prod" ? 1 : 0
name = "Secondary"
profile_id = azurerm_traffic_manager_profile.gsr_spoke_tm[0].id
weight = 1
priority = 2
target_resource_id = data.azurerm_public_ip.gsr_secondary_appgw_ukwest_ip[0].id
}
```
### Debug Output/Panic Output
```shell
Error: no schema available for azurerm_traffic_manager_endpoint.secondary_endpoint[0] while reading state; this is a bug in Terraform and should be reported
```
### Expected Behaviour
_No response_
### Actual Behaviour
_No response_
### Steps to Reproduce
_No response_
### Important Factoids
_No response_
### References
_No response_
|
1.0
|
Error: no schema available for azurerm_traffic_manager_endpoint.secondary_endpoint[0] while reading state; - ### Is there an existing issue for this?
- [X] I have searched the existing issues
### Community Note
<!--- Please keep this note for the community --->
* Please vote on this issue by adding a :thumbsup: [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request
* Please do not leave "+1" or "me too" comments, they generate extra noise for issue followers and do not help prioritize the request
* If you are interested in working on this issue or have submitted a pull request, please leave a comment
<!--- Thank you for keeping this note for the community --->
### Terraform Version
1.2.8
### AzureRM Provider Version
3.19.1
### Affected Resource(s)/Data Source(s)
traffic_manager_endpoint
### Terraform Configuration Files
```hcl
resource "azurerm_traffic_manager_azure_endpoint" "primary_endpoint" {
count = var.env == "prod" || var.env == "uat" ? 1 : 0
name = "Primary"
profile_id = azurerm_traffic_manager_profile.gsr_spoke_tm[0].id
weight = 1
priority = 1
target_resource_id = data.azurerm_public_ip.gsr_primary_appgw_uksouth_ip[0].id
}
resource "azurerm_traffic_manager_azure_endpoint" "secondary_endpoint" {
count = var.env == "prod" ? 1 : 0
name = "Secondary"
profile_id = azurerm_traffic_manager_profile.gsr_spoke_tm[0].id
weight = 1
priority = 2
target_resource_id = data.azurerm_public_ip.gsr_secondary_appgw_ukwest_ip[0].id
}
```
### Debug Output/Panic Output
```shell
Error: no schema available for azurerm_traffic_manager_endpoint.secondary_endpoint[0] while reading state; this is a bug in Terraform and should be reported
```
### Expected Behaviour
_No response_
### Actual Behaviour
_No response_
### Steps to Reproduce
_No response_
### Important Factoids
_No response_
### References
_No response_
|
non_defect
|
error no schema available for azurerm traffic manager endpoint secondary endpoint while reading state is there an existing issue for this i have searched the existing issues community note please vote on this issue by adding a thumbsup to the original issue to help the community and maintainers prioritize this request please do not leave or me too comments they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment terraform version azurerm provider version affected resource s data source s traffic manager endpoint terraform configuration files hcl resource azurerm traffic manager azure endpoint primary endpoint count var env prod var env uat name primary profile id azurerm traffic manager profile gsr spoke tm id weight priority target resource id data azurerm public ip gsr primary appgw uksouth ip id resource azurerm traffic manager azure endpoint secondary endpoint count var env prod name secondary profile id azurerm traffic manager profile gsr spoke tm id weight priority target resource id data azurerm public ip gsr secondary appgw ukwest ip id debug output panic output shell error no schema available for azurerm traffic manager endpoint secondary endpoint while reading state this is a bug in terraform and should be reported expected behaviour no response actual behaviour no response steps to reproduce no response important factoids no response references no response
| 0
|
18,320
| 24,438,849,687
|
IssuesEvent
|
2022-10-06 13:22:41
|
Ultimate-Hosts-Blacklist/whitelist
|
https://api.github.com/repos/Ultimate-Hosts-Blacklist/whitelist
|
opened
|
[FALSE-POSITIVE?]
|
whitelisting process
|
**Domains or links**
js.pusher.com
**More Information**
How did you discover your web site or domain was listed here?
2. Reported by another user.
**Have you requested removal from other sources?**
No
**Additional context**
Add any other context about the problem here.
:exclamation:
We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
|
1.0
|
[FALSE-POSITIVE?] - **Domains or links**
js.pusher.com
**More Information**
How did you discover your web site or domain was listed here?
2. Reported by another user.
**Have you requested removal from other sources?**
No
**Additional context**
Add any other context about the problem here.
:exclamation:
We understand being listed on a list like this can be frustrating and embarrassing for many web site owners. The first step is to remain calm. The second step is to rest assured one of our maintainers will address your issue as soon as possible. Please make sure you have provided as much information as possible to help speed up the process.
|
non_defect
|
domains or links js pusher com more information how did you discover your web site or domain was listed here reported by another user have you requested removal from other sources no additional context add any other context about the problem here exclamation we understand being listed on a list like this can be frustrating and embarrassing for many web site owners the first step is to remain calm the second step is to rest assured one of our maintainers will address your issue as soon as possible please make sure you have provided as much information as possible to help speed up the process
| 0
|
56,435
| 15,091,694,926
|
IssuesEvent
|
2021-02-06 16:31:56
|
dkfurrow/moneydance-investment-reports
|
https://api.github.com/repos/dkfurrow/moneydance-investment-reports
|
closed
|
Test initial issue report
|
Priority-Medium Type-Defect auto-migrated
|
```
Test initial issue report
```
Original issue reported on code.google.com by `dkfur...@gmail.com` on 23 Dec 2010 at 4:22
|
1.0
|
Test initial issue report - ```
Test initial issue report
```
Original issue reported on code.google.com by `dkfur...@gmail.com` on 23 Dec 2010 at 4:22
|
defect
|
test initial issue report test initial issue report original issue reported on code google com by dkfur gmail com on dec at
| 1
|
6,242
| 2,610,223,969
|
IssuesEvent
|
2015-02-26 19:10:59
|
chrsmith/somefinders
|
https://api.github.com/repos/chrsmith/somefinders
|
opened
|
kh 1171 инструкция на русском
|
auto-migrated Priority-Medium Type-Defect
|
```
'''Антонин Евсеев'''
Привет всем не подскажите где можно найти
.kh 1171 инструкция на русском. где то видел уже
'''Боеслав Овчинников'''
Вот держи линк http://bit.ly/16T1P32
'''Витольд Козлов'''
Спасибо вроде то но просит телефон вводить
'''Анастасий Калинин'''
Не это не влияет на баланс
'''Аверкий Пахомов'''
Неа все ок у меня ничего не списало
Информация о файле: kh 1171 инструкция на
русском
Загружен: В этом месяце
Скачан раз: 425
Рейтинг: 272
Средняя скорость скачивания: 693
Похожих файлов: 37
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 8:29
|
1.0
|
kh 1171 инструкция на русском - ```
'''Антонин Евсеев'''
Привет всем не подскажите где можно найти
.kh 1171 инструкция на русском. где то видел уже
'''Боеслав Овчинников'''
Вот держи линк http://bit.ly/16T1P32
'''Витольд Козлов'''
Спасибо вроде то но просит телефон вводить
'''Анастасий Калинин'''
Не это не влияет на баланс
'''Аверкий Пахомов'''
Неа все ок у меня ничего не списало
Информация о файле: kh 1171 инструкция на
русском
Загружен: В этом месяце
Скачан раз: 425
Рейтинг: 272
Средняя скорость скачивания: 693
Похожих файлов: 37
```
-----
Original issue reported on code.google.com by `kondense...@gmail.com` on 18 Dec 2013 at 8:29
|
defect
|
kh инструкция на русском антонин евсеев привет всем не подскажите где можно найти kh инструкция на русском где то видел уже боеслав овчинников вот держи линк витольд козлов спасибо вроде то но просит телефон вводить анастасий калинин не это не влияет на баланс аверкий пахомов неа все ок у меня ничего не списало информация о файле kh инструкция на русском загружен в этом месяце скачан раз рейтинг средняя скорость скачивания похожих файлов original issue reported on code google com by kondense gmail com on dec at
| 1
|
32,647
| 6,884,884,161
|
IssuesEvent
|
2017-11-21 14:31:29
|
tpfinal-pp1/tp-final
|
https://api.github.com/repos/tpfinal-pp1/tp-final
|
reopened
|
Contrato: Error en fecha de ingreso
|
bug Defecto medio Liberado por desarrollo
|
No deberia pasar (a mi criterio) que la fecha de celebracion sea despues de la fecha de ingreso 
|
1.0
|
Contrato: Error en fecha de ingreso - No deberia pasar (a mi criterio) que la fecha de celebracion sea despues de la fecha de ingreso 
|
defect
|
contrato error en fecha de ingreso no deberia pasar a mi criterio que la fecha de celebracion sea despues de la fecha de ingreso
| 1
|
18,283
| 3,040,177,737
|
IssuesEvent
|
2015-08-07 14:04:11
|
googlei18n/noto-fonts
|
https://api.github.com/repos/googlei18n/noto-fonts
|
opened
|
Imperial Aramaic: Tip of 100 and 10,000 characters should point right
|
Script-ImperialAramaic Type-Defect
|
moved from https://github.com/googlei18n/noto-alpha/issues/191
Imported from [Google Code issue #191](https://code.google.com/p/noto-alpha/issues/detail?id=191) created by [roozbeh@google.com](https://code.google.com/u/roozbeh@google.com/) on 2014-04-02T22:12:45.000Z:
----
The tip of ONE HUNDRED (U+1085D) and TEN THOUSAND (U+1085F) characters currently point to the left side. They should point to the right.
Compare with Figures 13 and 14 of the Unicode proposal for Imperial Aramaic, where attestation of the numbers are given: http://www.unicode.org/L2/L2007/07197r2-n3273r2-aramaic.pdf
Also, compare with the Unicode chart.
|
1.0
|
Imperial Aramaic: Tip of 100 and 10,000 characters should point right - moved from https://github.com/googlei18n/noto-alpha/issues/191
Imported from [Google Code issue #191](https://code.google.com/p/noto-alpha/issues/detail?id=191) created by [roozbeh@google.com](https://code.google.com/u/roozbeh@google.com/) on 2014-04-02T22:12:45.000Z:
----
The tip of ONE HUNDRED (U+1085D) and TEN THOUSAND (U+1085F) characters currently point to the left side. They should point to the right.
Compare with Figures 13 and 14 of the Unicode proposal for Imperial Aramaic, where attestation of the numbers are given: http://www.unicode.org/L2/L2007/07197r2-n3273r2-aramaic.pdf
Also, compare with the Unicode chart.
|
defect
|
imperial aramaic tip of and characters should point right moved from imported from created by on the tip of one hundred u and ten thousand u characters currently point to the left side they should point to the right compare with figures and of the unicode proposal for imperial aramaic where attestation of the numbers are given also compare with the unicode chart
| 1
|
72,903
| 24,351,475,056
|
IssuesEvent
|
2022-10-03 00:45:50
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
closed
|
BUG: 1.9.0rc1: `OptimizeResult` not populated when `optimize.milp` runs into `time_limit` even though a feasible solution was found.
|
defect scipy.optimize
|
### Describe your issue.
When `optimize.milp` terminates due to the time limit being reached, the `OptimizeResult` object is not populated, even though one or more feasible solutions have been found.
Example HiGHS log below shows that a feasible solution has been found (see solution status) when the 1 second time limit is reached.
[HiGHS log.txt](https://github.com/scipy/scipy/files/9039809/HiGHS.log.txt)
but the corresponding `OptimizeResult` object (see "Error message" section) is more or less empty and the only thing that is propagated correctly is the `message`, saying that the time limit was reached.
### Reproducing Code Example
```python
import numpy as np
from random import randint
from scipy.optimize import milp, LinearConstraint, Bounds
A = np.zeros((100, 100))
for c in range(100):
for v in range(100):
A[c, v] = randint(0, 5)
b_lb = [-1*np.inf for c in range(100)]
b_ub = [25 for c in range(100)]
constraints = LinearConstraint(A, b_lb, b_ub)
variable_lb = [0 for v in range(100)]
variable_ub = [1 for v in range(100)]
variable_bounds = Bounds(variable_lb, variable_ub)
integrality = [1 for v in range(100)]
c_vector = 100 * [-1]
res = milp(
c_vector,
integrality=integrality,
bounds=variable_bounds,
constraints=constraints,
options={"time_limit": 1, "disp": True}
)
```
### Error message
```shell
{'fun': None,
'message': 'Time limit reached. (HiGHS Status 13: model_status is Time limit '
"reached; primal_status is b'At upper bound')",
'mip_dual_bound': None,
'mip_gap': None,
'mip_node_count': None,
'status': 1,
'success': False,
'x': None}
```
### SciPy/NumPy/Python version information
1.9.0rc1 1.23.0 sys.version_info(major=3, minor=9, micro=13, releaselevel='final', serial=0)
|
1.0
|
BUG: 1.9.0rc1: `OptimizeResult` not populated when `optimize.milp` runs into `time_limit` even though a feasible solution was found. - ### Describe your issue.
When `optimize.milp` terminates due to the time limit being reached, the `OptimizeResult` object is not populated, even though one or more feasible solutions have been found.
Example HiGHS log below shows that a feasible solution has been found (see solution status) when the 1 second time limit is reached.
[HiGHS log.txt](https://github.com/scipy/scipy/files/9039809/HiGHS.log.txt)
but the corresponding `OptimizeResult` object (see "Error message" section) is more or less empty and the only thing that is propagated correctly is the `message`, saying that the time limit was reached.
### Reproducing Code Example
```python
import numpy as np
from random import randint
from scipy.optimize import milp, LinearConstraint, Bounds
A = np.zeros((100, 100))
for c in range(100):
for v in range(100):
A[c, v] = randint(0, 5)
b_lb = [-1*np.inf for c in range(100)]
b_ub = [25 for c in range(100)]
constraints = LinearConstraint(A, b_lb, b_ub)
variable_lb = [0 for v in range(100)]
variable_ub = [1 for v in range(100)]
variable_bounds = Bounds(variable_lb, variable_ub)
integrality = [1 for v in range(100)]
c_vector = 100 * [-1]
res = milp(
c_vector,
integrality=integrality,
bounds=variable_bounds,
constraints=constraints,
options={"time_limit": 1, "disp": True}
)
```
### Error message
```shell
{'fun': None,
'message': 'Time limit reached. (HiGHS Status 13: model_status is Time limit '
"reached; primal_status is b'At upper bound')",
'mip_dual_bound': None,
'mip_gap': None,
'mip_node_count': None,
'status': 1,
'success': False,
'x': None}
```
### SciPy/NumPy/Python version information
1.9.0rc1 1.23.0 sys.version_info(major=3, minor=9, micro=13, releaselevel='final', serial=0)
|
defect
|
bug optimizeresult not populated when optimize milp runs into time limit even though a feasible solution was found describe your issue when optimize milp terminates due to the time limit being reached the optimizeresult object is not populated even though one or more feasible solutions have been found example highs log below shows that a feasible solution has been found see solution status when the second time limit is reached but the corresponding optimizeresult object see error message section is more or less empty and the only thing that is propagated correctly is the message saying that the time limit was reached reproducing code example python import numpy as np from random import randint from scipy optimize import milp linearconstraint bounds a np zeros for c in range for v in range a randint b lb b ub constraints linearconstraint a b lb b ub variable lb variable ub variable bounds bounds variable lb variable ub integrality c vector res milp c vector integrality integrality bounds variable bounds constraints constraints options time limit disp true error message shell fun none message time limit reached highs status model status is time limit reached primal status is b at upper bound mip dual bound none mip gap none mip node count none status success false x none scipy numpy python version information sys version info major minor micro releaselevel final serial
| 1
|
8,951
| 2,612,934,070
|
IssuesEvent
|
2015-02-27 17:35:53
|
skelterjohn/go-wiki
|
https://api.github.com/repos/skelterjohn/go-wiki
|
closed
|
Broken url to video Practical_Go_Programming
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. On page
http://code.google.com/p/go-wiki/wiki/GoTalks#Practical_Go_Programming url to
video and slides is broken
2. URL to video at this moment is
http://www.youtube.com/watch?v=2-pPAvqylu but should be:
http://www.youtube.com/watch?v=2-pPAvqyluI
```
Original issue reported on code.google.com by `Dey...@gmail.com` on 15 Dec 2012 at 7:53
|
1.0
|
Broken url to video Practical_Go_Programming - ```
What steps will reproduce the problem?
1. On page
http://code.google.com/p/go-wiki/wiki/GoTalks#Practical_Go_Programming url to
video and slides is broken
2. URL to video at this moment is
http://www.youtube.com/watch?v=2-pPAvqylu but should be:
http://www.youtube.com/watch?v=2-pPAvqyluI
```
Original issue reported on code.google.com by `Dey...@gmail.com` on 15 Dec 2012 at 7:53
|
defect
|
broken url to video practical go programming what steps will reproduce the problem on page url to video and slides is broken url to video at this moment is but should be original issue reported on code google com by dey gmail com on dec at
| 1
|
57,580
| 15,866,618,468
|
IssuesEvent
|
2021-04-08 15:56:38
|
NREL/EnergyPlus
|
https://api.github.com/repos/NREL/EnergyPlus
|
closed
|
Airflow curve input defect in AirloopHVAC:UnitarySystem when MS DX Heating coils operate at Speed = 1
|
Defect
|
Issue overview
--------------
PR #8501 fixed the issue #8335 to pass correct mass flow rate to calculate flow fraction for DX cooling coil performance curves at Speed = 1. However, the same problem occurs for a MS DX heating coil at speed=1, when a UnitarySystem is used as a parent. The input values vary with node mass flow rate as show below.

It is expected to get a constant input value.
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus (if using an intermediate build, include SHA)
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [X] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
1.0
|
Airflow curve input defect in AirloopHVAC:UnitarySystem when MS DX Heating coils operate at Speed = 1 - Issue overview
--------------
PR #8501 fixed the issue #8335 to pass correct mass flow rate to calculate flow fraction for DX cooling coil performance curves at Speed = 1. However, the same problem occurs for a MS DX heating coil at speed=1, when a UnitarySystem is used as a parent. The input values vary with node mass flow rate as show below.

It is expected to get a constant input value.
### Details
Some additional details for this issue (if relevant):
- Platform (Operating system, version)
- Version of EnergyPlus (if using an intermediate build, include SHA)
- Unmethours link or helpdesk ticket number
### Checklist
Add to this list or remove from it as applicable. This is a simple templated set of guidelines.
- [X] Defect file added (list location of defect file here)
- [ ] Ticket added to Pivotal for defect (development team task)
- [ ] Pull request created (the pull request will have additional tasks related to reviewing changes that fix this defect)
|
defect
|
airflow curve input defect in airloophvac unitarysystem when ms dx heating coils operate at speed issue overview pr fixed the issue to pass correct mass flow rate to calculate flow fraction for dx cooling coil performance curves at speed however the same problem occurs for a ms dx heating coil at speed when a unitarysystem is used as a parent the input values vary with node mass flow rate as show below it is expected to get a constant input value details some additional details for this issue if relevant platform operating system version version of energyplus if using an intermediate build include sha unmethours link or helpdesk ticket number checklist add to this list or remove from it as applicable this is a simple templated set of guidelines defect file added list location of defect file here ticket added to pivotal for defect development team task pull request created the pull request will have additional tasks related to reviewing changes that fix this defect
| 1
|
19,710
| 3,248,226,238
|
IssuesEvent
|
2015-10-17 04:13:45
|
jimradford/superputty
|
https://api.github.com/repos/jimradford/superputty
|
closed
|
Automatic logon with -pw password
|
auto-migrated Priority-Medium Type-Defect
|
```
See the attached screenshot for more details!
sp version: 1.3.0.11 / Windows 7 Pro. SP1
```
Original issue reported on code.google.com by `strugare...@gmail.com` on 23 Nov 2012 at 2:19
Attachments:
* [sperror.png](https://storage.googleapis.com/google-code-attachments/superputty/issue-258/comment-0/sperror.png)
|
1.0
|
Automatic logon with -pw password - ```
See the attached screenshot for more details!
sp version: 1.3.0.11 / Windows 7 Pro. SP1
```
Original issue reported on code.google.com by `strugare...@gmail.com` on 23 Nov 2012 at 2:19
Attachments:
* [sperror.png](https://storage.googleapis.com/google-code-attachments/superputty/issue-258/comment-0/sperror.png)
|
defect
|
automatic logon with pw password see the attached screenshot for more details sp version windows pro original issue reported on code google com by strugare gmail com on nov at attachments
| 1
|
39,620
| 9,565,486,443
|
IssuesEvent
|
2019-05-05 12:55:58
|
jmc2obj/j-mc-2-obj
|
https://api.github.com/repos/jmc2obj/j-mc-2-obj
|
closed
|
r302 Launching for OS X
|
Priority-Medium Type-Defect auto-migrated
|
```
What steps will reproduce the problem?
1. Have Apple OS X Mountain Lion 10.8.5 with supplement update
2. Open the .jar file by double clicking or right click - Jar Launcher
3. Open the .jar file by following steps from Issue 78 in terminal.
What is the expected output? What do you see instead?
A conf file to be generated and the application GUI to display, but instead an
error report stating the jar file could not be launched.
What version of the product are you using? On what operating system?
jMc2Obj-dev_r302.jar on OS X 10.8.5 w/supplemental update
Please describe your error as detailed as possible.
Error Pop-up (Double Click): The Java JAR file "jMc2Obj-dev_r302.jar" could not
be launched. Check the Console for possible error messages.
Error in Terminal: new-host-6:Desktop Michael$ java -jar jMc2Obj-dev_r302.jar
Exception in thread "main" java.lang.UnsupportedClassVersionError: org/jmc/Main
: Unsupported major.minor version 51.0
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClassCond(ClassLoader.java:637)
at java.lang.ClassLoader.defineClass(ClassLoader.java:621)
at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
Please provide any additional information below.
Checked Console and found no errors, checked terminal and received the above
list. Also, r301.jar was working 100% beforehand (no resizing for GUI to
update, etc), with the exception of missing some blocks in MC 1.7.x. Initial
test was done through the built-in GUI updater, and then direct download from
the site.
I've also read on you GUI Redesign Reddit thread that there have been problems
with the GUI engine on OS X. I'm not sure if this is one of those issues. I
hope this helps! I also apologize for all the issues I've given you guys.
```
Original issue reported on code.google.com by `rochefor...@gmail.com` on 7 Jan 2014 at 11:43
Attachments:
- [Screen Shot 2014-01-07 at 6.41.15 PM.png](https://storage.googleapis.com/google-code-attachments/j-mc-2-obj/issue-81/comment-0/Screen Shot 2014-01-07 at 6.41.15 PM.png)
|
1.0
|
r302 Launching for OS X - ```
What steps will reproduce the problem?
1. Have Apple OS X Mountain Lion 10.8.5 with supplement update
2. Open the .jar file by double clicking or right click - Jar Launcher
3. Open the .jar file by following steps from Issue 78 in terminal.
What is the expected output? What do you see instead?
A conf file to be generated and the application GUI to display, but instead an
error report stating the jar file could not be launched.
What version of the product are you using? On what operating system?
jMc2Obj-dev_r302.jar on OS X 10.8.5 w/supplemental update
Please describe your error as detailed as possible.
Error Pop-up (Double Click): The Java JAR file "jMc2Obj-dev_r302.jar" could not
be launched. Check the Console for possible error messages.
Error in Terminal: new-host-6:Desktop Michael$ java -jar jMc2Obj-dev_r302.jar
Exception in thread "main" java.lang.UnsupportedClassVersionError: org/jmc/Main
: Unsupported major.minor version 51.0
at java.lang.ClassLoader.defineClass1(Native Method)
at java.lang.ClassLoader.defineClassCond(ClassLoader.java:637)
at java.lang.ClassLoader.defineClass(ClassLoader.java:621)
at java.security.SecureClassLoader.defineClass(SecureClassLoader.java:141)
at java.net.URLClassLoader.defineClass(URLClassLoader.java:283)
at java.net.URLClassLoader.access$000(URLClassLoader.java:58)
at java.net.URLClassLoader$1.run(URLClassLoader.java:197)
at java.security.AccessController.doPrivileged(Native Method)
at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
Please provide any additional information below.
Checked Console and found no errors, checked terminal and received the above
list. Also, r301.jar was working 100% beforehand (no resizing for GUI to
update, etc), with the exception of missing some blocks in MC 1.7.x. Initial
test was done through the built-in GUI updater, and then direct download from
the site.
I've also read on you GUI Redesign Reddit thread that there have been problems
with the GUI engine on OS X. I'm not sure if this is one of those issues. I
hope this helps! I also apologize for all the issues I've given you guys.
```
Original issue reported on code.google.com by `rochefor...@gmail.com` on 7 Jan 2014 at 11:43
Attachments:
- [Screen Shot 2014-01-07 at 6.41.15 PM.png](https://storage.googleapis.com/google-code-attachments/j-mc-2-obj/issue-81/comment-0/Screen Shot 2014-01-07 at 6.41.15 PM.png)
|
defect
|
launching for os x what steps will reproduce the problem have apple os x mountain lion with supplement update open the jar file by double clicking or right click jar launcher open the jar file by following steps from issue in terminal what is the expected output what do you see instead a conf file to be generated and the application gui to display but instead an error report stating the jar file could not be launched what version of the product are you using on what operating system dev jar on os x w supplemental update please describe your error as detailed as possible error pop up double click the java jar file dev jar could not be launched check the console for possible error messages error in terminal new host desktop michael java jar dev jar exception in thread main java lang unsupportedclassversionerror org jmc main unsupported major minor version at java lang classloader native method at java lang classloader defineclasscond classloader java at java lang classloader defineclass classloader java at java security secureclassloader defineclass secureclassloader java at java net urlclassloader defineclass urlclassloader java at java net urlclassloader access urlclassloader java at java net urlclassloader run urlclassloader java at java security accesscontroller doprivileged native method at java net urlclassloader findclass urlclassloader java at java lang classloader loadclass classloader java at sun misc launcher appclassloader loadclass launcher java at java lang classloader loadclass classloader java please provide any additional information below checked console and found no errors checked terminal and received the above list also jar was working beforehand no resizing for gui to update etc with the exception of missing some blocks in mc x initial test was done through the built in gui updater and then direct download from the site i ve also read on you gui redesign reddit thread that there have been problems with the gui engine on os x i m not sure if this is one of those issues i hope this helps i also apologize for all the issues i ve given you guys original issue reported on code google com by rochefor gmail com on jan at attachments shot at pm png
| 1
|
41,286
| 10,354,590,144
|
IssuesEvent
|
2019-09-05 14:02:16
|
vector-im/riot-web
|
https://api.github.com/repos/vector-im/riot-web
|
closed
|
Terms account data is meant to be additive, but currently sets only new URLs
|
bug defect phase:1 privacy privacy-sprint
|
We're meant to keep adding newly agreed URLs to account data, but a bug means we're only storing the ones you just agreed to this time.
|
1.0
|
Terms account data is meant to be additive, but currently sets only new URLs - We're meant to keep adding newly agreed URLs to account data, but a bug means we're only storing the ones you just agreed to this time.
|
defect
|
terms account data is meant to be additive but currently sets only new urls we re meant to keep adding newly agreed urls to account data but a bug means we re only storing the ones you just agreed to this time
| 1
|
23,793
| 3,851,867,809
|
IssuesEvent
|
2016-04-06 05:28:58
|
GPF/imame4all
|
https://api.github.com/repos/GPF/imame4all
|
closed
|
Button issues with Samurai Showdown 4 or maybe all NEO GEO games?
|
auto-migrated Priority-Medium Type-Defect
|
```
What steps will reproduce the problem?
1. Load Samurai Showdown 4
2. Enable 4-buttons or All buttons
3. Test all buttons.
What is the expected output? What do you see instead?
Samurai Showdown uses 4 buttons. A,B,C,D for LP, MP, HP, K.
In MAME4Droid, button L,R is not used. A is HP, B is LP, X is MP. Y opens up
NEO GEO GENIE. The button for Kick is missing.
Also, the Virtual button for Select is the coin button. But the onscreen coin
button is not working.
What version of the product are you using? On what operating system?
1.5 Android.
Please provide any additional information below.
I wonder about NEO GEO GENIE whether you can disable it. Maybe it is
conflicting with the last button for Neo Geo. esp since the portrait mode of
MAME4Droid only displays 4 buttons on my Galaxy S2.
I have only managed to test Samurai Showdown 4 since that was the only game I
found and the only one I really wanted to play.
```
Original issue reported on code.google.com by `alloutfo...@gmail.com` on 20 Dec 2011 at 8:16
|
1.0
|
Button issues with Samurai Showdown 4 or maybe all NEO GEO games? - ```
What steps will reproduce the problem?
1. Load Samurai Showdown 4
2. Enable 4-buttons or All buttons
3. Test all buttons.
What is the expected output? What do you see instead?
Samurai Showdown uses 4 buttons. A,B,C,D for LP, MP, HP, K.
In MAME4Droid, button L,R is not used. A is HP, B is LP, X is MP. Y opens up
NEO GEO GENIE. The button for Kick is missing.
Also, the Virtual button for Select is the coin button. But the onscreen coin
button is not working.
What version of the product are you using? On what operating system?
1.5 Android.
Please provide any additional information below.
I wonder about NEO GEO GENIE whether you can disable it. Maybe it is
conflicting with the last button for Neo Geo. esp since the portrait mode of
MAME4Droid only displays 4 buttons on my Galaxy S2.
I have only managed to test Samurai Showdown 4 since that was the only game I
found and the only one I really wanted to play.
```
Original issue reported on code.google.com by `alloutfo...@gmail.com` on 20 Dec 2011 at 8:16
|
defect
|
button issues with samurai showdown or maybe all neo geo games what steps will reproduce the problem load samurai showdown enable buttons or all buttons test all buttons what is the expected output what do you see instead samurai showdown uses buttons a b c d for lp mp hp k in button l r is not used a is hp b is lp x is mp y opens up neo geo genie the button for kick is missing also the virtual button for select is the coin button but the onscreen coin button is not working what version of the product are you using on what operating system android please provide any additional information below i wonder about neo geo genie whether you can disable it maybe it is conflicting with the last button for neo geo esp since the portrait mode of only displays buttons on my galaxy i have only managed to test samurai showdown since that was the only game i found and the only one i really wanted to play original issue reported on code google com by alloutfo gmail com on dec at
| 1
|
136,820
| 11,088,758,520
|
IssuesEvent
|
2019-12-14 13:40:08
|
BiancoRoyal/node-red-contrib-modbus
|
https://api.github.com/repos/BiancoRoyal/node-red-contrib-modbus
|
closed
|
Read Node stopped working
|
fixed pull request welcome testing unexpected behaviour
|
<!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MIGHT CLOSE YOUR ISSUE WITHOUT INVESTIGATING
PLEASE, DELETE NOT NEEDED PARTS FROM THAT ISSUE!
-->
**I'm submitting a ...** (check one with "x")
## Problem
- ( ) bug report => please search github for a similar issue or PR and test Modbus problems with the [modbus-serial](https://github.com/yaacov/node-modbus-serial) package before submitting
- (X) unexpected behavior => please visit the Wiki before
**Current behavior**
<!-- Describe how the bug or problem manifests. -->
Modbus Read node stopped working
**Expected behavior**
<!-- Describe what the behavior would be without the bug or problem. -->
Gets readings from modes devices
### Your Environment
**Please tell us about your environment:**
<!-- system, Modbus server or client type, package manager, equipment, vendor info -->
RPi 3+ with Hassos 2.12. Hassio 0.102.3. Node-red add-on 5.0.5
SDM120 and SDM230 energy meters connected to Pi via USB.
```
Add-on version: 5.0.5
You are running the latest version of this add-on.
System: HassOS 2.12 (armv7 / raspberrypi3)
Home Assistant version: 0.102.3
Supervisor version: 192
-----------------------------------------------------------
Please, share the above information when looking for help
or support in, e.g., GitHub, forums or the Discord chat.
-----------------------------------------------------------
[cont-init.d] 00-banner.sh: exited 0.
[cont-init.d] 01-log-level.sh: executing...
[cont-init.d] 01-log-level.sh: exited 0.
[cont-init.d] nginx.sh: executing...
[cont-init.d] nginx.sh: exited 0.
[cont-init.d] node-red.sh: executing...
patching file nodes/ui_base.html
Hunk #1 succeeded at 1243 (offset 712 lines).
audited 200 packages in 6.917s
found 0 vulnerabilities
[cont-init.d] node-red.sh: exited 0.
[cont-init.d] user.sh: executing...
[cont-init.d] user.sh: exited 0.
[cont-init.d] done.
[services.d] starting services
[services.d] done.
[10:09:46] INFO: Starting Node-RED...
> hassio-addon-node-red@ start /opt
> node $NODE_OPTIONS node_modules/node-red/red.js "--settings" "/etc/node-red/config.js"
7 Dec 10:09:51 - [info]
Welcome to Node-RED
===================
7 Dec 10:09:51 - [info] Node-RED version: v1.0.3
7 Dec 10:09:51 - [info] Node.js version: v10.16.3
7 Dec 10:09:51 - [info] Linux 4.14.98-v7 arm LE
7 Dec 10:09:53 - [info] Loading palette nodes
7 Dec 10:10:04 - [info] Dashboard version 2.14.0 started at /endpoint/ui
7 Dec 10:10:08 - [info] Settings file : /etc/node-red/config.js
7 Dec 10:10:08 - [info] Context store : 'default' [module=memory]
7 Dec 10:10:08 - [info] User directory : /config/node-red/
7 Dec 10:10:08 - [warn] Projects disabled : editorTheme.projects.enabled=false
7 Dec 10:10:08 - [info] Flows file : /config/node-red/flows.json
7 Dec 10:10:08 - [info] Starting flows
7 Dec 10:10:09 - [info] Started flows
7 Dec 10:10:09 - [info] Server now running at http://127.0.0.1:46836/
[10:10:09] INFO: Starting NGinx...
nginx: [alert] detected a LuaJIT version which is not OpenResty's; many optimizations will be disabled and performance will be compromised (see https://github.com/openresty/luajit2 for OpenResty's LuaJIT or, even better, consider using the OpenResty releases from https://openresty.org/en/download.html)
7 Dec 10:10:11 - [info] [mqtt-broker:c28a2dad.22663] Connected to broker: mqtt://192.168.1.200:1883
7 Dec 10:10:13 - [info] [server:Home Assistant] WebSocket Connected to http://hassio/homeassistant
7 Dec 10:10:21 - [error] [modbus-read:SDM 230 Voltage] Timed out
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Voltage] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Current] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Power] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Imp] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Exp] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Voltage] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Current] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Power] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Energy Sol] FSM Reset On State [object Object]
7 Dec 10:10:21 - [error] [modbus-read:SDM 120 Voltage] Timed out
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Voltage] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Current] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Power] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Imp] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Exp] FSM Reset On State [object Object]
```
* ( ) I am using just the Node-RED package and got [modbus-serial](https://github.com/yaacov/node-modbus-serial) package installed with it
* ( ) I have installed [modbus-serial](https://github.com/yaacov/node-modbus-serial) package global from source (using git clone)
* ( ) I have installed [modbus-serial](https://github.com/yaacov/node-modbus-serial) as a global package (using npm i -g)
Yes, modbus-serial is the name but it does all, TCP and Serial.
The node-modbus package is just for the Modbus Server node.
* Device :
- ( ) macOS: v
- ( ) Linux: v
- ( ) Windows: v
- ( ) Raspbian: v
- (X) Other: HASSOS on RPi3
- Name:
- Version:
* Description of the connecting Modbus system:
- Name :
- Version:
- Vendor :
- Link : https://
* **node-red-contrib-modbus version:** ``
<!-- Check which is the hash of the last commit from node-red-contrib-modbus that you have locally -->
* **Node:** (min. 6.x)
`node --version` = 5.1.0
|
1.0
|
Read Node stopped working - <!--
IF YOU DON'T FILL OUT THE FOLLOWING INFORMATION WE MIGHT CLOSE YOUR ISSUE WITHOUT INVESTIGATING
PLEASE, DELETE NOT NEEDED PARTS FROM THAT ISSUE!
-->
**I'm submitting a ...** (check one with "x")
## Problem
- ( ) bug report => please search github for a similar issue or PR and test Modbus problems with the [modbus-serial](https://github.com/yaacov/node-modbus-serial) package before submitting
- (X) unexpected behavior => please visit the Wiki before
**Current behavior**
<!-- Describe how the bug or problem manifests. -->
Modbus Read node stopped working
**Expected behavior**
<!-- Describe what the behavior would be without the bug or problem. -->
Gets readings from modes devices
### Your Environment
**Please tell us about your environment:**
<!-- system, Modbus server or client type, package manager, equipment, vendor info -->
RPi 3+ with Hassos 2.12. Hassio 0.102.3. Node-red add-on 5.0.5
SDM120 and SDM230 energy meters connected to Pi via USB.
```
Add-on version: 5.0.5
You are running the latest version of this add-on.
System: HassOS 2.12 (armv7 / raspberrypi3)
Home Assistant version: 0.102.3
Supervisor version: 192
-----------------------------------------------------------
Please, share the above information when looking for help
or support in, e.g., GitHub, forums or the Discord chat.
-----------------------------------------------------------
[cont-init.d] 00-banner.sh: exited 0.
[cont-init.d] 01-log-level.sh: executing...
[cont-init.d] 01-log-level.sh: exited 0.
[cont-init.d] nginx.sh: executing...
[cont-init.d] nginx.sh: exited 0.
[cont-init.d] node-red.sh: executing...
patching file nodes/ui_base.html
Hunk #1 succeeded at 1243 (offset 712 lines).
audited 200 packages in 6.917s
found 0 vulnerabilities
[cont-init.d] node-red.sh: exited 0.
[cont-init.d] user.sh: executing...
[cont-init.d] user.sh: exited 0.
[cont-init.d] done.
[services.d] starting services
[services.d] done.
[10:09:46] INFO: Starting Node-RED...
> hassio-addon-node-red@ start /opt
> node $NODE_OPTIONS node_modules/node-red/red.js "--settings" "/etc/node-red/config.js"
7 Dec 10:09:51 - [info]
Welcome to Node-RED
===================
7 Dec 10:09:51 - [info] Node-RED version: v1.0.3
7 Dec 10:09:51 - [info] Node.js version: v10.16.3
7 Dec 10:09:51 - [info] Linux 4.14.98-v7 arm LE
7 Dec 10:09:53 - [info] Loading palette nodes
7 Dec 10:10:04 - [info] Dashboard version 2.14.0 started at /endpoint/ui
7 Dec 10:10:08 - [info] Settings file : /etc/node-red/config.js
7 Dec 10:10:08 - [info] Context store : 'default' [module=memory]
7 Dec 10:10:08 - [info] User directory : /config/node-red/
7 Dec 10:10:08 - [warn] Projects disabled : editorTheme.projects.enabled=false
7 Dec 10:10:08 - [info] Flows file : /config/node-red/flows.json
7 Dec 10:10:08 - [info] Starting flows
7 Dec 10:10:09 - [info] Started flows
7 Dec 10:10:09 - [info] Server now running at http://127.0.0.1:46836/
[10:10:09] INFO: Starting NGinx...
nginx: [alert] detected a LuaJIT version which is not OpenResty's; many optimizations will be disabled and performance will be compromised (see https://github.com/openresty/luajit2 for OpenResty's LuaJIT or, even better, consider using the OpenResty releases from https://openresty.org/en/download.html)
7 Dec 10:10:11 - [info] [mqtt-broker:c28a2dad.22663] Connected to broker: mqtt://192.168.1.200:1883
7 Dec 10:10:13 - [info] [server:Home Assistant] WebSocket Connected to http://hassio/homeassistant
7 Dec 10:10:21 - [error] [modbus-read:SDM 230 Voltage] Timed out
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Voltage] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Current] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Power] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Imp] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Exp] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Voltage] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Current] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Power] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 120 Energy Sol] FSM Reset On State [object Object]
7 Dec 10:10:21 - [error] [modbus-read:SDM 120 Voltage] Timed out
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Voltage] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Current] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Power] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Imp] FSM Reset On State [object Object]
7 Dec 10:10:21 - [warn] [modbus-read:SDM 230 Energy Exp] FSM Reset On State [object Object]
```
* ( ) I am using just the Node-RED package and got [modbus-serial](https://github.com/yaacov/node-modbus-serial) package installed with it
* ( ) I have installed [modbus-serial](https://github.com/yaacov/node-modbus-serial) package global from source (using git clone)
* ( ) I have installed [modbus-serial](https://github.com/yaacov/node-modbus-serial) as a global package (using npm i -g)
Yes, modbus-serial is the name but it does all, TCP and Serial.
The node-modbus package is just for the Modbus Server node.
* Device :
- ( ) macOS: v
- ( ) Linux: v
- ( ) Windows: v
- ( ) Raspbian: v
- (X) Other: HASSOS on RPi3
- Name:
- Version:
* Description of the connecting Modbus system:
- Name :
- Version:
- Vendor :
- Link : https://
* **node-red-contrib-modbus version:** ``
<!-- Check which is the hash of the last commit from node-red-contrib-modbus that you have locally -->
* **Node:** (min. 6.x)
`node --version` = 5.1.0
|
non_defect
|
read node stopped working if you don t fill out the following information we might close your issue without investigating please delete not needed parts from that issue i m submitting a check one with x problem bug report please search github for a similar issue or pr and test modbus problems with the package before submitting x unexpected behavior please visit the wiki before current behavior modbus read node stopped working expected behavior gets readings from modes devices your environment please tell us about your environment rpi with hassos hassio node red add on and energy meters connected to pi via usb add on version you are running the latest version of this add on system hassos home assistant version supervisor version please share the above information when looking for help or support in e g github forums or the discord chat banner sh exited log level sh executing log level sh exited nginx sh executing nginx sh exited node red sh executing patching file nodes ui base html hunk succeeded at offset lines audited packages in found vulnerabilities node red sh exited user sh executing user sh exited done starting services done info starting node red hassio addon node red start opt node node options node modules node red red js settings etc node red config js dec welcome to node red dec node red version dec node js version dec linux arm le dec loading palette nodes dec dashboard version started at endpoint ui dec settings file etc node red config js dec context store default dec user directory config node red dec projects disabled editortheme projects enabled false dec flows file config node red flows json dec starting flows dec started flows dec server now running at info starting nginx nginx detected a luajit version which is not openresty s many optimizations will be disabled and performance will be compromised see for openresty s luajit or even better consider using the openresty releases from dec connected to broker mqtt dec websocket connected to dec timed out dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec timed out dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state dec fsm reset on state i am using just the node red package and got package installed with it i have installed package global from source using git clone i have installed as a global package using npm i g yes modbus serial is the name but it does all tcp and serial the node modbus package is just for the modbus server node device macos v linux v windows v raspbian v x other hassos on name version description of the connecting modbus system name version vendor link https node red contrib modbus version node min x node version
| 0
|
110,589
| 13,917,947,586
|
IssuesEvent
|
2020-10-21 06:55:24
|
vaadin/vaadin-avatar
|
https://api.github.com/repos/vaadin/vaadin-avatar
|
closed
|
The plus sign in "+5" when avatars are overflowing is not completely with a number larger than 9
|
needs design
|
Consider the following code
```
<vaadin-avatar-group id="avatarGroup" items={{items}} max-items-visible="3">
</vaadin-avatar-group>
```
Where `items` contains more than 13 items. At this point the overflow number is partly obscured.

I think this boils down to whether 10+ avatars in one group is realistic. Maybe the overflow circle could be shown on top (on the z axis)?
|
1.0
|
The plus sign in "+5" when avatars are overflowing is not completely with a number larger than 9 - Consider the following code
```
<vaadin-avatar-group id="avatarGroup" items={{items}} max-items-visible="3">
</vaadin-avatar-group>
```
Where `items` contains more than 13 items. At this point the overflow number is partly obscured.

I think this boils down to whether 10+ avatars in one group is realistic. Maybe the overflow circle could be shown on top (on the z axis)?
|
non_defect
|
the plus sign in when avatars are overflowing is not completely with a number larger than consider the following code where items contains more than items at this point the overflow number is partly obscured i think this boils down to whether avatars in one group is realistic maybe the overflow circle could be shown on top on the z axis
| 0
|
219,537
| 17,099,256,254
|
IssuesEvent
|
2021-07-09 08:54:09
|
tracim/tracim
|
https://api.github.com/repos/tracim/tracim
|
closed
|
Feat: Create TagList component
|
frontend manually tested not in changelog
|
## Feature description and goals
We add tags to contents (#4681), which requires a new component to render the tag list.
This component will be rendered in the apps file and html-document, in a dedicated tab above the other ones.
The goal of this feature is to let the user choose tags in the available tag list, or to create new ones in a form.
The selected tags will be displayed in the left part of the app, on top of the "download" icon.
This component should look like MemberList
The chosen tags will be marked as "checked", and the other ones will be "unchecked". The checked tags will appear on top of the unchecked ones.
When a new tag is created in the app, it will be marked as "checked" and it will be rendered in the list. This tag will be available in every content in its space, and will be displayed as an "unchecked" tag.
To do :
- [x] Create TagList component
- [x] Create NewTagForm component
- [x] Create Tag component
- [x] Render selected tags
- [x] Add translations
- [x] Link with backend endpoints
<!-- Explain why we want this feature and describe it. -->
## Prerequisites
Backend endpoint to GET the tag list: #4685
## Mockups

<!-- Optional - list the issues that must be solved or what needs to be done before handling this issue. -->
<!-- ## Required sections, if relevant ## -->
<!-- - To be discussed before development -->
<!-- - Interface -->
<!-- - Translations -->
<!-- - Workaround -->
<!-- - Extra information -->
|
1.0
|
Feat: Create TagList component - ## Feature description and goals
We add tags to contents (#4681), which requires a new component to render the tag list.
This component will be rendered in the apps file and html-document, in a dedicated tab above the other ones.
The goal of this feature is to let the user choose tags in the available tag list, or to create new ones in a form.
The selected tags will be displayed in the left part of the app, on top of the "download" icon.
This component should look like MemberList
The chosen tags will be marked as "checked", and the other ones will be "unchecked". The checked tags will appear on top of the unchecked ones.
When a new tag is created in the app, it will be marked as "checked" and it will be rendered in the list. This tag will be available in every content in its space, and will be displayed as an "unchecked" tag.
To do :
- [x] Create TagList component
- [x] Create NewTagForm component
- [x] Create Tag component
- [x] Render selected tags
- [x] Add translations
- [x] Link with backend endpoints
<!-- Explain why we want this feature and describe it. -->
## Prerequisites
Backend endpoint to GET the tag list: #4685
## Mockups

<!-- Optional - list the issues that must be solved or what needs to be done before handling this issue. -->
<!-- ## Required sections, if relevant ## -->
<!-- - To be discussed before development -->
<!-- - Interface -->
<!-- - Translations -->
<!-- - Workaround -->
<!-- - Extra information -->
|
non_defect
|
feat create taglist component feature description and goals we add tags to contents which requires a new component to render the tag list this component will be rendered in the apps file and html document in a dedicated tab above the other ones the goal of this feature is to let the user choose tags in the available tag list or to create new ones in a form the selected tags will be displayed in the left part of the app on top of the download icon this component should look like memberlist the chosen tags will be marked as checked and the other ones will be unchecked the checked tags will appear on top of the unchecked ones when a new tag is created in the app it will be marked as checked and it will be rendered in the list this tag will be available in every content in its space and will be displayed as an unchecked tag to do create taglist component create newtagform component create tag component render selected tags add translations link with backend endpoints prerequisites backend endpoint to get the tag list mockups
| 0
|
44,525
| 12,227,249,219
|
IssuesEvent
|
2020-05-03 14:31:36
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
SelectBooleanCheckbox: Disabled selectBooleanCheckbox with itemLabel defined
|
defect
|
## 1) Environment
- PrimeFaces version: 7.0
- Does it work on the newest released PrimeFaces version? Version?: No
- Does it work on the newest sources in GitHub? (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source)
- Application server + version: tomee 7.1.1
- Affected browsers: Firefox, Chrome, ...
## 2) Expected behavior
If SelectBooleanCheckbox is disabled and an ItemLabel is defined, the label should also be disabled.
(as in the 2. field of screenshot)
## 3) Actual behavior
The label looks like activated and has also the "click-finger" cursor (as the 3. field in the screenshot)

## 4) Steps to reproduce
..
## 5) Sample XHTML
```xml
<p:panelGrid columns="2">
<p:selectBooleanCheckbox id="cb1"
value="true"
disabled="false"
itemLabel="1. disabled='false'">
</p:selectBooleanCheckbox>
<p:selectBooleanCheckbox id="cb2"
value="true"
styleClass="ui-state-disabled"
disabled="false"
itemLabel="2. disabled='false' styleClass='ui-state-disabled'">
</p:selectBooleanCheckbox>
<p:selectBooleanCheckbox id="cb3"
value="true"
disabled="true"
itemLabel="3. disabled='true'">
</p:selectBooleanCheckbox>
<p:selectBooleanCheckbox id="cb4"
value="true"
styleClass="ui-state-disabled"
disabled="true"
itemLabel="4. disabled='true' styleClass='ui-state-disabled'">
</p:selectBooleanCheckbox>
</p:panelGrid>
```
## 6) Sample bean
..
|
1.0
|
SelectBooleanCheckbox: Disabled selectBooleanCheckbox with itemLabel defined - ## 1) Environment
- PrimeFaces version: 7.0
- Does it work on the newest released PrimeFaces version? Version?: No
- Does it work on the newest sources in GitHub? (Build by source -> https://github.com/primefaces/primefaces/wiki/Building-From-Source)
- Application server + version: tomee 7.1.1
- Affected browsers: Firefox, Chrome, ...
## 2) Expected behavior
If SelectBooleanCheckbox is disabled and an ItemLabel is defined, the label should also be disabled.
(as in the 2. field of screenshot)
## 3) Actual behavior
The label looks like activated and has also the "click-finger" cursor (as the 3. field in the screenshot)

## 4) Steps to reproduce
..
## 5) Sample XHTML
```xml
<p:panelGrid columns="2">
<p:selectBooleanCheckbox id="cb1"
value="true"
disabled="false"
itemLabel="1. disabled='false'">
</p:selectBooleanCheckbox>
<p:selectBooleanCheckbox id="cb2"
value="true"
styleClass="ui-state-disabled"
disabled="false"
itemLabel="2. disabled='false' styleClass='ui-state-disabled'">
</p:selectBooleanCheckbox>
<p:selectBooleanCheckbox id="cb3"
value="true"
disabled="true"
itemLabel="3. disabled='true'">
</p:selectBooleanCheckbox>
<p:selectBooleanCheckbox id="cb4"
value="true"
styleClass="ui-state-disabled"
disabled="true"
itemLabel="4. disabled='true' styleClass='ui-state-disabled'">
</p:selectBooleanCheckbox>
</p:panelGrid>
```
## 6) Sample bean
..
|
defect
|
selectbooleancheckbox disabled selectbooleancheckbox with itemlabel defined environment primefaces version does it work on the newest released primefaces version version no does it work on the newest sources in github build by source application server version tomee affected browsers firefox chrome expected behavior if selectbooleancheckbox is disabled and an itemlabel is defined the label should also be disabled as in the field of screenshot actual behavior the label looks like activated and has also the click finger cursor as the field in the screenshot steps to reproduce sample xhtml xml p selectbooleancheckbox id value true disabled false itemlabel disabled false p selectbooleancheckbox id value true styleclass ui state disabled disabled false itemlabel disabled false styleclass ui state disabled p selectbooleancheckbox id value true disabled true itemlabel disabled true p selectbooleancheckbox id value true styleclass ui state disabled disabled true itemlabel disabled true styleclass ui state disabled sample bean
| 1
|
45,435
| 12,799,854,314
|
IssuesEvent
|
2020-07-02 16:02:57
|
snowplow/snowplow-android-tracker
|
https://api.github.com/repos/snowplow/snowplow-android-tracker
|
closed
|
Fix importing of kotlin on gradle
|
priority:medium status:completed type:defect
|
This project is written 100% in Java, however the SDK ships with a dependency on [the Kotlin stdlib](https://github.com/snowplow/snowplow-android-tracker/blob/master/snowplow-tracker/build.gradle#L85) and [Kotlin Android extensions](https://github.com/snowplow/snowplow-android-tracker/blob/master/snowplow-tracker/build.gradle#L7). Kotlin was added in [this PR](https://github.com/snowplow/snowplow-android-tracker/pull/358), but seems unrelated?
Also as an aside it would be great if this library added nullability annotations to make Kotlin interoperability nicer! I can open up a separate issue for this if you'd prefer.
|
1.0
|
Fix importing of kotlin on gradle - This project is written 100% in Java, however the SDK ships with a dependency on [the Kotlin stdlib](https://github.com/snowplow/snowplow-android-tracker/blob/master/snowplow-tracker/build.gradle#L85) and [Kotlin Android extensions](https://github.com/snowplow/snowplow-android-tracker/blob/master/snowplow-tracker/build.gradle#L7). Kotlin was added in [this PR](https://github.com/snowplow/snowplow-android-tracker/pull/358), but seems unrelated?
Also as an aside it would be great if this library added nullability annotations to make Kotlin interoperability nicer! I can open up a separate issue for this if you'd prefer.
|
defect
|
fix importing of kotlin on gradle this project is written in java however the sdk ships with a dependency on and kotlin was added in but seems unrelated also as an aside it would be great if this library added nullability annotations to make kotlin interoperability nicer i can open up a separate issue for this if you d prefer
| 1
|
4,103
| 2,610,087,444
|
IssuesEvent
|
2015-02-26 18:26:33
|
chrsmith/dsdsdaadf
|
https://api.github.com/repos/chrsmith/dsdsdaadf
|
opened
|
深圳痤疮哪里祛比较好
|
auto-migrated Priority-Medium Type-Defect
|
```
深圳痤疮哪里祛比较好【深圳韩方科颜全国热线400-869-1818,24
小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩��
�秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,�
��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹
”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内��
�业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上�
��痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:20
|
1.0
|
深圳痤疮哪里祛比较好 - ```
深圳痤疮哪里祛比较好【深圳韩方科颜全国热线400-869-1818,24
小时QQ4008691818】深圳韩方科颜专业祛痘连锁机构,机构以韩��
�秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,�
��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹
”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内��
�业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上�
��痘痘。
```
-----
Original issue reported on code.google.com by `szft...@163.com` on 14 May 2014 at 7:20
|
defect
|
深圳痤疮哪里祛比较好 深圳痤疮哪里祛比较好【 , 】深圳韩方科颜专业祛痘连锁机构,机构以韩�� �秘方——韩方科颜这一国妆准字号治疗型权威,祛痘佳品,� ��方科颜专业祛痘连锁机构,采用韩国秘方配合专业“不反弹 ”健康祛痘技术并结合先进“先进豪华彩光”仪,开创国内�� �业治疗粉刺、痤疮签约包治先河,成功消除了许多顾客脸上� ��痘痘。 original issue reported on code google com by szft com on may at
| 1
|
795,706
| 28,083,324,494
|
IssuesEvent
|
2023-03-30 08:07:25
|
frequenz-floss/frequenz-sdk-python
|
https://api.github.com/repos/frequenz-floss/frequenz-sdk-python
|
opened
|
Update SDK dependency of channels to the new release 0.14
|
part:❓ priority:❓ type:enhancement
|
### What's needed?
Interfaces changed and need to be adjusted
### Proposed solution
_No response_
### Use cases
_No response_
### Alternatives and workarounds
_No response_
### Additional context
_No response_
|
1.0
|
Update SDK dependency of channels to the new release 0.14 - ### What's needed?
Interfaces changed and need to be adjusted
### Proposed solution
_No response_
### Use cases
_No response_
### Alternatives and workarounds
_No response_
### Additional context
_No response_
|
non_defect
|
update sdk dependency of channels to the new release what s needed interfaces changed and need to be adjusted proposed solution no response use cases no response alternatives and workarounds no response additional context no response
| 0
|
7,091
| 2,597,979,510
|
IssuesEvent
|
2015-02-22 01:06:43
|
JukkaL/mypy
|
https://api.github.com/repos/JukkaL/mypy
|
closed
|
For loop parse error with single-item tuple lvalue
|
bug priority
|
This is valid Python but mypy doesn't accept it (`Parse error before in`):
```python
for x, in [(1,), (2,), (3,)]:
print(x)
```
|
1.0
|
For loop parse error with single-item tuple lvalue - This is valid Python but mypy doesn't accept it (`Parse error before in`):
```python
for x, in [(1,), (2,), (3,)]:
print(x)
```
|
non_defect
|
for loop parse error with single item tuple lvalue this is valid python but mypy doesn t accept it parse error before in python for x in print x
| 0
|
13,406
| 2,755,583,219
|
IssuesEvent
|
2015-04-26 19:40:11
|
jakejustus/pct
|
https://api.github.com/repos/jakejustus/pct
|
closed
|
PCTCompileExt doesn't set resultProperty
|
auto-migrated Type-Defect wontfix
|
```
When running PCTCompileExt with failOnError="false" resultProperty="result",
the result property is not set at the end of the PCTCompileExt task. With
PCTCompile it works fine.
```
Original issue reported on code.google.com by `tif...@gmail.com` on 3 Oct 2013 at 10:21
|
1.0
|
PCTCompileExt doesn't set resultProperty - ```
When running PCTCompileExt with failOnError="false" resultProperty="result",
the result property is not set at the end of the PCTCompileExt task. With
PCTCompile it works fine.
```
Original issue reported on code.google.com by `tif...@gmail.com` on 3 Oct 2013 at 10:21
|
defect
|
pctcompileext doesn t set resultproperty when running pctcompileext with failonerror false resultproperty result the result property is not set at the end of the pctcompileext task with pctcompile it works fine original issue reported on code google com by tif gmail com on oct at
| 1
|
47,077
| 13,056,028,043
|
IssuesEvent
|
2020-07-30 03:26:08
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
[PROPOSAL] fatal in decay (Trac #2397)
|
Incomplete Migration Migrated from Trac combo simulation defect
|
Migrated from https://code.icecube.wisc.edu/ticket/2397
```json
{
"status": "closed",
"changetime": "2020-06-24T12:31:42",
"description": "I found this fatal error running PROPOSAL via the script [https://code.icecube.wisc.edu/projects/icecube/browser/IceCube/meta-projects/combo/trunk/simprod-scripts/resources/scripts/clsim.py]:\n\n\n{{{\nFATAL (PROPOSAL): Kaellen function is negative. Cannot caluclate momentum (Decay\nChannel.cxx:88 in static double PROPOSAL::DecayChannel::Momentum(double, double,\n double))\nERROR (I3Module): propagator_propagator: Exception thrown (I3Module.cxx:123 in v\noid I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/data/user/grenzi/combo-py3/src/simprod-scripts/resources/scripts/clsim.\npy\", line 188, in <module>\n main()\n File \"/data/user/grenzi/combo-py3/src/simprod-scripts/resources/scripts/clsim.\npy\", line 184, in main\n usegslrng=params['usegslrng'])\n File \"/data/user/grenzi/combo-py3/build/lib/icecube/simprod/util/simprodtray.py\", line 150, in RunI3Tray\n _execute(tray, executionmaxcount)\n File \"/data/user/grenzi/combo-py3/build/lib/icecube/simprod/util/simprodtray.py\", line 53, in _execute\n tray.Execute()\n File \"/data/user/grenzi/combo-py3/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\nRuntimeError: Kaellen function is negative. Cannot caluclate momentum (in static double PROPOSAL::DecayChannel::Momentum(double, double, double))\n}}}\n\nThe error seems to come from here at line 88 [https://code.icecube.wisc.edu/projects/icecube/browser/IceCube/meta-projects/combo/trunk/PROPOSAL/private/PROPOSAL/decay/DecayChannel.cxx]\n\n\n{{{\n79 double DecayChannel::Momentum(double m1, double m2, double m3)\n80\t{\n81\t double kaellen = (m1 - m2 - m3) * (m1 + m2 + m3) * (m1 - m2 + m3) * (m1 + m2 - m3);\n82\t\n83\t if (kaellen > 0.0)\n84\t {\n85\t return std::sqrt(kaellen) / (2.0 * m1);\n86\t } else\n87\t {\n88\t log_fatal(\"Kaellen function is negative. Cannot caluclate momentum\");\n89\t return 0.0;\n90\t }\n91\t}\n}}}\n\n\nThis error prevents the propagation process for all of the following frames. I think the right thing would be just skipping the single wrong one.\n",
"reporter": "grenzi",
"cc": "",
"resolution": "fixed",
"_ts": "1593001902142004",
"component": "combo simulation",
"summary": "[PROPOSAL] fatal in decay",
"priority": "major",
"keywords": "",
"time": "2020-01-08T09:47:52",
"milestone": "Autumnal Equinox 2020",
"owner": "jsoedingrekso",
"type": "defect"
}
```
|
1.0
|
[PROPOSAL] fatal in decay (Trac #2397) - Migrated from https://code.icecube.wisc.edu/ticket/2397
```json
{
"status": "closed",
"changetime": "2020-06-24T12:31:42",
"description": "I found this fatal error running PROPOSAL via the script [https://code.icecube.wisc.edu/projects/icecube/browser/IceCube/meta-projects/combo/trunk/simprod-scripts/resources/scripts/clsim.py]:\n\n\n{{{\nFATAL (PROPOSAL): Kaellen function is negative. Cannot caluclate momentum (Decay\nChannel.cxx:88 in static double PROPOSAL::DecayChannel::Momentum(double, double,\n double))\nERROR (I3Module): propagator_propagator: Exception thrown (I3Module.cxx:123 in v\noid I3Module::Do(void (I3Module::*)()))\nTraceback (most recent call last):\n File \"/data/user/grenzi/combo-py3/src/simprod-scripts/resources/scripts/clsim.\npy\", line 188, in <module>\n main()\n File \"/data/user/grenzi/combo-py3/src/simprod-scripts/resources/scripts/clsim.\npy\", line 184, in main\n usegslrng=params['usegslrng'])\n File \"/data/user/grenzi/combo-py3/build/lib/icecube/simprod/util/simprodtray.py\", line 150, in RunI3Tray\n _execute(tray, executionmaxcount)\n File \"/data/user/grenzi/combo-py3/build/lib/icecube/simprod/util/simprodtray.py\", line 53, in _execute\n tray.Execute()\n File \"/data/user/grenzi/combo-py3/build/lib/I3Tray.py\", line 256, in Execute\n super(I3Tray, self).Execute()\nRuntimeError: Kaellen function is negative. Cannot caluclate momentum (in static double PROPOSAL::DecayChannel::Momentum(double, double, double))\n}}}\n\nThe error seems to come from here at line 88 [https://code.icecube.wisc.edu/projects/icecube/browser/IceCube/meta-projects/combo/trunk/PROPOSAL/private/PROPOSAL/decay/DecayChannel.cxx]\n\n\n{{{\n79 double DecayChannel::Momentum(double m1, double m2, double m3)\n80\t{\n81\t double kaellen = (m1 - m2 - m3) * (m1 + m2 + m3) * (m1 - m2 + m3) * (m1 + m2 - m3);\n82\t\n83\t if (kaellen > 0.0)\n84\t {\n85\t return std::sqrt(kaellen) / (2.0 * m1);\n86\t } else\n87\t {\n88\t log_fatal(\"Kaellen function is negative. Cannot caluclate momentum\");\n89\t return 0.0;\n90\t }\n91\t}\n}}}\n\n\nThis error prevents the propagation process for all of the following frames. I think the right thing would be just skipping the single wrong one.\n",
"reporter": "grenzi",
"cc": "",
"resolution": "fixed",
"_ts": "1593001902142004",
"component": "combo simulation",
"summary": "[PROPOSAL] fatal in decay",
"priority": "major",
"keywords": "",
"time": "2020-01-08T09:47:52",
"milestone": "Autumnal Equinox 2020",
"owner": "jsoedingrekso",
"type": "defect"
}
```
|
defect
|
fatal in decay trac migrated from json status closed changetime description i found this fatal error running proposal via the script n n n nfatal proposal kaellen function is negative cannot caluclate momentum decay nchannel cxx in static double proposal decaychannel momentum double double n double nerror propagator propagator exception thrown cxx in v noid do void ntraceback most recent call last n file data user grenzi combo src simprod scripts resources scripts clsim npy line in n main n file data user grenzi combo src simprod scripts resources scripts clsim npy line in main n usegslrng params n file data user grenzi combo build lib icecube simprod util simprodtray py line in n execute tray executionmaxcount n file data user grenzi combo build lib icecube simprod util simprodtray py line in execute n tray execute n file data user grenzi combo build lib py line in execute n super self execute nruntimeerror kaellen function is negative cannot caluclate momentum in static double proposal decaychannel momentum double double double n n nthe error seems to come from here at line n n n double decaychannel momentum double double double t t double kaellen t t if kaellen t t return std sqrt kaellen t else t t log fatal kaellen function is negative cannot caluclate momentum t return t t n n n nthis error prevents the propagation process for all of the following frames i think the right thing would be just skipping the single wrong one n reporter grenzi cc resolution fixed ts component combo simulation summary fatal in decay priority major keywords time milestone autumnal equinox owner jsoedingrekso type defect
| 1
|
154,353
| 24,281,391,238
|
IssuesEvent
|
2022-09-28 17:44:15
|
microsoft/pylance-release
|
https://api.github.com/repos/microsoft/pylance-release
|
closed
|
Python language server no longer recommends auto imports from local files (still works for site-packages in venv)
|
by design
|
Type: <b>Bug</b>
Unsure, it worked 3 days ago, not it doesn't, didn't change any workspace settings what so ever.
Extension version: 2022.14.0
VS Code version: Code 1.71.2 (Universal) (74b1f979648cc44d385a2286793c226e611f59e7, 2022-09-14T21:05:37.721Z)
OS version: Darwin x64 21.6.0
Modes:
Sandboxed: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i9-9880H CPU @ 2.30GHz (16 x 2300)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off|
|Load (avg)|2, 2, 2|
|Memory (System)|32.00GB (6.37GB free)|
|Process Argv|--crash-reporter-id 0731fb54-df68-48c8-9bc7-cb94d92db4d8|
|Screen Reader|no|
|VM|0%|
</details>
<!-- generated by issue reporter -->
|
1.0
|
Python language server no longer recommends auto imports from local files (still works for site-packages in venv) -
Type: <b>Bug</b>
Unsure, it worked 3 days ago, not it doesn't, didn't change any workspace settings what so ever.
Extension version: 2022.14.0
VS Code version: Code 1.71.2 (Universal) (74b1f979648cc44d385a2286793c226e611f59e7, 2022-09-14T21:05:37.721Z)
OS version: Darwin x64 21.6.0
Modes:
Sandboxed: No
<details>
<summary>System Info</summary>
|Item|Value|
|---|---|
|CPUs|Intel(R) Core(TM) i9-9880H CPU @ 2.30GHz (16 x 2300)|
|GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>metal: disabled_off<br>multiple_raster_threads: enabled_on<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled<br>webgpu: disabled_off|
|Load (avg)|2, 2, 2|
|Memory (System)|32.00GB (6.37GB free)|
|Process Argv|--crash-reporter-id 0731fb54-df68-48c8-9bc7-cb94d92db4d8|
|Screen Reader|no|
|VM|0%|
</details>
<!-- generated by issue reporter -->
|
non_defect
|
python language server no longer recommends auto imports from local files still works for site packages in venv type bug unsure it worked days ago not it doesn t didn t change any workspace settings what so ever extension version vs code version code universal os version darwin modes sandboxed no system info item value cpus intel r core tm cpu x gpu status canvas enabled canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing enabled metal disabled off multiple raster threads enabled on opengl enabled on rasterization enabled raw draw disabled off ok skia renderer enabled on video decode enabled video encode enabled vulkan disabled off webgl enabled enabled webgpu disabled off load avg memory system free process argv crash reporter id screen reader no vm
| 0
|
78,426
| 27,516,800,032
|
IssuesEvent
|
2023-03-06 12:32:22
|
primefaces/primefaces
|
https://api.github.com/repos/primefaces/primefaces
|
closed
|
Tree: pfAjaxComplete and pfAjaxStart not working when filter a tree
|
:lady_beetle: defect :bangbang: needs-triage
|
### Describe the bug
When a Tree is filtered, an Ajax call is performed (see browser console), but event handlers for `pfAjaxStart` `pfAjaxComplete` don't react.
### Reproducer
[primefaces-ajax-trigger-missing.zip](https://github.com/primefaces/primefaces/files/10897430/primefaces-ajax-trigger-missing.zip)
```xhtml
<p:tree value="#{testView.root}" var="node" filterBy="#{node}">
<p:treeNode>
<h:outputText value="#{node}" />
</p:treeNode>
</p:tree>
<script>
$(document).on('pfAjaxStart',()=>console.info('Test'))
</script>
```
Also reproducable in PrimeFaces Showcase (type `$(document).on('pfAjaxStart',()=>console.info('Test'))` in the browser console and look for messages. Compare [Tree](https://www.primefaces.org/showcase/ui/data/tree/filter.xhtml) with [Data Table](https://www.primefaces.org/showcase/ui/data/datatable/filter.xhtml).
### Expected behavior
`pfAjaxStart` and `pfAjaxComplete` should be called.
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.2
### Java version
17
### Browser(s)
_No response_
|
1.0
|
Tree: pfAjaxComplete and pfAjaxStart not working when filter a tree - ### Describe the bug
When a Tree is filtered, an Ajax call is performed (see browser console), but event handlers for `pfAjaxStart` `pfAjaxComplete` don't react.
### Reproducer
[primefaces-ajax-trigger-missing.zip](https://github.com/primefaces/primefaces/files/10897430/primefaces-ajax-trigger-missing.zip)
```xhtml
<p:tree value="#{testView.root}" var="node" filterBy="#{node}">
<p:treeNode>
<h:outputText value="#{node}" />
</p:treeNode>
</p:tree>
<script>
$(document).on('pfAjaxStart',()=>console.info('Test'))
</script>
```
Also reproducable in PrimeFaces Showcase (type `$(document).on('pfAjaxStart',()=>console.info('Test'))` in the browser console and look for messages. Compare [Tree](https://www.primefaces.org/showcase/ui/data/tree/filter.xhtml) with [Data Table](https://www.primefaces.org/showcase/ui/data/datatable/filter.xhtml).
### Expected behavior
`pfAjaxStart` and `pfAjaxComplete` should be called.
### PrimeFaces edition
Community
### PrimeFaces version
12.0.0
### Theme
_No response_
### JSF implementation
Mojarra
### JSF version
2.2
### Java version
17
### Browser(s)
_No response_
|
defect
|
tree pfajaxcomplete and pfajaxstart not working when filter a tree describe the bug when a tree is filtered an ajax call is performed see browser console but event handlers for pfajaxstart pfajaxcomplete don t react reproducer xhtml document on pfajaxstart console info test also reproducable in primefaces showcase type document on pfajaxstart console info test in the browser console and look for messages compare with expected behavior pfajaxstart and pfajaxcomplete should be called primefaces edition community primefaces version theme no response jsf implementation mojarra jsf version java version browser s no response
| 1
|
4,879
| 2,610,159,473
|
IssuesEvent
|
2015-02-26 18:50:36
|
chrsmith/republic-at-war
|
https://api.github.com/repos/chrsmith/republic-at-war
|
closed
|
Map Issue
|
auto-migrated Priority-Medium Type-Defect
|
```
Naboo...
Some gungans trapped in trees.. possible passability issue
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 12:35
|
1.0
|
Map Issue - ```
Naboo...
Some gungans trapped in trees.. possible passability issue
```
-----
Original issue reported on code.google.com by `z3r0...@gmail.com` on 31 Jan 2011 at 12:35
|
defect
|
map issue naboo some gungans trapped in trees possible passability issue original issue reported on code google com by gmail com on jan at
| 1
|
64,365
| 18,506,943,317
|
IssuesEvent
|
2021-10-19 19:51:00
|
scipy/scipy
|
https://api.github.com/repos/scipy/scipy
|
opened
|
BUG: NumPy's `random` module should not be in the `scipy` namespace.
|
defect
|
### Describe your issue.
Currently, `scipy.random` is an alias for `numpy.random`:
```
In [12]: import numpy as np
In [13]: import scipy
In [14]: scipy.__version__
Out[14]: '1.8.0.dev0+1859.14341ea'
In [15]: scipy.random
Out[15]: <module 'numpy.random' from '/home/warren/mc39scipy/lib/python3.9/site-packages/numpy/random/__init__.py'>
In [16]: scipy.random is np.random
Out[16]: True
```
We have already deprecated other names in `scipy` that come from `numpy`, e.g.
```
In [17]: scipy.array([1, 2, 3])
<ipython-input-17-773c09c10413>:1: DeprecationWarning: scipy.array is deprecated and will be removed in SciPy 2.0.0, use numpy.array instead
scipy.array([1, 2, 3])
Out[17]: array([1, 2, 3])
```
We should to the same for `random`, so a call such as the following generates a deprecation warning:
```
In [18]: scipy.random.randint(10, size=3)
Out[18]: array([8, 5, 5])
```
### Reproducing Code Example
```python
See above.
```
### Error message
```shell
See above.
```
### SciPy/NumPy/Python version information
1.8.0.dev0+1859.14341ea 1.21.2 sys.version_info(major=3, minor=9, micro=5, releaselevel='final', serial=0)
|
1.0
|
BUG: NumPy's `random` module should not be in the `scipy` namespace. - ### Describe your issue.
Currently, `scipy.random` is an alias for `numpy.random`:
```
In [12]: import numpy as np
In [13]: import scipy
In [14]: scipy.__version__
Out[14]: '1.8.0.dev0+1859.14341ea'
In [15]: scipy.random
Out[15]: <module 'numpy.random' from '/home/warren/mc39scipy/lib/python3.9/site-packages/numpy/random/__init__.py'>
In [16]: scipy.random is np.random
Out[16]: True
```
We have already deprecated other names in `scipy` that come from `numpy`, e.g.
```
In [17]: scipy.array([1, 2, 3])
<ipython-input-17-773c09c10413>:1: DeprecationWarning: scipy.array is deprecated and will be removed in SciPy 2.0.0, use numpy.array instead
scipy.array([1, 2, 3])
Out[17]: array([1, 2, 3])
```
We should to the same for `random`, so a call such as the following generates a deprecation warning:
```
In [18]: scipy.random.randint(10, size=3)
Out[18]: array([8, 5, 5])
```
### Reproducing Code Example
```python
See above.
```
### Error message
```shell
See above.
```
### SciPy/NumPy/Python version information
1.8.0.dev0+1859.14341ea 1.21.2 sys.version_info(major=3, minor=9, micro=5, releaselevel='final', serial=0)
|
defect
|
bug numpy s random module should not be in the scipy namespace describe your issue currently scipy random is an alias for numpy random in import numpy as np in import scipy in scipy version out in scipy random out in scipy random is np random out true we have already deprecated other names in scipy that come from numpy e g in scipy array deprecationwarning scipy array is deprecated and will be removed in scipy use numpy array instead scipy array out array we should to the same for random so a call such as the following generates a deprecation warning in scipy random randint size out array reproducing code example python see above error message shell see above scipy numpy python version information sys version info major minor micro releaselevel final serial
| 1
|
67,535
| 20,980,637,225
|
IssuesEvent
|
2022-03-28 19:32:59
|
jccastillo0007/eFacturaT
|
https://api.github.com/repos/jccastillo0007/eFacturaT
|
closed
|
CFDI 4.0 - exportación y régimen. Enviar clave y descripción al PDF
|
resolved defect
|
Ahora mismo solo envía la clave, y se requiere la descripción también.
En general para todos los datos que son claves, enviar al PDF la descripción.
|
1.0
|
CFDI 4.0 - exportación y régimen. Enviar clave y descripción al PDF - Ahora mismo solo envía la clave, y se requiere la descripción también.
En general para todos los datos que son claves, enviar al PDF la descripción.
|
defect
|
cfdi exportación y régimen enviar clave y descripción al pdf ahora mismo solo envía la clave y se requiere la descripción también en general para todos los datos que son claves enviar al pdf la descripción
| 1
|
29,017
| 5,477,086,453
|
IssuesEvent
|
2017-03-12 03:45:37
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
postgresql: NUMERIC field in database appears as float when debug($entity)
|
Defect
|
This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.4.3 / Postgresql 9.4
### What you did
1- defined a field as type NUMERIC(1) in PostgreSQL database
2- did a find() over the table
### What happened
when I debug() over the query the field appears as float:
```
...
'sgn' => (float) -1,
...
```
### What you expected to happen
should appear as integer
|
1.0
|
postgresql: NUMERIC field in database appears as float when debug($entity) - This is a (multiple allowed):
* [x] bug
* [ ] enhancement
* [ ] feature-discussion (RFC)
* CakePHP Version: 3.4.3 / Postgresql 9.4
### What you did
1- defined a field as type NUMERIC(1) in PostgreSQL database
2- did a find() over the table
### What happened
when I debug() over the query the field appears as float:
```
...
'sgn' => (float) -1,
...
```
### What you expected to happen
should appear as integer
|
defect
|
postgresql numeric field in database appears as float when debug entity this is a multiple allowed bug enhancement feature discussion rfc cakephp version postgresql what you did defined a field as type numeric in postgresql database did a find over the table what happened when i debug over the query the field appears as float sgn float what you expected to happen should appear as integer
| 1
|
44,035
| 11,914,518,904
|
IssuesEvent
|
2020-03-31 13:43:18
|
SasView/sasview
|
https://api.github.com/repos/SasView/sasview
|
closed
|
When loading plugin model with a spurious unicode character plugin editor crashes
|
CCX: GUI defect
|
This a follow up to #1426. There is currently no handling for the a spurious unicode character in plugin model editor (Edit custom model), so it crashes when one tries to edit it.
|
1.0
|
When loading plugin model with a spurious unicode character plugin editor crashes - This a follow up to #1426. There is currently no handling for the a spurious unicode character in plugin model editor (Edit custom model), so it crashes when one tries to edit it.
|
defect
|
when loading plugin model with a spurious unicode character plugin editor crashes this a follow up to there is currently no handling for the a spurious unicode character in plugin model editor edit custom model so it crashes when one tries to edit it
| 1
|
234,143
| 7,717,198,671
|
IssuesEvent
|
2018-05-23 13:07:09
|
unfoldingWord-dev/translationCore
|
https://api.github.com/repos/unfoldingWord-dev/translationCore
|
closed
|
Again able to put GL words in wrong order
|
Kind/Bug Priority/Medium QA/Pass
|
beta.92 - same in rc2
[fr_tit_text_ulb.tstudio.zip](https://github.com/unfoldingWord-dev/translationCore/files/1381075/fr_tit_text_ulb.tstudio.zip)
1.

Move the French words into the target in this order: sévèrement, témoignage, reprimende, Ce
2.

Move the French words into the target in this order: saine, foi, une
|
1.0
|
Again able to put GL words in wrong order - beta.92 - same in rc2
[fr_tit_text_ulb.tstudio.zip](https://github.com/unfoldingWord-dev/translationCore/files/1381075/fr_tit_text_ulb.tstudio.zip)
1.

Move the French words into the target in this order: sévèrement, témoignage, reprimende, Ce
2.

Move the French words into the target in this order: saine, foi, une
|
non_defect
|
again able to put gl words in wrong order beta same in move the french words into the target in this order sévèrement témoignage reprimende ce move the french words into the target in this order saine foi une
| 0
|
108,214
| 16,762,817,311
|
IssuesEvent
|
2021-06-14 03:16:33
|
gms-ws-sandbox/nibrs-pr-test
|
https://api.github.com/repos/gms-ws-sandbox/nibrs-pr-test
|
opened
|
CVE-2020-11620 (High) detected in multiple libraries
|
security vulnerability
|
## CVE-2020-11620 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.8.0.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-summary-report-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.1.5.RELEASE.jar
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-validate-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.10.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: nibrs-pr-test/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.10.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-route/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs-pr-test/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs-pr-test/commit/860cc22f54e17594e32e303f0716fb065202fff5">860cc22f54e17594e32e303f0716fb065202fff5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly).
<p>Publish Date: 2020-04-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620>CVE-2020-11620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620</a></p>
<p>Release Date: 2020-04-07</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-route/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-validation/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11620","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
True
|
CVE-2020-11620 (High) detected in multiple libraries - ## CVE-2020-11620 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jackson-databind-2.8.0.jar</b>, <b>jackson-databind-2.9.8.jar</b>, <b>jackson-databind-2.9.5.jar</b>, <b>jackson-databind-2.8.10.jar</b>, <b>jackson-databind-2.9.6.jar</b></p></summary>
<p>
<details><summary><b>jackson-databind-2.8.0.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.0/jackson-databind-2.8.0.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.8.0.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.8.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-summary-report-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.8/jackson-databind-2.9.8.jar</p>
<p>
Dependency Hierarchy:
- spring-boot-starter-web-2.1.5.RELEASE.jar (Root Library)
- spring-boot-starter-json-2.1.5.RELEASE.jar
- :x: **jackson-databind-2.9.8.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.5.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-validate-common/pom.xml</p>
<p>Path to vulnerable library: /home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.5/jackson-databind-2.9.5.jar</p>
<p>
Dependency Hierarchy:
- tika-parsers-1.18.jar (Root Library)
- :x: **jackson-databind-2.9.5.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.8.10.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-fbi-service/pom.xml</p>
<p>Path to vulnerable library: nibrs-pr-test/tools/nibrs-fbi-service/target/nibrs-fbi-service-1.0.0/WEB-INF/lib/jackson-databind-2.8.10.jar,canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.8.10/jackson-databind-2.8.10.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.8.10.jar** (Vulnerable Library)
</details>
<details><summary><b>jackson-databind-2.9.6.jar</b></p></summary>
<p>General data-binding functionality for Jackson: works on core streaming API</p>
<p>Library home page: <a href="http://github.com/FasterXML/jackson">http://github.com/FasterXML/jackson</a></p>
<p>Path to dependency file: nibrs-pr-test/tools/nibrs-route/pom.xml</p>
<p>Path to vulnerable library: canner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,nibrs-pr-test/web/nibrs-web/target/nibrs-web/WEB-INF/lib/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar,/home/wss-scanner/.m2/repository/com/fasterxml/jackson/core/jackson-databind/2.9.6/jackson-databind-2.9.6.jar</p>
<p>
Dependency Hierarchy:
- :x: **jackson-databind-2.9.6.jar** (Vulnerable Library)
</details>
<p>Found in HEAD commit: <a href="https://github.com/gms-ws-sandbox/nibrs-pr-test/commit/860cc22f54e17594e32e303f0716fb065202fff5">860cc22f54e17594e32e303f0716fb065202fff5</a></p>
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly).
<p>Publish Date: 2020-04-07
<p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620>CVE-2020-11620</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.1</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: High
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: High
- Integrity Impact: High
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11620</a></p>
<p>Release Date: 2020-04-07</p>
<p>Fix Resolution: com.fasterxml.jackson.core:jackson-databind:2.9.10.4</p>
</p>
</details>
<p></p>
<!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.0","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.8.0","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.8","packageFilePaths":["/tools/nibrs-summary-report-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.springframework.boot:spring-boot-starter-web:2.1.5.RELEASE;org.springframework.boot:spring-boot-starter-json:2.1.5.RELEASE;com.fasterxml.jackson.core:jackson-databind:2.9.8","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.5","packageFilePaths":["/tools/nibrs-validate-common/pom.xml","/tools/nibrs-flatfile/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;com.fasterxml.jackson.core:jackson-databind:2.9.5","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.8.10","packageFilePaths":["/tools/nibrs-fbi-service/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"},{"packageType":"Java","groupId":"com.fasterxml.jackson.core","packageName":"jackson-databind","packageVersion":"2.9.6","packageFilePaths":["/tools/nibrs-route/pom.xml","/web/nibrs-web/pom.xml","/tools/nibrs-summary-report/pom.xml","/tools/nibrs-validation/pom.xml","/tools/nibrs-xmlfile/pom.xml","/tools/nibrs-staging-data/pom.xml","/tools/nibrs-staging-data-common/pom.xml"],"isTransitiveDependency":false,"dependencyTree":"com.fasterxml.jackson.core:jackson-databind:2.9.6","isMinimumFixVersionAvailable":true,"minimumFixVersion":"com.fasterxml.jackson.core:jackson-databind:2.9.10.4"}],"baseBranches":["master"],"vulnerabilityIdentifier":"CVE-2020-11620","vulnerabilityDetails":"FasterXML jackson-databind 2.x before 2.9.10.4 mishandles the interaction between serialization gadgets and typing, related to org.apache.commons.jelly.impl.Embedded (aka commons-jelly).","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11620","cvss3Severity":"high","cvss3Score":"8.1","cvss3Metrics":{"A":"High","AC":"High","PR":"None","S":"Unchanged","C":"High","UI":"None","AV":"Network","I":"High"},"extraData":{}}</REMEDIATE> -->
|
non_defect
|
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs summary report common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy spring boot starter web release jar root library spring boot starter json release jar x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs validate common pom xml path to vulnerable library home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy tika parsers jar root library x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs fbi service pom xml path to vulnerable library nibrs pr test tools nibrs fbi service target nibrs fbi service web inf lib jackson databind jar canner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library jackson databind jar general data binding functionality for jackson works on core streaming api library home page a href path to dependency file nibrs pr test tools nibrs route pom xml path to vulnerable library canner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar nibrs pr test web nibrs web target nibrs web web inf lib jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar home wss scanner repository com fasterxml jackson core jackson databind jackson databind jar dependency hierarchy x jackson databind jar vulnerable library found in head commit a href found in base branch master vulnerability details fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons jelly impl embedded aka commons jelly publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution com fasterxml jackson core jackson databind isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org springframework boot spring boot starter web release org springframework boot spring boot starter json release com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency true dependencytree org apache tika tika parsers com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind packagetype java groupid com fasterxml jackson core packagename jackson databind packageversion packagefilepaths istransitivedependency false dependencytree com fasterxml jackson core jackson databind isminimumfixversionavailable true minimumfixversion com fasterxml jackson core jackson databind basebranches vulnerabilityidentifier cve vulnerabilitydetails fasterxml jackson databind x before mishandles the interaction between serialization gadgets and typing related to org apache commons jelly impl embedded aka commons jelly vulnerabilityurl
| 0
|
1,154
| 2,598,004,861
|
IssuesEvent
|
2015-02-22 01:41:55
|
chrsmith/bwapi
|
https://api.github.com/repos/chrsmith/bwapi
|
opened
|
BWAPI 4 example projects still reference BWTA
|
auto-migrated Milestone-Release Priority-Critical Type-Defect
|
```
What steps will reproduce the problem?
1. Download the BWAPI 4.0.0 and install it as a full installation from
https://code.google.com/p/bwapi/downloads/list including the additional
gateways.
2. Open the example.sln in a compiler like Microsoft Visual Studio.
3. Attempt to build ExampleAIModule in release mode or debug mode.
What is the expected output? What do you see instead?
Expected output in release mode:
1>------ Build started: Project: ExampleAIModule, Configuration: Release Win32
------
1> Dll.cpp
1> ExampleAIModule.cpp
1>LINK : warning LNK4075: ignoring '/INCREMENTAL' due to '/LTCG' specification
1> Creating library C:\Program Files
(x86)\BWAPI\Release\ExampleAIModule.lib and object C:\Program Files
(x86)\BWAPI\Release\ExampleAIModule.exp
1> Generating code
1> Finished generating code
1> ExampleAIModule.vcxproj -> C:\Program Files
(x86)\BWAPI\Release\ExampleAIModule.dll
========== Build: 1 succeeded, 0 failed, 0 up-to-date, 0 skipped ==========
Instead it stops after:
1>LINK : fatal error LNK1104: cannot open file '../lib/BWTAd.lib'
What version of the product are you using? On what operating system?
Using BWAPI 4.0.0 beta on Windows 7 64 bit.
Please provide any additional information below.
If you remove the BWTA.lib from properties->linker->Input->Additional
Dependencies you will also get the errors of other libraries like CGAL.lib or
tinyxml.lib being missing. Other than BWAPI.lib, all other libs from previous
version are missing. Workaround is to copy the BWAPI 3.7.4 libs into BWAPI
4.0.0 and unlink the currently not working BWTA.lib and any others.
```
-----
Original issue reported on code.google.com by `georgetw...@gmail.com` on 24 Jun 2013 at 12:55
|
1.0
|
BWAPI 4 example projects still reference BWTA - ```
What steps will reproduce the problem?
1. Download the BWAPI 4.0.0 and install it as a full installation from
https://code.google.com/p/bwapi/downloads/list including the additional
gateways.
2. Open the example.sln in a compiler like Microsoft Visual Studio.
3. Attempt to build ExampleAIModule in release mode or debug mode.
What is the expected output? What do you see instead?
Expected output in release mode:
1>------ Build started: Project: ExampleAIModule, Configuration: Release Win32
------
1> Dll.cpp
1> ExampleAIModule.cpp
1>LINK : warning LNK4075: ignoring '/INCREMENTAL' due to '/LTCG' specification
1> Creating library C:\Program Files
(x86)\BWAPI\Release\ExampleAIModule.lib and object C:\Program Files
(x86)\BWAPI\Release\ExampleAIModule.exp
1> Generating code
1> Finished generating code
1> ExampleAIModule.vcxproj -> C:\Program Files
(x86)\BWAPI\Release\ExampleAIModule.dll
========== Build: 1 succeeded, 0 failed, 0 up-to-date, 0 skipped ==========
Instead it stops after:
1>LINK : fatal error LNK1104: cannot open file '../lib/BWTAd.lib'
What version of the product are you using? On what operating system?
Using BWAPI 4.0.0 beta on Windows 7 64 bit.
Please provide any additional information below.
If you remove the BWTA.lib from properties->linker->Input->Additional
Dependencies you will also get the errors of other libraries like CGAL.lib or
tinyxml.lib being missing. Other than BWAPI.lib, all other libs from previous
version are missing. Workaround is to copy the BWAPI 3.7.4 libs into BWAPI
4.0.0 and unlink the currently not working BWTA.lib and any others.
```
-----
Original issue reported on code.google.com by `georgetw...@gmail.com` on 24 Jun 2013 at 12:55
|
defect
|
bwapi example projects still reference bwta what steps will reproduce the problem download the bwapi and install it as a full installation from including the additional gateways open the example sln in a compiler like microsoft visual studio attempt to build exampleaimodule in release mode or debug mode what is the expected output what do you see instead expected output in release mode build started project exampleaimodule configuration release dll cpp exampleaimodule cpp link warning ignoring incremental due to ltcg specification creating library c program files bwapi release exampleaimodule lib and object c program files bwapi release exampleaimodule exp generating code finished generating code exampleaimodule vcxproj c program files bwapi release exampleaimodule dll build succeeded failed up to date skipped instead it stops after link fatal error cannot open file lib bwtad lib what version of the product are you using on what operating system using bwapi beta on windows bit please provide any additional information below if you remove the bwta lib from properties linker input additional dependencies you will also get the errors of other libraries like cgal lib or tinyxml lib being missing other than bwapi lib all other libs from previous version are missing workaround is to copy the bwapi libs into bwapi and unlink the currently not working bwta lib and any others original issue reported on code google com by georgetw gmail com on jun at
| 1
|
350,323
| 31,879,341,758
|
IssuesEvent
|
2023-09-16 07:10:58
|
depromeet/street-drop-server
|
https://api.github.com/repos/depromeet/street-drop-server
|
opened
|
Add Test Case to Geo Controller and Service
|
✅test
|
## DESCRIPTION
- Add Test Case to Geo Controller and Service
|
1.0
|
Add Test Case to Geo Controller and Service - ## DESCRIPTION
- Add Test Case to Geo Controller and Service
|
non_defect
|
add test case to geo controller and service description add test case to geo controller and service
| 0
|
80,605
| 30,386,748,880
|
IssuesEvent
|
2023-07-13 01:52:42
|
SeleniumHQ/selenium
|
https://api.github.com/repos/SeleniumHQ/selenium
|
opened
|
[🐛 Bug]: Grid closeed not wrigh, when click close button just close the remot desk page,but not disconnect the vnc.
|
I-defect needs-triaging
|
### What happened?
when click close button, just close the remot desk page,but not disconnect the vnc.
### How can we reproduce the issue?
```shell
1.connect remot desk window over grid session page.
2.open chrome develped tools,watch the network tab.
3.you can see new request over grid.
4.click close button.
5.you can see always request over grid vnc.
```
### Relevant log output
```shell
blank
```
### Operating System
linux
### Selenium version
4.8.3
### What are the browser(s) and version(s) where you see this issue?
chrome 111
### What are the browser driver(s) and version(s) where you see this issue?
use grid docker
### Are you using Selenium Grid?
4.8.3
|
1.0
|
[🐛 Bug]: Grid closeed not wrigh, when click close button just close the remot desk page,but not disconnect the vnc. - ### What happened?
when click close button, just close the remot desk page,but not disconnect the vnc.
### How can we reproduce the issue?
```shell
1.connect remot desk window over grid session page.
2.open chrome develped tools,watch the network tab.
3.you can see new request over grid.
4.click close button.
5.you can see always request over grid vnc.
```
### Relevant log output
```shell
blank
```
### Operating System
linux
### Selenium version
4.8.3
### What are the browser(s) and version(s) where you see this issue?
chrome 111
### What are the browser driver(s) and version(s) where you see this issue?
use grid docker
### Are you using Selenium Grid?
4.8.3
|
defect
|
grid closeed not wrigh when click close button just close the remot desk page but not disconnect the vnc what happened when click close button, just close the remot desk page but not disconnect the vnc how can we reproduce the issue shell connect remot desk window over grid session page open chrome develped tools watch the network tab you can see new request over grid click close button you can see always request over grid vnc relevant log output shell blank operating system linux selenium version what are the browser s and version s where you see this issue chrome what are the browser driver s and version s where you see this issue use grid docker are you using selenium grid
| 1
|
63,902
| 7,751,514,491
|
IssuesEvent
|
2018-05-30 17:18:22
|
researchstudio-sat/webofneeds
|
https://api.github.com/repos/researchstudio-sat/webofneeds
|
closed
|
Draft to Redux-State
|
UX User Story design decision enhancement suggestion
|
The draft used to be in the redux-state, it would be nice to have it there again. This would avoid data-loss if people leave the create-post view for a sec. Also it would enable later caching systems that allow to finish editing a draft later (e.g. by storing them in localstorage).
|
1.0
|
Draft to Redux-State - The draft used to be in the redux-state, it would be nice to have it there again. This would avoid data-loss if people leave the create-post view for a sec. Also it would enable later caching systems that allow to finish editing a draft later (e.g. by storing them in localstorage).
|
non_defect
|
draft to redux state the draft used to be in the redux state it would be nice to have it there again this would avoid data loss if people leave the create post view for a sec also it would enable later caching systems that allow to finish editing a draft later e g by storing them in localstorage
| 0
|
73,127
| 24,469,482,213
|
IssuesEvent
|
2022-10-07 18:16:36
|
BOINC/boinc
|
https://api.github.com/repos/BOINC/boinc
|
closed
|
Work Fetch - Leaves part of a GPU unused, when it should instead fetch work
|
C: Client - Work Fetch Policy P: Major R: wontfix T: Defect
|
**Reported by JacobKlein on 25 Mar 43272124 22:13 UTC**
If a project's GPU apps are setup to use only part of the GPU (ie: app_config.xml), then when the last remaining task(s) for that project are running and not utilizing the full GPU, work fetch should fetch more, but doesn't.
This issue was confirmed with both 7.0.60, as well as on 4/8/2013 in the simulator (which has several unreleased work fetch changes).
It would seem that the prerequisites to reproducing the bug are:
- use an app_config.xml file (to set an app to use part of a GPU, so multiple tasks could run at the same time on the same device).
- use a small buffer setting
I'm not certain if GPU Exclusions are necessary to create the issue, but I believe that using GPU Exclusions makes this problem worse.
As a workaround, I had to increase my buffer settings way above what I would normally expect. It feels like, in addition to work fetch not realizing a portion of the GPU is idle, it might also not be realizing that the tasks run 2-at-a-time.
Details, including examples in a simulation, are in the email below:
---
From: jacob_w_klein@msn.com[davea@ssl.berkeley.edu[[BR]([BR]]To:)]Subject: RE: job scheduling[Mon, 8 Apr 2013 09:51:16 -0400[[BR]([BR]]Date:)][ Thank you. I really appreciate you looking at these issues, and I'll try to verify they work.[[BR]([BR]])]Your WCG project sounds interesting; maybe they're going to support Android?[wish we had a Windows Phone platform, I'd love to test on it.[[BR]([BR]]I)][you remember Ed (Beyond) reporting a GPU Exclusion Work Fetch issue?[[BR]([BR]]Do)]I might have found examples of what he was trying to explain...[noticing an issue, both on my computer (7.0.60's work fetch algorithm), as well as the simulator (new work fetch algorithm).[[BR]([BR]][[BR]]I'm)]If a GPU is only partially-loaded (ie: 0.5 GPU) by the last remaining task(s) for a project that has GPU-Exclusions,[get into a scenario where GPUs are left part-idle, and work fetch won't fetch more.[[BR]([BR]]We)][task scheduler (correctly) schedules the workload, which is scheduled in a way where a GPU is left part-idle,[[BR]([BR]]The)]But work fetch thinks we have plenty of work, and sees no fully idle instances, so it doesn't ask for any.[are some examples where that occurred, even with our work fetch changes:[[BR]([BR]][[BR]]Here)][[[BR]([BR]]http://boinc.berkeley.edu/dev/sim_web.php?action=show_simulation&scen=86&sim=26)]2 days 17:03:00[days 14:33:00[[BR]([BR]]3)]6 days 06:13:00[days 16:43:00[[BR]([BR]]8)]9 days 16:07:00[fix might involve evaluating the project's GPU apps to see if it has any that use partial GPU[[BR]([BR]][[BR]]The)]... or maybe checking to see that all of its GPU apps use <= amount of currently idle GPU (to ensure we don't keep asking/getting work we cannot immediately use)[sounds to me like the fix for this one might be tricky instead of straight-forward, though I'm not sure.[[BR]([BR]][[BR]]It)]Do you plan on tackling this soon (fixed in short term), or should I create a ticket (fixed eventually, maybe months/years)?[[BR]][[BR]]Regards,[[BR]]Jacob
Migrated-From: http://boinc.berkeley.edu/trac/ticket/1239
|
1.0
|
Work Fetch - Leaves part of a GPU unused, when it should instead fetch work - **Reported by JacobKlein on 25 Mar 43272124 22:13 UTC**
If a project's GPU apps are setup to use only part of the GPU (ie: app_config.xml), then when the last remaining task(s) for that project are running and not utilizing the full GPU, work fetch should fetch more, but doesn't.
This issue was confirmed with both 7.0.60, as well as on 4/8/2013 in the simulator (which has several unreleased work fetch changes).
It would seem that the prerequisites to reproducing the bug are:
- use an app_config.xml file (to set an app to use part of a GPU, so multiple tasks could run at the same time on the same device).
- use a small buffer setting
I'm not certain if GPU Exclusions are necessary to create the issue, but I believe that using GPU Exclusions makes this problem worse.
As a workaround, I had to increase my buffer settings way above what I would normally expect. It feels like, in addition to work fetch not realizing a portion of the GPU is idle, it might also not be realizing that the tasks run 2-at-a-time.
Details, including examples in a simulation, are in the email below:
---
From: jacob_w_klein@msn.com[davea@ssl.berkeley.edu[[BR]([BR]]To:)]Subject: RE: job scheduling[Mon, 8 Apr 2013 09:51:16 -0400[[BR]([BR]]Date:)][ Thank you. I really appreciate you looking at these issues, and I'll try to verify they work.[[BR]([BR]])]Your WCG project sounds interesting; maybe they're going to support Android?[wish we had a Windows Phone platform, I'd love to test on it.[[BR]([BR]]I)][you remember Ed (Beyond) reporting a GPU Exclusion Work Fetch issue?[[BR]([BR]]Do)]I might have found examples of what he was trying to explain...[noticing an issue, both on my computer (7.0.60's work fetch algorithm), as well as the simulator (new work fetch algorithm).[[BR]([BR]][[BR]]I'm)]If a GPU is only partially-loaded (ie: 0.5 GPU) by the last remaining task(s) for a project that has GPU-Exclusions,[get into a scenario where GPUs are left part-idle, and work fetch won't fetch more.[[BR]([BR]]We)][task scheduler (correctly) schedules the workload, which is scheduled in a way where a GPU is left part-idle,[[BR]([BR]]The)]But work fetch thinks we have plenty of work, and sees no fully idle instances, so it doesn't ask for any.[are some examples where that occurred, even with our work fetch changes:[[BR]([BR]][[BR]]Here)][[[BR]([BR]]http://boinc.berkeley.edu/dev/sim_web.php?action=show_simulation&scen=86&sim=26)]2 days 17:03:00[days 14:33:00[[BR]([BR]]3)]6 days 06:13:00[days 16:43:00[[BR]([BR]]8)]9 days 16:07:00[fix might involve evaluating the project's GPU apps to see if it has any that use partial GPU[[BR]([BR]][[BR]]The)]... or maybe checking to see that all of its GPU apps use <= amount of currently idle GPU (to ensure we don't keep asking/getting work we cannot immediately use)[sounds to me like the fix for this one might be tricky instead of straight-forward, though I'm not sure.[[BR]([BR]][[BR]]It)]Do you plan on tackling this soon (fixed in short term), or should I create a ticket (fixed eventually, maybe months/years)?[[BR]][[BR]]Regards,[[BR]]Jacob
Migrated-From: http://boinc.berkeley.edu/trac/ticket/1239
|
defect
|
work fetch leaves part of a gpu unused when it should instead fetch work reported by jacobklein on mar utc if a project s gpu apps are setup to use only part of the gpu ie app config xml then when the last remaining task s for that project are running and not utilizing the full gpu work fetch should fetch more but doesn t this issue was confirmed with both as well as on in the simulator which has several unreleased work fetch changes it would seem that the prerequisites to reproducing the bug are use an app config xml file to set an app to use part of a gpu so multiple tasks could run at the same time on the same device use a small buffer setting i m not certain if gpu exclusions are necessary to create the issue but i believe that using gpu exclusions makes this problem worse as a workaround i had to increase my buffer settings way above what i would normally expect it feels like in addition to work fetch not realizing a portion of the gpu is idle it might also not be realizing that the tasks run at a time details including examples in a simulation are in the email below from jacob w klein msn com to subject re job scheduling date your wcg project sounds interesting maybe they re going to support android i do i might have found examples of what he was trying to explain i m if a gpu is only partially loaded ie gpu by the last remaining task s for a project that has gpu exclusions we the but work fetch thinks we have plenty of work and sees no fully idle instances so it doesn t ask for any here days days days the or maybe checking to see that all of its gpu apps use amount of currently idle gpu to ensure we don t keep asking getting work we cannot immediately use it do you plan on tackling this soon fixed in short term or should i create a ticket fixed eventually maybe months years regards jacob migrated from
| 1
|
212,720
| 16,494,949,770
|
IssuesEvent
|
2021-05-25 09:19:37
|
jeongukjae/tfds-korean
|
https://api.github.com/repos/jeongukjae/tfds-korean
|
closed
|
Add CONTRIBUTING.md
|
before-release documentation
|
- [ ] 프로젝트에서 사용하는 언어에 대한 설명. 사용법/데이터셋 설명은 가능하면 영어로 적되, 이슈/PR 소통은 한국어로 하는게 좋지 않을까?
- [ ] 데이터셋 추가하는 법
- [ ] 이슈/PR/Discussion 간단한 설명
- [ ] 추가로 같이 관리하고 싶은 분들에 대한 설명
- [ ] 데이터셋 라이선스에 대한 문제에 대한 설명
|
1.0
|
Add CONTRIBUTING.md - - [ ] 프로젝트에서 사용하는 언어에 대한 설명. 사용법/데이터셋 설명은 가능하면 영어로 적되, 이슈/PR 소통은 한국어로 하는게 좋지 않을까?
- [ ] 데이터셋 추가하는 법
- [ ] 이슈/PR/Discussion 간단한 설명
- [ ] 추가로 같이 관리하고 싶은 분들에 대한 설명
- [ ] 데이터셋 라이선스에 대한 문제에 대한 설명
|
non_defect
|
add contributing md 프로젝트에서 사용하는 언어에 대한 설명 사용법 데이터셋 설명은 가능하면 영어로 적되 이슈 pr 소통은 한국어로 하는게 좋지 않을까 데이터셋 추가하는 법 이슈 pr discussion 간단한 설명 추가로 같이 관리하고 싶은 분들에 대한 설명 데이터셋 라이선스에 대한 문제에 대한 설명
| 0
|
45,792
| 13,055,749,812
|
IssuesEvent
|
2020-07-30 02:37:26
|
icecube-trac/tix2
|
https://api.github.com/repos/icecube-trac/tix2
|
opened
|
cmake header-only projects. (Trac #62)
|
Incomplete Migration Migrated from Trac cmake defect
|
Migrated from https://code.icecube.wisc.edu/ticket/62
```json
{
"status": "closed",
"changetime": "2007-11-09T22:34:28",
"description": "header-only and executable-only projects aren't supported.",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"_ts": "1194647668000000",
"component": "cmake",
"summary": "cmake header-only projects.",
"priority": "normal",
"keywords": "",
"time": "2007-06-12T17:56:22",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
1.0
|
cmake header-only projects. (Trac #62) - Migrated from https://code.icecube.wisc.edu/ticket/62
```json
{
"status": "closed",
"changetime": "2007-11-09T22:34:28",
"description": "header-only and executable-only projects aren't supported.",
"reporter": "troy",
"cc": "",
"resolution": "duplicate",
"_ts": "1194647668000000",
"component": "cmake",
"summary": "cmake header-only projects.",
"priority": "normal",
"keywords": "",
"time": "2007-06-12T17:56:22",
"milestone": "",
"owner": "troy",
"type": "defect"
}
```
|
defect
|
cmake header only projects trac migrated from json status closed changetime description header only and executable only projects aren t supported reporter troy cc resolution duplicate ts component cmake summary cmake header only projects priority normal keywords time milestone owner troy type defect
| 1
|
731,796
| 25,231,336,992
|
IssuesEvent
|
2022-11-14 20:07:05
|
PrefectHQ/prefect
|
https://api.github.com/repos/PrefectHQ/prefect
|
closed
|
Ensure link styling gets consistently applied to all clickable links
|
ui priority:high
|
### First check
- [X] I added a descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I refreshed the page and this issue still occurred.
- [X] I checked if this issue was specific to the browser I was using by testing with a different browser.
### Bug summary
<img width="755" alt="image" src="https://user-images.githubusercontent.com/40272060/198748508-d7844d45-d03c-404b-9982-fc26f94c9312.png">
We have added link styling to work queues to make it more clear they are clickable links. We should add the same to deployment and flow run links.
In addition, let's clean up the Flow Run and Deployment wells so that they have a consistent order and group the "primary meta" details together. [Here's the figma design.](https://www.figma.com/file/WQwcuqBAxug8JqLigWaztu/Designs?node-id=1600%3A19270) Notice that date formatting is more consistent as well.
<img width="1307" alt="Screen Shot 2022-11-01 at 11 05 37 AM" src="https://user-images.githubusercontent.com/6776415/199266121-bb787d15-c162-4f23-b71f-a906b4fa3ee2.png">
### Reproduction
See image above
### Error
See image above
### Browers
- [X] Chrome
- [ ] Firefox
- [ ] Safari
- [ ] Edge
### Prefect version
_No response_
### Additional context
_No response_
|
1.0
|
Ensure link styling gets consistently applied to all clickable links - ### First check
- [X] I added a descriptive title to this issue.
- [X] I used the GitHub search to find a similar issue and didn't find it.
- [X] I refreshed the page and this issue still occurred.
- [X] I checked if this issue was specific to the browser I was using by testing with a different browser.
### Bug summary
<img width="755" alt="image" src="https://user-images.githubusercontent.com/40272060/198748508-d7844d45-d03c-404b-9982-fc26f94c9312.png">
We have added link styling to work queues to make it more clear they are clickable links. We should add the same to deployment and flow run links.
In addition, let's clean up the Flow Run and Deployment wells so that they have a consistent order and group the "primary meta" details together. [Here's the figma design.](https://www.figma.com/file/WQwcuqBAxug8JqLigWaztu/Designs?node-id=1600%3A19270) Notice that date formatting is more consistent as well.
<img width="1307" alt="Screen Shot 2022-11-01 at 11 05 37 AM" src="https://user-images.githubusercontent.com/6776415/199266121-bb787d15-c162-4f23-b71f-a906b4fa3ee2.png">
### Reproduction
See image above
### Error
See image above
### Browers
- [X] Chrome
- [ ] Firefox
- [ ] Safari
- [ ] Edge
### Prefect version
_No response_
### Additional context
_No response_
|
non_defect
|
ensure link styling gets consistently applied to all clickable links first check i added a descriptive title to this issue i used the github search to find a similar issue and didn t find it i refreshed the page and this issue still occurred i checked if this issue was specific to the browser i was using by testing with a different browser bug summary img width alt image src we have added link styling to work queues to make it more clear they are clickable links we should add the same to deployment and flow run links in addition let s clean up the flow run and deployment wells so that they have a consistent order and group the primary meta details together notice that date formatting is more consistent as well img width alt screen shot at am src reproduction see image above error see image above browers chrome firefox safari edge prefect version no response additional context no response
| 0
|
41,738
| 10,583,959,405
|
IssuesEvent
|
2019-10-08 14:37:46
|
zfsonlinux/zfs
|
https://api.github.com/repos/zfsonlinux/zfs
|
closed
|
After a week of running array, issuing zpool scrub causes system hang
|
Type: Defect
|
<!--
Thank you for reporting an issue.
*IMPORTANT* - Please search our issue tracker *before* making a new issue.
If you cannot find a similar issue, then create a new issue.
https://github.com/zfsonlinux/zfs/issues
*IMPORTANT* - This issue tracker is for *bugs* and *issues* only.
Please search the wiki and the mailing list archives before asking
questions on the mailing list.
https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Dell R620 | Sandy Bridge
--- | ---
Distribution Name | Gentoo
Distribution Version | Rolling
Linux Kernel | 4.15.16
Architecture | x86_64
ZFS Version | 0.7.9-r0-gentoo
SPL Version | 0.7.9-r0-gentoo
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
My array, which is made up of 10x5TB disks attached to an LSI 9300 SAS controller running RAIDz2, and is currently about 18.2TB full out of ~36TB of usable storage, is seeing a system hang after zpool scrub runs.
However, I can run zpool scrub on my pool after a fresh reboot, and the scrub runs to completion with no issues (and finds no problems). But if I have the scrub run out of cron once a week, as it has been running for about 2 years now, it will cause the system to become unresponsive. If I run the scrub manually after about a week of running the system, the same behavior occurs.
My system is a Dell R620 running Gentoo. It has 72GB of ECC RAM. I have checked SMART data on all the disks, and run other health checks against the RAM, and nothing has indicated a hardware issue. This started occurring after upgrading ZFS to 0.7.x at some point. I honestly don't know where the cutoff happened, since I wrote off the the strange crashes as anomalies until I noticed the pattern.
### Describe how to reproduce the problem
My system just has to run for about a week doing its normal workloads (a couple of VMs, serving data to my Plex server, etc.), and then kick off a zpool scrub on the pool. This can be done via cron or interactively. Either way, same issue.
### Include any warning/errors/backtraces from the system logs
I'm in the process of getting a serial console hooked up to capture this. Because my cron job is currently scheduled to run at 2am on Sunday's, I forget to get this configured until it's too late. I'm hoping someone has also seen this (Google'ing around did show some similar issues, but nothing specific).
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
|
1.0
|
After a week of running array, issuing zpool scrub causes system hang - <!--
Thank you for reporting an issue.
*IMPORTANT* - Please search our issue tracker *before* making a new issue.
If you cannot find a similar issue, then create a new issue.
https://github.com/zfsonlinux/zfs/issues
*IMPORTANT* - This issue tracker is for *bugs* and *issues* only.
Please search the wiki and the mailing list archives before asking
questions on the mailing list.
https://github.com/zfsonlinux/zfs/wiki/Mailing-Lists
Please fill in as much of the template as possible.
-->
### System information
<!-- add version after "|" character -->
Dell R620 | Sandy Bridge
--- | ---
Distribution Name | Gentoo
Distribution Version | Rolling
Linux Kernel | 4.15.16
Architecture | x86_64
ZFS Version | 0.7.9-r0-gentoo
SPL Version | 0.7.9-r0-gentoo
<!--
Commands to find ZFS/SPL versions:
modinfo zfs | grep -iw version
modinfo spl | grep -iw version
-->
### Describe the problem you're observing
My array, which is made up of 10x5TB disks attached to an LSI 9300 SAS controller running RAIDz2, and is currently about 18.2TB full out of ~36TB of usable storage, is seeing a system hang after zpool scrub runs.
However, I can run zpool scrub on my pool after a fresh reboot, and the scrub runs to completion with no issues (and finds no problems). But if I have the scrub run out of cron once a week, as it has been running for about 2 years now, it will cause the system to become unresponsive. If I run the scrub manually after about a week of running the system, the same behavior occurs.
My system is a Dell R620 running Gentoo. It has 72GB of ECC RAM. I have checked SMART data on all the disks, and run other health checks against the RAM, and nothing has indicated a hardware issue. This started occurring after upgrading ZFS to 0.7.x at some point. I honestly don't know where the cutoff happened, since I wrote off the the strange crashes as anomalies until I noticed the pattern.
### Describe how to reproduce the problem
My system just has to run for about a week doing its normal workloads (a couple of VMs, serving data to my Plex server, etc.), and then kick off a zpool scrub on the pool. This can be done via cron or interactively. Either way, same issue.
### Include any warning/errors/backtraces from the system logs
I'm in the process of getting a serial console hooked up to capture this. Because my cron job is currently scheduled to run at 2am on Sunday's, I forget to get this configured until it's too late. I'm hoping someone has also seen this (Google'ing around did show some similar issues, but nothing specific).
<!--
*IMPORTANT* - Please mark logs and text output from terminal commands
or else Github will not display them correctly.
An example is provided below.
Example:
```
this is an example how log text should be marked (wrap it with ```)
```
-->
|
defect
|
after a week of running array issuing zpool scrub causes system hang thank you for reporting an issue important please search our issue tracker before making a new issue if you cannot find a similar issue then create a new issue important this issue tracker is for bugs and issues only please search the wiki and the mailing list archives before asking questions on the mailing list please fill in as much of the template as possible system information dell sandy bridge distribution name gentoo distribution version rolling linux kernel architecture zfs version gentoo spl version gentoo commands to find zfs spl versions modinfo zfs grep iw version modinfo spl grep iw version describe the problem you re observing my array which is made up of disks attached to an lsi sas controller running and is currently about full out of of usable storage is seeing a system hang after zpool scrub runs however i can run zpool scrub on my pool after a fresh reboot and the scrub runs to completion with no issues and finds no problems but if i have the scrub run out of cron once a week as it has been running for about years now it will cause the system to become unresponsive if i run the scrub manually after about a week of running the system the same behavior occurs my system is a dell running gentoo it has of ecc ram i have checked smart data on all the disks and run other health checks against the ram and nothing has indicated a hardware issue this started occurring after upgrading zfs to x at some point i honestly don t know where the cutoff happened since i wrote off the the strange crashes as anomalies until i noticed the pattern describe how to reproduce the problem my system just has to run for about a week doing its normal workloads a couple of vms serving data to my plex server etc and then kick off a zpool scrub on the pool this can be done via cron or interactively either way same issue include any warning errors backtraces from the system logs i m in the process of getting a serial console hooked up to capture this because my cron job is currently scheduled to run at on sunday s i forget to get this configured until it s too late i m hoping someone has also seen this google ing around did show some similar issues but nothing specific important please mark logs and text output from terminal commands or else github will not display them correctly an example is provided below example this is an example how log text should be marked wrap it with
| 1
|
19,109
| 3,142,923,502
|
IssuesEvent
|
2015-09-14 00:08:22
|
CompEvol/beast2
|
https://api.github.com/repos/CompEvol/beast2
|
closed
|
Automatic detection of package directories gives false positives
|
defect MEDIUM priority
|
The method `AddOnManager.getBeastDirectories()` returns a list of directories that "may contain packages". All jar files within the lib/ subdirectory of these directories are loaded by `AddOnManager.loadExternalJars()`. If a jar file contains a class that is already in the class path, that jar file will not be loaded.
One of the criteria for a directory to be included in the list returned by `getBeastDirectories()` is that the directory is a subdirectory of the current directory (or whatever System.getProperty("user.dir") returns on your system) and that it contains a "/lib" or "/templates" directory.
I contend that this is not a strict enough criterion, as /lib in particular is a very common directory name on *nix systems. As the "user.dir" subdirectories are included before the official package directories (eg $HOME/.beast/2.1/*), jars detected here may prevent the installed package jars containing at least one identical class from being loaded. This could cause beast to behave in surprising ways if, for instance, an installed package used a different version of a library than the one picked up from a directory such as "user.dir"/BLAH/lib.
An alternative might be to look for (and ideally parse) "version.xml".
The related question of how to handle real packages using different versions of the same library would still exist though.
|
1.0
|
Automatic detection of package directories gives false positives - The method `AddOnManager.getBeastDirectories()` returns a list of directories that "may contain packages". All jar files within the lib/ subdirectory of these directories are loaded by `AddOnManager.loadExternalJars()`. If a jar file contains a class that is already in the class path, that jar file will not be loaded.
One of the criteria for a directory to be included in the list returned by `getBeastDirectories()` is that the directory is a subdirectory of the current directory (or whatever System.getProperty("user.dir") returns on your system) and that it contains a "/lib" or "/templates" directory.
I contend that this is not a strict enough criterion, as /lib in particular is a very common directory name on *nix systems. As the "user.dir" subdirectories are included before the official package directories (eg $HOME/.beast/2.1/*), jars detected here may prevent the installed package jars containing at least one identical class from being loaded. This could cause beast to behave in surprising ways if, for instance, an installed package used a different version of a library than the one picked up from a directory such as "user.dir"/BLAH/lib.
An alternative might be to look for (and ideally parse) "version.xml".
The related question of how to handle real packages using different versions of the same library would still exist though.
|
defect
|
automatic detection of package directories gives false positives the method addonmanager getbeastdirectories returns a list of directories that may contain packages all jar files within the lib subdirectory of these directories are loaded by addonmanager loadexternaljars if a jar file contains a class that is already in the class path that jar file will not be loaded one of the criteria for a directory to be included in the list returned by getbeastdirectories is that the directory is a subdirectory of the current directory or whatever system getproperty user dir returns on your system and that it contains a lib or templates directory i contend that this is not a strict enough criterion as lib in particular is a very common directory name on nix systems as the user dir subdirectories are included before the official package directories eg home beast jars detected here may prevent the installed package jars containing at least one identical class from being loaded this could cause beast to behave in surprising ways if for instance an installed package used a different version of a library than the one picked up from a directory such as user dir blah lib an alternative might be to look for and ideally parse version xml the related question of how to handle real packages using different versions of the same library would still exist though
| 1
|
74,152
| 24,978,172,181
|
IssuesEvent
|
2022-11-02 09:35:39
|
matrix-org/matrix-public-archive
|
https://api.github.com/repos/matrix-org/matrix-public-archive
|
closed
|
Don't change the URL as you scroll around
|
T-Defect A-archive-room-view
|
Even if we had https://github.com/matrix-org/matrix-public-archive/issues/72 to page-load the room with a stable URL, it still probably doesn't make sense to change the URL around as you scroll. The content on the page should be consistent with the URL.
> ## UX problems
>
> ### One person's link will not look like the other once they start scrolling
>
> Because we change the URL as you scroll around. Someone may link a page expecting all the same messages to be there but it will actually only be the current day they're scrolled to plus whatever is above it.
>
> *-- https://github.com/matrix-org/matrix-public-archive/pull/71*
It will probably also affect how search engines index the site attributing messages with a certain URL that when visiting will be missing.
---
I think it does make sense to still change the calendar and text around to make it easy to jump to another day from where you're currently viewing in the surrounding context.
|
1.0
|
Don't change the URL as you scroll around - Even if we had https://github.com/matrix-org/matrix-public-archive/issues/72 to page-load the room with a stable URL, it still probably doesn't make sense to change the URL around as you scroll. The content on the page should be consistent with the URL.
> ## UX problems
>
> ### One person's link will not look like the other once they start scrolling
>
> Because we change the URL as you scroll around. Someone may link a page expecting all the same messages to be there but it will actually only be the current day they're scrolled to plus whatever is above it.
>
> *-- https://github.com/matrix-org/matrix-public-archive/pull/71*
It will probably also affect how search engines index the site attributing messages with a certain URL that when visiting will be missing.
---
I think it does make sense to still change the calendar and text around to make it easy to jump to another day from where you're currently viewing in the surrounding context.
|
defect
|
don t change the url as you scroll around even if we had to page load the room with a stable url it still probably doesn t make sense to change the url around as you scroll the content on the page should be consistent with the url ux problems one person s link will not look like the other once they start scrolling because we change the url as you scroll around someone may link a page expecting all the same messages to be there but it will actually only be the current day they re scrolled to plus whatever is above it it will probably also affect how search engines index the site attributing messages with a certain url that when visiting will be missing i think it does make sense to still change the calendar and text around to make it easy to jump to another day from where you re currently viewing in the surrounding context
| 1
|
27,267
| 4,956,984,106
|
IssuesEvent
|
2016-12-02 01:53:33
|
cakephp/cakephp
|
https://api.github.com/repos/cakephp/cakephp
|
closed
|
ExistsIn Rule (since 3.3) causing SQL generation issue
|
Defect ORM validation
|
- [x] bug
- [ ] enhancement
- [ ] feature-discussion (RFC)
- CakePHP Version: 3.3.0
- Platform and Target: Apache 2.4.18,PHP 5.6,MYSQL 5.5.51
### What you did
table1 has 'hasOne' relationship with table2
``` php
$this->hasOne('table2', [
'className' => 'table2',
'foreignKey' => false,
'joinType' => 'inner',
'conditions' => ['table1.target_id = table2.id'],
]);
```
``` php
$rules->add($rules->existsIn(['target_id'], 'table2'));
```
### Expected Behavior
cake 3.2.14
before when inserting a value in table1 cake was checking in table2 if target_id existing without including the condition of the relation
``` sql
SELECT 1 AS `existing` FROM `table2` `table2` WHERE `table2`.`id` = 1 LIMIT 1
```
### Actual Behavior
cake 3.3.0
now cake add the relation to the query and broke the query because in this query table1 is unknown
``` sql
SELECT 1 AS `existing` FROM `table2` `table2` WHERE (`table2`.`id` = 1 AND table1.target_id = table2.id) LIMIT 1
```
|
1.0
|
ExistsIn Rule (since 3.3) causing SQL generation issue - - [x] bug
- [ ] enhancement
- [ ] feature-discussion (RFC)
- CakePHP Version: 3.3.0
- Platform and Target: Apache 2.4.18,PHP 5.6,MYSQL 5.5.51
### What you did
table1 has 'hasOne' relationship with table2
``` php
$this->hasOne('table2', [
'className' => 'table2',
'foreignKey' => false,
'joinType' => 'inner',
'conditions' => ['table1.target_id = table2.id'],
]);
```
``` php
$rules->add($rules->existsIn(['target_id'], 'table2'));
```
### Expected Behavior
cake 3.2.14
before when inserting a value in table1 cake was checking in table2 if target_id existing without including the condition of the relation
``` sql
SELECT 1 AS `existing` FROM `table2` `table2` WHERE `table2`.`id` = 1 LIMIT 1
```
### Actual Behavior
cake 3.3.0
now cake add the relation to the query and broke the query because in this query table1 is unknown
``` sql
SELECT 1 AS `existing` FROM `table2` `table2` WHERE (`table2`.`id` = 1 AND table1.target_id = table2.id) LIMIT 1
```
|
defect
|
existsin rule since causing sql generation issue bug enhancement feature discussion rfc cakephp version platform and target apache php mysql what you did has hasone relationship with php this hasone classname foreignkey false jointype inner conditions php rules add rules existsin expected behavior cake before when inserting a value in cake was checking in if target id existing without including the condition of the relation sql select as existing from where id limit actual behavior cake now cake add the relation to the query and broke the query because in this query is unknown sql select as existing from where id and target id id limit
| 1
|
252,565
| 27,245,824,211
|
IssuesEvent
|
2023-02-22 01:51:41
|
idmarinas/lotgd-modules
|
https://api.github.com/repos/idmarinas/lotgd-modules
|
closed
|
CVE-2021-35065 (High) detected in glob-parent-5.1.2.tgz - autoclosed
|
security vulnerability
|
## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- del-6.0.0.tgz (Root Library)
- globby-11.0.4.tgz
- fast-glob-3.2.7.tgz
- :x: **glob-parent-5.1.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent from 6.0.0 and before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
True
|
CVE-2021-35065 (High) detected in glob-parent-5.1.2.tgz - autoclosed - ## CVE-2021-35065 - High Severity Vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>glob-parent-5.1.2.tgz</b></p></summary>
<p>Extract the non-magic parent path from a glob string.</p>
<p>Library home page: <a href="https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz">https://registry.npmjs.org/glob-parent/-/glob-parent-5.1.2.tgz</a></p>
<p>Path to dependency file: /package.json</p>
<p>Path to vulnerable library: /node_modules/glob-parent/package.json</p>
<p>
Dependency Hierarchy:
- del-6.0.0.tgz (Root Library)
- globby-11.0.4.tgz
- fast-glob-3.2.7.tgz
- :x: **glob-parent-5.1.2.tgz** (Vulnerable Library)
<p>Found in base branch: <b>master</b></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary>
<p>
The package glob-parent from 6.0.0 and before 6.0.1 are vulnerable to Regular Expression Denial of Service (ReDoS)
<p>Publish Date: 2021-06-22
<p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-35065>CVE-2021-35065</a></p>
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary>
<p>
Base Score Metrics:
- Exploitability Metrics:
- Attack Vector: Network
- Attack Complexity: Low
- Privileges Required: None
- User Interaction: None
- Scope: Unchanged
- Impact Metrics:
- Confidentiality Impact: None
- Integrity Impact: None
- Availability Impact: High
</p>
For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>.
</p>
</details>
<p></p>
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary>
<p>
<p>Type: Upgrade version</p>
<p>Origin: <a href="https://github.com/advisories/GHSA-cj88-88mr-972w">https://github.com/advisories/GHSA-cj88-88mr-972w</a></p>
<p>Release Date: 2021-06-22</p>
<p>Fix Resolution: glob-parent - 6.0.1</p>
</p>
</details>
<p></p>
***
Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
|
non_defect
|
cve high detected in glob parent tgz autoclosed cve high severity vulnerability vulnerable library glob parent tgz extract the non magic parent path from a glob string library home page a href path to dependency file package json path to vulnerable library node modules glob parent package json dependency hierarchy del tgz root library globby tgz fast glob tgz x glob parent tgz vulnerable library found in base branch master vulnerability details the package glob parent from and before are vulnerable to regular expression denial of service redos publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution glob parent step up your open source security game with mend
| 0
|
6,516
| 14,707,438,935
|
IssuesEvent
|
2021-01-04 21:37:19
|
MicrosoftDocs/architecture-center
|
https://api.github.com/repos/MicrosoftDocs/architecture-center
|
closed
|
Change title to Reference Architecture or Create a new article for it!
|
architecture-center/svc cxp doc-idea example-scenario/subsvc triaged
|
A number of my customers are looking for reference architecture and this article is the closest to it (https://docs.microsoft.com/en-us/azure/architecture/example-scenario/wvd/windows-virtual-desktop ) Can the title of this article be changed to reference arch. or have a new article written for it?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f26be54d-0f67-5098-7839-98f5bb064188
* Version Independent ID: 187fa3ce-ade9-1b07-2af3-c874ec3d0aac
* Content: [Windows Virtual Desktop for the enterprise - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/wvd/windows-virtual-desktop)
* Content Source: [docs/example-scenario/wvd/windows-virtual-desktop.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/wvd/windows-virtual-desktop.yml)
* Service: **architecture-center**
* Sub-service: **example-scenario**
* GitHub Login: @doodlemania2
* Microsoft Alias: **pnp**
|
1.0
|
Change title to Reference Architecture or Create a new article for it! -
A number of my customers are looking for reference architecture and this article is the closest to it (https://docs.microsoft.com/en-us/azure/architecture/example-scenario/wvd/windows-virtual-desktop ) Can the title of this article be changed to reference arch. or have a new article written for it?
---
#### Document Details
⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.*
* ID: f26be54d-0f67-5098-7839-98f5bb064188
* Version Independent ID: 187fa3ce-ade9-1b07-2af3-c874ec3d0aac
* Content: [Windows Virtual Desktop for the enterprise - Azure Example Scenarios](https://docs.microsoft.com/en-us/azure/architecture/example-scenario/wvd/windows-virtual-desktop)
* Content Source: [docs/example-scenario/wvd/windows-virtual-desktop.yml](https://github.com/microsoftdocs/architecture-center/blob/master/docs/example-scenario/wvd/windows-virtual-desktop.yml)
* Service: **architecture-center**
* Sub-service: **example-scenario**
* GitHub Login: @doodlemania2
* Microsoft Alias: **pnp**
|
non_defect
|
change title to reference architecture or create a new article for it a number of my customers are looking for reference architecture and this article is the closest to it can the title of this article be changed to reference arch or have a new article written for it document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service architecture center sub service example scenario github login microsoft alias pnp
| 0
|
79,371
| 28,136,764,348
|
IssuesEvent
|
2023-04-01 13:21:56
|
dkfans/keeperfx
|
https://api.github.com/repos/dkfans/keeperfx
|
closed
|
Heap crash in log even though heap is gone
|
Type-Defect Priority-Low
|
Start any map and see this in the log: `=== Crash ===Sync: GraphicsHeap Size 50331648`
Most likely caused by 8443f4094f9accc6725ce39e9a6b9062472c5ceb
|
1.0
|
Heap crash in log even though heap is gone - Start any map and see this in the log: `=== Crash ===Sync: GraphicsHeap Size 50331648`
Most likely caused by 8443f4094f9accc6725ce39e9a6b9062472c5ceb
|
defect
|
heap crash in log even though heap is gone start any map and see this in the log crash sync graphicsheap size most likely caused by
| 1
|
307,187
| 23,188,305,257
|
IssuesEvent
|
2022-08-01 10:19:29
|
aitos-io/BoAT-X-Framework
|
https://api.github.com/repos/aitos-io/BoAT-X-Framework
|
closed
|
Some functions of utility lack annotation descriptions
|
documentation good first issue
|
As the title says:
<img width="626" alt="image" src="https://user-images.githubusercontent.com/44628575/178891105-ce13857f-c26b-44e7-8609-abdda89474d0.png">
|
1.0
|
Some functions of utility lack annotation descriptions - As the title says:
<img width="626" alt="image" src="https://user-images.githubusercontent.com/44628575/178891105-ce13857f-c26b-44e7-8609-abdda89474d0.png">
|
non_defect
|
some functions of utility lack annotation descriptions as the title says: img width alt image src
| 0
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.