Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
7
112
repo_url
stringlengths
36
141
action
stringclasses
3 values
title
stringlengths
1
744
labels
stringlengths
4
574
body
stringlengths
9
211k
index
stringclasses
10 values
text_combine
stringlengths
96
211k
label
stringclasses
2 values
text
stringlengths
96
188k
binary_label
int64
0
1
46,360
13,163,815,260
IssuesEvent
2020-08-11 01:40:34
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
dotnet implementation of AES decryption is slower than native call to bcrypt
area-System.Security tenet-performance
### Description I've compared 3 different implementations of AES 128 decrypt as a simple benchmark project. | Method | Time | Ratio | |-|-|-| | DecryptAesDotNetTransform | 5.76 µs | 1.00 baseline | | DecryptAesBCrypt | 4.35 µs | 0.76 | | DecryptAesDotNetStream | 9.32 µs | 1.62 | You can find implementation in https://github.com/vsevolodp/dotnet-aes-benchmark/blob/master/benchmark_simple/Program.cs To me this looks like significant performance issue. In other project in my repository I've compared dotnet core 3.1 to 5.0 preview and this perf issue is same on .net 5.0. Also it is much slower than OpenSSL implementation. ## Numbers from detailed benchmark project https://github.com/vsevolodp/dotnet-aes-benchmark/tree/master/benchmark AesDecryptDotNet128 is same implementation as DecryptAesDotNetStream above. AesDecryptDotNet128Type2 is same implementation as DecryptAesDotNetTransform above. ### Windows Host ``` BenchmarkDotNet=v0.12.1, OS=Windows 10.0.19041.329 (2004/?/20H1) Intel Core i7-6700 CPU 3.40GHz (Skylake), 1 CPU, 8 logical and 4 physical cores .NET Core SDK=5.0.100-preview.4.20258.7 [Host] : .NET Core 3.1.5 (CoreCLR 4.700.20.26901, CoreFX 4.700.20.27001), X64 RyuJIT .NET Core 3.1 : .NET Core 3.1.5 (CoreCLR 4.700.20.26901, CoreFX 4.700.20.27001), X64 RyuJIT .NET Core 5.0 : .NET Core 5.0.0 (CoreCLR 5.0.20.25106, CoreFX 5.0.20.25106), X64 RyuJIT ``` | Method | Runtime | Mean | Error | StdDev | Ratio | |------------------------- |-------------- |---------:|----------:|----------:|------:| | AesDecryptDotNet128 | .NET Core 3.1 | 9.925 us | 0.1985 us | 0.4906 us | 1.82 | | AesDecryptDotNet128Type2 | .NET Core 3.1 | 6.014 us | 0.1185 us | 0.1541 us | 1.03 | | AesDecryptBCryptWin128 | .NET Core 3.1 | 4.499 us | 0.0521 us | 0.0462 us | 0.78 | | AesDecryptOpenSsl128 | .NET Core 3.1 | 4.463 us | 0.0708 us | 0.0591 us | 0.77 | | AesDecryptDotNet128 | .NET Core 5.0 | 9.467 us | 0.0692 us | 0.0613 us | 1.63 | | AesDecryptDotNet128Type2 | .NET Core 5.0 | 5.804 us | 0.0465 us | 0.0412 us | 1.00 baseline | | AesDecryptBCryptWin128 | .NET Core 5.0 | 4.442 us | 0.0352 us | 0.0329 us | 0.77 | | AesDecryptOpenSsl128 | .NET Core 5.0 | 4.138 us | 0.0324 us | 0.0271 us | 0.71 | ### WSL2 (Ubuntu 18.04) on Windows Host ``` BenchmarkDotNet=v0.12.1, OS=ubuntu 18.04 Intel Core i7-6700 CPU 3.40GHz (Skylake), 1 CPU, 8 logical and 4 physical cores .NET Core SDK=5.0.100-preview.5.20279.10 [Host] : .NET Core 3.1.3 (CoreCLR 4.700.20.11803, CoreFX 4.700.20.12001), X64 RyuJIT .NET Core 3.1 : .NET Core 3.1.3 (CoreCLR 4.700.20.11803, CoreFX 4.700.20.12001), X64 RyuJIT .NET Core 5.0 : .NET Core 5.0.0 (CoreCLR 5.0.20.27801, CoreFX 5.0.20.27801), X64 RyuJIT ``` | Method | Runtime | Mean | Error | StdDev | Ratio | |------------------------- |-------------- |---------:|----------:|----------:|------:| | AesDecryptDotNet128 | .NET Core 3.1 | 9.572 us | 0.1426 us | 0.1585 us | 1.699 | | AesDecryptDotNet128Type2 | .NET Core 3.1 | 5.336 us | 0.1040 us | 0.1239 us | 0.946 | | AesDecryptOpenSsl128 | .NET Core 3.1 | 3.746 us | 0.0644 us | 0.0602 us | 0.669 | | AesDecryptDotNet128 | .NET Core 5.0 | 9.922 us | 0.1156 us | 0.1025 us | 1.771 | | AesDecryptDotNet128Type2 | .NET Core 5.0 | 5.577 us | 0.1040 us | 0.1900 us | 1.000 baseline | | AesDecryptOpenSsl128 | .NET Core 5.0 | 3.690 us | 0.0441 us | 0.0368 us | 0.655 | ### Regression? No
True
dotnet implementation of AES decryption is slower than native call to bcrypt - ### Description I've compared 3 different implementations of AES 128 decrypt as a simple benchmark project. | Method | Time | Ratio | |-|-|-| | DecryptAesDotNetTransform | 5.76 µs | 1.00 baseline | | DecryptAesBCrypt | 4.35 µs | 0.76 | | DecryptAesDotNetStream | 9.32 µs | 1.62 | You can find implementation in https://github.com/vsevolodp/dotnet-aes-benchmark/blob/master/benchmark_simple/Program.cs To me this looks like significant performance issue. In other project in my repository I've compared dotnet core 3.1 to 5.0 preview and this perf issue is same on .net 5.0. Also it is much slower than OpenSSL implementation. ## Numbers from detailed benchmark project https://github.com/vsevolodp/dotnet-aes-benchmark/tree/master/benchmark AesDecryptDotNet128 is same implementation as DecryptAesDotNetStream above. AesDecryptDotNet128Type2 is same implementation as DecryptAesDotNetTransform above. ### Windows Host ``` BenchmarkDotNet=v0.12.1, OS=Windows 10.0.19041.329 (2004/?/20H1) Intel Core i7-6700 CPU 3.40GHz (Skylake), 1 CPU, 8 logical and 4 physical cores .NET Core SDK=5.0.100-preview.4.20258.7 [Host] : .NET Core 3.1.5 (CoreCLR 4.700.20.26901, CoreFX 4.700.20.27001), X64 RyuJIT .NET Core 3.1 : .NET Core 3.1.5 (CoreCLR 4.700.20.26901, CoreFX 4.700.20.27001), X64 RyuJIT .NET Core 5.0 : .NET Core 5.0.0 (CoreCLR 5.0.20.25106, CoreFX 5.0.20.25106), X64 RyuJIT ``` | Method | Runtime | Mean | Error | StdDev | Ratio | |------------------------- |-------------- |---------:|----------:|----------:|------:| | AesDecryptDotNet128 | .NET Core 3.1 | 9.925 us | 0.1985 us | 0.4906 us | 1.82 | | AesDecryptDotNet128Type2 | .NET Core 3.1 | 6.014 us | 0.1185 us | 0.1541 us | 1.03 | | AesDecryptBCryptWin128 | .NET Core 3.1 | 4.499 us | 0.0521 us | 0.0462 us | 0.78 | | AesDecryptOpenSsl128 | .NET Core 3.1 | 4.463 us | 0.0708 us | 0.0591 us | 0.77 | | AesDecryptDotNet128 | .NET Core 5.0 | 9.467 us | 0.0692 us | 0.0613 us | 1.63 | | AesDecryptDotNet128Type2 | .NET Core 5.0 | 5.804 us | 0.0465 us | 0.0412 us | 1.00 baseline | | AesDecryptBCryptWin128 | .NET Core 5.0 | 4.442 us | 0.0352 us | 0.0329 us | 0.77 | | AesDecryptOpenSsl128 | .NET Core 5.0 | 4.138 us | 0.0324 us | 0.0271 us | 0.71 | ### WSL2 (Ubuntu 18.04) on Windows Host ``` BenchmarkDotNet=v0.12.1, OS=ubuntu 18.04 Intel Core i7-6700 CPU 3.40GHz (Skylake), 1 CPU, 8 logical and 4 physical cores .NET Core SDK=5.0.100-preview.5.20279.10 [Host] : .NET Core 3.1.3 (CoreCLR 4.700.20.11803, CoreFX 4.700.20.12001), X64 RyuJIT .NET Core 3.1 : .NET Core 3.1.3 (CoreCLR 4.700.20.11803, CoreFX 4.700.20.12001), X64 RyuJIT .NET Core 5.0 : .NET Core 5.0.0 (CoreCLR 5.0.20.27801, CoreFX 5.0.20.27801), X64 RyuJIT ``` | Method | Runtime | Mean | Error | StdDev | Ratio | |------------------------- |-------------- |---------:|----------:|----------:|------:| | AesDecryptDotNet128 | .NET Core 3.1 | 9.572 us | 0.1426 us | 0.1585 us | 1.699 | | AesDecryptDotNet128Type2 | .NET Core 3.1 | 5.336 us | 0.1040 us | 0.1239 us | 0.946 | | AesDecryptOpenSsl128 | .NET Core 3.1 | 3.746 us | 0.0644 us | 0.0602 us | 0.669 | | AesDecryptDotNet128 | .NET Core 5.0 | 9.922 us | 0.1156 us | 0.1025 us | 1.771 | | AesDecryptDotNet128Type2 | .NET Core 5.0 | 5.577 us | 0.1040 us | 0.1900 us | 1.000 baseline | | AesDecryptOpenSsl128 | .NET Core 5.0 | 3.690 us | 0.0441 us | 0.0368 us | 0.655 | ### Regression? No
non_process
dotnet implementation of aes decryption is slower than native call to bcrypt description i ve compared different implementations of aes decrypt as a simple benchmark project method time ratio decryptaesdotnettransform µs baseline decryptaesbcrypt µs decryptaesdotnetstream µs you can find implementation in to me this looks like significant performance issue in other project in my repository i ve compared dotnet core to preview and this perf issue is same on net also it is much slower than openssl implementation numbers from detailed benchmark project is same implementation as decryptaesdotnetstream above is same implementation as decryptaesdotnettransform above windows host benchmarkdotnet os windows intel core cpu skylake cpu logical and physical cores net core sdk preview net core coreclr corefx ryujit net core net core coreclr corefx ryujit net core net core coreclr corefx ryujit method runtime mean error stddev ratio net core us us us net core us us us net core us us us net core us us us net core us us us net core us us us baseline net core us us us net core us us us ubuntu on windows host benchmarkdotnet os ubuntu intel core cpu skylake cpu logical and physical cores net core sdk preview net core coreclr corefx ryujit net core net core coreclr corefx ryujit net core net core coreclr corefx ryujit method runtime mean error stddev ratio net core us us us net core us us us net core us us us net core us us us net core us us us baseline net core us us us regression no
0
9,407
12,405,672,359
IssuesEvent
2020-05-21 17:42:30
tokio-rs/tokio
https://api.github.com/repos/tokio-rs/tokio
closed
Several AsyncRead impls unnecessarily zero buffers in prepare_uninitialized_buffer
A-tokio C-bug E-easy E-help-wanted M-io M-process T-performance
<!-- Thank you for reporting an issue. Please fill in as much of the template below as you're able. --> ## Version <!-- List the versions of all `tokio` crates you are using. The easiest way to get this information is using `cargo-tree`. `cargo install cargo-tree` (see install here: https://github.com/sfackler/cargo-tree) Then: `cargo tree | grep tokio` --> `tokio` 0.2.13 ## Platform <!--- Output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) --> This is an issue for all platforms. ## Subcrates <!-- If known, please specify the affected Tokio sub crates. Otherwise, delete this section. --> I haven't reviewed the full source, but I've reviewed `process` and can confirm that the `ChildStdout` and `ChildStderr` types are affected. ## Description <!-- Enter your issue details below this comment. One way to structure the description: <short summary of the bug> I tried this code: <code sample that causes the bug> I expected to see this happen: <explanation> Instead, this happened: <explanation> --> `ChildStdout` and `ChildStderr` implement `AsyncRead` by deferring to an internal `PollEvented` of a platform specific type: UNIX uses the standard library's child streams with `EventedFd` and Windows uses `NamedPipe` from `mio_named_pipes`. `PollEvented` naturally cannot implement `prepare_uninitialized_buffer` in any way other than zeroing, due to `std::io::Read::initializer` being unstable, however `ChildStdout` and `ChildStderr` use well-behaved types internally and should be known not to read from the buffer. Thus, the buffer needs not be initialized and `prepare_uninitialized_buffer` should be overwritten for these types to be a no-op. Currently this bug affects me because I cannot call `prepare_uninitialized_buffer` due to the unacceptable zeroing, despite this being part of the contract of `AsyncRead` when an uninitialized buffer is involved. I just have to assume that the `tokio` implementation will remain well-behaved.
1.0
Several AsyncRead impls unnecessarily zero buffers in prepare_uninitialized_buffer - <!-- Thank you for reporting an issue. Please fill in as much of the template below as you're able. --> ## Version <!-- List the versions of all `tokio` crates you are using. The easiest way to get this information is using `cargo-tree`. `cargo install cargo-tree` (see install here: https://github.com/sfackler/cargo-tree) Then: `cargo tree | grep tokio` --> `tokio` 0.2.13 ## Platform <!--- Output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) --> This is an issue for all platforms. ## Subcrates <!-- If known, please specify the affected Tokio sub crates. Otherwise, delete this section. --> I haven't reviewed the full source, but I've reviewed `process` and can confirm that the `ChildStdout` and `ChildStderr` types are affected. ## Description <!-- Enter your issue details below this comment. One way to structure the description: <short summary of the bug> I tried this code: <code sample that causes the bug> I expected to see this happen: <explanation> Instead, this happened: <explanation> --> `ChildStdout` and `ChildStderr` implement `AsyncRead` by deferring to an internal `PollEvented` of a platform specific type: UNIX uses the standard library's child streams with `EventedFd` and Windows uses `NamedPipe` from `mio_named_pipes`. `PollEvented` naturally cannot implement `prepare_uninitialized_buffer` in any way other than zeroing, due to `std::io::Read::initializer` being unstable, however `ChildStdout` and `ChildStderr` use well-behaved types internally and should be known not to read from the buffer. Thus, the buffer needs not be initialized and `prepare_uninitialized_buffer` should be overwritten for these types to be a no-op. Currently this bug affects me because I cannot call `prepare_uninitialized_buffer` due to the unacceptable zeroing, despite this being part of the contract of `AsyncRead` when an uninitialized buffer is involved. I just have to assume that the `tokio` implementation will remain well-behaved.
process
several asyncread impls unnecessarily zero buffers in prepare uninitialized buffer thank you for reporting an issue please fill in as much of the template below as you re able version list the versions of all tokio crates you are using the easiest way to get this information is using cargo tree cargo install cargo tree see install here then cargo tree grep tokio tokio platform output of uname a unix or version and or bit windows this is an issue for all platforms subcrates if known please specify the affected tokio sub crates otherwise delete this section i haven t reviewed the full source but i ve reviewed process and can confirm that the childstdout and childstderr types are affected description enter your issue details below this comment one way to structure the description i tried this code i expected to see this happen instead this happened childstdout and childstderr implement asyncread by deferring to an internal pollevented of a platform specific type unix uses the standard library s child streams with eventedfd and windows uses namedpipe from mio named pipes pollevented naturally cannot implement prepare uninitialized buffer in any way other than zeroing due to std io read initializer being unstable however childstdout and childstderr use well behaved types internally and should be known not to read from the buffer thus the buffer needs not be initialized and prepare uninitialized buffer should be overwritten for these types to be a no op currently this bug affects me because i cannot call prepare uninitialized buffer due to the unacceptable zeroing despite this being part of the contract of asyncread when an uninitialized buffer is involved i just have to assume that the tokio implementation will remain well behaved
1
113,260
17,116,224,415
IssuesEvent
2021-07-11 12:10:06
theHinneh/ha
https://api.github.com/repos/theHinneh/ha
closed
CVE-2020-8203 (High) detected in lodash-4.17.15.tgz, lodash-3.10.1.tgz
security vulnerability
## CVE-2020-8203 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.17.15.tgz</b>, <b>lodash-3.10.1.tgz</b></p></summary> <p> <details><summary><b>lodash-4.17.15.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: ha/backend/package.json</p> <p>Path to vulnerable library: ha/backend/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - mosca-2.8.3.tgz (Root Library) - amqp-0.2.7.tgz - :x: **lodash-4.17.15.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: ha/backend/package.json</p> <p>Path to vulnerable library: ha/backend/node_modules/ioredis/node_modules/lodash/package.json,ha/backend/node_modules/kafka-node/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - mosca-2.8.3.tgz (Root Library) - ioredis-1.15.1.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/theHinneh/ha/commit/b67d33dd9df9e05b70466e310843976220230240">b67d33dd9df9e05b70466e310843976220230240</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-10-21</p> <p>Fix Resolution: lodash - 4.17.19</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8203 (High) detected in lodash-4.17.15.tgz, lodash-3.10.1.tgz - ## CVE-2020-8203 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-4.17.15.tgz</b>, <b>lodash-3.10.1.tgz</b></p></summary> <p> <details><summary><b>lodash-4.17.15.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: ha/backend/package.json</p> <p>Path to vulnerable library: ha/backend/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - mosca-2.8.3.tgz (Root Library) - amqp-0.2.7.tgz - :x: **lodash-4.17.15.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: ha/backend/package.json</p> <p>Path to vulnerable library: ha/backend/node_modules/ioredis/node_modules/lodash/package.json,ha/backend/node_modules/kafka-node/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - mosca-2.8.3.tgz (Root Library) - ioredis-1.15.1.tgz - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/theHinneh/ha/commit/b67d33dd9df9e05b70466e310843976220230240">b67d33dd9df9e05b70466e310843976220230240</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-10-21</p> <p>Fix Resolution: lodash - 4.17.19</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in lodash tgz lodash tgz cve high severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash modular utilities library home page a href path to dependency file ha backend package json path to vulnerable library ha backend node modules lodash package json dependency hierarchy mosca tgz root library amqp tgz x lodash tgz vulnerable library lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file ha backend package json path to vulnerable library ha backend node modules ioredis node modules lodash package json ha backend node modules kafka node node modules lodash package json dependency hierarchy mosca tgz root library ioredis tgz x lodash tgz vulnerable library found in head commit a href found in base branch main vulnerability details prototype pollution attack when using zipobjectdeep in lodash before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
0
34,399
7,450,985,989
IssuesEvent
2018-03-29 00:09:07
kerdokullamae/test_koik_issued
https://api.github.com/repos/kerdokullamae/test_koik_issued
closed
Täpsema otsingu tulemus
C: AIS P: highest R: fixed T: defect
**Reported by jelenag on 18 Feb 2013 09:34 UTC** Täpsem otsing -> määran lähteandmeteks midagi Isiku/Organisatsiooni lahtrisse - > tulemus tuleb kõikide väärtuste puhul üks ja sama, ja tundub, et ole seotud valitud Isikuga. Näide, Isik = Allohvitseride Kool Tulemus -> [= Akadeemiline keemia selts Tulemus -> [http://dev.raju.teepub/et/description_unit/list/?personId=200100007054](http://dev.raju.teepub/et/description_unit/list/?personId=122000000107] Isik)
1.0
Täpsema otsingu tulemus - **Reported by jelenag on 18 Feb 2013 09:34 UTC** Täpsem otsing -> määran lähteandmeteks midagi Isiku/Organisatsiooni lahtrisse - > tulemus tuleb kõikide väärtuste puhul üks ja sama, ja tundub, et ole seotud valitud Isikuga. Näide, Isik = Allohvitseride Kool Tulemus -> [= Akadeemiline keemia selts Tulemus -> [http://dev.raju.teepub/et/description_unit/list/?personId=200100007054](http://dev.raju.teepub/et/description_unit/list/?personId=122000000107] Isik)
non_process
täpsema otsingu tulemus reported by jelenag on feb utc täpsem otsing määran lähteandmeteks midagi isiku organisatsiooni lahtrisse tulemus tuleb kõikide väärtuste puhul üks ja sama ja tundub et ole seotud valitud isikuga näide isik allohvitseride kool tulemus akadeemiline keemia selts tulemus isik
0
122,386
12,150,410,057
IssuesEvent
2020-04-24 17:55:43
FraunhoferIOSB/FROST-Server
https://api.github.com/repos/FraunhoferIOSB/FROST-Server
closed
Dead link to Docker Documentation
documentation
Hi, Just to let you know that the link to "use Docker" page of the documentation runs to a 404 page. Correct link seems to be https://fraunhoferiosb.github.io/FROST-Server/deployment/docker.html The "deployement" part is missing. Best regards
1.0
Dead link to Docker Documentation - Hi, Just to let you know that the link to "use Docker" page of the documentation runs to a 404 page. Correct link seems to be https://fraunhoferiosb.github.io/FROST-Server/deployment/docker.html The "deployement" part is missing. Best regards
non_process
dead link to docker documentation hi just to let you know that the link to use docker page of the documentation runs to a page correct link seems to be the deployement part is missing best regards
0
74,120
9,754,250,892
IssuesEvent
2019-06-04 11:07:42
verdaccio/website
https://api.github.com/repos/verdaccio/website
closed
Webui wrong scope example in Configuration table
documentation
The webui scope example = `\@myscope` used to work pre v4.0. Specify it this way in >= v4.0 _config.yaml_ will fail the JSON.parse on the client in _index.html_ as illegal escape character. Sample should read: **`"@myscope"`** > Note: the sample in the black-box (full _web:_ example) is correct but not in the **Configuration table**. > Issue discover when migrating from v3 to v4 as behaviour changed.
1.0
Webui wrong scope example in Configuration table - The webui scope example = `\@myscope` used to work pre v4.0. Specify it this way in >= v4.0 _config.yaml_ will fail the JSON.parse on the client in _index.html_ as illegal escape character. Sample should read: **`"@myscope"`** > Note: the sample in the black-box (full _web:_ example) is correct but not in the **Configuration table**. > Issue discover when migrating from v3 to v4 as behaviour changed.
non_process
webui wrong scope example in configuration table the webui scope example myscope used to work pre specify it this way in config yaml will fail the json parse on the client in index html as illegal escape character sample should read myscope note the sample in the black box full web example is correct but not in the configuration table issue discover when migrating from to as behaviour changed
0
33,736
7,214,192,631
IssuesEvent
2018-02-08 00:54:56
jccastillo0007/eFacturaT
https://api.github.com/repos/jccastillo0007/eFacturaT
opened
recibo electrónico de pago (REP) - en pago en efectivo, no debe enviar los datos de cuenta bancaria
bug defect
Marca error. Aquí hay 2 formas de hacerlo: a) la correcta, si eligen forma de pago Efectivo al momento de capturar el REP, se deshabilita la captura de las cuentas y todo lo relativo a ello. b) la chacarrona... adicional a ello, si capturan como forma de pago efectivo, no se deben enviar los números de cuenta al XML, aún cuando hayan capturado las cuentas (eso es precisamente lo chaca).
1.0
recibo electrónico de pago (REP) - en pago en efectivo, no debe enviar los datos de cuenta bancaria - Marca error. Aquí hay 2 formas de hacerlo: a) la correcta, si eligen forma de pago Efectivo al momento de capturar el REP, se deshabilita la captura de las cuentas y todo lo relativo a ello. b) la chacarrona... adicional a ello, si capturan como forma de pago efectivo, no se deben enviar los números de cuenta al XML, aún cuando hayan capturado las cuentas (eso es precisamente lo chaca).
non_process
recibo electrónico de pago rep en pago en efectivo no debe enviar los datos de cuenta bancaria marca error aquí hay formas de hacerlo a la correcta si eligen forma de pago efectivo al momento de capturar el rep se deshabilita la captura de las cuentas y todo lo relativo a ello b la chacarrona adicional a ello si capturan como forma de pago efectivo no se deben enviar los números de cuenta al xml aún cuando hayan capturado las cuentas eso es precisamente lo chaca
0
362,580
10,729,530,855
IssuesEvent
2019-10-28 15:45:41
steve8x8/geotoad
https://api.github.com/repos/steve8x8/geotoad
closed
GeoToad (3.29.1 and before) randomly reuses text for additional waypoints
Priority-Medium bug fixed
Sometimes, GeoToad does not extract the proper "Note" text from the Additional Waypoints table, and uses the one of the previous entry instead - this can happen multiple times in a row for the same cache (example: GC6GHP4; instructions for some virtual waypoint tasks are missing from the GPX output - fortunately they are repeated in the long description). Reloading the cache details within c:geo (of course) recitifies the problem, but (of course) that completely misses the point of running GeoToad before.
1.0
GeoToad (3.29.1 and before) randomly reuses text for additional waypoints - Sometimes, GeoToad does not extract the proper "Note" text from the Additional Waypoints table, and uses the one of the previous entry instead - this can happen multiple times in a row for the same cache (example: GC6GHP4; instructions for some virtual waypoint tasks are missing from the GPX output - fortunately they are repeated in the long description). Reloading the cache details within c:geo (of course) recitifies the problem, but (of course) that completely misses the point of running GeoToad before.
non_process
geotoad and before randomly reuses text for additional waypoints sometimes geotoad does not extract the proper note text from the additional waypoints table and uses the one of the previous entry instead this can happen multiple times in a row for the same cache example instructions for some virtual waypoint tasks are missing from the gpx output fortunately they are repeated in the long description reloading the cache details within c geo of course recitifies the problem but of course that completely misses the point of running geotoad before
0
14,046
16,850,624,492
IssuesEvent
2021-06-20 12:37:14
rladstaetter/LogoRRR
https://api.github.com/repos/rladstaetter/LogoRRR
closed
Move to gluonfx-maven-plugin
release process
client-maven-plugin was renamed to gluonfx-maven-plugin. The necessary changes should be applied to the project.
1.0
Move to gluonfx-maven-plugin - client-maven-plugin was renamed to gluonfx-maven-plugin. The necessary changes should be applied to the project.
process
move to gluonfx maven plugin client maven plugin was renamed to gluonfx maven plugin the necessary changes should be applied to the project
1
255,659
21,942,722,079
IssuesEvent
2022-05-23 19:57:48
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
closed
Incorrect Pending message for uphold
bug feature/rewards QA/Yes QA/Test-Plan-Specified closed/not-actionable OS/Desktop
<!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Incorrect Pending message for uphold ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Follow the steps from https://github.com/brave/brave-browser/issues/22131#issue-1194953525 ## Actual result: <!--Please add screenshots if needed--> ![image](https://user-images.githubusercontent.com/38657976/162042065-01bbb90e-045f-40c6-a0d0-4b4ad4bc8ed3.png) ## Expected result: ![image](https://user-images.githubusercontent.com/38657976/162042132-ef0edded-f421-450a-872b-8cef37b980bd.png) ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easy ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? Yes - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/qa-team @Miyayes @gpestana @szilardszaloki @LaurenWags Please refer slack thread for more info https://bravesoftware.slack.com/archives/C0NPFB6H5/p1649265976224599?thread_ts=1648750147.596179&cid=C0NPFB6H5
1.0
Incorrect Pending message for uphold - <!-- Have you searched for similar issues? Before submitting this issue, please check the open issues and add a note before logging a new issue. PLEASE USE THE TEMPLATE BELOW TO PROVIDE INFORMATION ABOUT THE ISSUE. INSUFFICIENT INFO WILL GET THE ISSUE CLOSED. IT WILL ONLY BE REOPENED AFTER SUFFICIENT INFO IS PROVIDED--> ## Description <!--Provide a brief description of the issue--> Incorrect Pending message for uphold ## Steps to Reproduce <!--Please add a series of steps to reproduce the issue--> 1. Follow the steps from https://github.com/brave/brave-browser/issues/22131#issue-1194953525 ## Actual result: <!--Please add screenshots if needed--> ![image](https://user-images.githubusercontent.com/38657976/162042065-01bbb90e-045f-40c6-a0d0-4b4ad4bc8ed3.png) ## Expected result: ![image](https://user-images.githubusercontent.com/38657976/162042132-ef0edded-f421-450a-872b-8cef37b980bd.png) ## Reproduces how often: <!--[Easily reproduced/Intermittent issue/No steps to reproduce]--> Easy ## Brave version (brave://version info) <!--For installed build, please copy Brave, Revision and OS from brave://version and paste here. If building from source please mention it along with brave://version details--> ## Version/Channel Information: <!--Does this issue happen on any other channels? Or is it specific to a certain channel?--> - Can you reproduce this issue with the current release? Yes - Can you reproduce this issue with the beta channel? Yes - Can you reproduce this issue with the nightly channel? Yes ## Other Additional Information: - Does the issue resolve itself when disabling Brave Shields? NA - Does the issue resolve itself when disabling Brave Rewards? NA - Is the issue reproducible on the latest version of Chrome? NA ## Miscellaneous Information: <!--Any additional information, related issues, extra QA steps, configuration or data that might be necessary to reproduce the issue--> cc: @brave/qa-team @Miyayes @gpestana @szilardszaloki @LaurenWags Please refer slack thread for more info https://bravesoftware.slack.com/archives/C0NPFB6H5/p1649265976224599?thread_ts=1648750147.596179&cid=C0NPFB6H5
non_process
incorrect pending message for uphold have you searched for similar issues before submitting this issue please check the open issues and add a note before logging a new issue please use the template below to provide information about the issue insufficient info will get the issue closed it will only be reopened after sufficient info is provided description incorrect pending message for uphold steps to reproduce follow the steps from actual result expected result reproduces how often easy brave version brave version info version channel information can you reproduce this issue with the current release yes can you reproduce this issue with the beta channel yes can you reproduce this issue with the nightly channel yes other additional information does the issue resolve itself when disabling brave shields na does the issue resolve itself when disabling brave rewards na is the issue reproducible on the latest version of chrome na miscellaneous information cc brave qa team miyayes gpestana szilardszaloki laurenwags please refer slack thread for more info
0
7,133
10,278,469,532
IssuesEvent
2019-08-25 14:41:33
nextmoov/nextmoov
https://api.github.com/repos/nextmoov/nextmoov
closed
DX on Sterroids
#Dev Tools & Processes
nextmoov on Sterroids is a huge internal project @ nextmoov, having as a target to become an even more efficient company, despite scaling heavily. This project touches every part of nextmoov — the way we handle accounting, PM, day-to-day @ HQ and remotely, responsabilities inside the (core) team,... **DX on Sterroids is the dev-centric scope of this project.** This Issue is meant to give a high-level overview of all stuff to be done related to this goal, as well as a place of genera discussion about it. ## DX on Sterroids... ### 1. Prerequisite : Changing our toolset nextmoov/nextmoov#3 ### 2. Every step of the Dev Flow — Design Flow nextmoov/nextmoov#10 — Time tracking nextmoov/nextmoov#9 — Stronger Convention & Getting Started Templates nextmoov/nextmoov#1 — CI Deployment nextmoov/nextmoov#7 — PR Flow — Monitoring & Reporting ### 3. Every type of project — Frontend — RN Mobile Apps — Backend Node Projects
1.0
DX on Sterroids - nextmoov on Sterroids is a huge internal project @ nextmoov, having as a target to become an even more efficient company, despite scaling heavily. This project touches every part of nextmoov — the way we handle accounting, PM, day-to-day @ HQ and remotely, responsabilities inside the (core) team,... **DX on Sterroids is the dev-centric scope of this project.** This Issue is meant to give a high-level overview of all stuff to be done related to this goal, as well as a place of genera discussion about it. ## DX on Sterroids... ### 1. Prerequisite : Changing our toolset nextmoov/nextmoov#3 ### 2. Every step of the Dev Flow — Design Flow nextmoov/nextmoov#10 — Time tracking nextmoov/nextmoov#9 — Stronger Convention & Getting Started Templates nextmoov/nextmoov#1 — CI Deployment nextmoov/nextmoov#7 — PR Flow — Monitoring & Reporting ### 3. Every type of project — Frontend — RN Mobile Apps — Backend Node Projects
process
dx on sterroids nextmoov on sterroids is a huge internal project nextmoov having as a target to become an even more efficient company despite scaling heavily this project touches every part of nextmoov — the way we handle accounting pm day to day hq and remotely responsabilities inside the core team dx on sterroids is the dev centric scope of this project this issue is meant to give a high level overview of all stuff to be done related to this goal as well as a place of genera discussion about it dx on sterroids prerequisite changing our toolset nextmoov nextmoov every step of the dev flow — design flow nextmoov nextmoov — time tracking nextmoov nextmoov — stronger convention getting started templates nextmoov nextmoov — ci deployment nextmoov nextmoov — pr flow — monitoring reporting every type of project — frontend — rn mobile apps — backend node projects
1
699,151
24,006,628,923
IssuesEvent
2022-09-14 15:15:24
COPRS/rs-issues
https://api.github.com/repos/COPRS/rs-issues
closed
[BUG] [PRO] rs-addon s1-l0aiop: tasktable drops all RAW sessions files
bug WERUM dev ivv pro CCB priority:minor
<!-- Note: Please search to see if an issue already exists for the bug you encountered. Note: A closed bug can be reopened and affected to a new version of the software. --> **Environment:** <!-- - Delivery tag: release/0.1.0 - Platform: IVV Orange Cloud - Configuration: --> - Delivery tag: 1.4.0-rc1 - Platform: IVV Orange Cloud **Test:** <!-- - Name: TST_INFRA_DEP_orange - Traçability (requirements): NA --> - Name: TST_PRO_WF_S1L0 **Current Behavior:** <!-- A concise description of what you're experiencing. --> When a file has been ingested and sent to the s1-l0aiop preparation worker, the preparation worker drops it. **Expected Behavior:** <!-- A concise description of what you expected to happen. --> The file should be sent to an execution worker. **Test execution artefacts (i.e. logs, screenshots…)** Tip: You can attach images or log files by dragging & dropping, selecting or pasting them. ``` {"header":{"type":"REPORT","timestamp":"2022-08-10T10:34:56.469000Z","level":"INFO","mission":"S1","workflow":"NOMINAL","rs_chain_name":"S1-L0AIOP","rs_chain_version":"1.4.0-rc1"},"message":{"content":"Check if any jobs can be finalized for the IPF"},"task":{"uid":"b88ab97e-5418-46d5-9316-5da9a2a233f7","name":"PreparationWorkerService","event":"BEGIN","input":{"filename_string":"DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"follows_from_task":"5fed5147-5447-4256-923d-edff7d2cf020"}} {"header":{"type":"REPORT","timestamp":"2022-08-10T10:34:56.481000Z","level":"INFO","mission":"S1","workflow":"NOMINAL","rs_chain_name":"S1-L0AIOP","rs_chain_version":"1.4.0-rc1"},"message":{"content":"Received CatalogEvent for S1A_20200120185900030888/DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"task":{"uid":"b88ab97e-5418-46d5-9316-5da9a2a233f7","name":"PreparationWorkerService","event":"BEGIN","input":{"filename_string":"DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"follows_from_task":"5fed5147-5447-4256-923d-edff7d2cf020"}} {"header":{"type":"REPORT","timestamp":"2022-08-10T10:34:56.568000Z","level":"INFO","mission":"S1","workflow":"NOMINAL","rs_chain_name":"S1-L0AIOP","rs_chain_version":"1.4.0-rc1"},"message":{"content":"Product S1A_20200120185900030888/DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw is not intersecting EW slice mask, skipping"},"task":{"uid":"b88ab97e-5418-46d5-9316-5da9a2a233f7","name":"PreparationWorkerService","event":"END","status":"OK","output":{},"input":{"filename_string":"DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"quality":{},"error_code":0,"duration_in_seconds":0.085999,"missing_output":[]}} {"header":{"type":"LOG","timestamp":"2022-08-10T10:34:56.574238Z","level":"INFO","line":84,"file":"TaskTableMapperService.java","thread":"KafkaConsumerDestination{consumerDestinationName='s1-l0aiop-part1.message-filter', partitions=1, dlqName='error-warning'}.container-0-C-1"},"message":{"content":"Dispatching product S1A_20200120185900030888/DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"custom":{"logger_string":"esa.s1pdgs.cpoc.preparation.worker.service.TaskTableMapperService"}} {"header":{"type":"REPORT","timestamp":"2022-08-10T10:34:57.183000Z","level":"INFO","mission":"S1","workflow":"NOMINAL","rs_chain_name":"S1-L0AIOP","rs_chain_version":"1.4.0-rc1"},"message":{"content":"End preparation of new execution jobs"},"task":{"uid":"b88ab97e-5418-46d5-9316-5da9a2a233f7","name":"PreparationWorkerService","event":"END","status":"OK","input":{},"quality":{},"error_code":0,"duration_in_seconds":0.0,"missing_output":[]}} ``` **Bug Generic Definition of Ready (DoR)** - [X] The affect version in which the bug has been found is mentioned - [X] The context and environment of the bug is detailed - [X] The description of the bug is clear and unambiguous - [ ] The procedure (steps) to reproduce the bug is clearly detailed - [ ] The failed tests is linked to the bug : failed result % expected result - [ ] The tested User Story / features is linked to the bug - [X] Logs are attached if available - [ ] A data set attached if available - [X] Category label is link to the bug <!-- infra, mon, pro, perfo, hmi, secu --> **Bug Generic Definition of Done (DoD)** - [ ] the modification implemented (the solution to fix the bug) is described in the bug. - [ ] Unit tests & Continuous integration performed - Test results available - Structural Test coverage reported by SONAR - [ ] Code committed in GIT with right tag or Analysis/Trade Off documentation up-to-date in reference-system-documentation repository - [ ] Code is compliant with coding rules (SONAR Report as evidence) - [ ] Acceptance criteria of the related User story are checked and Passed
1.0
[BUG] [PRO] rs-addon s1-l0aiop: tasktable drops all RAW sessions files - <!-- Note: Please search to see if an issue already exists for the bug you encountered. Note: A closed bug can be reopened and affected to a new version of the software. --> **Environment:** <!-- - Delivery tag: release/0.1.0 - Platform: IVV Orange Cloud - Configuration: --> - Delivery tag: 1.4.0-rc1 - Platform: IVV Orange Cloud **Test:** <!-- - Name: TST_INFRA_DEP_orange - Traçability (requirements): NA --> - Name: TST_PRO_WF_S1L0 **Current Behavior:** <!-- A concise description of what you're experiencing. --> When a file has been ingested and sent to the s1-l0aiop preparation worker, the preparation worker drops it. **Expected Behavior:** <!-- A concise description of what you expected to happen. --> The file should be sent to an execution worker. **Test execution artefacts (i.e. logs, screenshots…)** Tip: You can attach images or log files by dragging & dropping, selecting or pasting them. ``` {"header":{"type":"REPORT","timestamp":"2022-08-10T10:34:56.469000Z","level":"INFO","mission":"S1","workflow":"NOMINAL","rs_chain_name":"S1-L0AIOP","rs_chain_version":"1.4.0-rc1"},"message":{"content":"Check if any jobs can be finalized for the IPF"},"task":{"uid":"b88ab97e-5418-46d5-9316-5da9a2a233f7","name":"PreparationWorkerService","event":"BEGIN","input":{"filename_string":"DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"follows_from_task":"5fed5147-5447-4256-923d-edff7d2cf020"}} {"header":{"type":"REPORT","timestamp":"2022-08-10T10:34:56.481000Z","level":"INFO","mission":"S1","workflow":"NOMINAL","rs_chain_name":"S1-L0AIOP","rs_chain_version":"1.4.0-rc1"},"message":{"content":"Received CatalogEvent for S1A_20200120185900030888/DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"task":{"uid":"b88ab97e-5418-46d5-9316-5da9a2a233f7","name":"PreparationWorkerService","event":"BEGIN","input":{"filename_string":"DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"follows_from_task":"5fed5147-5447-4256-923d-edff7d2cf020"}} {"header":{"type":"REPORT","timestamp":"2022-08-10T10:34:56.568000Z","level":"INFO","mission":"S1","workflow":"NOMINAL","rs_chain_name":"S1-L0AIOP","rs_chain_version":"1.4.0-rc1"},"message":{"content":"Product S1A_20200120185900030888/DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw is not intersecting EW slice mask, skipping"},"task":{"uid":"b88ab97e-5418-46d5-9316-5da9a2a233f7","name":"PreparationWorkerService","event":"END","status":"OK","output":{},"input":{"filename_string":"DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"quality":{},"error_code":0,"duration_in_seconds":0.085999,"missing_output":[]}} {"header":{"type":"LOG","timestamp":"2022-08-10T10:34:56.574238Z","level":"INFO","line":84,"file":"TaskTableMapperService.java","thread":"KafkaConsumerDestination{consumerDestinationName='s1-l0aiop-part1.message-filter', partitions=1, dlqName='error-warning'}.container-0-C-1"},"message":{"content":"Dispatching product S1A_20200120185900030888/DCS_01_S1A_20200120185900030888_ch1_DSDB_00035.raw"},"custom":{"logger_string":"esa.s1pdgs.cpoc.preparation.worker.service.TaskTableMapperService"}} {"header":{"type":"REPORT","timestamp":"2022-08-10T10:34:57.183000Z","level":"INFO","mission":"S1","workflow":"NOMINAL","rs_chain_name":"S1-L0AIOP","rs_chain_version":"1.4.0-rc1"},"message":{"content":"End preparation of new execution jobs"},"task":{"uid":"b88ab97e-5418-46d5-9316-5da9a2a233f7","name":"PreparationWorkerService","event":"END","status":"OK","input":{},"quality":{},"error_code":0,"duration_in_seconds":0.0,"missing_output":[]}} ``` **Bug Generic Definition of Ready (DoR)** - [X] The affect version in which the bug has been found is mentioned - [X] The context and environment of the bug is detailed - [X] The description of the bug is clear and unambiguous - [ ] The procedure (steps) to reproduce the bug is clearly detailed - [ ] The failed tests is linked to the bug : failed result % expected result - [ ] The tested User Story / features is linked to the bug - [X] Logs are attached if available - [ ] A data set attached if available - [X] Category label is link to the bug <!-- infra, mon, pro, perfo, hmi, secu --> **Bug Generic Definition of Done (DoD)** - [ ] the modification implemented (the solution to fix the bug) is described in the bug. - [ ] Unit tests & Continuous integration performed - Test results available - Structural Test coverage reported by SONAR - [ ] Code committed in GIT with right tag or Analysis/Trade Off documentation up-to-date in reference-system-documentation repository - [ ] Code is compliant with coding rules (SONAR Report as evidence) - [ ] Acceptance criteria of the related User story are checked and Passed
non_process
rs addon tasktable drops all raw sessions files note please search to see if an issue already exists for the bug you encountered note a closed bug can be reopened and affected to a new version of the software environment delivery tag release platform ivv orange cloud configuration delivery tag platform ivv orange cloud test name tst infra dep orange traçability requirements na name tst pro wf current behavior when a file has been ingested and sent to the preparation worker the preparation worker drops it expected behavior the file should be sent to an execution worker test execution artefacts i e logs screenshots… tip you can attach images or log files by dragging dropping selecting or pasting them header type report timestamp level info mission workflow nominal rs chain name rs chain version message content check if any jobs can be finalized for the ipf task uid name preparationworkerservice event begin input filename string dcs dsdb raw follows from task header type report timestamp level info mission workflow nominal rs chain name rs chain version message content received catalogevent for dcs dsdb raw task uid name preparationworkerservice event begin input filename string dcs dsdb raw follows from task header type report timestamp level info mission workflow nominal rs chain name rs chain version message content product dcs dsdb raw is not intersecting ew slice mask skipping task uid name preparationworkerservice event end status ok output input filename string dcs dsdb raw quality error code duration in seconds missing output header type log timestamp level info line file tasktablemapperservice java thread kafkaconsumerdestination consumerdestinationname message filter partitions dlqname error warning container c message content dispatching product dcs dsdb raw custom logger string esa cpoc preparation worker service tasktablemapperservice header type report timestamp level info mission workflow nominal rs chain name rs chain version message content end preparation of new execution jobs task uid name preparationworkerservice event end status ok input quality error code duration in seconds missing output bug generic definition of ready dor the affect version in which the bug has been found is mentioned the context and environment of the bug is detailed the description of the bug is clear and unambiguous the procedure steps to reproduce the bug is clearly detailed the failed tests is linked to the bug failed result expected result the tested user story features is linked to the bug logs are attached if available a data set attached if available category label is link to the bug bug generic definition of done dod the modification implemented the solution to fix the bug is described in the bug unit tests continuous integration performed test results available structural test coverage reported by sonar code committed in git with right tag or analysis trade off documentation up to date in reference system documentation repository code is compliant with coding rules sonar report as evidence acceptance criteria of the related user story are checked and passed
0
21,123
28,091,451,332
IssuesEvent
2023-03-30 13:18:53
googleapis/nodejs-storage
https://api.github.com/repos/googleapis/nodejs-storage
closed
test: update signed url conformance test cases
type: process api: storage
Please update your conformance test v4_signatures.json file to match with https://github.com/googleapis/conformance-tests/blob/main/storage/v1/v4_signatures.json and make sure tests pass. PR was made to the conformance test repo adding a test case: https://github.com/googleapis/conformance-tests/pull/75
1.0
test: update signed url conformance test cases - Please update your conformance test v4_signatures.json file to match with https://github.com/googleapis/conformance-tests/blob/main/storage/v1/v4_signatures.json and make sure tests pass. PR was made to the conformance test repo adding a test case: https://github.com/googleapis/conformance-tests/pull/75
process
test update signed url conformance test cases please update your conformance test signatures json file to match with and make sure tests pass pr was made to the conformance test repo adding a test case
1
42,133
22,310,554,069
IssuesEvent
2022-06-13 16:35:10
tstack/lnav
https://api.github.com/repos/tstack/lnav
closed
Load/index files in fragments
performance
Currently, lnav indexes all available file content on startup so you can start working with the full data set. But, this doesn't work well if the files are large since it can take awhile to index and the user is left looking at a blank screen. So, we should look into loading the files in fragments so that the UI pops up right away and the user has something to look at. Initial thoughts: - Initial scan of a file should find the format and the fragmentation points. Fragmentation should happen at line breaks with a fragment size of a few megs. - The last fragment should be indexed so we have something to display at startup. - The preceding fragments should be indexed one-by-one with pauses to accept user input. We might want to load fragments in a background process at some point. ## <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/10789767-load-index-files-in-fragments?utm_campaign=plugin&utm_content=tracker%2F449456&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F449456&utm_medium=issues&utm_source=github). </bountysource-plugin>
True
Load/index files in fragments - Currently, lnav indexes all available file content on startup so you can start working with the full data set. But, this doesn't work well if the files are large since it can take awhile to index and the user is left looking at a blank screen. So, we should look into loading the files in fragments so that the UI pops up right away and the user has something to look at. Initial thoughts: - Initial scan of a file should find the format and the fragmentation points. Fragmentation should happen at line breaks with a fragment size of a few megs. - The last fragment should be indexed so we have something to display at startup. - The preceding fragments should be indexed one-by-one with pauses to accept user input. We might want to load fragments in a background process at some point. ## <bountysource-plugin> --- Want to back this issue? **[Post a bounty on it!](https://www.bountysource.com/issues/10789767-load-index-files-in-fragments?utm_campaign=plugin&utm_content=tracker%2F449456&utm_medium=issues&utm_source=github)** We accept bounties via [Bountysource](https://www.bountysource.com/?utm_campaign=plugin&utm_content=tracker%2F449456&utm_medium=issues&utm_source=github). </bountysource-plugin>
non_process
load index files in fragments currently lnav indexes all available file content on startup so you can start working with the full data set but this doesn t work well if the files are large since it can take awhile to index and the user is left looking at a blank screen so we should look into loading the files in fragments so that the ui pops up right away and the user has something to look at initial thoughts initial scan of a file should find the format and the fragmentation points fragmentation should happen at line breaks with a fragment size of a few megs the last fragment should be indexed so we have something to display at startup the preceding fragments should be indexed one by one with pauses to accept user input we might want to load fragments in a background process at some point want to back this issue we accept bounties via
0
5,626
8,481,858,711
IssuesEvent
2018-10-25 16:50:59
easy-software-ufal/annotations_repos
https://api.github.com/repos/easy-software-ufal/annotations_repos
opened
dotnet/BenchmarkDotNet Benchmark attributed with "HardwareCounters" throws an exception
ADA C# wrong processing
Issue: `https://github.com/dotnet/BenchmarkDotNet/issues/879` PR: `https://github.com/dotnet/BenchmarkDotNet/pull/878/files`
1.0
dotnet/BenchmarkDotNet Benchmark attributed with "HardwareCounters" throws an exception - Issue: `https://github.com/dotnet/BenchmarkDotNet/issues/879` PR: `https://github.com/dotnet/BenchmarkDotNet/pull/878/files`
process
dotnet benchmarkdotnet benchmark attributed with hardwarecounters throws an exception issue pr
1
221,137
16,995,125,036
IssuesEvent
2021-07-01 04:55:31
faridlesosibirsk/TrainingRepository
https://api.github.com/repos/faridlesosibirsk/TrainingRepository
opened
40.01 Список удаленных веток
documentation
В конце файла написать заголовок и ответ, всё сохранить в файл "40. Удаленные ветки.md".
1.0
40.01 Список удаленных веток - В конце файла написать заголовок и ответ, всё сохранить в файл "40. Удаленные ветки.md".
non_process
список удаленных веток в конце файла написать заголовок и ответ всё сохранить в файл удаленные ветки md
0
126,314
12,289,927,842
IssuesEvent
2020-05-10 00:23:38
scanapi/scanapi
https://api.github.com/repos/scanapi/scanapi
opened
Templates Marketplace
documentation reporter
## Description Add a section to README.md with examples of report templates and asking for people to send their templates. Something like danger's plugins section: https://danger.systems/ruby/ ![image](https://user-images.githubusercontent.com/2728804/81488133-47969c80-923b-11ea-9992-fd508968e3fe.png)
1.0
Templates Marketplace - ## Description Add a section to README.md with examples of report templates and asking for people to send their templates. Something like danger's plugins section: https://danger.systems/ruby/ ![image](https://user-images.githubusercontent.com/2728804/81488133-47969c80-923b-11ea-9992-fd508968e3fe.png)
non_process
templates marketplace description add a section to readme md with examples of report templates and asking for people to send their templates something like danger s plugins section
0
7,572
10,685,455,613
IssuesEvent
2019-10-22 12:43:41
didi/mpx
https://api.github.com/repos/didi/mpx
closed
条件编译下引入第三方库打包出错
processing
**问题描述** 在mpx文件中,使用条件编译引入支付宝的第三方组件库,编译后缺少引入组件的axml文件和acss文件 **复现步骤** 如下图 **期望的表现** 使用条件编译引入对应平台的第三方组件,编译正常 **截图** ![image](https://user-images.githubusercontent.com/26195239/66902501-57e2f200-f033-11e9-8f21-e33fe2f09378.png) ![image](https://user-images.githubusercontent.com/26195239/66902509-5adde280-f033-11e9-9b3a-eac1d1482617.png) ![image](https://user-images.githubusercontent.com/26195239/66902478-4d285d00-f033-11e9-888b-313eca6cb9e5.png) **系统及环境** "@mpxjs/api-proxy": "^2.2.30", "@mpxjs/core": "^2.2.30", "@mpxjs/fetch": "^2.2.30", "mini-antui": "^0.4.34", **附件** 光标选中这里并拖拽demo到此上传到github
1.0
条件编译下引入第三方库打包出错 - **问题描述** 在mpx文件中,使用条件编译引入支付宝的第三方组件库,编译后缺少引入组件的axml文件和acss文件 **复现步骤** 如下图 **期望的表现** 使用条件编译引入对应平台的第三方组件,编译正常 **截图** ![image](https://user-images.githubusercontent.com/26195239/66902501-57e2f200-f033-11e9-8f21-e33fe2f09378.png) ![image](https://user-images.githubusercontent.com/26195239/66902509-5adde280-f033-11e9-9b3a-eac1d1482617.png) ![image](https://user-images.githubusercontent.com/26195239/66902478-4d285d00-f033-11e9-888b-313eca6cb9e5.png) **系统及环境** "@mpxjs/api-proxy": "^2.2.30", "@mpxjs/core": "^2.2.30", "@mpxjs/fetch": "^2.2.30", "mini-antui": "^0.4.34", **附件** 光标选中这里并拖拽demo到此上传到github
process
条件编译下引入第三方库打包出错 问题描述 在mpx文件中,使用条件编译引入支付宝的第三方组件库,编译后缺少引入组件的axml文件和acss文件 复现步骤 如下图 期望的表现 使用条件编译引入对应平台的第三方组件,编译正常 截图 系统及环境 mpxjs api proxy mpxjs core mpxjs fetch mini antui 附件 光标选中这里并拖拽demo到此上传到github
1
6,310
9,311,396,267
IssuesEvent
2019-03-25 21:15:46
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Sporadic failure in System.Diagnostics.Tests.ProcessTests.TestEnableRaiseEvents on Linux
area-System.Diagnostics.Process
Debian.8.Amd64-x64 and Alpine.38.Amd64-x64 today. Previously I see SLES and OpenSUSE. Each of the two tests (null and false) passed on rerun on one configuration and failed on the other, suggesting general flakiness. It looks like historically they are slightly flaky. The "true" version is not. The failure is because the `Exited` delegate was invoked even though `p.EnableRaisingEvents` was set to false or not set at all. ``` Assert.False() Failure Expected: False Actual: True Stack Trace : at System.Diagnostics.Tests.ProcessTests.TestEnableRaiseEvents(Nullable`1 enable) in /__w/1/s/src/System.Diagnostics.Process/tests/ProcessTests.cs:line 121 ``` https://mc.dot.net/#/product/netcore/30/source/official~2Fdotnet~2Fcorefx~2Frefs~2Fheads~2Fmaster/type/test~2Ffunctional~2Fcli~2F/build/20190324.4/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessTests~2FTestEnableRaiseEvents(enable:%20null) https://mc.dot.net/#/product/netcore/30/source/official~2Fdotnet~2Fcorefx~2Frefs~2Fheads~2Fmaster/type/test~2Ffunctional~2Fcli~2F/build/20190324.4/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessTests~2FTestEnableRaiseEvents(enable:%20False) In Process.cs, `EnableRaisingEvents` sets `_watchForExit`. In this case we do not set it (or set to false) => `_watchForExit` is false. The `Exited` delegate is called by `OnExited()` which is called by `RaiseOnExited()` which is called synchronously by `HasExited` (not used here) or synchronously by `WaitForExit` (if `_watchForExit` set) or by `CompletionCallback`. => `CompletionCallback` must be the source of this callback. `CompletionCallback` is set up by `EnsureWatchingForExit()` to be called when the process handle is signaled. This is the only only place it is set up =>we know `EnsureWatchingForExit()` was called. If `_watchForExit` is false the only other effect is to prevent `SetProcessHandle` from calling `EnsureWatchingForExit()`. Other than by `EnableRaisingEvents` this is the only place that `EnsureWatchingForExit()` is called (on Windows). => on Windows, if `EnableRaisingEvents` is not true, `Exited` will only be invoked if you use `HasExited` and this test does not. Incidentally MSDN says >Note that the Exited event is raised even if the value of EnableRaisingEvents is false when the process exits during or before the user performs a HasExited check. That comment seems wrong. On Unix `EnsureWatchingForExit()` is called in one other place, `GetProcessHandle`, and `GetProcessHandle` is always called in `ForkAndExecProcess` if the fork succeeds. => on Unix, whatever the value of `EnableRaisingEvents`, the callback will be called when the `_waitHandle` is signaled 1. Question then is why the test ever passes - apparently most times, the handle does not get signaled.? The test after killing the process does `p.WaitForExit(30 sec)` before checking that `Exited` was called. The simplest explanation is that the process is usually not terminating within that 30 seconds. Seems unlikely? The `ProcessWaitState` class is complex and it's not clear to me what other causes there may be. @tmds 2. Either way it seems that we should modify the Unix implementation so that, like Windows, it does not raise the event ever, if `EnableRaisingEvents` is false, and you do not call `HasExited`. @stephentoub
1.0
Sporadic failure in System.Diagnostics.Tests.ProcessTests.TestEnableRaiseEvents on Linux - Debian.8.Amd64-x64 and Alpine.38.Amd64-x64 today. Previously I see SLES and OpenSUSE. Each of the two tests (null and false) passed on rerun on one configuration and failed on the other, suggesting general flakiness. It looks like historically they are slightly flaky. The "true" version is not. The failure is because the `Exited` delegate was invoked even though `p.EnableRaisingEvents` was set to false or not set at all. ``` Assert.False() Failure Expected: False Actual: True Stack Trace : at System.Diagnostics.Tests.ProcessTests.TestEnableRaiseEvents(Nullable`1 enable) in /__w/1/s/src/System.Diagnostics.Process/tests/ProcessTests.cs:line 121 ``` https://mc.dot.net/#/product/netcore/30/source/official~2Fdotnet~2Fcorefx~2Frefs~2Fheads~2Fmaster/type/test~2Ffunctional~2Fcli~2F/build/20190324.4/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessTests~2FTestEnableRaiseEvents(enable:%20null) https://mc.dot.net/#/product/netcore/30/source/official~2Fdotnet~2Fcorefx~2Frefs~2Fheads~2Fmaster/type/test~2Ffunctional~2Fcli~2F/build/20190324.4/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessTests~2FTestEnableRaiseEvents(enable:%20False) In Process.cs, `EnableRaisingEvents` sets `_watchForExit`. In this case we do not set it (or set to false) => `_watchForExit` is false. The `Exited` delegate is called by `OnExited()` which is called by `RaiseOnExited()` which is called synchronously by `HasExited` (not used here) or synchronously by `WaitForExit` (if `_watchForExit` set) or by `CompletionCallback`. => `CompletionCallback` must be the source of this callback. `CompletionCallback` is set up by `EnsureWatchingForExit()` to be called when the process handle is signaled. This is the only only place it is set up =>we know `EnsureWatchingForExit()` was called. If `_watchForExit` is false the only other effect is to prevent `SetProcessHandle` from calling `EnsureWatchingForExit()`. Other than by `EnableRaisingEvents` this is the only place that `EnsureWatchingForExit()` is called (on Windows). => on Windows, if `EnableRaisingEvents` is not true, `Exited` will only be invoked if you use `HasExited` and this test does not. Incidentally MSDN says >Note that the Exited event is raised even if the value of EnableRaisingEvents is false when the process exits during or before the user performs a HasExited check. That comment seems wrong. On Unix `EnsureWatchingForExit()` is called in one other place, `GetProcessHandle`, and `GetProcessHandle` is always called in `ForkAndExecProcess` if the fork succeeds. => on Unix, whatever the value of `EnableRaisingEvents`, the callback will be called when the `_waitHandle` is signaled 1. Question then is why the test ever passes - apparently most times, the handle does not get signaled.? The test after killing the process does `p.WaitForExit(30 sec)` before checking that `Exited` was called. The simplest explanation is that the process is usually not terminating within that 30 seconds. Seems unlikely? The `ProcessWaitState` class is complex and it's not clear to me what other causes there may be. @tmds 2. Either way it seems that we should modify the Unix implementation so that, like Windows, it does not raise the event ever, if `EnableRaisingEvents` is false, and you do not call `HasExited`. @stephentoub
process
sporadic failure in system diagnostics tests processtests testenableraiseevents on linux debian and alpine today previously i see sles and opensuse each of the two tests null and false passed on rerun on one configuration and failed on the other suggesting general flakiness it looks like historically they are slightly flaky the true version is not the failure is because the exited delegate was invoked even though p enableraisingevents was set to false or not set at all assert false failure expected false actual true stack trace at system diagnostics tests processtests testenableraiseevents nullable enable in w s src system diagnostics process tests processtests cs line in process cs enableraisingevents sets watchforexit in this case we do not set it or set to false watchforexit is false the exited delegate is called by onexited which is called by raiseonexited which is called synchronously by hasexited not used here or synchronously by waitforexit if watchforexit set or by completioncallback completioncallback must be the source of this callback completioncallback is set up by ensurewatchingforexit to be called when the process handle is signaled this is the only only place it is set up we know ensurewatchingforexit was called if watchforexit is false the only other effect is to prevent setprocesshandle from calling ensurewatchingforexit other than by enableraisingevents this is the only place that ensurewatchingforexit is called on windows on windows if enableraisingevents is not true exited will only be invoked if you use hasexited and this test does not incidentally msdn says note that the exited event is raised even if the value of enableraisingevents is false when the process exits during or before the user performs a hasexited check that comment seems wrong on unix ensurewatchingforexit is called in one other place getprocesshandle and getprocesshandle is always called in forkandexecprocess if the fork succeeds on unix whatever the value of enableraisingevents the callback will be called when the waithandle is signaled question then is why the test ever passes apparently most times the handle does not get signaled the test after killing the process does p waitforexit sec before checking that exited was called the simplest explanation is that the process is usually not terminating within that seconds seems unlikely the processwaitstate class is complex and it s not clear to me what other causes there may be tmds either way it seems that we should modify the unix implementation so that like windows it does not raise the event ever if enableraisingevents is false and you do not call hasexited stephentoub
1
9,588
12,539,705,339
IssuesEvent
2020-06-05 09:04:37
GoogleCloudPlatform/dotnet-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/dotnet-docs-samples
closed
Reactivate Sudukumb WebApp tests after fixing Chrome Driver issue.
priority: p1 type: process
We are getting the following error when trying to run Sudokumb WebApp tests which use Selenium. > This version of chromedriver only supports chrome version 77 I've deactivated these tests (#922) until the issue is fixed. I want to check all other possible failures and get CI as green as possible to be able to merge several PRs we have pending. @SurferJeffAtGoogle any pointers as to what this error may be are welcomed.
1.0
Reactivate Sudukumb WebApp tests after fixing Chrome Driver issue. - We are getting the following error when trying to run Sudokumb WebApp tests which use Selenium. > This version of chromedriver only supports chrome version 77 I've deactivated these tests (#922) until the issue is fixed. I want to check all other possible failures and get CI as green as possible to be able to merge several PRs we have pending. @SurferJeffAtGoogle any pointers as to what this error may be are welcomed.
process
reactivate sudukumb webapp tests after fixing chrome driver issue we are getting the following error when trying to run sudokumb webapp tests which use selenium this version of chromedriver only supports chrome version i ve deactivated these tests until the issue is fixed i want to check all other possible failures and get ci as green as possible to be able to merge several prs we have pending surferjeffatgoogle any pointers as to what this error may be are welcomed
1
99,441
30,454,299,379
IssuesEvent
2023-07-16 17:37:48
SigNoz/signoz
https://api.github.com/repos/SigNoz/signoz
closed
Add support for limit on Time series
query-builder customer demand
Currently limit on time series is not supported. Limit on time series limits on number of series after ordering by Aggregation(key).
1.0
Add support for limit on Time series - Currently limit on time series is not supported. Limit on time series limits on number of series after ordering by Aggregation(key).
non_process
add support for limit on time series currently limit on time series is not supported limit on time series limits on number of series after ordering by aggregation key
0
824,755
31,169,367,215
IssuesEvent
2023-08-16 23:02:14
janus-idp/backstage-showcase
https://api.github.com/repos/janus-idp/backstage-showcase
opened
Pygments vulnerable to ReDoS
priority/critical kind/security
### What is the issue? Pygments vulnerable to ReDoS ### Is there a CVE Mitre link? https://quay.io/repository/janus-idp/backstage-showcase/manifest/sha256:97d53948633234ce8111bdd9ba7223da834db83dbf9e845c651839d20ff83515?tab=vulnerabilities ### What is the mitigation if known? Upgrade from 2.14 to 2.15
1.0
Pygments vulnerable to ReDoS - ### What is the issue? Pygments vulnerable to ReDoS ### Is there a CVE Mitre link? https://quay.io/repository/janus-idp/backstage-showcase/manifest/sha256:97d53948633234ce8111bdd9ba7223da834db83dbf9e845c651839d20ff83515?tab=vulnerabilities ### What is the mitigation if known? Upgrade from 2.14 to 2.15
non_process
pygments vulnerable to redos what is the issue pygments vulnerable to redos is there a cve mitre link what is the mitigation if known upgrade from to
0
2,337
5,143,253,307
IssuesEvent
2017-01-12 15:33:33
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process execFile / spawn throw non-descript exception on windows if exe requires elevation
child_process libuv windows
* **Version**: v6.5.0 * **Platform**: Windows 10 (x64) * **Subsystem**: child_process ``` let cp = require('child_process'); let util = require('util'); try { cp.execFileSync('SomeTool.exe') } catch (err) { console.log('err', util.inspect(err)); } ``` outputs ``` err { Error: spawnSync SomeTool.exe UNKNOWN at exports._errnoException (util.js:1026:11) at spawnSync (child_process.js:461:20) at Object.execFileSync (child_process.js:498:13) at repl:2:4 at sigintHandlersWrap (vm.js:22:35) at sigintHandlersWrap (vm.js:96:12) at ContextifyScript.Script.runInThisContext (vm.js:21:12) at REPLServer.defaultEval (repl.js:313:29) at bound (domain.js:280:14) at REPLServer.runBound [as eval] (domain.js:293:12) code: 'UNKNOWN', errno: 'UNKNOWN', ... } ``` spawn generates the same error. Obviously it would be better if this error was returned through the callback but I would also love to have a proper error code to react to the situation. Not sure how execFile and spawn are implemented but it should be possible to report a proper errorcode as CreateProcess generates errorcode 740 if elevation is required.
1.0
child_process execFile / spawn throw non-descript exception on windows if exe requires elevation - * **Version**: v6.5.0 * **Platform**: Windows 10 (x64) * **Subsystem**: child_process ``` let cp = require('child_process'); let util = require('util'); try { cp.execFileSync('SomeTool.exe') } catch (err) { console.log('err', util.inspect(err)); } ``` outputs ``` err { Error: spawnSync SomeTool.exe UNKNOWN at exports._errnoException (util.js:1026:11) at spawnSync (child_process.js:461:20) at Object.execFileSync (child_process.js:498:13) at repl:2:4 at sigintHandlersWrap (vm.js:22:35) at sigintHandlersWrap (vm.js:96:12) at ContextifyScript.Script.runInThisContext (vm.js:21:12) at REPLServer.defaultEval (repl.js:313:29) at bound (domain.js:280:14) at REPLServer.runBound [as eval] (domain.js:293:12) code: 'UNKNOWN', errno: 'UNKNOWN', ... } ``` spawn generates the same error. Obviously it would be better if this error was returned through the callback but I would also love to have a proper error code to react to the situation. Not sure how execFile and spawn are implemented but it should be possible to report a proper errorcode as CreateProcess generates errorcode 740 if elevation is required.
process
child process execfile spawn throw non descript exception on windows if exe requires elevation version platform windows subsystem child process let cp require child process let util require util try cp execfilesync sometool exe catch err console log err util inspect err outputs err error spawnsync sometool exe unknown at exports errnoexception util js at spawnsync child process js at object execfilesync child process js at repl at siginthandlerswrap vm js at siginthandlerswrap vm js at contextifyscript script runinthiscontext vm js at replserver defaulteval repl js at bound domain js at replserver runbound domain js code unknown errno unknown spawn generates the same error obviously it would be better if this error was returned through the callback but i would also love to have a proper error code to react to the situation not sure how execfile and spawn are implemented but it should be possible to report a proper errorcode as createprocess generates errorcode if elevation is required
1
19,094
25,147,995,119
IssuesEvent
2022-11-10 07:40:52
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Fit Cells/Nodes selector is missing in SAGA tools
Processing Bug
### What is the bug or the crash? In QGIS 3.20 Versions with SAGA 7.8.2 implemented I experience that in several Raster algorithms like Features to Raster, Resampling, Multilevel B-Spline-Interpolation the selector "Fit cells" / "Fit nodes" is missing. The result is that "Fit nodes" is selected by default and the calculated raster is shifted by half of the cell size in both dicections even if XMIN/XMAX/YMIN/YMAX are defined correctly. In older Versions with SAGA 2.3.2 this selector was there. Without it a lot of SAGA algorithms are not usable ### Steps to reproduce the issue 1. Go to SAGA (7.8.2) 2. Click on Raster - Rasterizing - Features to Raster (or other raster/grid algorithms described above) 3. Fit cells/nodes selector is missing ### Versions 3.20.2-Odense ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
1.0
Fit Cells/Nodes selector is missing in SAGA tools - ### What is the bug or the crash? In QGIS 3.20 Versions with SAGA 7.8.2 implemented I experience that in several Raster algorithms like Features to Raster, Resampling, Multilevel B-Spline-Interpolation the selector "Fit cells" / "Fit nodes" is missing. The result is that "Fit nodes" is selected by default and the calculated raster is shifted by half of the cell size in both dicections even if XMIN/XMAX/YMIN/YMAX are defined correctly. In older Versions with SAGA 2.3.2 this selector was there. Without it a lot of SAGA algorithms are not usable ### Steps to reproduce the issue 1. Go to SAGA (7.8.2) 2. Click on Raster - Rasterizing - Features to Raster (or other raster/grid algorithms described above) 3. Fit cells/nodes selector is missing ### Versions 3.20.2-Odense ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [ ] I tried with a new QGIS profile ### Additional context _No response_
process
fit cells nodes selector is missing in saga tools what is the bug or the crash in qgis versions with saga implemented i experience that in several raster algorithms like features to raster resampling multilevel b spline interpolation the selector fit cells fit nodes is missing the result is that fit nodes is selected by default and the calculated raster is shifted by half of the cell size in both dicections even if xmin xmax ymin ymax are defined correctly in older versions with saga this selector was there without it a lot of saga algorithms are not usable steps to reproduce the issue go to saga click on raster rasterizing features to raster or other raster grid algorithms described above fit cells nodes selector is missing versions odense supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
16,934
22,283,039,942
IssuesEvent
2022-06-11 06:52:45
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
wrong terminal startup folder when open vscode with path param
needs more info confirmation-pending terminal-process
Issue Type: <b>Bug</b> When open vscode from terminal with a path, e.g. `code repo-x`, the intergrated terminal will start at the original folder instead of from `repo-x`. A typical use case: ```sh # working on one repo, try to open another repo the first time # open integrated terminal repo-a> code ../repo-b ``` When open the integrated terminal in the new window, it still points to `repo-a` VS Code version: Code 1.67.2 (c3511e6c69bb39013c4a4b7b9566ec1ca73fc4d5, 2022-05-17T18:15:52.058Z) OS version: Windows_NT x64 10.0.22000 Restricted Mode: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|AMD Ryzen 7 5800X 8-Core Processor (16 x 3800)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|undefined| |Memory (System)|63.94GB (51.22GB free)| |Process Argv|--crash-reporter-id b492549b-40cf-406b-b510-34194e061545| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (67)</summary> Extension|Author (truncated)|Version ---|---|--- better-comments|aar|3.0.0 codesnap|adp|1.3.4 vscode-css-formatter|aes|1.0.2 vscode-nginx-conf|ahm|0.1.3 vscode-zipfs|arc|3.0.0 solidjs|ber|3.0.7 emojisense|bie|0.9.0 vscode-tailwindcss|bra|0.8.2 better-toml|bun|0.3.2 postcss|css|1.0.9 vscode-markdownlint|Dav|0.47.0 vscode-eslint|dba|2.2.2 eslint-disable-snippets|drK|1.3.0 gitlens|eam|12.0.6 filter-lines|ear|1.0.0 EditorConfig|Edi|0.16.4 vsc-material-theme|Equ|33.4.0 vsc-material-theme-icons|equ|2.2.1 prettier-vscode|esb|9.5.0 vscode-reveal|evi|4.1.3 code-runner|for|0.11.7 shell-format|fox|7.2.2 remotehub|Git|0.30.0 vscode-pull-request-github|Git|0.40.0 rest-client|hum|0.24.6 code-eol|jef|1.0.11 vscode-advanced-open-file|jit|0.3.0 svg|joc|1.4.18 vscode-random|jre|1.11.0 rust-analyzer|mat|0.2.1022 hex-hover-converter|maz|1.2.1 git-graph|mhu|1.30.0 prettify-json|moh|0.0.3 theme-monokai-pro-vscode|mon|1.1.20 vscode-json5|mrm|1.0.0 vscode-scss|mrm|0.10.0 remote-containers|ms-|0.231.6 remote-wsl|ms-|0.66.0 hexeditor|ms-|1.9.6 powershell|ms-|2021.12.0 remote-repositories|ms-|0.4.0 resourcemonitor|mut|1.0.7 indent-rainbow|ode|8.3.1 fix-json|oli|0.1.2 advanced-new-file|pat|1.2.2 emoji|Per|1.0.1 vscode-versionlens|pfl|1.0.9 material-icon-theme|PKi|4.16.0 vscode-xml|red|0.20.0 vscode-yaml|red|1.6.0 vscode-sort-json|ric|1.20.0 LiveServer|rit|5.7.5 vscode-open|san|0.1.0 vscode-scss-formatter|sib|2.4.2 mdx|sil|0.1.0 code-spell-checker|str|2.1.11 svelte-vscode|sve|105.16.0 html-preview-vscode|tht|0.2.5 sort-lines|Tyr|1.9.1 vscode-sort-package-json|uni|1.3.0 vscode-ltex|val|13.1.0 quokka-vscode|Wal|1.0.459 browserslist|web|1.1.0 vscode-mdx-preview|xyc|0.3.3 vscode-ol-syntax|yur|0.1.65 markdown-all-in-one|yzh|3.4.2 material-theme|zhu|3.13.24 (5 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vsreu685:30147344 python383cf:30185419 vspor879:30202332 vspor708:30202333 vspor363:30204092 pythonvspyl392:30443607 pythontb:30283811 pythonptprofiler:30281270 vsdfh931cf:30280410 vshan820:30294714 vstes263:30335439 vscoreces:30445986 pythondataviewer:30285071 vscod805:30301674 pythonvspyt200:30340761 binariesv615:30325510 bridge0708:30335490 bridge0723:30353136 vsaa593:30376534 vsc1dst:30438360 pythonvs932:30410667 wslgetstarted:30449410 pythonvsnew555:30457759 vscscmwlcmt:30465135 cppdebug:30492333 vsclangdf:30486550 ``` </details> <!-- generated by issue reporter -->
1.0
wrong terminal startup folder when open vscode with path param - Issue Type: <b>Bug</b> When open vscode from terminal with a path, e.g. `code repo-x`, the intergrated terminal will start at the original folder instead of from `repo-x`. A typical use case: ```sh # working on one repo, try to open another repo the first time # open integrated terminal repo-a> code ../repo-b ``` When open the integrated terminal in the new window, it still points to `repo-a` VS Code version: Code 1.67.2 (c3511e6c69bb39013c4a4b7b9566ec1ca73fc4d5, 2022-05-17T18:15:52.058Z) OS version: Windows_NT x64 10.0.22000 Restricted Mode: No <details> <summary>System Info</summary> |Item|Value| |---|---| |CPUs|AMD Ryzen 7 5800X 8-Core Processor (16 x 3800)| |GPU Status|2d_canvas: enabled<br>canvas_oop_rasterization: disabled_off<br>direct_rendering_display_compositor: disabled_off_ok<br>gpu_compositing: enabled<br>multiple_raster_threads: enabled_on<br>oop_rasterization: enabled<br>opengl: enabled_on<br>rasterization: enabled<br>raw_draw: disabled_off_ok<br>skia_renderer: enabled_on<br>video_decode: enabled<br>video_encode: enabled<br>vulkan: disabled_off<br>webgl: enabled<br>webgl2: enabled| |Load (avg)|undefined| |Memory (System)|63.94GB (51.22GB free)| |Process Argv|--crash-reporter-id b492549b-40cf-406b-b510-34194e061545| |Screen Reader|no| |VM|0%| </details><details><summary>Extensions (67)</summary> Extension|Author (truncated)|Version ---|---|--- better-comments|aar|3.0.0 codesnap|adp|1.3.4 vscode-css-formatter|aes|1.0.2 vscode-nginx-conf|ahm|0.1.3 vscode-zipfs|arc|3.0.0 solidjs|ber|3.0.7 emojisense|bie|0.9.0 vscode-tailwindcss|bra|0.8.2 better-toml|bun|0.3.2 postcss|css|1.0.9 vscode-markdownlint|Dav|0.47.0 vscode-eslint|dba|2.2.2 eslint-disable-snippets|drK|1.3.0 gitlens|eam|12.0.6 filter-lines|ear|1.0.0 EditorConfig|Edi|0.16.4 vsc-material-theme|Equ|33.4.0 vsc-material-theme-icons|equ|2.2.1 prettier-vscode|esb|9.5.0 vscode-reveal|evi|4.1.3 code-runner|for|0.11.7 shell-format|fox|7.2.2 remotehub|Git|0.30.0 vscode-pull-request-github|Git|0.40.0 rest-client|hum|0.24.6 code-eol|jef|1.0.11 vscode-advanced-open-file|jit|0.3.0 svg|joc|1.4.18 vscode-random|jre|1.11.0 rust-analyzer|mat|0.2.1022 hex-hover-converter|maz|1.2.1 git-graph|mhu|1.30.0 prettify-json|moh|0.0.3 theme-monokai-pro-vscode|mon|1.1.20 vscode-json5|mrm|1.0.0 vscode-scss|mrm|0.10.0 remote-containers|ms-|0.231.6 remote-wsl|ms-|0.66.0 hexeditor|ms-|1.9.6 powershell|ms-|2021.12.0 remote-repositories|ms-|0.4.0 resourcemonitor|mut|1.0.7 indent-rainbow|ode|8.3.1 fix-json|oli|0.1.2 advanced-new-file|pat|1.2.2 emoji|Per|1.0.1 vscode-versionlens|pfl|1.0.9 material-icon-theme|PKi|4.16.0 vscode-xml|red|0.20.0 vscode-yaml|red|1.6.0 vscode-sort-json|ric|1.20.0 LiveServer|rit|5.7.5 vscode-open|san|0.1.0 vscode-scss-formatter|sib|2.4.2 mdx|sil|0.1.0 code-spell-checker|str|2.1.11 svelte-vscode|sve|105.16.0 html-preview-vscode|tht|0.2.5 sort-lines|Tyr|1.9.1 vscode-sort-package-json|uni|1.3.0 vscode-ltex|val|13.1.0 quokka-vscode|Wal|1.0.459 browserslist|web|1.1.0 vscode-mdx-preview|xyc|0.3.3 vscode-ol-syntax|yur|0.1.65 markdown-all-in-one|yzh|3.4.2 material-theme|zhu|3.13.24 (5 theme extensions excluded) </details><details> <summary>A/B Experiments</summary> ``` vsliv368cf:30146710 vsreu685:30147344 python383cf:30185419 vspor879:30202332 vspor708:30202333 vspor363:30204092 pythonvspyl392:30443607 pythontb:30283811 pythonptprofiler:30281270 vsdfh931cf:30280410 vshan820:30294714 vstes263:30335439 vscoreces:30445986 pythondataviewer:30285071 vscod805:30301674 pythonvspyt200:30340761 binariesv615:30325510 bridge0708:30335490 bridge0723:30353136 vsaa593:30376534 vsc1dst:30438360 pythonvs932:30410667 wslgetstarted:30449410 pythonvsnew555:30457759 vscscmwlcmt:30465135 cppdebug:30492333 vsclangdf:30486550 ``` </details> <!-- generated by issue reporter -->
process
wrong terminal startup folder when open vscode with path param issue type bug when open vscode from terminal with a path e g code repo x the intergrated terminal will start at the original folder instead of from repo x a typical use case sh working on one repo try to open another repo the first time open integrated terminal repo a code repo b when open the integrated terminal in the new window it still points to repo a vs code version code os version windows nt restricted mode no system info item value cpus amd ryzen core processor x gpu status canvas enabled canvas oop rasterization disabled off direct rendering display compositor disabled off ok gpu compositing enabled multiple raster threads enabled on oop rasterization enabled opengl enabled on rasterization enabled raw draw disabled off ok skia renderer enabled on video decode enabled video encode enabled vulkan disabled off webgl enabled enabled load avg undefined memory system free process argv crash reporter id screen reader no vm extensions extension author truncated version better comments aar codesnap adp vscode css formatter aes vscode nginx conf ahm vscode zipfs arc solidjs ber emojisense bie vscode tailwindcss bra better toml bun postcss css vscode markdownlint dav vscode eslint dba eslint disable snippets drk gitlens eam filter lines ear editorconfig edi vsc material theme equ vsc material theme icons equ prettier vscode esb vscode reveal evi code runner for shell format fox remotehub git vscode pull request github git rest client hum code eol jef vscode advanced open file jit svg joc vscode random jre rust analyzer mat hex hover converter maz git graph mhu prettify json moh theme monokai pro vscode mon vscode mrm vscode scss mrm remote containers ms remote wsl ms hexeditor ms powershell ms remote repositories ms resourcemonitor mut indent rainbow ode fix json oli advanced new file pat emoji per vscode versionlens pfl material icon theme pki vscode xml red vscode yaml red vscode sort json ric liveserver rit vscode open san vscode scss formatter sib mdx sil code spell checker str svelte vscode sve html preview vscode tht sort lines tyr vscode sort package json uni vscode ltex val quokka vscode wal browserslist web vscode mdx preview xyc vscode ol syntax yur markdown all in one yzh material theme zhu theme extensions excluded a b experiments pythontb pythonptprofiler vscoreces pythondataviewer wslgetstarted vscscmwlcmt cppdebug vsclangdf
1
16,842
22,092,067,895
IssuesEvent
2022-06-01 07:00:05
ethereum/EIPs
https://api.github.com/repos/ethereum/EIPs
closed
Introduce more automation
type: EIP1 (Process)
We have the following pieces of automation currently: 1. automerger bot 2. syntax tests a. eipv (preplaced the old validator not long ago](https://github.com/ethereum/EIPs/pull/2860)) b. spellcheck c. hmlcheck 3. greeter bot 4. stale bot ([added lately](https://github.com/ethereum/EIPs/pull/2949)) The automerger can automatically merge PRs, except those changing statuses or adding new files. It is a Python-based code run on Google AppEngine and triggered by Travis. The syntax tests are executed via Travis and are displayed on github with a red or green "continuous integration" icons. This is sometimes overlooked by authors. The automerger will also not act, if the status is red. The most important check we have is the eipv one: it validates a lot of formatting requirements, but not every single requirement. For outstanding issues [check here](https://github.com/lightclient/eipv/issues). The greeter bot reminds new contributors to read EIP-1, while the stale bot ensures that old issues are closed if there is no activity. --- Now, this looks great, but it could be better. In order to speed up the review process we should make sure that `eipv` checks for a lot more mistakes. The most common mistake is not having all the appropriate sections: the correct names, ordering, indentation, or having non-standard sections. The second biggest problem is that a lot of contributors are not familiar with travis and/or continuous integration. We should introduce a bot (eipv-bot) which creates comments for each mistake. Then mostly editors would be restricted to assigning EIP numbers. Once #2941 is properly decided on we could introduce another bot to mark long untouched drafts as `status=Withdrawn reason="withdrawn due to inactivity"`. It should create a PR, and ideally merge it after some period of inactivity, or let editors merge it. --- I think we should move over to Github Actions, because: - We can define all the rules of execution in this repository in the `.github/workflows` directory - Can place all the bots in a separate repository I propose the following projects to undertake: 1. eipv-bot: github action compatible bot which executes eipv and creates comments for each issue 2. abandon-bot: github action compatible bot which creates a PR marking old drafts as withdrawn 3. automerger-bot: github action compatible bot which replaces the automerger
1.0
Introduce more automation - We have the following pieces of automation currently: 1. automerger bot 2. syntax tests a. eipv (preplaced the old validator not long ago](https://github.com/ethereum/EIPs/pull/2860)) b. spellcheck c. hmlcheck 3. greeter bot 4. stale bot ([added lately](https://github.com/ethereum/EIPs/pull/2949)) The automerger can automatically merge PRs, except those changing statuses or adding new files. It is a Python-based code run on Google AppEngine and triggered by Travis. The syntax tests are executed via Travis and are displayed on github with a red or green "continuous integration" icons. This is sometimes overlooked by authors. The automerger will also not act, if the status is red. The most important check we have is the eipv one: it validates a lot of formatting requirements, but not every single requirement. For outstanding issues [check here](https://github.com/lightclient/eipv/issues). The greeter bot reminds new contributors to read EIP-1, while the stale bot ensures that old issues are closed if there is no activity. --- Now, this looks great, but it could be better. In order to speed up the review process we should make sure that `eipv` checks for a lot more mistakes. The most common mistake is not having all the appropriate sections: the correct names, ordering, indentation, or having non-standard sections. The second biggest problem is that a lot of contributors are not familiar with travis and/or continuous integration. We should introduce a bot (eipv-bot) which creates comments for each mistake. Then mostly editors would be restricted to assigning EIP numbers. Once #2941 is properly decided on we could introduce another bot to mark long untouched drafts as `status=Withdrawn reason="withdrawn due to inactivity"`. It should create a PR, and ideally merge it after some period of inactivity, or let editors merge it. --- I think we should move over to Github Actions, because: - We can define all the rules of execution in this repository in the `.github/workflows` directory - Can place all the bots in a separate repository I propose the following projects to undertake: 1. eipv-bot: github action compatible bot which executes eipv and creates comments for each issue 2. abandon-bot: github action compatible bot which creates a PR marking old drafts as withdrawn 3. automerger-bot: github action compatible bot which replaces the automerger
process
introduce more automation we have the following pieces of automation currently automerger bot syntax tests a eipv preplaced the old validator not long ago b spellcheck c hmlcheck greeter bot stale bot the automerger can automatically merge prs except those changing statuses or adding new files it is a python based code run on google appengine and triggered by travis the syntax tests are executed via travis and are displayed on github with a red or green continuous integration icons this is sometimes overlooked by authors the automerger will also not act if the status is red the most important check we have is the eipv one it validates a lot of formatting requirements but not every single requirement for outstanding issues the greeter bot reminds new contributors to read eip while the stale bot ensures that old issues are closed if there is no activity now this looks great but it could be better in order to speed up the review process we should make sure that eipv checks for a lot more mistakes the most common mistake is not having all the appropriate sections the correct names ordering indentation or having non standard sections the second biggest problem is that a lot of contributors are not familiar with travis and or continuous integration we should introduce a bot eipv bot which creates comments for each mistake then mostly editors would be restricted to assigning eip numbers once is properly decided on we could introduce another bot to mark long untouched drafts as status withdrawn reason withdrawn due to inactivity it should create a pr and ideally merge it after some period of inactivity or let editors merge it i think we should move over to github actions because we can define all the rules of execution in this repository in the github workflows directory can place all the bots in a separate repository i propose the following projects to undertake eipv bot github action compatible bot which executes eipv and creates comments for each issue abandon bot github action compatible bot which creates a pr marking old drafts as withdrawn automerger bot github action compatible bot which replaces the automerger
1
252,608
27,247,532,023
IssuesEvent
2023-02-22 04:12:08
a23au/awe-base-images
https://api.github.com/repos/a23au/awe-base-images
closed
Resolve vulnerability alerts
security
Tracking resolution of the following alerts: ## Vulnerability CVE-2023-23914 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/50 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | LOW | libcurl | 7.87.0-r2 | [CVE-2023-23914](https://avd.aquasec.com/nvd/cve-2023-23914) | HSTS ignored on multiple requests ## Vulnerability CVE-2023-23915 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/51 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | LOW | libcurl | 7.87.0-r2 | [CVE-2023-23915](https://avd.aquasec.com/nvd/cve-2023-23915) | HSTS amnesia with --parallel ## Vulnerability CVE-2022-4203 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/42 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/34 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | MEDIUM | libssl3 | 3.0.8-r0 | CVE-2022-4203 | X.509 Name Constraints Read Buffer Overflow ## Vulnerability CVE-2022-4304 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/43 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/35 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | MEDIUM | libssl3 | 3.0.8-r0 | CVE-2022-4304 | A timing based side channel exists in the OpenSSL RSA Decryption implementation which could be sufficient to recover a plaintext across a network in a Bleichenbacher style attack. To achieve a successful decryption an attacker would have to be able to send a very large number of trial messages for decryption. The vulnerability affects all RSA padding modes: PKCS#1 v1.5, RSA-OEAP and RSASVE. For example, in a TLS connection, RSA is commonly used by a client to send an encrypted pre-master secret to the server. An attacker that had observed a genuine connection between a client and a server could use this flaw to send trial messages to the server and record the time taken to process them. After a sufficiently large number of messages the attacker could recover the pre-master secret used for the original connection and thus be able to decrypt the application data sent over that connection. ## Vulnerability CVE-2023-23916 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/49 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | MEDIUM | libcurl | 7.87.0-r2 | [CVE-2023-23916](https://avd.aquasec.com/nvd/cve-2023-23916) | HTTP multi-header compression denial of service ## Vulnerability CVE-2022-4450 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/44 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/36 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2022-4450 | The function PEM_read_bio_ex() reads a PEM file from a BIO and parses and decodes the "name" (e.g. "CERTIFICATE"), any header data and the payload data. If the function succeeds then the "name_out", "header" and "data" arguments are populated with pointers to buffers containing the relevant decoded data. The caller is responsible for freeing those buffers. It is possible to construct a PEM file that results in 0 bytes of payload data. In this case PEM_read_bio_ex() will return a failure code but will populate the header argument with a pointer to a buffer that has already been freed. If the caller also frees this buffer then a double free will occur. This will most likely lead to a crash. This could be exploited by an attacker who has the ability to supply malicious PEM files for parsing to achieve a denial of service attack. The functions PEM_read_bio() and PEM_read() are simple wrappers around PEM_read_bio_ex() and therefore these functions are also directly affected. These functions are also called indirectly by a number of other OpenSSL functions including PEM_X509_INFO_read_bio_ex() and SSL_CTX_use_serverinfo_file() which are also vulnerable. Some OpenSSL internal uses of these functions are not vulnerable because the caller does not free the header argument if PEM_read_bio_ex() returns a failure code. These locations include the PEM_read_bio_TYPE() functions as well as the decoders introduced in OpenSSL 3.0. The OpenSSL asn1parse command line application is also impacted by this issue. ## Vulnerability CVE-2022-25147 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/32 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/31 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/29 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | CRITICAL | apr-util-ldap | 1.6.3-r0 | CVE-2022-25147 | Integer Overflow or Wraparound vulnerability in apr_base64 functions of Apache Portable Runtime Utility (APR-util) allows an attacker to write beyond bounds of a buffer. This issue affects Apache Portable Runtime Utility (APR-util) 1.6.1 and prior versions. ## Vulnerability CVE-2022-24963 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/28 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | CRITICAL | apr | 1.7.1-r0 | CVE-2022-24963 | Integer Overflow or Wraparound vulnerability in apr_encode functions of Apache Portable Runtime (APR) allows an attacker to write beyond bounds of a buffer. This issue affects Apache Portable Runtime (APR) version 1.7.0. ## Vulnerability CVE-2022-28331 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/30 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | CRITICAL | apr | 1.7.1-r0 | CVE-2022-28331 | On Windows, Apache Portable Runtime 1.7.0 and earlier may write beyond the end of a stack based buffer in apr_socket_sendv(). This is a result of integer overflow. #### Vulnerability CVE-2023-0286 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/41 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/33 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | CRITICAL | libssl3 3.0.8-r0 | CVE-2023-0286 | There is a type confusion vulnerability relating to X.400 address processing inside an X.509 GeneralName. X.400 addresses were parsed as an ASN1_STRING but the public structure definition for GENERAL_NAME incorrectly specified the type of the x400Address field as ASN1_TYPE. This field is subsequently interpreted by the OpenSSL function GENERAL_NAME_cmp as an ASN1_TYPE rather than an ASN1_STRING. When CRL checking is enabled (i.e. the application sets the X509_V_FLAG_CRL_CHECK flag), this vulnerability may allow an attacker to pass arbitrary pointers to a memcmp call, enabling them to read memory contents or enact a denial of service. In most cases, the attack requires the attacker to provide both the certificate chain and CRL, neither of which need to have a valid signature. If the attacker only controls one of these inputs, the other input must already contain an X.400 address as a CRL distribution point, which is uncommon. As such, this vulnerability is most likely to only affect applications which have implemented their own functionality for retrieving CRLs over a network. ## Vulnerability CVE-2023-0401 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/48 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/40 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2023-0401 | A NULL pointer can be dereferenced when signatures are being verified on PKCS7 signed or signedAndEnveloped data. In case the hash algorithm used for the signature is known to the OpenSSL library but the implementation of the hash algorithm is not available the digest initialization will fail. There is a missing check for the return value from the initialization function which later leads to invalid usage of the digest API most likely leading to a crash. The unavailability of an algorithm can be caused by using FIPS enabled configuration of providers or more commonly by not loading the legacy provider. PKCS7 data is processed by the SMIME library calls and also by the time stamp (TS) library calls. The TLS implementation in OpenSSL does not call these functions however third party applications would be affected if they call these functions to verify signatures on untrusted data. ## Vulnerability CVE-2023-0217 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/47 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/39 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2023-0217 | An invalid pointer dereference on read can be triggered when an application tries to check a malformed DSA public key by the EVP_PKEY_public_check() function. This will most likely lead to an application crash. This function can be called on public keys supplied from untrusted sources which could allow an attacker to cause a denial of service attack. The TLS implementation in OpenSSL does not call this function but applications might call the function if there are additional security requirements imposed by standards such as FIPS 140-3. ## Vulnerability CVE-2023-0216 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/46 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/38 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2023-0216 | An invalid pointer dereference on read can be triggered when an application tries to load malformed PKCS7 data with the d2i_PKCS7(), d2i_PKCS7_bio() or d2i_PKCS7_fp() functions. The result of the dereference is an application crash which could lead to a denial of service attack. The TLS implementation in OpenSSL does not call this function however third party applications might call these functions on untrusted data. ## Vulnerability CVE-2023-0215 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/45 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/37 | Severity | Package | Fixed Version| Link | | ----- | ----- | ----- | ----- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2023-0215 | The public API function BIO_new_NDEF is a helper function used for streaming ASN.1 data via a BIO. It is primarily used internally to OpenSSL to support the SMIME, CMS and PKCS7 streaming capabilities, but may also be called directly by end user applications. The function receives a BIO from the caller, prepends a new BIO_f_asn1 filter BIO onto the front of it to form a BIO chain, and then returns the new head of the BIO chain to the caller. Under certain conditions, for example if a CMS recipient public key is invalid, the new filter BIO is freed and the function returns a NULL result indicating a failure. However, in this case, the BIO chain is not properly cleaned up and the BIO passed by the caller still retains internal pointers to the previously freed filter BIO. If the caller then goes on to call BIO_pop() on the BIO then a use-after-free will occur. This will most likely result in a crash. This scenario occurs directly in the internal function B64_write_ASN1() which may cause BIO_new_NDEF() to be called and will subsequently call BIO_pop() on the BIO. This internal function is in turn called by the public API functions PEM_write_bio_ASN1_stream, PEM_write_bio_CMS_stream, PEM_write_bio_PKCS7_stream, SMIME_write_ASN1, SMIME_write_CMS and SMIME_write_PKCS7. Other public API functions that may be impacted by this include i2d_ASN1_bio_stream, BIO_new_CMS, BIO_new_PKCS7, i2d_CMS_bio_stream and i2d_PKCS7_bio_stream. The OpenSSL cms and smime command line applications are similarly affected.
True
Resolve vulnerability alerts - Tracking resolution of the following alerts: ## Vulnerability CVE-2023-23914 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/50 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | LOW | libcurl | 7.87.0-r2 | [CVE-2023-23914](https://avd.aquasec.com/nvd/cve-2023-23914) | HSTS ignored on multiple requests ## Vulnerability CVE-2023-23915 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/51 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | LOW | libcurl | 7.87.0-r2 | [CVE-2023-23915](https://avd.aquasec.com/nvd/cve-2023-23915) | HSTS amnesia with --parallel ## Vulnerability CVE-2022-4203 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/42 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/34 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | MEDIUM | libssl3 | 3.0.8-r0 | CVE-2022-4203 | X.509 Name Constraints Read Buffer Overflow ## Vulnerability CVE-2022-4304 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/43 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/35 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | MEDIUM | libssl3 | 3.0.8-r0 | CVE-2022-4304 | A timing based side channel exists in the OpenSSL RSA Decryption implementation which could be sufficient to recover a plaintext across a network in a Bleichenbacher style attack. To achieve a successful decryption an attacker would have to be able to send a very large number of trial messages for decryption. The vulnerability affects all RSA padding modes: PKCS#1 v1.5, RSA-OEAP and RSASVE. For example, in a TLS connection, RSA is commonly used by a client to send an encrypted pre-master secret to the server. An attacker that had observed a genuine connection between a client and a server could use this flaw to send trial messages to the server and record the time taken to process them. After a sufficiently large number of messages the attacker could recover the pre-master secret used for the original connection and thus be able to decrypt the application data sent over that connection. ## Vulnerability CVE-2023-23916 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/49 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | MEDIUM | libcurl | 7.87.0-r2 | [CVE-2023-23916](https://avd.aquasec.com/nvd/cve-2023-23916) | HTTP multi-header compression denial of service ## Vulnerability CVE-2022-4450 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/44 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/36 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2022-4450 | The function PEM_read_bio_ex() reads a PEM file from a BIO and parses and decodes the "name" (e.g. "CERTIFICATE"), any header data and the payload data. If the function succeeds then the "name_out", "header" and "data" arguments are populated with pointers to buffers containing the relevant decoded data. The caller is responsible for freeing those buffers. It is possible to construct a PEM file that results in 0 bytes of payload data. In this case PEM_read_bio_ex() will return a failure code but will populate the header argument with a pointer to a buffer that has already been freed. If the caller also frees this buffer then a double free will occur. This will most likely lead to a crash. This could be exploited by an attacker who has the ability to supply malicious PEM files for parsing to achieve a denial of service attack. The functions PEM_read_bio() and PEM_read() are simple wrappers around PEM_read_bio_ex() and therefore these functions are also directly affected. These functions are also called indirectly by a number of other OpenSSL functions including PEM_X509_INFO_read_bio_ex() and SSL_CTX_use_serverinfo_file() which are also vulnerable. Some OpenSSL internal uses of these functions are not vulnerable because the caller does not free the header argument if PEM_read_bio_ex() returns a failure code. These locations include the PEM_read_bio_TYPE() functions as well as the decoders introduced in OpenSSL 3.0. The OpenSSL asn1parse command line application is also impacted by this issue. ## Vulnerability CVE-2022-25147 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/32 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/31 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/29 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | CRITICAL | apr-util-ldap | 1.6.3-r0 | CVE-2022-25147 | Integer Overflow or Wraparound vulnerability in apr_base64 functions of Apache Portable Runtime Utility (APR-util) allows an attacker to write beyond bounds of a buffer. This issue affects Apache Portable Runtime Utility (APR-util) 1.6.1 and prior versions. ## Vulnerability CVE-2022-24963 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/28 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | CRITICAL | apr | 1.7.1-r0 | CVE-2022-24963 | Integer Overflow or Wraparound vulnerability in apr_encode functions of Apache Portable Runtime (APR) allows an attacker to write beyond bounds of a buffer. This issue affects Apache Portable Runtime (APR) version 1.7.0. ## Vulnerability CVE-2022-28331 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/30 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | CRITICAL | apr | 1.7.1-r0 | CVE-2022-28331 | On Windows, Apache Portable Runtime 1.7.0 and earlier may write beyond the end of a stack based buffer in apr_socket_sendv(). This is a result of integer overflow. #### Vulnerability CVE-2023-0286 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/41 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/33 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | CRITICAL | libssl3 3.0.8-r0 | CVE-2023-0286 | There is a type confusion vulnerability relating to X.400 address processing inside an X.509 GeneralName. X.400 addresses were parsed as an ASN1_STRING but the public structure definition for GENERAL_NAME incorrectly specified the type of the x400Address field as ASN1_TYPE. This field is subsequently interpreted by the OpenSSL function GENERAL_NAME_cmp as an ASN1_TYPE rather than an ASN1_STRING. When CRL checking is enabled (i.e. the application sets the X509_V_FLAG_CRL_CHECK flag), this vulnerability may allow an attacker to pass arbitrary pointers to a memcmp call, enabling them to read memory contents or enact a denial of service. In most cases, the attack requires the attacker to provide both the certificate chain and CRL, neither of which need to have a valid signature. If the attacker only controls one of these inputs, the other input must already contain an X.400 address as a CRL distribution point, which is uncommon. As such, this vulnerability is most likely to only affect applications which have implemented their own functionality for retrieving CRLs over a network. ## Vulnerability CVE-2023-0401 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/48 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/40 | Severity | Package | Fixed Version | Link | | ------------- | ------------- | ------------- | ------------- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2023-0401 | A NULL pointer can be dereferenced when signatures are being verified on PKCS7 signed or signedAndEnveloped data. In case the hash algorithm used for the signature is known to the OpenSSL library but the implementation of the hash algorithm is not available the digest initialization will fail. There is a missing check for the return value from the initialization function which later leads to invalid usage of the digest API most likely leading to a crash. The unavailability of an algorithm can be caused by using FIPS enabled configuration of providers or more commonly by not loading the legacy provider. PKCS7 data is processed by the SMIME library calls and also by the time stamp (TS) library calls. The TLS implementation in OpenSSL does not call these functions however third party applications would be affected if they call these functions to verify signatures on untrusted data. ## Vulnerability CVE-2023-0217 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/47 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/39 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2023-0217 | An invalid pointer dereference on read can be triggered when an application tries to check a malformed DSA public key by the EVP_PKEY_public_check() function. This will most likely lead to an application crash. This function can be called on public keys supplied from untrusted sources which could allow an attacker to cause a denial of service attack. The TLS implementation in OpenSSL does not call this function but applications might call the function if there are additional security requirements imposed by standards such as FIPS 140-3. ## Vulnerability CVE-2023-0216 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/46 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/38 | Severity | Package | Fixed Version | Link | | ----- | ----- | ----- | ----- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2023-0216 | An invalid pointer dereference on read can be triggered when an application tries to load malformed PKCS7 data with the d2i_PKCS7(), d2i_PKCS7_bio() or d2i_PKCS7_fp() functions. The result of the dereference is an application crash which could lead to a denial of service attack. The TLS implementation in OpenSSL does not call this function however third party applications might call these functions on untrusted data. ## Vulnerability CVE-2023-0215 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/45 - [ ] https://github.com/a23au/awe-base-images/security/code-scanning/37 | Severity | Package | Fixed Version| Link | | ----- | ----- | ----- | ----- | | HIGH | libssl3 | 3.0.8-r0 | CVE-2023-0215 | The public API function BIO_new_NDEF is a helper function used for streaming ASN.1 data via a BIO. It is primarily used internally to OpenSSL to support the SMIME, CMS and PKCS7 streaming capabilities, but may also be called directly by end user applications. The function receives a BIO from the caller, prepends a new BIO_f_asn1 filter BIO onto the front of it to form a BIO chain, and then returns the new head of the BIO chain to the caller. Under certain conditions, for example if a CMS recipient public key is invalid, the new filter BIO is freed and the function returns a NULL result indicating a failure. However, in this case, the BIO chain is not properly cleaned up and the BIO passed by the caller still retains internal pointers to the previously freed filter BIO. If the caller then goes on to call BIO_pop() on the BIO then a use-after-free will occur. This will most likely result in a crash. This scenario occurs directly in the internal function B64_write_ASN1() which may cause BIO_new_NDEF() to be called and will subsequently call BIO_pop() on the BIO. This internal function is in turn called by the public API functions PEM_write_bio_ASN1_stream, PEM_write_bio_CMS_stream, PEM_write_bio_PKCS7_stream, SMIME_write_ASN1, SMIME_write_CMS and SMIME_write_PKCS7. Other public API functions that may be impacted by this include i2d_ASN1_bio_stream, BIO_new_CMS, BIO_new_PKCS7, i2d_CMS_bio_stream and i2d_PKCS7_bio_stream. The OpenSSL cms and smime command line applications are similarly affected.
non_process
resolve vulnerability alerts tracking resolution of the following alerts vulnerability cve severity package fixed version link low libcurl hsts ignored on multiple requests vulnerability cve severity package fixed version link low libcurl hsts amnesia with parallel vulnerability cve severity package fixed version link medium cve x name constraints read buffer overflow vulnerability cve severity package fixed version link medium cve a timing based side channel exists in the openssl rsa decryption implementation which could be sufficient to recover a plaintext across a network in a bleichenbacher style attack to achieve a successful decryption an attacker would have to be able to send a very large number of trial messages for decryption the vulnerability affects all rsa padding modes pkcs rsa oeap and rsasve for example in a tls connection rsa is commonly used by a client to send an encrypted pre master secret to the server an attacker that had observed a genuine connection between a client and a server could use this flaw to send trial messages to the server and record the time taken to process them after a sufficiently large number of messages the attacker could recover the pre master secret used for the original connection and thus be able to decrypt the application data sent over that connection vulnerability cve severity package fixed version link medium libcurl http multi header compression denial of service vulnerability cve severity package fixed version link high cve the function pem read bio ex reads a pem file from a bio and parses and decodes the name e g certificate any header data and the payload data if the function succeeds then the name out header and data arguments are populated with pointers to buffers containing the relevant decoded data the caller is responsible for freeing those buffers it is possible to construct a pem file that results in bytes of payload data in this case pem read bio ex will return a failure code but will populate the header argument with a pointer to a buffer that has already been freed if the caller also frees this buffer then a double free will occur this will most likely lead to a crash this could be exploited by an attacker who has the ability to supply malicious pem files for parsing to achieve a denial of service attack the functions pem read bio and pem read are simple wrappers around pem read bio ex and therefore these functions are also directly affected these functions are also called indirectly by a number of other openssl functions including pem info read bio ex and ssl ctx use serverinfo file which are also vulnerable some openssl internal uses of these functions are not vulnerable because the caller does not free the header argument if pem read bio ex returns a failure code these locations include the pem read bio type functions as well as the decoders introduced in openssl the openssl command line application is also impacted by this issue vulnerability cve severity package fixed version link critical apr util ldap cve integer overflow or wraparound vulnerability in apr functions of apache portable runtime utility apr util allows an attacker to write beyond bounds of a buffer this issue affects apache portable runtime utility apr util and prior versions vulnerability cve severity package fixed version link critical apr cve integer overflow or wraparound vulnerability in apr encode functions of apache portable runtime apr allows an attacker to write beyond bounds of a buffer this issue affects apache portable runtime apr version vulnerability cve severity package fixed version link critical apr cve on windows apache portable runtime and earlier may write beyond the end of a stack based buffer in apr socket sendv this is a result of integer overflow vulnerability cve severity package fixed version link critical cve there is a type confusion vulnerability relating to x address processing inside an x generalname x addresses were parsed as an string but the public structure definition for general name incorrectly specified the type of the field as type this field is subsequently interpreted by the openssl function general name cmp as an type rather than an string when crl checking is enabled i e the application sets the v flag crl check flag this vulnerability may allow an attacker to pass arbitrary pointers to a memcmp call enabling them to read memory contents or enact a denial of service in most cases the attack requires the attacker to provide both the certificate chain and crl neither of which need to have a valid signature if the attacker only controls one of these inputs the other input must already contain an x address as a crl distribution point which is uncommon as such this vulnerability is most likely to only affect applications which have implemented their own functionality for retrieving crls over a network vulnerability cve severity package fixed version link high cve a null pointer can be dereferenced when signatures are being verified on signed or signedandenveloped data in case the hash algorithm used for the signature is known to the openssl library but the implementation of the hash algorithm is not available the digest initialization will fail there is a missing check for the return value from the initialization function which later leads to invalid usage of the digest api most likely leading to a crash the unavailability of an algorithm can be caused by using fips enabled configuration of providers or more commonly by not loading the legacy provider data is processed by the smime library calls and also by the time stamp ts library calls the tls implementation in openssl does not call these functions however third party applications would be affected if they call these functions to verify signatures on untrusted data vulnerability cve severity package fixed version link high cve an invalid pointer dereference on read can be triggered when an application tries to check a malformed dsa public key by the evp pkey public check function this will most likely lead to an application crash this function can be called on public keys supplied from untrusted sources which could allow an attacker to cause a denial of service attack the tls implementation in openssl does not call this function but applications might call the function if there are additional security requirements imposed by standards such as fips vulnerability cve severity package fixed version link high cve an invalid pointer dereference on read can be triggered when an application tries to load malformed data with the bio or fp functions the result of the dereference is an application crash which could lead to a denial of service attack the tls implementation in openssl does not call this function however third party applications might call these functions on untrusted data vulnerability cve severity package fixed version link high cve the public api function bio new ndef is a helper function used for streaming asn data via a bio it is primarily used internally to openssl to support the smime cms and streaming capabilities but may also be called directly by end user applications the function receives a bio from the caller prepends a new bio f filter bio onto the front of it to form a bio chain and then returns the new head of the bio chain to the caller under certain conditions for example if a cms recipient public key is invalid the new filter bio is freed and the function returns a null result indicating a failure however in this case the bio chain is not properly cleaned up and the bio passed by the caller still retains internal pointers to the previously freed filter bio if the caller then goes on to call bio pop on the bio then a use after free will occur this will most likely result in a crash this scenario occurs directly in the internal function write which may cause bio new ndef to be called and will subsequently call bio pop on the bio this internal function is in turn called by the public api functions pem write bio stream pem write bio cms stream pem write bio stream smime write smime write cms and smime write other public api functions that may be impacted by this include bio stream bio new cms bio new cms bio stream and bio stream the openssl cms and smime command line applications are similarly affected
0
263,679
8,300,542,667
IssuesEvent
2018-09-21 08:27:47
prometheus/prometheus
https://api.github.com/repos/prometheus/prometheus
closed
retention not working in version 2.3.1
component/promql kind/bug low hanging fruit priority/P3
Hey, i set prometheus retention to be 1h with flag --storage.tsdb.retention="1h" can see it in the ui but still can query data longer than 1h. i am using the latest version ui shows 1h: https://screenshot.net/3dk9riq more than 1h data: https://screenshot.net/20plrad i have been asked to open a bug when asked in prometheus-users
1.0
retention not working in version 2.3.1 - Hey, i set prometheus retention to be 1h with flag --storage.tsdb.retention="1h" can see it in the ui but still can query data longer than 1h. i am using the latest version ui shows 1h: https://screenshot.net/3dk9riq more than 1h data: https://screenshot.net/20plrad i have been asked to open a bug when asked in prometheus-users
non_process
retention not working in version hey i set prometheus retention to be with flag storage tsdb retention can see it in the ui but still can query data longer than i am using the latest version ui shows more than data i have been asked to open a bug when asked in prometheus users
0
9,139
12,203,187,289
IssuesEvent
2020-04-30 10:10:18
MHRA/products
https://api.github.com/repos/MHRA/products
closed
AUTO BATCH - Support XML Requests
EPIC - Auto Batch Process :oncoming_automobile: HIGH PRIORITY :arrow_double_up:
Accenture require an XML interface to the `doc-index-updater` API, as the Sentinel system is unable to serialize or deserialize JSON. **Acceptance Criteria** - The API should check the `Content-Type` or `Accept` headers. If the header is `application/xml`, then we should deserialize body & serialize response assuming XML rather than JSON; - This should be implemented so that it's easy to carry forward into future endpoints. **Stages** - [x] Backlog - [x] Discovery - [x] Dev - [x] Review - [ ] QA - [ ] UAT
1.0
AUTO BATCH - Support XML Requests - Accenture require an XML interface to the `doc-index-updater` API, as the Sentinel system is unable to serialize or deserialize JSON. **Acceptance Criteria** - The API should check the `Content-Type` or `Accept` headers. If the header is `application/xml`, then we should deserialize body & serialize response assuming XML rather than JSON; - This should be implemented so that it's easy to carry forward into future endpoints. **Stages** - [x] Backlog - [x] Discovery - [x] Dev - [x] Review - [ ] QA - [ ] UAT
process
auto batch support xml requests accenture require an xml interface to the doc index updater api as the sentinel system is unable to serialize or deserialize json acceptance criteria the api should check the content type or accept headers if the header is application xml then we should deserialize body serialize response assuming xml rather than json this should be implemented so that it s easy to carry forward into future endpoints stages backlog discovery dev review qa uat
1
418,515
12,198,907,903
IssuesEvent
2020-04-30 00:05:31
microsoft/terminal
https://api.github.com/repos/microsoft/terminal
closed
Crashes if zoom font size up and down rapidly with ctrl + scroll wheel
Area-Rendering In-PR Issue-Bug Priority-1 Product-Terminal Severity-Crash
# Environment Win10 1909 x64 Windows Terminal version (if applicable): 0.11.1121.0 # Steps to reproduce Increase and decrease the font size rapidly in the first tab using ctrl and mouse scroll wheel # Expected behavior Not to crash # Actual behavior Crashes --- Crashes every time when I use Ctrl and the mouse wheel to increase and decrease the font rapidly. Only happens with the first tab and regardless of what is running in there. I have crash dumps. Doesn't crash when I try and record a video of it in Snagit. ![image](https://user-images.githubusercontent.com/11651939/80043624-22dfbc80-84fa-11ea-8725-1b62874966cf.png)
1.0
Crashes if zoom font size up and down rapidly with ctrl + scroll wheel - # Environment Win10 1909 x64 Windows Terminal version (if applicable): 0.11.1121.0 # Steps to reproduce Increase and decrease the font size rapidly in the first tab using ctrl and mouse scroll wheel # Expected behavior Not to crash # Actual behavior Crashes --- Crashes every time when I use Ctrl and the mouse wheel to increase and decrease the font rapidly. Only happens with the first tab and regardless of what is running in there. I have crash dumps. Doesn't crash when I try and record a video of it in Snagit. ![image](https://user-images.githubusercontent.com/11651939/80043624-22dfbc80-84fa-11ea-8725-1b62874966cf.png)
non_process
crashes if zoom font size up and down rapidly with ctrl scroll wheel environment windows terminal version if applicable steps to reproduce increase and decrease the font size rapidly in the first tab using ctrl and mouse scroll wheel expected behavior not to crash actual behavior crashes crashes every time when i use ctrl and the mouse wheel to increase and decrease the font rapidly only happens with the first tab and regardless of what is running in there i have crash dumps doesn t crash when i try and record a video of it in snagit
0
50,550
7,611,010,183
IssuesEvent
2018-05-01 11:45:35
has2k1/plotnine
https://api.github.com/repos/has2k1/plotnine
closed
Changing limits in scale_y_continuous errors out:
documentation
qplot('currency', 'gini', data=df, geom="bar", stat="identity") +\ scale_y_continuous(range=(0.8, 1)) +\ theme(axis_text_x=element_text(rotation=45)) it looks like range is trying to call "train" on the object. From a wee df I can provide I think, returns this error: error here --> self.range.train(x) AttributeError: 'tuple' object has no attribute 'train'
1.0
Changing limits in scale_y_continuous errors out: - qplot('currency', 'gini', data=df, geom="bar", stat="identity") +\ scale_y_continuous(range=(0.8, 1)) +\ theme(axis_text_x=element_text(rotation=45)) it looks like range is trying to call "train" on the object. From a wee df I can provide I think, returns this error: error here --> self.range.train(x) AttributeError: 'tuple' object has no attribute 'train'
non_process
changing limits in scale y continuous errors out qplot currency gini data df geom bar stat identity scale y continuous range theme axis text x element text rotation it looks like range is trying to call train on the object from a wee df i can provide i think returns this error error here self range train x attributeerror tuple object has no attribute train
0
18,086
24,108,196,804
IssuesEvent
2022-09-20 09:11:33
cloudfoundry/korifi
https://api.github.com/repos/cloudfoundry/korifi
closed
[Feature]: Developer can push apps using the top-level `memory` field in the manifest
Top-level process config
### Background **As a** developer **I want** top-level process configuration in manifests to be supported **So that** I can use shortcut `cf push` flags like `-c`, `-i`, `-m` etc. ### Acceptance Criteria * **GIVEN** I have the sources of an application (e.g. `tests/smoke/assets/test-node-app`) **AND** `manifest.yml` looks like this: ```yaml --- applications: - name: my-app memory: 1G ``` **WHEN I** `cf push` **THEN I** see the push succeeds with an output similar to this: ``` name: test requested state: started routes: test.vcap.me last uploaded: Mon 29 Aug 16:28:36 UTC 2022 stack: cflinuxfs3 buildpacks: name version detect output buildpack name nodejs_buildpack 1.7.61 nodejs nodejs type: web sidecars: instances: 1/1 memory usage: 1G start command: npm start state since cpu memory disk details #0 running 2022-08-29T16:28:54Z 1.6% 42.3M of 1G 115.7M of 1G ``` * **GIVEN** I have the same app with the following manifest: ```yaml --- applications: - name: my-app memory: 512M processes: type: web memory: 1G ``` **WHEN I** `cf push` **THEN I** see the push succeeds with the same output as above ### Dev Notes This field is already [partially supported](https://github.com/cloudfoundry/korifi/blob/55a5e62991c7e45bb6d2c99c099d2a197ba7d758/api/actions/manifest.go#L52-L66). It's still not quite there as it gets ignored in presence of a `web` process, even if that doesn't have a `memory` field.
1.0
[Feature]: Developer can push apps using the top-level `memory` field in the manifest - ### Background **As a** developer **I want** top-level process configuration in manifests to be supported **So that** I can use shortcut `cf push` flags like `-c`, `-i`, `-m` etc. ### Acceptance Criteria * **GIVEN** I have the sources of an application (e.g. `tests/smoke/assets/test-node-app`) **AND** `manifest.yml` looks like this: ```yaml --- applications: - name: my-app memory: 1G ``` **WHEN I** `cf push` **THEN I** see the push succeeds with an output similar to this: ``` name: test requested state: started routes: test.vcap.me last uploaded: Mon 29 Aug 16:28:36 UTC 2022 stack: cflinuxfs3 buildpacks: name version detect output buildpack name nodejs_buildpack 1.7.61 nodejs nodejs type: web sidecars: instances: 1/1 memory usage: 1G start command: npm start state since cpu memory disk details #0 running 2022-08-29T16:28:54Z 1.6% 42.3M of 1G 115.7M of 1G ``` * **GIVEN** I have the same app with the following manifest: ```yaml --- applications: - name: my-app memory: 512M processes: type: web memory: 1G ``` **WHEN I** `cf push` **THEN I** see the push succeeds with the same output as above ### Dev Notes This field is already [partially supported](https://github.com/cloudfoundry/korifi/blob/55a5e62991c7e45bb6d2c99c099d2a197ba7d758/api/actions/manifest.go#L52-L66). It's still not quite there as it gets ignored in presence of a `web` process, even if that doesn't have a `memory` field.
process
developer can push apps using the top level memory field in the manifest background as a developer i want top level process configuration in manifests to be supported so that i can use shortcut cf push flags like c i m etc acceptance criteria given i have the sources of an application e g tests smoke assets test node app and manifest yml looks like this yaml applications name my app memory when i cf push then i see the push succeeds with an output similar to this name test requested state started routes test vcap me last uploaded mon aug utc stack buildpacks name version detect output buildpack name nodejs buildpack nodejs nodejs type web sidecars instances memory usage start command npm start state since cpu memory disk details running of of given i have the same app with the following manifest yaml applications name my app memory processes type web memory when i cf push then i see the push succeeds with the same output as above dev notes this field is already it s still not quite there as it gets ignored in presence of a web process even if that doesn t have a memory field
1
9,440
12,425,143,066
IssuesEvent
2020-05-24 15:01:53
jyn514/rcc
https://api.github.com/repos/jyn514/rcc
opened
Missing predefined macros
enhancement preprocessor
[6.10.8.1 Mandatory macros](http://port70.net/~nsz/c/c11/n1570.html#6.10.8.1) - [ ] `__LINE__` - tricky - can't be passed through to `replace()` literally. This might be as simple as updating `self.definitions` whenever someone calls `line()`? - [ ] `__COLUMN__` - tricky since we don't keep track of this explicitly and recalculate it. Probably needs yet more preprocessing before passing things through to `replace()`. - [ ] `__FILE__` - easy, but might need a refactor to put `PreProcessor` and `FileProcessor` back together. - [ ] `__DATE__` - easy, but needs a date-time dependency - [ ] `__TIME__` - same as `__DATE__` [6.10.8.3 Conditional feature macros](http://port70.net/~nsz/c/c11/n1570.html#6.10.8.3) - [ ] ` __STDC_IEC_559__ ` - trivial, good first issue. Code should go near https://github.com/jyn514/rcc/blob/180b27bb90bfcc02e2d4cba3afdd0a4dedccd662/src/lex/cpp.rs#L330. Note that `__LINE__` should be the line at time of replacement, not within a definition. For example: ```c $ clang -x c - -E -P #define a __LINE__ a 2 ```
1.0
Missing predefined macros - [6.10.8.1 Mandatory macros](http://port70.net/~nsz/c/c11/n1570.html#6.10.8.1) - [ ] `__LINE__` - tricky - can't be passed through to `replace()` literally. This might be as simple as updating `self.definitions` whenever someone calls `line()`? - [ ] `__COLUMN__` - tricky since we don't keep track of this explicitly and recalculate it. Probably needs yet more preprocessing before passing things through to `replace()`. - [ ] `__FILE__` - easy, but might need a refactor to put `PreProcessor` and `FileProcessor` back together. - [ ] `__DATE__` - easy, but needs a date-time dependency - [ ] `__TIME__` - same as `__DATE__` [6.10.8.3 Conditional feature macros](http://port70.net/~nsz/c/c11/n1570.html#6.10.8.3) - [ ] ` __STDC_IEC_559__ ` - trivial, good first issue. Code should go near https://github.com/jyn514/rcc/blob/180b27bb90bfcc02e2d4cba3afdd0a4dedccd662/src/lex/cpp.rs#L330. Note that `__LINE__` should be the line at time of replacement, not within a definition. For example: ```c $ clang -x c - -E -P #define a __LINE__ a 2 ```
process
missing predefined macros line tricky can t be passed through to replace literally this might be as simple as updating self definitions whenever someone calls line column tricky since we don t keep track of this explicitly and recalculate it probably needs yet more preprocessing before passing things through to replace file easy but might need a refactor to put preprocessor and fileprocessor back together date easy but needs a date time dependency time same as date stdc iec trivial good first issue code should go near note that line should be the line at time of replacement not within a definition for example c clang x c e p define a line a
1
179,757
30,294,624,249
IssuesEvent
2023-07-09 17:51:34
sars-cov-2-variants/lineage-proposals
https://api.github.com/repos/sars-cov-2-variants/lineage-proposals
closed
EU.1.1 with Orf8: A51V and its sublineage with S:F157L emerging in Germany.
Discussion Mutations outside spike designated
Mutations on top of EU.1.1: Seemingly just Orf8:A51V Gisaid Inquiry: Spike_I410V,P521S, F486P +C28045T Sequences: EPI_ISL_17070943, EPI_ISL_17080159, EPI_ISL_17113008, EPI_ISL_17113070, EPI_ISL_17184064, EPI_ISL_17184110, EPI_ISL_17184731, EPI_ISL_17186627, EPI_ISL_17189053, EPI_ISL_17189727, EPI_ISL_17209859, EPI_ISL_17210328, EPI_ISL_17247464, EPI_ISL_17254193, EPI_ISL_17256409, EPI_ISL_17257719, EPI_ISL_17257725, EPI_ISL_17258824, EPI_ISL_17258856, EPI_ISL_17258867, EPI_ISL_17258871, EPI_ISL_17264453, EPI_ISL_17264729, EPI_ISL_17269020, EPI_ISL_17269098, EPI_ISL_17269145, EPI_ISL_17274196, EPI_ISL_17293263, EPI_ISL_17293268, EPI_ISL_17296694, EPI_ISL_17307289, EPI_ISL_17315222, EPI_ISL_17315650, EPI_ISL_17315776, EPI_ISL_17326805, EPI_ISL_17327319, EPI_ISL_17344347, EPI_ISL_17344437, EPI_ISL_17348202, EPI_ISL_17375588, EPI_ISL_17389819, EPI_ISL_17390163, EPI_ISL_17394255, EPI_ISL_17394273, EPI_ISL_17397852, EPI_ISL_17397861, EPI_ISL_17397875, EPI_ISL_17397878, EPI_ISL_17398093, EPI_ISL_17398138, EPI_ISL_17398161, EPI_ISL_17407395, EPI_ISL_17410155, EPI_ISL_17414522, EPI_ISL_17440463, EPI_ISL_17442569, EPI_ISL_17463547, EPI_ISL_17464666, EPI_ISL_17468239-17468240, EPI_ISL_17468245, EPI_ISL_17468978, EPI_ISL_17469510, EPI_ISL_17472154, EPI_ISL_17472339, EPI_ISL_17472996, EPI_ISL_17479759, EPI_ISL_17479773, EPI_ISL_17479789, EPI_ISL_17479796, EPI_ISL_17486216, EPI_ISL_17495725, EPI_ISL_17500222, EPI_ISL_17500231, EPI_ISL_17510436, EPI_ISL_17513880, EPI_ISL_17513897, EPI_ISL_17513979, EPI_ISL_17514036, EPI_ISL_17514081, EPI_ISL_17515603, EPI_ISL_17515723, EPI_ISL_17515785, EPI_ISL_17518756, EPI_ISL_17519958, EPI_ISL_17521277, EPI_ISL_17525535, EPI_ISL_17525546, EPI_ISL_17525566, EPI_ISL_17525597, EPI_ISL_17528546, EPI_ISL_17528548, EPI_ISL_17534789, EPI_ISL_17538356, EPI_ISL_17540658, EPI_ISL_17541356, EPI_ISL_17547763, EPI_ISL_17547794, EPI_ISL_17547873, EPI_ISL_17553208, EPI_ISL_17554952, EPI_ISL_17556024, EPI_ISL_17562745-17562746, EPI_ISL_17562758, EPI_ISL_17566113, EPI_ISL_17590797, EPI_ISL_17598039, EPI_ISL_17598106, EPI_ISL_17598246, EPI_ISL_17607643, EPI_ISL_17615685, EPI_ISL_17621307, EPI_ISL_17621537, EPI_ISL_17621620, EPI_ISL_17623544, EPI_ISL_17625217, EPI_ISL_17625495, EPI_ISL_17629574, EPI_ISL_17634144-17634145, EPI_ISL_17637468-17637472, EPI_ISL_17637476, EPI_ISL_17637478-17637479, EPI_ISL_17639585, EPI_ISL_17653975, EPI_ISL_17654084, EPI_ISL_17654208, EPI_ISL_17654216-17654217, EPI_ISL_17654220, EPI_ISL_17654226, EPI_ISL_17662342, EPI_ISL_17662357, EPI_ISL_17669094, EPI_ISL_17669109, EPI_ISL_17672116, EPI_ISL_17679514, EPI_ISL_17680110, EPI_ISL_17685531, EPI_ISL_17685635, EPI_ISL_17687076-17687077, EPI_ISL_17689197, EPI_ISL_17695136, EPI_ISL_17700382, EPI_ISL_17700422, EPI_ISL_17700431, EPI_ISL_17701166, EPI_ISL_17701171, EPI_ISL_17701277, EPI_ISL_17701359-17701360, EPI_ISL_17701364, EPI_ISL_17701384-17701385, EPI_ISL_17701391, EPI_ISL_17701408-17701409, EPI_ISL_17701490-17701491, EPI_ISL_17704028, EPI_ISL_17714378, EPI_ISL_17716653, EPI_ISL_17723535, EPI_ISL_17731392, EPI_ISL_17732709, EPI_ISL_17742515, EPI_ISL_17744615, EPI_ISL_17763218, EPI_ISL_17763400, EPI_ISL_17764136, EPI_ISL_17765825, EPI_ISL_17769744, EPI_ISL_17770416, EPI_ISL_17775297, EPI_ISL_17775340, EPI_ISL_17775526, EPI_ISL_17777366, EPI_ISL_17777370-17777372, EPI_ISL_17778116, EPI_ISL_17781969, EPI_ISL_17782526, EPI_ISL_17783467, EPI_ISL_17785250, EPI_ISL_17785859, EPI_ISL_17787480, EPI_ISL_17789472, EPI_ISL_17794036, EPI_ISL_17794109, EPI_ISL_17795414, EPI_ISL_17795575-17795576, EPI_ISL_17795683, EPI_ISL_17795754-17795755, EPI_ISL_17795761, EPI_ISL_17795764, EPI_ISL_17795771, EPI_ISL_17795842, EPI_ISL_17795851, EPI_ISL_17795859, EPI_ISL_17795893, EPI_ISL_17796124, EPI_ISL_17796145, EPI_ISL_17796153, EPI_ISL_17796242, EPI_ISL_17796359, EPI_ISL_17796409, EPI_ISL_17803414, EPI_ISL_17803439, EPI_ISL_17804455, EPI_ISL_17815870, EPI_ISL_17826702, EPI_ISL_17826738, EPI_ISL_17826777, EPI_ISL_17826963, EPI_ISL_17827001, EPI_ISL_17827098, EPI_ISL_17827105, EPI_ISL_17827111, EPI_ISL_17831364, EPI_ISL_17831367, EPI_ISL_17832609, EPI_ISL_17833139, EPI_ISL_17833158, EPI_ISL_17836379, EPI_ISL_17836612, EPI_ISL_17837081, EPI_ISL_17837085, EPI_ISL_17838493, EPI_ISL_17849208-17849210, EPI_ISL_17849244-17849248, EPI_ISL_17850919, EPI_ISL_17850921, EPI_ISL_17852904, EPI_ISL_17855095, EPI_ISL_17855995, EPI_ISL_17859517, EPI_ISL_17859519, EPI_ISL_17859524, EPI_ISL_17859532, EPI_ISL_17859547, EPI_ISL_17880364, EPI_ISL_17880920, EPI_ISL_17880998, EPI_ISL_17881022, EPI_ISL_17881070, EPI_ISL_17881120, EPI_ISL_17881224, EPI_ISL_17881260, EPI_ISL_17883447, EPI_ISL_17950463, EPI_ISL_17954916, Earliest Sequence: EPI_ISL_17080159 Netherland 02-07,2023 Latest Sequence: EPI_ISL_17881224 Austria 06-21, 2023 Usher Tree: https://nextstrain.org/fetch/genome.ucsc.edu/trash/ct/subtreeAuspice1_genome_88af_363070.json?f_userOrOld=uploaded%20sample&label=id:node_2991533 Comment: This is weird because I would consider Orf8 mutations after XBB.1* gaining Orf8:G8* to be quasi-silent, yet Orf8:A51V seems to randomly appear among different variants. For example, it is spotted in #33 as part of the mutation chain. This strain seems to be the parent of my lineage B in #274 . It has shown quite significant growth compared to EU.1.1* base, and could be the reason behind EU.1.1*'s recent recovery in variant share. <img width="890" alt="2E6B2098-C5BD-4377-AC9C-3EF1733DF90C" src="https://github.com/sars-cov-2-variants/lineage-proposals/assets/131021067/0500254b-9d9e-41ad-93b7-679c480f24ec"> Edited Recently a sublineage of this popped up in GERMANY: Mutations:C28045T(ORF8:A51V)+C22033A(S:157L)+C14322T on EU.1.1* Inquiry: C22033A,C28045T,C14322T Distribution: Austria,GERMANY, NORWAY,USA Sequences: 28 (Updated) EPI_ISL_17685531, EPI_ISL_17700382, EPI_ISL_17700422, EPI_ISL_17782526, EPI_ISL_17803439, EPI_ISL_17810615, EPI_ISL_17826702, EPI_ISL_17826738, EPI_ISL_17826777, EPI_ISL_17826963, EPI_ISL_17827001, EPI_ISL_17827098, EPI_ISL_17827105, EPI_ISL_17827111, EPI_ISL_17833139, EPI_ISL_17837081, EPI_ISL_17837085 Earliest: EPI_ISL_17685531 ,GERMANY, 05-02,2023 Latest: EPI_ISL_17827111 Austria 06-13,2023 Usher Individual Tree: https://nextstrain.org/fetch/genome.ucsc.edu/trash/ct/subtreeAuspice1_genome_26cf5_ca3a80.json?f_userOrOld=uploaded%20sample&label=id:node_4922117
1.0
EU.1.1 with Orf8: A51V and its sublineage with S:F157L emerging in Germany. - Mutations on top of EU.1.1: Seemingly just Orf8:A51V Gisaid Inquiry: Spike_I410V,P521S, F486P +C28045T Sequences: EPI_ISL_17070943, EPI_ISL_17080159, EPI_ISL_17113008, EPI_ISL_17113070, EPI_ISL_17184064, EPI_ISL_17184110, EPI_ISL_17184731, EPI_ISL_17186627, EPI_ISL_17189053, EPI_ISL_17189727, EPI_ISL_17209859, EPI_ISL_17210328, EPI_ISL_17247464, EPI_ISL_17254193, EPI_ISL_17256409, EPI_ISL_17257719, EPI_ISL_17257725, EPI_ISL_17258824, EPI_ISL_17258856, EPI_ISL_17258867, EPI_ISL_17258871, EPI_ISL_17264453, EPI_ISL_17264729, EPI_ISL_17269020, EPI_ISL_17269098, EPI_ISL_17269145, EPI_ISL_17274196, EPI_ISL_17293263, EPI_ISL_17293268, EPI_ISL_17296694, EPI_ISL_17307289, EPI_ISL_17315222, EPI_ISL_17315650, EPI_ISL_17315776, EPI_ISL_17326805, EPI_ISL_17327319, EPI_ISL_17344347, EPI_ISL_17344437, EPI_ISL_17348202, EPI_ISL_17375588, EPI_ISL_17389819, EPI_ISL_17390163, EPI_ISL_17394255, EPI_ISL_17394273, EPI_ISL_17397852, EPI_ISL_17397861, EPI_ISL_17397875, EPI_ISL_17397878, EPI_ISL_17398093, EPI_ISL_17398138, EPI_ISL_17398161, EPI_ISL_17407395, EPI_ISL_17410155, EPI_ISL_17414522, EPI_ISL_17440463, EPI_ISL_17442569, EPI_ISL_17463547, EPI_ISL_17464666, EPI_ISL_17468239-17468240, EPI_ISL_17468245, EPI_ISL_17468978, EPI_ISL_17469510, EPI_ISL_17472154, EPI_ISL_17472339, EPI_ISL_17472996, EPI_ISL_17479759, EPI_ISL_17479773, EPI_ISL_17479789, EPI_ISL_17479796, EPI_ISL_17486216, EPI_ISL_17495725, EPI_ISL_17500222, EPI_ISL_17500231, EPI_ISL_17510436, EPI_ISL_17513880, EPI_ISL_17513897, EPI_ISL_17513979, EPI_ISL_17514036, EPI_ISL_17514081, EPI_ISL_17515603, EPI_ISL_17515723, EPI_ISL_17515785, EPI_ISL_17518756, EPI_ISL_17519958, EPI_ISL_17521277, EPI_ISL_17525535, EPI_ISL_17525546, EPI_ISL_17525566, EPI_ISL_17525597, EPI_ISL_17528546, EPI_ISL_17528548, EPI_ISL_17534789, EPI_ISL_17538356, EPI_ISL_17540658, EPI_ISL_17541356, EPI_ISL_17547763, EPI_ISL_17547794, EPI_ISL_17547873, EPI_ISL_17553208, EPI_ISL_17554952, EPI_ISL_17556024, EPI_ISL_17562745-17562746, EPI_ISL_17562758, EPI_ISL_17566113, EPI_ISL_17590797, EPI_ISL_17598039, EPI_ISL_17598106, EPI_ISL_17598246, EPI_ISL_17607643, EPI_ISL_17615685, EPI_ISL_17621307, EPI_ISL_17621537, EPI_ISL_17621620, EPI_ISL_17623544, EPI_ISL_17625217, EPI_ISL_17625495, EPI_ISL_17629574, EPI_ISL_17634144-17634145, EPI_ISL_17637468-17637472, EPI_ISL_17637476, EPI_ISL_17637478-17637479, EPI_ISL_17639585, EPI_ISL_17653975, EPI_ISL_17654084, EPI_ISL_17654208, EPI_ISL_17654216-17654217, EPI_ISL_17654220, EPI_ISL_17654226, EPI_ISL_17662342, EPI_ISL_17662357, EPI_ISL_17669094, EPI_ISL_17669109, EPI_ISL_17672116, EPI_ISL_17679514, EPI_ISL_17680110, EPI_ISL_17685531, EPI_ISL_17685635, EPI_ISL_17687076-17687077, EPI_ISL_17689197, EPI_ISL_17695136, EPI_ISL_17700382, EPI_ISL_17700422, EPI_ISL_17700431, EPI_ISL_17701166, EPI_ISL_17701171, EPI_ISL_17701277, EPI_ISL_17701359-17701360, EPI_ISL_17701364, EPI_ISL_17701384-17701385, EPI_ISL_17701391, EPI_ISL_17701408-17701409, EPI_ISL_17701490-17701491, EPI_ISL_17704028, EPI_ISL_17714378, EPI_ISL_17716653, EPI_ISL_17723535, EPI_ISL_17731392, EPI_ISL_17732709, EPI_ISL_17742515, EPI_ISL_17744615, EPI_ISL_17763218, EPI_ISL_17763400, EPI_ISL_17764136, EPI_ISL_17765825, EPI_ISL_17769744, EPI_ISL_17770416, EPI_ISL_17775297, EPI_ISL_17775340, EPI_ISL_17775526, EPI_ISL_17777366, EPI_ISL_17777370-17777372, EPI_ISL_17778116, EPI_ISL_17781969, EPI_ISL_17782526, EPI_ISL_17783467, EPI_ISL_17785250, EPI_ISL_17785859, EPI_ISL_17787480, EPI_ISL_17789472, EPI_ISL_17794036, EPI_ISL_17794109, EPI_ISL_17795414, EPI_ISL_17795575-17795576, EPI_ISL_17795683, EPI_ISL_17795754-17795755, EPI_ISL_17795761, EPI_ISL_17795764, EPI_ISL_17795771, EPI_ISL_17795842, EPI_ISL_17795851, EPI_ISL_17795859, EPI_ISL_17795893, EPI_ISL_17796124, EPI_ISL_17796145, EPI_ISL_17796153, EPI_ISL_17796242, EPI_ISL_17796359, EPI_ISL_17796409, EPI_ISL_17803414, EPI_ISL_17803439, EPI_ISL_17804455, EPI_ISL_17815870, EPI_ISL_17826702, EPI_ISL_17826738, EPI_ISL_17826777, EPI_ISL_17826963, EPI_ISL_17827001, EPI_ISL_17827098, EPI_ISL_17827105, EPI_ISL_17827111, EPI_ISL_17831364, EPI_ISL_17831367, EPI_ISL_17832609, EPI_ISL_17833139, EPI_ISL_17833158, EPI_ISL_17836379, EPI_ISL_17836612, EPI_ISL_17837081, EPI_ISL_17837085, EPI_ISL_17838493, EPI_ISL_17849208-17849210, EPI_ISL_17849244-17849248, EPI_ISL_17850919, EPI_ISL_17850921, EPI_ISL_17852904, EPI_ISL_17855095, EPI_ISL_17855995, EPI_ISL_17859517, EPI_ISL_17859519, EPI_ISL_17859524, EPI_ISL_17859532, EPI_ISL_17859547, EPI_ISL_17880364, EPI_ISL_17880920, EPI_ISL_17880998, EPI_ISL_17881022, EPI_ISL_17881070, EPI_ISL_17881120, EPI_ISL_17881224, EPI_ISL_17881260, EPI_ISL_17883447, EPI_ISL_17950463, EPI_ISL_17954916, Earliest Sequence: EPI_ISL_17080159 Netherland 02-07,2023 Latest Sequence: EPI_ISL_17881224 Austria 06-21, 2023 Usher Tree: https://nextstrain.org/fetch/genome.ucsc.edu/trash/ct/subtreeAuspice1_genome_88af_363070.json?f_userOrOld=uploaded%20sample&label=id:node_2991533 Comment: This is weird because I would consider Orf8 mutations after XBB.1* gaining Orf8:G8* to be quasi-silent, yet Orf8:A51V seems to randomly appear among different variants. For example, it is spotted in #33 as part of the mutation chain. This strain seems to be the parent of my lineage B in #274 . It has shown quite significant growth compared to EU.1.1* base, and could be the reason behind EU.1.1*'s recent recovery in variant share. <img width="890" alt="2E6B2098-C5BD-4377-AC9C-3EF1733DF90C" src="https://github.com/sars-cov-2-variants/lineage-proposals/assets/131021067/0500254b-9d9e-41ad-93b7-679c480f24ec"> Edited Recently a sublineage of this popped up in GERMANY: Mutations:C28045T(ORF8:A51V)+C22033A(S:157L)+C14322T on EU.1.1* Inquiry: C22033A,C28045T,C14322T Distribution: Austria,GERMANY, NORWAY,USA Sequences: 28 (Updated) EPI_ISL_17685531, EPI_ISL_17700382, EPI_ISL_17700422, EPI_ISL_17782526, EPI_ISL_17803439, EPI_ISL_17810615, EPI_ISL_17826702, EPI_ISL_17826738, EPI_ISL_17826777, EPI_ISL_17826963, EPI_ISL_17827001, EPI_ISL_17827098, EPI_ISL_17827105, EPI_ISL_17827111, EPI_ISL_17833139, EPI_ISL_17837081, EPI_ISL_17837085 Earliest: EPI_ISL_17685531 ,GERMANY, 05-02,2023 Latest: EPI_ISL_17827111 Austria 06-13,2023 Usher Individual Tree: https://nextstrain.org/fetch/genome.ucsc.edu/trash/ct/subtreeAuspice1_genome_26cf5_ca3a80.json?f_userOrOld=uploaded%20sample&label=id:node_4922117
non_process
eu with and its sublineage with s emerging in germany mutations on top of eu seemingly just gisaid inquiry spike sequences epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl earliest sequence epi isl netherland latest sequence epi isl austria usher tree comment this is weird because i would consider mutations after xbb gaining to be quasi silent yet seems to randomly appear among different variants for example it is spotted in as part of the mutation chain this strain seems to be the parent of my lineage b in it has shown quite significant growth compared to eu base and could be the reason behind eu s recent recovery in variant share img width alt src edited recently a sublineage of this popped up in germany mutations s on eu inquiry distribution austria germany norway usa sequences updated epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl epi isl earliest epi isl germany latest epi isl austria usher individual tree
0
11,445
14,264,499,745
IssuesEvent
2020-11-20 15:50:17
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
StartInfo_TextFile_ShellExecute failing consistently on Server Core
area-System.Diagnostics.Process bug
ERROR: type should be string, got "https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180521.01/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessStartInfoTests~2FStartInfo_TextFile_ShellExecute\r\n\r\nApparently Process.Start on foo.txt with UseShellExecute=true is returning a null process object. \r\n\r\n```\r\nCould not start C:\\\\Users\\\\runner\\\\AppData\\\\Local\\\\Temp\\\\ProcessStartInfoTests_x1e54hcw.eri\\\\StartInfo_TextFile_ShellExecute_993_88c6de27.txt UseShellExecute=True\r\nAssociation details for '.txt'\r\n------------------------------\r\nOpen command: C:\\\\Windows\\\\system32\\\\NOTEPAD.EXE %1\r\nProgID: txtfile\r\n\r\nExpected: True\r\nActual: False\r\nStack Trace :\r\n at System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_TextFile_ShellExecute() in E:\\A\\_work\\63\\s\\corefx\\src\\System.Diagnostics.Process\\tests\\ProcessStartInfoTests.cs:line 990\r\n```"
1.0
StartInfo_TextFile_ShellExecute failing consistently on Server Core - https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fcli~2F/build/20180521.01/workItem/System.Diagnostics.Process.Tests/analysis/xunit/System.Diagnostics.Tests.ProcessStartInfoTests~2FStartInfo_TextFile_ShellExecute Apparently Process.Start on foo.txt with UseShellExecute=true is returning a null process object. ``` Could not start C:\\Users\\runner\\AppData\\Local\\Temp\\ProcessStartInfoTests_x1e54hcw.eri\\StartInfo_TextFile_ShellExecute_993_88c6de27.txt UseShellExecute=True Association details for '.txt' ------------------------------ Open command: C:\\Windows\\system32\\NOTEPAD.EXE %1 ProgID: txtfile Expected: True Actual: False Stack Trace : at System.Diagnostics.Tests.ProcessStartInfoTests.StartInfo_TextFile_ShellExecute() in E:\A\_work\63\s\corefx\src\System.Diagnostics.Process\tests\ProcessStartInfoTests.cs:line 990 ```
process
startinfo textfile shellexecute failing consistently on server core apparently process start on foo txt with useshellexecute true is returning a null process object could not start c users runner appdata local temp processstartinfotests eri startinfo textfile shellexecute txt useshellexecute true association details for txt open command c windows notepad exe progid txtfile expected true actual false stack trace at system diagnostics tests processstartinfotests startinfo textfile shellexecute in e a work s corefx src system diagnostics process tests processstartinfotests cs line
1
18,490
24,550,908,479
IssuesEvent
2022-10-12 12:32:40
GoogleCloudPlatform/fda-mystudies
https://api.github.com/repos/GoogleCloudPlatform/fda-mystudies
closed
[iOS] [Offline indicator] Sign in screen > Empty screen is getting displayed
Bug P1 iOS Process: Fixed Process: Tested QA Process: Tested dev
Steps: 1. Install the app and open the app 2. Click on the 'Get started' 3. Click on the Hamburger menu button 4. Click on 'Signin' 5. Blank screen is getting displayed with a 'You are offline' message 6. Turn on the data and observe AR: Blank screen is getting displayed even though user turn on their mobile data ER: Sign in page should get a load Refer to the attached video https://user-images.githubusercontent.com/71445210/178304851-3e80bd15-49c3-40f8-9aed-f60f73332919.MOV
3.0
[iOS] [Offline indicator] Sign in screen > Empty screen is getting displayed - Steps: 1. Install the app and open the app 2. Click on the 'Get started' 3. Click on the Hamburger menu button 4. Click on 'Signin' 5. Blank screen is getting displayed with a 'You are offline' message 6. Turn on the data and observe AR: Blank screen is getting displayed even though user turn on their mobile data ER: Sign in page should get a load Refer to the attached video https://user-images.githubusercontent.com/71445210/178304851-3e80bd15-49c3-40f8-9aed-f60f73332919.MOV
process
sign in screen empty screen is getting displayed steps install the app and open the app click on the get started click on the hamburger menu button click on signin blank screen is getting displayed with a you are offline message turn on the data and observe ar blank screen is getting displayed even though user turn on their mobile data er sign in page should get a load refer to the attached video
1
8,104
11,299,519,037
IssuesEvent
2020-01-17 11:23:39
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
SAGA Fill sinks (Wang & Liu) gives error in 3.4.13
Bug Feedback Processing
**Describe the bug** In QGIS 3.4.13 the SAGA Fill sink (Wang & Liu) processing tool is not working. It gives an error. See screenshot. We tested it against QGIS 3.4.10 where it was still working with the same procedure and files. ![error_fill_sinks](https://user-images.githubusercontent.com/1172662/69155490-189c3b00-0ae2-11ea-9a0c-0ea4498273c9.jpg) **How to Reproduce** I used the "catchment delineation procedure" as explained in the book QGIS for Hydrological Applications and this video (https://youtu.be/Ro-RRzMMw-c). Here I'll only put the summary: 1. In the Processing Tools go to SAGA -> Terrain Analysis - Hydrology -> Fill sinks (Wang & Liu) 2. Select an DEM tif file (we use SRTM 1 arc second, reprojected to UTM/WGS-84) 3. Keep defaults and Save output to dem_fill.tif. You can uncheck the boxes for Flow Direction and Watershed Basins. Click Run. 4. See error --> screenshot above **QGIS and OS versions** 3.4.13 for Windows 64 bit **Additional context** It was working in 3.4.10 and older versions
1.0
SAGA Fill sinks (Wang & Liu) gives error in 3.4.13 - **Describe the bug** In QGIS 3.4.13 the SAGA Fill sink (Wang & Liu) processing tool is not working. It gives an error. See screenshot. We tested it against QGIS 3.4.10 where it was still working with the same procedure and files. ![error_fill_sinks](https://user-images.githubusercontent.com/1172662/69155490-189c3b00-0ae2-11ea-9a0c-0ea4498273c9.jpg) **How to Reproduce** I used the "catchment delineation procedure" as explained in the book QGIS for Hydrological Applications and this video (https://youtu.be/Ro-RRzMMw-c). Here I'll only put the summary: 1. In the Processing Tools go to SAGA -> Terrain Analysis - Hydrology -> Fill sinks (Wang & Liu) 2. Select an DEM tif file (we use SRTM 1 arc second, reprojected to UTM/WGS-84) 3. Keep defaults and Save output to dem_fill.tif. You can uncheck the boxes for Flow Direction and Watershed Basins. Click Run. 4. See error --> screenshot above **QGIS and OS versions** 3.4.13 for Windows 64 bit **Additional context** It was working in 3.4.10 and older versions
process
saga fill sinks wang liu gives error in describe the bug in qgis the saga fill sink wang liu processing tool is not working it gives an error see screenshot we tested it against qgis where it was still working with the same procedure and files how to reproduce i used the catchment delineation procedure as explained in the book qgis for hydrological applications and this video here i ll only put the summary in the processing tools go to saga terrain analysis hydrology fill sinks wang liu select an dem tif file we use srtm arc second reprojected to utm wgs keep defaults and save output to dem fill tif you can uncheck the boxes for flow direction and watershed basins click run see error screenshot above qgis and os versions for windows bit additional context it was working in and older versions
1
12
2,496,237,293
IssuesEvent
2015-01-06 18:02:29
vivo-isf/vivo-isf-ontology
https://api.github.com/repos/vivo-isf/vivo-isf-ontology
closed
DNA replication
biological_process imported
_From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 15:57:52_ GO:0006260 Parent: cellular biosynthetic process \<a href="http://purl.obolibrary.org/obo/GO_0044249" rel="nofollow">http://purl.obolibrary.org/obo/GO_0044249</a>&#13; &#13; for this record:&#13; \<a href="http://ohsu.eagle-i.net/i/0000012c-0e30-ff12-20c4-4fd180000000" rel="nofollow">http://ohsu.eagle-i.net/i/0000012c-0e30-ff12-20c4-4fd180000000</a> _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=163_
1.0
DNA replication - _From [vasil...@ohsu.edu](https://code.google.com/u/108803237899917466626/) on November 12, 2012 15:57:52_ GO:0006260 Parent: cellular biosynthetic process \<a href="http://purl.obolibrary.org/obo/GO_0044249" rel="nofollow">http://purl.obolibrary.org/obo/GO_0044249</a>&#13; &#13; for this record:&#13; \<a href="http://ohsu.eagle-i.net/i/0000012c-0e30-ff12-20c4-4fd180000000" rel="nofollow">http://ohsu.eagle-i.net/i/0000012c-0e30-ff12-20c4-4fd180000000</a> _Original issue: http://code.google.com/p/eagle-i/issues/detail?id=163_
process
dna replication from on november go parent cellular biosynthetic process for this record original issue
1
13,446
15,883,719,332
IssuesEvent
2021-04-09 17:48:04
MicrosoftDocs/azure-devops-docs
https://api.github.com/repos/MicrosoftDocs/azure-devops-docs
closed
Documentation is incorrect for issecret, fails to note secret variables do *not* get set as env vars
Pri1 devops-cicd-process/tech devops/prod doc-enhancement
The documentation mentions: > This doesn't update the environment variables, but it does make the new variable available to downstream steps within the same job. > > When issecret is set to true, the value of the variable will be saved as secret and masked from the log. > ... > Subsequent steps will also have the pipeline variable added to their environment. This is confusing an inaccurate in two ways: 1. Environment variables *are* updated, just not immediately for the running job. Subsequent jobs have access to the new variable in YAML via `$(MACRO_SYNTAX)` **and** in tasks via environment variable. 2. Secret variables are only available via `$(MACRO_SYNTAX)` not via environment variable, and this is not mentioned in this documentation page. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: dd7e0bd3-1f7d-d7b6-cc72-5ef63c31b46a * Version Independent ID: dae87abd-b73d-9120-bcdb-6097d4b40f2a * Content: [Define variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?view=azure-devops&tabs=yaml%2Cbatch#set-variables-in-scripts) * Content Source: [docs/pipelines/process/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/variables.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
1.0
Documentation is incorrect for issecret, fails to note secret variables do *not* get set as env vars - The documentation mentions: > This doesn't update the environment variables, but it does make the new variable available to downstream steps within the same job. > > When issecret is set to true, the value of the variable will be saved as secret and masked from the log. > ... > Subsequent steps will also have the pipeline variable added to their environment. This is confusing an inaccurate in two ways: 1. Environment variables *are* updated, just not immediately for the running job. Subsequent jobs have access to the new variable in YAML via `$(MACRO_SYNTAX)` **and** in tasks via environment variable. 2. Secret variables are only available via `$(MACRO_SYNTAX)` not via environment variable, and this is not mentioned in this documentation page. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: dd7e0bd3-1f7d-d7b6-cc72-5ef63c31b46a * Version Independent ID: dae87abd-b73d-9120-bcdb-6097d4b40f2a * Content: [Define variables - Azure Pipelines](https://docs.microsoft.com/en-us/azure/devops/pipelines/process/variables?view=azure-devops&tabs=yaml%2Cbatch#set-variables-in-scripts) * Content Source: [docs/pipelines/process/variables.md](https://github.com/MicrosoftDocs/azure-devops-docs/blob/master/docs/pipelines/process/variables.md) * Product: **devops** * Technology: **devops-cicd-process** * GitHub Login: @juliakm * Microsoft Alias: **jukullam**
process
documentation is incorrect for issecret fails to note secret variables do not get set as env vars the documentation mentions this doesn t update the environment variables but it does make the new variable available to downstream steps within the same job when issecret is set to true the value of the variable will be saved as secret and masked from the log subsequent steps will also have the pipeline variable added to their environment this is confusing an inaccurate in two ways environment variables are updated just not immediately for the running job subsequent jobs have access to the new variable in yaml via macro syntax and in tasks via environment variable secret variables are only available via macro syntax not via environment variable and this is not mentioned in this documentation page document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id bcdb content content source product devops technology devops cicd process github login juliakm microsoft alias jukullam
1
7,568
10,683,362,182
IssuesEvent
2019-10-22 08:11:35
prisma/prisma2
https://api.github.com/repos/prisma/prisma2
closed
`prisma2 studio`
kind/feature process/candidate
It would be nice to have a CLI command to start Studio out of the context of Watchmode (`prisma2 dev`), so without any possible side effects of migration, the `dev` UI complications and so on. Possible output: ``` $ prisma2 studio Prisma Studio is now available at http://localhost:5555/ Ctrl + C to exit ^C $ ``` Notes: - If you want to get fancy, it could use https://www.npmjs.com/package/open to automatically open the URL in the default browser as well (as this command is explicitly about using Studio, so this matches the user's expectation). - As this command is a first experiment, it should probably not yet be surfaced in any help commands, release notes or similar.
1.0
`prisma2 studio` - It would be nice to have a CLI command to start Studio out of the context of Watchmode (`prisma2 dev`), so without any possible side effects of migration, the `dev` UI complications and so on. Possible output: ``` $ prisma2 studio Prisma Studio is now available at http://localhost:5555/ Ctrl + C to exit ^C $ ``` Notes: - If you want to get fancy, it could use https://www.npmjs.com/package/open to automatically open the URL in the default browser as well (as this command is explicitly about using Studio, so this matches the user's expectation). - As this command is a first experiment, it should probably not yet be surfaced in any help commands, release notes or similar.
process
studio it would be nice to have a cli command to start studio out of the context of watchmode dev so without any possible side effects of migration the dev ui complications and so on possible output studio prisma studio is now available at ctrl c to exit c notes if you want to get fancy it could use to automatically open the url in the default browser as well as this command is explicitly about using studio so this matches the user s expectation as this command is a first experiment it should probably not yet be surfaced in any help commands release notes or similar
1
11,034
13,850,497,674
IssuesEvent
2020-10-15 01:23:20
kubeflow/community
https://api.github.com/repos/kubeflow/community
closed
Additional calendar admins needed
kind/process
We need additional admins for the [Kubeflow Calendar](https://calendar.google.com/calendar/embed?src=kubeflow.org_7l5vnbn8suj2se10sen81d9428%40group.calendar.google.com&ctz=America%2FLos_Angeles) We currently have some partial tooling for managing the calendar. The [script](https://github.com/kubeflow/community/blob/master/scripts/calendar_import.py) configures the calendar based on [calendar.yaml](https://github.com/kubeflow/community/blob/master/calendar.yaml) The script currently needs to be run manually by someone with admin privileges to the calendar. Expectations for would be volunteers * Figure out what needs to happen to gain access to the calendar * Commit to helping SIGs/WGs/ etc... get meetings added to the calendar * Document the process * Work on automating the process to make it more scalable * Move calendar.yaml into its OWN folder and create an OWNERs file * Review and approve updates to the calendar
1.0
Additional calendar admins needed - We need additional admins for the [Kubeflow Calendar](https://calendar.google.com/calendar/embed?src=kubeflow.org_7l5vnbn8suj2se10sen81d9428%40group.calendar.google.com&ctz=America%2FLos_Angeles) We currently have some partial tooling for managing the calendar. The [script](https://github.com/kubeflow/community/blob/master/scripts/calendar_import.py) configures the calendar based on [calendar.yaml](https://github.com/kubeflow/community/blob/master/calendar.yaml) The script currently needs to be run manually by someone with admin privileges to the calendar. Expectations for would be volunteers * Figure out what needs to happen to gain access to the calendar * Commit to helping SIGs/WGs/ etc... get meetings added to the calendar * Document the process * Work on automating the process to make it more scalable * Move calendar.yaml into its OWN folder and create an OWNERs file * Review and approve updates to the calendar
process
additional calendar admins needed we need additional admins for the we currently have some partial tooling for managing the calendar the configures the calendar based on the script currently needs to be run manually by someone with admin privileges to the calendar expectations for would be volunteers figure out what needs to happen to gain access to the calendar commit to helping sigs wgs etc get meetings added to the calendar document the process work on automating the process to make it more scalable move calendar yaml into its own folder and create an owners file review and approve updates to the calendar
1
21,630
30,034,010,939
IssuesEvent
2023-06-27 11:32:40
NationalSecurityAgency/ghidra
https://api.github.com/repos/NationalSecurityAgency/ghidra
closed
dsPIC30F6014 Emulation Error
Feature: Processor/PIC Status: Internal
**Describe the bug** Attempting to emulate a dsPIC30F6014 file causes: ``` java.lang.IllegalArgumentException: Memory-mapped register PC does not map to space register: ``` **To Reproduce** Steps to reproduce the behavior: 1. Compile program with the [XC16 compiler](https://www.microchip.com/en-us/tools-resources/develop/mplab-xc-compilers/downloads-documentation#XC16): ```C #include <stdio.h> int main() { printf("Hello World"); return 0; } ``` 2. Compile for dsPIC30F6014: ```bash xc16-gcc -mcpu=30f6014 -o hello_world2.elf hello_world2.c ``` 3. Load program into Ghidra and analyze. 4. Navigate to `_main` 5. Right click on instruction and click `Emulate Program in new Trace` **Expected behavior** The ability to emulate the instructions. **Screenshots** If applicable, add screenshots to help explain your problem. **Attachments** [error2.log](https://github.com/NationalSecurityAgency/ghidra/files/11638020/error2.log) **Environment (please complete the following information):** - OS: Ubuntu 20.04 Linux 5.15.0-71-generic amd64 - Java Version: openjdk 17.0.7 - Ghidra Version: 10.2.3 - Ghidra Origin: Official GitHub distro Let me know if you need me to provide anything else.
1.0
dsPIC30F6014 Emulation Error - **Describe the bug** Attempting to emulate a dsPIC30F6014 file causes: ``` java.lang.IllegalArgumentException: Memory-mapped register PC does not map to space register: ``` **To Reproduce** Steps to reproduce the behavior: 1. Compile program with the [XC16 compiler](https://www.microchip.com/en-us/tools-resources/develop/mplab-xc-compilers/downloads-documentation#XC16): ```C #include <stdio.h> int main() { printf("Hello World"); return 0; } ``` 2. Compile for dsPIC30F6014: ```bash xc16-gcc -mcpu=30f6014 -o hello_world2.elf hello_world2.c ``` 3. Load program into Ghidra and analyze. 4. Navigate to `_main` 5. Right click on instruction and click `Emulate Program in new Trace` **Expected behavior** The ability to emulate the instructions. **Screenshots** If applicable, add screenshots to help explain your problem. **Attachments** [error2.log](https://github.com/NationalSecurityAgency/ghidra/files/11638020/error2.log) **Environment (please complete the following information):** - OS: Ubuntu 20.04 Linux 5.15.0-71-generic amd64 - Java Version: openjdk 17.0.7 - Ghidra Version: 10.2.3 - Ghidra Origin: Official GitHub distro Let me know if you need me to provide anything else.
process
emulation error describe the bug attempting to emulate a file causes java lang illegalargumentexception memory mapped register pc does not map to space register to reproduce steps to reproduce the behavior compile program with the c include int main printf hello world return compile for bash gcc mcpu o hello elf hello c load program into ghidra and analyze navigate to main right click on instruction and click emulate program in new trace expected behavior the ability to emulate the instructions screenshots if applicable add screenshots to help explain your problem attachments environment please complete the following information os ubuntu linux generic java version openjdk ghidra version ghidra origin official github distro let me know if you need me to provide anything else
1
1,869
4,697,561,610
IssuesEvent
2016-10-12 09:46:27
nodejs/node
https://api.github.com/repos/nodejs/node
closed
child_process.spawn will either exit or not control stdin when detached and using inherit
child_process os x
This may be multiple issues in node or just discrepancies between what the docs suggest and how it actually operates. > When using the detached option to start a long-running process, ... If the parent's stdio is inherited, the child will remain attached to the controlling terminal. ### conosle.log() causes early exit This fails by exiting immediately (with success error code) `fail-on-spawn.js`: ```javascript 'use strict'; var spawn = require('child_process').spawn; console.log('hello'); var child = spawn( 'ping' , [ '-c', '3', 'google.com' ] , { detached: true, stdio: 'inherit' } ); child.unref(); process.on('exit', function () { console.log('goodbye'); }); ``` ```bash node fail-on-spawn.js hello goodbye echo $? 0 ``` ### Detaches and stdin is lost If we change nothing but removing the initial `console.log('hello')` (or even just put it after `spawn` is called) then we get more desirable results. `fails-to-foreground.js`: ```javascript 'use strict'; var spawn = require('child_process').spawn; // just comment that out there don't ya know //console.log('hello'); var child = spawn( 'ping' , [ '-c', '3', 'google.com' ] , { detached: true, stdio: 'inherit' } ); child.unref(); process.on('exit', function () { console.log('goodbye'); }); ``` ```bash node fails-to-foreground.js goodbye aj@brunchfast ~> PING google.com (216.58.217.46): 56 data bytes 64 bytes from 216.58.217.46: icmp_seq=0 ttl=55 time=21.414 ms 64 bytes from 216.58.217.46: icmp_seq=1 ttl=55 time=21.091 ms 64 bytes from 216.58.217.46: icmp_seq=2 ttl=55 time=21.613 ms --- google.com ping statistics --- 3 packets transmitted, 3 packets received, 0.0% packet loss round-trip min/avg/max/stddev = 21.091/21.373/21.613/0.215 ms ``` As you can see, as soon as node exits process.stdin goes back to bash rather than being handled by ping. ### runs in foreground when not detached (this works as documented, I'm using it as an example as how I expect the above to work) From the description of the documentation, I would expect the user experience to be just like this: node has control, then gives control to the next program to be opened - except that in this example the child process is not detached and hence node cannot exit. `foreground-without-detach.js`: ```javascript 'use strict'; var fs = require('fs'); var child_process = require('child_process'); var file = '/tmp/node-editor-test'; fs.writeFile(file, "Node Editor", function () { var child = child_process.spawn('vim', [file], {stdio: 'inherit'}); // would like to add detached: true and child.unref(), but that will kill this }); ``` If we do try to detach it it fails dramatically since this program actually depends on stdin whereas ping does not. ### More examples of unexpected(?) behavior ```javascript 'use strict'; var fs = require('fs'); var child_process = require('child_process'); var file = '/tmp/node-editor-test'; fs.writeFile(file, "Node Editor", function () { var child = child_process.spawn('vim', [file], {stdio: 'inherit'}); child.unref(); }); ``` (both die) or ```javascript 'use strict'; var fs = require('fs'); var child_process = require('child_process'); var file = '/tmp/node-editor-test'; fs.writeFile(file, "Node Editor", function () { var child = child_process.spawn('vim', [file], { detached: true, stdio: 'inherit'}); process.exit(0); }); ``` (this one above is neat because both vim and terminal get control of stdin and it creates all sorts of mess) or ```javascript 'use strict'; var fs = require('fs'); var child_process = require('child_process'); var file = '/tmp/node-editor-test'; fs.writeFile(file, "Node Editor", function () { var child = child_process.spawn('vim', [file], { detached: true, stdio: 'inherit'}); child.unref(); }); ``` (both die) * **Version**: v4.3.2 (also tried in v5.2 or so) * **Platform**: Darwin brunchfast.local 13.0.0 Darwin Kernel Version 13.0.0: Thu Sep 19 22:22:27 PDT 2013; root:xnu-2422.1.72~6/RELEASE_X86_64 x86_64 * **Subsystem**: child_process
1.0
child_process.spawn will either exit or not control stdin when detached and using inherit - This may be multiple issues in node or just discrepancies between what the docs suggest and how it actually operates. > When using the detached option to start a long-running process, ... If the parent's stdio is inherited, the child will remain attached to the controlling terminal. ### conosle.log() causes early exit This fails by exiting immediately (with success error code) `fail-on-spawn.js`: ```javascript 'use strict'; var spawn = require('child_process').spawn; console.log('hello'); var child = spawn( 'ping' , [ '-c', '3', 'google.com' ] , { detached: true, stdio: 'inherit' } ); child.unref(); process.on('exit', function () { console.log('goodbye'); }); ``` ```bash node fail-on-spawn.js hello goodbye echo $? 0 ``` ### Detaches and stdin is lost If we change nothing but removing the initial `console.log('hello')` (or even just put it after `spawn` is called) then we get more desirable results. `fails-to-foreground.js`: ```javascript 'use strict'; var spawn = require('child_process').spawn; // just comment that out there don't ya know //console.log('hello'); var child = spawn( 'ping' , [ '-c', '3', 'google.com' ] , { detached: true, stdio: 'inherit' } ); child.unref(); process.on('exit', function () { console.log('goodbye'); }); ``` ```bash node fails-to-foreground.js goodbye aj@brunchfast ~> PING google.com (216.58.217.46): 56 data bytes 64 bytes from 216.58.217.46: icmp_seq=0 ttl=55 time=21.414 ms 64 bytes from 216.58.217.46: icmp_seq=1 ttl=55 time=21.091 ms 64 bytes from 216.58.217.46: icmp_seq=2 ttl=55 time=21.613 ms --- google.com ping statistics --- 3 packets transmitted, 3 packets received, 0.0% packet loss round-trip min/avg/max/stddev = 21.091/21.373/21.613/0.215 ms ``` As you can see, as soon as node exits process.stdin goes back to bash rather than being handled by ping. ### runs in foreground when not detached (this works as documented, I'm using it as an example as how I expect the above to work) From the description of the documentation, I would expect the user experience to be just like this: node has control, then gives control to the next program to be opened - except that in this example the child process is not detached and hence node cannot exit. `foreground-without-detach.js`: ```javascript 'use strict'; var fs = require('fs'); var child_process = require('child_process'); var file = '/tmp/node-editor-test'; fs.writeFile(file, "Node Editor", function () { var child = child_process.spawn('vim', [file], {stdio: 'inherit'}); // would like to add detached: true and child.unref(), but that will kill this }); ``` If we do try to detach it it fails dramatically since this program actually depends on stdin whereas ping does not. ### More examples of unexpected(?) behavior ```javascript 'use strict'; var fs = require('fs'); var child_process = require('child_process'); var file = '/tmp/node-editor-test'; fs.writeFile(file, "Node Editor", function () { var child = child_process.spawn('vim', [file], {stdio: 'inherit'}); child.unref(); }); ``` (both die) or ```javascript 'use strict'; var fs = require('fs'); var child_process = require('child_process'); var file = '/tmp/node-editor-test'; fs.writeFile(file, "Node Editor", function () { var child = child_process.spawn('vim', [file], { detached: true, stdio: 'inherit'}); process.exit(0); }); ``` (this one above is neat because both vim and terminal get control of stdin and it creates all sorts of mess) or ```javascript 'use strict'; var fs = require('fs'); var child_process = require('child_process'); var file = '/tmp/node-editor-test'; fs.writeFile(file, "Node Editor", function () { var child = child_process.spawn('vim', [file], { detached: true, stdio: 'inherit'}); child.unref(); }); ``` (both die) * **Version**: v4.3.2 (also tried in v5.2 or so) * **Platform**: Darwin brunchfast.local 13.0.0 Darwin Kernel Version 13.0.0: Thu Sep 19 22:22:27 PDT 2013; root:xnu-2422.1.72~6/RELEASE_X86_64 x86_64 * **Subsystem**: child_process
process
child process spawn will either exit or not control stdin when detached and using inherit this may be multiple issues in node or just discrepancies between what the docs suggest and how it actually operates when using the detached option to start a long running process if the parent s stdio is inherited the child will remain attached to the controlling terminal conosle log causes early exit this fails by exiting immediately with success error code fail on spawn js javascript use strict var spawn require child process spawn console log hello var child spawn ping detached true stdio inherit child unref process on exit function console log goodbye bash node fail on spawn js hello goodbye echo detaches and stdin is lost if we change nothing but removing the initial console log hello or even just put it after spawn is called then we get more desirable results fails to foreground js javascript use strict var spawn require child process spawn just comment that out there don t ya know console log hello var child spawn ping detached true stdio inherit child unref process on exit function console log goodbye bash node fails to foreground js goodbye aj brunchfast ping google com data bytes bytes from icmp seq ttl time ms bytes from icmp seq ttl time ms bytes from icmp seq ttl time ms google com ping statistics packets transmitted packets received packet loss round trip min avg max stddev ms as you can see as soon as node exits process stdin goes back to bash rather than being handled by ping runs in foreground when not detached this works as documented i m using it as an example as how i expect the above to work from the description of the documentation i would expect the user experience to be just like this node has control then gives control to the next program to be opened except that in this example the child process is not detached and hence node cannot exit foreground without detach js javascript use strict var fs require fs var child process require child process var file tmp node editor test fs writefile file node editor function var child child process spawn vim stdio inherit would like to add detached true and child unref but that will kill this if we do try to detach it it fails dramatically since this program actually depends on stdin whereas ping does not more examples of unexpected behavior javascript use strict var fs require fs var child process require child process var file tmp node editor test fs writefile file node editor function var child child process spawn vim stdio inherit child unref both die or javascript use strict var fs require fs var child process require child process var file tmp node editor test fs writefile file node editor function var child child process spawn vim detached true stdio inherit process exit this one above is neat because both vim and terminal get control of stdin and it creates all sorts of mess or javascript use strict var fs require fs var child process require child process var file tmp node editor test fs writefile file node editor function var child child process spawn vim detached true stdio inherit child unref both die version also tried in or so platform darwin brunchfast local darwin kernel version thu sep pdt root xnu release subsystem child process
1
14,252
17,188,390,574
IssuesEvent
2021-07-16 07:23:10
medic/cht-core
https://api.github.com/repos/medic/cht-core
opened
Release 3.10.5
Type: Internal process
# Planning - [ ] Create an GH Milestone and add this issue to it. - [ ] Add all the issues to be worked on to the Milestone. # Development When development is ready to begin one of the engineers should be nominated as a Release Manager. They will be responsible for making sure the following tasks are completed though not necessarily completing them. - [ ] Set the version number in `package.json` and `package-lock.json` and submit a PR to the release branch. The easiest way to do this is to use `npm --no-git-tag-version version patch`. - [ ] Write an update in the weekly Product Team call agenda summarising development and acceptance testing progress and identifying any blockers. The release manager is to update this every week until the version is released. # Releasing Once all issues have passed acceptance testing and have been merged into `master` and backported to the release branch release testing can begin. - [ ] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing. - [ ] Create a new document in the [release-notes folder](https://github.com/medic/cht-core/tree/master/release-notes) in `master`. Ensure all issues are in the GH Milestone, that they're correct labelled, and have human readable descriptions. Use [this script](https://github.com/medic/cht-core/blob/master/scripts/release-notes/) to export the issues into our release note format. Manually document any known migration steps and known issues. - [ ] Until release testing passes, make sure regressions are fixed in `master`, cherry-pick them into the release branch, and release another beta. - [ ] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/cht-core/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Link to the release notes in the description of the release. - [ ] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>` - [ ] Announce the release in #products using this template: ``` @channel *Announcing the release of {{version}}* This release fixes {{number of bugs}}. Read the release notes for full details: {{url}} ``` - [ ] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category. You can use the previous message and omit `@channel`. - [ ] Mark this issue "done" and close the Milestone.
1.0
Release 3.10.5 - # Planning - [ ] Create an GH Milestone and add this issue to it. - [ ] Add all the issues to be worked on to the Milestone. # Development When development is ready to begin one of the engineers should be nominated as a Release Manager. They will be responsible for making sure the following tasks are completed though not necessarily completing them. - [ ] Set the version number in `package.json` and `package-lock.json` and submit a PR to the release branch. The easiest way to do this is to use `npm --no-git-tag-version version patch`. - [ ] Write an update in the weekly Product Team call agenda summarising development and acceptance testing progress and identifying any blockers. The release manager is to update this every week until the version is released. # Releasing Once all issues have passed acceptance testing and have been merged into `master` and backported to the release branch release testing can begin. - [ ] Build a beta named `<major>.<minor>.<patch>-beta.1` by pushing a git tag and when CI completes successfully notify the QA team that it's ready for release testing. - [ ] Create a new document in the [release-notes folder](https://github.com/medic/cht-core/tree/master/release-notes) in `master`. Ensure all issues are in the GH Milestone, that they're correct labelled, and have human readable descriptions. Use [this script](https://github.com/medic/cht-core/blob/master/scripts/release-notes/) to export the issues into our release note format. Manually document any known migration steps and known issues. - [ ] Until release testing passes, make sure regressions are fixed in `master`, cherry-pick them into the release branch, and release another beta. - [ ] Create a release in GitHub from the release branch so it shows up under the [Releases tab](https://github.com/medic/cht-core/releases) with the naming convention `<major>.<minor>.<patch>`. This will create the git tag automatically. Link to the release notes in the description of the release. - [ ] Confirm the release build completes successfully and the new release is available on the [market](https://staging.dev.medicmobile.org/builds/releases). Make sure that the document has new entry with `id: medic:medic:<major>.<minor>.<patch>` - [ ] Announce the release in #products using this template: ``` @channel *Announcing the release of {{version}}* This release fixes {{number of bugs}}. Read the release notes for full details: {{url}} ``` - [ ] Announce the release on the [CHT forum](https://forum.communityhealthtoolkit.org/), under the "Product - Releases" category. You can use the previous message and omit `@channel`. - [ ] Mark this issue "done" and close the Milestone.
process
release planning create an gh milestone and add this issue to it add all the issues to be worked on to the milestone development when development is ready to begin one of the engineers should be nominated as a release manager they will be responsible for making sure the following tasks are completed though not necessarily completing them set the version number in package json and package lock json and submit a pr to the release branch the easiest way to do this is to use npm no git tag version version patch write an update in the weekly product team call agenda summarising development and acceptance testing progress and identifying any blockers the release manager is to update this every week until the version is released releasing once all issues have passed acceptance testing and have been merged into master and backported to the release branch release testing can begin build a beta named beta by pushing a git tag and when ci completes successfully notify the qa team that it s ready for release testing create a new document in the in master ensure all issues are in the gh milestone that they re correct labelled and have human readable descriptions use to export the issues into our release note format manually document any known migration steps and known issues until release testing passes make sure regressions are fixed in master cherry pick them into the release branch and release another beta create a release in github from the release branch so it shows up under the with the naming convention this will create the git tag automatically link to the release notes in the description of the release confirm the release build completes successfully and the new release is available on the make sure that the document has new entry with id medic medic announce the release in products using this template channel announcing the release of version this release fixes number of bugs read the release notes for full details url announce the release on the under the product releases category you can use the previous message and omit channel mark this issue done and close the milestone
1
1,342
3,901,423,010
IssuesEvent
2016-04-18 10:47:23
DevExpress/testcafe-hammerhead
https://api.github.com/repos/DevExpress/testcafe-hammerhead
closed
PageProcessor parse binaries as a html page
AREA: server SYSTEM: resource processing TYPE: bug
Reproduces in the playground when try to download a file. For example reproduced on the http://testcafe.devexpress.com page Click `Download TestCafe` button. We have the following `Content-Type` in response: `application/octet-stream` But `RequestPipelineContext.isPage` is `true` because it depends on request headers ([see here](https://github.com/DevExpress/testcafe-hammerhead/blob/master/src/request-pipeline/context.js#L55)) As a result `PageProcessor.shouldProcessResource(ctx)` returns `true` and parse response via `parse5` module ([see here](https://github.com/DevExpress/testcafe-hammerhead/blob/master/src/processing/resources/page.js#L129))
1.0
PageProcessor parse binaries as a html page - Reproduces in the playground when try to download a file. For example reproduced on the http://testcafe.devexpress.com page Click `Download TestCafe` button. We have the following `Content-Type` in response: `application/octet-stream` But `RequestPipelineContext.isPage` is `true` because it depends on request headers ([see here](https://github.com/DevExpress/testcafe-hammerhead/blob/master/src/request-pipeline/context.js#L55)) As a result `PageProcessor.shouldProcessResource(ctx)` returns `true` and parse response via `parse5` module ([see here](https://github.com/DevExpress/testcafe-hammerhead/blob/master/src/processing/resources/page.js#L129))
process
pageprocessor parse binaries as a html page reproduces in the playground when try to download a file for example reproduced on the page click download testcafe button we have the following content type in response application octet stream but requestpipelinecontext ispage is true because it depends on request headers as a result pageprocessor shouldprocessresource ctx returns true and parse response via module
1
20,916
27,754,022,465
IssuesEvent
2023-03-15 23:55:36
dDevTech/tapas-top-frontend
https://api.github.com/repos/dDevTech/tapas-top-frontend
closed
Crear pagina register-account-info 20/03/2023
pending in process require testing
Crear nuevo package en webapp/app/modules/account nombrado como register-account-info En el se crearán los archivos necesarios para: Crear pagina nueva account-info donde se podrán introducir los campos: - Nombre (requerido) - Apellido 1 text (opcional) - Apellido 2 text (opcional) - Adjuntar foto como archivo blob (opcional) > Debe conventirse a una foto cuadrada de 128px - Breve introduccion Multitext (opcional) - País [categoría] - Ubicación [categoría] OPCIONAL: Seleccionar ubicación con api de google maps Se añadirá barra de progreso Se añadirá un botón para crear la cuenta ![image](https://user-images.githubusercontent.com/18512841/224556953-791fd7c3-80f3-4b8c-a759-0cdc20ef3eae.png)
1.0
Crear pagina register-account-info 20/03/2023 - Crear nuevo package en webapp/app/modules/account nombrado como register-account-info En el se crearán los archivos necesarios para: Crear pagina nueva account-info donde se podrán introducir los campos: - Nombre (requerido) - Apellido 1 text (opcional) - Apellido 2 text (opcional) - Adjuntar foto como archivo blob (opcional) > Debe conventirse a una foto cuadrada de 128px - Breve introduccion Multitext (opcional) - País [categoría] - Ubicación [categoría] OPCIONAL: Seleccionar ubicación con api de google maps Se añadirá barra de progreso Se añadirá un botón para crear la cuenta ![image](https://user-images.githubusercontent.com/18512841/224556953-791fd7c3-80f3-4b8c-a759-0cdc20ef3eae.png)
process
crear pagina register account info crear nuevo package en webapp app modules account nombrado como register account info en el se crearán los archivos necesarios para crear pagina nueva account info donde se podrán introducir los campos nombre requerido apellido text opcional apellido text opcional adjuntar foto como archivo blob opcional debe conventirse a una foto cuadrada de breve introduccion multitext opcional país ubicación opcional seleccionar ubicación con api de google maps se añadirá barra de progreso se añadirá un botón para crear la cuenta
1
589,116
17,690,020,340
IssuesEvent
2021-08-24 08:48:24
magento/magento2
https://api.github.com/repos/magento/magento2
closed
Refactor codebase to fix problem of reserved keyword "match" for PHP 8
Progress: PR in progress Priority: P2 Project: PHP8 Project: Platform Health
### Description (*) Magento has places where use keyword "match" that is reserved PHP from version 8 according to [this document](https://www.php.net/manual/en/reserved.keywords.php). These is classes provided below app/code/Magento/CustomerSegment/Controller/Adminhtml/Index/Match.php app/code/Magento/Elasticsearch/SearchAdapter/Query/Builder/Match.php lib/internal/Magento/Framework/Search/Request/Query/Match.php Using this keyword as class name makes unpossible compatible with PHP 8. We need eliminate this problem.
1.0
Refactor codebase to fix problem of reserved keyword "match" for PHP 8 - ### Description (*) Magento has places where use keyword "match" that is reserved PHP from version 8 according to [this document](https://www.php.net/manual/en/reserved.keywords.php). These is classes provided below app/code/Magento/CustomerSegment/Controller/Adminhtml/Index/Match.php app/code/Magento/Elasticsearch/SearchAdapter/Query/Builder/Match.php lib/internal/Magento/Framework/Search/Request/Query/Match.php Using this keyword as class name makes unpossible compatible with PHP 8. We need eliminate this problem.
non_process
refactor codebase to fix problem of reserved keyword match for php description magento has places where use keyword match that is reserved php from version according to these is classes provided below app code magento customersegment controller adminhtml index match php app code magento elasticsearch searchadapter query builder match php lib internal magento framework search request query match php using this keyword as class name makes unpossible compatible with php we need eliminate this problem
0
225,790
24,881,366,618
IssuesEvent
2022-10-28 01:37:40
praneethpanasala/linux
https://api.github.com/repos/praneethpanasala/linux
opened
CVE-2022-3621 (High) detected in linuxlinux-4.19.6
security vulnerability
## CVE-2022-3621 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in Linux Kernel. It has been classified as problematic. Affected is the function nilfs_bmap_lookup_at_level of the file fs/nilfs2/inode.c of the component nilfs2. The manipulation leads to null pointer dereference. It is possible to launch the attack remotely. It is recommended to apply a patch to fix this issue. The identifier of this vulnerability is VDB-211920. <p>Publish Date: 2022-10-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3621>CVE-2022-3621</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3621">https://www.linuxkernelcves.com/cves/CVE-2022-3621</a></p> <p>Release Date: 2022-10-20</p> <p>Fix Resolution: v5.4.218,v5.10.148,v5.15.74,v6.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-3621 (High) detected in linuxlinux-4.19.6 - ## CVE-2022-3621 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.19.6</b></p></summary> <p> <p>Apache Software Foundation (ASF)</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A vulnerability was found in Linux Kernel. It has been classified as problematic. Affected is the function nilfs_bmap_lookup_at_level of the file fs/nilfs2/inode.c of the component nilfs2. The manipulation leads to null pointer dereference. It is possible to launch the attack remotely. It is recommended to apply a patch to fix this issue. The identifier of this vulnerability is VDB-211920. <p>Publish Date: 2022-10-20 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-3621>CVE-2022-3621</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2022-3621">https://www.linuxkernelcves.com/cves/CVE-2022-3621</a></p> <p>Release Date: 2022-10-20</p> <p>Fix Resolution: v5.4.218,v5.10.148,v5.15.74,v6.0.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in linuxlinux cve high severity vulnerability vulnerable library linuxlinux apache software foundation asf library home page a href found in base branch master vulnerable source files vulnerability details a vulnerability was found in linux kernel it has been classified as problematic affected is the function nilfs bmap lookup at level of the file fs inode c of the component the manipulation leads to null pointer dereference it is possible to launch the attack remotely it is recommended to apply a patch to fix this issue the identifier of this vulnerability is vdb publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
92,369
26,666,790,146
IssuesEvent
2023-01-26 05:22:37
runatlantis/atlantis
https://api.github.com/repos/runatlantis/atlantis
closed
Add permissions to all github actions
build github-actions
<!--- Please keep this note for the community ---> ### Community Note - Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request. Searching for pre-existing feature requests helps us consolidate datapoints for identical requirements into a single place, thank you! - Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request. - If you are interested in working on this issue or have submitted a pull request, please leave a comment. <!--- Thank you for keeping this note for the community ---> --- - [ ] I'd be willing to implement this feature ([contributing guide](https://github.com/runatlantis/atlantis/blob/main/CONTRIBUTING.md)) **Describe the user story** <!-- A clear and concise description of what workflow is meant to be improved. Example: "As a developer, I often want to do <something>, but I often face <problem>". --> Keep things secure by adding necessary permissions. This can be used to generate the permissions. https://app.stepsecurity.io/secureworkflow/runatlantis/atlantis
1.0
Add permissions to all github actions - <!--- Please keep this note for the community ---> ### Community Note - Please vote on this issue by adding a 👍 [reaction](https://blog.github.com/2016-03-10-add-reactions-to-pull-requests-issues-and-comments/) to the original issue to help the community and maintainers prioritize this request. Searching for pre-existing feature requests helps us consolidate datapoints for identical requirements into a single place, thank you! - Please do not leave "+1" or other comments that do not add relevant new information or questions, they generate extra noise for issue followers and do not help prioritize the request. - If you are interested in working on this issue or have submitted a pull request, please leave a comment. <!--- Thank you for keeping this note for the community ---> --- - [ ] I'd be willing to implement this feature ([contributing guide](https://github.com/runatlantis/atlantis/blob/main/CONTRIBUTING.md)) **Describe the user story** <!-- A clear and concise description of what workflow is meant to be improved. Example: "As a developer, I often want to do <something>, but I often face <problem>". --> Keep things secure by adding necessary permissions. This can be used to generate the permissions. https://app.stepsecurity.io/secureworkflow/runatlantis/atlantis
non_process
add permissions to all github actions community note please vote on this issue by adding a 👍 to the original issue to help the community and maintainers prioritize this request searching for pre existing feature requests helps us consolidate datapoints for identical requirements into a single place thank you please do not leave or other comments that do not add relevant new information or questions they generate extra noise for issue followers and do not help prioritize the request if you are interested in working on this issue or have submitted a pull request please leave a comment i d be willing to implement this feature describe the user story a clear and concise description of what workflow is meant to be improved example as a developer i often want to do but i often face keep things secure by adding necessary permissions this can be used to generate the permissions
0
126,756
12,298,732,535
IssuesEvent
2020-05-11 11:03:18
BA-HanseML/NF_Prj_MIMII_Dataset
https://api.github.com/repos/BA-HanseML/NF_Prj_MIMII_Dataset
closed
MEL and PSD grid picture cards
documentation mi
a nice high res pcture with PSD and MEL per MAchine part and ID machine part ID normal ,also normal, abnormal, alose abnorma PSD MEL plus mp 3
1.0
MEL and PSD grid picture cards - a nice high res pcture with PSD and MEL per MAchine part and ID machine part ID normal ,also normal, abnormal, alose abnorma PSD MEL plus mp 3
non_process
mel and psd grid picture cards a nice high res pcture with psd and mel per machine part and id machine part id normal also normal abnormal alose abnorma psd mel plus mp
0
18,956
24,918,406,616
IssuesEvent
2022-10-30 17:23:47
aiidateam/aiida-core
https://api.github.com/repos/aiidateam/aiida-core
closed
`get_function_source_code` gives `FileNotFoundError` when `calcfunction` is not in a file.
type/bug priority/nice-to-have topic/processes
### Steps to reproduce In `verdi shell`: ```python from aiida.engine import calcfunction @calcfunction def my_func(inp): return Int(4) res=my_func(Int(1)) res.creator.get_function_source_code() ``` ### Expected behavior Not sure what is the behavior we want here...because the source code is not stored anywhere I believe. The minimal fix would be to follow what the doc-string says: `returns: the absolute path of the source file in the repository, or None if it does not exist` i.e. implement to return `None`. ### Your environment - Operating system [e.g. Linux]: Linux - Python version [e.g. 3.7.1]: 3.9.13 - aiida-core version [e.g. 1.2.1]: 2.0.1 ### Additional context Of course, in production, no-one would define `calcfunctions` in the shell, but many basic tutorials show `calcfunctions` defined in the shell. Not nice to see errors there.
1.0
`get_function_source_code` gives `FileNotFoundError` when `calcfunction` is not in a file. - ### Steps to reproduce In `verdi shell`: ```python from aiida.engine import calcfunction @calcfunction def my_func(inp): return Int(4) res=my_func(Int(1)) res.creator.get_function_source_code() ``` ### Expected behavior Not sure what is the behavior we want here...because the source code is not stored anywhere I believe. The minimal fix would be to follow what the doc-string says: `returns: the absolute path of the source file in the repository, or None if it does not exist` i.e. implement to return `None`. ### Your environment - Operating system [e.g. Linux]: Linux - Python version [e.g. 3.7.1]: 3.9.13 - aiida-core version [e.g. 1.2.1]: 2.0.1 ### Additional context Of course, in production, no-one would define `calcfunctions` in the shell, but many basic tutorials show `calcfunctions` defined in the shell. Not nice to see errors there.
process
get function source code gives filenotfounderror when calcfunction is not in a file steps to reproduce in verdi shell python from aiida engine import calcfunction calcfunction def my func inp return int res my func int res creator get function source code expected behavior not sure what is the behavior we want here because the source code is not stored anywhere i believe the minimal fix would be to follow what the doc string says returns the absolute path of the source file in the repository or none if it does not exist i e implement to return none your environment operating system linux python version aiida core version additional context of course in production no one would define calcfunctions in the shell but many basic tutorials show calcfunctions defined in the shell not nice to see errors there
1
13,125
15,526,203,480
IssuesEvent
2021-03-13 00:25:28
googleapis/python-binary-authorization
https://api.github.com/repos/googleapis/python-binary-authorization
opened
Add v1
type: process
https://github.com/googleapis/googleapis/tree/master/google/cloud/binaryauthorization There are no v1 protos in googleapis/googleapis. Probably need to ping the API team to sync v1 protos.
1.0
Add v1 - https://github.com/googleapis/googleapis/tree/master/google/cloud/binaryauthorization There are no v1 protos in googleapis/googleapis. Probably need to ping the API team to sync v1 protos.
process
add there are no protos in googleapis googleapis probably need to ping the api team to sync protos
1
36,297
17,604,381,813
IssuesEvent
2021-08-17 15:18:58
smith-chem-wisc/Spritz
https://api.github.com/repos/smith-chem-wisc/Spritz
closed
gatk MarkDuplicates Exception in thread "main" java.lang.OutOfMemoryError: GC overhead limit exceeded
Question Performance
Dear Spritz developers, thank you very much for your amazing work. We tried to run Spritz for 12 samples on a local machine with 40G RAM and many cores, but computation stopped with GATK MarkDuplicates command. `INFO 2021-07-20 06:16:14 MarkDuplicates Tracking 65403162 as yet unmatched pairs. 65358276 records in RAM. [Tue Jul 20 08:46:37 CEST 2021] picard.sam.markduplicates.MarkDuplicates done. Elapsed time: 989.15 minutes. Runtime.totalMemory()=30150754304 To get help, see http://broadinstitute.github.io/picard/index.html#GettingHelp Exception in thread "main" java.lang.OutOfMemoryError: GC overhead limit exceeded at java.util.HashMap.newNode(HashMap.java:1750) at java.util.HashMap.putVal(HashMap.java:631) at java.util.HashMap.put(HashMap.java:612) at htsjdk.samtools.SAMRecord.setTransientAttribute(SAMRecord.java:2318) at htsjdk.samtools.DuplicateScoringStrategy.computeDuplicateScore(DuplicateScoringStrategy.java:112) at htsjdk.samtools.DuplicateScoringStrategy.computeDuplicateScore(DuplicateScoringStrategy.java:62) at picard.sam.markduplicates.MarkDuplicates.buildReadEnds(MarkDuplicates.java:650) at picard.sam.markduplicates.MarkDuplicates.buildSortedReadEndLists(MarkDuplicates.java:552) at picard.sam.markduplicates.MarkDuplicates.doWork(MarkDuplicates.java:257) at picard.cmdline.CommandLineProgram.instanceMain(CommandLineProgram.java:301) at org.broadinstitute.hellbender.cmdline.PicardCommandLineProgramExecutor.instanceMain(PicardCommandLineProgramExecutor.java:37) at org.broadinstitute.hellbender.Main.runCommandLineProgram(Main.java:160) at org.broadinstitute.hellbender.Main.mainEntry(Main.java:203) at org.broadinstitute.hellbender.Main.main(Main.java:289) Using GATK jar /mnt/bin/miniconda3/envs/spritzenv/share/gatk4-4.1.9.0-0/gatk-package-4.1.9.0-local.jar ` We tried to figure out what the problem could be and we noticed that we were outofmemory in correspondence of mitochondrial genome due to the large number of mapped reads. The output of samtools idxstats 1 248956422 151486569 0 2 242193529 129975569 0 3 198295559 121739113 0 4 190214555 80953331 0 5 181538259 96426281 0 6 170805979 94787913 0 7 159345973 79069305 0 8 145138636 64688331 0 9 138394717 85252265 0 10 133797422 64796009 0 11 135086622 88895141 0 12 133275309 73509958 0 13 114364328 52786721 0 14 107043718 134493736 0 15 101991189 52094551 0 16 90338345 44675919 0 17 83257441 54111496 0 18 80373285 32061298 0 19 58617616 32868629 0 20 64444167 30170370 0 21 46709983 43678658 0 22 50818468 18853590 0 X 156040895 50604944 0 Y 57227415 1872438 0 MT 16569 844752497 0 Our memory is bounded by 32G RAM and default value for --MAX_RECORDS_IN_RAM argument. Is there any solution for this issue? Thank you very much for your support
True
gatk MarkDuplicates Exception in thread "main" java.lang.OutOfMemoryError: GC overhead limit exceeded - Dear Spritz developers, thank you very much for your amazing work. We tried to run Spritz for 12 samples on a local machine with 40G RAM and many cores, but computation stopped with GATK MarkDuplicates command. `INFO 2021-07-20 06:16:14 MarkDuplicates Tracking 65403162 as yet unmatched pairs. 65358276 records in RAM. [Tue Jul 20 08:46:37 CEST 2021] picard.sam.markduplicates.MarkDuplicates done. Elapsed time: 989.15 minutes. Runtime.totalMemory()=30150754304 To get help, see http://broadinstitute.github.io/picard/index.html#GettingHelp Exception in thread "main" java.lang.OutOfMemoryError: GC overhead limit exceeded at java.util.HashMap.newNode(HashMap.java:1750) at java.util.HashMap.putVal(HashMap.java:631) at java.util.HashMap.put(HashMap.java:612) at htsjdk.samtools.SAMRecord.setTransientAttribute(SAMRecord.java:2318) at htsjdk.samtools.DuplicateScoringStrategy.computeDuplicateScore(DuplicateScoringStrategy.java:112) at htsjdk.samtools.DuplicateScoringStrategy.computeDuplicateScore(DuplicateScoringStrategy.java:62) at picard.sam.markduplicates.MarkDuplicates.buildReadEnds(MarkDuplicates.java:650) at picard.sam.markduplicates.MarkDuplicates.buildSortedReadEndLists(MarkDuplicates.java:552) at picard.sam.markduplicates.MarkDuplicates.doWork(MarkDuplicates.java:257) at picard.cmdline.CommandLineProgram.instanceMain(CommandLineProgram.java:301) at org.broadinstitute.hellbender.cmdline.PicardCommandLineProgramExecutor.instanceMain(PicardCommandLineProgramExecutor.java:37) at org.broadinstitute.hellbender.Main.runCommandLineProgram(Main.java:160) at org.broadinstitute.hellbender.Main.mainEntry(Main.java:203) at org.broadinstitute.hellbender.Main.main(Main.java:289) Using GATK jar /mnt/bin/miniconda3/envs/spritzenv/share/gatk4-4.1.9.0-0/gatk-package-4.1.9.0-local.jar ` We tried to figure out what the problem could be and we noticed that we were outofmemory in correspondence of mitochondrial genome due to the large number of mapped reads. The output of samtools idxstats 1 248956422 151486569 0 2 242193529 129975569 0 3 198295559 121739113 0 4 190214555 80953331 0 5 181538259 96426281 0 6 170805979 94787913 0 7 159345973 79069305 0 8 145138636 64688331 0 9 138394717 85252265 0 10 133797422 64796009 0 11 135086622 88895141 0 12 133275309 73509958 0 13 114364328 52786721 0 14 107043718 134493736 0 15 101991189 52094551 0 16 90338345 44675919 0 17 83257441 54111496 0 18 80373285 32061298 0 19 58617616 32868629 0 20 64444167 30170370 0 21 46709983 43678658 0 22 50818468 18853590 0 X 156040895 50604944 0 Y 57227415 1872438 0 MT 16569 844752497 0 Our memory is bounded by 32G RAM and default value for --MAX_RECORDS_IN_RAM argument. Is there any solution for this issue? Thank you very much for your support
non_process
gatk markduplicates exception in thread main java lang outofmemoryerror gc overhead limit exceeded dear spritz developers thank you very much for your amazing work we tried to run spritz for samples on a local machine with ram and many cores but computation stopped with gatk markduplicates command info markduplicates tracking as yet unmatched pairs records in ram picard sam markduplicates markduplicates done elapsed time minutes runtime totalmemory to get help see exception in thread main java lang outofmemoryerror gc overhead limit exceeded at java util hashmap newnode hashmap java at java util hashmap putval hashmap java at java util hashmap put hashmap java at htsjdk samtools samrecord settransientattribute samrecord java at htsjdk samtools duplicatescoringstrategy computeduplicatescore duplicatescoringstrategy java at htsjdk samtools duplicatescoringstrategy computeduplicatescore duplicatescoringstrategy java at picard sam markduplicates markduplicates buildreadends markduplicates java at picard sam markduplicates markduplicates buildsortedreadendlists markduplicates java at picard sam markduplicates markduplicates dowork markduplicates java at picard cmdline commandlineprogram instancemain commandlineprogram java at org broadinstitute hellbender cmdline picardcommandlineprogramexecutor instancemain picardcommandlineprogramexecutor java at org broadinstitute hellbender main runcommandlineprogram main java at org broadinstitute hellbender main mainentry main java at org broadinstitute hellbender main main main java using gatk jar mnt bin envs spritzenv share gatk package local jar we tried to figure out what the problem could be and we noticed that we were outofmemory in correspondence of mitochondrial genome due to the large number of mapped reads the output of samtools idxstats x y mt our memory is bounded by ram and default value for max records in ram argument is there any solution for this issue thank you very much for your support
0
255,323
21,918,159,229
IssuesEvent
2022-05-22 06:24:08
MohistMC/Mohist
https://api.github.com/repos/MohistMC/Mohist
closed
New version not working on linux
1.16.5 Wait Needs Testing
<!-- ISSUE_TEMPLATE_3 -> IMPORTANT: DO NOT DELETE THIS LINE.--> <!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.--> <!-- If you don't know where to upload your logs and crash reports, you can use these websites : --> <!-- https://gist.github.com (recommended) --> <!-- https://mclo.gs --> <!-- https://haste.mohistmc.com --> <!-- https://pastebin.com --> <!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT --> **Minecraft Version :** 1.16.5 **Mohist Version :** 996 **Operating System :** Linux centos **Logs :** https://haste.mohistmc.com/sumiyejugi **Mod list :** none **Plugin list :** none **Description of issue :** download mohist jar from website, upload to centos server, use java -jar mohist.jar, server starts downloading libraries etc. and then crash with the logs on top. this build works on windows
1.0
New version not working on linux - <!-- ISSUE_TEMPLATE_3 -> IMPORTANT: DO NOT DELETE THIS LINE.--> <!-- Thank you for reporting ! Please note that issues can take a lot of time to be fixed and there is no eta.--> <!-- If you don't know where to upload your logs and crash reports, you can use these websites : --> <!-- https://gist.github.com (recommended) --> <!-- https://mclo.gs --> <!-- https://haste.mohistmc.com --> <!-- https://pastebin.com --> <!-- TO FILL THIS TEMPLATE, YOU NEED TO REPLACE THE {} BY WHAT YOU WANT --> **Minecraft Version :** 1.16.5 **Mohist Version :** 996 **Operating System :** Linux centos **Logs :** https://haste.mohistmc.com/sumiyejugi **Mod list :** none **Plugin list :** none **Description of issue :** download mohist jar from website, upload to centos server, use java -jar mohist.jar, server starts downloading libraries etc. and then crash with the logs on top. this build works on windows
non_process
new version not working on linux important do not delete this line minecraft version mohist version operating system linux centos logs mod list none plugin list none description of issue download mohist jar from website upload to centos server use java jar mohist jar server starts downloading libraries etc and then crash with the logs on top this build works on windows
0
325,282
27,862,746,811
IssuesEvent
2023-03-21 08:03:23
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix ndarray.test_numpy_instance_ior__
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4475563861/jobs/7865105584" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4475563861/jobs/7865105584" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4475563861/jobs/7865105584" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4476610484/jobs/7867147669" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
1.0
Fix ndarray.test_numpy_instance_ior__ - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4475563861/jobs/7865105584" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4475563861/jobs/7865105584" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4475563861/jobs/7865105584" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4476610484/jobs/7867147669" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a>
non_process
fix ndarray test numpy instance ior tensorflow img src torch img src numpy img src jax img src
0
10,641
13,446,164,910
IssuesEvent
2020-09-08 12:34:18
MHRA/products
https://api.github.com/repos/MHRA/products
closed
PARs - File upload on every edit
BUG :bug: EPIC - PARs process
### User want As a Medical Writer I would like my file to remain uploaded even after editing my upload details ### Acceptance Criteria **Customer acceptance criteria** - [ ] Document remains uploaded following edit of details - [ ] Document remains uploaded following deletion of another PARs document **Technical acceptance criteria** **Data acceptance criteria** **Testing acceptance criteria** **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
1.0
PARs - File upload on every edit - ### User want As a Medical Writer I would like my file to remain uploaded even after editing my upload details ### Acceptance Criteria **Customer acceptance criteria** - [ ] Document remains uploaded following edit of details - [ ] Document remains uploaded following deletion of another PARs document **Technical acceptance criteria** **Data acceptance criteria** **Testing acceptance criteria** **Size** **Value** **Effort** ### Exit Criteria met - [ ] Backlog - [ ] Discovery - [ ] DUXD - [ ] Development - [ ] Quality Assurance - [ ] Release and Validate
process
pars file upload on every edit user want as a medical writer i would like my file to remain uploaded even after editing my upload details acceptance criteria customer acceptance criteria document remains uploaded following edit of details document remains uploaded following deletion of another pars document technical acceptance criteria data acceptance criteria testing acceptance criteria size value effort exit criteria met backlog discovery duxd development quality assurance release and validate
1
1,782
4,513,786,188
IssuesEvent
2016-09-04 13:54:23
AnalyticalGraphicsInc/cesium
https://api.github.com/repos/AnalyticalGraphicsInc/cesium
opened
Run WebGL tests in CI
dev process
As discussed with @mramato offline: * Replace all read pixels expectations with a function that can have a no-op expectation when the tests are ran with a "no WebGL" flag, e.g., ```javascript expect(scene.renderForSpecs()).toEqual([0, 0, 0, 255]); ``` becomes ```javascript scene.expectRenderForSpecs([0, 0, 0, 255]); ``` * Replace the object returned by `getContext` with a mock object with GL functions that are no-ops, `function() {}`, except for `getExtension`, which should return mocked objects for the extensions we care about. * Likewise, all `gl.get*` functions should be mocked to return reasonable values. * Run the tests and fix things I forgot. This should only take a few hours and will be more reliable than [mesa](https://github.com/AnalyticalGraphicsInc/cesium/compare/mesa).
1.0
Run WebGL tests in CI - As discussed with @mramato offline: * Replace all read pixels expectations with a function that can have a no-op expectation when the tests are ran with a "no WebGL" flag, e.g., ```javascript expect(scene.renderForSpecs()).toEqual([0, 0, 0, 255]); ``` becomes ```javascript scene.expectRenderForSpecs([0, 0, 0, 255]); ``` * Replace the object returned by `getContext` with a mock object with GL functions that are no-ops, `function() {}`, except for `getExtension`, which should return mocked objects for the extensions we care about. * Likewise, all `gl.get*` functions should be mocked to return reasonable values. * Run the tests and fix things I forgot. This should only take a few hours and will be more reliable than [mesa](https://github.com/AnalyticalGraphicsInc/cesium/compare/mesa).
process
run webgl tests in ci as discussed with mramato offline replace all read pixels expectations with a function that can have a no op expectation when the tests are ran with a no webgl flag e g javascript expect scene renderforspecs toequal becomes javascript scene expectrenderforspecs replace the object returned by getcontext with a mock object with gl functions that are no ops function except for getextension which should return mocked objects for the extensions we care about likewise all gl get functions should be mocked to return reasonable values run the tests and fix things i forgot this should only take a few hours and will be more reliable than
1
35,856
12,393,588,171
IssuesEvent
2020-05-20 15:38:16
wallanpsantos/apache-camel-alura
https://api.github.com/repos/wallanpsantos/apache-camel-alura
opened
CVE-2009-2625 (Medium) detected in xercesImpl-2.8.0.jar
security vulnerability
## CVE-2009-2625 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.8.0.jar</b></p></summary> <p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI), a complete framework for building parser components and configurations that is extremely modular and easy to program.</p> <p>Path to dependency file: /tmp/ws-scm/apache-camel-alura/camel-alura/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/xerces/xercesImpl/2.8.0/xercesImpl-2.8.0.jar</p> <p> Dependency Hierarchy: - xom-1.2.5.jar (Root Library) - :x: **xercesImpl-2.8.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/wallanpsantos/apache-camel-alura/commit/1c082812a144fc7b9f16c14f56f305c26ef10ca5">1c082812a144fc7b9f16c14f56f305c26ef10ca5</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XMLScanner.java in Apache Xerces2 Java, as used in Sun Java Runtime Environment (JRE) in JDK and JRE 6 before Update 15 and JDK and JRE 5.0 before Update 20, and in other products, allows remote attackers to cause a denial of service (infinite loop and application hang) via malformed XML input, as demonstrated by the Codenomicon XML fuzzing framework. <p>Publish Date: 2009-08-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2009-2625>CVE-2009-2625</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://www.securitytracker.com/id?1022680">http://www.securitytracker.com/id?1022680</a></p> <p>Release Date: 2017-12-31</p> <p>Fix Resolution: The vendor has issued a fix for Windows, Solaris, and Linux: * JDK and JRE 6 Update 15 or later * JDK and JRE 5.0 Update 20 or later Java SE releases are available at: JDK and JRE 6 Update 15: http://java.sun.com/javase/downloads/index.jsp JRE 6 Update 15: http://java.com/ through the Java Update tool for Microsoft Windows users. JDK 6 Update 15 for Solaris is available in the following patches: * Java SE 6 Update 15 (as delivered in patch 125136-16) * Java SE 6 Update 15 (as delivered in patch 125137-16 (64bit)) * Java SE 6_x86 Update 15 (as delivered in patch 125138-16) * Java SE 6_x86 Update 15 (as delivered in patch 125139-16 (64bit)) JDK and JRE 5.0 Update 20: http://java.sun.com/javase/downloads/index_jdk5.jsp JDK 5.0 Update 20 for Solaris is available in the following patches: * J2SE 5.0 Update 18 (as delivered in patch 118666-21) * J2SE 5.0 Update 18 (as delivered in patch 118667-21 (64bit)) * J2SE 5.0_x86 Update 18 (as delivered in patch 118668-21) * J2SE 5.0_x86 Update 18 (as delivered in patch 118669-21 (64bit)) Java SE for Business releases are available at: http://www.sun.com/software/javaseforbusiness/getit_download.jsp Note: When installing a new version of the product from a source other than a Solaris patch, it is recommended that the old affected versions be removed from your system. To remove old affected versions on the Windows platform, please see: http://www.java.com/en/download/help/5000010800.xml The vendor's advisory is available at: http://sunsolve.sun.com/search/document.do?assetkey=1-66-263489-1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2009-2625 (Medium) detected in xercesImpl-2.8.0.jar - ## CVE-2009-2625 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>xercesImpl-2.8.0.jar</b></p></summary> <p>Xerces2 is the next generation of high performance, fully compliant XML parsers in the Apache Xerces family. This new version of Xerces introduces the Xerces Native Interface (XNI), a complete framework for building parser components and configurations that is extremely modular and easy to program.</p> <p>Path to dependency file: /tmp/ws-scm/apache-camel-alura/camel-alura/pom.xml</p> <p>Path to vulnerable library: /root/.m2/repository/xerces/xercesImpl/2.8.0/xercesImpl-2.8.0.jar</p> <p> Dependency Hierarchy: - xom-1.2.5.jar (Root Library) - :x: **xercesImpl-2.8.0.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/wallanpsantos/apache-camel-alura/commit/1c082812a144fc7b9f16c14f56f305c26ef10ca5">1c082812a144fc7b9f16c14f56f305c26ef10ca5</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> XMLScanner.java in Apache Xerces2 Java, as used in Sun Java Runtime Environment (JRE) in JDK and JRE 6 before Update 15 and JDK and JRE 5.0 before Update 20, and in other products, allows remote attackers to cause a denial of service (infinite loop and application hang) via malformed XML input, as demonstrated by the Codenomicon XML fuzzing framework. <p>Publish Date: 2009-08-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2009-2625>CVE-2009-2625</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 2 Score Details (<b>5.0</b>)</summary> <p> Base Score Metrics not available</p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="http://www.securitytracker.com/id?1022680">http://www.securitytracker.com/id?1022680</a></p> <p>Release Date: 2017-12-31</p> <p>Fix Resolution: The vendor has issued a fix for Windows, Solaris, and Linux: * JDK and JRE 6 Update 15 or later * JDK and JRE 5.0 Update 20 or later Java SE releases are available at: JDK and JRE 6 Update 15: http://java.sun.com/javase/downloads/index.jsp JRE 6 Update 15: http://java.com/ through the Java Update tool for Microsoft Windows users. JDK 6 Update 15 for Solaris is available in the following patches: * Java SE 6 Update 15 (as delivered in patch 125136-16) * Java SE 6 Update 15 (as delivered in patch 125137-16 (64bit)) * Java SE 6_x86 Update 15 (as delivered in patch 125138-16) * Java SE 6_x86 Update 15 (as delivered in patch 125139-16 (64bit)) JDK and JRE 5.0 Update 20: http://java.sun.com/javase/downloads/index_jdk5.jsp JDK 5.0 Update 20 for Solaris is available in the following patches: * J2SE 5.0 Update 18 (as delivered in patch 118666-21) * J2SE 5.0 Update 18 (as delivered in patch 118667-21 (64bit)) * J2SE 5.0_x86 Update 18 (as delivered in patch 118668-21) * J2SE 5.0_x86 Update 18 (as delivered in patch 118669-21 (64bit)) Java SE for Business releases are available at: http://www.sun.com/software/javaseforbusiness/getit_download.jsp Note: When installing a new version of the product from a source other than a Solaris patch, it is recommended that the old affected versions be removed from your system. To remove old affected versions on the Windows platform, please see: http://www.java.com/en/download/help/5000010800.xml The vendor's advisory is available at: http://sunsolve.sun.com/search/document.do?assetkey=1-66-263489-1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve medium detected in xercesimpl jar cve medium severity vulnerability vulnerable library xercesimpl jar is the next generation of high performance fully compliant xml parsers in the apache xerces family this new version of xerces introduces the xerces native interface xni a complete framework for building parser components and configurations that is extremely modular and easy to program path to dependency file tmp ws scm apache camel alura camel alura pom xml path to vulnerable library root repository xerces xercesimpl xercesimpl jar dependency hierarchy xom jar root library x xercesimpl jar vulnerable library found in head commit a href vulnerability details xmlscanner java in apache java as used in sun java runtime environment jre in jdk and jre before update and jdk and jre before update and in other products allows remote attackers to cause a denial of service infinite loop and application hang via malformed xml input as demonstrated by the codenomicon xml fuzzing framework publish date url a href cvss score details base score metrics not available suggested fix type upgrade version origin a href release date fix resolution the vendor has issued a fix for windows solaris and linux jdk and jre update or later jdk and jre update or later java se releases are available at jdk and jre update jre update through the java update tool for microsoft windows users jdk update for solaris is available in the following patches java se update as delivered in patch java se update as delivered in patch java se update as delivered in patch java se update as delivered in patch jdk and jre update jdk update for solaris is available in the following patches update as delivered in patch update as delivered in patch update as delivered in patch update as delivered in patch java se for business releases are available at note when installing a new version of the product from a source other than a solaris patch it is recommended that the old affected versions be removed from your system to remove old affected versions on the windows platform please see the vendor s advisory is available at step up your open source security game with whitesource
0
462
2,902,653,419
IssuesEvent
2015-06-18 08:34:56
haskell-distributed/distributed-process
https://api.github.com/repos/haskell-distributed/distributed-process
closed
The SimpleLocalnet backend should take (RemoteTable -> RemoteTable) rather than RemoteTable
Bug distributed-process-simplelocalnet
This avoids the need for Control.Distributed.Process.Node in applications that use it.
1.0
The SimpleLocalnet backend should take (RemoteTable -> RemoteTable) rather than RemoteTable - This avoids the need for Control.Distributed.Process.Node in applications that use it.
process
the simplelocalnet backend should take remotetable remotetable rather than remotetable this avoids the need for control distributed process node in applications that use it
1
19,382
25,518,987,817
IssuesEvent
2022-11-28 18:45:18
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Networking Tab is missing
automation/svc triaged assigned-to-author doc-bug process-automation/subsvc Pri2
Networking Tab is missing in documentation. ![image](https://user-images.githubusercontent.com/110971297/183857999-f9d3b633-7636-4a7b-a7ed-d2fa343a2eff.png) There is no explanation/guideline which option should be selected and their further impacts. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9b4440e0-1ff5-0fd3-6983-d5f6ed86e818 * Version Independent ID: 8d6aecae-1a58-83aa-45f7-306fb6c92d38 * Content: [Create a standalone Azure Automation account](https://docs.microsoft.com/en-us/azure/automation/automation-create-standalone-account?tabs=azureportal#feedback) * Content Source: [articles/automation/automation-create-standalone-account.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-create-standalone-account.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SnehaSudhirG * Microsoft Alias: **sudhirsneha**
1.0
Networking Tab is missing - Networking Tab is missing in documentation. ![image](https://user-images.githubusercontent.com/110971297/183857999-f9d3b633-7636-4a7b-a7ed-d2fa343a2eff.png) There is no explanation/guideline which option should be selected and their further impacts. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 9b4440e0-1ff5-0fd3-6983-d5f6ed86e818 * Version Independent ID: 8d6aecae-1a58-83aa-45f7-306fb6c92d38 * Content: [Create a standalone Azure Automation account](https://docs.microsoft.com/en-us/azure/automation/automation-create-standalone-account?tabs=azureportal#feedback) * Content Source: [articles/automation/automation-create-standalone-account.md](https://github.com/MicrosoftDocs/azure-docs/blob/main/articles/automation/automation-create-standalone-account.md) * Service: **automation** * Sub-service: **process-automation** * GitHub Login: @SnehaSudhirG * Microsoft Alias: **sudhirsneha**
process
networking tab is missing networking tab is missing in documentation there is no explanation guideline which option should be selected and their further impacts document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source service automation sub service process automation github login snehasudhirg microsoft alias sudhirsneha
1
1,180
3,681,568,915
IssuesEvent
2016-02-24 04:12:09
18F/FEC
https://api.github.com/repos/18F/FEC
closed
Positive: Calendar features
processed
## What were you trying to do and how can we improve it? I was looking at your new calendar features ## General feedback? I like the new layout ## Tell us about yourself I&#39;m new to the FEC website ## Details * URL: https://fec-proxy.18f.gov/calendar/ * User Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.10; rv:44.0) Gecko/20100101 Firefox/44.0
1.0
Positive: Calendar features - ## What were you trying to do and how can we improve it? I was looking at your new calendar features ## General feedback? I like the new layout ## Tell us about yourself I&#39;m new to the FEC website ## Details * URL: https://fec-proxy.18f.gov/calendar/ * User Agent: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.10; rv:44.0) Gecko/20100101 Firefox/44.0
process
positive calendar features what were you trying to do and how can we improve it i was looking at your new calendar features general feedback i like the new layout tell us about yourself i m new to the fec website details url user agent mozilla macintosh intel mac os x rv gecko firefox
1
18,286
24,380,782,985
IssuesEvent
2022-10-04 07:38:14
nedbat/coveragepy
https://api.github.com/repos/nedbat/coveragepy
closed
Some multiprocessing processes not covered (but were covered by pytest-cov 3.0.0)
bug multiprocessing needs triage
**Describe the bug** I am one of the developers of [qtile](https://github.com/qtile/qtile). We have a fairly large test suite that we ran with `pytest-cov`. Since they dropped `multiprocessing` support in 4.0.0, I have followed [their advice](https://pytest-cov.readthedocs.io/en/latest/changelog.html#id1) to use `coverage`. The `.coveragerc` file looks like this: ``` [run] source = libqtile concurrency = multiprocessing parallel = true sigterm = true [report] omit = libqtile/interactive/* libqtile/scripts/* libqtile/ffi_build.py test/* ``` I am getting a few discrepancies which I'm struggling to debug but I'm pretty sure they arise from lines that were executed by a subprocess not being covered. **To Reproduce** How can we reproduce the problem? Please *be specific*. Don't link to a failing CI job. To be honest, I'm not expecting you to recreate the problem but, if you do, you would need to clone our repo and run the suite and run the following test: `test/widgets/test_openweather.py`). Previously the file `libqtile/widget/open_weather.py` had 91.3% coverage, now it's 40%. At this stage, though, I'd be happy with some pointers as to how debug this more myself. Answer the questions below: 1. What version of Python are you using? `3.10` 1. What version of coverage.py shows the problem? The output of `coverage debug sys` is helpful. `coverage_version: 6.5.0` 1. What versions of what packages do you have installed? The output of `pip freeze` is helpful. ``` appdirs==1.4.4 attrs==22.1.0 bowler==0.9.0 cairocffi==1.4.0 cffi==1.15.1 click==8.1.3 coverage==6.5.0 dbus-next==0.2.3 fissix==21.11.13 iniconfig==1.1.1 moreorless==0.4.0 packaging==21.3 pluggy==1.0.0 py==1.11.0 pycairo==1.21.0 pycparser==2.21 PyGObject==3.42.2 pyparsing==3.0.9 pytest==7.1.3 pywayland==0.4.14 pywlroots==0.15.22 qtile==0.22.2.dev7+gfec84938.d20221001 six==1.16.0 tomli==2.0.1 volatile==2.1.0 xcffib==0.12.1 xkbcommon==0.4 ``` 1. What code shows the problem? Give us a specific commit of a specific repo that we can check out. If you've already worked around the problem, please provide a commit before that fix. My current working on this issue is here: https://github.com/elParaguayo/qtile/tree/pin-pytest-cov 1. What commands did you run? ``` coverage run -m pytest --backend=x11 --backend=wayland -k test_openweather coverage combine -q coverage report -m ``` **Expected behavior** I would expect the coverage to remain unchanged at 91.3%. The [existing coverage](https://coveralls.io/builds/52732848/source?filename=libqtile%2Fwidget%2Fopen_weather.py) shows multiple hits per line but this drops when using coverage. **Additional context** We launch subprocesses via a fixture with `multiprocessing.Process`. When the test ends, the fixture cleans up with a `terminate` call to that process which, I understand, should trigger the writing of the coverage data for that process.
1.0
Some multiprocessing processes not covered (but were covered by pytest-cov 3.0.0) - **Describe the bug** I am one of the developers of [qtile](https://github.com/qtile/qtile). We have a fairly large test suite that we ran with `pytest-cov`. Since they dropped `multiprocessing` support in 4.0.0, I have followed [their advice](https://pytest-cov.readthedocs.io/en/latest/changelog.html#id1) to use `coverage`. The `.coveragerc` file looks like this: ``` [run] source = libqtile concurrency = multiprocessing parallel = true sigterm = true [report] omit = libqtile/interactive/* libqtile/scripts/* libqtile/ffi_build.py test/* ``` I am getting a few discrepancies which I'm struggling to debug but I'm pretty sure they arise from lines that were executed by a subprocess not being covered. **To Reproduce** How can we reproduce the problem? Please *be specific*. Don't link to a failing CI job. To be honest, I'm not expecting you to recreate the problem but, if you do, you would need to clone our repo and run the suite and run the following test: `test/widgets/test_openweather.py`). Previously the file `libqtile/widget/open_weather.py` had 91.3% coverage, now it's 40%. At this stage, though, I'd be happy with some pointers as to how debug this more myself. Answer the questions below: 1. What version of Python are you using? `3.10` 1. What version of coverage.py shows the problem? The output of `coverage debug sys` is helpful. `coverage_version: 6.5.0` 1. What versions of what packages do you have installed? The output of `pip freeze` is helpful. ``` appdirs==1.4.4 attrs==22.1.0 bowler==0.9.0 cairocffi==1.4.0 cffi==1.15.1 click==8.1.3 coverage==6.5.0 dbus-next==0.2.3 fissix==21.11.13 iniconfig==1.1.1 moreorless==0.4.0 packaging==21.3 pluggy==1.0.0 py==1.11.0 pycairo==1.21.0 pycparser==2.21 PyGObject==3.42.2 pyparsing==3.0.9 pytest==7.1.3 pywayland==0.4.14 pywlroots==0.15.22 qtile==0.22.2.dev7+gfec84938.d20221001 six==1.16.0 tomli==2.0.1 volatile==2.1.0 xcffib==0.12.1 xkbcommon==0.4 ``` 1. What code shows the problem? Give us a specific commit of a specific repo that we can check out. If you've already worked around the problem, please provide a commit before that fix. My current working on this issue is here: https://github.com/elParaguayo/qtile/tree/pin-pytest-cov 1. What commands did you run? ``` coverage run -m pytest --backend=x11 --backend=wayland -k test_openweather coverage combine -q coverage report -m ``` **Expected behavior** I would expect the coverage to remain unchanged at 91.3%. The [existing coverage](https://coveralls.io/builds/52732848/source?filename=libqtile%2Fwidget%2Fopen_weather.py) shows multiple hits per line but this drops when using coverage. **Additional context** We launch subprocesses via a fixture with `multiprocessing.Process`. When the test ends, the fixture cleans up with a `terminate` call to that process which, I understand, should trigger the writing of the coverage data for that process.
process
some multiprocessing processes not covered but were covered by pytest cov describe the bug i am one of the developers of we have a fairly large test suite that we ran with pytest cov since they dropped multiprocessing support in i have followed to use coverage the coveragerc file looks like this source libqtile concurrency multiprocessing parallel true sigterm true omit libqtile interactive libqtile scripts libqtile ffi build py test i am getting a few discrepancies which i m struggling to debug but i m pretty sure they arise from lines that were executed by a subprocess not being covered to reproduce how can we reproduce the problem please be specific don t link to a failing ci job to be honest i m not expecting you to recreate the problem but if you do you would need to clone our repo and run the suite and run the following test test widgets test openweather py previously the file libqtile widget open weather py had coverage now it s at this stage though i d be happy with some pointers as to how debug this more myself answer the questions below what version of python are you using what version of coverage py shows the problem the output of coverage debug sys is helpful coverage version what versions of what packages do you have installed the output of pip freeze is helpful appdirs attrs bowler cairocffi cffi click coverage dbus next fissix iniconfig moreorless packaging pluggy py pycairo pycparser pygobject pyparsing pytest pywayland pywlroots qtile six tomli volatile xcffib xkbcommon what code shows the problem give us a specific commit of a specific repo that we can check out if you ve already worked around the problem please provide a commit before that fix my current working on this issue is here what commands did you run coverage run m pytest backend backend wayland k test openweather coverage combine q coverage report m expected behavior i would expect the coverage to remain unchanged at the shows multiple hits per line but this drops when using coverage additional context we launch subprocesses via a fixture with multiprocessing process when the test ends the fixture cleans up with a terminate call to that process which i understand should trigger the writing of the coverage data for that process
1
346,948
10,422,127,315
IssuesEvent
2019-09-16 08:16:45
zdnscloud/singlecloud
https://api.github.com/repos/zdnscloud/singlecloud
closed
error should return if deleting storagecluster which is used by pod
bug priority: Medium todo
1. create a storagecluster, 2. create a pod use the storagecluster, 3. delete storagecluser, the following occurs: 204 NoContent returns and the storagecluster is deleted. expected behavior: user can't delete storagecluster if storagecluster is in use. It's better to returns an error.
1.0
error should return if deleting storagecluster which is used by pod - 1. create a storagecluster, 2. create a pod use the storagecluster, 3. delete storagecluser, the following occurs: 204 NoContent returns and the storagecluster is deleted. expected behavior: user can't delete storagecluster if storagecluster is in use. It's better to returns an error.
non_process
error should return if deleting storagecluster which is used by pod create a storagecluster create a pod use the storagecluster delete storagecluser the following occurs nocontent returns and the storagecluster is deleted expected behavior user can t delete storagecluster if storagecluster is in use it s better to returns an error
0
8,625
7,343,225,893
IssuesEvent
2018-03-07 10:37:40
NationalBankBelgium/stark
https://api.github.com/repos/NationalBankBelgium/stark
opened
developer guide: document how to protect against HTML extraction
comp: developer-guide comp: docs security should type: doc
Goal: review and add to our doc: https://chloe.re/2016/07/19/protect-against-html-extraction/
True
developer guide: document how to protect against HTML extraction - Goal: review and add to our doc: https://chloe.re/2016/07/19/protect-against-html-extraction/
non_process
developer guide document how to protect against html extraction goal review and add to our doc
0
20,598
27,265,255,776
IssuesEvent
2023-02-22 17:31:25
googleapis/google-cloud-node
https://api.github.com/repos/googleapis/google-cloud-node
closed
Your .repo-metadata.json files have a problem 🤒
type: process repo-metadata: lint
You have a problem with your .repo-metadata.json files: Result of scan 📈: * release_level must be equal to one of the allowed values in packages/gapic-node-templating/templates/bootstrap-templates/.repo-metadata.json * api_shortname field missing from packages/gapic-node-templating/templates/bootstrap-templates/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-api-apikeys/.repo-metadata.json * api_shortname field missing from packages/google-api-apikeys/.repo-metadata.json * api_shortname 'asset' invalid in packages/google-cloud-asset/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-batch/.repo-metadata.json * api_shortname field missing from packages/google-cloud-batch/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-appconnections/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-appconnections/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-appconnectors/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-appconnectors/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-appgateways/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-appgateways/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-clientconnectorservices/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-clientconnectorservices/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-clientgateways/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-clientgateways/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-bigquery-analyticshub/.repo-metadata.json * api_shortname field missing from packages/google-cloud-bigquery-analyticshub/.repo-metadata.json * api_shortname field missing from packages/google-cloud-bigquery-dataexchange/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-bigquery-datapolicies/.repo-metadata.json * api_shortname field missing from packages/google-cloud-bigquery-datapolicies/.repo-metadata.json * api_shortname 'dms' invalid in packages/google-cloud-clouddms/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-contentwarehouse/.repo-metadata.json * api_shortname field missing from packages/google-cloud-contentwarehouse/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-datacatalog-lineage/.repo-metadata.json * api_shortname field missing from packages/google-cloud-datacatalog-lineage/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-discoveryengine/.repo-metadata.json * api_shortname field missing from packages/google-cloud-discoveryengine/.repo-metadata.json * api_shortname 'filestore' invalid in packages/google-cloud-filestore/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-gkemulticloud/.repo-metadata.json * api_shortname field missing from packages/google-cloud-gkemulticloud/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-gsuiteaddons/.repo-metadata.json * api_shortname field missing from packages/google-cloud-gsuiteaddons/.repo-metadata.json * must have required property 'library_type' in packages/google-cloud-run/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-run/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-security-publicca/.repo-metadata.json * api_shortname field missing from packages/google-cloud-security-publicca/.repo-metadata.json * must have required property 'library_type' in packages/google-cloud-video-stitcher/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-video-stitcher/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-vmwareengine/.repo-metadata.json * api_shortname field missing from packages/google-cloud-vmwareengine/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-iam/.repo-metadata.json * api_shortname field missing from packages/google-iam/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-maps-addressvalidation/.repo-metadata.json * api_shortname field missing from packages/google-maps-addressvalidation/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-maps-mapsplatformdatasets/.repo-metadata.json * api_shortname field missing from packages/google-maps-mapsplatformdatasets/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-maps-routing/.repo-metadata.json * api_shortname field missing from packages/google-maps-routing/.repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
1.0
Your .repo-metadata.json files have a problem 🤒 - You have a problem with your .repo-metadata.json files: Result of scan 📈: * release_level must be equal to one of the allowed values in packages/gapic-node-templating/templates/bootstrap-templates/.repo-metadata.json * api_shortname field missing from packages/gapic-node-templating/templates/bootstrap-templates/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-api-apikeys/.repo-metadata.json * api_shortname field missing from packages/google-api-apikeys/.repo-metadata.json * api_shortname 'asset' invalid in packages/google-cloud-asset/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-batch/.repo-metadata.json * api_shortname field missing from packages/google-cloud-batch/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-appconnections/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-appconnections/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-appconnectors/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-appconnectors/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-appgateways/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-appgateways/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-clientconnectorservices/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-clientconnectorservices/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-beyondcorp-clientgateways/.repo-metadata.json * api_shortname field missing from packages/google-cloud-beyondcorp-clientgateways/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-bigquery-analyticshub/.repo-metadata.json * api_shortname field missing from packages/google-cloud-bigquery-analyticshub/.repo-metadata.json * api_shortname field missing from packages/google-cloud-bigquery-dataexchange/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-bigquery-datapolicies/.repo-metadata.json * api_shortname field missing from packages/google-cloud-bigquery-datapolicies/.repo-metadata.json * api_shortname 'dms' invalid in packages/google-cloud-clouddms/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-contentwarehouse/.repo-metadata.json * api_shortname field missing from packages/google-cloud-contentwarehouse/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-datacatalog-lineage/.repo-metadata.json * api_shortname field missing from packages/google-cloud-datacatalog-lineage/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-discoveryengine/.repo-metadata.json * api_shortname field missing from packages/google-cloud-discoveryengine/.repo-metadata.json * api_shortname 'filestore' invalid in packages/google-cloud-filestore/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-gkemulticloud/.repo-metadata.json * api_shortname field missing from packages/google-cloud-gkemulticloud/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-gsuiteaddons/.repo-metadata.json * api_shortname field missing from packages/google-cloud-gsuiteaddons/.repo-metadata.json * must have required property 'library_type' in packages/google-cloud-run/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-run/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-security-publicca/.repo-metadata.json * api_shortname field missing from packages/google-cloud-security-publicca/.repo-metadata.json * must have required property 'library_type' in packages/google-cloud-video-stitcher/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-video-stitcher/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-cloud-vmwareengine/.repo-metadata.json * api_shortname field missing from packages/google-cloud-vmwareengine/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-iam/.repo-metadata.json * api_shortname field missing from packages/google-iam/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-maps-addressvalidation/.repo-metadata.json * api_shortname field missing from packages/google-maps-addressvalidation/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-maps-mapsplatformdatasets/.repo-metadata.json * api_shortname field missing from packages/google-maps-mapsplatformdatasets/.repo-metadata.json * release_level must be equal to one of the allowed values in packages/google-maps-routing/.repo-metadata.json * api_shortname field missing from packages/google-maps-routing/.repo-metadata.json ☝️ Once you address these problems, you can close this issue. ### Need help? * [Schema definition](https://github.com/googleapis/repo-automation-bots/blob/main/packages/repo-metadata-lint/src/repo-metadata-schema.json): lists valid options for each field. * [API index](https://github.com/googleapis/googleapis/blob/master/api-index-v1.json): for gRPC libraries **api_shortname** should match the subdomain of an API's **hostName**. * Reach out to **go/github-automation** if you have any questions.
process
your repo metadata json files have a problem 🤒 you have a problem with your repo metadata json files result of scan 📈 release level must be equal to one of the allowed values in packages gapic node templating templates bootstrap templates repo metadata json api shortname field missing from packages gapic node templating templates bootstrap templates repo metadata json release level must be equal to one of the allowed values in packages google api apikeys repo metadata json api shortname field missing from packages google api apikeys repo metadata json api shortname asset invalid in packages google cloud asset repo metadata json release level must be equal to one of the allowed values in packages google cloud batch repo metadata json api shortname field missing from packages google cloud batch repo metadata json release level must be equal to one of the allowed values in packages google cloud beyondcorp appconnections repo metadata json api shortname field missing from packages google cloud beyondcorp appconnections repo metadata json release level must be equal to one of the allowed values in packages google cloud beyondcorp appconnectors repo metadata json api shortname field missing from packages google cloud beyondcorp appconnectors repo metadata json release level must be equal to one of the allowed values in packages google cloud beyondcorp appgateways repo metadata json api shortname field missing from packages google cloud beyondcorp appgateways repo metadata json release level must be equal to one of the allowed values in packages google cloud beyondcorp clientconnectorservices repo metadata json api shortname field missing from packages google cloud beyondcorp clientconnectorservices repo metadata json release level must be equal to one of the allowed values in packages google cloud beyondcorp clientgateways repo metadata json api shortname field missing from packages google cloud beyondcorp clientgateways repo metadata json release level must be equal to one of the allowed values in packages google cloud bigquery analyticshub repo metadata json api shortname field missing from packages google cloud bigquery analyticshub repo metadata json api shortname field missing from packages google cloud bigquery dataexchange repo metadata json release level must be equal to one of the allowed values in packages google cloud bigquery datapolicies repo metadata json api shortname field missing from packages google cloud bigquery datapolicies repo metadata json api shortname dms invalid in packages google cloud clouddms repo metadata json release level must be equal to one of the allowed values in packages google cloud contentwarehouse repo metadata json api shortname field missing from packages google cloud contentwarehouse repo metadata json release level must be equal to one of the allowed values in packages google cloud datacatalog lineage repo metadata json api shortname field missing from packages google cloud datacatalog lineage repo metadata json release level must be equal to one of the allowed values in packages google cloud discoveryengine repo metadata json api shortname field missing from packages google cloud discoveryengine repo metadata json api shortname filestore invalid in packages google cloud filestore repo metadata json release level must be equal to one of the allowed values in packages google cloud gkemulticloud repo metadata json api shortname field missing from packages google cloud gkemulticloud repo metadata json release level must be equal to one of the allowed values in packages google cloud gsuiteaddons repo metadata json api shortname field missing from packages google cloud gsuiteaddons repo metadata json must have required property library type in packages google cloud run repo metadata json release level must be equal to one of the allowed values in packages google cloud run repo metadata json release level must be equal to one of the allowed values in packages google cloud security publicca repo metadata json api shortname field missing from packages google cloud security publicca repo metadata json must have required property library type in packages google cloud video stitcher repo metadata json release level must be equal to one of the allowed values in packages google cloud video stitcher repo metadata json release level must be equal to one of the allowed values in packages google cloud vmwareengine repo metadata json api shortname field missing from packages google cloud vmwareengine repo metadata json release level must be equal to one of the allowed values in packages google iam repo metadata json api shortname field missing from packages google iam repo metadata json release level must be equal to one of the allowed values in packages google maps addressvalidation repo metadata json api shortname field missing from packages google maps addressvalidation repo metadata json release level must be equal to one of the allowed values in packages google maps mapsplatformdatasets repo metadata json api shortname field missing from packages google maps mapsplatformdatasets repo metadata json release level must be equal to one of the allowed values in packages google maps routing repo metadata json api shortname field missing from packages google maps routing repo metadata json ☝️ once you address these problems you can close this issue need help lists valid options for each field for grpc libraries api shortname should match the subdomain of an api s hostname reach out to go github automation if you have any questions
1
3,117
6,149,228,540
IssuesEvent
2017-06-27 19:38:07
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
closed
Reenable the tests ActiveIssue'd on 19909 this in Uap (not UapAot)
area-System.Diagnostics.Process os-mac-os-x-10.13 os-windows-uwp test-run-uwp-coreclr
``` System.IO.MemoryMappedFiles.Tests -method System.IO.MemoryMappedFiles.Tests.DataShared System.Net.Primitives.Functional.Tests -method System.Net.Primitives.Functional.Tests.EventSource_EventsRaisedAsExpected System.Net.Security.Tests -method System.Net.Security.Tests.EventSource_EventsRaisedAsExpected ``` These tests are ActiveIssue'd against this issue but for Uap, not UapAot. Chances are they were intended to be ActiveIssued against some other process-related issue that prevents them from running in UWP (non-AOT) but since we've fix the aot-related issues, zombie'ing this issue to track reenabling these two tests.
1.0
Reenable the tests ActiveIssue'd on 19909 this in Uap (not UapAot) - ``` System.IO.MemoryMappedFiles.Tests -method System.IO.MemoryMappedFiles.Tests.DataShared System.Net.Primitives.Functional.Tests -method System.Net.Primitives.Functional.Tests.EventSource_EventsRaisedAsExpected System.Net.Security.Tests -method System.Net.Security.Tests.EventSource_EventsRaisedAsExpected ``` These tests are ActiveIssue'd against this issue but for Uap, not UapAot. Chances are they were intended to be ActiveIssued against some other process-related issue that prevents them from running in UWP (non-AOT) but since we've fix the aot-related issues, zombie'ing this issue to track reenabling these two tests.
process
reenable the tests activeissue d on this in uap not uapaot system io memorymappedfiles tests method system io memorymappedfiles tests datashared system net primitives functional tests method system net primitives functional tests eventsource eventsraisedasexpected system net security tests method system net security tests eventsource eventsraisedasexpected these tests are activeissue d against this issue but for uap not uapaot chances are they were intended to be activeissued against some other process related issue that prevents them from running in uwp non aot but since we ve fix the aot related issues zombie ing this issue to track reenabling these two tests
1
1,901
4,117,504,667
IssuesEvent
2016-06-08 07:46:38
deb761/Solstice
https://api.github.com/repos/deb761/Solstice
closed
Create a procedure that returns a students missed problems
Requirement
Create a stored procedure that returns the problems of a given type that a student has missed in the current level.
1.0
Create a procedure that returns a students missed problems - Create a stored procedure that returns the problems of a given type that a student has missed in the current level.
non_process
create a procedure that returns a students missed problems create a stored procedure that returns the problems of a given type that a student has missed in the current level
0
21,499
29,661,828,611
IssuesEvent
2023-06-10 08:50:41
nextflow-io/nextflow
https://api.github.com/repos/nextflow-io/nextflow
closed
Files (symlinks) not found when using regex and `followLinks: false`
stale lang/processes
Not completely sure if this is a bug: I am running into an issue when using the option followLinks in the output section and combining it with the regex. Here is the simplest example to showcase the issue. The first example works as expected: ``` #!/usr/bin/env nextflow process createMetaFiles { output: path "link*" shell: """ echo "1" > ~/MetaFile1 ln -s ~/MetaFile1 link1 """ } workflow { metaFiles = createMetaFiles() metaFiles.view { it -> "MetaFiles: $it" } } ``` Now if I add `followLinks: false` it fails. I expect that `followLinks: false` shouldn't have an effect: ``` #!/usr/bin/env nextflow process createMetaFiles { output: path "link*", followLinks: false shell: """ echo "1" > ~/MetaFile1 ln -s ~/MetaFile1 link1 """ } workflow { metaFiles = createMetaFiles() metaFiles.view { it -> "MetaFiles: $it" } } ``` The error is: ``` Caused by: Missing output file(s) `link*` expected by process `createMetaFiles` ``` Any suggestion would be appreciated. Note: If no glob is used in the path section, the previous example works fine: ``` #!/usr/bin/env nextflow process createMetaFiles { output: path "link1", followLinks: false shell: """ echo "1" > ~/MetaFile1 ln -s ~/MetaFile1 link1 """ } workflow { metaFiles = createMetaFiles() metaFiles.view { it -> "MetaFiles: $it" } } ```
1.0
Files (symlinks) not found when using regex and `followLinks: false` - Not completely sure if this is a bug: I am running into an issue when using the option followLinks in the output section and combining it with the regex. Here is the simplest example to showcase the issue. The first example works as expected: ``` #!/usr/bin/env nextflow process createMetaFiles { output: path "link*" shell: """ echo "1" > ~/MetaFile1 ln -s ~/MetaFile1 link1 """ } workflow { metaFiles = createMetaFiles() metaFiles.view { it -> "MetaFiles: $it" } } ``` Now if I add `followLinks: false` it fails. I expect that `followLinks: false` shouldn't have an effect: ``` #!/usr/bin/env nextflow process createMetaFiles { output: path "link*", followLinks: false shell: """ echo "1" > ~/MetaFile1 ln -s ~/MetaFile1 link1 """ } workflow { metaFiles = createMetaFiles() metaFiles.view { it -> "MetaFiles: $it" } } ``` The error is: ``` Caused by: Missing output file(s) `link*` expected by process `createMetaFiles` ``` Any suggestion would be appreciated. Note: If no glob is used in the path section, the previous example works fine: ``` #!/usr/bin/env nextflow process createMetaFiles { output: path "link1", followLinks: false shell: """ echo "1" > ~/MetaFile1 ln -s ~/MetaFile1 link1 """ } workflow { metaFiles = createMetaFiles() metaFiles.view { it -> "MetaFiles: $it" } } ```
process
files symlinks not found when using regex and followlinks false not completely sure if this is a bug i am running into an issue when using the option followlinks in the output section and combining it with the regex here is the simplest example to showcase the issue the first example works as expected usr bin env nextflow process createmetafiles output path link shell echo ln s workflow metafiles createmetafiles metafiles view it metafiles it now if i add followlinks false it fails i expect that followlinks false shouldn t have an effect usr bin env nextflow process createmetafiles output path link followlinks false shell echo ln s workflow metafiles createmetafiles metafiles view it metafiles it the error is caused by missing output file s link expected by process createmetafiles any suggestion would be appreciated note if no glob is used in the path section the previous example works fine usr bin env nextflow process createmetafiles output path followlinks false shell echo ln s workflow metafiles createmetafiles metafiles view it metafiles it
1
19,835
26,230,140,757
IssuesEvent
2023-01-04 22:59:30
open-telemetry/opentelemetry-collector-contrib
https://api.github.com/repos/open-telemetry/opentelemetry-collector-contrib
closed
Remove deprecated 'gke' and 'gce' resource detectors after v0.54.0 is released
admin issues comp:google processor/resourcedetection
The replacement is to use the single 'gcp' resource detector. See https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/10347.
1.0
Remove deprecated 'gke' and 'gce' resource detectors after v0.54.0 is released - The replacement is to use the single 'gcp' resource detector. See https://github.com/open-telemetry/opentelemetry-collector-contrib/pull/10347.
process
remove deprecated gke and gce resource detectors after is released the replacement is to use the single gcp resource detector see
1
14,490
17,603,832,787
IssuesEvent
2021-08-17 14:46:50
qgis/QGIS-Documentation
https://api.github.com/repos/qgis/QGIS-Documentation
closed
Algorithm parameters - use of include and handling advanced parameters
Processing
## Description **First**: We have been "experimenting" with includes for algorithm parameters, and have discovered that the RST `.. include:` has some important limitations. It does not seem to be possible to use it to import a group of elements of a list-table. Below are some alternatives that we have tested, and I hope we can discuss. The alternatives are implemented in PR #4627 (in the documentation of the QGIS Network algorithms - probably the only place where includes have a real potential for algorithm parameters). * Include 1: "Pasting" tables together. Demonstrated in "Service area (from layer)". * Include 2: Placing the "repeating" group of parameters in a separate table. Demonstrated for the Advanced parameters in the "Service area (from point)" algorithm. **Second**: To help the users, it would be nice to have the "Advanced parameters" group exposed in the parameter tables. Three approaches: * Advanced 1: Adding a row in the table before and after the Advanced parameters group, as demonstrated in the "Shortest path (layer to point)" algorithm. * Advanced 2: Including "Advanced" in one of the existing columns. In the "Shortest path (point to layer)" algorithm, it is included in the Label column. * Advanced 3: A separate column for Advanced in the table. Demonstrated in the "Shortest path (point to point)" algorithm "Advanced" alternative 2 is easiest to implement, and seems quite logical (placing "Optional" and "Advanced" in the same ("Label") column. <!-- Cleaning the queue is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. Include sentences with details describing the issue you have encountered (e.g., actual behavior, expected behavior, steps to reproduce). --> Page URL: https://docs.qgis.org/testing/en/docs/user_manual/processing_algs/qgis/networkanalysis.html
1.0
Algorithm parameters - use of include and handling advanced parameters - ## Description **First**: We have been "experimenting" with includes for algorithm parameters, and have discovered that the RST `.. include:` has some important limitations. It does not seem to be possible to use it to import a group of elements of a list-table. Below are some alternatives that we have tested, and I hope we can discuss. The alternatives are implemented in PR #4627 (in the documentation of the QGIS Network algorithms - probably the only place where includes have a real potential for algorithm parameters). * Include 1: "Pasting" tables together. Demonstrated in "Service area (from layer)". * Include 2: Placing the "repeating" group of parameters in a separate table. Demonstrated for the Advanced parameters in the "Service area (from point)" algorithm. **Second**: To help the users, it would be nice to have the "Advanced parameters" group exposed in the parameter tables. Three approaches: * Advanced 1: Adding a row in the table before and after the Advanced parameters group, as demonstrated in the "Shortest path (layer to point)" algorithm. * Advanced 2: Including "Advanced" in one of the existing columns. In the "Shortest path (point to layer)" algorithm, it is included in the Label column. * Advanced 3: A separate column for Advanced in the table. Demonstrated in the "Shortest path (point to point)" algorithm "Advanced" alternative 2 is easiest to implement, and seems quite logical (placing "Optional" and "Advanced" in the same ("Label") column. <!-- Cleaning the queue is a process done by project maintainers, mostly on a volunteer basis. We try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items. Include sentences with details describing the issue you have encountered (e.g., actual behavior, expected behavior, steps to reproduce). --> Page URL: https://docs.qgis.org/testing/en/docs/user_manual/processing_algs/qgis/networkanalysis.html
process
algorithm parameters use of include and handling advanced parameters description first we have been experimenting with includes for algorithm parameters and have discovered that the rst include has some important limitations it does not seem to be possible to use it to import a group of elements of a list table below are some alternatives that we have tested and i hope we can discuss the alternatives are implemented in pr in the documentation of the qgis network algorithms probably the only place where includes have a real potential for algorithm parameters include pasting tables together demonstrated in service area from layer include placing the repeating group of parameters in a separate table demonstrated for the advanced parameters in the service area from point algorithm second to help the users it would be nice to have the advanced parameters group exposed in the parameter tables three approaches advanced adding a row in the table before and after the advanced parameters group as demonstrated in the shortest path layer to point algorithm advanced including advanced in one of the existing columns in the shortest path point to layer algorithm it is included in the label column advanced a separate column for advanced in the table demonstrated in the shortest path point to point algorithm advanced alternative is easiest to implement and seems quite logical placing optional and advanced in the same label column cleaning the queue is a process done by project maintainers mostly on a volunteer basis we try to keep the overhead as small as possible and appreciate if you help us to do so by completing the following items include sentences with details describing the issue you have encountered e g actual behavior expected behavior steps to reproduce page url
1
126,392
12,291,507,803
IssuesEvent
2020-05-10 10:22:11
i-on-project/core
https://api.github.com/repos/i-on-project/core
closed
Docs: Compose API Spec
documentation
Compose a basic documentation of the API. Some topics to include: - endpoints - available operations - properties - specific errors - routes - error handling - message formats - [ ] Read API Spec - [ ] Write API Spec
1.0
Docs: Compose API Spec - Compose a basic documentation of the API. Some topics to include: - endpoints - available operations - properties - specific errors - routes - error handling - message formats - [ ] Read API Spec - [ ] Write API Spec
non_process
docs compose api spec compose a basic documentation of the api some topics to include endpoints available operations properties specific errors routes error handling message formats read api spec write api spec
0
69,096
17,571,287,217
IssuesEvent
2021-08-14 18:56:14
haskell/cabal
https://api.github.com/repos/haskell/cabal
closed
`new-haddock`'s file monitoring broken
type: bug cabal-install: nix-local-build
Consider the following simple reproduction instructions: ``` $ cabal get sort-1.0.0.1 && cd sort-0.0.0.1/ Unpacking to sort-0.0.0.1/ ``` The first invocation of `new-haddock` works as expected: ``` $ cabal new-haddock Resolving dependencies... In order, the following will be built (use -v for more details): - sort-0.0.0.1 (lib) (first run) Configuring library for sort-0.0.0.1.. Preprocessing library for sort-0.0.0.1.. Running Haddock on library for sort-0.0.0.1.. Haddock coverage: 67% ( 2 / 3) in 'Data.Sort' Missing documentation for: Module header Documentation created: /tmp/sort-0.0.0.1/dist-newstyle/build/x86_64-linux/ghc-8.0.2/sort-0.0.0.1/doc/html/sort/index.html ``` However, the 2nd invocation of `new-haddock` rebuilds the documentation (NB: the install-plan wasn't resolved, and the action plan says "(first run)"): ``` $ cabal new-haddock In order, the following will be built (use -v for more details): - sort-0.0.0.1 (lib) (first run) Preprocessing library for sort-0.0.0.1.. Running Haddock on library for sort-0.0.0.1.. Haddock coverage: 67% ( 2 / 3) in 'Data.Sort' Missing documentation for: Module header Documentation created: /tmp/sort-0.0.0.1/dist-newstyle/build/x86_64-linux/ghc-8.0.2/sort-0.0.0.1/doc/html/sort/index.html ``` What's really bad though: if we remove `dist-newstyle` again, and start with `cabal new-build`, then `new-haddock` doesn't generate any documentation at all: ``` $ rm -rf dist-newstyle/ $ cabal new-build Resolving dependencies... In order, the following will be built (use -v for more details): - sort-0.0.0.1 (lib) (first run) Configuring library for sort-0.0.0.1.. Preprocessing library for sort-0.0.0.1.. Building library for sort-0.0.0.1.. [1 of 1] Compiling Data.Sort ( src/Data/Sort.hs, /tmp/sort-0.0.0.1/dist-newstyle/build/x86_64-linux/ghc-8.0.2/sort-0.0.0.1/build/Data/Sort.o ) $ cabal new-haddock Up to date ```
1.0
`new-haddock`'s file monitoring broken - Consider the following simple reproduction instructions: ``` $ cabal get sort-1.0.0.1 && cd sort-0.0.0.1/ Unpacking to sort-0.0.0.1/ ``` The first invocation of `new-haddock` works as expected: ``` $ cabal new-haddock Resolving dependencies... In order, the following will be built (use -v for more details): - sort-0.0.0.1 (lib) (first run) Configuring library for sort-0.0.0.1.. Preprocessing library for sort-0.0.0.1.. Running Haddock on library for sort-0.0.0.1.. Haddock coverage: 67% ( 2 / 3) in 'Data.Sort' Missing documentation for: Module header Documentation created: /tmp/sort-0.0.0.1/dist-newstyle/build/x86_64-linux/ghc-8.0.2/sort-0.0.0.1/doc/html/sort/index.html ``` However, the 2nd invocation of `new-haddock` rebuilds the documentation (NB: the install-plan wasn't resolved, and the action plan says "(first run)"): ``` $ cabal new-haddock In order, the following will be built (use -v for more details): - sort-0.0.0.1 (lib) (first run) Preprocessing library for sort-0.0.0.1.. Running Haddock on library for sort-0.0.0.1.. Haddock coverage: 67% ( 2 / 3) in 'Data.Sort' Missing documentation for: Module header Documentation created: /tmp/sort-0.0.0.1/dist-newstyle/build/x86_64-linux/ghc-8.0.2/sort-0.0.0.1/doc/html/sort/index.html ``` What's really bad though: if we remove `dist-newstyle` again, and start with `cabal new-build`, then `new-haddock` doesn't generate any documentation at all: ``` $ rm -rf dist-newstyle/ $ cabal new-build Resolving dependencies... In order, the following will be built (use -v for more details): - sort-0.0.0.1 (lib) (first run) Configuring library for sort-0.0.0.1.. Preprocessing library for sort-0.0.0.1.. Building library for sort-0.0.0.1.. [1 of 1] Compiling Data.Sort ( src/Data/Sort.hs, /tmp/sort-0.0.0.1/dist-newstyle/build/x86_64-linux/ghc-8.0.2/sort-0.0.0.1/build/Data/Sort.o ) $ cabal new-haddock Up to date ```
non_process
new haddock s file monitoring broken consider the following simple reproduction instructions cabal get sort cd sort unpacking to sort the first invocation of new haddock works as expected cabal new haddock resolving dependencies in order the following will be built use v for more details sort lib first run configuring library for sort preprocessing library for sort running haddock on library for sort haddock coverage in data sort missing documentation for module header documentation created tmp sort dist newstyle build linux ghc sort doc html sort index html however the invocation of new haddock rebuilds the documentation nb the install plan wasn t resolved and the action plan says first run cabal new haddock in order the following will be built use v for more details sort lib first run preprocessing library for sort running haddock on library for sort haddock coverage in data sort missing documentation for module header documentation created tmp sort dist newstyle build linux ghc sort doc html sort index html what s really bad though if we remove dist newstyle again and start with cabal new build then new haddock doesn t generate any documentation at all rm rf dist newstyle cabal new build resolving dependencies in order the following will be built use v for more details sort lib first run configuring library for sort preprocessing library for sort building library for sort compiling data sort src data sort hs tmp sort dist newstyle build linux ghc sort build data sort o cabal new haddock up to date
0
14,201
17,100,505,467
IssuesEvent
2021-07-09 10:31:15
prisma/prisma
https://api.github.com/repos/prisma/prisma
opened
`$queryRaw` availability isn't derived from the DMMF
process/candidate
Blocks/depends on: https://github.com/prisma/prisma-engines/pull/2087 If raw operations are not available for a connector, they are not rendered in the DMMF: ```js // ... "mappings": { "modelOperations": [ { "model": "Parent", // ... }, { "model": "Child", // ... } ], "otherOperations": { "read": [], "write": [] } } // ... ``` However, the generated client still allows `$queryRaw` without complaining. This should be based on the available queries of the engine. This is the object for a connector with raw query availability: ```js "otherOperations": { "read": [], "write": [ "executeRaw", "queryRaw" ] } ```
1.0
`$queryRaw` availability isn't derived from the DMMF - Blocks/depends on: https://github.com/prisma/prisma-engines/pull/2087 If raw operations are not available for a connector, they are not rendered in the DMMF: ```js // ... "mappings": { "modelOperations": [ { "model": "Parent", // ... }, { "model": "Child", // ... } ], "otherOperations": { "read": [], "write": [] } } // ... ``` However, the generated client still allows `$queryRaw` without complaining. This should be based on the available queries of the engine. This is the object for a connector with raw query availability: ```js "otherOperations": { "read": [], "write": [ "executeRaw", "queryRaw" ] } ```
process
queryraw availability isn t derived from the dmmf blocks depends on if raw operations are not available for a connector they are not rendered in the dmmf js mappings modeloperations model parent model child otheroperations read write however the generated client still allows queryraw without complaining this should be based on the available queries of the engine this is the object for a connector with raw query availability js otheroperations read write executeraw queryraw
1
1,632
4,242,410,618
IssuesEvent
2016-07-06 19:23:44
Eibriel/rdany
https://api.github.com/repos/Eibriel/rdany
opened
Add MAXS processor
Add processor Español
MAXS allows you to receive notifications and remote control your Android device over XMPP.
1.0
Add MAXS processor - MAXS allows you to receive notifications and remote control your Android device over XMPP.
process
add maxs processor maxs allows you to receive notifications and remote control your android device over xmpp
1
577,942
17,139,522,823
IssuesEvent
2021-07-13 08:03:59
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
startpage.com - desktop site instead of mobile site
browser-firefox-ios os-ios priority-normal
<!-- @browser: Firefox iOS 34.2 --> <!-- @ua_header: Mozilla/5.0 (iPhone; CPU iPhone OS 14_6 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/34.2 Mobile/15E148 Safari/605.1.15 --> <!-- @reported_with: unknown --> **URL**: https://startpage.com **Browser / Version**: Firefox iOS 34.2 **Operating System**: iOS 14.6 **Tested Another Browser**: Yes Other **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: https://https//www.bing.com/images/search?view=detailV2&ccid=ZDoew7%2fI&id=49077E78DD4736B2E23372A203FE8BE595426A8D&thid=OIP.ZDoew7_IoDs7Fo_Rd9BtTgHaIW&mediaurl=https%3a%2f%2fwww.pinclipart.com%2fpicdir%2fmiddle%2f55-553025_black-check-mark-png-clipart.png&cdnurl=https%3a%2f%2fth.bing.com%2fth%2fid%2fR.643a1ec3bfc8a03b3b168fd177d06d4e%3frik%3djWpCleWL%252fgOicg%26pid%3dImgRaw&exph=993&expw=880&q=black+check+mark+emoji&simid=608038146553575890&ck=9AF7567CDB692500162ABDC3E6215190&selectedIndex=76&adlt=strict&FORM=IRPRST <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
startpage.com - desktop site instead of mobile site - <!-- @browser: Firefox iOS 34.2 --> <!-- @ua_header: Mozilla/5.0 (iPhone; CPU iPhone OS 14_6 like Mac OS X) AppleWebKit/605.1.15 (KHTML, like Gecko) FxiOS/34.2 Mobile/15E148 Safari/605.1.15 --> <!-- @reported_with: unknown --> **URL**: https://startpage.com **Browser / Version**: Firefox iOS 34.2 **Operating System**: iOS 14.6 **Tested Another Browser**: Yes Other **Problem type**: Desktop site instead of mobile site **Description**: Desktop site instead of mobile site **Steps to Reproduce**: https://https//www.bing.com/images/search?view=detailV2&ccid=ZDoew7%2fI&id=49077E78DD4736B2E23372A203FE8BE595426A8D&thid=OIP.ZDoew7_IoDs7Fo_Rd9BtTgHaIW&mediaurl=https%3a%2f%2fwww.pinclipart.com%2fpicdir%2fmiddle%2f55-553025_black-check-mark-png-clipart.png&cdnurl=https%3a%2f%2fth.bing.com%2fth%2fid%2fR.643a1ec3bfc8a03b3b168fd177d06d4e%3frik%3djWpCleWL%252fgOicg%26pid%3dImgRaw&exph=993&expw=880&q=black+check+mark+emoji&simid=608038146553575890&ck=9AF7567CDB692500162ABDC3E6215190&selectedIndex=76&adlt=strict&FORM=IRPRST <details> <summary>Browser Configuration</summary> <ul> <li>None</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_process
startpage com desktop site instead of mobile site url browser version firefox ios operating system ios tested another browser yes other problem type desktop site instead of mobile site description desktop site instead of mobile site steps to reproduce browser configuration none from with ❤️
0
157,838
12,392,975,556
IssuesEvent
2020-05-20 14:46:02
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] RemoteClusterClientTests testEnsureWeReconnect failing with NoSuchRemoteClusterException
:Distributed/Network >test-failure Team:Distributed
This test has failed 4 times in the past three days after passing basically 100% of the time for the past month. Look suspicious. Happening on both `master` and `7.x`. ``` :server:test » org.elasticsearch.transport.RemoteClusterClientTests » testEnsureWeReconnect (1.662s) org.elasticsearch.transport.NoSuchRemoteClusterException: no such remote cluster: [test] java.util.concurrent.ExecutionException: org.elasticsearch.transport.NoSuchRemoteClusterException: no such remote cluster: [test]Open stacktrace Caused by: org.elasticsearch.transport.NoSuchRemoteClusterException: no such remote cluster: [test]Open stacktrace [2020-02-07T11:10:26,252][INFO ][o.e.t.RemoteClusterClientTests] [testEnsureWeReconnect] before test [2020-02-07T11:10:26,808][INFO ][o.e.t.TransportService ] [testEnsureWeReconnect] publish_address {127.0.0.1:14000}, bound_addresses {[::1]:14000}, {127.0.0.1:14000} [2020-02-07T11:10:27,048][INFO ][o.e.t.TransportService ] [testEnsureWeReconnect] publish_address {127.0.0.1:14001}, bound_addresses {[::1]:14001}, {127.0.0.1:14001} [2020-02-07T11:10:27,519][INFO ][o.e.t.RemoteClusterClientTests] [testEnsureWeReconnect] after test REPRODUCE WITH: ./gradlew ':server:test' --tests "org.elasticsearch.transport.RemoteClusterClientTests.testEnsureWeReconnect" -Dtests.seed=EE4961D2D47604F4 -Dtests.security.manager=true -Dtests.locale=zh-Hans-SG -Dtests.timezone=Asia/Pontianak -Dcompiler.java=13 ``` This same time period has also seen a rather significant jump in average test execution times so perhaps there is something going on here. ![image](https://user-images.githubusercontent.com/4106672/74005594-3da88900-492e-11ea-911d-e35cdc2c28c7.png) https://gradle-enterprise.elastic.co/scans/tests?failures.failureClassification=non_verification&list.offset=0&list.size=50&list.sortColumn=startTime&list.sortOrder=desc&search.buildToolType=gradle&search.buildToolType=maven&search.startTimeMax=1581055741639&search.startTimeMin=1580450941631&search.tags=CI&search.tags=not:nested&search.tags=not:pull-request&tests.container=org.elasticsearch.transport.RemoteClusterClientTests&tests.sortField=FAILED&tests.test=testEnsureWeReconnect&tests.unstableOnly&trends.section=overview&trends.timeResolution=day&viewer.tzOffset=-480
1.0
[CI] RemoteClusterClientTests testEnsureWeReconnect failing with NoSuchRemoteClusterException - This test has failed 4 times in the past three days after passing basically 100% of the time for the past month. Look suspicious. Happening on both `master` and `7.x`. ``` :server:test » org.elasticsearch.transport.RemoteClusterClientTests » testEnsureWeReconnect (1.662s) org.elasticsearch.transport.NoSuchRemoteClusterException: no such remote cluster: [test] java.util.concurrent.ExecutionException: org.elasticsearch.transport.NoSuchRemoteClusterException: no such remote cluster: [test]Open stacktrace Caused by: org.elasticsearch.transport.NoSuchRemoteClusterException: no such remote cluster: [test]Open stacktrace [2020-02-07T11:10:26,252][INFO ][o.e.t.RemoteClusterClientTests] [testEnsureWeReconnect] before test [2020-02-07T11:10:26,808][INFO ][o.e.t.TransportService ] [testEnsureWeReconnect] publish_address {127.0.0.1:14000}, bound_addresses {[::1]:14000}, {127.0.0.1:14000} [2020-02-07T11:10:27,048][INFO ][o.e.t.TransportService ] [testEnsureWeReconnect] publish_address {127.0.0.1:14001}, bound_addresses {[::1]:14001}, {127.0.0.1:14001} [2020-02-07T11:10:27,519][INFO ][o.e.t.RemoteClusterClientTests] [testEnsureWeReconnect] after test REPRODUCE WITH: ./gradlew ':server:test' --tests "org.elasticsearch.transport.RemoteClusterClientTests.testEnsureWeReconnect" -Dtests.seed=EE4961D2D47604F4 -Dtests.security.manager=true -Dtests.locale=zh-Hans-SG -Dtests.timezone=Asia/Pontianak -Dcompiler.java=13 ``` This same time period has also seen a rather significant jump in average test execution times so perhaps there is something going on here. ![image](https://user-images.githubusercontent.com/4106672/74005594-3da88900-492e-11ea-911d-e35cdc2c28c7.png) https://gradle-enterprise.elastic.co/scans/tests?failures.failureClassification=non_verification&list.offset=0&list.size=50&list.sortColumn=startTime&list.sortOrder=desc&search.buildToolType=gradle&search.buildToolType=maven&search.startTimeMax=1581055741639&search.startTimeMin=1580450941631&search.tags=CI&search.tags=not:nested&search.tags=not:pull-request&tests.container=org.elasticsearch.transport.RemoteClusterClientTests&tests.sortField=FAILED&tests.test=testEnsureWeReconnect&tests.unstableOnly&trends.section=overview&trends.timeResolution=day&viewer.tzOffset=-480
non_process
remoteclusterclienttests testensurewereconnect failing with nosuchremoteclusterexception this test has failed times in the past three days after passing basically of the time for the past month look suspicious happening on both master and x server test » org elasticsearch transport remoteclusterclienttests » testensurewereconnect org elasticsearch transport nosuchremoteclusterexception no such remote cluster java util concurrent executionexception org elasticsearch transport nosuchremoteclusterexception no such remote cluster open stacktrace caused by org elasticsearch transport nosuchremoteclusterexception no such remote cluster open stacktrace before test publish address bound addresses publish address bound addresses after test reproduce with gradlew server test tests org elasticsearch transport remoteclusterclienttests testensurewereconnect dtests seed dtests security manager true dtests locale zh hans sg dtests timezone asia pontianak dcompiler java this same time period has also seen a rather significant jump in average test execution times so perhaps there is something going on here
0
1,570
4,165,441,673
IssuesEvent
2016-06-19 14:02:57
sysown/proxysql
https://api.github.com/repos/sysown/proxysql
opened
Debug mode query rewrite
ADMIN MYSQL PROTOCOL QUERY PROCESSOR
ProxySQL should have a way to understand how queries will be rewritten if a rewrite rule is applied, without applying the rules yet. Implementation still TBD
1.0
Debug mode query rewrite - ProxySQL should have a way to understand how queries will be rewritten if a rewrite rule is applied, without applying the rules yet. Implementation still TBD
process
debug mode query rewrite proxysql should have a way to understand how queries will be rewritten if a rewrite rule is applied without applying the rules yet implementation still tbd
1
3,936
6,875,248,094
IssuesEvent
2017-11-19 11:40:42
nuclio/nuclio
https://api.github.com/repos/nuclio/nuclio
closed
Python logger should support structured logging
area/processor priority/medium
The Python logger should conform to the Go logger interface: ``` logger.info('Some {0} string', 'formatted') logger.info_with('Unformatted string', arg1='foo', arg2=30) ``` This should be relayed all the way back to the processor.
1.0
Python logger should support structured logging - The Python logger should conform to the Go logger interface: ``` logger.info('Some {0} string', 'formatted') logger.info_with('Unformatted string', arg1='foo', arg2=30) ``` This should be relayed all the way back to the processor.
process
python logger should support structured logging the python logger should conform to the go logger interface logger info some string formatted logger info with unformatted string foo this should be relayed all the way back to the processor
1
239,938
18,288,200,677
IssuesEvent
2021-10-05 12:43:26
Azure/Azure-Functions
https://api.github.com/repos/Azure/Azure-Functions
closed
Update Docs - Linux Dedicated supports WEBSITE_RUN_FROM_PACKAGE
documentation
According to #1076, Linux Dedicated supports `WEBSITE_RUN_FROM_PACKAGE` = <url>, and it's noted in the [Functions Deployment Technologies](https://docs.microsoft.com/en-us/azure/azure-functions/functions-deployment-technologies#deployment-technology-availability) matrix. These pages need updating: - https://github.com/Azure/Azure-Functions/wiki/Azure-Functions-V2-hosting-options#supported-deployment-methods (Deployment Matrix) - https://docs.microsoft.com/en-us/azure/azure-functions/run-functions-from-deployment-package (Top of the page... `The functionality described in this article is not available for function apps running on Linux in an App Service plan.`)
1.0
Update Docs - Linux Dedicated supports WEBSITE_RUN_FROM_PACKAGE - According to #1076, Linux Dedicated supports `WEBSITE_RUN_FROM_PACKAGE` = <url>, and it's noted in the [Functions Deployment Technologies](https://docs.microsoft.com/en-us/azure/azure-functions/functions-deployment-technologies#deployment-technology-availability) matrix. These pages need updating: - https://github.com/Azure/Azure-Functions/wiki/Azure-Functions-V2-hosting-options#supported-deployment-methods (Deployment Matrix) - https://docs.microsoft.com/en-us/azure/azure-functions/run-functions-from-deployment-package (Top of the page... `The functionality described in this article is not available for function apps running on Linux in an App Service plan.`)
non_process
update docs linux dedicated supports website run from package according to linux dedicated supports website run from package and it s noted in the matrix these pages need updating deployment matrix top of the page the functionality described in this article is not available for function apps running on linux in an app service plan
0
21,627
30,028,721,601
IssuesEvent
2023-06-27 08:10:54
h4sh5/pypi-auto-scanner
https://api.github.com/repos/h4sh5/pypi-auto-scanner
opened
reprepbuild 0.10.1 has 1 GuardDog issues
guarddog silent-process-execution
https://pypi.org/project/reprepbuild https://inspector.pypi.io/project/reprepbuild ```{ "dependency": "reprepbuild", "version": "0.10.1", "result": { "issues": 1, "errors": {}, "results": { "silent-process-execution": [ { "location": "RepRepBuild-0.10.1/src/reprepbuild/latexdep.py:83", "code": " subprocess.run(\n args,\n cwd=workdir,\n check=False,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpj4_rm0wv/reprepbuild" } }```
1.0
reprepbuild 0.10.1 has 1 GuardDog issues - https://pypi.org/project/reprepbuild https://inspector.pypi.io/project/reprepbuild ```{ "dependency": "reprepbuild", "version": "0.10.1", "result": { "issues": 1, "errors": {}, "results": { "silent-process-execution": [ { "location": "RepRepBuild-0.10.1/src/reprepbuild/latexdep.py:83", "code": " subprocess.run(\n args,\n cwd=workdir,\n check=False,\n stdin=subprocess.DEVNULL,\n stdout=subprocess.DEVNULL,\n stderr=subprocess.DEVNULL,\n )", "message": "This package is silently executing an external binary, redirecting stdout, stderr and stdin to /dev/null" } ] }, "path": "/tmp/tmpj4_rm0wv/reprepbuild" } }```
process
reprepbuild has guarddog issues dependency reprepbuild version result issues errors results silent process execution location reprepbuild src reprepbuild latexdep py code subprocess run n args n cwd workdir n check false n stdin subprocess devnull n stdout subprocess devnull n stderr subprocess devnull n message this package is silently executing an external binary redirecting stdout stderr and stdin to dev null path tmp reprepbuild
1
19,230
25,384,211,372
IssuesEvent
2022-11-21 20:19:03
carbon-design-system/ibm-cloud-cognitive
https://api.github.com/repos/carbon-design-system/ibm-cloud-cognitive
closed
Add EIM labels
type: process improvement
Add the Experience Issue Management labels for better issue management. Maybe require merging/removing duplicate labels.
1.0
Add EIM labels - Add the Experience Issue Management labels for better issue management. Maybe require merging/removing duplicate labels.
process
add eim labels add the experience issue management labels for better issue management maybe require merging removing duplicate labels
1
206,958
16,062,411,909
IssuesEvent
2021-04-23 14:16:10
getndazn/chaos-squirrel
https://api.github.com/repos/getndazn/chaos-squirrel
opened
Document production deployment guidelines
documentation
### What would you like to discuss? Produce document explaining how Chaos Squirrel should work in pre-production and production environments. For example, we'd expect services to deploy with Chaos Squirrel included, but disabled in production. Clarify why and re-assure how this works (performance, reliability, etc) ### Checklist - [x] I have read the docs
1.0
Document production deployment guidelines - ### What would you like to discuss? Produce document explaining how Chaos Squirrel should work in pre-production and production environments. For example, we'd expect services to deploy with Chaos Squirrel included, but disabled in production. Clarify why and re-assure how this works (performance, reliability, etc) ### Checklist - [x] I have read the docs
non_process
document production deployment guidelines what would you like to discuss produce document explaining how chaos squirrel should work in pre production and production environments for example we d expect services to deploy with chaos squirrel included but disabled in production clarify why and re assure how this works performance reliability etc checklist i have read the docs
0
319
2,765,301,356
IssuesEvent
2015-04-29 19:56:05
scieloorg/search-journals
https://api.github.com/repos/scieloorg/search-journals
opened
Indexar de forma separadas os dados da referência bibliografica
Nova interface Processamento
Atualmente temos os dados da referência bibliografica indexado de forma a não permitir uma granularidade, segue: ```json "fo": [ "Sci. agric. (Piracicaba, Braz.); 67(2); 0-0; 2010-04" ], ``` Devemos alterar isso para itens mais específicos, sugiro: ```json [ {"abrev_source": "Sci. agric. (Piracicaba, Braz.)"}, {"volume": 67}, {"number": 2}, {"start_page": 0}, {"end_page": 0}, {"publication_date": 2010-04}, ] ``` Ano de publicação devemos utilizar recurso do sistema de template para traduzir para meses de forma literal.
1.0
Indexar de forma separadas os dados da referência bibliografica - Atualmente temos os dados da referência bibliografica indexado de forma a não permitir uma granularidade, segue: ```json "fo": [ "Sci. agric. (Piracicaba, Braz.); 67(2); 0-0; 2010-04" ], ``` Devemos alterar isso para itens mais específicos, sugiro: ```json [ {"abrev_source": "Sci. agric. (Piracicaba, Braz.)"}, {"volume": 67}, {"number": 2}, {"start_page": 0}, {"end_page": 0}, {"publication_date": 2010-04}, ] ``` Ano de publicação devemos utilizar recurso do sistema de template para traduzir para meses de forma literal.
process
indexar de forma separadas os dados da referência bibliografica atualmente temos os dados da referência bibliografica indexado de forma a não permitir uma granularidade segue json fo sci agric piracicaba braz devemos alterar isso para itens mais específicos sugiro json abrev source sci agric piracicaba braz volume number start page end page publication date ano de publicação devemos utilizar recurso do sistema de template para traduzir para meses de forma literal
1
20,020
26,493,272,446
IssuesEvent
2023-01-18 01:42:49
apache/arrow-ballista
https://api.github.com/repos/apache/arrow-ballista
closed
Ballista 0.9.0 Release (October 2022)
enhancement development-process
**Is your feature request related to a problem or challenge? Please describe what you are trying to do.** I plan on cutting the release candidate on October 14th. It would be nice to get as many of these features in as possible, but none should block the release. - [ ] Features - [x] https://github.com/apache/arrow-ballista/pull/242 - [x] https://github.com/apache/arrow-ballista/pull/188 - [x] https://github.com/apache/arrow-ballista/pull/290 - [ ] https://github.com/apache/arrow-ballista/issues/363 - [ ] https://github.com/apache/arrow-ballista/issues/368 - [ ] Bug Fixes - [x] https://github.com/apache/arrow-ballista/issues/287 - [x] https://github.com/apache/arrow-ballista/issues/353 - [x] https://github.com/apache/arrow-ballista/issues/360 - [ ] Improve User Guide - [ ] Fix references to DataFusion - [x] https://github.com/apache/arrow-ballista/issues/362 - [x] https://github.com/apache/arrow-ballista/issues/272 - [x] Add scheduler UI screen shot and documentation - [x] Explain how to use query graphs in tuning guide - [ ] Update README - [ ] Update project status information - [ ] Add scheduler UI screenshot - [ ] Release Process - [x] Upgrade to DataFusion 13.0.0 - [ ] https://github.com/apache/arrow-ballista/issues/359 - [ ] Prepare the release candidate - [ ] Publish the release to crates.io once the vote passes - [ ] Publish the Python bindings, if possible. - [ ] Publish the [blog post](https://github.com/apache/arrow-site/pull/257) **Describe the solution you'd like** :point_up: **Describe alternatives you've considered** None **Additional context** None
1.0
Ballista 0.9.0 Release (October 2022) - **Is your feature request related to a problem or challenge? Please describe what you are trying to do.** I plan on cutting the release candidate on October 14th. It would be nice to get as many of these features in as possible, but none should block the release. - [ ] Features - [x] https://github.com/apache/arrow-ballista/pull/242 - [x] https://github.com/apache/arrow-ballista/pull/188 - [x] https://github.com/apache/arrow-ballista/pull/290 - [ ] https://github.com/apache/arrow-ballista/issues/363 - [ ] https://github.com/apache/arrow-ballista/issues/368 - [ ] Bug Fixes - [x] https://github.com/apache/arrow-ballista/issues/287 - [x] https://github.com/apache/arrow-ballista/issues/353 - [x] https://github.com/apache/arrow-ballista/issues/360 - [ ] Improve User Guide - [ ] Fix references to DataFusion - [x] https://github.com/apache/arrow-ballista/issues/362 - [x] https://github.com/apache/arrow-ballista/issues/272 - [x] Add scheduler UI screen shot and documentation - [x] Explain how to use query graphs in tuning guide - [ ] Update README - [ ] Update project status information - [ ] Add scheduler UI screenshot - [ ] Release Process - [x] Upgrade to DataFusion 13.0.0 - [ ] https://github.com/apache/arrow-ballista/issues/359 - [ ] Prepare the release candidate - [ ] Publish the release to crates.io once the vote passes - [ ] Publish the Python bindings, if possible. - [ ] Publish the [blog post](https://github.com/apache/arrow-site/pull/257) **Describe the solution you'd like** :point_up: **Describe alternatives you've considered** None **Additional context** None
process
ballista release october is your feature request related to a problem or challenge please describe what you are trying to do i plan on cutting the release candidate on october it would be nice to get as many of these features in as possible but none should block the release features bug fixes improve user guide fix references to datafusion add scheduler ui screen shot and documentation explain how to use query graphs in tuning guide update readme update project status information add scheduler ui screenshot release process upgrade to datafusion prepare the release candidate publish the release to crates io once the vote passes publish the python bindings if possible publish the describe the solution you d like point up describe alternatives you ve considered none additional context none
1
10,274
13,128,631,376
IssuesEvent
2020-08-06 12:38:18
keep-network/tbtc
https://api.github.com/repos/keep-network/tbtc
closed
Failing npm install
process & client team
`npm install` command fails with a following error: ``` npm ERR! Could not install from "node_modules/web3/bignumber.js@git+https:/github.com/debris/bignumber.js.git#94d7146671b9719e00a09c29b01a691bc85048c2" as it does not contain a package.json file ``` This cause a fail of the [nightly end-to-end test](https://app.circleci.com/pipelines/github/keep-network/local-setup/106/workflows/29060cee-cf29-4890-b8b6-1a451db03b25/jobs/58)
1.0
Failing npm install - `npm install` command fails with a following error: ``` npm ERR! Could not install from "node_modules/web3/bignumber.js@git+https:/github.com/debris/bignumber.js.git#94d7146671b9719e00a09c29b01a691bc85048c2" as it does not contain a package.json file ``` This cause a fail of the [nightly end-to-end test](https://app.circleci.com/pipelines/github/keep-network/local-setup/106/workflows/29060cee-cf29-4890-b8b6-1a451db03b25/jobs/58)
process
failing npm install npm install command fails with a following error npm err could not install from node modules bignumber js git https github com debris bignumber js git as it does not contain a package json file this cause a fail of the
1
1,876
4,700,488,971
IssuesEvent
2016-10-12 18:41:55
metabase/metabase
https://api.github.com/repos/metabase/metabase
closed
MySQL should use sample standard deviation
Correctness Database/MySQL Query Processor Won't Fix
MySQL's `stddev()`is an alias of `stddev_pop()`, i.e. a population standard deviation. All of our other databases that support standard deviations do sample ones, which is available in MySQL as `stddev_samp()`. We should use this function instead to bring MySQL's behavior in line with our other databases. This would probably let us eliminate some of the special-casing in the unit tests around sample deviation aggregations where MySQL returns slightly different results from the other DBs.
1.0
MySQL should use sample standard deviation - MySQL's `stddev()`is an alias of `stddev_pop()`, i.e. a population standard deviation. All of our other databases that support standard deviations do sample ones, which is available in MySQL as `stddev_samp()`. We should use this function instead to bring MySQL's behavior in line with our other databases. This would probably let us eliminate some of the special-casing in the unit tests around sample deviation aggregations where MySQL returns slightly different results from the other DBs.
process
mysql should use sample standard deviation mysql s stddev is an alias of stddev pop i e a population standard deviation all of our other databases that support standard deviations do sample ones which is available in mysql as stddev samp we should use this function instead to bring mysql s behavior in line with our other databases this would probably let us eliminate some of the special casing in the unit tests around sample deviation aggregations where mysql returns slightly different results from the other dbs
1
162,906
20,257,614,709
IssuesEvent
2022-02-15 01:54:10
kapseliboi/mimic
https://api.github.com/repos/kapseliboi/mimic
opened
CVE-2020-36048 (High) detected in engine.io-1.8.3.tgz
security vulnerability
## CVE-2020-36048 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>engine.io-1.8.3.tgz</b></p></summary> <p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p> <p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-1.8.3.tgz">https://registry.npmjs.org/engine.io/-/engine.io-1.8.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/engine.io/package.json</p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - socket.io-1.7.3.tgz - :x: **engine.io-1.8.3.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Engine.IO before 4.0.0 allows attackers to cause a denial of service (resource consumption) via a POST request to the long polling transport. <p>Publish Date: 2021-01-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36048>CVE-2020-36048</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048</a></p> <p>Release Date: 2021-01-08</p> <p>Fix Resolution (engine.io): 4.0.0-alpha.0</p> <p>Direct dependency fix Resolution (karma): 6.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-36048 (High) detected in engine.io-1.8.3.tgz - ## CVE-2020-36048 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>engine.io-1.8.3.tgz</b></p></summary> <p>The realtime engine behind Socket.IO. Provides the foundation of a bidirectional connection between client and server</p> <p>Library home page: <a href="https://registry.npmjs.org/engine.io/-/engine.io-1.8.3.tgz">https://registry.npmjs.org/engine.io/-/engine.io-1.8.3.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/engine.io/package.json</p> <p> Dependency Hierarchy: - karma-1.7.0.tgz (Root Library) - socket.io-1.7.3.tgz - :x: **engine.io-1.8.3.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Engine.IO before 4.0.0 allows attackers to cause a denial of service (resource consumption) via a POST request to the long polling transport. <p>Publish Date: 2021-01-08 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-36048>CVE-2020-36048</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-36048</a></p> <p>Release Date: 2021-01-08</p> <p>Fix Resolution (engine.io): 4.0.0-alpha.0</p> <p>Direct dependency fix Resolution (karma): 6.0.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in engine io tgz cve high severity vulnerability vulnerable library engine io tgz the realtime engine behind socket io provides the foundation of a bidirectional connection between client and server library home page a href path to dependency file package json path to vulnerable library node modules engine io package json dependency hierarchy karma tgz root library socket io tgz x engine io tgz vulnerable library found in base branch master vulnerability details engine io before allows attackers to cause a denial of service resource consumption via a post request to the long polling transport publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution engine io alpha direct dependency fix resolution karma step up your open source security game with whitesource
0
45,117
11,590,702,839
IssuesEvent
2020-02-24 07:36:29
rokups/rbfx
https://api.github.com/repos/rokups/rbfx
closed
WEB is not exported as C++ define
area: build system bug
Apparently WEB is not exported as define. Therefore, we may have invalid logics https://github.com/rokups/rbfx/blob/master/Source/Urho3D/IO/Log.cpp#L237 This code snippet doesn't fire when I build Web. ``` #if defined(WEB) #error fu #endif ``` I wonder if other places are broken too
1.0
WEB is not exported as C++ define - Apparently WEB is not exported as define. Therefore, we may have invalid logics https://github.com/rokups/rbfx/blob/master/Source/Urho3D/IO/Log.cpp#L237 This code snippet doesn't fire when I build Web. ``` #if defined(WEB) #error fu #endif ``` I wonder if other places are broken too
non_process
web is not exported as c define apparently web is not exported as define therefore we may have invalid logics this code snippet doesn t fire when i build web if defined web error fu endif i wonder if other places are broken too
0
18,112
24,143,779,617
IssuesEvent
2022-09-21 16:51:31
qgis/QGIS
https://api.github.com/repos/qgis/QGIS
closed
Suspected reduced precision in "QGIS" vector Check Validity method causing false negatives.
Feedback Processing Bug
### What is the bug or the crash? QGIS will report the polygon `POLYGON ((-998000 -292000,-998000 -284096.11976889,-992722.998884085 -285410.073094352,-992299.787312271 -281787.828810877,-993492.945112339 -292000,-998000 -292000))` (ESRI:102003 - USA_Contiguous_Albers_Equal_Area_Conic coordinate system) as invalid using its own method, claiming to suffer from an intersection. I noticed I was able to trigger the same error manually in Python when performing intersection tests using 4 byte floats instead of the native 8 byte doubles. Does QGIS perform these validity checks with floats instead of doubles? If so, I recommend to, at the least, respect the storage size when choosing what level of precision to preform arithmetic on. Evidenced below, GEOS via OGR and Shapely confirm the polygon's validity, presumably due to the increased precision. I leafed through 3 pages of related issues to see if this is a duplicate but did not find a match. ```python import sys import osgeo from numpy import float32 as f32 from osgeo import ogr import shapely import shapely.wkt ogr.UseExceptions() # QGIS: This polygon is invalid - segments 1 and 3 of line 0 intersect at -992723, -285410 #POLYGON ((-998000 -292000,-998000 -284096.11976889,-992722.998884085 -285410.073094352,-992299.787312271 -281787.828810877,-993492.945112339 -292000,-998000 -292000)) def main(): # floating point tolerance for coordinate system epsilon = float(0.0001) print(f'epsilon = {epsilon}') wkt = 'POLYGON ((-998000 -292000,-998000 -284096.11976889,-992722.998884085 -285410.073094352,-992299.787312271 -281787.828810877,-993492.945112339 -292000,-998000 -292000))' # OGR sez it's valid try: polygon = ogr.CreateGeometryFromWkt(wkt) except Exception as e: print(f'CreateGeometryFromWkt raised exception {str(e)}') sys.exit(1) if polygon is None: print(f'CreateGeometryFromWkt failed') sys.exit(1) if not polygon.IsValid(): print(f'CreateGeometryFromWkt generated an invalid polygon') sys.exit(1) print(f'CreateGeometryFromWkt generated a polygon of area {polygon.Area()}') # Shapely also sez it's valid try: spolygon = shapely.wkt.loads(wkt) except Exception as e: print(f'shapely.wkt.loads raised exception {str(e)}') sys.exit(1) if spolygon is None: print(f'shapely.wkt.loads failed') sys.exit(1) if not spolygon.is_valid: print(f'shapely.wkt.loads generated an invalid polygon') sys.exit(1) print(f'shapely.wkt.loads generated a polygon of area {spolygon.area}') # form the line segments between each pair of adjacent vertices ring = polygon.GetGeometryRef(0) npts = ring.GetPointCount() linesegs = [] for p in range(npts - 1): # form the line segment between vertex p and p+1 pt1 = ring.GetPoint(p) pt2 = ring.GetPoint(p+1) lineseg = ogr.Geometry(ogr.wkbLineString) lineseg.AddPoint(pt1[0],pt1[1]) lineseg.AddPoint(pt2[0],pt2[1]) linesegs.append(lineseg) # check each line segment for crossing with all of the others- Crosses does not complain for l in range(npts - 1): this_lineseg = linesegs[l] for m in range(npts - 1): if m!=l: if this_lineseg.Crosses(linesegs[m]): crossing_point = this_lineseg.Intersection(linesegs[m]) print(f'Line segment {l+1} crosses line segment {m+1} at point {crossing_point.GetPoint(0)}') # the crossing lines connect points 1 and 2 (line A), and points 3 and 4 (line B), indexing from 0 x1, y1, _ = ring.GetPoint(1) x2, y2, _ = ring.GetPoint(2) x3, y3, _ = ring.GetPoint(3) x4, y4, _ = ring.GetPoint(4) # slope and intercept for line containing each segment slopeA = (y2 - y1)/(x2 - x1) intA = y1 - slopeA*x1 slopeB = (y4 - y3)/(x4 - x3) intB = y3 - slopeB*x3 # intersection point between the lines ipx = (intA - intB)/(slopeB-slopeA) ipy = slopeA*ipx + intA print(f"Potential crossing point at ({ipx},{ipy})") # is the intersect point between the first line segment endpoints? # if so, it's a crossing point crossprod = (ipy - y1)*(x2 - x1) - (ipx - x1)*(y2 - y1) print(f'crossprod = {crossprod}') if abs(crossprod) > epsilon: print(f'The lines do not cross, cross product {crossprod} is greater than fp zero {epsilon}') sys.exit(0) dotprod = (ipx - x1)*(x2 - x1) + (ipy - y1)*(y2 - y1) if dotprod < 0: print(f'The lines do not cross, dot product {dotprod} is < 0') sys.exit(0) squaredlength = (x2 - x1)*(x2 - x1) + (y2 - y1)*(y2 - y1) if dotprod > squaredlength: diff = dotprod - squaredlength print(f'The lines do not cross, dot product {dotprod} is > squared length {squaredlength} (diff {diff})') else: print(f'The lines intersect at point ({ipx},{ipy})') # slope and intercept for line containing each segment # 4 byte floats x1, y1 = f32(x1), f32(y1) x2, y2 = f32(x2), f32(y2) x3, y3 = f32(x3), f32(y3) x4, y4 = f32(x4), f32(y4) slopeA = (y2 - y1)/(x2 - x1) intA = y1 - slopeA*x1 slopeB = (y4 - y3)/(x4 - x3) intB = y3 - slopeB*x3 # intersection point between the lines ipx = (intA - intB)/(slopeB-slopeA) ipy = slopeA*ipx + intA print(f"Potential crossing point at ({ipx},{ipy})") # is the intersect point between the first line segment endpoints? # if so, it's a crossing point crossprod = (ipy - y1)*(x2 - x1) - (ipx - x1)*(y2 - y1) print(f'crossprod = {crossprod}') if abs(crossprod) > epsilon: print(f'The lines do not cross, cross product {crossprod} is greater than fp zero {epsilon}') sys.exit(0) dotprod = (ipx - x1)*(x2 - x1) + (ipy - y1)*(y2 - y1) if dotprod < 0: print(f'The lines do not cross, dot product {dotprod} is < 0') sys.exit(0) squaredlength = (x2 - x1)*(x2 - x1) + (y2 - y1)*(y2 - y1) if dotprod > squaredlength: diff = dotprod - squaredlength print(f'The lines do not cross, dot product {dotprod} is > squared length {squaredlength} (diff {diff})') sys.exit(0) print(f'The lines intersect at point ({ipx},{ipy})') if __name__ == '__main__': main() ``` ### Steps to reproduce the issue Happy to upload a GPKG of the polygon. Vector -> Geometry Tools -> Check Validity -> QGIS method (Does not occur with neither GEOS nor manual algebraic method) ### Versions Tried 3.22.8 LTR and 3.26.3 PR on OSX 12.6 and ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
1.0
Suspected reduced precision in "QGIS" vector Check Validity method causing false negatives. - ### What is the bug or the crash? QGIS will report the polygon `POLYGON ((-998000 -292000,-998000 -284096.11976889,-992722.998884085 -285410.073094352,-992299.787312271 -281787.828810877,-993492.945112339 -292000,-998000 -292000))` (ESRI:102003 - USA_Contiguous_Albers_Equal_Area_Conic coordinate system) as invalid using its own method, claiming to suffer from an intersection. I noticed I was able to trigger the same error manually in Python when performing intersection tests using 4 byte floats instead of the native 8 byte doubles. Does QGIS perform these validity checks with floats instead of doubles? If so, I recommend to, at the least, respect the storage size when choosing what level of precision to preform arithmetic on. Evidenced below, GEOS via OGR and Shapely confirm the polygon's validity, presumably due to the increased precision. I leafed through 3 pages of related issues to see if this is a duplicate but did not find a match. ```python import sys import osgeo from numpy import float32 as f32 from osgeo import ogr import shapely import shapely.wkt ogr.UseExceptions() # QGIS: This polygon is invalid - segments 1 and 3 of line 0 intersect at -992723, -285410 #POLYGON ((-998000 -292000,-998000 -284096.11976889,-992722.998884085 -285410.073094352,-992299.787312271 -281787.828810877,-993492.945112339 -292000,-998000 -292000)) def main(): # floating point tolerance for coordinate system epsilon = float(0.0001) print(f'epsilon = {epsilon}') wkt = 'POLYGON ((-998000 -292000,-998000 -284096.11976889,-992722.998884085 -285410.073094352,-992299.787312271 -281787.828810877,-993492.945112339 -292000,-998000 -292000))' # OGR sez it's valid try: polygon = ogr.CreateGeometryFromWkt(wkt) except Exception as e: print(f'CreateGeometryFromWkt raised exception {str(e)}') sys.exit(1) if polygon is None: print(f'CreateGeometryFromWkt failed') sys.exit(1) if not polygon.IsValid(): print(f'CreateGeometryFromWkt generated an invalid polygon') sys.exit(1) print(f'CreateGeometryFromWkt generated a polygon of area {polygon.Area()}') # Shapely also sez it's valid try: spolygon = shapely.wkt.loads(wkt) except Exception as e: print(f'shapely.wkt.loads raised exception {str(e)}') sys.exit(1) if spolygon is None: print(f'shapely.wkt.loads failed') sys.exit(1) if not spolygon.is_valid: print(f'shapely.wkt.loads generated an invalid polygon') sys.exit(1) print(f'shapely.wkt.loads generated a polygon of area {spolygon.area}') # form the line segments between each pair of adjacent vertices ring = polygon.GetGeometryRef(0) npts = ring.GetPointCount() linesegs = [] for p in range(npts - 1): # form the line segment between vertex p and p+1 pt1 = ring.GetPoint(p) pt2 = ring.GetPoint(p+1) lineseg = ogr.Geometry(ogr.wkbLineString) lineseg.AddPoint(pt1[0],pt1[1]) lineseg.AddPoint(pt2[0],pt2[1]) linesegs.append(lineseg) # check each line segment for crossing with all of the others- Crosses does not complain for l in range(npts - 1): this_lineseg = linesegs[l] for m in range(npts - 1): if m!=l: if this_lineseg.Crosses(linesegs[m]): crossing_point = this_lineseg.Intersection(linesegs[m]) print(f'Line segment {l+1} crosses line segment {m+1} at point {crossing_point.GetPoint(0)}') # the crossing lines connect points 1 and 2 (line A), and points 3 and 4 (line B), indexing from 0 x1, y1, _ = ring.GetPoint(1) x2, y2, _ = ring.GetPoint(2) x3, y3, _ = ring.GetPoint(3) x4, y4, _ = ring.GetPoint(4) # slope and intercept for line containing each segment slopeA = (y2 - y1)/(x2 - x1) intA = y1 - slopeA*x1 slopeB = (y4 - y3)/(x4 - x3) intB = y3 - slopeB*x3 # intersection point between the lines ipx = (intA - intB)/(slopeB-slopeA) ipy = slopeA*ipx + intA print(f"Potential crossing point at ({ipx},{ipy})") # is the intersect point between the first line segment endpoints? # if so, it's a crossing point crossprod = (ipy - y1)*(x2 - x1) - (ipx - x1)*(y2 - y1) print(f'crossprod = {crossprod}') if abs(crossprod) > epsilon: print(f'The lines do not cross, cross product {crossprod} is greater than fp zero {epsilon}') sys.exit(0) dotprod = (ipx - x1)*(x2 - x1) + (ipy - y1)*(y2 - y1) if dotprod < 0: print(f'The lines do not cross, dot product {dotprod} is < 0') sys.exit(0) squaredlength = (x2 - x1)*(x2 - x1) + (y2 - y1)*(y2 - y1) if dotprod > squaredlength: diff = dotprod - squaredlength print(f'The lines do not cross, dot product {dotprod} is > squared length {squaredlength} (diff {diff})') else: print(f'The lines intersect at point ({ipx},{ipy})') # slope and intercept for line containing each segment # 4 byte floats x1, y1 = f32(x1), f32(y1) x2, y2 = f32(x2), f32(y2) x3, y3 = f32(x3), f32(y3) x4, y4 = f32(x4), f32(y4) slopeA = (y2 - y1)/(x2 - x1) intA = y1 - slopeA*x1 slopeB = (y4 - y3)/(x4 - x3) intB = y3 - slopeB*x3 # intersection point between the lines ipx = (intA - intB)/(slopeB-slopeA) ipy = slopeA*ipx + intA print(f"Potential crossing point at ({ipx},{ipy})") # is the intersect point between the first line segment endpoints? # if so, it's a crossing point crossprod = (ipy - y1)*(x2 - x1) - (ipx - x1)*(y2 - y1) print(f'crossprod = {crossprod}') if abs(crossprod) > epsilon: print(f'The lines do not cross, cross product {crossprod} is greater than fp zero {epsilon}') sys.exit(0) dotprod = (ipx - x1)*(x2 - x1) + (ipy - y1)*(y2 - y1) if dotprod < 0: print(f'The lines do not cross, dot product {dotprod} is < 0') sys.exit(0) squaredlength = (x2 - x1)*(x2 - x1) + (y2 - y1)*(y2 - y1) if dotprod > squaredlength: diff = dotprod - squaredlength print(f'The lines do not cross, dot product {dotprod} is > squared length {squaredlength} (diff {diff})') sys.exit(0) print(f'The lines intersect at point ({ipx},{ipy})') if __name__ == '__main__': main() ``` ### Steps to reproduce the issue Happy to upload a GPKG of the polygon. Vector -> Geometry Tools -> Check Validity -> QGIS method (Does not occur with neither GEOS nor manual algebraic method) ### Versions Tried 3.22.8 LTR and 3.26.3 PR on OSX 12.6 and ### Supported QGIS version - [X] I'm running a supported QGIS version according to the roadmap. ### New profile - [X] I tried with a new QGIS profile ### Additional context _No response_
process
suspected reduced precision in qgis vector check validity method causing false negatives what is the bug or the crash qgis will report the polygon polygon esri usa contiguous albers equal area conic coordinate system as invalid using its own method claiming to suffer from an intersection i noticed i was able to trigger the same error manually in python when performing intersection tests using byte floats instead of the native byte doubles does qgis perform these validity checks with floats instead of doubles if so i recommend to at the least respect the storage size when choosing what level of precision to preform arithmetic on evidenced below geos via ogr and shapely confirm the polygon s validity presumably due to the increased precision i leafed through pages of related issues to see if this is a duplicate but did not find a match python import sys import osgeo from numpy import as from osgeo import ogr import shapely import shapely wkt ogr useexceptions qgis this polygon is invalid segments and of line intersect at polygon def main floating point tolerance for coordinate system epsilon float print f epsilon epsilon wkt polygon ogr sez it s valid try polygon ogr creategeometryfromwkt wkt except exception as e print f creategeometryfromwkt raised exception str e sys exit if polygon is none print f creategeometryfromwkt failed sys exit if not polygon isvalid print f creategeometryfromwkt generated an invalid polygon sys exit print f creategeometryfromwkt generated a polygon of area polygon area shapely also sez it s valid try spolygon shapely wkt loads wkt except exception as e print f shapely wkt loads raised exception str e sys exit if spolygon is none print f shapely wkt loads failed sys exit if not spolygon is valid print f shapely wkt loads generated an invalid polygon sys exit print f shapely wkt loads generated a polygon of area spolygon area form the line segments between each pair of adjacent vertices ring polygon getgeometryref npts ring getpointcount linesegs for p in range npts form the line segment between vertex p and p ring getpoint p ring getpoint p lineseg ogr geometry ogr wkblinestring lineseg addpoint lineseg addpoint linesegs append lineseg check each line segment for crossing with all of the others crosses does not complain for l in range npts this lineseg linesegs for m in range npts if m l if this lineseg crosses linesegs crossing point this lineseg intersection linesegs print f line segment l crosses line segment m at point crossing point getpoint the crossing lines connect points and line a and points and line b indexing from ring getpoint ring getpoint ring getpoint ring getpoint slope and intercept for line containing each segment slopea inta slopea slopeb intb slopeb intersection point between the lines ipx inta intb slopeb slopea ipy slopea ipx inta print f potential crossing point at ipx ipy is the intersect point between the first line segment endpoints if so it s a crossing point crossprod ipy ipx print f crossprod crossprod if abs crossprod epsilon print f the lines do not cross cross product crossprod is greater than fp zero epsilon sys exit dotprod ipx ipy if dotprod print f the lines do not cross dot product dotprod is sys exit squaredlength if dotprod squaredlength diff dotprod squaredlength print f the lines do not cross dot product dotprod is squared length squaredlength diff diff else print f the lines intersect at point ipx ipy slope and intercept for line containing each segment byte floats slopea inta slopea slopeb intb slopeb intersection point between the lines ipx inta intb slopeb slopea ipy slopea ipx inta print f potential crossing point at ipx ipy is the intersect point between the first line segment endpoints if so it s a crossing point crossprod ipy ipx print f crossprod crossprod if abs crossprod epsilon print f the lines do not cross cross product crossprod is greater than fp zero epsilon sys exit dotprod ipx ipy if dotprod print f the lines do not cross dot product dotprod is sys exit squaredlength if dotprod squaredlength diff dotprod squaredlength print f the lines do not cross dot product dotprod is squared length squaredlength diff diff sys exit print f the lines intersect at point ipx ipy if name main main steps to reproduce the issue happy to upload a gpkg of the polygon vector geometry tools check validity qgis method does not occur with neither geos nor manual algebraic method versions tried ltr and pr on osx and supported qgis version i m running a supported qgis version according to the roadmap new profile i tried with a new qgis profile additional context no response
1
290,434
32,078,885,503
IssuesEvent
2023-09-25 12:48:37
pazhanivel07/Packages_apps_Nfc_AOSP_10_r33
https://api.github.com/repos/pazhanivel07/Packages_apps_Nfc_AOSP_10_r33
opened
CVE-2023-20945 (High) detected in Nfcandroid-10.0.0_r31
Mend: dependency security vulnerability
## CVE-2023-20945 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Nfcandroid-10.0.0_r31</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Nfc>https://android.googlesource.com/platform/packages/apps/Nfc</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/Packages_apps_Nfc_AOSP_10_r33/commit/844e6886a9ac506964bf64428ef202c74ef29dbf">844e6886a9ac506964bf64428ef202c74ef29dbf</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/nci/jni/extns/pn54x/src/mifare/phNxpExtns_MifareStd.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In phNciNfc_MfCreateXchgDataHdr of phNxpExtns_MifareStd.cpp, there is a possible out of bounds write due to a missing bounds check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-246932269 <p>Publish Date: 2023-02-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-20945>CVE-2023-20945</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2023-20945 (High) detected in Nfcandroid-10.0.0_r31 - ## CVE-2023-20945 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Nfcandroid-10.0.0_r31</b></p></summary> <p> <p>Library home page: <a href=https://android.googlesource.com/platform/packages/apps/Nfc>https://android.googlesource.com/platform/packages/apps/Nfc</a></p> <p>Found in HEAD commit: <a href="https://github.com/pazhanivel07/Packages_apps_Nfc_AOSP_10_r33/commit/844e6886a9ac506964bf64428ef202c74ef29dbf">844e6886a9ac506964bf64428ef202c74ef29dbf</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/nci/jni/extns/pn54x/src/mifare/phNxpExtns_MifareStd.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> In phNciNfc_MfCreateXchgDataHdr of phNxpExtns_MifareStd.cpp, there is a possible out of bounds write due to a missing bounds check. This could lead to local escalation of privilege with no additional execution privileges needed. User interaction is not needed for exploitation.Product: AndroidVersions: Android-10Android ID: A-246932269 <p>Publish Date: 2023-02-28 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2023-20945>CVE-2023-20945</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_process
cve high detected in nfcandroid cve high severity vulnerability vulnerable library nfcandroid library home page a href found in head commit a href found in base branch main vulnerable source files nci jni extns src mifare phnxpextns mifarestd cpp vulnerability details in phncinfc mfcreatexchgdatahdr of phnxpextns mifarestd cpp there is a possible out of bounds write due to a missing bounds check this could lead to local escalation of privilege with no additional execution privileges needed user interaction is not needed for exploitation product androidversions android id a publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href step up your open source security game with mend
0
91,131
10,710,008,092
IssuesEvent
2019-10-25 00:19:18
Varying-Vagrant-Vagrants/VVV
https://api.github.com/repos/Varying-Vagrant-Vagrants/VVV
closed
Document vagrant disk size in changelog
documentation hacktoberfest
Related to #1915 we need to document this in the changelog for 3.2
1.0
Document vagrant disk size in changelog - Related to #1915 we need to document this in the changelog for 3.2
non_process
document vagrant disk size in changelog related to we need to document this in the changelog for
0
39,674
6,760,668,661
IssuesEvent
2017-10-24 21:29:14
18F/web-design-standards
https://api.github.com/repos/18F/web-design-standards
closed
Clarify what step #4 of CONTRIBUTING.md
[Priority] Minor [Type] Documentation [Type] Enhancement
<!-- --> ## Issue type Documentation ## Description Documentation could be more clear for new comers. This is like text looks right now: "Ensure that your contribution works via npm, if applicable. See below under Install the package locally via nom-link." **My questions:** 1. How to know if it is "applicable"? 2. During step 4, should we building the project? Should we move immediately to the next section: "Building the project locally with gulp"? 3. And there is no link provided to "Install the package locally via nom-link." Should that link to the following section? ## Steps to reproduce the issue none. Just opening Contributing.md ## Additional information [optional] none
1.0
Clarify what step #4 of CONTRIBUTING.md - <!-- --> ## Issue type Documentation ## Description Documentation could be more clear for new comers. This is like text looks right now: "Ensure that your contribution works via npm, if applicable. See below under Install the package locally via nom-link." **My questions:** 1. How to know if it is "applicable"? 2. During step 4, should we building the project? Should we move immediately to the next section: "Building the project locally with gulp"? 3. And there is no link provided to "Install the package locally via nom-link." Should that link to the following section? ## Steps to reproduce the issue none. Just opening Contributing.md ## Additional information [optional] none
non_process
clarify what step of contributing md issue type documentation description documentation could be more clear for new comers this is like text looks right now ensure that your contribution works via npm if applicable see below under install the package locally via nom link my questions how to know if it is applicable during step should we building the project should we move immediately to the next section building the project locally with gulp and there is no link provided to install the package locally via nom link should that link to the following section steps to reproduce the issue none just opening contributing md additional information none
0
686,144
23,478,664,334
IssuesEvent
2022-08-17 08:38:28
opendatahub-io/odh-dashboard
https://api.github.com/repos/opendatahub-io/odh-dashboard
closed
[KFNBC]: Add GPU soft anti-affinity to KFNBC
kind/enhancement feature/notebook-controller priority/high
### Feature description Add a feature to KFNBC for notebooks that don't request a GPU to avoid them Port of https://issues.redhat.com/browse/RHODS-3074 ### Describe alternatives you've considered _No response_ ### Anything else? _No response_
1.0
[KFNBC]: Add GPU soft anti-affinity to KFNBC - ### Feature description Add a feature to KFNBC for notebooks that don't request a GPU to avoid them Port of https://issues.redhat.com/browse/RHODS-3074 ### Describe alternatives you've considered _No response_ ### Anything else? _No response_
non_process
add gpu soft anti affinity to kfnbc feature description add a feature to kfnbc for notebooks that don t request a gpu to avoid them port of describe alternatives you ve considered no response anything else no response
0
754,732
26,399,617,020
IssuesEvent
2023-01-12 23:17:10
GoogleCloudPlatform/emblem
https://api.github.com/repos/GoogleCloudPlatform/emblem
closed
Clean up Client app setup script and README instructions
type: feature request priority: p2
- We should explicitly run ./scripts/configure_auth.sh before ./scripts/startup.sh. - `npm build` should be `npm run build` in setup instructions. - DonationPage.js should be donationPage.js as linked in src/containers/donation/donation-page.js
1.0
Clean up Client app setup script and README instructions - - We should explicitly run ./scripts/configure_auth.sh before ./scripts/startup.sh. - `npm build` should be `npm run build` in setup instructions. - DonationPage.js should be donationPage.js as linked in src/containers/donation/donation-page.js
non_process
clean up client app setup script and readme instructions we should explicitly run scripts configure auth sh before scripts startup sh npm build should be npm run build in setup instructions donationpage js should be donationpage js as linked in src containers donation donation page js
0
385,689
26,649,351,845
IssuesEvent
2023-01-25 12:32:18
lumapu/ahoy
https://api.github.com/repos/lumapu/ahoy
closed
MQTT -> Mosquitto -> Telegraf -> InfluxDB
documentation question resolved
Hi, habe Probleme mit der Konfiguration von Telegraf betreffend JSON Objekte der Ahoy. Ich würde ja gerne das Plugin im Telegraf Consumer erstellen. Mir fehlt aber die Basis wie die JSON Strukturen am Input von Telegraf aussehen. Ich sehe zwar Strukturen im MQTTFx, weiss aber nicht wirklich ob diese passen. Hat jemand schon mal das telegraf.conf File entsprechend konfiguriert? Wäre für Hinweise, Tipps, oder gar einen Auszug aus einem telegraf.conf File dankbar.
1.0
MQTT -> Mosquitto -> Telegraf -> InfluxDB - Hi, habe Probleme mit der Konfiguration von Telegraf betreffend JSON Objekte der Ahoy. Ich würde ja gerne das Plugin im Telegraf Consumer erstellen. Mir fehlt aber die Basis wie die JSON Strukturen am Input von Telegraf aussehen. Ich sehe zwar Strukturen im MQTTFx, weiss aber nicht wirklich ob diese passen. Hat jemand schon mal das telegraf.conf File entsprechend konfiguriert? Wäre für Hinweise, Tipps, oder gar einen Auszug aus einem telegraf.conf File dankbar.
non_process
mqtt mosquitto telegraf influxdb hi habe probleme mit der konfiguration von telegraf betreffend json objekte der ahoy ich würde ja gerne das plugin im telegraf consumer erstellen mir fehlt aber die basis wie die json strukturen am input von telegraf aussehen ich sehe zwar strukturen im mqttfx weiss aber nicht wirklich ob diese passen hat jemand schon mal das telegraf conf file entsprechend konfiguriert wäre für hinweise tipps oder gar einen auszug aus einem telegraf conf file dankbar
0
5,620
12,828,426,789
IssuesEvent
2020-07-06 20:31:09
dotnet/docs
https://api.github.com/repos/dotnet/docs
closed
Missing tag in example output
:book: guide - Blazor :books: Area - .NET Architecture Guide doc-bug
In the template parameter example component output, it appears to be missing the &lt;p&gt; tag from the item template. I would expect to see ul &gt; li &gt; p instead of ul &gt; li. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 60c195c0-99f6-64f5-66a3-4e69109dc8ee * Version Independent ID: 24368808-4b90-f75a-e887-eb72a1c85f5b * Content: [Build reusable UI components with Blazor](https://docs.microsoft.com/en-us/dotnet/architecture/blazor-for-web-forms-developers/components#feedback) * Content Source: [docs/architecture/blazor-for-web-forms-developers/components.md](https://github.com/dotnet/docs/blob/master/docs/architecture/blazor-for-web-forms-developers/components.md) * Product: **dotnet-architecture** * Technology: **blazor** * GitHub Login: @danroth27 * Microsoft Alias: **daroth**
1.0
Missing tag in example output - In the template parameter example component output, it appears to be missing the &lt;p&gt; tag from the item template. I would expect to see ul &gt; li &gt; p instead of ul &gt; li. --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: 60c195c0-99f6-64f5-66a3-4e69109dc8ee * Version Independent ID: 24368808-4b90-f75a-e887-eb72a1c85f5b * Content: [Build reusable UI components with Blazor](https://docs.microsoft.com/en-us/dotnet/architecture/blazor-for-web-forms-developers/components#feedback) * Content Source: [docs/architecture/blazor-for-web-forms-developers/components.md](https://github.com/dotnet/docs/blob/master/docs/architecture/blazor-for-web-forms-developers/components.md) * Product: **dotnet-architecture** * Technology: **blazor** * GitHub Login: @danroth27 * Microsoft Alias: **daroth**
non_process
missing tag in example output in the template parameter example component output it appears to be missing the lt p gt tag from the item template i would expect to see ul gt li gt p instead of ul gt li document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id content content source product dotnet architecture technology blazor github login microsoft alias daroth
0
12,469
14,939,718,807
IssuesEvent
2021-01-25 17:18:07
xatkit-bot-platform/xatkit-runtime
https://api.github.com/repos/xatkit-bot-platform/xatkit-runtime
closed
Processor interfaces could be @FunctionalInterface
Enhancement Low Priority Processors
Each interface contains two methods: - `init()`: there is a default implementation for this that should be enough for most of the use cases. If not then it's a good reason to actually implement the interface. - `process(intent/input, context)`: this method could be the one to specify as part of the functional interface. This would allow code like this: ```java IntentPostProcessor processor = (intent, context) -> [return a new version of intent] ```
1.0
Processor interfaces could be @FunctionalInterface - Each interface contains two methods: - `init()`: there is a default implementation for this that should be enough for most of the use cases. If not then it's a good reason to actually implement the interface. - `process(intent/input, context)`: this method could be the one to specify as part of the functional interface. This would allow code like this: ```java IntentPostProcessor processor = (intent, context) -> [return a new version of intent] ```
process
processor interfaces could be functionalinterface each interface contains two methods init there is a default implementation for this that should be enough for most of the use cases if not then it s a good reason to actually implement the interface process intent input context this method could be the one to specify as part of the functional interface this would allow code like this java intentpostprocessor processor intent context
1
21,410
29,351,206,194
IssuesEvent
2023-05-27 00:34:46
devssa/onde-codar-em-salvador
https://api.github.com/repos/devssa/onde-codar-em-salvador
closed
[Belo Horizonte, Minas Gerais, Brazil] Systems Development Corporate Coordinator na Coodesh
SALVADOR DESENVOLVIMENTO DE SOFTWARE TELECOM REQUISITOS PROCESSOS INOVAÇÃO GITHUB CI EXCEL UMA QUALIDADE ITIL METODOLOGIAS ÁGEIS PMBOK SUPORTE ALOCADO Stale
## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/coordenador-corporativo-desenvolvimento-sistemas-152805091?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Telemont</strong> está em busca de <strong><ins>Systems Development Corporate Coordinator</ins></strong> para compor seu time!</p> <p>Há mais de 45 anos, a Telemont realiza serviços essenciais para o desenvolvimento do Brasil, se destacando nos segmentos de telecom, energia e TI. Atuamos de ponta a ponta para oferecer comunicação de voz, banda larga e dados, tecnologia da informação, transporte multimídia e gestão de sistemas de energia, conectando milhões de brasileiros. Contamos com um time de profissionais que traduzem em ações os nossos valores, garantindo a excelência operacional e a qualidade em seus atendimentos.&nbsp;</p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Assegurar a evolução dos sistemas por meio da coordenação dos trabalhos de: refinamento das necessidades dos usuários, planejamento do desenvolvimento, testes, homologação e disponibilização em produção;</li> <li>Desenvolver novos projetos de sistemas de informação alinhados aos objetivos da empresa, por meio de prospecção das soluções, escolha de fornecedores, planejamento, alocação de equipe e acompanhamento da implementação;</li> <li>Contribuir na identificação de melhorias em processos da área ou do próprio sistema, através da análise crítica dos mesmos propondo alternativas e/ou propostas de solução;</li> <li>Assessorar, quando necessário, a escolha correta de fornecedores de desenvolvimento de sistemas, através da análise dos mesmos;</li> <li>Supervisionar o ambiente dos sistemas de informação da empresa por meio da atuação junto às áreas de negócio definindo e propondo melhorias sistêmicas ao processo e alinhando as necessidades com a equipe responsável ou com consultores terceirizados;</li> <li>Sustentar os sistemas da empresa e suas integrações, gerenciando a qualidade do suporte a usuários, fila, prazo e idade de chamados bem como diagnosticando problemas e realizando implantação da solução definitiva encontrada;</li> <li>Gerir a esteira de desenvolvimentos mantendo alta performance e produtividade, alinhando prioridades e mantendo os usuários-chave atualizados sobre a evolução das demandas;</li> <li>Executar outras atividades de mesma natureza, nível de complexidade e responsabilidade.</li> </ul> ## Telemont: <p>Há mais de 45 anos, a Telemont realiza serviços essenciais para o desenvolvimento do Brasil, se destacando nos segmentos de telecom, energia e TI. Atuamos de ponta a ponta para oferecer comunicação de voz, banda larga e dados, tecnologia da informação, transporte multimídia e gestão de sistemas de energia, conectando milhões de brasileiros. Contamos com um time de profissionais que traduzem em ações os nossos valores, garantindo a excelência operacional e a qualidade em seus atendimentos. Além disso, somos uma empresa que busca promover uma maior diversidade e inclusão, pois acreditamos que a pluralidade das nossas equipes traz mais conhecimento, respeito às múltiplas possibilidades de vivência, inovação e integração entre as áreas. Essa incansável procura por vencer desafios e conquistar novas oportunidades faz a Telemont ser uma empresa onde cada pessoa tem a possibilidade de crescer e se desenvolver profissionalmente, construindo assim, junto com a organização, uma trajetória de sucesso. Nossos colaboradores são orientados por valores que norteiam todas as nossas ações e garantem a sustentabilidade do negócio, bem como de quem faz parte dele.</p> </p> ## Habilidades: - Metodologias ágeis - Gestão de times de tecnologia - Gestão de contratos - Gestão e Negociação com Cliente ## Local: Belo Horizonte, Minas Gerais, Brazil ## Requisitos: - Superior completo em Ciência da Computação ou cursos relacionados; - Domínio em processo de desenvolvimento de software; - Noção de supervisão profissional de projetos (PMBOK); - Facilidade com gestão de pessoas; - Conhecimento em padronização de processos e procedimentos; - Conhecimento em Metodologia ITIL: Incidentes, Demandas, Problemas. ## Benefícios: - VR; - VT; - Seguro de vida; - Gympass; - Descontos estabelecimentos conveniados; - Plano de saúde; - Plano odontológico. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Systems Development Corporate Coordinator na Telemont](https://coodesh.com/vagas/coordenador-corporativo-desenvolvimento-sistemas-152805091?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Categoria Gestão em TI
1.0
[Belo Horizonte, Minas Gerais, Brazil] Systems Development Corporate Coordinator na Coodesh - ## Descrição da vaga: Esta é uma vaga de um parceiro da plataforma Coodesh, ao candidatar-se você terá acesso as informações completas sobre a empresa e benefícios. Fique atento ao redirecionamento que vai te levar para uma url [https://coodesh.com](https://coodesh.com/vagas/coordenador-corporativo-desenvolvimento-sistemas-152805091?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) com o pop-up personalizado de candidatura. 👋 <p>A <strong>Telemont</strong> está em busca de <strong><ins>Systems Development Corporate Coordinator</ins></strong> para compor seu time!</p> <p>Há mais de 45 anos, a Telemont realiza serviços essenciais para o desenvolvimento do Brasil, se destacando nos segmentos de telecom, energia e TI. Atuamos de ponta a ponta para oferecer comunicação de voz, banda larga e dados, tecnologia da informação, transporte multimídia e gestão de sistemas de energia, conectando milhões de brasileiros. Contamos com um time de profissionais que traduzem em ações os nossos valores, garantindo a excelência operacional e a qualidade em seus atendimentos.&nbsp;</p> <p><strong>Responsabilidades:</strong></p> <ul> <li>Assegurar a evolução dos sistemas por meio da coordenação dos trabalhos de: refinamento das necessidades dos usuários, planejamento do desenvolvimento, testes, homologação e disponibilização em produção;</li> <li>Desenvolver novos projetos de sistemas de informação alinhados aos objetivos da empresa, por meio de prospecção das soluções, escolha de fornecedores, planejamento, alocação de equipe e acompanhamento da implementação;</li> <li>Contribuir na identificação de melhorias em processos da área ou do próprio sistema, através da análise crítica dos mesmos propondo alternativas e/ou propostas de solução;</li> <li>Assessorar, quando necessário, a escolha correta de fornecedores de desenvolvimento de sistemas, através da análise dos mesmos;</li> <li>Supervisionar o ambiente dos sistemas de informação da empresa por meio da atuação junto às áreas de negócio definindo e propondo melhorias sistêmicas ao processo e alinhando as necessidades com a equipe responsável ou com consultores terceirizados;</li> <li>Sustentar os sistemas da empresa e suas integrações, gerenciando a qualidade do suporte a usuários, fila, prazo e idade de chamados bem como diagnosticando problemas e realizando implantação da solução definitiva encontrada;</li> <li>Gerir a esteira de desenvolvimentos mantendo alta performance e produtividade, alinhando prioridades e mantendo os usuários-chave atualizados sobre a evolução das demandas;</li> <li>Executar outras atividades de mesma natureza, nível de complexidade e responsabilidade.</li> </ul> ## Telemont: <p>Há mais de 45 anos, a Telemont realiza serviços essenciais para o desenvolvimento do Brasil, se destacando nos segmentos de telecom, energia e TI. Atuamos de ponta a ponta para oferecer comunicação de voz, banda larga e dados, tecnologia da informação, transporte multimídia e gestão de sistemas de energia, conectando milhões de brasileiros. Contamos com um time de profissionais que traduzem em ações os nossos valores, garantindo a excelência operacional e a qualidade em seus atendimentos. Além disso, somos uma empresa que busca promover uma maior diversidade e inclusão, pois acreditamos que a pluralidade das nossas equipes traz mais conhecimento, respeito às múltiplas possibilidades de vivência, inovação e integração entre as áreas. Essa incansável procura por vencer desafios e conquistar novas oportunidades faz a Telemont ser uma empresa onde cada pessoa tem a possibilidade de crescer e se desenvolver profissionalmente, construindo assim, junto com a organização, uma trajetória de sucesso. Nossos colaboradores são orientados por valores que norteiam todas as nossas ações e garantem a sustentabilidade do negócio, bem como de quem faz parte dele.</p> </p> ## Habilidades: - Metodologias ágeis - Gestão de times de tecnologia - Gestão de contratos - Gestão e Negociação com Cliente ## Local: Belo Horizonte, Minas Gerais, Brazil ## Requisitos: - Superior completo em Ciência da Computação ou cursos relacionados; - Domínio em processo de desenvolvimento de software; - Noção de supervisão profissional de projetos (PMBOK); - Facilidade com gestão de pessoas; - Conhecimento em padronização de processos e procedimentos; - Conhecimento em Metodologia ITIL: Incidentes, Demandas, Problemas. ## Benefícios: - VR; - VT; - Seguro de vida; - Gympass; - Descontos estabelecimentos conveniados; - Plano de saúde; - Plano odontológico. ## Como se candidatar: Candidatar-se exclusivamente através da plataforma Coodesh no link a seguir: [Systems Development Corporate Coordinator na Telemont](https://coodesh.com/vagas/coordenador-corporativo-desenvolvimento-sistemas-152805091?utm_source=github&utm_medium=devssa-onde-codar-em-salvador&modal=open) Após candidatar-se via plataforma Coodesh e validar o seu login, você poderá acompanhar e receber todas as interações do processo por lá. Utilize a opção **Pedir Feedback** entre uma etapa e outra na vaga que se candidatou. Isso fará com que a pessoa **Recruiter** responsável pelo processo na empresa receba a notificação. ## Labels #### Alocação Alocado #### Categoria Gestão em TI
process
systems development corporate coordinator na coodesh descrição da vaga esta é uma vaga de um parceiro da plataforma coodesh ao candidatar se você terá acesso as informações completas sobre a empresa e benefícios fique atento ao redirecionamento que vai te levar para uma url com o pop up personalizado de candidatura 👋 a telemont está em busca de systems development corporate coordinator para compor seu time há mais de anos a telemont realiza serviços essenciais para o desenvolvimento do brasil se destacando nos segmentos de telecom energia e ti atuamos de ponta a ponta para oferecer comunicação de voz banda larga e dados tecnologia da informação transporte multimídia e gestão de sistemas de energia conectando milhões de brasileiros contamos com um time de profissionais que traduzem em ações os nossos valores garantindo a excelência operacional e a qualidade em seus atendimentos nbsp responsabilidades assegurar a evolução dos sistemas por meio da coordenação dos trabalhos de refinamento das necessidades dos usuários planejamento do desenvolvimento testes homologação e disponibilização em produção desenvolver novos projetos de sistemas de informação alinhados aos objetivos da empresa por meio de prospecção das soluções escolha de fornecedores planejamento alocação de equipe e acompanhamento da implementação contribuir na identificação de melhorias em processos da área ou do próprio sistema através da análise crítica dos mesmos propondo alternativas e ou propostas de solução assessorar quando necessário a escolha correta de fornecedores de desenvolvimento de sistemas através da análise dos mesmos supervisionar o ambiente dos sistemas de informação da empresa por meio da atuação junto às áreas de negócio definindo e propondo melhorias sistêmicas ao processo e alinhando as necessidades com a equipe responsável ou com consultores terceirizados sustentar os sistemas da empresa e suas integrações gerenciando a qualidade do suporte a usuários fila prazo e idade de chamados bem como diagnosticando problemas e realizando implantação da solução definitiva encontrada gerir a esteira de desenvolvimentos mantendo alta performance e produtividade alinhando prioridades e mantendo os usuários chave atualizados sobre a evolução das demandas executar outras atividades de mesma natureza nível de complexidade e responsabilidade telemont há mais de anos a telemont realiza serviços essenciais para o desenvolvimento do brasil se destacando nos segmentos de telecom energia e ti atuamos de ponta a ponta para oferecer comunicação de voz banda larga e dados tecnologia da informação transporte multimídia e gestão de sistemas de energia conectando milhões de brasileiros contamos com um time de profissionais que traduzem em ações os nossos valores garantindo a excelência operacional e a qualidade em seus atendimentos além disso somos uma empresa que busca promover uma maior diversidade e inclusão pois acreditamos que a pluralidade das nossas equipes traz mais conhecimento respeito às múltiplas possibilidades de vivência inovação e integração entre as áreas essa incansável procura por vencer desafios e conquistar novas oportunidades faz a telemont ser uma empresa onde cada pessoa tem a possibilidade de crescer e se desenvolver profissionalmente construindo assim junto com a organização uma trajetória de sucesso nossos colaboradores são orientados por valores que norteiam todas as nossas ações e garantem a sustentabilidade do negócio bem como de quem faz parte dele habilidades metodologias ágeis gestão de times de tecnologia gestão de contratos gestão e negociação com cliente local belo horizonte minas gerais brazil requisitos superior completo em ciência da computação ou cursos relacionados domínio em processo de desenvolvimento de software noção de supervisão profissional de projetos pmbok facilidade com gestão de pessoas conhecimento em padronização de processos e procedimentos conhecimento em metodologia itil incidentes demandas problemas benefícios vr vt seguro de vida gympass descontos estabelecimentos conveniados plano de saúde plano odontológico como se candidatar candidatar se exclusivamente através da plataforma coodesh no link a seguir após candidatar se via plataforma coodesh e validar o seu login você poderá acompanhar e receber todas as interações do processo por lá utilize a opção pedir feedback entre uma etapa e outra na vaga que se candidatou isso fará com que a pessoa recruiter responsável pelo processo na empresa receba a notificação labels alocação alocado categoria gestão em ti
1
224,284
7,468,675,884
IssuesEvent
2018-04-02 19:51:31
concrete5/concrete5
https://api.github.com/repos/concrete5/concrete5
closed
[Performance] Delete user attribute values on user delete
help wanted priority:like to have type:bug
## Issue conrete5 doesn't delete custom attribute entries from `at*` tables on user delete. It just deletes the entries from `UserAttributeValues` table. ## Impact - The database contains a large number of unused data. - Sometimes it could be a legal issue. Because, we shouldn't keep the data of any deleted user. ## Proposal Let's remove the values from `at*` tables on user delete. ## How to reproduce - Add a custom user attribute. e.g- fullname with text type. - Add an user. Make sure to insert the attribute value. - Get the user id and check the entries on `UserAttributeValues` table. - Delete the user. - Check `atDefault` table. None of the values of this table hasn't deleted.
1.0
[Performance] Delete user attribute values on user delete - ## Issue conrete5 doesn't delete custom attribute entries from `at*` tables on user delete. It just deletes the entries from `UserAttributeValues` table. ## Impact - The database contains a large number of unused data. - Sometimes it could be a legal issue. Because, we shouldn't keep the data of any deleted user. ## Proposal Let's remove the values from `at*` tables on user delete. ## How to reproduce - Add a custom user attribute. e.g- fullname with text type. - Add an user. Make sure to insert the attribute value. - Get the user id and check the entries on `UserAttributeValues` table. - Delete the user. - Check `atDefault` table. None of the values of this table hasn't deleted.
non_process
delete user attribute values on user delete issue doesn t delete custom attribute entries from at tables on user delete it just deletes the entries from userattributevalues table impact the database contains a large number of unused data sometimes it could be a legal issue because we shouldn t keep the data of any deleted user proposal let s remove the values from at tables on user delete how to reproduce add a custom user attribute e g fullname with text type add an user make sure to insert the attribute value get the user id and check the entries on userattributevalues table delete the user check atdefault table none of the values of this table hasn t deleted
0
6,239
9,195,858,884
IssuesEvent
2019-03-07 04:25:21
dev-register-medical-examination/Register-medical-examination-Project2-
https://api.github.com/repos/dev-register-medical-examination/Register-medical-examination-Project2-
reopened
mockup-user
In Process
Xem và tìm kiếm thông tin bệnh Chọn khoa khám Chọn giờ đến khám Chọn Bác sĩ khám để phân loại mức giá Thông tin đăng ký(hóa đơn,thời gian khám ,bác sĩ ,trạng thái , thời gian hết hạn đến khám)
1.0
mockup-user - Xem và tìm kiếm thông tin bệnh Chọn khoa khám Chọn giờ đến khám Chọn Bác sĩ khám để phân loại mức giá Thông tin đăng ký(hóa đơn,thời gian khám ,bác sĩ ,trạng thái , thời gian hết hạn đến khám)
process
mockup user xem và tìm kiếm thông tin bệnh chọn khoa khám chọn giờ đến khám chọn bác sĩ khám để phân loại mức giá thông tin đăng ký hóa đơn thời gian khám bác sĩ trạng thái thời gian hết hạn đến khám
1