Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
146,402
11,735,164,658
IssuesEvent
2020-03-11 10:37:57
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: kv0bench/nodes=20/cpu=8/sequential failed
C-test-failure O-roachtest O-robot branch-release-19.2 release-blocker
[(roachtest).kv0bench/nodes=20/cpu=8/sequential failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1721939&tab=buildLog) on [release-19.2@6fab3dc4cb79e95bd8b301eb399220e5331dbf7d](https://github.com/cockroachdb/cockroach/commits/6fab3dc4cb79e95bd8b301eb399220e5331dbf7d): ``` The test failed on branch=release-19.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20200201-1721939/kv0bench/nodes=20/cpu=8/sequential/run_1 cluster.go:1927,kvbench.go:220,search.go:43,search.go:173,kvbench.go:330,kvbench.go:87,test_runner.go:734: error with attached stack trace: main.execCmd /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:406 main.(*cluster).WipeE /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:1917 main.(*cluster).Wipe /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:1926 main.runKVBench.func1 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kvbench.go:220 github.com/cockroachdb/cockroach/pkg/util/search.searchWithSearcher /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/util/search/search.go:43 github.com/cockroachdb/cockroach/pkg/util/search.(*lineSearcher).Search /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/util/search/search.go:173 main.runKVBench /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kvbench.go:330 main.registerKVBenchSpec.func1 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kvbench.go:87 main.(*testRunner).runTest.func2 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:734 runtime.goexit /usr/local/go/src/runtime/asm_amd64.s:1357 - error with embedded safe details: %s returned: stderr: %s stdout: %s -- arg 1: <string> -- arg 2: <string> -- arg 3: <string> - /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod wipe teamcity-1721939-1580542446-68-n21cpu8:1-20 returned: stderr: stdout: teamcity-1721939-1580542446-68-n21cpu8: stopping and waiting teamcity-1721939-1580542446-68-n21cpu8: wiping 4: exit status 255: I200201 12:16:03.112635 1 cluster_synced.go:1635 command failed: - exit status 1 ``` <details><summary>More</summary><p> Artifacts: [/kv0bench/nodes=20/cpu=8/sequential](https://teamcity.cockroachdb.com/viewLog.html?buildId=1721939&tab=artifacts#/kv0bench/nodes=20/cpu=8/sequential) Related: - #44330 roachtest: kv0bench/nodes=10/cpu=8/shards=20/sequential failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) - #44109 roachtest: kv0bench/nodes=10/cpu=8/shards=20/sequential failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) - #43551 roachtest: kv0bench/nodes=20/cpu=8/shards=80/sequential failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) - #43519 roachtest: kv0bench/nodes=20/cpu=8/sequential failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Akv0bench%2Fnodes%3D20%2Fcpu%3D8%2Fsequential.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: kv0bench/nodes=20/cpu=8/sequential failed - [(roachtest).kv0bench/nodes=20/cpu=8/sequential failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=1721939&tab=buildLog) on [release-19.2@6fab3dc4cb79e95bd8b301eb399220e5331dbf7d](https://github.com/cockroachdb/cockroach/commits/6fab3dc4cb79e95bd8b301eb399220e5331dbf7d): ``` The test failed on branch=release-19.2, cloud=gce: test artifacts and logs in: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/artifacts/20200201-1721939/kv0bench/nodes=20/cpu=8/sequential/run_1 cluster.go:1927,kvbench.go:220,search.go:43,search.go:173,kvbench.go:330,kvbench.go:87,test_runner.go:734: error with attached stack trace: main.execCmd /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:406 main.(*cluster).WipeE /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:1917 main.(*cluster).Wipe /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:1926 main.runKVBench.func1 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kvbench.go:220 github.com/cockroachdb/cockroach/pkg/util/search.searchWithSearcher /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/util/search/search.go:43 github.com/cockroachdb/cockroach/pkg/util/search.(*lineSearcher).Search /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/util/search/search.go:173 main.runKVBench /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kvbench.go:330 main.registerKVBenchSpec.func1 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/kvbench.go:87 main.(*testRunner).runTest.func2 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:734 runtime.goexit /usr/local/go/src/runtime/asm_amd64.s:1357 - error with embedded safe details: %s returned: stderr: %s stdout: %s -- arg 1: <string> -- arg 2: <string> -- arg 3: <string> - /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod wipe teamcity-1721939-1580542446-68-n21cpu8:1-20 returned: stderr: stdout: teamcity-1721939-1580542446-68-n21cpu8: stopping and waiting teamcity-1721939-1580542446-68-n21cpu8: wiping 4: exit status 255: I200201 12:16:03.112635 1 cluster_synced.go:1635 command failed: - exit status 1 ``` <details><summary>More</summary><p> Artifacts: [/kv0bench/nodes=20/cpu=8/sequential](https://teamcity.cockroachdb.com/viewLog.html?buildId=1721939&tab=artifacts#/kv0bench/nodes=20/cpu=8/sequential) Related: - #44330 roachtest: kv0bench/nodes=10/cpu=8/shards=20/sequential failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-release-19.1](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-release-19.1) - #44109 roachtest: kv0bench/nodes=10/cpu=8/shards=20/sequential failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) - #43551 roachtest: kv0bench/nodes=20/cpu=8/shards=80/sequential failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) - #43519 roachtest: kv0bench/nodes=20/cpu=8/sequential failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Akv0bench%2Fnodes%3D20%2Fcpu%3D8%2Fsequential.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest nodes cpu sequential failed on the test failed on branch release cloud gce test artifacts and logs in home agent work go src github com cockroachdb cockroach artifacts nodes cpu sequential run cluster go kvbench go search go search go kvbench go kvbench go test runner go error with attached stack trace main execcmd home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main cluster wipee home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main cluster wipe home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main runkvbench home agent work go src github com cockroachdb cockroach pkg cmd roachtest kvbench go github com cockroachdb cockroach pkg util search searchwithsearcher home agent work go src github com cockroachdb cockroach pkg util search search go github com cockroachdb cockroach pkg util search linesearcher search home agent work go src github com cockroachdb cockroach pkg util search search go main runkvbench home agent work go src github com cockroachdb cockroach pkg cmd roachtest kvbench go main registerkvbenchspec home agent work go src github com cockroachdb cockroach pkg cmd roachtest kvbench go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go runtime goexit usr local go src runtime asm s error with embedded safe details s returned stderr s stdout s arg arg arg home agent work go src github com cockroachdb cockroach bin roachprod wipe teamcity returned stderr stdout teamcity stopping and waiting teamcity wiping exit status cluster synced go command failed exit status more artifacts related roachtest nodes cpu shards sequential failed roachtest nodes cpu shards sequential failed roachtest nodes cpu shards sequential failed roachtest nodes cpu sequential failed powered by
1
24,443
4,082,165,144
IssuesEvent
2016-05-31 11:50:47
mantidproject/mantid
https://api.github.com/repos/mantidproject/mantid
closed
Reorganize unit tests of functions, category "General", chunk 4
Component: Fitting Component: Framework Misc: Maintenance Quality: Unit Tests
This is one chunk of #16267. Move the test cases that use the algorithm Fit to the unit test file of Fit (FitTest). Add tests to test the function itself when necessary. Tidy up the includes wherever possible, as for example in #16470. Functions included in this chunk: - ThermalNeutronBk2BkExpSigma, - ThermalNeutronDtoTOFFunction, - UserFunction, - UserFunctionMD, - VesuvioResolution, - Voigt
1.0
Reorganize unit tests of functions, category "General", chunk 4 - This is one chunk of #16267. Move the test cases that use the algorithm Fit to the unit test file of Fit (FitTest). Add tests to test the function itself when necessary. Tidy up the includes wherever possible, as for example in #16470. Functions included in this chunk: - ThermalNeutronBk2BkExpSigma, - ThermalNeutronDtoTOFFunction, - UserFunction, - UserFunctionMD, - VesuvioResolution, - Voigt
test
reorganize unit tests of functions category general chunk this is one chunk of move the test cases that use the algorithm fit to the unit test file of fit fittest add tests to test the function itself when necessary tidy up the includes wherever possible as for example in functions included in this chunk thermalneutrondtotoffunction userfunction userfunctionmd vesuvioresolution voigt
1
43,394
7,044,452,926
IssuesEvent
2018-01-01 01:17:18
jessesquires/JSQDataSourcesKit
https://api.github.com/repos/jessesquires/JSQDataSourcesKit
closed
Update changelog for TableEditingController (#67)
documentation
Update changelog for TableEditingController (#67) for the 6.1.0 release
1.0
Update changelog for TableEditingController (#67) - Update changelog for TableEditingController (#67) for the 6.1.0 release
non_test
update changelog for tableeditingcontroller update changelog for tableeditingcontroller for the release
0
140,240
18,900,637,928
IssuesEvent
2021-11-16 00:14:08
pustovitDmytro/code-chronicle
https://api.github.com/repos/pustovitDmytro/code-chronicle
opened
CVE-2020-11023 (Medium) detected in jquery-1.8.1.min.js
security vulnerability
## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: code-chronicle/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: /node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/pustovitDmytro/code-chronicle/commit/6e512d3b26c0461e41894cae6558b35a24e5b20e">6e512d3b26c0461e41894cae6558b35a24e5b20e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-11023 (Medium) detected in jquery-1.8.1.min.js - ## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: code-chronicle/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: /node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/pustovitDmytro/code-chronicle/commit/6e512d3b26c0461e41894cae6558b35a24e5b20e">6e512d3b26c0461e41894cae6558b35a24e5b20e</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440">https://github.com/jquery/jquery/security/advisories/GHSA-jpcq-cgw6-v4j6,https://github.com/rails/jquery-rails/blob/master/CHANGELOG.md#440</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0;jquery-rails - 4.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in jquery min js cve medium severity vulnerability vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file code chronicle node modules redeyed examples browser index html path to vulnerable library node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library found in head commit a href found in base branch master vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery jquery rails step up your open source security game with whitesource
0
218,551
16,762,853,207
IssuesEvent
2021-06-14 03:23:42
bee-queue/bee-queue
https://api.github.com/repos/bee-queue/bee-queue
reopened
In no way job is ever retried
needs-documentation
Code to reproduce ```javascript const Queue = require("bee-queue"); const queue = new Queue("http-sender"); const job = queue.createJob({ timestamp: Date.now(), }); job.backoff("exponential", 1000); job.retries(3); job.save(); job.on("succeeded", (result) => { console.log(`${job.id}: sent`); }); queue.process((job, done) => { console.log(`Processing job ${job.id}`); return done(new Error("test")); }); queue.on("failed", (job, err) => { console.log(`Job ${job.id} failed with error ${err.message}`); }); queue.on("retrying", (job, err) => { console.log( `Job ${job.id} failed with error ${err.message} but is being retried!` ); }); ```
1.0
In no way job is ever retried - Code to reproduce ```javascript const Queue = require("bee-queue"); const queue = new Queue("http-sender"); const job = queue.createJob({ timestamp: Date.now(), }); job.backoff("exponential", 1000); job.retries(3); job.save(); job.on("succeeded", (result) => { console.log(`${job.id}: sent`); }); queue.process((job, done) => { console.log(`Processing job ${job.id}`); return done(new Error("test")); }); queue.on("failed", (job, err) => { console.log(`Job ${job.id} failed with error ${err.message}`); }); queue.on("retrying", (job, err) => { console.log( `Job ${job.id} failed with error ${err.message} but is being retried!` ); }); ```
non_test
in no way job is ever retried code to reproduce javascript const queue require bee queue const queue new queue http sender const job queue createjob timestamp date now job backoff exponential job retries job save job on succeeded result console log job id sent queue process job done console log processing job job id return done new error test queue on failed job err console log job job id failed with error err message queue on retrying job err console log job job id failed with error err message but is being retried
0
131,355
18,244,682,905
IssuesEvent
2021-10-01 16:46:23
ibm-skills-network/eslint-config-apset
https://api.github.com/repos/ibm-skills-network/eslint-config-apset
opened
CVE-2021-23337 (High) detected in lodash-4.17.15.tgz
security vulnerability
## CVE-2021-23337 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: eslint-config-apset/package.json</p> <p>Path to vulnerable library: eslint-config-apset/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - eslint-plugin-flowtype-3.13.0.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ibm-skills-network/eslint-config-apset/commit/f87e56dd9ce1995d4e58eef9e03128254285c8d3">f87e56dd9ce1995d4e58eef9e03128254285c8d3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-23337 (High) detected in lodash-4.17.15.tgz - ## CVE-2021-23337 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>lodash-4.17.15.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.15.tgz</a></p> <p>Path to dependency file: eslint-config-apset/package.json</p> <p>Path to vulnerable library: eslint-config-apset/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - eslint-plugin-flowtype-3.13.0.tgz (Root Library) - :x: **lodash-4.17.15.tgz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/ibm-skills-network/eslint-config-apset/commit/f87e56dd9ce1995d4e58eef9e03128254285c8d3">f87e56dd9ce1995d4e58eef9e03128254285c8d3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Lodash versions prior to 4.17.21 are vulnerable to Command Injection via the template function. <p>Publish Date: 2021-02-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-23337>CVE-2021-23337</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.2</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c">https://github.com/lodash/lodash/commit/3469357cff396a26c363f8c1b5a91dde28ba4b1c</a></p> <p>Release Date: 2021-02-15</p> <p>Fix Resolution: lodash - 4.17.21</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in lodash tgz cve high severity vulnerability vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file eslint config apset package json path to vulnerable library eslint config apset node modules lodash package json dependency hierarchy eslint plugin flowtype tgz root library x lodash tgz vulnerable library found in head commit a href found in base branch master vulnerability details lodash versions prior to are vulnerable to command injection via the template function publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash step up your open source security game with whitesource
0
210,379
23,754,648,287
IssuesEvent
2022-09-01 01:04:37
venkateshreddypala/enrollee-service
https://api.github.com/repos/venkateshreddypala/enrollee-service
opened
CVE-2022-25857 (High) detected in snakeyaml-1.26.jar
security vulnerability
## CVE-2022-25857 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.26.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.26/snakeyaml-1.26.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-mongodb-2.4.0-SNAPSHOT.jar (Root Library) - spring-boot-starter-2.4.0-SNAPSHOT.jar - :x: **snakeyaml-1.26.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package org.yaml:snakeyaml from 0 and before 1.31 are vulnerable to Denial of Service (DoS) due missing to nested depth limitation for collections. <p>Publish Date: 2022-08-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25857>CVE-2022-25857</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25857">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25857</a></p> <p>Release Date: 2022-08-30</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-25857 (High) detected in snakeyaml-1.26.jar - ## CVE-2022-25857 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.26.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.26/snakeyaml-1.26.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-data-mongodb-2.4.0-SNAPSHOT.jar (Root Library) - spring-boot-starter-2.4.0-SNAPSHOT.jar - :x: **snakeyaml-1.26.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package org.yaml:snakeyaml from 0 and before 1.31 are vulnerable to Denial of Service (DoS) due missing to nested depth limitation for collections. <p>Publish Date: 2022-08-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25857>CVE-2022-25857</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25857">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25857</a></p> <p>Release Date: 2022-08-30</p> <p>Fix Resolution: org.yaml:snakeyaml:1.31</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in snakeyaml jar cve high severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter data mongodb snapshot jar root library spring boot starter snapshot jar x snakeyaml jar vulnerable library found in base branch master vulnerability details the package org yaml snakeyaml from and before are vulnerable to denial of service dos due missing to nested depth limitation for collections publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml step up your open source security game with mend
0
250,400
21,263,544,524
IssuesEvent
2022-04-13 07:44:10
hzi-braunschweig/SORMAS-Project
https://api.github.com/repos/hzi-braunschweig/SORMAS-Project
closed
Add loggers to reflect user steps at BDD level into automation framework
testing task e2e-tests
Currently the framework has logging only at the core level: actions performed on webelements. For a better debuging, please update the BDD methods and add steps to reflect the actions of the user in order to determine failing steps and debug more easy further issues. Please don't implement other logging libraries, use the existing one, and add fix to expose the logging file into allure report.
2.0
Add loggers to reflect user steps at BDD level into automation framework - Currently the framework has logging only at the core level: actions performed on webelements. For a better debuging, please update the BDD methods and add steps to reflect the actions of the user in order to determine failing steps and debug more easy further issues. Please don't implement other logging libraries, use the existing one, and add fix to expose the logging file into allure report.
test
add loggers to reflect user steps at bdd level into automation framework currently the framework has logging only at the core level actions performed on webelements for a better debuging please update the bdd methods and add steps to reflect the actions of the user in order to determine failing steps and debug more easy further issues please don t implement other logging libraries use the existing one and add fix to expose the logging file into allure report
1
630,188
20,100,184,328
IssuesEvent
2022-02-07 02:24:43
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
m.youtube.com - Scrolling not responding when watching videos in landscape mode
browser-firefox-mobile priority-normal severity-important engine-gecko
<!-- @browser: Firefox Mobile 98.0 --> <!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:98.0) Gecko/98.0 Firefox/98.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/99010 --> <!-- @extra_labels: browser-fenix --> **URL**: https://m.youtube.com/watch?v=UnFOv0tdEmk&t=2s **Browser / Version**: Firefox Mobile 98.0 **Operating System**: Android 12 **Tested Another Browser**: Yes Other **Problem type**: Something else **Description**: video is cut off even before going full screen, scrolling does not work, and I cant even reach full screen button, this happens on several but not all videos on youtube **Steps to Reproduce**: Ok so I am watching videos on youtube in landscape mode, most work fine, but several end up being too zoomed in, cutting the edges of the video and that is even before I hit fullscreen, I can't even scroll the video page and have to use android back button to escape, I am not sure why it only happens on some videos. Expected wantrd behaviour is for whole video to be viewable, and scrolling to work, in any orientation, and for full screen to not cut a single pixel off any edge of all videos, I suspect youtube is messing with firefox again intentionally on android, to make people use their app. Firefox on desktop works fine on all videos. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/2/8dc7f486-2e6e-4914-beba-676da9c6dd0a.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220130093554</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/2/a039d424-dc30-4a92-8b58-f67bf3cf1c7d) _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
m.youtube.com - Scrolling not responding when watching videos in landscape mode - <!-- @browser: Firefox Mobile 98.0 --> <!-- @ua_header: Mozilla/5.0 (Android 12; Mobile; rv:98.0) Gecko/98.0 Firefox/98.0 --> <!-- @reported_with: android-components-reporter --> <!-- @public_url: https://github.com/webcompat/web-bugs/issues/99010 --> <!-- @extra_labels: browser-fenix --> **URL**: https://m.youtube.com/watch?v=UnFOv0tdEmk&t=2s **Browser / Version**: Firefox Mobile 98.0 **Operating System**: Android 12 **Tested Another Browser**: Yes Other **Problem type**: Something else **Description**: video is cut off even before going full screen, scrolling does not work, and I cant even reach full screen button, this happens on several but not all videos on youtube **Steps to Reproduce**: Ok so I am watching videos on youtube in landscape mode, most work fine, but several end up being too zoomed in, cutting the edges of the video and that is even before I hit fullscreen, I can't even scroll the video page and have to use android back button to escape, I am not sure why it only happens on some videos. Expected wantrd behaviour is for whole video to be viewable, and scrolling to work, in any orientation, and for full screen to not cut a single pixel off any edge of all videos, I suspect youtube is messing with firefox again intentionally on android, to make people use their app. Firefox on desktop works fine on all videos. <details> <summary>View the screenshot</summary> <img alt="Screenshot" src="https://webcompat.com/uploads/2022/2/8dc7f486-2e6e-4914-beba-676da9c6dd0a.jpeg"> </details> <details> <summary>Browser Configuration</summary> <ul> <li>gfx.webrender.all: false</li><li>gfx.webrender.blob-images: true</li><li>gfx.webrender.enabled: false</li><li>image.mem.shared: true</li><li>buildID: 20220130093554</li><li>channel: nightly</li><li>hasTouchScreen: true</li><li>mixed active content blocked: false</li><li>mixed passive content blocked: false</li><li>tracking content blocked: false</li> </ul> </details> [View console log messages](https://webcompat.com/console_logs/2022/2/a039d424-dc30-4a92-8b58-f67bf3cf1c7d) _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
m youtube com scrolling not responding when watching videos in landscape mode url browser version firefox mobile operating system android tested another browser yes other problem type something else description video is cut off even before going full screen scrolling does not work and i cant even reach full screen button this happens on several but not all videos on youtube steps to reproduce ok so i am watching videos on youtube in landscape mode most work fine but several end up being too zoomed in cutting the edges of the video and that is even before i hit fullscreen i can t even scroll the video page and have to use android back button to escape i am not sure why it only happens on some videos expected wantrd behaviour is for whole video to be viewable and scrolling to work in any orientation and for full screen to not cut a single pixel off any edge of all videos i suspect youtube is messing with firefox again intentionally on android to make people use their app firefox on desktop works fine on all videos view the screenshot img alt screenshot src browser configuration gfx webrender all false gfx webrender blob images true gfx webrender enabled false image mem shared true buildid channel nightly hastouchscreen true mixed active content blocked false mixed passive content blocked false tracking content blocked false from with ❤️
0
417,147
12,156,140,335
IssuesEvent
2020-04-25 16:04:56
code4romania/stam-acasa
https://api.github.com/repos/code4romania/stam-acasa
closed
Alte persoane în grijă details should only show for account holder
enhancement front-end medium-priority
The `Alte persoane în grijă:` info if now available in the main account holder profile details, but also in the `alte persoane` profile details. I think it would be best to hide them for `alte persoane`.
1.0
Alte persoane în grijă details should only show for account holder - The `Alte persoane în grijă:` info if now available in the main account holder profile details, but also in the `alte persoane` profile details. I think it would be best to hide them for `alte persoane`.
non_test
alte persoane în grijă details should only show for account holder the alte persoane în grijă info if now available in the main account holder profile details but also in the alte persoane profile details i think it would be best to hide them for alte persoane
0
173,829
13,446,433,516
IssuesEvent
2020-09-08 12:57:36
citusdata/citus
https://api.github.com/repos/citusdata/citus
opened
Remove almost duplicate tests
regression tests
We have some tests that are almost copy-pasted for testing mx structure as well. With mx we have some more commands as we connect to the workers but most of the queries are the same. It is hard to maintain two test files. With some conditions (such as is_mx) in the test file, we can have a single file, of course with an alternative output. However this would have the advantage that whenever we update the single file, we would also see the effect on the mx side (as the alternative output will change). With the current structure it is possible to forget to update one of the test files. See: https://github.com/citusdata/citus/pull/4133#discussion_r484422432
1.0
Remove almost duplicate tests - We have some tests that are almost copy-pasted for testing mx structure as well. With mx we have some more commands as we connect to the workers but most of the queries are the same. It is hard to maintain two test files. With some conditions (such as is_mx) in the test file, we can have a single file, of course with an alternative output. However this would have the advantage that whenever we update the single file, we would also see the effect on the mx side (as the alternative output will change). With the current structure it is possible to forget to update one of the test files. See: https://github.com/citusdata/citus/pull/4133#discussion_r484422432
test
remove almost duplicate tests we have some tests that are almost copy pasted for testing mx structure as well with mx we have some more commands as we connect to the workers but most of the queries are the same it is hard to maintain two test files with some conditions such as is mx in the test file we can have a single file of course with an alternative output however this would have the advantage that whenever we update the single file we would also see the effect on the mx side as the alternative output will change with the current structure it is possible to forget to update one of the test files see
1
14,312
8,554,640,033
IssuesEvent
2018-11-08 07:20:20
ckfinder/ckfinder
https://api.github.com/repos/ckfinder/ckfinder
closed
Very slow to load when folder contains many directories
Performance UI Tweak
CkFinder 3 hangs for a long time (1-2 minutes) when the image folder contains thousands of directories. CkFinder was much quicker.
True
Very slow to load when folder contains many directories - CkFinder 3 hangs for a long time (1-2 minutes) when the image folder contains thousands of directories. CkFinder was much quicker.
non_test
very slow to load when folder contains many directories ckfinder hangs for a long time minutes when the image folder contains thousands of directories ckfinder was much quicker
0
249,576
21,178,273,432
IssuesEvent
2022-04-08 04:09:50
tinhthanhvo/api-symfony-unlock
https://api.github.com/repos/tinhthanhvo/api-symfony-unlock
closed
[Report] Export Order info with CSV extension
Testing
This feature support Admin export Order data ### Flow - Handle data input: Filter condition - Return output data ### Input (Payload & token) 1. Payload - fileName (optional): Specific CSV file name - status (optional): Order status - fromDate (optional, format: Y-m-d): Start date (to get data) - toDate (optional, format: Y-m-d): End date (to get data) 2. Token ### Output - success: String message with CSV file (format like image below) ![image](https://user-images.githubusercontent.com/99877956/162242077-afc7bb85-df55-499a-a74e-aa1e018f2173.png) - error: String message ### Validation 1. fileName - Only contain alphabet letter and length cannot be more than 30 characters 2. status - Value only [1, 2, 3, 4] correspond with [Pending, Approved, Canceled, Completed] 3. fromDate - Format Y-m-d - Date value cannot be greater than today 4. toDate - Format Y-m-d - Date value cannot be smaller than fromDate
1.0
[Report] Export Order info with CSV extension - This feature support Admin export Order data ### Flow - Handle data input: Filter condition - Return output data ### Input (Payload & token) 1. Payload - fileName (optional): Specific CSV file name - status (optional): Order status - fromDate (optional, format: Y-m-d): Start date (to get data) - toDate (optional, format: Y-m-d): End date (to get data) 2. Token ### Output - success: String message with CSV file (format like image below) ![image](https://user-images.githubusercontent.com/99877956/162242077-afc7bb85-df55-499a-a74e-aa1e018f2173.png) - error: String message ### Validation 1. fileName - Only contain alphabet letter and length cannot be more than 30 characters 2. status - Value only [1, 2, 3, 4] correspond with [Pending, Approved, Canceled, Completed] 3. fromDate - Format Y-m-d - Date value cannot be greater than today 4. toDate - Format Y-m-d - Date value cannot be smaller than fromDate
test
export order info with csv extension this feature support admin export order data flow handle data input filter condition return output data input payload token payload filename optional specific csv file name status optional order status fromdate optional format y m d start date to get data todate optional format y m d end date to get data token output success string message with csv file format like image below error string message validation filename only contain alphabet letter and length cannot be more than characters status value only correspond with fromdate format y m d date value cannot be greater than today todate format y m d date value cannot be smaller than fromdate
1
163,739
25,866,679,176
IssuesEvent
2022-12-13 21:34:18
EscolaDeSaudePublica/DesignLab
https://api.github.com/repos/EscolaDeSaudePublica/DesignLab
closed
Atualização de Conteúdo | site Felicilab | Métodos
Site PROJ: Felicilab Prioridade Design: Alta
## **Objetivo** **Como** designer **Quero** atualizar o novo site do Felicilab com conteúdos desenvolvidos pelas narrativas **Para** lançá-lo até o fim do defeso eleitoral ## **Contexto** O novo site do Felicilab está em sua fase final de desenvolvimento. Com isso, novas seções foram pensadas e o conteúdo de cada uma delas está em desenvolvimento pelo time de narrativas. Assim que cada conteúdo for entregue, a seção do site precisa ser atualizada e disponibilizada. ## **Escopo** Métodos ( @ericabpinho ) Metodologias ágeis (ver no guia do colaborador, mas resumir consideravelmente). Breve história da board integrada. SCRUM/papeis. Cerimônias. Modelo híbrido/assíncrono - nossa construção (até o momento). Ver o texto/modelo do Rani.
1.0
Atualização de Conteúdo | site Felicilab | Métodos - ## **Objetivo** **Como** designer **Quero** atualizar o novo site do Felicilab com conteúdos desenvolvidos pelas narrativas **Para** lançá-lo até o fim do defeso eleitoral ## **Contexto** O novo site do Felicilab está em sua fase final de desenvolvimento. Com isso, novas seções foram pensadas e o conteúdo de cada uma delas está em desenvolvimento pelo time de narrativas. Assim que cada conteúdo for entregue, a seção do site precisa ser atualizada e disponibilizada. ## **Escopo** Métodos ( @ericabpinho ) Metodologias ágeis (ver no guia do colaborador, mas resumir consideravelmente). Breve história da board integrada. SCRUM/papeis. Cerimônias. Modelo híbrido/assíncrono - nossa construção (até o momento). Ver o texto/modelo do Rani.
non_test
atualização de conteúdo site felicilab métodos objetivo como designer quero atualizar o novo site do felicilab com conteúdos desenvolvidos pelas narrativas para lançá lo até o fim do defeso eleitoral contexto o novo site do felicilab está em sua fase final de desenvolvimento com isso novas seções foram pensadas e o conteúdo de cada uma delas está em desenvolvimento pelo time de narrativas assim que cada conteúdo for entregue a seção do site precisa ser atualizada e disponibilizada escopo métodos ericabpinho metodologias ágeis ver no guia do colaborador mas resumir consideravelmente breve história da board integrada scrum papeis cerimônias modelo híbrido assíncrono nossa construção até o momento ver o texto modelo do rani
0
341,124
30,567,461,865
IssuesEvent
2023-07-20 18:57:41
darbaidzeavto/ci_final_exam_test
https://api.github.com/repos/darbaidzeavto/ci_final_exam_test
opened
fe35c89 failed unit tests.
ci-pytest
Automatically generated message fe35c899c3e3cfff4bcf57cf67cc41921e591ed9 failed unit tests. first bad commit for pytest was cdd9ed4c690b17c376a77c386ba5401d46ba3482 Pytest report: https://darbaidzeavto.github.io/ci_final_report/fe35c899c3e3cfff4bcf57cf67cc41921e591ed9-1689879459/pytest.html Black report: https://darbaidzeavto.github.io/ci_final_report/fe35c899c3e3cfff4bcf57cf67cc41921e591ed9-1689879459/black.html
1.0
fe35c89 failed unit tests. - Automatically generated message fe35c899c3e3cfff4bcf57cf67cc41921e591ed9 failed unit tests. first bad commit for pytest was cdd9ed4c690b17c376a77c386ba5401d46ba3482 Pytest report: https://darbaidzeavto.github.io/ci_final_report/fe35c899c3e3cfff4bcf57cf67cc41921e591ed9-1689879459/pytest.html Black report: https://darbaidzeavto.github.io/ci_final_report/fe35c899c3e3cfff4bcf57cf67cc41921e591ed9-1689879459/black.html
test
failed unit tests automatically generated message failed unit tests first bad commit for pytest was pytest report black report
1
168,858
14,174,326,232
IssuesEvent
2020-11-12 19:43:41
OpenMined/OM-Welcome-Package
https://api.github.com/repos/OpenMined/OM-Welcome-Package
opened
Can't Access PySyft Tutorials
Type: Documentation :books:
## Description When I click on the PySyft Tutorials link, It gives me a 404 Page Not Found Error ## Screenshots ![openmined](https://user-images.githubusercontent.com/47494475/98985968-4a57ae80-254a-11eb-94b7-0fff76925944.jpg)
1.0
Can't Access PySyft Tutorials - ## Description When I click on the PySyft Tutorials link, It gives me a 404 Page Not Found Error ## Screenshots ![openmined](https://user-images.githubusercontent.com/47494475/98985968-4a57ae80-254a-11eb-94b7-0fff76925944.jpg)
non_test
can t access pysyft tutorials description when i click on the pysyft tutorials link it gives me a page not found error screenshots
0
2,320
2,525,203,644
IssuesEvent
2015-01-20 22:53:06
AtlasOfLivingAustralia/biocache-service
https://api.github.com/repos/AtlasOfLivingAustralia/biocache-service
opened
Rationalise i18n mappings
enhancement priority-medium status-accepted type-enhancement
_From @djtfmartin on August 19, 2014 13:20_ *migrated from:* https://code.google.com/p/ala/issues/detail?id=697 *date:* Thu Jun 12 18:33:27 2014 *author:* moyesyside --- 1. Take out duplicate from messages.properties & download.properties (with download.properties being the primary source). 2. Take out all references to EL and CL layers from both. 3. Support darwin core fields in downloads with a dwcHeaders=true request params - leaving the default as is. 4. Add the description into the /search/grouped/facets service. 5. Move i18n properties in message.properties into download.properties (e.g. [http://biocache.ala.org.au/ws/index/fields](http://biocache.ala.org.au/ws/index/fields)) 6. If dwcHeaders=true, then use cl### field for headers for sampled fields. 7. Provide dwc.***** mapping in download.properties for index fields and include a dwcTerm field in /index/fields and /search/grouped/facets 8. Add sorts values (index or count). 9. Rename download.properties to fields.properties 10. Consider merging message.properties and download.properties and adding doco. _Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#83_
1.0
Rationalise i18n mappings - _From @djtfmartin on August 19, 2014 13:20_ *migrated from:* https://code.google.com/p/ala/issues/detail?id=697 *date:* Thu Jun 12 18:33:27 2014 *author:* moyesyside --- 1. Take out duplicate from messages.properties & download.properties (with download.properties being the primary source). 2. Take out all references to EL and CL layers from both. 3. Support darwin core fields in downloads with a dwcHeaders=true request params - leaving the default as is. 4. Add the description into the /search/grouped/facets service. 5. Move i18n properties in message.properties into download.properties (e.g. [http://biocache.ala.org.au/ws/index/fields](http://biocache.ala.org.au/ws/index/fields)) 6. If dwcHeaders=true, then use cl### field for headers for sampled fields. 7. Provide dwc.***** mapping in download.properties for index fields and include a dwcTerm field in /index/fields and /search/grouped/facets 8. Add sorts values (index or count). 9. Rename download.properties to fields.properties 10. Consider merging message.properties and download.properties and adding doco. _Copied from original issue: AtlasOfLivingAustralia/biocache-hubs#83_
non_test
rationalise mappings from djtfmartin on august migrated from date thu jun author moyesyside take out duplicate from messages properties download properties with download properties being the primary source take out all references to el and cl layers from both support darwin core fields in downloads with a dwcheaders true request params leaving the default as is add the description into the search grouped facets service move properties in message properties into download properties e g if dwcheaders true then use cl field for headers for sampled fields provide dwc mapping in download properties for index fields and include a dwcterm field in index fields and search grouped facets add sorts values index or count rename download properties to fields properties consider merging message properties and download properties and adding doco copied from original issue atlasoflivingaustralia biocache hubs
0
250,769
21,335,179,204
IssuesEvent
2022-04-18 13:47:14
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
opened
[Flaky Test] Inserts the filtered hello world block even when filter added after block registration
[Type] Flaky Test
<!-- __META_DATA__:{"failedTimes":1,"totalCommits":1,"baseCommit":"33a5d514892c254760b5fceb17e72a9faefa1218"} --> **Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.** ## Test title Inserts the filtered hello world block even when filter added after block registration ## Test path `/home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/plugins/block-api.spec.js` ## Flaky rate (_estimated_) `1 / 2` runs ## Errors
1.0
[Flaky Test] Inserts the filtered hello world block even when filter added after block registration - <!-- __META_DATA__:{"failedTimes":1,"totalCommits":1,"baseCommit":"33a5d514892c254760b5fceb17e72a9faefa1218"} --> **Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.** ## Test title Inserts the filtered hello world block even when filter added after block registration ## Test path `/home/runner/work/gutenberg/gutenberg/test/e2e/specs/editor/plugins/block-api.spec.js` ## Flaky rate (_estimated_) `1 / 2` runs ## Errors
test
inserts the filtered hello world block even when filter added after block registration flaky test detected this is an auto generated issue by github actions please do not edit this manually test title inserts the filtered hello world block even when filter added after block registration test path home runner work gutenberg gutenberg test specs editor plugins block api spec js flaky rate estimated runs errors
1
324,861
27,826,108,036
IssuesEvent
2023-03-19 19:37:01
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
kvcoord: flake in TestMultiRangeScanReverseScanInconsistent
C-test-failure A-kv-transactions A-kv skipped-test GA-blocker T-kv branch-release-23.1
``` $ ./dev test --stress //pkg/kv/kvclient/kvcoord --filter=TestMultiRangeScanReverseScanInconsistent ... --- FAIL: TestMultiRangeScanReverseScanInconsistent (1.52s) test_log_scope.go:161: test logs captured to: /tmp/_tmp/e1d742d7ffb97ae3adc81944eac18765/logTestMultiRangeScanReverseScanInconsistent3519145941 test_log_scope.go:79: use -show-logs to present logs inline --- FAIL: TestMultiRangeScanReverseScanInconsistent/INCONSISTENT (0.70s) dist_sender_server_test.go:1160: 0: expected 1 row; got 0 [] dist_sender_server_test.go:1168: -- test log scope end -- FAIL ``` Also found here: https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_UnitTests_BazelUnitTests/7504394?buildTab=overview&showRootCauses=false&expandBuildProblemsSection=true&expandBuildTestsSection=true&expandBuildChangesSection=true&expandBuildDeploymentsSection=true#%2Ftmp cc @nvanbenschoten for triage Jira issue: CRDB-21459
2.0
kvcoord: flake in TestMultiRangeScanReverseScanInconsistent - ``` $ ./dev test --stress //pkg/kv/kvclient/kvcoord --filter=TestMultiRangeScanReverseScanInconsistent ... --- FAIL: TestMultiRangeScanReverseScanInconsistent (1.52s) test_log_scope.go:161: test logs captured to: /tmp/_tmp/e1d742d7ffb97ae3adc81944eac18765/logTestMultiRangeScanReverseScanInconsistent3519145941 test_log_scope.go:79: use -show-logs to present logs inline --- FAIL: TestMultiRangeScanReverseScanInconsistent/INCONSISTENT (0.70s) dist_sender_server_test.go:1160: 0: expected 1 row; got 0 [] dist_sender_server_test.go:1168: -- test log scope end -- FAIL ``` Also found here: https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_UnitTests_BazelUnitTests/7504394?buildTab=overview&showRootCauses=false&expandBuildProblemsSection=true&expandBuildTestsSection=true&expandBuildChangesSection=true&expandBuildDeploymentsSection=true#%2Ftmp cc @nvanbenschoten for triage Jira issue: CRDB-21459
test
kvcoord flake in testmultirangescanreversescaninconsistent dev test stress pkg kv kvclient kvcoord filter testmultirangescanreversescaninconsistent fail testmultirangescanreversescaninconsistent test log scope go test logs captured to tmp tmp test log scope go use show logs to present logs inline fail testmultirangescanreversescaninconsistent inconsistent dist sender server test go expected row got dist sender server test go test log scope end fail also found here cc nvanbenschoten for triage jira issue crdb
1
199,371
22,693,317,141
IssuesEvent
2022-07-05 01:12:12
attesch/swapi
https://api.github.com/repos/attesch/swapi
closed
CVE-2018-16984 (Medium) detected in Django-1.7.4-py2.py3-none-any.whl - autoclosed
security vulnerability
## CVE-2018-16984 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.7.4-py2.py3-none-any.whl</b></p></summary> <p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/c9/1e/66f185ca0d4d0ca11b94caeac96a33a13954963a8b563b67d11f50bfeee7/Django-1.7.4-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/c9/1e/66f185ca0d4d0ca11b94caeac96a33a13954963a8b563b67d11f50bfeee7/Django-1.7.4-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /swapi/requirements.txt</p> <p>Path to vulnerable library: teSource-ArchiveExtractor_cd0131d4-2ba0-4601-beec-b5f2a7e3636b/20190620051939_56417/20190620051812_depth_0/13/Django-2.2.2.tar/Django-2.2.2/django</p> <p> Dependency Hierarchy: - :x: **Django-1.7.4-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/attesch/swapi/commit/4f94ce0b66c94cc8d0908d27fdc9d39faa534139">4f94ce0b66c94cc8d0908d27fdc9d39faa534139</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Django 2.1 before 2.1.2, in which unprivileged users can read the password hashes of arbitrary accounts. The read-only password widget used by the Django Admin to display an obfuscated password hash was bypassed if a user has only the "view" permission (new in Django 2.1), resulting in display of the entire password hash to those users. This may result in a vulnerability for sites with legacy user accounts using insecure hashes. <p>Publish Date: 2018-10-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16984>CVE-2018-16984</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-16984">https://nvd.nist.gov/vuln/detail/CVE-2018-16984</a></p> <p>Release Date: 2018-10-02</p> <p>Fix Resolution: 2.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-16984 (Medium) detected in Django-1.7.4-py2.py3-none-any.whl - autoclosed - ## CVE-2018-16984 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>Django-1.7.4-py2.py3-none-any.whl</b></p></summary> <p>A high-level Python Web framework that encourages rapid development and clean, pragmatic design.</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/c9/1e/66f185ca0d4d0ca11b94caeac96a33a13954963a8b563b67d11f50bfeee7/Django-1.7.4-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/c9/1e/66f185ca0d4d0ca11b94caeac96a33a13954963a8b563b67d11f50bfeee7/Django-1.7.4-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /swapi/requirements.txt</p> <p>Path to vulnerable library: teSource-ArchiveExtractor_cd0131d4-2ba0-4601-beec-b5f2a7e3636b/20190620051939_56417/20190620051812_depth_0/13/Django-2.2.2.tar/Django-2.2.2/django</p> <p> Dependency Hierarchy: - :x: **Django-1.7.4-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/attesch/swapi/commit/4f94ce0b66c94cc8d0908d27fdc9d39faa534139">4f94ce0b66c94cc8d0908d27fdc9d39faa534139</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue was discovered in Django 2.1 before 2.1.2, in which unprivileged users can read the password hashes of arbitrary accounts. The read-only password widget used by the Django Admin to display an obfuscated password hash was bypassed if a user has only the "view" permission (new in Django 2.1), resulting in display of the entire password hash to those users. This may result in a vulnerability for sites with legacy user accounts using insecure hashes. <p>Publish Date: 2018-10-02 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-16984>CVE-2018-16984</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: High - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2018-16984">https://nvd.nist.gov/vuln/detail/CVE-2018-16984</a></p> <p>Release Date: 2018-10-02</p> <p>Fix Resolution: 2.1.2</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in django none any whl autoclosed cve medium severity vulnerability vulnerable library django none any whl a high level python web framework that encourages rapid development and clean pragmatic design library home page a href path to dependency file swapi requirements txt path to vulnerable library tesource archiveextractor beec depth django tar django django dependency hierarchy x django none any whl vulnerable library found in head commit a href vulnerability details an issue was discovered in django before in which unprivileged users can read the password hashes of arbitrary accounts the read only password widget used by the django admin to display an obfuscated password hash was bypassed if a user has only the view permission new in django resulting in display of the entire password hash to those users this may result in a vulnerability for sites with legacy user accounts using insecure hashes publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required high user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
118,749
10,007,255,431
IssuesEvent
2019-07-14 09:05:55
inf3rno/patterns
https://api.github.com/repos/inf3rno/patterns
opened
Alt(ernatives) by patterns
do:2 - test do:3 - implement do:4 - document
I already use this in #27 I believe. It is already designed, I just need to test and document it.
1.0
Alt(ernatives) by patterns - I already use this in #27 I believe. It is already designed, I just need to test and document it.
test
alt ernatives by patterns i already use this in i believe it is already designed i just need to test and document it
1
193,163
6,882,072,762
IssuesEvent
2017-11-21 01:41:47
Marri/glowfic
https://api.github.com/repos/Marri/glowfic
opened
Audits for character & icon replace
2. high priority 9. hard dev enhancement
Currently we don't track the state prior to the replace, which means we can lose data if someone accidentally does a replace (especially to something used in the same thread, so we can't just readily disentangle it on a thread-by-thread basis). This is also a prerequisite to moderators getting the ability to do character / icon replace, since we want to be able to track that nothing silly has gone on (and revert it if a user makes a compelling case for it) – #3. It would also lead into us allowing users to see their recent replacements, and revert them as necessary, or perhaps display it on the relevant reply histories.
1.0
Audits for character & icon replace - Currently we don't track the state prior to the replace, which means we can lose data if someone accidentally does a replace (especially to something used in the same thread, so we can't just readily disentangle it on a thread-by-thread basis). This is also a prerequisite to moderators getting the ability to do character / icon replace, since we want to be able to track that nothing silly has gone on (and revert it if a user makes a compelling case for it) – #3. It would also lead into us allowing users to see their recent replacements, and revert them as necessary, or perhaps display it on the relevant reply histories.
non_test
audits for character icon replace currently we don t track the state prior to the replace which means we can lose data if someone accidentally does a replace especially to something used in the same thread so we can t just readily disentangle it on a thread by thread basis this is also a prerequisite to moderators getting the ability to do character icon replace since we want to be able to track that nothing silly has gone on and revert it if a user makes a compelling case for it – it would also lead into us allowing users to see their recent replacements and revert them as necessary or perhaps display it on the relevant reply histories
0
203,595
15,376,218,272
IssuesEvent
2021-03-02 15:44:55
avandvik/masters-thesis
https://api.github.com/repos/avandvik/masters-thesis
closed
Skrive tester til Evaluator
test
- Test isFeasibleLoad - Test isFeasibleDuration - Test instInMoreThanOneSequence - Test isIllegalPattern (få den skilt ut i submetode først)
1.0
Skrive tester til Evaluator - - Test isFeasibleLoad - Test isFeasibleDuration - Test instInMoreThanOneSequence - Test isIllegalPattern (få den skilt ut i submetode først)
test
skrive tester til evaluator test isfeasibleload test isfeasibleduration test instinmorethanonesequence test isillegalpattern få den skilt ut i submetode først
1
106,337
13,262,453,231
IssuesEvent
2020-08-20 21:49:34
elastic/eui
https://api.github.com/repos/elastic/eui
opened
Replace highlight.js as the engine for EuiCodeBlock
assign:designer
Although popular, highlight.js is pretty slow and getting a little old in its methodology. I often see it fail in complex syntax blocks and it'd be nice if we had something that provided virtualization out of the gate. I'd like to look into replacing it with [react-syntax-highlighter](https://github.com/react-syntax-highlighter/react-syntax-highlighter) which is backed by Prism JS (which I have some love and familiarity with) and comes with virtualization (which would close https://github.com/elastic/eui/issues/1208). I can give this a shot myself, but please give a yell if anything looks out of sort from that dependency. The bulk of the work I think will be transferring our styling over.
1.0
Replace highlight.js as the engine for EuiCodeBlock - Although popular, highlight.js is pretty slow and getting a little old in its methodology. I often see it fail in complex syntax blocks and it'd be nice if we had something that provided virtualization out of the gate. I'd like to look into replacing it with [react-syntax-highlighter](https://github.com/react-syntax-highlighter/react-syntax-highlighter) which is backed by Prism JS (which I have some love and familiarity with) and comes with virtualization (which would close https://github.com/elastic/eui/issues/1208). I can give this a shot myself, but please give a yell if anything looks out of sort from that dependency. The bulk of the work I think will be transferring our styling over.
non_test
replace highlight js as the engine for euicodeblock although popular highlight js is pretty slow and getting a little old in its methodology i often see it fail in complex syntax blocks and it d be nice if we had something that provided virtualization out of the gate i d like to look into replacing it with which is backed by prism js which i have some love and familiarity with and comes with virtualization which would close i can give this a shot myself but please give a yell if anything looks out of sort from that dependency the bulk of the work i think will be transferring our styling over
0
78,585
7,655,030,366
IssuesEvent
2018-05-10 11:38:11
minishift/minishift
https://api.github.com/repos/minishift/minishift
closed
Chdir to the test-dir as part of the beforeFeature()
component/integration-test kind/task priority/major
Right now the working directory of the integration tests is where `integration-test.go` file is located. However in some Gherkin steps we do define manipulation with files, which we want to do in the `test-dir` directory, where the tests are being created, run and deleted. This makes both `test-dir` and `integration-test.go` dependent on its relative locations. For example: ``` When file from "https://raw.githubusercontent.com/minishift/minishift/master/addons/anyuid/anyuid.addon" is downloaded into location "download/anyuid" # (this step prepends test-dir path to the path defined in feature) And executing "minishift addons install ../../out/integration-test/download/anyuid" succeeds ``` This is because the workdir is at `minishift/test/integration` but we want to work at `minishift/out/integration-test`. Problem is that if we would use `cmd-addons.feature` from custom location, it would fail. The feature should describe the paths from the directory in which the tests are executed. As fix we should add os.Chdir(test-dir) to the beforeFeature() so all the commands used in gherkin steps will start from the directory where we actually run the tests. Resulting in: ``` Scenario: Installing add-on from a folder Note: working directory when executing Minishift commands is: /test/integration. When file from "https://raw.githubusercontent.com/minishift/minishift/master/addons/anyuid/anyuid.addon" is downloaded into location "download/anyuid" And executing "minishift addons install download/anyuid" succeeds Then executing "minishift addons list" succeeds And stdout should contain "anyuid" ```
1.0
Chdir to the test-dir as part of the beforeFeature() - Right now the working directory of the integration tests is where `integration-test.go` file is located. However in some Gherkin steps we do define manipulation with files, which we want to do in the `test-dir` directory, where the tests are being created, run and deleted. This makes both `test-dir` and `integration-test.go` dependent on its relative locations. For example: ``` When file from "https://raw.githubusercontent.com/minishift/minishift/master/addons/anyuid/anyuid.addon" is downloaded into location "download/anyuid" # (this step prepends test-dir path to the path defined in feature) And executing "minishift addons install ../../out/integration-test/download/anyuid" succeeds ``` This is because the workdir is at `minishift/test/integration` but we want to work at `minishift/out/integration-test`. Problem is that if we would use `cmd-addons.feature` from custom location, it would fail. The feature should describe the paths from the directory in which the tests are executed. As fix we should add os.Chdir(test-dir) to the beforeFeature() so all the commands used in gherkin steps will start from the directory where we actually run the tests. Resulting in: ``` Scenario: Installing add-on from a folder Note: working directory when executing Minishift commands is: /test/integration. When file from "https://raw.githubusercontent.com/minishift/minishift/master/addons/anyuid/anyuid.addon" is downloaded into location "download/anyuid" And executing "minishift addons install download/anyuid" succeeds Then executing "minishift addons list" succeeds And stdout should contain "anyuid" ```
test
chdir to the test dir as part of the beforefeature right now the working directory of the integration tests is where integration test go file is located however in some gherkin steps we do define manipulation with files which we want to do in the test dir directory where the tests are being created run and deleted this makes both test dir and integration test go dependent on its relative locations for example when file from is downloaded into location download anyuid this step prepends test dir path to the path defined in feature and executing minishift addons install out integration test download anyuid succeeds this is because the workdir is at minishift test integration but we want to work at minishift out integration test problem is that if we would use cmd addons feature from custom location it would fail the feature should describe the paths from the directory in which the tests are executed as fix we should add os chdir test dir to the beforefeature so all the commands used in gherkin steps will start from the directory where we actually run the tests resulting in scenario installing add on from a folder note working directory when executing minishift commands is test integration when file from is downloaded into location download anyuid and executing minishift addons install download anyuid succeeds then executing minishift addons list succeeds and stdout should contain anyuid
1
52,522
6,260,033,638
IssuesEvent
2017-07-14 19:34:06
openbmc/openbmc-test-automation
https://api.github.com/repos/openbmc/openbmc-test-automation
closed
Skip tests/test_boot_policies.robot test
Test
Due to https://github.com/openbmc/openbmc/issues/1967 , we will need to skip this boot test for now
1.0
Skip tests/test_boot_policies.robot test - Due to https://github.com/openbmc/openbmc/issues/1967 , we will need to skip this boot test for now
test
skip tests test boot policies robot test due to we will need to skip this boot test for now
1
28,851
11,705,951,741
IssuesEvent
2020-03-07 19:00:13
franzbischoff/franzwebsite
https://api.github.com/repos/franzbischoff/franzwebsite
opened
CVE-2019-11358 (Medium) detected in jquery-2.2.2.js
security vulnerability
## CVE-2019-11358 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-2.2.2.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.2/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.2/jquery.js</a></p> <p>Path to vulnerable library: /franzwebsite/themes/robjhyndman/static/js/jquery.js,/franzwebsite/public/js/jquery.js</p> <p> Dependency Hierarchy: - :x: **jquery-2.2.2.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/franzbischoff/franzwebsite/commit/0a047e974d0c36ca7fd96cb3155b01b87f512170">0a047e974d0c36ca7fd96cb3155b01b87f512170</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-11358 (Medium) detected in jquery-2.2.2.js - ## CVE-2019-11358 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>jquery-2.2.2.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.2/jquery.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/2.2.2/jquery.js</a></p> <p>Path to vulnerable library: /franzwebsite/themes/robjhyndman/static/js/jquery.js,/franzwebsite/public/js/jquery.js</p> <p> Dependency Hierarchy: - :x: **jquery-2.2.2.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/franzbischoff/franzwebsite/commit/0a047e974d0c36ca7fd96cb3155b01b87f512170">0a047e974d0c36ca7fd96cb3155b01b87f512170</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> jQuery before 3.4.0, as used in Drupal, Backdrop CMS, and other products, mishandles jQuery.extend(true, {}, ...) because of Object.prototype pollution. If an unsanitized source object contained an enumerable __proto__ property, it could extend the native Object.prototype. <p>Publish Date: 2019-04-20 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-11358>CVE-2019-11358</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2019-11358</a></p> <p>Release Date: 2019-04-20</p> <p>Fix Resolution: 3.4.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in jquery js cve medium severity vulnerability vulnerable library jquery js javascript library for dom operations library home page a href path to vulnerable library franzwebsite themes robjhyndman static js jquery js franzwebsite public js jquery js dependency hierarchy x jquery js vulnerable library found in head commit a href vulnerability details jquery before as used in drupal backdrop cms and other products mishandles jquery extend true because of object prototype pollution if an unsanitized source object contained an enumerable proto property it could extend the native object prototype publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
29,233
4,479,537,582
IssuesEvent
2016-08-27 17:29:42
artiator/artiator-database
https://api.github.com/repos/artiator/artiator-database
opened
Add migration generator
refactoring / automation / testing
Add Rake task `db:migrate:new` which takes one argument - snake-cased migration name. Raise exception if migration name is invalid. Generate migration file like Rails does it.
1.0
Add migration generator - Add Rake task `db:migrate:new` which takes one argument - snake-cased migration name. Raise exception if migration name is invalid. Generate migration file like Rails does it.
test
add migration generator add rake task db migrate new which takes one argument snake cased migration name raise exception if migration name is invalid generate migration file like rails does it
1
129,740
12,417,891,507
IssuesEvent
2020-05-22 21:59:58
FormidableLabs/badges
https://api.github.com/repos/FormidableLabs/badges
opened
Infra: Add docs/scripts to completely nuke a deployment.
documentation enhancement
Including: - [ ] All TF resources - [ ] The state bucket / DDB locks for TF. For reference see how we nuke versioned buckets in the reference app: https://github.com/FormidableLabs/aws-lambda-serverless-reference/blob/master/package.json#L35-L38
1.0
Infra: Add docs/scripts to completely nuke a deployment. - Including: - [ ] All TF resources - [ ] The state bucket / DDB locks for TF. For reference see how we nuke versioned buckets in the reference app: https://github.com/FormidableLabs/aws-lambda-serverless-reference/blob/master/package.json#L35-L38
non_test
infra add docs scripts to completely nuke a deployment including all tf resources the state bucket ddb locks for tf for reference see how we nuke versioned buckets in the reference app
0
664,728
22,286,527,340
IssuesEvent
2022-06-11 18:15:11
ctm/mb2-doc
https://api.github.com/repos/ctm/mb2-doc
closed
Enhance tag functionality to support inclusion
chore high priority easy
Add the ability to create tags that are used for identification, but not (necessarily) restriction. Currently adding a tag to an event has the effect of restricting the event to players who also have that tag. That worked well for various ARGE events, but is not what we want for WSOPS. We want to be able to add the `wsops_2022` tag and still allow everyone to participate. Adding an `inclusive` `BOOLEAN` column to `tags` should work, assuming we add the logic that omits the restriction when `inclusive` is true. An alternative would be to have a `auto_add` column that automatically adds a tag to a newly created player and then run a migration that adds that tag to all the users. The downside is this means that we have to add a row in `player_tags` for each player. OTOH, the exclusion code doesn't have to be changed and that representation lends itself to allowing an administrator or the player itself to remove a player from a series. A third alternative would be to add both an `inclusive` column and then have a separate table for noting players who are not to be included. Although this solution is the most complex, it can be added after implementing `inclusive`, which is why initially simply implementing `inclusive` is what I'll do.
1.0
Enhance tag functionality to support inclusion - Add the ability to create tags that are used for identification, but not (necessarily) restriction. Currently adding a tag to an event has the effect of restricting the event to players who also have that tag. That worked well for various ARGE events, but is not what we want for WSOPS. We want to be able to add the `wsops_2022` tag and still allow everyone to participate. Adding an `inclusive` `BOOLEAN` column to `tags` should work, assuming we add the logic that omits the restriction when `inclusive` is true. An alternative would be to have a `auto_add` column that automatically adds a tag to a newly created player and then run a migration that adds that tag to all the users. The downside is this means that we have to add a row in `player_tags` for each player. OTOH, the exclusion code doesn't have to be changed and that representation lends itself to allowing an administrator or the player itself to remove a player from a series. A third alternative would be to add both an `inclusive` column and then have a separate table for noting players who are not to be included. Although this solution is the most complex, it can be added after implementing `inclusive`, which is why initially simply implementing `inclusive` is what I'll do.
non_test
enhance tag functionality to support inclusion add the ability to create tags that are used for identification but not necessarily restriction currently adding a tag to an event has the effect of restricting the event to players who also have that tag that worked well for various arge events but is not what we want for wsops we want to be able to add the wsops tag and still allow everyone to participate adding an inclusive boolean column to tags should work assuming we add the logic that omits the restriction when inclusive is true an alternative would be to have a auto add column that automatically adds a tag to a newly created player and then run a migration that adds that tag to all the users the downside is this means that we have to add a row in player tags for each player otoh the exclusion code doesn t have to be changed and that representation lends itself to allowing an administrator or the player itself to remove a player from a series a third alternative would be to add both an inclusive column and then have a separate table for noting players who are not to be included although this solution is the most complex it can be added after implementing inclusive which is why initially simply implementing inclusive is what i ll do
0
184,363
14,977,152,062
IssuesEvent
2021-01-28 09:06:51
telerik/kendo-react
https://api.github.com/repos/telerik/kendo-react
opened
TileLayout Controlling the Position demo uses an undefined argument.
documentation pkg:layout
TileLayout Controlling the Position demo uses an undefined argument. The demo uses `e.positions` where the correct argument is value.
1.0
TileLayout Controlling the Position demo uses an undefined argument. - TileLayout Controlling the Position demo uses an undefined argument. The demo uses `e.positions` where the correct argument is value.
non_test
tilelayout controlling the position demo uses an undefined argument tilelayout controlling the position demo uses an undefined argument the demo uses e positions where the correct argument is value
0
113,258
9,633,946,435
IssuesEvent
2019-05-15 19:58:34
Daedeross/Subterra
https://api.github.com/repos/Daedeross/Subterra
opened
Test Bobs Mods Compatibility
help wanted testing
Test Compatiblity with Bob's suit of mods: - [ ] Bob's Adjustable Inserters - [ ] Bob's Ores - [ ] Bob's Assembling Machines - [ ] Bob's Electronics - [ ] Bob's Power - [ ] Bob's Greenhouse
1.0
Test Bobs Mods Compatibility - Test Compatiblity with Bob's suit of mods: - [ ] Bob's Adjustable Inserters - [ ] Bob's Ores - [ ] Bob's Assembling Machines - [ ] Bob's Electronics - [ ] Bob's Power - [ ] Bob's Greenhouse
test
test bobs mods compatibility test compatiblity with bob s suit of mods bob s adjustable inserters bob s ores bob s assembling machines bob s electronics bob s power bob s greenhouse
1
244,536
26,421,856,742
IssuesEvent
2023-01-13 21:24:40
opensearch-project/data-prepper
https://api.github.com/repos/opensearch-project/data-prepper
closed
CVE-2022-40899 (High) detected in future-0.18.2.tar.gz - autoclosed
security vulnerability
## CVE-2022-40899 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>future-0.18.2.tar.gz</b></p></summary> <p>Clean single-source support for Python 3 and 2</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/45/0b/38b06fd9b92dc2b68d58b75f900e97884c45bedd2ff83203d933cf5851c9/future-0.18.2.tar.gz">https://files.pythonhosted.org/packages/45/0b/38b06fd9b92dc2b68d58b75f900e97884c45bedd2ff83203d933cf5851c9/future-0.18.2.tar.gz</a></p> <p>Path to dependency file: /examples/trace-analytics-sample-app/sample-app/requirements.txt</p> <p>Path to vulnerable library: /examples/trace-analytics-sample-app/sample-app/requirements.txt</p> <p> Dependency Hierarchy: - dash-1.17.0.tar.gz (Root Library) - :x: **future-0.18.2.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/data-prepper/commit/90bdaa7e7833bdd504c817e49d4434b4d8880f56">90bdaa7e7833bdd504c817e49d4434b4d8880f56</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue discovered in Python Charmers Future 0.18.2 and earlier allows remote attackers to cause a denial of service via crafted Set-Cookie header from malicious web server. <p>Publish Date: 2022-12-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40899>CVE-2022-40899</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
True
CVE-2022-40899 (High) detected in future-0.18.2.tar.gz - autoclosed - ## CVE-2022-40899 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>future-0.18.2.tar.gz</b></p></summary> <p>Clean single-source support for Python 3 and 2</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/45/0b/38b06fd9b92dc2b68d58b75f900e97884c45bedd2ff83203d933cf5851c9/future-0.18.2.tar.gz">https://files.pythonhosted.org/packages/45/0b/38b06fd9b92dc2b68d58b75f900e97884c45bedd2ff83203d933cf5851c9/future-0.18.2.tar.gz</a></p> <p>Path to dependency file: /examples/trace-analytics-sample-app/sample-app/requirements.txt</p> <p>Path to vulnerable library: /examples/trace-analytics-sample-app/sample-app/requirements.txt</p> <p> Dependency Hierarchy: - dash-1.17.0.tar.gz (Root Library) - :x: **future-0.18.2.tar.gz** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/opensearch-project/data-prepper/commit/90bdaa7e7833bdd504c817e49d4434b4d8880f56">90bdaa7e7833bdd504c817e49d4434b4d8880f56</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> An issue discovered in Python Charmers Future 0.18.2 and earlier allows remote attackers to cause a denial of service via crafted Set-Cookie header from malicious web server. <p>Publish Date: 2022-12-23 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-40899>CVE-2022-40899</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p>
non_test
cve high detected in future tar gz autoclosed cve high severity vulnerability vulnerable library future tar gz clean single source support for python and library home page a href path to dependency file examples trace analytics sample app sample app requirements txt path to vulnerable library examples trace analytics sample app sample app requirements txt dependency hierarchy dash tar gz root library x future tar gz vulnerable library found in head commit a href found in base branch main vulnerability details an issue discovered in python charmers future and earlier allows remote attackers to cause a denial of service via crafted set cookie header from malicious web server publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href
0
105,359
11,449,097,535
IssuesEvent
2020-02-06 06:03:14
acord-robotics/manacaster
https://api.github.com/repos/acord-robotics/manacaster
opened
New Plan --> UnityDev by ACo/0RD
documentation enhancement unity
Also in I spoke to my computer science teacher, Mr Musovic today. We also have a new part-time member, my friend Martin from my comp sci class. We have decided on a plan & and a timeline for the game. **Keep it light-scale** **Follow storyline** - http://acord-robotics.github.io/premium/starsailors.html (remember /premium = /acord) **Use Blender** for assets --> Martin can help me with this **Sit down and make paper models**
1.0
New Plan --> UnityDev by ACo/0RD - Also in I spoke to my computer science teacher, Mr Musovic today. We also have a new part-time member, my friend Martin from my comp sci class. We have decided on a plan & and a timeline for the game. **Keep it light-scale** **Follow storyline** - http://acord-robotics.github.io/premium/starsailors.html (remember /premium = /acord) **Use Blender** for assets --> Martin can help me with this **Sit down and make paper models**
non_test
new plan unitydev by aco also in i spoke to my computer science teacher mr musovic today we also have a new part time member my friend martin from my comp sci class we have decided on a plan and a timeline for the game keep it light scale follow storyline remember premium acord use blender for assets martin can help me with this sit down and make paper models
0
65,852
6,976,455,023
IssuesEvent
2017-12-12 11:07:11
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Ceph RBD support in Rancher storage
area/storage kind/feature status/reopened status/resolved status/to-test team/cn
Some users expect to use RBD storage under Cattle. **Jewel** is the latest LTS version, so I will give priority to support it and test it well. Here are some usage information. Rancher RBD catalog item: Repo: https://github.com/niusmallnan/rancher-rbd-catalog Use this repo to deploy the rancher-rbd driver. You can use the following compose files to test. **Stack Scoped**: ``` version: '2' services: foo: image: nginx volumes: - bar:/var/lib/storage volumes: bar: driver: rancher-rbd driver_opts: size: 2G pool: rbd ``` For stack scoped: The containers for any service using the RBD volume will automatically be scheduled onto the same host as where the RBD volume is attached to. As we all know, Ceph's configuration file is very complex, so it is difficult to automatically generate it. You need to manually synchronize the configuration file in the /etc/ceph directory on all rancher-agent nodes, rancher-rbd will mount this directory. If you do not have a Ceph setup, you can use [ceph-docker](https://github.com/ceph/ceph-docker/tree/master/ceph-releases/jewel/ubuntu/16.04/demo) to quickly build your setup. ``` docker run -d --net=host -v /etc/ceph:/etc/ceph -e MON_IP=192.168.0.20 -e CEPH_PUBLIC_NETWORK=192.168.0.0/24 ceph/demo:tag-build-master-jewel-ubuntu-16.04 ``` Due to librbd compatibility with the kernel, it is recommended to use ubuntu16.04. It is recommended to use Rancher v1.6.0+ .
1.0
Ceph RBD support in Rancher storage - Some users expect to use RBD storage under Cattle. **Jewel** is the latest LTS version, so I will give priority to support it and test it well. Here are some usage information. Rancher RBD catalog item: Repo: https://github.com/niusmallnan/rancher-rbd-catalog Use this repo to deploy the rancher-rbd driver. You can use the following compose files to test. **Stack Scoped**: ``` version: '2' services: foo: image: nginx volumes: - bar:/var/lib/storage volumes: bar: driver: rancher-rbd driver_opts: size: 2G pool: rbd ``` For stack scoped: The containers for any service using the RBD volume will automatically be scheduled onto the same host as where the RBD volume is attached to. As we all know, Ceph's configuration file is very complex, so it is difficult to automatically generate it. You need to manually synchronize the configuration file in the /etc/ceph directory on all rancher-agent nodes, rancher-rbd will mount this directory. If you do not have a Ceph setup, you can use [ceph-docker](https://github.com/ceph/ceph-docker/tree/master/ceph-releases/jewel/ubuntu/16.04/demo) to quickly build your setup. ``` docker run -d --net=host -v /etc/ceph:/etc/ceph -e MON_IP=192.168.0.20 -e CEPH_PUBLIC_NETWORK=192.168.0.0/24 ceph/demo:tag-build-master-jewel-ubuntu-16.04 ``` Due to librbd compatibility with the kernel, it is recommended to use ubuntu16.04. It is recommended to use Rancher v1.6.0+ .
test
ceph rbd support in rancher storage some users expect to use rbd storage under cattle jewel is the latest lts version so i will give priority to support it and test it well here are some usage information rancher rbd catalog item repo use this repo to deploy the rancher rbd driver you can use the following compose files to test stack scoped version services foo image nginx volumes bar var lib storage volumes bar driver rancher rbd driver opts size pool rbd for stack scoped the containers for any service using the rbd volume will automatically be scheduled onto the same host as where the rbd volume is attached to as we all know ceph s configuration file is very complex so it is difficult to automatically generate it you need to manually synchronize the configuration file in the etc ceph directory on all rancher agent nodes rancher rbd will mount this directory if you do not have a ceph setup you can use to quickly build your setup docker run d net host v etc ceph etc ceph e mon ip e ceph public network ceph demo tag build master jewel ubuntu due to librbd compatibility with the kernel it is recommended to use it is recommended to use rancher
1
43,756
5,559,537,651
IssuesEvent
2017-03-24 17:10:26
brangerbriz/liBB.js
https://api.github.com/repos/brangerbriz/liBB.js
closed
Test polarToCartesian(...) and cartesianToPolar(...) functions in BB.MathUtils.
test
These functions appear to be correct as [Processing explains](https://processing.org/examples/polartocartesian.html) and uses them however they do not provide the expected results from online polar-to-cartesian conversion resources like [this](http://www.engineeringtoolbox.com/converting-cartesian-polar-coordinates-d_1347.html).
1.0
Test polarToCartesian(...) and cartesianToPolar(...) functions in BB.MathUtils. - These functions appear to be correct as [Processing explains](https://processing.org/examples/polartocartesian.html) and uses them however they do not provide the expected results from online polar-to-cartesian conversion resources like [this](http://www.engineeringtoolbox.com/converting-cartesian-polar-coordinates-d_1347.html).
test
test polartocartesian and cartesiantopolar functions in bb mathutils these functions appear to be correct as and uses them however they do not provide the expected results from online polar to cartesian conversion resources like
1
254,734
8,087,340,006
IssuesEvent
2018-08-09 01:08:48
C3DSU/e-DefPR
https://api.github.com/repos/C3DSU/e-DefPR
closed
Implementar CRUD de permissões
Category: Backend Priority: Medium Stage: Review Type: New-feature
A execução dessa tarefa depende da #165. Implementar as rotas CRUD (Create, Read, Update, Delete) de permissões seguindo o padrão atual (verificar a implementação das rotas de User em Controllers/Api/UserController). required by #276
1.0
Implementar CRUD de permissões - A execução dessa tarefa depende da #165. Implementar as rotas CRUD (Create, Read, Update, Delete) de permissões seguindo o padrão atual (verificar a implementação das rotas de User em Controllers/Api/UserController). required by #276
non_test
implementar crud de permissões a execução dessa tarefa depende da implementar as rotas crud create read update delete de permissões seguindo o padrão atual verificar a implementação das rotas de user em controllers api usercontroller required by
0
190,305
14,541,498,138
IssuesEvent
2020-12-15 14:38:31
cyfronet-fid/sat4envi
https://api.github.com/repos/cyfronet-fid/sat4envi
closed
The city name is not highlighted during navigating with the arrows
bug dev-env-test front-end pre-prod-test
Tested on instance https://sok.grid.cyfronet.pl SOK version v13.2.0 Steps: 1. Go to the /map/products 2. Enter "War" in search input 3. Use the down arrow to select a location ↓ **What happends?** Selected city is not highlighted **What was expected to happend?** Selected city is highlighted
2.0
The city name is not highlighted during navigating with the arrows - Tested on instance https://sok.grid.cyfronet.pl SOK version v13.2.0 Steps: 1. Go to the /map/products 2. Enter "War" in search input 3. Use the down arrow to select a location ↓ **What happends?** Selected city is not highlighted **What was expected to happend?** Selected city is highlighted
test
the city name is not highlighted during navigating with the arrows tested on instance sok version steps go to the map products enter war in search input use the down arrow to select a location ↓ what happends selected city is not highlighted what was expected to happend selected city is highlighted
1
117,797
9,959,540,323
IssuesEvent
2019-07-06 07:51:15
KSP-CKAN/CKAN
https://api.github.com/repos/KSP-CKAN/CKAN
closed
MainModList test is not working correctly
Bug GUI Tests
Problem ------- This test passes for the wrong reasons: https://github.com/KSP-CKAN/CKAN/blob/7b9da8c6bb167cb64dd7f2f7836d5d16b41e6826/Tests/GUI/MainModList.cs#L56-L81 It is trying to have one installed module and one module-to-be-installed that conflict with one another, and it wants this to cause an `InconsistentKraken` to be thrown. However, what actually happens is that it throws `InconsistentKraken` because its installed module is missing a dependency. It simulates having FireSpitter installed, but the source file has this dependency: https://github.com/KSP-CKAN/CKAN/blob/7b9da8c6bb167cb64dd7f2f7836d5d16b41e6826/Tests/Data/Firespitter-6.3.5.ckan#L12-L17 ... which is not added to the registry. This is an inconsistent state. If you add this to the test code to remove the dependency: ```csharp module.depends = null; ``` ... then the test starts failing: ``` 1) Failed : Tests.GUI.MainModListTests.ComputeChangeSetFromModList_WithConflictingMods_ThrowsInconsistentKraken Expected exception of type: CKAN.InconsistentKraken at Tests.UtilStatic.<Throws>d__4`1.MoveNext() in C:\Users\Paul\github\CKAN\Tests\Util.cs:line 95 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Runtime.CompilerServices.TaskAwaiter.GetResult() at Tests.GUI.MainModListTests.<ComputeChangeSetFromModList_WithConflictingMods_ThrowsInconsistentKraken>d__4.MoveNext() in C:\Users\Paul\github\CKAN\Tests\GUI\MainModList.cs:line 81 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at NUnit.Framework.Internal.AsyncInvocationRegion.AsyncTaskInvocationRegion.WaitForPendingOperationsToComplete(Object invocationResult) at NUnit.Framework.Internal.Commands.TestMethodCommand.RunAsyncTestMethod(TestExecutionContext context) ``` This shouldn't happen if the conflict relationship is what's causing the exception to be thrown. This means we're not testing what we think we're testing. Found while looking into test failures of #2740.
1.0
MainModList test is not working correctly - Problem ------- This test passes for the wrong reasons: https://github.com/KSP-CKAN/CKAN/blob/7b9da8c6bb167cb64dd7f2f7836d5d16b41e6826/Tests/GUI/MainModList.cs#L56-L81 It is trying to have one installed module and one module-to-be-installed that conflict with one another, and it wants this to cause an `InconsistentKraken` to be thrown. However, what actually happens is that it throws `InconsistentKraken` because its installed module is missing a dependency. It simulates having FireSpitter installed, but the source file has this dependency: https://github.com/KSP-CKAN/CKAN/blob/7b9da8c6bb167cb64dd7f2f7836d5d16b41e6826/Tests/Data/Firespitter-6.3.5.ckan#L12-L17 ... which is not added to the registry. This is an inconsistent state. If you add this to the test code to remove the dependency: ```csharp module.depends = null; ``` ... then the test starts failing: ``` 1) Failed : Tests.GUI.MainModListTests.ComputeChangeSetFromModList_WithConflictingMods_ThrowsInconsistentKraken Expected exception of type: CKAN.InconsistentKraken at Tests.UtilStatic.<Throws>d__4`1.MoveNext() in C:\Users\Paul\github\CKAN\Tests\Util.cs:line 95 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at System.Runtime.CompilerServices.TaskAwaiter.HandleNonSuccessAndDebuggerNotification(Task task) at System.Runtime.CompilerServices.TaskAwaiter.GetResult() at Tests.GUI.MainModListTests.<ComputeChangeSetFromModList_WithConflictingMods_ThrowsInconsistentKraken>d__4.MoveNext() in C:\Users\Paul\github\CKAN\Tests\GUI\MainModList.cs:line 81 --- End of stack trace from previous location where exception was thrown --- at System.Runtime.ExceptionServices.ExceptionDispatchInfo.Throw() at NUnit.Framework.Internal.AsyncInvocationRegion.AsyncTaskInvocationRegion.WaitForPendingOperationsToComplete(Object invocationResult) at NUnit.Framework.Internal.Commands.TestMethodCommand.RunAsyncTestMethod(TestExecutionContext context) ``` This shouldn't happen if the conflict relationship is what's causing the exception to be thrown. This means we're not testing what we think we're testing. Found while looking into test failures of #2740.
test
mainmodlist test is not working correctly problem this test passes for the wrong reasons it is trying to have one installed module and one module to be installed that conflict with one another and it wants this to cause an inconsistentkraken to be thrown however what actually happens is that it throws inconsistentkraken because its installed module is missing a dependency it simulates having firespitter installed but the source file has this dependency which is not added to the registry this is an inconsistent state if you add this to the test code to remove the dependency csharp module depends null then the test starts failing failed tests gui mainmodlisttests computechangesetfrommodlist withconflictingmods throwsinconsistentkraken expected exception of type ckan inconsistentkraken at tests utilstatic d movenext in c users paul github ckan tests util cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at system runtime compilerservices taskawaiter handlenonsuccessanddebuggernotification task task at system runtime compilerservices taskawaiter getresult at tests gui mainmodlisttests d movenext in c users paul github ckan tests gui mainmodlist cs line end of stack trace from previous location where exception was thrown at system runtime exceptionservices exceptiondispatchinfo throw at nunit framework internal asyncinvocationregion asynctaskinvocationregion waitforpendingoperationstocomplete object invocationresult at nunit framework internal commands testmethodcommand runasynctestmethod testexecutioncontext context this shouldn t happen if the conflict relationship is what s causing the exception to be thrown this means we re not testing what we think we re testing found while looking into test failures of
1
34,304
7,808,990,518
IssuesEvent
2018-06-11 22:09:32
Microsoft/vscode-python
https://api.github.com/repos/Microsoft/vscode-python
closed
Wrong folding for construction with multiline strings:
feature-code navigation needs more info type-bug
## Environment data - VS Code version: 1.19.3 - Extension version (available under the Extensions sidebar): 2018.4.0 - OS and version: Win7 x64 - Python version (& distribution if applicable, e.g. Anaconda): Python 3.6 - Type of virtual environment used (N/A | venv | virtualenv | conda | ...): NA - Relevant/affected Python packages and their versions: - ## Actual behavior If I enter code like this: ```python class Test(): def test(self): template = """SOME TEXT HERE""" # <- Collapsing brokes here pass def test(): pass ``` And collapse definition of Test class I expect to see entire class collapsed. But it collapse it without counting quotes. (see attachment ![collapsed](https://user-images.githubusercontent.com/6580639/39887161-83e77126-545f-11e8-96c9-70843fc84875.png) ) ## Expected behavior An entire class definition will be collapsed properly
1.0
Wrong folding for construction with multiline strings: - ## Environment data - VS Code version: 1.19.3 - Extension version (available under the Extensions sidebar): 2018.4.0 - OS and version: Win7 x64 - Python version (& distribution if applicable, e.g. Anaconda): Python 3.6 - Type of virtual environment used (N/A | venv | virtualenv | conda | ...): NA - Relevant/affected Python packages and their versions: - ## Actual behavior If I enter code like this: ```python class Test(): def test(self): template = """SOME TEXT HERE""" # <- Collapsing brokes here pass def test(): pass ``` And collapse definition of Test class I expect to see entire class collapsed. But it collapse it without counting quotes. (see attachment ![collapsed](https://user-images.githubusercontent.com/6580639/39887161-83e77126-545f-11e8-96c9-70843fc84875.png) ) ## Expected behavior An entire class definition will be collapsed properly
non_test
wrong folding for construction with multiline strings environment data vs code version extension version available under the extensions sidebar os and version python version distribution if applicable e g anaconda python type of virtual environment used n a venv virtualenv conda na relevant affected python packages and their versions actual behavior if i enter code like this python class test def test self template some text here collapsing brokes here pass def test pass and collapse definition of test class i expect to see entire class collapsed but it collapse it without counting quotes see attachment expected behavior an entire class definition will be collapsed properly
0
563,602
16,701,682,214
IssuesEvent
2021-06-09 04:00:47
googleapis/python-api-core
https://api.github.com/repos/googleapis/python-api-core
closed
ServiceUnavailable missing in polling retries
priority: p2 type: bug
"503 service unavailable" errors are not being retried when long running operations (e.g. [dataproc create_cluster](https://googleapis.dev/python/dataproc/latest/dataproc_v1/services.html?highlight=create_cluster#google.cloud.dataproc_v1.services.cluster_controller.ClusterControllerClient.create_cluster)) are polled. Currently, the [default retry predicate (retry.py)](https://github.com/googleapis/python-api-core/blob/ed092271ab94c7ef53c40ce8e73ca2c48e97159c/google/api_core/retry.py#L100) includes `google.api_core.exceptions.ServiceUnavailable`, but the [polling retry predicate (polling.py)](https://github.com/googleapis/python-api-core/blob/ed092271ab94c7ef53c40ce8e73ca2c48e97159c/google/api_core/future/polling.py#L32) overrides the former and doesn't include ServiceUnavailable. This doesn't look intentional: if retry.py considers ServiceUnavailable to be transient and retriable, so should polling.py. This is causing issues for customers: e.g. their Cloud Function crashes during polling when they encounter a single `connection reset`.
1.0
ServiceUnavailable missing in polling retries - "503 service unavailable" errors are not being retried when long running operations (e.g. [dataproc create_cluster](https://googleapis.dev/python/dataproc/latest/dataproc_v1/services.html?highlight=create_cluster#google.cloud.dataproc_v1.services.cluster_controller.ClusterControllerClient.create_cluster)) are polled. Currently, the [default retry predicate (retry.py)](https://github.com/googleapis/python-api-core/blob/ed092271ab94c7ef53c40ce8e73ca2c48e97159c/google/api_core/retry.py#L100) includes `google.api_core.exceptions.ServiceUnavailable`, but the [polling retry predicate (polling.py)](https://github.com/googleapis/python-api-core/blob/ed092271ab94c7ef53c40ce8e73ca2c48e97159c/google/api_core/future/polling.py#L32) overrides the former and doesn't include ServiceUnavailable. This doesn't look intentional: if retry.py considers ServiceUnavailable to be transient and retriable, so should polling.py. This is causing issues for customers: e.g. their Cloud Function crashes during polling when they encounter a single `connection reset`.
non_test
serviceunavailable missing in polling retries service unavailable errors are not being retried when long running operations e g are polled currently the includes google api core exceptions serviceunavailable but the overrides the former and doesn t include serviceunavailable this doesn t look intentional if retry py considers serviceunavailable to be transient and retriable so should polling py this is causing issues for customers e g their cloud function crashes during polling when they encounter a single connection reset
0
31,426
4,706,373,747
IssuesEvent
2016-10-13 16:57:55
Microsoft/vsts-tasks
https://api.github.com/repos/Microsoft/vsts-tasks
reopened
Run Functional Test runs slowly compared to MTM Environments
Area: Test
- Running functional tests in TFS 15 with Vnext in comparison to the old system with MTM and Tests Environments, it is awfully slow. it takes like 10 minutes after initial test start, before the first tests are started. And while running the tests, they take longer as normal. - Distribution of tests is slightly "unhappy", tests get distributed at the beginning of the test run, but if 1 machine is finished, while the other one has still 5 long test runs doesn't make sense. Bucket size was a way more intelligent system because it distributed the tests one after another, with the test distribution now the tests get distributed at the start of the test run. Seems like Bucket size get's ignored at RunSettings file now. is there are a way to improve this? We have updated to RC2 and we are not happy with the test outcome. Feeling like test tast is a bottleneck [not implying that tfs 15 rc2 is bad, we realy love it! :)]
1.0
Run Functional Test runs slowly compared to MTM Environments - - Running functional tests in TFS 15 with Vnext in comparison to the old system with MTM and Tests Environments, it is awfully slow. it takes like 10 minutes after initial test start, before the first tests are started. And while running the tests, they take longer as normal. - Distribution of tests is slightly "unhappy", tests get distributed at the beginning of the test run, but if 1 machine is finished, while the other one has still 5 long test runs doesn't make sense. Bucket size was a way more intelligent system because it distributed the tests one after another, with the test distribution now the tests get distributed at the start of the test run. Seems like Bucket size get's ignored at RunSettings file now. is there are a way to improve this? We have updated to RC2 and we are not happy with the test outcome. Feeling like test tast is a bottleneck [not implying that tfs 15 rc2 is bad, we realy love it! :)]
test
run functional test runs slowly compared to mtm environments running functional tests in tfs with vnext in comparison to the old system with mtm and tests environments it is awfully slow it takes like minutes after initial test start before the first tests are started and while running the tests they take longer as normal distribution of tests is slightly unhappy tests get distributed at the beginning of the test run but if machine is finished while the other one has still long test runs doesn t make sense bucket size was a way more intelligent system because it distributed the tests one after another with the test distribution now the tests get distributed at the start of the test run seems like bucket size get s ignored at runsettings file now is there are a way to improve this we have updated to and we are not happy with the test outcome feeling like test tast is a bottleneck
1
483,598
13,926,795,165
IssuesEvent
2020-10-21 18:49:24
open-telemetry/opentelemetry-specification
https://api.github.com/repos/open-telemetry/opentelemetry-specification
opened
Rename SpanReference to SpanContext (revert #1075)
area:api priority:p1 release:required-for-ga spec:trace
**What are you trying to achieve?** Revert #1075 after it was badly received in the actual implementations. Overall, the reasons behind going back to `SpanContext` are: * `SpanContext` is an already *very* well known term. * `SpanReference` may make people think of pointers/references. * `SpanReference` includes additional information to `Span`s, such as `tracestate`. For this change, we will actually expect more than the usual reviews, in order to verify with SIG maintainers/approvers this is a good thing.
1.0
Rename SpanReference to SpanContext (revert #1075) - **What are you trying to achieve?** Revert #1075 after it was badly received in the actual implementations. Overall, the reasons behind going back to `SpanContext` are: * `SpanContext` is an already *very* well known term. * `SpanReference` may make people think of pointers/references. * `SpanReference` includes additional information to `Span`s, such as `tracestate`. For this change, we will actually expect more than the usual reviews, in order to verify with SIG maintainers/approvers this is a good thing.
non_test
rename spanreference to spancontext revert what are you trying to achieve revert after it was badly received in the actual implementations overall the reasons behind going back to spancontext are spancontext is an already very well known term spanreference may make people think of pointers references spanreference includes additional information to span s such as tracestate for this change we will actually expect more than the usual reviews in order to verify with sig maintainers approvers this is a good thing
0
165,715
12,879,868,177
IssuesEvent
2020-07-12 01:25:30
osquery/osquery
https://api.github.com/repos/osquery/osquery
closed
Create in-memory numeric monitoring plugin
feature table test
## In-memory numeric monitoring plugin Create simple in-memory numeric monitoring plugin to test that certain keys are being bumped in the integration tests. Filesystem plugin for example: `osquery/numeric_monitoring/plugins/filesystem.h` Need to implement another plugin with global access interface to the records by key.
1.0
Create in-memory numeric monitoring plugin - ## In-memory numeric monitoring plugin Create simple in-memory numeric monitoring plugin to test that certain keys are being bumped in the integration tests. Filesystem plugin for example: `osquery/numeric_monitoring/plugins/filesystem.h` Need to implement another plugin with global access interface to the records by key.
test
create in memory numeric monitoring plugin in memory numeric monitoring plugin create simple in memory numeric monitoring plugin to test that certain keys are being bumped in the integration tests filesystem plugin for example osquery numeric monitoring plugins filesystem h need to implement another plugin with global access interface to the records by key
1
608,268
18,820,331,628
IssuesEvent
2021-11-10 07:25:47
redhat-developer/vscode-openshift-tools
https://api.github.com/repos/redhat-developer/vscode-openshift-tools
closed
Context folders are not visible in OpenShift Application Explorer if they are form not existing project or/and application
resolution/wontfix priority/major kind/bug
If odo context folder added to workspace and project used is not available in current cluster, that component is not visible in OpenShift Application Explorer.
1.0
Context folders are not visible in OpenShift Application Explorer if they are form not existing project or/and application - If odo context folder added to workspace and project used is not available in current cluster, that component is not visible in OpenShift Application Explorer.
non_test
context folders are not visible in openshift application explorer if they are form not existing project or and application if odo context folder added to workspace and project used is not available in current cluster that component is not visible in openshift application explorer
0
32,509
4,774,285,552
IssuesEvent
2016-10-27 05:52:45
linkedpipes/etl
https://api.github.com/repos/linkedpipes/etl
closed
2 template + instance inheritance broken
bug test
I will demonstrate on one property - file name Settings: 1. Root template (non-editable): file.csv (I guess - cant see) 2. T2: fileL2.csv 3. T3: fileL3.csv 4. Instance: instancefile.csv 1. When I set T3 to overwrite, I get fileL3.csv, which is OK. 2. When I then set T2 to overwrite, I get fileL2.csv, which is also OK. 3. When I then set T2 to inherit, I would expect to get file.csv, however, I get fileL3.csv
1.0
2 template + instance inheritance broken - I will demonstrate on one property - file name Settings: 1. Root template (non-editable): file.csv (I guess - cant see) 2. T2: fileL2.csv 3. T3: fileL3.csv 4. Instance: instancefile.csv 1. When I set T3 to overwrite, I get fileL3.csv, which is OK. 2. When I then set T2 to overwrite, I get fileL2.csv, which is also OK. 3. When I then set T2 to inherit, I would expect to get file.csv, however, I get fileL3.csv
test
template instance inheritance broken i will demonstrate on one property file name settings root template non editable file csv i guess cant see csv csv instance instancefile csv when i set to overwrite i get csv which is ok when i then set to overwrite i get csv which is also ok when i then set to inherit i would expect to get file csv however i get csv
1
53,579
28,289,443,693
IssuesEvent
2023-04-09 02:19:34
yamoo9/likelion-FEQA
https://api.github.com/repos/yamoo9/likelion-FEQA
closed
[LAB-5] card객체의 변화에 따른 리랜더링 문제 발생
React 프로젝트 Q/A Performance
## 질문 작성자 김서현 ## 문제 상황 ![2023-04-09 01;29;39](https://user-images.githubusercontent.com/38703262/230732345-4ff1831b-26d5-4699-9159-d18003860bb6.gif) 제목을 검색한 뒤, 카드를 클릭했을때 그 해당 카드의 정보가 랜더링되게 최대한 구현해보려 했고, 구현은 하였으나 랜더링 문제가 생겨 글을 남깁니다. 프로파일러 돌려보았을때 너무 랜더링이 많이 일어나고, (232회 발생) 콘솔창을 보면 계속 찍혀 나오는 것을 확인할 수 있었습니다.. ![image](https://user-images.githubusercontent.com/38703262/230732519-f3792199-963c-452b-ae1a-72f5b9a19a68.png) card는 현재 recoil의 atom형태로 상태관리 하고 있습니다. 문제의 원인이 card때문인듯 한데, card는 바뀐것이 없는데 왜 계속 리랜더링 되는 상황이 만들어지는지 모르겠습니다 ㅠ 프로파일러를 확인해보면, ReadMeeting파일에 있는 card가 계속 리랜더링 되는것 같습니다.. ## 프로젝트 저장소 URL https://github.com/React-Project-lab5/React-Project-lab5 develop브랜치 - 현재 파일 위치 ![image](https://user-images.githubusercontent.com/38703262/230732660-faf2ad39-c53a-4f2e-ba80-d36336144136.png) ![image](https://user-images.githubusercontent.com/38703262/230732674-73715809-6e06-4dbc-aaf8-48bc68aaa1bc.png) - src파일 : [src.zip](https://github.com/yamoo9/likelion-FEQA/files/11183692/src.zip) - 로그인은 구글로그인 혹은 아이디 : babjo@naver.com / 비번: aaaa1111 입니다. - .env파일에 작성하시면 되는 코드를 아래에 적어드립니다. VITE_API_KEY = AIzaSyB-16F5fhuvgb5lKC6618iuSZPvEOwWhLk VITE_AUTH_DOMAIN = calkolab5.firebaseapp.com VITE_PROJECT_ID = calkolab5 VITE_STORAGE_BUCKET = calkolab5.appspot.com VITE_MESSAGE_SENDER_ID = 869641050054 VITE_APP_ID = 1:869641050054:web:0d0b56dac5093949fdc4c7 VITE_MEASUREMENT_ID = G-LCH8VZEKV3 VITE_SERVICE_KEY=M4E1Vdrm%2Ba%2FjVfatJPdEbBibLbYBeDaGm7tgF%2BxisSinPYTEEBDkFgSO0gmqjAZGOzfrSffYbsOF3XshtdL6lg%3D%3D VITE_KAKAO_API_KEY=ebf9334d7db030f4f026a0d53f08a62a ## 환경 정보 - 운영체제 : Windows - Node.js : v16.14.0 - 라이브러리 : React v18.2.0
True
[LAB-5] card객체의 변화에 따른 리랜더링 문제 발생 - ## 질문 작성자 김서현 ## 문제 상황 ![2023-04-09 01;29;39](https://user-images.githubusercontent.com/38703262/230732345-4ff1831b-26d5-4699-9159-d18003860bb6.gif) 제목을 검색한 뒤, 카드를 클릭했을때 그 해당 카드의 정보가 랜더링되게 최대한 구현해보려 했고, 구현은 하였으나 랜더링 문제가 생겨 글을 남깁니다. 프로파일러 돌려보았을때 너무 랜더링이 많이 일어나고, (232회 발생) 콘솔창을 보면 계속 찍혀 나오는 것을 확인할 수 있었습니다.. ![image](https://user-images.githubusercontent.com/38703262/230732519-f3792199-963c-452b-ae1a-72f5b9a19a68.png) card는 현재 recoil의 atom형태로 상태관리 하고 있습니다. 문제의 원인이 card때문인듯 한데, card는 바뀐것이 없는데 왜 계속 리랜더링 되는 상황이 만들어지는지 모르겠습니다 ㅠ 프로파일러를 확인해보면, ReadMeeting파일에 있는 card가 계속 리랜더링 되는것 같습니다.. ## 프로젝트 저장소 URL https://github.com/React-Project-lab5/React-Project-lab5 develop브랜치 - 현재 파일 위치 ![image](https://user-images.githubusercontent.com/38703262/230732660-faf2ad39-c53a-4f2e-ba80-d36336144136.png) ![image](https://user-images.githubusercontent.com/38703262/230732674-73715809-6e06-4dbc-aaf8-48bc68aaa1bc.png) - src파일 : [src.zip](https://github.com/yamoo9/likelion-FEQA/files/11183692/src.zip) - 로그인은 구글로그인 혹은 아이디 : babjo@naver.com / 비번: aaaa1111 입니다. - .env파일에 작성하시면 되는 코드를 아래에 적어드립니다. VITE_API_KEY = AIzaSyB-16F5fhuvgb5lKC6618iuSZPvEOwWhLk VITE_AUTH_DOMAIN = calkolab5.firebaseapp.com VITE_PROJECT_ID = calkolab5 VITE_STORAGE_BUCKET = calkolab5.appspot.com VITE_MESSAGE_SENDER_ID = 869641050054 VITE_APP_ID = 1:869641050054:web:0d0b56dac5093949fdc4c7 VITE_MEASUREMENT_ID = G-LCH8VZEKV3 VITE_SERVICE_KEY=M4E1Vdrm%2Ba%2FjVfatJPdEbBibLbYBeDaGm7tgF%2BxisSinPYTEEBDkFgSO0gmqjAZGOzfrSffYbsOF3XshtdL6lg%3D%3D VITE_KAKAO_API_KEY=ebf9334d7db030f4f026a0d53f08a62a ## 환경 정보 - 운영체제 : Windows - Node.js : v16.14.0 - 라이브러리 : React v18.2.0
non_test
card객체의 변화에 따른 리랜더링 문제 발생 질문 작성자 김서현 문제 상황 제목을 검색한 뒤 카드를 클릭했을때 그 해당 카드의 정보가 랜더링되게 최대한 구현해보려 했고 구현은 하였으나 랜더링 문제가 생겨 글을 남깁니다 프로파일러 돌려보았을때 너무 랜더링이 많이 일어나고 발생 콘솔창을 보면 계속 찍혀 나오는 것을 확인할 수 있었습니다 card는 현재 recoil의 atom형태로 상태관리 하고 있습니다 문제의 원인이 card때문인듯 한데 card는 바뀐것이 없는데 왜 계속 리랜더링 되는 상황이 만들어지는지 모르겠습니다 ㅠ 프로파일러를 확인해보면 readmeeting파일에 있는 card가 계속 리랜더링 되는것 같습니다 프로젝트 저장소 url develop브랜치 현재 파일 위치 src파일 로그인은 구글로그인 혹은 아이디 babjo naver com 비번 입니다 env파일에 작성하시면 되는 코드를 아래에 적어드립니다 vite api key aizasyb vite auth domain firebaseapp com vite project id vite storage bucket appspot com vite message sender id vite app id web vite measurement id g vite service key vite kakao api key 환경 정보 운영체제 windows node js 라이브러리 react
0
465,739
13,391,249,549
IssuesEvent
2020-09-02 22:07:57
microsoft/PowerToys
https://api.github.com/repos/microsoft/PowerToys
closed
[PT RUN] PowerLauncher.exe crashes on launch due to unhandled exception from Wox
Issue-Bug Priority-0 Product-Launcher Resolution-Fix-Committed
## ℹ Computer information - Windows build number: 19041.388 - PowerToys version: 0.20.1 - PowerToy module: PowerToys Run ## 📝 Provide detailed reproduction steps (if any) 1. Start PowerToys or manually launch PowerLauncher.exe ### ✔️ Expected result PowerLauncher.exe runs in the background. Pressing the keyboard shortcut for PowerToys Run brings up the search bar. ### ❌ Actual result PowerLauncher.exe appears in task manager briefly before disappearing. Pressing the keyboard shortcut for PowerToys Run does nothing. ## Additional Details The following unhandled exception is logged in Event Viewer: Application: PowerLauncher.exe CoreCLR Version: 4.700.20.20201 .NET Core Version: 3.1.4 Description: The process was terminated due to an unhandled exception. Exception Info: System.TypeLoadException: Method 'PushResults' in type 'Wox.PublicAPIInstance' from assembly 'PowerLauncher, Version=0.20.1.0, Culture=neutral, PublicKeyToken=null' does not have an implementation. at PowerLauncher.App.<>c__DisplayClass16_0.<OnStartup>b__1() at Wox.Infrastructure.Stopwatch.Normal(String message, Action action) at PowerLauncher.App.OnStartup(Object sender, StartupEventArgs e) at System.Windows.Application.OnStartup(StartupEventArgs e) at System.Windows.Application.<.ctor>b__1_0(Object unused) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(Delegate callback, Object args, Int32 numArgs) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(Object source, Delegate callback, Object args, Int32 numArgs, Delegate catchHandler) at System.Windows.Threading.DispatcherOperation.InvokeImpl() at System.Windows.Threading.DispatcherOperation.InvokeInSecurityContext(Object state) at MS.Internal.CulturePreservingExecutionContext.CallbackWrapper(Object obj) at System.Threading.ExecutionContext.RunInternal(ExecutionContext executionContext, ContextCallback callback, Object state) --- End of stack trace from previous location where exception was thrown --- at System.Threading.ExecutionContext.RunInternal(ExecutionContext executionContext, ContextCallback callback, Object state) at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state) at MS.Internal.CulturePreservingExecutionContext.Run(CulturePreservingExecutionContext executionContext, ContextCallback callback, Object state) at System.Windows.Threading.DispatcherOperation.Invoke() at System.Windows.Threading.Dispatcher.ProcessQueue() at System.Windows.Threading.Dispatcher.WndProcHook(IntPtr hwnd, Int32 msg, IntPtr wParam, IntPtr lParam, Boolean& handled) at MS.Win32.HwndWrapper.WndProc(IntPtr hwnd, Int32 msg, IntPtr wParam, IntPtr lParam, Boolean& handled) at MS.Win32.HwndSubclass.DispatcherCallbackOperation(Object o) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(Delegate callback, Object args, Int32 numArgs) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(Object source, Delegate callback, Object args, Int32 numArgs, Delegate catchHandler) at System.Windows.Threading.Dispatcher.LegacyInvokeImpl(DispatcherPriority priority, TimeSpan timeout, Delegate method, Object args, Int32 numArgs) at MS.Win32.HwndSubclass.SubclassWndProc(IntPtr hwnd, Int32 msg, IntPtr wParam, IntPtr lParam) at MS.Win32.UnsafeNativeMethods.DispatchMessage(MSG& msg) at System.Windows.Threading.Dispatcher.PushFrameImpl(DispatcherFrame frame) at System.Windows.Threading.Dispatcher.PushFrame(DispatcherFrame frame) at System.Windows.Threading.Dispatcher.Run() at System.Windows.Application.RunDispatcher(Object ignore) at System.Windows.Application.RunInternal(Window window) at System.Windows.Application.Run() at PowerLauncher.App.Main(String[] args)
1.0
[PT RUN] PowerLauncher.exe crashes on launch due to unhandled exception from Wox - ## ℹ Computer information - Windows build number: 19041.388 - PowerToys version: 0.20.1 - PowerToy module: PowerToys Run ## 📝 Provide detailed reproduction steps (if any) 1. Start PowerToys or manually launch PowerLauncher.exe ### ✔️ Expected result PowerLauncher.exe runs in the background. Pressing the keyboard shortcut for PowerToys Run brings up the search bar. ### ❌ Actual result PowerLauncher.exe appears in task manager briefly before disappearing. Pressing the keyboard shortcut for PowerToys Run does nothing. ## Additional Details The following unhandled exception is logged in Event Viewer: Application: PowerLauncher.exe CoreCLR Version: 4.700.20.20201 .NET Core Version: 3.1.4 Description: The process was terminated due to an unhandled exception. Exception Info: System.TypeLoadException: Method 'PushResults' in type 'Wox.PublicAPIInstance' from assembly 'PowerLauncher, Version=0.20.1.0, Culture=neutral, PublicKeyToken=null' does not have an implementation. at PowerLauncher.App.<>c__DisplayClass16_0.<OnStartup>b__1() at Wox.Infrastructure.Stopwatch.Normal(String message, Action action) at PowerLauncher.App.OnStartup(Object sender, StartupEventArgs e) at System.Windows.Application.OnStartup(StartupEventArgs e) at System.Windows.Application.<.ctor>b__1_0(Object unused) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(Delegate callback, Object args, Int32 numArgs) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(Object source, Delegate callback, Object args, Int32 numArgs, Delegate catchHandler) at System.Windows.Threading.DispatcherOperation.InvokeImpl() at System.Windows.Threading.DispatcherOperation.InvokeInSecurityContext(Object state) at MS.Internal.CulturePreservingExecutionContext.CallbackWrapper(Object obj) at System.Threading.ExecutionContext.RunInternal(ExecutionContext executionContext, ContextCallback callback, Object state) --- End of stack trace from previous location where exception was thrown --- at System.Threading.ExecutionContext.RunInternal(ExecutionContext executionContext, ContextCallback callback, Object state) at System.Threading.ExecutionContext.Run(ExecutionContext executionContext, ContextCallback callback, Object state) at MS.Internal.CulturePreservingExecutionContext.Run(CulturePreservingExecutionContext executionContext, ContextCallback callback, Object state) at System.Windows.Threading.DispatcherOperation.Invoke() at System.Windows.Threading.Dispatcher.ProcessQueue() at System.Windows.Threading.Dispatcher.WndProcHook(IntPtr hwnd, Int32 msg, IntPtr wParam, IntPtr lParam, Boolean& handled) at MS.Win32.HwndWrapper.WndProc(IntPtr hwnd, Int32 msg, IntPtr wParam, IntPtr lParam, Boolean& handled) at MS.Win32.HwndSubclass.DispatcherCallbackOperation(Object o) at System.Windows.Threading.ExceptionWrapper.InternalRealCall(Delegate callback, Object args, Int32 numArgs) at System.Windows.Threading.ExceptionWrapper.TryCatchWhen(Object source, Delegate callback, Object args, Int32 numArgs, Delegate catchHandler) at System.Windows.Threading.Dispatcher.LegacyInvokeImpl(DispatcherPriority priority, TimeSpan timeout, Delegate method, Object args, Int32 numArgs) at MS.Win32.HwndSubclass.SubclassWndProc(IntPtr hwnd, Int32 msg, IntPtr wParam, IntPtr lParam) at MS.Win32.UnsafeNativeMethods.DispatchMessage(MSG& msg) at System.Windows.Threading.Dispatcher.PushFrameImpl(DispatcherFrame frame) at System.Windows.Threading.Dispatcher.PushFrame(DispatcherFrame frame) at System.Windows.Threading.Dispatcher.Run() at System.Windows.Application.RunDispatcher(Object ignore) at System.Windows.Application.RunInternal(Window window) at System.Windows.Application.Run() at PowerLauncher.App.Main(String[] args)
non_test
powerlauncher exe crashes on launch due to unhandled exception from wox ℹ computer information windows build number powertoys version powertoy module powertoys run 📝 provide detailed reproduction steps if any start powertoys or manually launch powerlauncher exe ✔️ expected result powerlauncher exe runs in the background pressing the keyboard shortcut for powertoys run brings up the search bar ❌ actual result powerlauncher exe appears in task manager briefly before disappearing pressing the keyboard shortcut for powertoys run does nothing additional details the following unhandled exception is logged in event viewer application powerlauncher exe coreclr version net core version description the process was terminated due to an unhandled exception exception info system typeloadexception method pushresults in type wox publicapiinstance from assembly powerlauncher version culture neutral publickeytoken null does not have an implementation at powerlauncher app c b at wox infrastructure stopwatch normal string message action action at powerlauncher app onstartup object sender startupeventargs e at system windows application onstartup startupeventargs e at system windows application b object unused at system windows threading exceptionwrapper internalrealcall delegate callback object args numargs at system windows threading exceptionwrapper trycatchwhen object source delegate callback object args numargs delegate catchhandler at system windows threading dispatcheroperation invokeimpl at system windows threading dispatcheroperation invokeinsecuritycontext object state at ms internal culturepreservingexecutioncontext callbackwrapper object obj at system threading executioncontext runinternal executioncontext executioncontext contextcallback callback object state end of stack trace from previous location where exception was thrown at system threading executioncontext runinternal executioncontext executioncontext contextcallback callback object state at system threading executioncontext run executioncontext executioncontext contextcallback callback object state at ms internal culturepreservingexecutioncontext run culturepreservingexecutioncontext executioncontext contextcallback callback object state at system windows threading dispatcheroperation invoke at system windows threading dispatcher processqueue at system windows threading dispatcher wndprochook intptr hwnd msg intptr wparam intptr lparam boolean handled at ms hwndwrapper wndproc intptr hwnd msg intptr wparam intptr lparam boolean handled at ms hwndsubclass dispatchercallbackoperation object o at system windows threading exceptionwrapper internalrealcall delegate callback object args numargs at system windows threading exceptionwrapper trycatchwhen object source delegate callback object args numargs delegate catchhandler at system windows threading dispatcher legacyinvokeimpl dispatcherpriority priority timespan timeout delegate method object args numargs at ms hwndsubclass subclasswndproc intptr hwnd msg intptr wparam intptr lparam at ms unsafenativemethods dispatchmessage msg msg at system windows threading dispatcher pushframeimpl dispatcherframe frame at system windows threading dispatcher pushframe dispatcherframe frame at system windows threading dispatcher run at system windows application rundispatcher object ignore at system windows application runinternal window window at system windows application run at powerlauncher app main string args
0
97,590
8,663,163,903
IssuesEvent
2018-11-28 16:40:11
MachoThemes/modula-lite
https://api.github.com/repos/MachoThemes/modula-lite
closed
See LITE vs PRO difference upsell button doesn't link anywhere
bug tested
<img width="1192" alt="image" src="https://user-images.githubusercontent.com/7841556/46799959-dd67fe00-cd5e-11e8-8f58-5c4bd89bcbac.png">
1.0
See LITE vs PRO difference upsell button doesn't link anywhere - <img width="1192" alt="image" src="https://user-images.githubusercontent.com/7841556/46799959-dd67fe00-cd5e-11e8-8f58-5c4bd89bcbac.png">
test
see lite vs pro difference upsell button doesn t link anywhere img width alt image src
1
102,354
8,825,095,547
IssuesEvent
2019-01-02 19:32:04
medic/medic-webapp
https://api.github.com/repos/medic/medic-webapp
closed
Add a pregnancy visit to pregnant woman on standard config as restricted user on mobile app
Release Test
**User**: Restricted User **Site**: Standard config site **Platform**: Mobile App **Test Steps**: Using a pregnant woman Process a pregnancy visit form. **Expected Result**: The pregnancy condition card should update The report should be associated to the woman. The targets should be updated.
1.0
Add a pregnancy visit to pregnant woman on standard config as restricted user on mobile app - **User**: Restricted User **Site**: Standard config site **Platform**: Mobile App **Test Steps**: Using a pregnant woman Process a pregnancy visit form. **Expected Result**: The pregnancy condition card should update The report should be associated to the woman. The targets should be updated.
test
add a pregnancy visit to pregnant woman on standard config as restricted user on mobile app user restricted user site standard config site platform mobile app test steps using a pregnant woman process a pregnancy visit form expected result the pregnancy condition card should update the report should be associated to the woman the targets should be updated
1
89,679
8,212,038,819
IssuesEvent
2018-09-04 15:16:01
nasa-gibs/worldview
https://api.github.com/repos/nasa-gibs/worldview
closed
Build on RHEL7
technical testing
Ensure that the RPM builds and deploys properly on RHEL7. Fix any issues with the spec file if not.
1.0
Build on RHEL7 - Ensure that the RPM builds and deploys properly on RHEL7. Fix any issues with the spec file if not.
test
build on ensure that the rpm builds and deploys properly on fix any issues with the spec file if not
1
292,595
25,224,670,717
IssuesEvent
2022-11-14 15:11:08
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: acceptance/status-server failed
C-test-failure O-robot O-roachtest release-blocker T-observability-inf A-observability-inf branch-release-22.2
roachtest.acceptance/status-server [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7466987?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7466987?buildTab=artifacts#/acceptance/status-server) on release-22.2 @ [9ec2ddb8dc118ce91bf806612bc4c0f86dd7b32b](https://github.com/cockroachdb/cockroach/commits/9ec2ddb8dc118ce91bf806612bc4c0f86dd7b32b): ``` test artifacts and logs in: /artifacts/acceptance/status-server/run_1 cluster.go:1904,status_server.go:32,acceptance.go:111,test_runner.go:930: one or more parallel execution failure (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2286 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2167 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start | github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:155 | github.com/cockroachdb/cockroach/pkg/roachprod.Start | github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:662 | main.(*clusterImpl).StartE | main/pkg/cmd/roachtest/cluster.go:1858 | main.(*clusterImpl).Start | main/pkg/cmd/roachtest/cluster.go:1903 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runStatusServer | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/status_server.go:32 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerAcceptance.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/acceptance.go:111 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:930 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1594 Wraps: (2) one or more parallel execution failure Error types: (1) *withstack.withStack (2) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/obs-inf-prs <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*acceptance/status-server.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-21428
2.0
roachtest: acceptance/status-server failed - roachtest.acceptance/status-server [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7466987?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/7466987?buildTab=artifacts#/acceptance/status-server) on release-22.2 @ [9ec2ddb8dc118ce91bf806612bc4c0f86dd7b32b](https://github.com/cockroachdb/cockroach/commits/9ec2ddb8dc118ce91bf806612bc4c0f86dd7b32b): ``` test artifacts and logs in: /artifacts/acceptance/status-server/run_1 cluster.go:1904,status_server.go:32,acceptance.go:111,test_runner.go:930: one or more parallel execution failure (1) attached stack trace -- stack trace: | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).ParallelE | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2286 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Parallel | github.com/cockroachdb/cockroach/pkg/roachprod/install/cluster_synced.go:2167 | github.com/cockroachdb/cockroach/pkg/roachprod/install.(*SyncedCluster).Start | github.com/cockroachdb/cockroach/pkg/roachprod/install/cockroach.go:155 | github.com/cockroachdb/cockroach/pkg/roachprod.Start | github.com/cockroachdb/cockroach/pkg/roachprod/roachprod.go:662 | main.(*clusterImpl).StartE | main/pkg/cmd/roachtest/cluster.go:1858 | main.(*clusterImpl).Start | main/pkg/cmd/roachtest/cluster.go:1903 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.runStatusServer | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/status_server.go:32 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerAcceptance.func1 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/acceptance.go:111 | main.(*testRunner).runTest.func2 | main/pkg/cmd/roachtest/test_runner.go:930 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1594 Wraps: (2) one or more parallel execution failure Error types: (1) *withstack.withStack (2) *errutil.leafError ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_encrypted=false</code> , <code>ROACHTEST_fs=ext4</code> , <code>ROACHTEST_localSSD=true</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/obs-inf-prs <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*acceptance/status-server.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub> Jira issue: CRDB-21428
test
roachtest acceptance status server failed roachtest acceptance status server with on release test artifacts and logs in artifacts acceptance status server run cluster go status server go acceptance go test runner go one or more parallel execution failure attached stack trace stack trace github com cockroachdb cockroach pkg roachprod install syncedcluster parallele github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster parallel github com cockroachdb cockroach pkg roachprod install cluster synced go github com cockroachdb cockroach pkg roachprod install syncedcluster start github com cockroachdb cockroach pkg roachprod install cockroach go github com cockroachdb cockroach pkg roachprod start github com cockroachdb cockroach pkg roachprod roachprod go main clusterimpl starte main pkg cmd roachtest cluster go main clusterimpl start main pkg cmd roachtest cluster go github com cockroachdb cockroach pkg cmd roachtest tests runstatusserver github com cockroachdb cockroach pkg cmd roachtest tests status server go github com cockroachdb cockroach pkg cmd roachtest tests registeracceptance github com cockroachdb cockroach pkg cmd roachtest tests acceptance go main testrunner runtest main pkg cmd roachtest test runner go runtime goexit goroot src runtime asm s wraps one or more parallel execution failure error types withstack withstack errutil leaferror parameters roachtest cloud gce roachtest cpu roachtest encrypted false roachtest fs roachtest localssd true roachtest ssd help see see cc cockroachdb obs inf prs jira issue crdb
1
182,896
21,678,028,326
IssuesEvent
2022-05-09 01:09:21
prafullkotecha/azure-sdk-for-java
https://api.github.com/repos/prafullkotecha/azure-sdk-for-java
opened
CVE-2022-24823 (Medium) detected in netty-all-4.0.56.Final.jar
security vulnerability
## CVE-2022-24823 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-all-4.0.56.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p> <p>Path to vulnerable library: /sdk/spring/azure-spring-data-gremlin/package/apache-tinkerpop-gremlin-server-minimal-3.3.4.tar/apache-tinkerpop-gremlin-server-minimal-3.3.4/lib/netty-all-4.0.56.Final.jar</p> <p> Dependency Hierarchy: - :x: **netty-all-4.0.56.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/prafullkotecha/azure-sdk-for-java/commit/d00274c5406c57214dbd8d18882a072c71e83d85">d00274c5406c57214dbd8d18882a072c71e83d85</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user. <p>Publish Date: 2022-05-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823>CVE-2022-24823</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p> <p>Release Date: 2022-05-06</p> <p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-24823 (Medium) detected in netty-all-4.0.56.Final.jar - ## CVE-2022-24823 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>netty-all-4.0.56.Final.jar</b></p></summary> <p>Netty is an asynchronous event-driven network application framework for rapid development of maintainable high performance protocol servers and clients.</p> <p>Library home page: <a href="http://netty.io/">http://netty.io/</a></p> <p>Path to vulnerable library: /sdk/spring/azure-spring-data-gremlin/package/apache-tinkerpop-gremlin-server-minimal-3.3.4.tar/apache-tinkerpop-gremlin-server-minimal-3.3.4/lib/netty-all-4.0.56.Final.jar</p> <p> Dependency Hierarchy: - :x: **netty-all-4.0.56.Final.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/prafullkotecha/azure-sdk-for-java/commit/d00274c5406c57214dbd8d18882a072c71e83d85">d00274c5406c57214dbd8d18882a072c71e83d85</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Netty is an open-source, asynchronous event-driven network application framework. The package `io.netty:netty-codec-http` prior to version 4.1.77.Final contains an insufficient fix for CVE-2021-21290. When Netty's multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled. This only impacts applications running on Java version 6 and lower. Additionally, this vulnerability impacts code running on Unix-like systems, and very old versions of Mac OSX and Windows as they all share the system temporary directory between all users. Version 4.1.77.Final contains a patch for this vulnerability. As a workaround, specify one's own `java.io.tmpdir` when starting the JVM or use DefaultHttpDataFactory.setBaseDir(...) to set the directory to something that is only readable by the current user. <p>Publish Date: 2022-05-06 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-24823>CVE-2022-24823</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: None - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-24823</a></p> <p>Release Date: 2022-05-06</p> <p>Fix Resolution: io.netty:netty-all;io.netty:netty-common - 4.1.77.Final</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in netty all final jar cve medium severity vulnerability vulnerable library netty all final jar netty is an asynchronous event driven network application framework for rapid development of maintainable high performance protocol servers and clients library home page a href path to vulnerable library sdk spring azure spring data gremlin package apache tinkerpop gremlin server minimal tar apache tinkerpop gremlin server minimal lib netty all final jar dependency hierarchy x netty all final jar vulnerable library found in head commit a href found in base branch master vulnerability details netty is an open source asynchronous event driven network application framework the package io netty netty codec http prior to version final contains an insufficient fix for cve when netty s multipart decoders are used local information disclosure can occur via the local system temporary directory if temporary storing uploads on the disk is enabled this only impacts applications running on java version and lower additionally this vulnerability impacts code running on unix like systems and very old versions of mac osx and windows as they all share the system temporary directory between all users version final contains a patch for this vulnerability as a workaround specify one s own java io tmpdir when starting the jvm or use defaulthttpdatafactory setbasedir to set the directory to something that is only readable by the current user publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact none availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution io netty netty all io netty netty common final step up your open source security game with whitesource
0
132,449
12,508,369,821
IssuesEvent
2020-06-02 15:27:28
Prometee/SyliusPayumStripeCheckoutSessionPlugin
https://api.github.com/repos/Prometee/SyliusPayumStripeCheckoutSessionPlugin
closed
Gateway "stripe_checkout_session" does not exist
documentation good first issue question
Hello @Prometee you have done a great job, thanks 👍 Payment with Stripe works well, in dev mode. But i am facing an issue on incoming webhooks : `500 - Gateway "stripe_checkout_session" does not exist` My payment method followed the documentation provided with the lib : - **code** _stripe_session_checkout_with_sca_ - **gateway** _stripe_checkout_session_ I am confused with the gateway name `stripe_checkout_session`, in database (table `sylius_gateway_config`) it sounds different : - **gateway_name** _stripe_session_checkout_with_sca_ - **factory** _stripe_checkout_session_ I would appreciate some help 😉 ### versions sylius/sylius **v1.7.0** prometee/payum-stripe **v1.1.1** prometee/payum-stripe-checkout-session-bundle **v1.1.0** prometee/sylius-payum-stripe-checkout-session-plugin **v1.1.2**
1.0
Gateway "stripe_checkout_session" does not exist - Hello @Prometee you have done a great job, thanks 👍 Payment with Stripe works well, in dev mode. But i am facing an issue on incoming webhooks : `500 - Gateway "stripe_checkout_session" does not exist` My payment method followed the documentation provided with the lib : - **code** _stripe_session_checkout_with_sca_ - **gateway** _stripe_checkout_session_ I am confused with the gateway name `stripe_checkout_session`, in database (table `sylius_gateway_config`) it sounds different : - **gateway_name** _stripe_session_checkout_with_sca_ - **factory** _stripe_checkout_session_ I would appreciate some help 😉 ### versions sylius/sylius **v1.7.0** prometee/payum-stripe **v1.1.1** prometee/payum-stripe-checkout-session-bundle **v1.1.0** prometee/sylius-payum-stripe-checkout-session-plugin **v1.1.2**
non_test
gateway stripe checkout session does not exist hello prometee you have done a great job thanks 👍 payment with stripe works well in dev mode but i am facing an issue on incoming webhooks gateway stripe checkout session does not exist my payment method followed the documentation provided with the lib code stripe session checkout with sca gateway stripe checkout session i am confused with the gateway name stripe checkout session in database table sylius gateway config it sounds different gateway name stripe session checkout with sca factory stripe checkout session i would appreciate some help 😉 versions sylius sylius prometee payum stripe prometee payum stripe checkout session bundle prometee sylius payum stripe checkout session plugin
0
288,171
24,884,547,905
IssuesEvent
2022-10-28 06:23:43
ValveSoftware/Source-1-Games
https://api.github.com/repos/ValveSoftware/Source-1-Games
closed
[TF2] Re-entering Halloween Karts forces players into an incorrect view angle
Team Fortress 2 Need Retest
Right now, you can give players the halloween kart with condition 82 `(TF_COND_HALLOWEEN_KART)` and control it just fine. The condition overrides your pitch and roll angles to provide a locked over-the-shoulder angle in third person, which is intended. However, your yaw view angle can be set to unpredictable and unintended values. From testing, It seems karts have a variable in memory for last stored yaw angle of the kart. By default this value is 0, snapping the player into yaw 0 when they first gain condition 82. This is viewable when using the `cl_showpos 1` command in console. Every time after this though, this default yaw value will be set to whatever angle the player last had when _losing_ condition 82. For example, if you give yourself condition 82 (`addcond 82` in console), you'll be set to yaw 0. Turn 90 degrees, and remove condition 82 (`removecond 82` in console.) You'll lose the bumper kart, but preserve your view angles. Now, turn until facing, say, 270, and give yourself condition 82 again. You'll see that your view angle has been snapped back to 90 degrees, completely irrelevant of what angle you had as a player. You also cannot use the `setang` command to fix this, since condition 82 actively overwrites your view angle at all times. This leads to having extremely little control over what direction a player faces when repeatedly granting them a bumper kart for things such as race tracks, minigames, or otherwise. The ideal function of this condition should still overwrite your pitch and roll angles, but the yaw angle should be whatever yaw angle the player itself had before gaining the condition, rather than whatever the last stored yaw of the kart itself was. I recorded a video here of how to reproduce, and what the ideal fix should be- [![Re-entering Halloween Karts forces players into an incorrect view angle](https://i.imgur.com/UUDBi4X.png)](https://www.youtube.com/watch?v=2ysZRb38KYI "[TF2] Re-entering Halloween Karts forces players into an incorrect view angle")
1.0
[TF2] Re-entering Halloween Karts forces players into an incorrect view angle - Right now, you can give players the halloween kart with condition 82 `(TF_COND_HALLOWEEN_KART)` and control it just fine. The condition overrides your pitch and roll angles to provide a locked over-the-shoulder angle in third person, which is intended. However, your yaw view angle can be set to unpredictable and unintended values. From testing, It seems karts have a variable in memory for last stored yaw angle of the kart. By default this value is 0, snapping the player into yaw 0 when they first gain condition 82. This is viewable when using the `cl_showpos 1` command in console. Every time after this though, this default yaw value will be set to whatever angle the player last had when _losing_ condition 82. For example, if you give yourself condition 82 (`addcond 82` in console), you'll be set to yaw 0. Turn 90 degrees, and remove condition 82 (`removecond 82` in console.) You'll lose the bumper kart, but preserve your view angles. Now, turn until facing, say, 270, and give yourself condition 82 again. You'll see that your view angle has been snapped back to 90 degrees, completely irrelevant of what angle you had as a player. You also cannot use the `setang` command to fix this, since condition 82 actively overwrites your view angle at all times. This leads to having extremely little control over what direction a player faces when repeatedly granting them a bumper kart for things such as race tracks, minigames, or otherwise. The ideal function of this condition should still overwrite your pitch and roll angles, but the yaw angle should be whatever yaw angle the player itself had before gaining the condition, rather than whatever the last stored yaw of the kart itself was. I recorded a video here of how to reproduce, and what the ideal fix should be- [![Re-entering Halloween Karts forces players into an incorrect view angle](https://i.imgur.com/UUDBi4X.png)](https://www.youtube.com/watch?v=2ysZRb38KYI "[TF2] Re-entering Halloween Karts forces players into an incorrect view angle")
test
re entering halloween karts forces players into an incorrect view angle right now you can give players the halloween kart with condition tf cond halloween kart and control it just fine the condition overrides your pitch and roll angles to provide a locked over the shoulder angle in third person which is intended however your yaw view angle can be set to unpredictable and unintended values from testing it seems karts have a variable in memory for last stored yaw angle of the kart by default this value is snapping the player into yaw when they first gain condition this is viewable when using the cl showpos command in console every time after this though this default yaw value will be set to whatever angle the player last had when losing condition for example if you give yourself condition addcond in console you ll be set to yaw turn degrees and remove condition removecond in console you ll lose the bumper kart but preserve your view angles now turn until facing say and give yourself condition again you ll see that your view angle has been snapped back to degrees completely irrelevant of what angle you had as a player you also cannot use the setang command to fix this since condition actively overwrites your view angle at all times this leads to having extremely little control over what direction a player faces when repeatedly granting them a bumper kart for things such as race tracks minigames or otherwise the ideal function of this condition should still overwrite your pitch and roll angles but the yaw angle should be whatever yaw angle the player itself had before gaining the condition rather than whatever the last stored yaw of the kart itself was i recorded a video here of how to reproduce and what the ideal fix should be re entering halloween karts forces players into an incorrect view angle
1
296,453
9,116,099,973
IssuesEvent
2019-02-22 07:55:49
aiidateam/aiida_core
https://api.github.com/repos/aiidateam/aiida_core
closed
Recent move of `DbLogHandler` to `aiida.orm.utils` makes `verdi` super slow
priority/nice to have topic/VerdiCommandLine type/accepted feature
[In a recent commit](https://github.com/aiidateam/aiida_core/commit/b22f55155e7a602ac42b00edf62e57e4a458f55b) the definition of the `DbLogHandler` was moved from `aiida.common.log` to `aiida.orm.utils.log` because `aiida.common` should be importable without having to load the `orm`, which will slow down `verdi` and `DbLogHandler` depends on the ORM. However, because the logging is configured upon the init, the `dblogger` handler is configured, which means it will dereference the class string `'aiida.orm.utils.log.DBLogHandler'`, causing the entire ORM to be loaded. Result: ``` (aiida_prov) sphuber@theos:~/code/aiida/env/prov/aiida-core$ time verdi real 0m1.130s user 0m1.216s sys 0m0.780s ``` We need to find a way to have the `dblogger` handler configured lazily when the ORM is loaded
1.0
Recent move of `DbLogHandler` to `aiida.orm.utils` makes `verdi` super slow - [In a recent commit](https://github.com/aiidateam/aiida_core/commit/b22f55155e7a602ac42b00edf62e57e4a458f55b) the definition of the `DbLogHandler` was moved from `aiida.common.log` to `aiida.orm.utils.log` because `aiida.common` should be importable without having to load the `orm`, which will slow down `verdi` and `DbLogHandler` depends on the ORM. However, because the logging is configured upon the init, the `dblogger` handler is configured, which means it will dereference the class string `'aiida.orm.utils.log.DBLogHandler'`, causing the entire ORM to be loaded. Result: ``` (aiida_prov) sphuber@theos:~/code/aiida/env/prov/aiida-core$ time verdi real 0m1.130s user 0m1.216s sys 0m0.780s ``` We need to find a way to have the `dblogger` handler configured lazily when the ORM is loaded
non_test
recent move of dbloghandler to aiida orm utils makes verdi super slow the definition of the dbloghandler was moved from aiida common log to aiida orm utils log because aiida common should be importable without having to load the orm which will slow down verdi and dbloghandler depends on the orm however because the logging is configured upon the init the dblogger handler is configured which means it will dereference the class string aiida orm utils log dbloghandler causing the entire orm to be loaded result aiida prov sphuber theos code aiida env prov aiida core time verdi real user sys we need to find a way to have the dblogger handler configured lazily when the orm is loaded
0
23,918
4,052,697,696
IssuesEvent
2016-05-24 04:32:04
coreos/etcd
https://api.github.com/repos/coreos/etcd
closed
functional-tester: recover RPC call times out
component/functional-tester
``` 2016-05-08 20:10:44.060406 I | etcd-tester: [round#2931] compacted storage 2016-05-08 20:10:44.060433 I | etcd-tester: [round#2931] check compaction at 1332894215 2016-05-08 20:10:44.269451 I | etcd-tester: [round#2931] confirmed compaction at 1332894215 2016-05-08 20:10:44.353723 I | etcd-tester: [round#2932 case#0] start failure kill all members 2016-05-08 20:10:44.353793 I | etcd-tester: [round#2932 case#0] start injecting failure... 2016-05-08 20:10:44.438626 I | etcd-tester: [round#2932 case#0] injected failure 2016-05-08 20:10:44.438670 I | etcd-tester: [round#2932 case#0] start recovering failure... 2016-05-08 20:12:44.487122 I | etcd-tester: [round#2932 case#0] recovery error: rpc error: code = 4 desc = context deadline exceeded 2016-05-08 20:12:44.487196 I | etcd-tester: [round#2932 case#0] cleaning up... ``` ``` 2016-05-11 10:51:57.902438 I | etcd-tester: [round#84 case#9] all members are consistent with current revisions [revisions: map[10.240.0.19:2379:148713120 10.240.0.17:2379:148713120 10.240.0.18:2379:148713120]] 2016-05-11 10:51:57.902446 I | etcd-tester: [round#84 case#9] checking current storage hashes... 2016-05-11 10:51:57.902453 I | etcd-tester: [round#84 case#9] all members are consistent with storage hashes 2016-05-11 10:51:57.902461 I | etcd-tester: [round#84 case#9] starting the stressers... 2016-05-11 10:51:57.902470 I | etcd-tester: [round#84 case#9] started stressers 2016-05-11 10:51:57.902475 I | etcd-tester: [round#84 case#9] succeed! 2016-05-11 10:51:57.970426 I | etcd-tester: [round#84 case#10] starting failure slow down all members' network 2016-05-11 10:51:57.970464 I | etcd-tester: [round#84 case#10] injecting failure... 2016-05-11 10:51:59.557837 I | etcd-tester: [round#84 case#10] injected failure 2016-05-11 10:51:59.557876 I | etcd-tester: [round#84 case#10] recovering failure... 2016-05-11 10:54:08.161527 I | etcd-tester: [round#84 case#10] recovery error: rpc error: code = 4 desc = context deadline exceeded 2016-05-11 10:54:08.161591 I | etcd-tester: [round#84 case#10] cleaning up... ```
1.0
functional-tester: recover RPC call times out - ``` 2016-05-08 20:10:44.060406 I | etcd-tester: [round#2931] compacted storage 2016-05-08 20:10:44.060433 I | etcd-tester: [round#2931] check compaction at 1332894215 2016-05-08 20:10:44.269451 I | etcd-tester: [round#2931] confirmed compaction at 1332894215 2016-05-08 20:10:44.353723 I | etcd-tester: [round#2932 case#0] start failure kill all members 2016-05-08 20:10:44.353793 I | etcd-tester: [round#2932 case#0] start injecting failure... 2016-05-08 20:10:44.438626 I | etcd-tester: [round#2932 case#0] injected failure 2016-05-08 20:10:44.438670 I | etcd-tester: [round#2932 case#0] start recovering failure... 2016-05-08 20:12:44.487122 I | etcd-tester: [round#2932 case#0] recovery error: rpc error: code = 4 desc = context deadline exceeded 2016-05-08 20:12:44.487196 I | etcd-tester: [round#2932 case#0] cleaning up... ``` ``` 2016-05-11 10:51:57.902438 I | etcd-tester: [round#84 case#9] all members are consistent with current revisions [revisions: map[10.240.0.19:2379:148713120 10.240.0.17:2379:148713120 10.240.0.18:2379:148713120]] 2016-05-11 10:51:57.902446 I | etcd-tester: [round#84 case#9] checking current storage hashes... 2016-05-11 10:51:57.902453 I | etcd-tester: [round#84 case#9] all members are consistent with storage hashes 2016-05-11 10:51:57.902461 I | etcd-tester: [round#84 case#9] starting the stressers... 2016-05-11 10:51:57.902470 I | etcd-tester: [round#84 case#9] started stressers 2016-05-11 10:51:57.902475 I | etcd-tester: [round#84 case#9] succeed! 2016-05-11 10:51:57.970426 I | etcd-tester: [round#84 case#10] starting failure slow down all members' network 2016-05-11 10:51:57.970464 I | etcd-tester: [round#84 case#10] injecting failure... 2016-05-11 10:51:59.557837 I | etcd-tester: [round#84 case#10] injected failure 2016-05-11 10:51:59.557876 I | etcd-tester: [round#84 case#10] recovering failure... 2016-05-11 10:54:08.161527 I | etcd-tester: [round#84 case#10] recovery error: rpc error: code = 4 desc = context deadline exceeded 2016-05-11 10:54:08.161591 I | etcd-tester: [round#84 case#10] cleaning up... ```
test
functional tester recover rpc call times out i etcd tester compacted storage i etcd tester check compaction at i etcd tester confirmed compaction at i etcd tester start failure kill all members i etcd tester start injecting failure i etcd tester injected failure i etcd tester start recovering failure i etcd tester recovery error rpc error code desc context deadline exceeded i etcd tester cleaning up i etcd tester all members are consistent with current revisions i etcd tester checking current storage hashes i etcd tester all members are consistent with storage hashes i etcd tester starting the stressers i etcd tester started stressers i etcd tester succeed i etcd tester starting failure slow down all members network i etcd tester injecting failure i etcd tester injected failure i etcd tester recovering failure i etcd tester recovery error rpc error code desc context deadline exceeded i etcd tester cleaning up
1
156,099
13,641,895,791
IssuesEvent
2020-09-25 14:47:09
maxitg/SetReplace
https://api.github.com/repos/maxitg/SetReplace
closed
Local multiway system description
documentation soon writing
Before implementing the local multiway system, I need to describe what it's going to do, which is just going to be a .md document. This should mention problems with the current `MultiwaySystem`, the explanation of how the local multiway system will solve them. It should also explain the different options that will be available to control evolution and the implementation milestones.
1.0
Local multiway system description - Before implementing the local multiway system, I need to describe what it's going to do, which is just going to be a .md document. This should mention problems with the current `MultiwaySystem`, the explanation of how the local multiway system will solve them. It should also explain the different options that will be available to control evolution and the implementation milestones.
non_test
local multiway system description before implementing the local multiway system i need to describe what it s going to do which is just going to be a md document this should mention problems with the current multiwaysystem the explanation of how the local multiway system will solve them it should also explain the different options that will be available to control evolution and the implementation milestones
0
667,191
22,421,472,957
IssuesEvent
2022-06-20 04:04:30
pvs-hd-tea/LapsPython
https://api.github.com/repos/pvs-hd-tea/LapsPython
closed
feat: Remove comments from parsed primitives
Feature Request Suggestion Low Priority
Translations can currently be cluttered with unneeded comments, use regex to remove them in each line 2 cases: 1. Remove entire line if it only contains a comment 2. Only remove the comment if it follows code in the same line
1.0
feat: Remove comments from parsed primitives - Translations can currently be cluttered with unneeded comments, use regex to remove them in each line 2 cases: 1. Remove entire line if it only contains a comment 2. Only remove the comment if it follows code in the same line
non_test
feat remove comments from parsed primitives translations can currently be cluttered with unneeded comments use regex to remove them in each line cases remove entire line if it only contains a comment only remove the comment if it follows code in the same line
0
831,210
32,041,264,302
IssuesEvent
2023-09-22 19:33:41
ctrl-space-labs/gendox-core
https://api.github.com/repos/ctrl-space-labs/gendox-core
opened
As a Backend developer I want to associate Uploaded documents with the project in the RequestPara
bug backend priority-1
### Expected functionality In the `documents/upload` `projectId` is a request param. Uploaded document need to be associated with this project in DB ### Actual Functionality The documents are associated only with the Organization, and not the projectID
1.0
As a Backend developer I want to associate Uploaded documents with the project in the RequestPara - ### Expected functionality In the `documents/upload` `projectId` is a request param. Uploaded document need to be associated with this project in DB ### Actual Functionality The documents are associated only with the Organization, and not the projectID
non_test
as a backend developer i want to associate uploaded documents with the project in the requestpara expected functionality in the documents upload projectid is a request param uploaded document need to be associated with this project in db actual functionality the documents are associated only with the organization and not the projectid
0
21,889
14,918,487,813
IssuesEvent
2021-01-22 21:47:32
PavlidisLab/Gemma
https://api.github.com/repos/PavlidisLab/Gemma
opened
Use Tomcat Manager App to deploy/redeploy Gemma
enhancement infrastructure
This is less disruptive than restarting Tomcat and safer than auto-reloading the app when then WAR changes. - [ ] disable auto-reload - [ ] add jenkins to the list of authorized users in Tomcat configuration More details here: https://tomcat.apache.org/tomcat-8.5-doc/manager-howto.html#Supported_Manager_Commands It should be first tested on the development instance.
1.0
Use Tomcat Manager App to deploy/redeploy Gemma - This is less disruptive than restarting Tomcat and safer than auto-reloading the app when then WAR changes. - [ ] disable auto-reload - [ ] add jenkins to the list of authorized users in Tomcat configuration More details here: https://tomcat.apache.org/tomcat-8.5-doc/manager-howto.html#Supported_Manager_Commands It should be first tested on the development instance.
non_test
use tomcat manager app to deploy redeploy gemma this is less disruptive than restarting tomcat and safer than auto reloading the app when then war changes disable auto reload add jenkins to the list of authorized users in tomcat configuration more details here it should be first tested on the development instance
0
139,764
31,776,032,723
IssuesEvent
2023-09-12 14:28:18
ArctosDB/arctos
https://api.github.com/repos/ArctosDB/arctos
opened
Code Table Request - New identification attribute: verbatim identification
Priority-High (Needed for work) Function-CodeTables
### Goal _Describe what you're trying to accomplish. This is the only necessary step to start this process. The Committee is available to assist with all other steps. Please clearly indicate any uncertainty or desired guidance if you proceed beyond this step._ Record the identification as written that resulted in the taxon name being associated with a catalog record in a manner that makes it easy to understand. ### Context _Describe why this new value is necessary and existing values are not._ https://github.com/ArctosDB/arctos/issues/4827#issuecomment-1710709904 It is clear from that discussion that this would be useful to a few collections (MSB, CHAS, OGL) and possibly others if they were aware of the conversation. I understand the concept of simply using a second identification, however, our current method of ranking identifications does not allow a method for marking an identification as "verbatim" and I do not believe it is clear to anyone outside of (and perhaps even within) the Arctos community that an A {string} identification = verbatim. In addition, using an A {string} identification still requires a remark to flag the identification as "verbatim" adding to the [overburden of identification remark](https://github.com/ArctosDB/arctos/issues/4922). The alternate that I could see to this is replacing our current identification numerical ranks with values such as accepted, unaccepted, verbatim. The numbers 2-10 have no meaning and in some ways are useless except to whoever assigned the order and knows why they assigned it unless it is laid out in the identification remark adding to the [overburden of identification remark](https://github.com/ArctosDB/arctos/issues/4922). I could argue that 0 and 1 are also meaningless to anyone but "us". ### Table _Code Tables are http://arctos.database.museum/info/ctDocumentation.cfm. Link to the specific table or value. This may involve multiple tables and will control datatype for Attributes. OtherID requests require BaseURL (and example) or explanation. Please ask for assistance if unsure._ https://arctos.database.museum/info/ctDocumentation.cfm?table=ctidentification_attribute_type ### Proposed Value _Proposed new value. This should be clear and compatible with similar values in the relevant table and across Arctos._ verbatim identification ### Proposed Definition _Clear, complete, non-collection-type-specific **functional** definition of the value. Avoid discipline-specific terminology if possible, include parenthetically if unavoidable._ A string representing the taxonomic identification as it appeared in the original record. - from Darwin Core [verbatimIdentification](https://dwc.tdwg.org/terms/#dwc:verbatimIdentification) ### Collection type _Some code tables contain collection-type-specific values. ``collection_cde`` may be found from https://arctos.database.museum/home.cfm_ N/A ### Attribute Extras #### Attribute data type _If the request is for an attribute, what values will be allowed? free-text, categorical, or number+units depending upon the attribute (TBA)_ free-text #### Attribute controlled values _If the values are categorical (to be controlled by a code table), add a link to the appropriate code table. If a new table or set of values is needed, please elaborate._ N/A #### Attribute units _if numerical values should be accompanied by units, provide a link to the appropriate units table._ N/A ### Priority _Please describe the urgency and/or choose a priority-label to the right. You should expect a response within two working days, and may utilize [Arctos Contacts](https://arctosdb.org/contacts/) if you feel response is lacking._ High - we need a resolution for the incoming OGL data. ### Example Data _Requests with clarifying sample data are generally much easier to understand and prioritize. Please attach or link to any representative data, in any form or format, which might help clarify the request._ We always list verbatim name, even if it is identical to the 'official' scientific name, and having that as 2 official identifications seems confusing not clarifying. - https://github.com/ArctosDB/arctos/issues/4827#issuecomment-1715736330 ### Available for Public View _Most data are by default publicly available. Describe any necessary access restrictions._ yes ### Helpful Actions - [ ] Add the issue to the [Code Table Management Project](https://github.com/ArctosDB/arctos/projects/13#card-31628184). - [ ] Please reach out to anyone who might be affected by this change. Leave a comment or add this to the Committee agenda if you believe more focused conversation is necessary. Tagging all participants in the other issue: @happiah-madson @falco-rk @campmlc @wellerjes @DerekSikes Others who may have an interest: @sharpphyl @genevieve-anderegg (this would be applicable to your identification order = 2 identifications) @ArctosDB/arctos-code-table-administrators ## Approval _All of the following must be checked before this may proceed._ _The [How-To Document](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html) should be followed. Pay particular attention to terminology (with emphasis on consistency) and documentation (with emphasis on functionality). **No person should act in multiple roles**; the submitter cannot also serve as a Code Table Administrator, for example._ - [ ] Code Table Administrator[1] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval - [ ] Code Table Administrator[2] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval - [ ] DBA - The request is functionally acceptable. The term is not a functional duplicate, and is compatible with existing data and code. - [ ] DBA - Appropriate code or handlers are in place as necessary. (ID_References, Media Relationships, Encumbrances, etc. require particular attention) ## Rejection _If you believe this request should not proceed, explain why here. Suggest any changes that would make the change acceptable, alternate (usually existing) paths to the same goals, etc._ 1. _Can a suitable solution be found here? If not, proceed to (2)_ 2. _Can a suitable solution be found by Code Table Committee discussion? If not, proceed to (3)_ 3. _Take the discussion to a monthly Arctos Working Group meeting for final resolution._ ## Implementation _Once all of the Approval Checklist is appropriately checked and there are no Rejection comments, or in special circumstances by decree of the Arctos Working Group, the change may be made._ - [ ] Review everything one last time. Ensure the How-To has been followed. Ensure all checks have been made by appropriate personnel. - [ ] Add or revise the code table term/definition as described above. Ensure the URL of this Issue is included in the definition. _Close this Issue._ _**DO NOT** modify Arctos Authorities in any way before all points in this Issue have been fully addressed; data loss may result._ ## Special Exemptions _In very specific cases and by prior approval of The Committee, the approval process may be skipped, and implementation requirements may be slightly altered. Please note here if you are proceeding under one of these use cases._ 1. _Adding an existing term to additional collection types may proceed immediately and without discussion, but doing so may also subject users to future cleanup efforts. If time allows, please review the term and definition as part of this step._ 2. _The Committee may grant special access on particular tables to particular users. This should be exercised with great caution only after several smooth test cases, and generally limited to "taxonomy-like" data such as International Commission on Stratigraphy terminology._
1.0
Code Table Request - New identification attribute: verbatim identification - ### Goal _Describe what you're trying to accomplish. This is the only necessary step to start this process. The Committee is available to assist with all other steps. Please clearly indicate any uncertainty or desired guidance if you proceed beyond this step._ Record the identification as written that resulted in the taxon name being associated with a catalog record in a manner that makes it easy to understand. ### Context _Describe why this new value is necessary and existing values are not._ https://github.com/ArctosDB/arctos/issues/4827#issuecomment-1710709904 It is clear from that discussion that this would be useful to a few collections (MSB, CHAS, OGL) and possibly others if they were aware of the conversation. I understand the concept of simply using a second identification, however, our current method of ranking identifications does not allow a method for marking an identification as "verbatim" and I do not believe it is clear to anyone outside of (and perhaps even within) the Arctos community that an A {string} identification = verbatim. In addition, using an A {string} identification still requires a remark to flag the identification as "verbatim" adding to the [overburden of identification remark](https://github.com/ArctosDB/arctos/issues/4922). The alternate that I could see to this is replacing our current identification numerical ranks with values such as accepted, unaccepted, verbatim. The numbers 2-10 have no meaning and in some ways are useless except to whoever assigned the order and knows why they assigned it unless it is laid out in the identification remark adding to the [overburden of identification remark](https://github.com/ArctosDB/arctos/issues/4922). I could argue that 0 and 1 are also meaningless to anyone but "us". ### Table _Code Tables are http://arctos.database.museum/info/ctDocumentation.cfm. Link to the specific table or value. This may involve multiple tables and will control datatype for Attributes. OtherID requests require BaseURL (and example) or explanation. Please ask for assistance if unsure._ https://arctos.database.museum/info/ctDocumentation.cfm?table=ctidentification_attribute_type ### Proposed Value _Proposed new value. This should be clear and compatible with similar values in the relevant table and across Arctos._ verbatim identification ### Proposed Definition _Clear, complete, non-collection-type-specific **functional** definition of the value. Avoid discipline-specific terminology if possible, include parenthetically if unavoidable._ A string representing the taxonomic identification as it appeared in the original record. - from Darwin Core [verbatimIdentification](https://dwc.tdwg.org/terms/#dwc:verbatimIdentification) ### Collection type _Some code tables contain collection-type-specific values. ``collection_cde`` may be found from https://arctos.database.museum/home.cfm_ N/A ### Attribute Extras #### Attribute data type _If the request is for an attribute, what values will be allowed? free-text, categorical, or number+units depending upon the attribute (TBA)_ free-text #### Attribute controlled values _If the values are categorical (to be controlled by a code table), add a link to the appropriate code table. If a new table or set of values is needed, please elaborate._ N/A #### Attribute units _if numerical values should be accompanied by units, provide a link to the appropriate units table._ N/A ### Priority _Please describe the urgency and/or choose a priority-label to the right. You should expect a response within two working days, and may utilize [Arctos Contacts](https://arctosdb.org/contacts/) if you feel response is lacking._ High - we need a resolution for the incoming OGL data. ### Example Data _Requests with clarifying sample data are generally much easier to understand and prioritize. Please attach or link to any representative data, in any form or format, which might help clarify the request._ We always list verbatim name, even if it is identical to the 'official' scientific name, and having that as 2 official identifications seems confusing not clarifying. - https://github.com/ArctosDB/arctos/issues/4827#issuecomment-1715736330 ### Available for Public View _Most data are by default publicly available. Describe any necessary access restrictions._ yes ### Helpful Actions - [ ] Add the issue to the [Code Table Management Project](https://github.com/ArctosDB/arctos/projects/13#card-31628184). - [ ] Please reach out to anyone who might be affected by this change. Leave a comment or add this to the Committee agenda if you believe more focused conversation is necessary. Tagging all participants in the other issue: @happiah-madson @falco-rk @campmlc @wellerjes @DerekSikes Others who may have an interest: @sharpphyl @genevieve-anderegg (this would be applicable to your identification order = 2 identifications) @ArctosDB/arctos-code-table-administrators ## Approval _All of the following must be checked before this may proceed._ _The [How-To Document](https://handbook.arctosdb.org/how_to/How-To-Manage-Code-Table-Requests.html) should be followed. Pay particular attention to terminology (with emphasis on consistency) and documentation (with emphasis on functionality). **No person should act in multiple roles**; the submitter cannot also serve as a Code Table Administrator, for example._ - [ ] Code Table Administrator[1] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval - [ ] Code Table Administrator[2] - check and initial, comment, or thumbs-up to indicate that the request complies with the how-to documentation and has your approval - [ ] DBA - The request is functionally acceptable. The term is not a functional duplicate, and is compatible with existing data and code. - [ ] DBA - Appropriate code or handlers are in place as necessary. (ID_References, Media Relationships, Encumbrances, etc. require particular attention) ## Rejection _If you believe this request should not proceed, explain why here. Suggest any changes that would make the change acceptable, alternate (usually existing) paths to the same goals, etc._ 1. _Can a suitable solution be found here? If not, proceed to (2)_ 2. _Can a suitable solution be found by Code Table Committee discussion? If not, proceed to (3)_ 3. _Take the discussion to a monthly Arctos Working Group meeting for final resolution._ ## Implementation _Once all of the Approval Checklist is appropriately checked and there are no Rejection comments, or in special circumstances by decree of the Arctos Working Group, the change may be made._ - [ ] Review everything one last time. Ensure the How-To has been followed. Ensure all checks have been made by appropriate personnel. - [ ] Add or revise the code table term/definition as described above. Ensure the URL of this Issue is included in the definition. _Close this Issue._ _**DO NOT** modify Arctos Authorities in any way before all points in this Issue have been fully addressed; data loss may result._ ## Special Exemptions _In very specific cases and by prior approval of The Committee, the approval process may be skipped, and implementation requirements may be slightly altered. Please note here if you are proceeding under one of these use cases._ 1. _Adding an existing term to additional collection types may proceed immediately and without discussion, but doing so may also subject users to future cleanup efforts. If time allows, please review the term and definition as part of this step._ 2. _The Committee may grant special access on particular tables to particular users. This should be exercised with great caution only after several smooth test cases, and generally limited to "taxonomy-like" data such as International Commission on Stratigraphy terminology._
non_test
code table request new identification attribute verbatim identification goal describe what you re trying to accomplish this is the only necessary step to start this process the committee is available to assist with all other steps please clearly indicate any uncertainty or desired guidance if you proceed beyond this step record the identification as written that resulted in the taxon name being associated with a catalog record in a manner that makes it easy to understand context describe why this new value is necessary and existing values are not it is clear from that discussion that this would be useful to a few collections msb chas ogl and possibly others if they were aware of the conversation i understand the concept of simply using a second identification however our current method of ranking identifications does not allow a method for marking an identification as verbatim and i do not believe it is clear to anyone outside of and perhaps even within the arctos community that an a string identification verbatim in addition using an a string identification still requires a remark to flag the identification as verbatim adding to the the alternate that i could see to this is replacing our current identification numerical ranks with values such as accepted unaccepted verbatim the numbers have no meaning and in some ways are useless except to whoever assigned the order and knows why they assigned it unless it is laid out in the identification remark adding to the i could argue that and are also meaningless to anyone but us table code tables are link to the specific table or value this may involve multiple tables and will control datatype for attributes otherid requests require baseurl and example or explanation please ask for assistance if unsure proposed value proposed new value this should be clear and compatible with similar values in the relevant table and across arctos verbatim identification proposed definition clear complete non collection type specific functional definition of the value avoid discipline specific terminology if possible include parenthetically if unavoidable a string representing the taxonomic identification as it appeared in the original record from darwin core collection type some code tables contain collection type specific values collection cde may be found from n a attribute extras attribute data type if the request is for an attribute what values will be allowed free text categorical or number units depending upon the attribute tba free text attribute controlled values if the values are categorical to be controlled by a code table add a link to the appropriate code table if a new table or set of values is needed please elaborate n a attribute units if numerical values should be accompanied by units provide a link to the appropriate units table n a priority please describe the urgency and or choose a priority label to the right you should expect a response within two working days and may utilize if you feel response is lacking high we need a resolution for the incoming ogl data example data requests with clarifying sample data are generally much easier to understand and prioritize please attach or link to any representative data in any form or format which might help clarify the request we always list verbatim name even if it is identical to the official scientific name and having that as official identifications seems confusing not clarifying available for public view most data are by default publicly available describe any necessary access restrictions yes helpful actions add the issue to the please reach out to anyone who might be affected by this change leave a comment or add this to the committee agenda if you believe more focused conversation is necessary tagging all participants in the other issue happiah madson falco rk campmlc wellerjes dereksikes others who may have an interest sharpphyl genevieve anderegg this would be applicable to your identification order identifications arctosdb arctos code table administrators approval all of the following must be checked before this may proceed the should be followed pay particular attention to terminology with emphasis on consistency and documentation with emphasis on functionality no person should act in multiple roles the submitter cannot also serve as a code table administrator for example code table administrator check and initial comment or thumbs up to indicate that the request complies with the how to documentation and has your approval code table administrator check and initial comment or thumbs up to indicate that the request complies with the how to documentation and has your approval dba the request is functionally acceptable the term is not a functional duplicate and is compatible with existing data and code dba appropriate code or handlers are in place as necessary id references media relationships encumbrances etc require particular attention rejection if you believe this request should not proceed explain why here suggest any changes that would make the change acceptable alternate usually existing paths to the same goals etc can a suitable solution be found here if not proceed to can a suitable solution be found by code table committee discussion if not proceed to take the discussion to a monthly arctos working group meeting for final resolution implementation once all of the approval checklist is appropriately checked and there are no rejection comments or in special circumstances by decree of the arctos working group the change may be made review everything one last time ensure the how to has been followed ensure all checks have been made by appropriate personnel add or revise the code table term definition as described above ensure the url of this issue is included in the definition close this issue do not modify arctos authorities in any way before all points in this issue have been fully addressed data loss may result special exemptions in very specific cases and by prior approval of the committee the approval process may be skipped and implementation requirements may be slightly altered please note here if you are proceeding under one of these use cases adding an existing term to additional collection types may proceed immediately and without discussion but doing so may also subject users to future cleanup efforts if time allows please review the term and definition as part of this step the committee may grant special access on particular tables to particular users this should be exercised with great caution only after several smooth test cases and generally limited to taxonomy like data such as international commission on stratigraphy terminology
0
28,044
5,167,306,408
IssuesEvent
2017-01-17 18:26:27
eliasferreyra/googlesitemapgenerator
https://api.github.com/repos/eliasferreyra/googlesitemapgenerator
closed
error while installing googlesitemapgenerator
auto-migrated Priority-Medium Type-Defect
``` What steps will reproduce the problem? 1. click install go through steps last step 2. error appears 3. What is the expected output? What do you see instead? install complete error a component didnt install What version of the product are you using? On what operating system? google site map generator Please provide any additional information below. im useing windows vista ``` Original issue reported on code.google.com by `ddhidub...@yahoo.com` on 1 Oct 2009 at 12:25 Attachments: - [sitemap_win-beta1-20090225.msi](https://storage.googleapis.com/google-code-attachments/googlesitemapgenerator/issue-76/comment-0/sitemap_win-beta1-20090225.msi)
1.0
error while installing googlesitemapgenerator - ``` What steps will reproduce the problem? 1. click install go through steps last step 2. error appears 3. What is the expected output? What do you see instead? install complete error a component didnt install What version of the product are you using? On what operating system? google site map generator Please provide any additional information below. im useing windows vista ``` Original issue reported on code.google.com by `ddhidub...@yahoo.com` on 1 Oct 2009 at 12:25 Attachments: - [sitemap_win-beta1-20090225.msi](https://storage.googleapis.com/google-code-attachments/googlesitemapgenerator/issue-76/comment-0/sitemap_win-beta1-20090225.msi)
non_test
error while installing googlesitemapgenerator what steps will reproduce the problem click install go through steps last step error appears what is the expected output what do you see instead install complete error a component didnt install what version of the product are you using on what operating system google site map generator please provide any additional information below im useing windows vista original issue reported on code google com by ddhidub yahoo com on oct at attachments
0
457,951
13,165,602,941
IssuesEvent
2020-08-11 06:58:21
buddyboss/buddyboss-platform
https://api.github.com/repos/buddyboss/buddyboss-platform
opened
Private Network issue
bug priority: medium
**Describe the bug** Earlier If we want to make sub-pages accessible under a certain path eg site.com/groups/test-group while private network in enabled. Removing the trailing slash from the URL eg site.com/groups was doing the job but It's not working anymore. **To Reproduce** Steps to reproduce the behavior: 1. Go to Admin Dashboard >> BuddyBoss >> Settings. 2. Turn On the Private Network option. 3. Now add any URL in Public content area without trailing slash eg www.site.com/groups 4. See on the front end, you will not be able to access subpages of the groups www.site.com/groups/test-group **Expected behavior** If we don't use trailing slash, it should make subpages accessible for public. It was like this before. But it's not working anymore. **Support ticket links** https://secure.helpscout.net/conversation/1246713962/86891?folderId=3701248
1.0
Private Network issue - **Describe the bug** Earlier If we want to make sub-pages accessible under a certain path eg site.com/groups/test-group while private network in enabled. Removing the trailing slash from the URL eg site.com/groups was doing the job but It's not working anymore. **To Reproduce** Steps to reproduce the behavior: 1. Go to Admin Dashboard >> BuddyBoss >> Settings. 2. Turn On the Private Network option. 3. Now add any URL in Public content area without trailing slash eg www.site.com/groups 4. See on the front end, you will not be able to access subpages of the groups www.site.com/groups/test-group **Expected behavior** If we don't use trailing slash, it should make subpages accessible for public. It was like this before. But it's not working anymore. **Support ticket links** https://secure.helpscout.net/conversation/1246713962/86891?folderId=3701248
non_test
private network issue describe the bug earlier if we want to make sub pages accessible under a certain path eg site com groups test group while private network in enabled removing the trailing slash from the url eg site com groups was doing the job but it s not working anymore to reproduce steps to reproduce the behavior go to admin dashboard buddyboss settings turn on the private network option now add any url in public content area without trailing slash eg see on the front end you will not be able to access subpages of the groups expected behavior if we don t use trailing slash it should make subpages accessible for public it was like this before but it s not working anymore support ticket links
0
161,018
12,529,885,335
IssuesEvent
2020-06-04 12:09:34
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
MixedClusterClientYamlTestSuiteIT fails with "Partial shards failure" instead of "Trying to create too many buckets"
:Search/Search >test-failure Team:Search
**Build scan**: https://gradle-enterprise.elastic.co/s/3bcaij5hfyzhs **Repro line**: `./gradlew ':qa:mixed-cluster:v7.2.0#mixedClusterTest' --tests "org.elasticsearch.backwards.MixedClusterClientYamlTestSuiteIT" \ -Dtests.method="test {p0=search.aggregation/240_max_buckets/Max bucket}" \ -Dtests.seed=D1E3CB4F243CA3C9 \ -Dtests.security.manager=true \ -Dtests.locale=fr-FR \ -Dtests.timezone=America/Regina \ -Druntime.java=8` **Reproduces locally?**: Yes. **Applicable branches**: `7.x` **Failure history**: Nothing obvious in the last week. **Failure excerpt**: ``` org.elasticsearch.backwards.MixedClusterClientYamlTestSuiteIT > test {p0=search.aggregation/240_max_buckets/Max bucket} FAILED |   -- | --   | java.lang.AssertionError: Failure at [search.aggregation/240_max_buckets:88]: the error message was expected to match the provided regex but didn't |     | Expected: .*Trying to create too many buckets.* |     | but: was "{root_cause=[], type=search_phase_execution_exception, reason=Partial shards failure, phase=query, grouped=true, failed_shards=[], stack_trace=Failed to execute phase [query], Partial shards failure\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseFailure(AbstractSearchAsyncAction.java:296)\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.executeNextPhase(AbstractSearchAsyncAction.java:151)\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseDone(AbstractSearchAsyncAction.java:259)\n\tat org.elasticsearch.action.search.InitialSearchPhase.successfulShardExecution(InitialSearchPhase.java:320)\n\tat org.elasticsearch.action.search.InitialSearchPhase.onShardResult(InitialSearchPhase.java:308)\n\tat org.elasticsearch.action.search.InitialSearchPhase.access$000(InitialSearchPhase.java:50)\n\tat org.elasticsearch.action.search.InitialSearchPhase$2.innerOnResponse(InitialSearchPhase.java:264)\n\tat org.elasticsearch.action.search.SearchActionListener.onResponse(SearchActionListener.java:45)\n\tat org.elasticsearch.action.search.SearchActionListener.onResponse(SearchActionListener.java:29)\n\tat org.elasticsearch.action.search.SearchExecutionStatsCollector.onResponse(SearchExecutionStatsCollector.java:68)\n\tat org.elasticsearch.action.search.SearchExecutionStatsCollector.onResponse(SearchExecutionStatsCollector.java:36)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.action.search.SearchTransportService$ConnectionCountingHandler.handleResponse(SearchTransportService.java:432)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1101)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:224)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:193)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:216)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:141)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:105)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:660)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:323)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:297)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:682)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:582)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:536)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:906)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:835)\n}" |     | at __randomizedtesting.SeedInfo.seed([D1E3CB4F243CA3C9:59B7F4958AC0CE31]:0) |     | at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:405) |     | at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:382) |     | at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source) |     | at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) |     | at java.lang.reflect.Method.invoke(Method.java:498) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) |     | at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) |     | at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) |     | at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) |     | at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) |     | at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) |     | at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894) |     | at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) |     | at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) |     | at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) |     | at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) |     | at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) |     | at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) |     | at java.lang.Thread.run(Thread.java:748) |     |   |     | Caused by: |     | java.lang.AssertionError: the error message was expected to match the provided regex but didn't |     | Expected: .*Trying to create too many buckets.* |     | but: was "{root_cause=[], type=search_phase_execution_exception, reason=Partial shards failure, phase=query, grouped=true, failed_shards=[], stack_trace=Failed to execute phase [query], Partial shards failure\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseFailure(AbstractSearchAsyncAction.java:296)\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.executeNextPhase(AbstractSearchAsyncAction.java:151)\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseDone(AbstractSearchAsyncAction.java:259)\n\tat org.elasticsearch.action.search.InitialSearchPhase.successfulShardExecution(InitialSearchPhase.java:320)\n\tat org.elasticsearch.action.search.InitialSearchPhase.onShardResult(InitialSearchPhase.java:308)\n\tat org.elasticsearch.action.search.InitialSearchPhase.access$000(InitialSearchPhase.java:50)\n\tat org.elasticsearch.action.search.InitialSearchPhase$2.innerOnResponse(InitialSearchPhase.java:264)\n\tat org.elasticsearch.action.search.SearchActionListener.onResponse(SearchActionListener.java:45)\n\tat org.elasticsearch.action.search.SearchActionListener.onResponse(SearchActionListener.java:29)\n\tat org.elasticsearch.action.search.SearchExecutionStatsCollector.onResponse(SearchExecutionStatsCollector.java:68)\n\tat org.elasticsearch.action.search.SearchExecutionStatsCollector.onResponse(SearchExecutionStatsCollector.java:36)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.action.search.SearchTransportService$ConnectionCountingHandler.handleResponse(SearchTransportService.java:432)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1101)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:224)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:193)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:216)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:141)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:105)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:660)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:323)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:297)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:682)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:582)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:536)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:906)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:835)\n}" |     | at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) |     | at org.junit.Assert.assertThat(Assert.java:956) |     | at org.elasticsearch.test.rest.yaml.section.DoSection.execute(DoSection.java:308) |     | at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:398) |     | ... 37 more |     | REPRODUCE WITH: ./gradlew ':qa:mixed-cluster:v7.2.0#mixedClusterTest' --tests "org.elasticsearch.backwards.MixedClusterClientYamlTestSuiteIT" -Dtests.method="test {p0=search.aggregation/240_max_buckets/Max bucket}" -Dtests.seed=D1E3CB4F243CA3C9 -Dtests.security.manager=true -Dtests.locale=fr-FR -Dtests.timezone=America/Regina -Druntime.java=8 |     |   ``` Not sure if this is related, but it looks wrong too: ``` » WARN ][r.suppressed ] [v7.2.0-0] path: _index_template/*, params: {name=*} » org.elasticsearch.transport.RemoteTransportException: [v7.2.0-2][127.0.0.1:45944][indices:admin/index_template/delete] » Caused by: org.elasticsearch.transport.ActionNotFoundTransportException: No handler for action [indices:admin/index_template/delete] » at org.elasticsearch.transport.InboundHandler.handleRequest(InboundHandler.java:169) ~[elasticsearch-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:121) ~[elasticsearch-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:105) ~[elasticsearch-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:660) ~[elasticsearch-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62) ~[transport-netty4-client-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:323) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final] » at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:297) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241) ~[netty-handler-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:682) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:582) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:536) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:906) ~[?:?] » at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) [netty-common-4.1.49.Final.jar:4.1.49.Final] » at java.lang.Thread.run(Thread.java:835) [?:1.8.0_241] ```
1.0
MixedClusterClientYamlTestSuiteIT fails with "Partial shards failure" instead of "Trying to create too many buckets" - **Build scan**: https://gradle-enterprise.elastic.co/s/3bcaij5hfyzhs **Repro line**: `./gradlew ':qa:mixed-cluster:v7.2.0#mixedClusterTest' --tests "org.elasticsearch.backwards.MixedClusterClientYamlTestSuiteIT" \ -Dtests.method="test {p0=search.aggregation/240_max_buckets/Max bucket}" \ -Dtests.seed=D1E3CB4F243CA3C9 \ -Dtests.security.manager=true \ -Dtests.locale=fr-FR \ -Dtests.timezone=America/Regina \ -Druntime.java=8` **Reproduces locally?**: Yes. **Applicable branches**: `7.x` **Failure history**: Nothing obvious in the last week. **Failure excerpt**: ``` org.elasticsearch.backwards.MixedClusterClientYamlTestSuiteIT > test {p0=search.aggregation/240_max_buckets/Max bucket} FAILED |   -- | --   | java.lang.AssertionError: Failure at [search.aggregation/240_max_buckets:88]: the error message was expected to match the provided regex but didn't |     | Expected: .*Trying to create too many buckets.* |     | but: was "{root_cause=[], type=search_phase_execution_exception, reason=Partial shards failure, phase=query, grouped=true, failed_shards=[], stack_trace=Failed to execute phase [query], Partial shards failure\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseFailure(AbstractSearchAsyncAction.java:296)\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.executeNextPhase(AbstractSearchAsyncAction.java:151)\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseDone(AbstractSearchAsyncAction.java:259)\n\tat org.elasticsearch.action.search.InitialSearchPhase.successfulShardExecution(InitialSearchPhase.java:320)\n\tat org.elasticsearch.action.search.InitialSearchPhase.onShardResult(InitialSearchPhase.java:308)\n\tat org.elasticsearch.action.search.InitialSearchPhase.access$000(InitialSearchPhase.java:50)\n\tat org.elasticsearch.action.search.InitialSearchPhase$2.innerOnResponse(InitialSearchPhase.java:264)\n\tat org.elasticsearch.action.search.SearchActionListener.onResponse(SearchActionListener.java:45)\n\tat org.elasticsearch.action.search.SearchActionListener.onResponse(SearchActionListener.java:29)\n\tat org.elasticsearch.action.search.SearchExecutionStatsCollector.onResponse(SearchExecutionStatsCollector.java:68)\n\tat org.elasticsearch.action.search.SearchExecutionStatsCollector.onResponse(SearchExecutionStatsCollector.java:36)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.action.search.SearchTransportService$ConnectionCountingHandler.handleResponse(SearchTransportService.java:432)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1101)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:224)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:193)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:216)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:141)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:105)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:660)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:323)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:297)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:682)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:582)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:536)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:906)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:835)\n}" |     | at __randomizedtesting.SeedInfo.seed([D1E3CB4F243CA3C9:59B7F4958AC0CE31]:0) |     | at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:405) |     | at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.test(ESClientYamlSuiteTestCase.java:382) |     | at sun.reflect.GeneratedMethodAccessor15.invoke(Unknown Source) |     | at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) |     | at java.lang.reflect.Method.invoke(Method.java:498) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1750) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:938) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:974) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:988) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) |     | at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) |     | at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) |     | at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) |     | at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) |     | at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:817) |     | at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:468) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:947) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:832) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:883) |     | at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:894) |     | at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) |     | at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) |     | at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) |     | at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) |     | at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) |     | at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) |     | at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) |     | at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:368) |     | at java.lang.Thread.run(Thread.java:748) |     |   |     | Caused by: |     | java.lang.AssertionError: the error message was expected to match the provided regex but didn't |     | Expected: .*Trying to create too many buckets.* |     | but: was "{root_cause=[], type=search_phase_execution_exception, reason=Partial shards failure, phase=query, grouped=true, failed_shards=[], stack_trace=Failed to execute phase [query], Partial shards failure\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseFailure(AbstractSearchAsyncAction.java:296)\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.executeNextPhase(AbstractSearchAsyncAction.java:151)\n\tat org.elasticsearch.action.search.AbstractSearchAsyncAction.onPhaseDone(AbstractSearchAsyncAction.java:259)\n\tat org.elasticsearch.action.search.InitialSearchPhase.successfulShardExecution(InitialSearchPhase.java:320)\n\tat org.elasticsearch.action.search.InitialSearchPhase.onShardResult(InitialSearchPhase.java:308)\n\tat org.elasticsearch.action.search.InitialSearchPhase.access$000(InitialSearchPhase.java:50)\n\tat org.elasticsearch.action.search.InitialSearchPhase$2.innerOnResponse(InitialSearchPhase.java:264)\n\tat org.elasticsearch.action.search.SearchActionListener.onResponse(SearchActionListener.java:45)\n\tat org.elasticsearch.action.search.SearchActionListener.onResponse(SearchActionListener.java:29)\n\tat org.elasticsearch.action.search.SearchExecutionStatsCollector.onResponse(SearchExecutionStatsCollector.java:68)\n\tat org.elasticsearch.action.search.SearchExecutionStatsCollector.onResponse(SearchExecutionStatsCollector.java:36)\n\tat org.elasticsearch.action.ActionListenerResponseHandler.handleResponse(ActionListenerResponseHandler.java:54)\n\tat org.elasticsearch.action.search.SearchTransportService$ConnectionCountingHandler.handleResponse(SearchTransportService.java:432)\n\tat org.elasticsearch.transport.TransportService$ContextRestoreResponseHandler.handleResponse(TransportService.java:1101)\n\tat org.elasticsearch.transport.InboundHandler$1.doRun(InboundHandler.java:224)\n\tat org.elasticsearch.common.util.concurrent.AbstractRunnable.run(AbstractRunnable.java:37)\n\tat org.elasticsearch.common.util.concurrent.EsExecutors$DirectExecutorService.execute(EsExecutors.java:193)\n\tat org.elasticsearch.transport.InboundHandler.handleResponse(InboundHandler.java:216)\n\tat org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:141)\n\tat org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:105)\n\tat org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:660)\n\tat org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:323)\n\tat io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:297)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352)\n\tat io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374)\n\tat io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360)\n\tat io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930)\n\tat io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:682)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:582)\n\tat io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:536)\n\tat io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496)\n\tat io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:906)\n\tat io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74)\n\tat java.base/java.lang.Thread.run(Thread.java:835)\n}" |     | at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) |     | at org.junit.Assert.assertThat(Assert.java:956) |     | at org.elasticsearch.test.rest.yaml.section.DoSection.execute(DoSection.java:308) |     | at org.elasticsearch.test.rest.yaml.ESClientYamlSuiteTestCase.executeSection(ESClientYamlSuiteTestCase.java:398) |     | ... 37 more |     | REPRODUCE WITH: ./gradlew ':qa:mixed-cluster:v7.2.0#mixedClusterTest' --tests "org.elasticsearch.backwards.MixedClusterClientYamlTestSuiteIT" -Dtests.method="test {p0=search.aggregation/240_max_buckets/Max bucket}" -Dtests.seed=D1E3CB4F243CA3C9 -Dtests.security.manager=true -Dtests.locale=fr-FR -Dtests.timezone=America/Regina -Druntime.java=8 |     |   ``` Not sure if this is related, but it looks wrong too: ``` » WARN ][r.suppressed ] [v7.2.0-0] path: _index_template/*, params: {name=*} » org.elasticsearch.transport.RemoteTransportException: [v7.2.0-2][127.0.0.1:45944][indices:admin/index_template/delete] » Caused by: org.elasticsearch.transport.ActionNotFoundTransportException: No handler for action [indices:admin/index_template/delete] » at org.elasticsearch.transport.InboundHandler.handleRequest(InboundHandler.java:169) ~[elasticsearch-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at org.elasticsearch.transport.InboundHandler.messageReceived(InboundHandler.java:121) ~[elasticsearch-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at org.elasticsearch.transport.InboundHandler.inboundMessage(InboundHandler.java:105) ~[elasticsearch-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at org.elasticsearch.transport.TcpTransport.inboundMessage(TcpTransport.java:660) ~[elasticsearch-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at org.elasticsearch.transport.netty4.Netty4MessageChannelHandler.channelRead(Netty4MessageChannelHandler.java:62) ~[transport-netty4-client-7.9.0-SNAPSHOT.jar:7.9.0-SNAPSHOT] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.handler.codec.ByteToMessageDecoder.fireChannelRead(ByteToMessageDecoder.java:323) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final] » at io.netty.handler.codec.ByteToMessageDecoder.channelRead(ByteToMessageDecoder.java:297) ~[netty-codec-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.handler.logging.LoggingHandler.channelRead(LoggingHandler.java:241) ~[netty-handler-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.fireChannelRead(AbstractChannelHandlerContext.java:352) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.DefaultChannelPipeline$HeadContext.channelRead(DefaultChannelPipeline.java:1408) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:374) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.AbstractChannelHandlerContext.invokeChannelRead(AbstractChannelHandlerContext.java:360) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.DefaultChannelPipeline.fireChannelRead(DefaultChannelPipeline.java:930) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.AbstractNioByteChannel$NioByteUnsafe.read(AbstractNioByteChannel.java:163) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.NioEventLoop.processSelectedKey(NioEventLoop.java:682) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.NioEventLoop.processSelectedKeysPlain(NioEventLoop.java:582) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.NioEventLoop.processSelectedKeys(NioEventLoop.java:536) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.channel.nio.NioEventLoop.run(NioEventLoop.java:496) ~[netty-transport-4.1.49.Final.jar:4.1.49.Final] » at io.netty.util.concurrent.SingleThreadEventExecutor$5.run(SingleThreadEventExecutor.java:906) ~[?:?] » at io.netty.util.internal.ThreadExecutorMap$2.run(ThreadExecutorMap.java:74) [netty-common-4.1.49.Final.jar:4.1.49.Final] » at java.lang.Thread.run(Thread.java:835) [?:1.8.0_241] ```
test
mixedclusterclientyamltestsuiteit fails with partial shards failure instead of trying to create too many buckets build scan repro line gradlew qa mixed cluster mixedclustertest tests org elasticsearch backwards mixedclusterclientyamltestsuiteit dtests method test search aggregation max buckets max bucket dtests seed dtests security manager true dtests locale fr fr dtests timezone america regina druntime java reproduces locally yes applicable branches x failure history nothing obvious in the last week failure excerpt org elasticsearch backwards mixedclusterclientyamltestsuiteit test search aggregation max buckets max bucket failed     java lang assertionerror failure at the error message was expected to match the provided regex but didn t     expected trying to create too many buckets     but was root cause type search phase execution exception reason partial shards failure phase query grouped true failed shards stack trace failed to execute phase partial shards failure n tat org elasticsearch action search abstractsearchasyncaction onphasefailure abstractsearchasyncaction java n tat org elasticsearch action search abstractsearchasyncaction executenextphase abstractsearchasyncaction java n tat org elasticsearch action search abstractsearchasyncaction onphasedone abstractsearchasyncaction java n tat org elasticsearch action search initialsearchphase successfulshardexecution initialsearchphase java n tat org elasticsearch action search initialsearchphase onshardresult initialsearchphase java n tat org elasticsearch action search initialsearchphase access initialsearchphase java n tat org elasticsearch action search initialsearchphase inneronresponse initialsearchphase java n tat org elasticsearch action search searchactionlistener onresponse searchactionlistener java n tat org elasticsearch action search searchactionlistener onresponse searchactionlistener java n tat org elasticsearch action search searchexecutionstatscollector onresponse searchexecutionstatscollector java n tat org elasticsearch action search searchexecutionstatscollector onresponse searchexecutionstatscollector java n tat org elasticsearch action actionlistenerresponsehandler handleresponse actionlistenerresponsehandler java n tat org elasticsearch action search searchtransportservice connectioncountinghandler handleresponse searchtransportservice java n tat org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch transport inboundhandler dorun inboundhandler java n tat org elasticsearch common util concurrent abstractrunnable run abstractrunnable java n tat org elasticsearch common util concurrent esexecutors directexecutorservice execute esexecutors java n tat org elasticsearch transport inboundhandler handleresponse inboundhandler java n tat org elasticsearch transport inboundhandler messagereceived inboundhandler java n tat org elasticsearch transport inboundhandler inboundmessage inboundhandler java n tat org elasticsearch transport tcptransport inboundmessage tcptransport java n tat org elasticsearch transport channelread java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java n tat io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler logging logginghandler channelread logginghandler java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java n tat io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java n tat io netty channel nio nioeventloop processselectedkey nioeventloop java n tat io netty channel nio nioeventloop processselectedkeysplain nioeventloop java n tat io netty channel nio nioeventloop processselectedkeys nioeventloop java n tat io netty channel nio nioeventloop run nioeventloop java n tat io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java n tat io netty util internal threadexecutormap run threadexecutormap java n tat java base java lang thread run thread java n     at randomizedtesting seedinfo seed     at org elasticsearch test rest yaml esclientyamlsuitetestcase executesection esclientyamlsuitetestcase java     at org elasticsearch test rest yaml esclientyamlsuitetestcase test esclientyamlsuitetestcase java     at sun reflect invoke unknown source     at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java     at java lang reflect method invoke method java     at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java     at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java     at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java     at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java     at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java     at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java     at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java     at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java     at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java     at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java     at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java     at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java     at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java     at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java     at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java     at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java     at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java     at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java     at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java     at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java     at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java     at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java     at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java     at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java     at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java     at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java     at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java     at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java     at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java     at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java     at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java     at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java     at java lang thread run thread java           caused by     java lang assertionerror the error message was expected to match the provided regex but didn t     expected trying to create too many buckets     but was root cause type search phase execution exception reason partial shards failure phase query grouped true failed shards stack trace failed to execute phase partial shards failure n tat org elasticsearch action search abstractsearchasyncaction onphasefailure abstractsearchasyncaction java n tat org elasticsearch action search abstractsearchasyncaction executenextphase abstractsearchasyncaction java n tat org elasticsearch action search abstractsearchasyncaction onphasedone abstractsearchasyncaction java n tat org elasticsearch action search initialsearchphase successfulshardexecution initialsearchphase java n tat org elasticsearch action search initialsearchphase onshardresult initialsearchphase java n tat org elasticsearch action search initialsearchphase access initialsearchphase java n tat org elasticsearch action search initialsearchphase inneronresponse initialsearchphase java n tat org elasticsearch action search searchactionlistener onresponse searchactionlistener java n tat org elasticsearch action search searchactionlistener onresponse searchactionlistener java n tat org elasticsearch action search searchexecutionstatscollector onresponse searchexecutionstatscollector java n tat org elasticsearch action search searchexecutionstatscollector onresponse searchexecutionstatscollector java n tat org elasticsearch action actionlistenerresponsehandler handleresponse actionlistenerresponsehandler java n tat org elasticsearch action search searchtransportservice connectioncountinghandler handleresponse searchtransportservice java n tat org elasticsearch transport transportservice contextrestoreresponsehandler handleresponse transportservice java n tat org elasticsearch transport inboundhandler dorun inboundhandler java n tat org elasticsearch common util concurrent abstractrunnable run abstractrunnable java n tat org elasticsearch common util concurrent esexecutors directexecutorservice execute esexecutors java n tat org elasticsearch transport inboundhandler handleresponse inboundhandler java n tat org elasticsearch transport inboundhandler messagereceived inboundhandler java n tat org elasticsearch transport inboundhandler inboundmessage inboundhandler java n tat org elasticsearch transport tcptransport inboundmessage tcptransport java n tat org elasticsearch transport channelread java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java n tat io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty handler logging logginghandler channelread logginghandler java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java n tat io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java n tat io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java n tat io netty channel nio nioeventloop processselectedkey nioeventloop java n tat io netty channel nio nioeventloop processselectedkeysplain nioeventloop java n tat io netty channel nio nioeventloop processselectedkeys nioeventloop java n tat io netty channel nio nioeventloop run nioeventloop java n tat io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java n tat io netty util internal threadexecutormap run threadexecutormap java n tat java base java lang thread run thread java n     at org hamcrest matcherassert assertthat matcherassert java     at org junit assert assertthat assert java     at org elasticsearch test rest yaml section dosection execute dosection java     at org elasticsearch test rest yaml esclientyamlsuitetestcase executesection esclientyamlsuitetestcase java     more     reproduce with gradlew qa mixed cluster mixedclustertest tests org elasticsearch backwards mixedclusterclientyamltestsuiteit dtests method test search aggregation max buckets max bucket dtests seed dtests security manager true dtests locale fr fr dtests timezone america regina druntime java       not sure if this is related but it looks wrong too » warn path index template params name » org elasticsearch transport remotetransportexception » caused by org elasticsearch transport actionnotfoundtransportexception no handler for action » at org elasticsearch transport inboundhandler handlerequest inboundhandler java » at org elasticsearch transport inboundhandler messagereceived inboundhandler java » at org elasticsearch transport inboundhandler inboundmessage inboundhandler java » at org elasticsearch transport tcptransport inboundmessage tcptransport java » at org elasticsearch transport channelread java » at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java » at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java » at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java » at io netty handler codec bytetomessagedecoder firechannelread bytetomessagedecoder java » at io netty handler codec bytetomessagedecoder channelread bytetomessagedecoder java » at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java » at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java » at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java » at io netty handler logging logginghandler channelread logginghandler java » at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java » at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java » at io netty channel abstractchannelhandlercontext firechannelread abstractchannelhandlercontext java » at io netty channel defaultchannelpipeline headcontext channelread defaultchannelpipeline java » at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java » at io netty channel abstractchannelhandlercontext invokechannelread abstractchannelhandlercontext java » at io netty channel defaultchannelpipeline firechannelread defaultchannelpipeline java » at io netty channel nio abstractniobytechannel niobyteunsafe read abstractniobytechannel java » at io netty channel nio nioeventloop processselectedkey nioeventloop java » at io netty channel nio nioeventloop processselectedkeysplain nioeventloop java » at io netty channel nio nioeventloop processselectedkeys nioeventloop java » at io netty channel nio nioeventloop run nioeventloop java » at io netty util concurrent singlethreadeventexecutor run singlethreadeventexecutor java » at io netty util internal threadexecutormap run threadexecutormap java » at java lang thread run thread java
1
344,568
30,751,699,921
IssuesEvent
2023-07-28 19:55:57
saltstack/salt
https://api.github.com/repos/saltstack/salt
opened
[Increase Test Coverage] Batch 128
Tests
Increase the code coverage percent on the following files to at least 80%. Please be aware that currently the percentage might be inaccurate if the module uses salt due to #64696 File | Percent salt/output/overstatestage.py 12 salt/output/progress.py 34 salt/pillar/cmd_json.py 40 salt/pillar/cmd_yaml.py 72 salt/pillar/cmd_yamlex.py 40
1.0
[Increase Test Coverage] Batch 128 - Increase the code coverage percent on the following files to at least 80%. Please be aware that currently the percentage might be inaccurate if the module uses salt due to #64696 File | Percent salt/output/overstatestage.py 12 salt/output/progress.py 34 salt/pillar/cmd_json.py 40 salt/pillar/cmd_yaml.py 72 salt/pillar/cmd_yamlex.py 40
test
batch increase the code coverage percent on the following files to at least please be aware that currently the percentage might be inaccurate if the module uses salt due to file percent salt output overstatestage py salt output progress py salt pillar cmd json py salt pillar cmd yaml py salt pillar cmd yamlex py
1
219,080
17,062,842,623
IssuesEvent
2021-07-07 00:57:54
ClickHouse/ClickHouse
https://api.github.com/repos/ClickHouse/ClickHouse
closed
Evaluate Coderrect
help wanted testing
Idea suggested by @myrrc https://coderrect.com/documentation/ Subscribe to their open-source program: https://coderrect.com/openscan/
1.0
Evaluate Coderrect - Idea suggested by @myrrc https://coderrect.com/documentation/ Subscribe to their open-source program: https://coderrect.com/openscan/
test
evaluate coderrect idea suggested by myrrc subscribe to their open source program
1
214,921
7,279,515,922
IssuesEvent
2018-02-22 05:02:17
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
codepen.io - design is broken
browser-firefox priority-important
<!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: web --> **URL**: https://codepen.io/lassediercks/pen/NywMxp **Browser / Version**: Firefox 58.0 **Operating System**: Mac OS X 10.12 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: The padding-top: 100%; causes the height to explode **Steps to Reproduce**: The Codepen url contains the bug in an isolated environment. _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
codepen.io - design is broken - <!-- @browser: Firefox 58.0 --> <!-- @ua_header: Mozilla/5.0 (Macintosh; Intel Mac OS X 10.12; rv:58.0) Gecko/20100101 Firefox/58.0 --> <!-- @reported_with: web --> **URL**: https://codepen.io/lassediercks/pen/NywMxp **Browser / Version**: Firefox 58.0 **Operating System**: Mac OS X 10.12 **Tested Another Browser**: Yes **Problem type**: Design is broken **Description**: The padding-top: 100%; causes the height to explode **Steps to Reproduce**: The Codepen url contains the bug in an isolated environment. _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
codepen io design is broken url browser version firefox operating system mac os x tested another browser yes problem type design is broken description the padding top causes the height to explode steps to reproduce the codepen url contains the bug in an isolated environment from with ❤️
0
183,257
14,219,015,405
IssuesEvent
2020-11-17 12:41:07
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: exports a "ip_range" list from an uploaded file - value lists management modal export list types exports a "ip_range" list from an uploaded file
failed-test
A test failed on a tracked branch ``` AssertionError: expected '' to include '192.168.100.14/24' at Context.eval (http://localhost:6131/__cypress/tests?p=cypress/integration/value_lists.spec.ts:13648:53) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/9722/) <!-- kibanaCiData = {"failed-test":{"test.class":"exports a \"ip_range\" list from an uploaded file","test.name":"value lists management modal export list types exports a \"ip_range\" list from an uploaded file","test.failCount":1}} -->
1.0
Failing test: exports a "ip_range" list from an uploaded file - value lists management modal export list types exports a "ip_range" list from an uploaded file - A test failed on a tracked branch ``` AssertionError: expected '' to include '192.168.100.14/24' at Context.eval (http://localhost:6131/__cypress/tests?p=cypress/integration/value_lists.spec.ts:13648:53) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/9722/) <!-- kibanaCiData = {"failed-test":{"test.class":"exports a \"ip_range\" list from an uploaded file","test.name":"value lists management modal export list types exports a \"ip_range\" list from an uploaded file","test.failCount":1}} -->
test
failing test exports a ip range list from an uploaded file value lists management modal export list types exports a ip range list from an uploaded file a test failed on a tracked branch assertionerror expected to include at context eval first failure
1
145,981
5,591,724,498
IssuesEvent
2017-03-30 01:01:44
ywwg/ffagc
https://api.github.com/repos/ywwg/ffagc
closed
"show scores" is not sorting by score, but only on the prod server :(
bug priority:high
No idea why. rails versioning issue? ruby versioning issue?
1.0
"show scores" is not sorting by score, but only on the prod server :( - No idea why. rails versioning issue? ruby versioning issue?
non_test
show scores is not sorting by score but only on the prod server no idea why rails versioning issue ruby versioning issue
0
142,626
11,488,089,195
IssuesEvent
2020-02-11 13:17:20
joeyfrog/hooktest
https://api.github.com/repos/joeyfrog/hooktest
closed
[XRAY] Vulnerability in artifact: jjjjjjjjjjjjjjjjjjj
test xray
This is an automated issue made via XRAY Github webhook. The deployed artifact **jjjjjjjjjjjjjjjjjjj** contains the following vaulnerable dependencies: aopalliance-repackaged-2.4.0-b09.jar Here is the sent JSON from XRAY: [ { "depth": 0, "display_name": "test:6639", "infected_files": [ { "depth": 0, "display_name": "ant-1.9.4.jar", "name": "ant-1.9.4.jar", "parent_sha": "c9be3f74c49d2f3ea273de9c9e172ea99be696d995f31876d43185113bbe91bb", "path": "", "pkg_type": "Generic", "sha256": "649ae0730251de07b8913f49286d46bba7b92d47c5f332610aa426c4f02161d8" }, { "depth": 0, "display_name": "org.glassfish.hk2.external:aopalliance-repackaged:2.4.0-b09", "name": "aopalliance-repackaged-2.4.0-b09.jar", "parent_sha": "c9be3f74c49d2f3ea273de9c9e172ea99be696d995f31876d43185113bbe91bb", "path": "", "pkg_type": "Maven", "sha256": "a97667a617fa5d427c2e95ce6f3eab5cf2d21d00c69ad2a7524ff6d9a9144f58" } ], "name": "jjjjjjjjjjjjjjjjjjj", "parent_sha": "c9be3f74c49d2f3ea273de9c9e172ea99be696d995f31876d43185113bbe91bb", "path": "artifactory-xray/builds/", "pkg_type": "Build", "sha1": "737145943754ac99a678d366269dcafc205233ba", "sha256": "c9be3f74c49d2f3ea273de9c9e172ea99be696d995f31876d43185113bbe91bb" } ]
1.0
[XRAY] Vulnerability in artifact: jjjjjjjjjjjjjjjjjjj - This is an automated issue made via XRAY Github webhook. The deployed artifact **jjjjjjjjjjjjjjjjjjj** contains the following vaulnerable dependencies: aopalliance-repackaged-2.4.0-b09.jar Here is the sent JSON from XRAY: [ { "depth": 0, "display_name": "test:6639", "infected_files": [ { "depth": 0, "display_name": "ant-1.9.4.jar", "name": "ant-1.9.4.jar", "parent_sha": "c9be3f74c49d2f3ea273de9c9e172ea99be696d995f31876d43185113bbe91bb", "path": "", "pkg_type": "Generic", "sha256": "649ae0730251de07b8913f49286d46bba7b92d47c5f332610aa426c4f02161d8" }, { "depth": 0, "display_name": "org.glassfish.hk2.external:aopalliance-repackaged:2.4.0-b09", "name": "aopalliance-repackaged-2.4.0-b09.jar", "parent_sha": "c9be3f74c49d2f3ea273de9c9e172ea99be696d995f31876d43185113bbe91bb", "path": "", "pkg_type": "Maven", "sha256": "a97667a617fa5d427c2e95ce6f3eab5cf2d21d00c69ad2a7524ff6d9a9144f58" } ], "name": "jjjjjjjjjjjjjjjjjjj", "parent_sha": "c9be3f74c49d2f3ea273de9c9e172ea99be696d995f31876d43185113bbe91bb", "path": "artifactory-xray/builds/", "pkg_type": "Build", "sha1": "737145943754ac99a678d366269dcafc205233ba", "sha256": "c9be3f74c49d2f3ea273de9c9e172ea99be696d995f31876d43185113bbe91bb" } ]
test
vulnerability in artifact jjjjjjjjjjjjjjjjjjj this is an automated issue made via xray github webhook the deployed artifact jjjjjjjjjjjjjjjjjjj contains the following vaulnerable dependencies aopalliance repackaged jar here is the sent json from xray depth display name test infected files depth display name ant jar name ant jar parent sha path pkg type generic depth display name org glassfish external aopalliance repackaged name aopalliance repackaged jar parent sha path pkg type maven name jjjjjjjjjjjjjjjjjjj parent sha path artifactory xray builds pkg type build
1
338,499
30,302,728,857
IssuesEvent
2023-07-10 07:22:58
ray-project/ray
https://api.github.com/repos/ray-project/ray
closed
Release test tune_cloud_durable_upload_rllib_trainer.aws failed
bug P0 release-test ml unstable-release-test
Release test **tune_cloud_durable_upload_rllib_trainer.aws** failed. See https://buildkite.com/ray-project/release-tests-branch/builds/1873#01892f3a-37ef-4da3-ad2d-c6a6869d6470 for more details. Managed by OSS Test Policy
2.0
Release test tune_cloud_durable_upload_rllib_trainer.aws failed - Release test **tune_cloud_durable_upload_rllib_trainer.aws** failed. See https://buildkite.com/ray-project/release-tests-branch/builds/1873#01892f3a-37ef-4da3-ad2d-c6a6869d6470 for more details. Managed by OSS Test Policy
test
release test tune cloud durable upload rllib trainer aws failed release test tune cloud durable upload rllib trainer aws failed see for more details managed by oss test policy
1
37,886
5,147,513,457
IssuesEvent
2017-01-13 07:41:52
dotnet/corefx
https://api.github.com/repos/dotnet/corefx
reopened
Test failure: System.IO.Tests.WaitForChangedTests/CreatedDeleted_Success(changeType: Created)
area-System.IO test bug test-run-desktop
Opened on behalf of @jiangzeng The test `System.IO.Tests.WaitForChangedTests/CreatedDeleted_Success(changeType: Created)` has failed. Assert.Equal() Failure\r Expected: Created\r Actual: 0 Stack Trace: at System.IO.Tests.WaitForChangedTests.CreatedDeleted_Success(WatcherChangeTypes changeType) in D:\A\_work\26\s\corefx\src\System.IO.FileSystem.Watcher\tests\FileSystemWatcher.WaitForChanged.cs:line 149 Build : Master - 20161102.03 (Full Framework Tests) Failing configurations: - Windows.10.Amd64 - AnyCPU-Release Details: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fdesktop~2Fcli~2F/build/20161102.03/workItem/System.IO.FileSystem.Watcher.Tests/analysis/xunit/System.IO.Tests.WaitForChangedTests~2FCreatedDeleted_Success(changeType:%20Created)
2.0
Test failure: System.IO.Tests.WaitForChangedTests/CreatedDeleted_Success(changeType: Created) - Opened on behalf of @jiangzeng The test `System.IO.Tests.WaitForChangedTests/CreatedDeleted_Success(changeType: Created)` has failed. Assert.Equal() Failure\r Expected: Created\r Actual: 0 Stack Trace: at System.IO.Tests.WaitForChangedTests.CreatedDeleted_Success(WatcherChangeTypes changeType) in D:\A\_work\26\s\corefx\src\System.IO.FileSystem.Watcher\tests\FileSystemWatcher.WaitForChanged.cs:line 149 Build : Master - 20161102.03 (Full Framework Tests) Failing configurations: - Windows.10.Amd64 - AnyCPU-Release Details: https://mc.dot.net/#/product/netcore/master/source/official~2Fcorefx~2Fmaster~2F/type/test~2Ffunctional~2Fdesktop~2Fcli~2F/build/20161102.03/workItem/System.IO.FileSystem.Watcher.Tests/analysis/xunit/System.IO.Tests.WaitForChangedTests~2FCreatedDeleted_Success(changeType:%20Created)
test
test failure system io tests waitforchangedtests createddeleted success changetype created opened on behalf of jiangzeng the test system io tests waitforchangedtests createddeleted success changetype created has failed assert equal failure r expected created r actual stack trace at system io tests waitforchangedtests createddeleted success watcherchangetypes changetype in d a work s corefx src system io filesystem watcher tests filesystemwatcher waitforchanged cs line build master full framework tests failing configurations windows anycpu release details
1
136,678
11,072,417,934
IssuesEvent
2019-12-12 10:15:30
microsoft/AzureStorageExplorer
https://api.github.com/repos/microsoft/AzureStorageExplorer
opened
The popped dialog title isn't localized when downloading one disk
:gear: disks 🌐 localization 🧪 testing
**Storage Explorer Version:** 1.11.1 **Build:** [20191212.4](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3315517) **Branch:** master **Language**: Chinese(zh-CN) / Chinese(zh-TW) **Platform/OS:** Windows 10/ Linux Ubuntu 18.04/macOS High Sierra **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Launch Storage Explorer. 2. Open 'Settings' -> Application (Regional Settings) -> Select '中文(简体)' -> Restart Storage Explorer. 3. Expand one resource group -> Try to download one disk. 4. Check the popped dialog. **Expect Experience:** The popped dialog title is localized. **Actual Experience:** The popped dialog title isn't localized. ![image](https://user-images.githubusercontent.com/54055206/70700730-23737580-1d06-11ea-83a3-dcbcb9b096bd.png)
1.0
The popped dialog title isn't localized when downloading one disk - **Storage Explorer Version:** 1.11.1 **Build:** [20191212.4](https://devdiv.visualstudio.com/DevDiv/_build/results?buildId=3315517) **Branch:** master **Language**: Chinese(zh-CN) / Chinese(zh-TW) **Platform/OS:** Windows 10/ Linux Ubuntu 18.04/macOS High Sierra **Architecture:** ia32/x64 **Regression From:** Not a regression **Steps to reproduce:** 1. Launch Storage Explorer. 2. Open 'Settings' -> Application (Regional Settings) -> Select '中文(简体)' -> Restart Storage Explorer. 3. Expand one resource group -> Try to download one disk. 4. Check the popped dialog. **Expect Experience:** The popped dialog title is localized. **Actual Experience:** The popped dialog title isn't localized. ![image](https://user-images.githubusercontent.com/54055206/70700730-23737580-1d06-11ea-83a3-dcbcb9b096bd.png)
test
the popped dialog title isn t localized when downloading one disk storage explorer version build branch master language chinese zh cn chinese zh tw platform os windows linux ubuntu macos high sierra architecture regression from not a regression steps to reproduce launch storage explorer open settings application regional settings select 中文 简体 restart storage explorer expand one resource group try to download one disk check the popped dialog expect experience the popped dialog title is localized actual experience the popped dialog title isn t localized
1
120,602
10,130,067,729
IssuesEvent
2019-08-01 16:03:51
Matte23/server-status-page
https://api.github.com/repos/Matte23/server-status-page
closed
Implement checks for configuration passed to test's execute method
enhancement test
**Feature description** Implement checks for configuration passed to test's execute method, to prevent difficult configuration debug and unexpected behaviour.
1.0
Implement checks for configuration passed to test's execute method - **Feature description** Implement checks for configuration passed to test's execute method, to prevent difficult configuration debug and unexpected behaviour.
test
implement checks for configuration passed to test s execute method feature description implement checks for configuration passed to test s execute method to prevent difficult configuration debug and unexpected behaviour
1
105,692
9,100,203,334
IssuesEvent
2019-02-20 07:48:38
humera987/FXLabs-Test-Automation
https://api.github.com/repos/humera987/FXLabs-Test-Automation
opened
Test : ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue
test
Project : Test Job : Default Env : Default Category : Negative Tags : No tags Severity : Major Region : US_WEST Result : fail Status Code : 500 Headers : {} Endpoint : http://13.56.210.25/api/v1/api/v1/runs/job/{id}/runNo/null?nav=FgdOPpwM Request : Response : Not enough variable values available to expand 'id' Logs : 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : URL [http://13.56.210.25/api/v1/api/v1/runs/job/{id}/runNo/null?nav=FgdOPpwM] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Method [GET] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Request [] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Request-Headers [{Content-Type=[application/json], Accept=[application/json], Authorization=[Basic SHVtZXJhLy9odW1lcmFAZnhsYWJzLmlvOmh1bWVyYTEyMyQ=]}] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Response [Not enough variable values available to expand 'id'] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Response-Headers [{}] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : StatusCode [500] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Time [3] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Size [51] 2019-02-20 07:48:29 INFO [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Assertion [@StatusCode != 401] resolved-to [500 != 401] result [Passed] 2019-02-20 07:48:29 ERROR [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Assertion [@StatusCode != 500] resolved-to [500 != 500] result [Failed] 2019-02-20 07:48:29 INFO [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Assertion [@StatusCode != 200] resolved-to [500 != 200] result [Passed] 2019-02-20 07:48:29 INFO [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Assertion [@StatusCode != 404] resolved-to [500 != 404] result [Passed] --- FX Bot ---
1.0
Test : ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue - Project : Test Job : Default Env : Default Category : Negative Tags : No tags Severity : Major Region : US_WEST Result : fail Status Code : 500 Headers : {} Endpoint : http://13.56.210.25/api/v1/api/v1/runs/job/{id}/runNo/null?nav=FgdOPpwM Request : Response : Not enough variable values available to expand 'id' Logs : 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : URL [http://13.56.210.25/api/v1/api/v1/runs/job/{id}/runNo/null?nav=FgdOPpwM] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Method [GET] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Request [] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Request-Headers [{Content-Type=[application/json], Accept=[application/json], Authorization=[Basic SHVtZXJhLy9odW1lcmFAZnhsYWJzLmlvOmh1bWVyYTEyMyQ=]}] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Response [Not enough variable values available to expand 'id'] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Response-Headers [{}] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : StatusCode [500] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Time [3] 2019-02-20 07:48:29 DEBUG [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Size [51] 2019-02-20 07:48:29 INFO [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Assertion [@StatusCode != 401] resolved-to [500 != 401] result [Passed] 2019-02-20 07:48:29 ERROR [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Assertion [@StatusCode != 500] resolved-to [500 != 500] result [Failed] 2019-02-20 07:48:29 INFO [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Assertion [@StatusCode != 200] resolved-to [500 != 200] result [Passed] 2019-02-20 07:48:29 INFO [ApiV1RunsJobIdRunnoRunnoGetPathParamRunnoNullValue] : Assertion [@StatusCode != 404] resolved-to [500 != 404] result [Passed] --- FX Bot ---
test
test project test job default env default category negative tags no tags severity major region us west result fail status code headers endpoint request response not enough variable values available to expand id logs debug url debug method debug request debug request headers accept authorization debug response debug response headers debug statuscode debug time debug size info assertion resolved to result error assertion resolved to result info assertion resolved to result info assertion resolved to result fx bot
1
471,363
13,565,545,341
IssuesEvent
2020-09-18 11:54:06
arfc/fhr-benchmark
https://api.github.com/repos/arfc/fhr-benchmark
opened
FHR Benchmark Phase 1 wrap up
Comp:Output Difficulty:1-Beginner Priority:2-Normal Status:4-In Progress
This issue can be closed when the latest changes to the fhr benchmark phase 1 results are PRed into this repository. These results reflect what was submitted for the M&C paper.
1.0
FHR Benchmark Phase 1 wrap up - This issue can be closed when the latest changes to the fhr benchmark phase 1 results are PRed into this repository. These results reflect what was submitted for the M&C paper.
non_test
fhr benchmark phase wrap up this issue can be closed when the latest changes to the fhr benchmark phase results are pred into this repository these results reflect what was submitted for the m c paper
0
552,762
16,326,299,157
IssuesEvent
2021-05-12 01:27:07
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
media.transcoder.job_test: test_create_job_from_template failed
api: mediatranslation flakybot: flaky flakybot: issue priority: p1 samples type: bug
Note: #5401 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 045f999a22ce8aa2a4f5346b2252e48e8ca4e929 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2a4940f9-0ee5-4c97-bf01-efed786f033d), [Sponge](http://sponge2/2a4940f9-0ee5-4c97-bf01-efed786f033d) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 73, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNKNOWN details = "Unknown Error." debug_error_string = "{"created":"@1620565201.182027262","description":"Error received from peer ipv4:172.253.117.95:443","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Unknown Error.","grpc_status":2}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/media/transcoder/job_test.py", line 147, in test_create_job_from_template _assert_job_state_succeeded(capsys, job_id) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/backoff/_sync.py", line 94, in retry ret = target(*args, **kwargs) File "/workspace/media/transcoder/job_test.py", line 366, in _assert_job_state_succeeded get_job_state.get_job_state(project_id, location, job_id) File "/workspace/media/transcoder/get_job_state.py", line 43, in get_job_state response = client.get_job(name=name) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/video/transcoder_v1beta1/services/transcoder_service/client.py", line 605, in get_job response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 75, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.Unknown: None Unknown Error.</pre></details>
1.0
media.transcoder.job_test: test_create_job_from_template failed - Note: #5401 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 045f999a22ce8aa2a4f5346b2252e48e8ca4e929 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/2a4940f9-0ee5-4c97-bf01-efed786f033d), [Sponge](http://sponge2/2a4940f9-0ee5-4c97-bf01-efed786f033d) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 73, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.UNKNOWN details = "Unknown Error." debug_error_string = "{"created":"@1620565201.182027262","description":"Error received from peer ipv4:172.253.117.95:443","file":"src/core/lib/surface/call.cc","file_line":1067,"grpc_message":"Unknown Error.","grpc_status":2}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/media/transcoder/job_test.py", line 147, in test_create_job_from_template _assert_job_state_succeeded(capsys, job_id) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/backoff/_sync.py", line 94, in retry ret = target(*args, **kwargs) File "/workspace/media/transcoder/job_test.py", line 366, in _assert_job_state_succeeded get_job_state.get_job_state(project_id, location, job_id) File "/workspace/media/transcoder/get_job_state.py", line 43, in get_job_state response = client.get_job(name=name) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/video/transcoder_v1beta1/services/transcoder_service/client.py", line 605, in get_job response = rpc(request, retry=retry, timeout=timeout, metadata=metadata,) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/gapic_v1/method.py", line 145, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/media/transcoder/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 75, in error_remapped_callable six.raise_from(exceptions.from_grpc_error(exc), exc) File "<string>", line 3, in raise_from google.api_core.exceptions.Unknown: None Unknown Error.</pre></details>
non_test
media transcoder job test test create job from template failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace media transcoder nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace media transcoder nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace media transcoder nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode unknown details unknown error debug error string created description error received from peer file src core lib surface call cc file line grpc message unknown error grpc status the above exception was the direct cause of the following exception traceback most recent call last file workspace media transcoder job test py line in test create job from template assert job state succeeded capsys job id file workspace media transcoder nox py lib site packages backoff sync py line in retry ret target args kwargs file workspace media transcoder job test py line in assert job state succeeded get job state get job state project id location job id file workspace media transcoder get job state py line in get job state response client get job name name file workspace media transcoder nox py lib site packages google cloud video transcoder services transcoder service client py line in get job response rpc request retry retry timeout timeout metadata metadata file workspace media transcoder nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace media transcoder nox py lib site packages google api core grpc helpers py line in error remapped callable six raise from exceptions from grpc error exc exc file line in raise from google api core exceptions unknown none unknown error
0
56,011
6,498,151,241
IssuesEvent
2017-08-22 16:17:41
openbmc/openbmc-test-automation
https://api.github.com/repos/openbmc/openbmc-test-automation
closed
Systest: inventory ignore list should be a parameter
SYS_TEST Test
For htx_hardbootme_test systest, the inventory items to ignore should be a parameter. This would allow testers to choose what to ignore on the particular system they are testing. Currently the ignore_list is only processor:speed.
2.0
Systest: inventory ignore list should be a parameter - For htx_hardbootme_test systest, the inventory items to ignore should be a parameter. This would allow testers to choose what to ignore on the particular system they are testing. Currently the ignore_list is only processor:speed.
test
systest inventory ignore list should be a parameter for htx hardbootme test systest the inventory items to ignore should be a parameter this would allow testers to choose what to ignore on the particular system they are testing currently the ignore list is only processor speed
1
327,311
28,052,363,307
IssuesEvent
2023-03-29 07:00:39
ALTA-LapakUMKM-Group-2/LapakUMKM-APITesting
https://api.github.com/repos/ALTA-LapakUMKM-Group-2/LapakUMKM-APITesting
closed
[Products-A025]PUT update a product invalid json request body
Manual Api Testing
"Given PUT update a product invalid json request body When Send update a product parameter Then Status code should be 400 Bad Request And Validate update a product invalid req body resources json schema"
1.0
[Products-A025]PUT update a product invalid json request body - "Given PUT update a product invalid json request body When Send update a product parameter Then Status code should be 400 Bad Request And Validate update a product invalid req body resources json schema"
test
put update a product invalid json request body given put update a product invalid json request body when send update a product parameter then status code should be bad request and validate update a product invalid req body resources json schema
1
102,741
11,306,484,476
IssuesEvent
2020-01-18 14:27:55
SPGoding/datapack-language-server
https://api.github.com/repos/SPGoding/datapack-language-server
opened
Move cache files to `context.storagePath`
:lipstick: UI :pencil: documentation
Move cache files to [`context.storagePath`](https://code.visualstudio.com/api/extension-capabilities/common-capabilities#data-storage). This means that the ugly `.datapack` folder in you workspace will no long be needed.
1.0
Move cache files to `context.storagePath` - Move cache files to [`context.storagePath`](https://code.visualstudio.com/api/extension-capabilities/common-capabilities#data-storage). This means that the ugly `.datapack` folder in you workspace will no long be needed.
non_test
move cache files to context storagepath move cache files to this means that the ugly datapack folder in you workspace will no long be needed
0
159,371
12,474,534,600
IssuesEvent
2020-05-29 09:49:42
ICIJ/datashare
https://api.github.com/repos/ICIJ/datashare
reopened
Put 'Frequency' by default in the Sort By of Batch Search's list of queries
front need testing
At the moment, we have 'default' by default. We would like to have 'frequency' by default. ![Screenshot 2020-04-06 at 15 02 34](https://user-images.githubusercontent.com/17233829/78561298-c1d89900-7817-11ea-9fbc-be51d7e10094.png)
1.0
Put 'Frequency' by default in the Sort By of Batch Search's list of queries - At the moment, we have 'default' by default. We would like to have 'frequency' by default. ![Screenshot 2020-04-06 at 15 02 34](https://user-images.githubusercontent.com/17233829/78561298-c1d89900-7817-11ea-9fbc-be51d7e10094.png)
test
put frequency by default in the sort by of batch search s list of queries at the moment we have default by default we would like to have frequency by default
1
277,828
24,105,666,812
IssuesEvent
2022-09-20 07:13:51
milvus-io/milvus
https://api.github.com/repos/milvus-io/milvus
opened
[Bug]: [benchmark][standalone] Milvus search failed,raise an error"fail to search on all shard leaders, err=fail to Search, QueryNode ID=1, reason=stream operation failed: deadline exceed"
kind/bug priority/urgent needs-triage test/benchmark
### Is there an existing issue for this? - [X] I have searched the existing issues ### Environment ```markdown - Milvus version:master-20220920-5143e2a7 - Deployment mode(standalone or cluster):standalone - SDK version(e.g. pymilvus v2.0.0rc2):2.2.0dev32 - OS(Ubuntu or CentOS): - CPU/Memory: - GPU: - Others: ``` ### Current Behavior server-instance fouram-w5ccc-1 server-configmap server-single-8c16m client-configmap client-acc-sift-flat sever: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES fouram-w5ccc-1-etcd-0 1/1 Running 0 112s 10.104.5.82 4am-node12 <none> <none> fouram-w5ccc-1-milvus-standalone-79c595d867-bvc6h 1/1 Running 0 112s 10.104.6.121 4am-node13 <none> <none> fouram-w5ccc-1-minio-65968fd9-dh5nf 1/1 Running 0 112s 10.104.6.120 4am-node13 <none> <none> ``` client log: ``` 2022-09-20 03:33:52,111] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:276) [2022-09-20 03:33:57,188] [ DEBUG] - Milvus load_collection run in 5.0775s (milvus_benchmark.client:57) [2022-09-20 03:33:57,188] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:279) [2022-09-20 03:33:57,189] [ DEBUG] - Milvus get_query_segment_info run in 0.0008s (milvus_benchmark.client:57) [2022-09-20 03:33:57,189] [ DEBUG] - query_segment_info: [] (milvus_benchmark.runners.accuracy:282) [2022-09-20 03:33:57,191] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-09-20 03:33:58,088] [ ERROR] - RPC error: [search], <MilvusException: (code=1, message=fail to search on all shard leaders, err=fail to Search, QueryNode ID=1, reason=stream operation failed: deadline exceed)>, <Time:{'RPC start': '2022-09-20 03:33:57.191366', 'RPC error': '2022-09-20 03:33:58.088386'}> (pymilvus.decorators:112) [2022-09-20 03:33:58,089] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 292, in run_case self.milvus.query(case_param["vector_query"], filter_query=case_param["filter_query"], File "/src/milvus_benchmark/client.py", line 53, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 346, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/stub.py", line 844, in search return handler.search(collection_name, data, anns_field, param, limit, expression, File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 113, in handler raise e File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 109, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 139, in handler ret = func(self, *args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 89, in handler raise e File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 51, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 472, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 436, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 427, in _execute_search_requests raise MilvusException(response.status.error_code, response.status.reason) pymilvus.exceptions.MilvusException: <MilvusException: (code=1, message=fail to search on all shard leaders, err=fail to Search, QueryNode ID=1, reason=stream operation failed: deadline exceed)> (milvus_benchmark.main:98) ``` ### Expected Behavior _No response_ ### Steps To Reproduce ```markdown 1.create an collection 2.insert sift1m data 3.create flat index, then load 4.search raise error ``` ### Milvus Log _No response_ ### Anything else? client-acc-sift-flat: { "config.yaml": "ann_accuracy: collections: - milvus: cache_config.cpu_cache_capacity: 16GB engine_config.use_blas_threshold: 1100 server: cpus: 12 source_file: /test/milvus/ann_hdf5/sift-128-euclidean.hdf5 collection_name: sift_128_euclidean index_types: ['flat'] index_params: nlist: [1024] top_ks: [10] nqs: [10000] search_params: nprobe: [1] " }
1.0
[Bug]: [benchmark][standalone] Milvus search failed,raise an error"fail to search on all shard leaders, err=fail to Search, QueryNode ID=1, reason=stream operation failed: deadline exceed" - ### Is there an existing issue for this? - [X] I have searched the existing issues ### Environment ```markdown - Milvus version:master-20220920-5143e2a7 - Deployment mode(standalone or cluster):standalone - SDK version(e.g. pymilvus v2.0.0rc2):2.2.0dev32 - OS(Ubuntu or CentOS): - CPU/Memory: - GPU: - Others: ``` ### Current Behavior server-instance fouram-w5ccc-1 server-configmap server-single-8c16m client-configmap client-acc-sift-flat sever: ``` NAME READY STATUS RESTARTS AGE IP NODE NOMINATED NODE READINESS GATES fouram-w5ccc-1-etcd-0 1/1 Running 0 112s 10.104.5.82 4am-node12 <none> <none> fouram-w5ccc-1-milvus-standalone-79c595d867-bvc6h 1/1 Running 0 112s 10.104.6.121 4am-node13 <none> <none> fouram-w5ccc-1-minio-65968fd9-dh5nf 1/1 Running 0 112s 10.104.6.120 4am-node13 <none> <none> ``` client log: ``` 2022-09-20 03:33:52,111] [ INFO] - Start load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:276) [2022-09-20 03:33:57,188] [ DEBUG] - Milvus load_collection run in 5.0775s (milvus_benchmark.client:57) [2022-09-20 03:33:57,188] [ INFO] - End load collection: sift_128_euclidean (milvus_benchmark.runners.accuracy:279) [2022-09-20 03:33:57,189] [ DEBUG] - Milvus get_query_segment_info run in 0.0008s (milvus_benchmark.client:57) [2022-09-20 03:33:57,189] [ DEBUG] - query_segment_info: [] (milvus_benchmark.runners.accuracy:282) [2022-09-20 03:33:57,191] [ INFO] - Start run case (milvus_benchmark.main:88) [2022-09-20 03:33:58,088] [ ERROR] - RPC error: [search], <MilvusException: (code=1, message=fail to search on all shard leaders, err=fail to Search, QueryNode ID=1, reason=stream operation failed: deadline exceed)>, <Time:{'RPC start': '2022-09-20 03:33:57.191366', 'RPC error': '2022-09-20 03:33:58.088386'}> (pymilvus.decorators:112) [2022-09-20 03:33:58,089] [ ERROR] - Traceback (most recent call last): File "main.py", line 95, in run_suite result = runner.run_case(case_metric, **case) File "/src/milvus_benchmark/runners/accuracy.py", line 292, in run_case self.milvus.query(case_param["vector_query"], filter_query=case_param["filter_query"], File "/src/milvus_benchmark/client.py", line 53, in wrapper result = func(*args, **kwargs) File "/src/milvus_benchmark/client.py", line 346, in query result = self._milvus.search(tmp_collection_name, **params) File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/stub.py", line 844, in search return handler.search(collection_name, data, anns_field, param, limit, expression, File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 113, in handler raise e File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 109, in handler return func(*args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 139, in handler ret = func(self, *args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 89, in handler raise e File "/usr/local/lib/python3.8/dist-packages/pymilvus/decorators.py", line 51, in handler return func(self, *args, **kwargs) File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 472, in search return self._execute_search_requests(requests, timeout, **_kwargs) File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 436, in _execute_search_requests raise pre_err File "/usr/local/lib/python3.8/dist-packages/pymilvus/client/grpc_handler.py", line 427, in _execute_search_requests raise MilvusException(response.status.error_code, response.status.reason) pymilvus.exceptions.MilvusException: <MilvusException: (code=1, message=fail to search on all shard leaders, err=fail to Search, QueryNode ID=1, reason=stream operation failed: deadline exceed)> (milvus_benchmark.main:98) ``` ### Expected Behavior _No response_ ### Steps To Reproduce ```markdown 1.create an collection 2.insert sift1m data 3.create flat index, then load 4.search raise error ``` ### Milvus Log _No response_ ### Anything else? client-acc-sift-flat: { "config.yaml": "ann_accuracy: collections: - milvus: cache_config.cpu_cache_capacity: 16GB engine_config.use_blas_threshold: 1100 server: cpus: 12 source_file: /test/milvus/ann_hdf5/sift-128-euclidean.hdf5 collection_name: sift_128_euclidean index_types: ['flat'] index_params: nlist: [1024] top_ks: [10] nqs: [10000] search_params: nprobe: [1] " }
test
milvus search failed raise an error fail to search on all shard leaders err fail to search querynode id reason stream operation failed deadline exceed is there an existing issue for this i have searched the existing issues environment markdown milvus version master deployment mode standalone or cluster standalone sdk version e g pymilvus os ubuntu or centos cpu memory gpu others current behavior server instance fouram server configmap server single client configmap client acc sift flat sever name ready status restarts age ip node nominated node readiness gates fouram etcd running fouram milvus standalone running fouram minio running client log start load collection sift euclidean milvus benchmark runners accuracy milvus load collection run in milvus benchmark client end load collection sift euclidean milvus benchmark runners accuracy milvus get query segment info run in milvus benchmark client query segment info milvus benchmark runners accuracy start run case milvus benchmark main rpc error pymilvus decorators traceback most recent call last file main py line in run suite result runner run case case metric case file src milvus benchmark runners accuracy py line in run case self milvus query case param filter query case param file src milvus benchmark client py line in wrapper result func args kwargs file src milvus benchmark client py line in query result self milvus search tmp collection name params file usr local lib dist packages pymilvus client stub py line in search return handler search collection name data anns field param limit expression file usr local lib dist packages pymilvus decorators py line in handler raise e file usr local lib dist packages pymilvus decorators py line in handler return func args kwargs file usr local lib dist packages pymilvus decorators py line in handler ret func self args kwargs file usr local lib dist packages pymilvus decorators py line in handler raise e file usr local lib dist packages pymilvus decorators py line in handler return func self args kwargs file usr local lib dist packages pymilvus client grpc handler py line in search return self execute search requests requests timeout kwargs file usr local lib dist packages pymilvus client grpc handler py line in execute search requests raise pre err file usr local lib dist packages pymilvus client grpc handler py line in execute search requests raise milvusexception response status error code response status reason pymilvus exceptions milvusexception milvus benchmark main expected behavior no response steps to reproduce markdown create an collection insert data create flat index then load search raise error milvus log no response anything else client acc sift flat config yaml ann accuracy collections milvus cache config cpu cache capacity engine config use blas threshold server cpus source file test milvus ann sift euclidean collection name sift euclidean index types index params nlist top ks nqs search params nprobe
1
185,033
21,785,054,383
IssuesEvent
2022-05-14 02:18:46
Yash-Handa/GitHub-Org-Geographics
https://api.github.com/repos/Yash-Handa/GitHub-Org-Geographics
closed
CVE-2021-3807 (High) detected in ansi-regex-4.1.0.tgz, ansi-regex-3.0.0.tgz - autoclosed
security vulnerability
## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-4.1.0.tgz</b>, <b>ansi-regex-3.0.0.tgz</b></p></summary> <p> <details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/inquirer/node_modules/strip-ansi/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - cli-7.3.8.tgz (Root Library) - inquirer-6.2.1.tgz - strip-ansi-5.2.0.tgz - :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-3.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/webpack-cli/node_modules/ansi-regex/package.json,/node_modules/inquirer/node_modules/ansi-regex/package.json,/node_modules/@angular/compiler-cli/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - webpack-cli-3.3.0.tgz (Root Library) - yargs-12.0.5.tgz - cliui-4.1.0.tgz - strip-ansi-4.0.0.tgz - :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Yash-Handa/GitHub-Org-Geographics/commit/b8f8164cc4941846663e45ae47eb053813efeb8d">b8f8164cc4941846663e45ae47eb053813efeb8d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-3807 (High) detected in ansi-regex-4.1.0.tgz, ansi-regex-3.0.0.tgz - autoclosed - ## CVE-2021-3807 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>ansi-regex-4.1.0.tgz</b>, <b>ansi-regex-3.0.0.tgz</b></p></summary> <p> <details><summary><b>ansi-regex-4.1.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-4.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/inquirer/node_modules/strip-ansi/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - cli-7.3.8.tgz (Root Library) - inquirer-6.2.1.tgz - strip-ansi-5.2.0.tgz - :x: **ansi-regex-4.1.0.tgz** (Vulnerable Library) </details> <details><summary><b>ansi-regex-3.0.0.tgz</b></p></summary> <p>Regular expression for matching ANSI escape codes</p> <p>Library home page: <a href="https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz">https://registry.npmjs.org/ansi-regex/-/ansi-regex-3.0.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/webpack-cli/node_modules/ansi-regex/package.json,/node_modules/inquirer/node_modules/ansi-regex/package.json,/node_modules/@angular/compiler-cli/node_modules/ansi-regex/package.json</p> <p> Dependency Hierarchy: - webpack-cli-3.3.0.tgz (Root Library) - yargs-12.0.5.tgz - cliui-4.1.0.tgz - strip-ansi-4.0.0.tgz - :x: **ansi-regex-3.0.0.tgz** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/Yash-Handa/GitHub-Org-Geographics/commit/b8f8164cc4941846663e45ae47eb053813efeb8d">b8f8164cc4941846663e45ae47eb053813efeb8d</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> ansi-regex is vulnerable to Inefficient Regular Expression Complexity <p>Publish Date: 2021-09-17 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2021-3807>CVE-2021-3807</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/">https://huntr.dev/bounties/5b3cf33b-ede0-4398-9974-800876dfd994/</a></p> <p>Release Date: 2021-09-17</p> <p>Fix Resolution: ansi-regex - 5.0.1,6.0.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in ansi regex tgz ansi regex tgz autoclosed cve high severity vulnerability vulnerable libraries ansi regex tgz ansi regex tgz ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file package json path to vulnerable library node modules inquirer node modules strip ansi node modules ansi regex package json dependency hierarchy cli tgz root library inquirer tgz strip ansi tgz x ansi regex tgz vulnerable library ansi regex tgz regular expression for matching ansi escape codes library home page a href path to dependency file package json path to vulnerable library node modules webpack cli node modules ansi regex package json node modules inquirer node modules ansi regex package json node modules angular compiler cli node modules ansi regex package json dependency hierarchy webpack cli tgz root library yargs tgz cliui tgz strip ansi tgz x ansi regex tgz vulnerable library found in head commit a href vulnerability details ansi regex is vulnerable to inefficient regular expression complexity publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution ansi regex step up your open source security game with whitesource
0
101,875
8,806,664,862
IssuesEvent
2018-12-27 05:46:07
humera987/FXLabs-Test-Automation
https://api.github.com/repos/humera987/FXLabs-Test-Automation
closed
testing 3 : ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue
testing 3
Project : testing 3 Job : Default Env : Default Region : US_WEST Result : fail Status Code : 200 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NjI2MzI0ODQtZjc5NC00NzZiLTkyNzUtODYwMjgwZGMwZGUz; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 27 Dec 2018 05:38:15 GMT]} Endpoint : http://13.56.210.25/api/v1/projects/null/clear-all-suggestions Request : Response : { "requestId" : "None", "requestTime" : "2018-12-27T05:38:16.617+0000", "errors" : false, "messages" : [ ], "data" : true, "totalPages" : 0, "totalElements" : 0 } Logs : 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : URL [http://13.56.210.25/api/v1/projects/null/clear-all-suggestions] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Method [GET] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Request [] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Request-Headers [{Content-Type=[application/json], Accept=[application/json], Authorization=[Basic SHVtZXJhLy9odW1lcmFAZnhsYWJzLmlvOmh1bWVyYTEyMyQ=]}] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Response [{ "requestId" : "None", "requestTime" : "2018-12-27T05:38:16.617+0000", "errors" : false, "messages" : [ ], "data" : true, "totalPages" : 0, "totalElements" : 0 }] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Response-Headers [{X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NjI2MzI0ODQtZjc5NC00NzZiLTkyNzUtODYwMjgwZGMwZGUz; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 27 Dec 2018 05:38:15 GMT]}] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : StatusCode [200] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Time [1286] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Size [139] 2018-12-27 05:38:16 INFO [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Assertion [@StatusCode != 401] resolved-to [200 != 401] result [Passed] 2018-12-27 05:38:16 INFO [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Assertion [@StatusCode != 500] resolved-to [200 != 500] result [Passed] 2018-12-27 05:38:16 INFO [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Assertion [@StatusCode != 404] resolved-to [200 != 404] result [Passed] 2018-12-27 05:38:16 ERROR [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Assertion [@StatusCode != 200] resolved-to [200 != 200] result [Failed] --- FX Bot ---
1.0
testing 3 : ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue - Project : testing 3 Job : Default Env : Default Region : US_WEST Result : fail Status Code : 200 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NjI2MzI0ODQtZjc5NC00NzZiLTkyNzUtODYwMjgwZGMwZGUz; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 27 Dec 2018 05:38:15 GMT]} Endpoint : http://13.56.210.25/api/v1/projects/null/clear-all-suggestions Request : Response : { "requestId" : "None", "requestTime" : "2018-12-27T05:38:16.617+0000", "errors" : false, "messages" : [ ], "data" : true, "totalPages" : 0, "totalElements" : 0 } Logs : 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : URL [http://13.56.210.25/api/v1/projects/null/clear-all-suggestions] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Method [GET] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Request [] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Request-Headers [{Content-Type=[application/json], Accept=[application/json], Authorization=[Basic SHVtZXJhLy9odW1lcmFAZnhsYWJzLmlvOmh1bWVyYTEyMyQ=]}] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Response [{ "requestId" : "None", "requestTime" : "2018-12-27T05:38:16.617+0000", "errors" : false, "messages" : [ ], "data" : true, "totalPages" : 0, "totalElements" : 0 }] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Response-Headers [{X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=NjI2MzI0ODQtZjc5NC00NzZiLTkyNzUtODYwMjgwZGMwZGUz; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Thu, 27 Dec 2018 05:38:15 GMT]}] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : StatusCode [200] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Time [1286] 2018-12-27 05:38:16 DEBUG [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Size [139] 2018-12-27 05:38:16 INFO [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Assertion [@StatusCode != 401] resolved-to [200 != 401] result [Passed] 2018-12-27 05:38:16 INFO [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Assertion [@StatusCode != 500] resolved-to [200 != 500] result [Passed] 2018-12-27 05:38:16 INFO [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Assertion [@StatusCode != 404] resolved-to [200 != 404] result [Passed] 2018-12-27 05:38:16 ERROR [ApiV1ProjectsIdClearAllSuggestionsGetPathParamIdNullValue] : Assertion [@StatusCode != 200] resolved-to [200 != 200] result [Failed] --- FX Bot ---
test
testing project testing job default env default region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response requestid none requesttime errors false messages data true totalpages totalelements logs debug url debug method debug request debug request headers accept authorization debug response requestid none requesttime errors false messages data true totalpages totalelements debug response headers x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date debug statuscode debug time debug size info assertion resolved to result info assertion resolved to result info assertion resolved to result error assertion resolved to result fx bot
1
141,526
11,424,142,143
IssuesEvent
2020-02-03 17:08:35
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
globalRoles - Issue while adding a Global Role
[zube]: To Test kind/bug-qa team/ui
**What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** 1. Go to Global --> Security --> Roles 2. Click on Add Global Role 3. User is presented with UI <img width="1144" alt="Screen Shot 2020-01-23 at 4 49 03 PM" src="https://user-images.githubusercontent.com/26032343/73036391-4b272480-3e00-11ea-8d8b-6dd1b0bb6b3f.png"> 4. Click on Cancel. 5. Click on Admin --> Clone 6. Click on Cancel 7. Click on Add Global Role 8. User is presented with UI <img width="1127" alt="Screen Shot 2020-01-23 at 4 50 03 PM" src="https://user-images.githubusercontent.com/26032343/73036443-6b56e380-3e00-11ea-80ad-222c62ddaa44.png"> **Expected Result:** User should be presented with UI as in Step 3 **Actual Result:** UI seems to cache the UI from the previous page load. **Note:** Refresh of the page, makes this issue go away. **Other details that may be helpful:** **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master-head - commit id: `47bf996c2` - Installation option (single install/HA): single
1.0
globalRoles - Issue while adding a Global Role - **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** 1. Go to Global --> Security --> Roles 2. Click on Add Global Role 3. User is presented with UI <img width="1144" alt="Screen Shot 2020-01-23 at 4 49 03 PM" src="https://user-images.githubusercontent.com/26032343/73036391-4b272480-3e00-11ea-8d8b-6dd1b0bb6b3f.png"> 4. Click on Cancel. 5. Click on Admin --> Clone 6. Click on Cancel 7. Click on Add Global Role 8. User is presented with UI <img width="1127" alt="Screen Shot 2020-01-23 at 4 50 03 PM" src="https://user-images.githubusercontent.com/26032343/73036443-6b56e380-3e00-11ea-80ad-222c62ddaa44.png"> **Expected Result:** User should be presented with UI as in Step 3 **Actual Result:** UI seems to cache the UI from the previous page load. **Note:** Refresh of the page, makes this issue go away. **Other details that may be helpful:** **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): master-head - commit id: `47bf996c2` - Installation option (single install/HA): single
test
globalroles issue while adding a global role what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible go to global security roles click on add global role user is presented with ui img width alt screen shot at pm src click on cancel click on admin clone click on cancel click on add global role user is presented with ui img width alt screen shot at pm src expected result user should be presented with ui as in step actual result ui seems to cache the ui from the previous page load note refresh of the page makes this issue go away other details that may be helpful environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui master head commit id installation option single install ha single
1
262,205
22,823,416,858
IssuesEvent
2022-07-12 06:10:25
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: backupTPCC failed
C-test-failure O-robot O-roachtest branch-master release-blocker
roachtest.backupTPCC [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5718837?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5718837?buildTab=artifacts#/backupTPCC) on master @ [571bfa3afb3858ae84d8a8fcdbb0a38e058402a5](https://github.com/cockroachdb/cockroach/commits/571bfa3afb3858ae84d8a8fcdbb0a38e058402a5): ``` test artifacts and logs in: /artifacts/backupTPCC/run_1 monitor.go:127,backup.go:1095,test_runner.go:896: monitor failure: monitor task failed: dial tcp 34.139.144.3:26257: connect: connection refused (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerBackup.func4 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/backup.go:1095 | [...repeated from below...] Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func2 | main/pkg/cmd/roachtest/monitor.go:171 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (4) monitor task failed Wraps: (5) dial tcp 34.139.144.3:26257 Wraps: (6) connect Wraps: (7) connection refused Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *net.OpError (6) *os.SyscallError (7) syscall.Errno ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/bulk-io <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*backupTPCC.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
2.0
roachtest: backupTPCC failed - roachtest.backupTPCC [failed](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5718837?buildTab=log) with [artifacts](https://teamcity.cockroachdb.com/buildConfiguration/Cockroach_Nightlies_RoachtestNightlyGceBazel/5718837?buildTab=artifacts#/backupTPCC) on master @ [571bfa3afb3858ae84d8a8fcdbb0a38e058402a5](https://github.com/cockroachdb/cockroach/commits/571bfa3afb3858ae84d8a8fcdbb0a38e058402a5): ``` test artifacts and logs in: /artifacts/backupTPCC/run_1 monitor.go:127,backup.go:1095,test_runner.go:896: monitor failure: monitor task failed: dial tcp 34.139.144.3:26257: connect: connection refused (1) attached stack trace -- stack trace: | main.(*monitorImpl).WaitE | main/pkg/cmd/roachtest/monitor.go:115 | main.(*monitorImpl).Wait | main/pkg/cmd/roachtest/monitor.go:123 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests.registerBackup.func4 | github.com/cockroachdb/cockroach/pkg/cmd/roachtest/tests/backup.go:1095 | [...repeated from below...] Wraps: (2) monitor failure Wraps: (3) attached stack trace -- stack trace: | main.(*monitorImpl).wait.func2 | main/pkg/cmd/roachtest/monitor.go:171 | runtime.goexit | GOROOT/src/runtime/asm_amd64.s:1581 Wraps: (4) monitor task failed Wraps: (5) dial tcp 34.139.144.3:26257 Wraps: (6) connect Wraps: (7) connection refused Error types: (1) *withstack.withStack (2) *errutil.withPrefix (3) *withstack.withStack (4) *errutil.withPrefix (5) *net.OpError (6) *os.SyscallError (7) syscall.Errno ``` <p>Parameters: <code>ROACHTEST_cloud=gce</code> , <code>ROACHTEST_cpu=4</code> , <code>ROACHTEST_ssd=0</code> </p> <details><summary>Help</summary> <p> See: [roachtest README](https://github.com/cockroachdb/cockroach/blob/master/pkg/cmd/roachtest/README.md) See: [How To Investigate \(internal\)](https://cockroachlabs.atlassian.net/l/c/SSSBr8c7) </p> </details> /cc @cockroachdb/bulk-io <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*backupTPCC.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
roachtest backuptpcc failed roachtest backuptpcc with on master test artifacts and logs in artifacts backuptpcc run monitor go backup go test runner go monitor failure monitor task failed dial tcp connect connection refused attached stack trace stack trace main monitorimpl waite main pkg cmd roachtest monitor go main monitorimpl wait main pkg cmd roachtest monitor go github com cockroachdb cockroach pkg cmd roachtest tests registerbackup github com cockroachdb cockroach pkg cmd roachtest tests backup go wraps monitor failure wraps attached stack trace stack trace main monitorimpl wait main pkg cmd roachtest monitor go runtime goexit goroot src runtime asm s wraps monitor task failed wraps dial tcp wraps connect wraps connection refused error types withstack withstack errutil withprefix withstack withstack errutil withprefix net operror os syscallerror syscall errno parameters roachtest cloud gce roachtest cpu roachtest ssd help see see cc cockroachdb bulk io
1
723,103
24,884,914,043
IssuesEvent
2022-10-28 06:48:25
GoogleCloudPlatform/pgadapter
https://api.github.com/repos/GoogleCloudPlatform/pgadapter
opened
Errors are not propagated correctly in a DDL batch with an if [not] exists clause
type: bug priority: p2
A DDL batch that contains the following: 1. The first statement is a statement with an `if not exists` clause that does not get executed, because the object already exists. 2. The second statement is executed, but contains an error. The error from the second statement is not propagated correctly to the client by PGAdapter.
1.0
Errors are not propagated correctly in a DDL batch with an if [not] exists clause - A DDL batch that contains the following: 1. The first statement is a statement with an `if not exists` clause that does not get executed, because the object already exists. 2. The second statement is executed, but contains an error. The error from the second statement is not propagated correctly to the client by PGAdapter.
non_test
errors are not propagated correctly in a ddl batch with an if exists clause a ddl batch that contains the following the first statement is a statement with an if not exists clause that does not get executed because the object already exists the second statement is executed but contains an error the error from the second statement is not propagated correctly to the client by pgadapter
0
143,194
19,143,494,178
IssuesEvent
2021-12-02 03:25:47
TreyM-WSS/Struts2-Examples
https://api.github.com/repos/TreyM-WSS/Struts2-Examples
opened
CVE-2016-4433 (High) detected in struts2-core-2.3.20.1.jar, xwork-core-2.3.20.1.jar
security vulnerability
## CVE-2016-4433 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>struts2-core-2.3.20.1.jar</b>, <b>xwork-core-2.3.20.1.jar</b></p></summary> <p> <details><summary><b>struts2-core-2.3.20.1.jar</b></p></summary> <p>Apache Struts 2</p> <p>Path to dependency file: Struts2-Examples/Struts2RestConvention/pom.xml</p> <p>Path to vulnerable library: tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar</p> <p> Dependency Hierarchy: - :x: **struts2-core-2.3.20.1.jar** (Vulnerable Library) </details> <details><summary><b>xwork-core-2.3.20.1.jar</b></p></summary> <p>Apache Struts 2</p> <p>Library home page: <a href="http://struts.apache.org/">http://struts.apache.org/</a></p> <p>Path to dependency file: Struts2-Examples/Struts2RestConvention/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar</p> <p> Dependency Hierarchy: - struts2-core-2.3.20.1.jar (Root Library) - :x: **xwork-core-2.3.20.1.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/Struts2-Examples/commit/40920eab4f14f1e7436aafb936c150a59b8e1940">40920eab4f14f1e7436aafb936c150a59b8e1940</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Struts 2 2.3.20 through 2.3.28.1 allows remote attackers to bypass intended access restrictions and conduct redirection attacks via a crafted request. <p>Publish Date: 2016-07-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4433>CVE-2016-4433</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/struts/tree/STRUTS_2_3_29">https://github.com/apache/struts/tree/STRUTS_2_3_29</a></p> <p>Release Date: 2016-07-04</p> <p>Fix Resolution: org.apache.struts:struts2-core:2.3.29, org.apache.struts.xwork:xwork-core:2.3.29</p> </p> </details> <p></p>
True
CVE-2016-4433 (High) detected in struts2-core-2.3.20.1.jar, xwork-core-2.3.20.1.jar - ## CVE-2016-4433 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>struts2-core-2.3.20.1.jar</b>, <b>xwork-core-2.3.20.1.jar</b></p></summary> <p> <details><summary><b>struts2-core-2.3.20.1.jar</b></p></summary> <p>Apache Struts 2</p> <p>Path to dependency file: Struts2-Examples/Struts2RestConvention/pom.xml</p> <p>Path to vulnerable library: tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar,tory/org/apache/struts/struts2-core/2.3.20.1/struts2-core-2.3.20.1.jar</p> <p> Dependency Hierarchy: - :x: **struts2-core-2.3.20.1.jar** (Vulnerable Library) </details> <details><summary><b>xwork-core-2.3.20.1.jar</b></p></summary> <p>Apache Struts 2</p> <p>Library home page: <a href="http://struts.apache.org/">http://struts.apache.org/</a></p> <p>Path to dependency file: Struts2-Examples/Struts2RestConvention/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar,/home/wss-scanner/.m2/repository/org/apache/struts/xwork/xwork-core/2.3.20.1/xwork-core-2.3.20.1.jar</p> <p> Dependency Hierarchy: - struts2-core-2.3.20.1.jar (Root Library) - :x: **xwork-core-2.3.20.1.jar** (Vulnerable Library) </details> <p>Found in HEAD commit: <a href="https://github.com/TreyM-WSS/Struts2-Examples/commit/40920eab4f14f1e7436aafb936c150a59b8e1940">40920eab4f14f1e7436aafb936c150a59b8e1940</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache Struts 2 2.3.20 through 2.3.28.1 allows remote attackers to bypass intended access restrictions and conduct redirection attacks via a crafted request. <p>Publish Date: 2016-07-04 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-4433>CVE-2016-4433</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/apache/struts/tree/STRUTS_2_3_29">https://github.com/apache/struts/tree/STRUTS_2_3_29</a></p> <p>Release Date: 2016-07-04</p> <p>Fix Resolution: org.apache.struts:struts2-core:2.3.29, org.apache.struts.xwork:xwork-core:2.3.29</p> </p> </details> <p></p>
non_test
cve high detected in core jar xwork core jar cve high severity vulnerability vulnerable libraries core jar xwork core jar core jar apache struts path to dependency file examples pom xml path to vulnerable library tory org apache struts core core jar tory org apache struts core core jar tory org apache struts core core jar tory org apache struts core core jar tory org apache struts core core jar tory org apache struts core core jar tory org apache struts core core jar tory org apache struts core core jar dependency hierarchy x core jar vulnerable library xwork core jar apache struts library home page a href path to dependency file examples pom xml path to vulnerable library home wss scanner repository org apache struts xwork xwork core xwork core jar home wss scanner repository org apache struts xwork xwork core xwork core jar home wss scanner repository org apache struts xwork xwork core xwork core jar home wss scanner repository org apache struts xwork xwork core xwork core jar home wss scanner repository org apache struts xwork xwork core xwork core jar home wss scanner repository org apache struts xwork xwork core xwork core jar home wss scanner repository org apache struts xwork xwork core xwork core jar home wss scanner repository org apache struts xwork xwork core xwork core jar dependency hierarchy core jar root library x xwork core jar vulnerable library found in head commit a href vulnerability details apache struts through allows remote attackers to bypass intended access restrictions and conduct redirection attacks via a crafted request publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache struts core org apache struts xwork xwork core
0
44,670
9,621,577,984
IssuesEvent
2019-05-14 10:59:48
Yoast/wordpress-seo
https://api.github.com/repos/Yoast/wordpress-seo
closed
A fix for our "Person cannot be a Publisher" problem
0.5 component: schema (JSON+LD) needs-code-review
Pages (articles) on sites where the site represents a `Person` throw an error because the SDTT does not permit a `Person` to be a `Publisher`. See [this example](https://search.google.com/structured-data/testing-tool/#url=https%3A%2F%2Fwww.jonoalderson.com%2Fconjecture%2Fhow-to-do-seo-in-20xx%2F). Google require that the `Publisher` be an `Organization` (despite that being [valid schema](https://schema.org/publisher)). Whilst this is undoubtedly a problem/issue with Google / the SDTT, it's creating noise and support overheads for us. We can solve (or more specifically, _bypass_) this by: - Changing the _type_ of the `Person` piece to an array of `["Person","Organization"]` - Add a `logo` property to the `Person`, referencing the ID of the `#personlogo` This is a slightly ugly fix, but it works, and doesn't compromise (too much) on the integrity of the graph. When a site represents a person, we're now saying that the person in question is both a Person _and_ a/the Organization. This feels like a reasonable compromise/approach.
1.0
A fix for our "Person cannot be a Publisher" problem - Pages (articles) on sites where the site represents a `Person` throw an error because the SDTT does not permit a `Person` to be a `Publisher`. See [this example](https://search.google.com/structured-data/testing-tool/#url=https%3A%2F%2Fwww.jonoalderson.com%2Fconjecture%2Fhow-to-do-seo-in-20xx%2F). Google require that the `Publisher` be an `Organization` (despite that being [valid schema](https://schema.org/publisher)). Whilst this is undoubtedly a problem/issue with Google / the SDTT, it's creating noise and support overheads for us. We can solve (or more specifically, _bypass_) this by: - Changing the _type_ of the `Person` piece to an array of `["Person","Organization"]` - Add a `logo` property to the `Person`, referencing the ID of the `#personlogo` This is a slightly ugly fix, but it works, and doesn't compromise (too much) on the integrity of the graph. When a site represents a person, we're now saying that the person in question is both a Person _and_ a/the Organization. This feels like a reasonable compromise/approach.
non_test
a fix for our person cannot be a publisher problem pages articles on sites where the site represents a person throw an error because the sdtt does not permit a person to be a publisher see google require that the publisher be an organization despite that being whilst this is undoubtedly a problem issue with google the sdtt it s creating noise and support overheads for us we can solve or more specifically bypass this by changing the type of the person piece to an array of add a logo property to the person referencing the id of the personlogo this is a slightly ugly fix but it works and doesn t compromise too much on the integrity of the graph when a site represents a person we re now saying that the person in question is both a person and a the organization this feels like a reasonable compromise approach
0
852
10,963,357,781
IssuesEvent
2019-11-27 19:29:10
Azure/azure-functions-host
https://api.github.com/repos/Azure/azure-functions-host
closed
[V1] Investigate race condition where placeholder never specializes
1.x Supportability
#### Conditions 1. Functions V1 2. Run From Package 3. Placeholders enabled. #### Expected behavior Functions V1 placeholder properly specializes and by the time the Functions host is started, it will be able to start processing events. #### Actual behavior Functions V1 host appears to be starting before customer's site content is downloaded, resulting in a state where the customer's app is assigned a worker that doesn't actually have their site content. #### Example Timeline: All times are in UTC: 2019-11-03 07:08:58.9927151: DWAS has registered placeholder mawsFnPlaceholder77_f_v1_x86 2019-11-03 07:09:00.3082544: DWAS claims that mawsFnPlaceholder77_f_v1_x86 is warmed up 2019-11-03 07:09:01.9075569: DWAS sees request for customer site 2019-11-03 07:09:02.3681568: Dynamic cache identifies customer's run from package ZIP, begins download. 2019-11-03 07:09:04.9026224: Functions host starts with customer app name 2019-11-03 07:09:05.0471987: Functions host logs empty host configuration file read. 2019-11-03 07:09:06.3418355: Functions host sees /api/Warmup request 2019-11-03 07:09:06.6142375: Dynamic cache claims run from package zip has finished downloading 2019-11-03 07:09:06.6664812: DWAS claims customer site is specialized 2019-11-03 07:09:07.7217008: Functions host claims it has started. #### Known workarounds Turn off placeholders.
True
[V1] Investigate race condition where placeholder never specializes - #### Conditions 1. Functions V1 2. Run From Package 3. Placeholders enabled. #### Expected behavior Functions V1 placeholder properly specializes and by the time the Functions host is started, it will be able to start processing events. #### Actual behavior Functions V1 host appears to be starting before customer's site content is downloaded, resulting in a state where the customer's app is assigned a worker that doesn't actually have their site content. #### Example Timeline: All times are in UTC: 2019-11-03 07:08:58.9927151: DWAS has registered placeholder mawsFnPlaceholder77_f_v1_x86 2019-11-03 07:09:00.3082544: DWAS claims that mawsFnPlaceholder77_f_v1_x86 is warmed up 2019-11-03 07:09:01.9075569: DWAS sees request for customer site 2019-11-03 07:09:02.3681568: Dynamic cache identifies customer's run from package ZIP, begins download. 2019-11-03 07:09:04.9026224: Functions host starts with customer app name 2019-11-03 07:09:05.0471987: Functions host logs empty host configuration file read. 2019-11-03 07:09:06.3418355: Functions host sees /api/Warmup request 2019-11-03 07:09:06.6142375: Dynamic cache claims run from package zip has finished downloading 2019-11-03 07:09:06.6664812: DWAS claims customer site is specialized 2019-11-03 07:09:07.7217008: Functions host claims it has started. #### Known workarounds Turn off placeholders.
non_test
investigate race condition where placeholder never specializes conditions functions run from package placeholders enabled expected behavior functions placeholder properly specializes and by the time the functions host is started it will be able to start processing events actual behavior functions host appears to be starting before customer s site content is downloaded resulting in a state where the customer s app is assigned a worker that doesn t actually have their site content example timeline all times are in utc dwas has registered placeholder f dwas claims that f is warmed up dwas sees request for customer site dynamic cache identifies customer s run from package zip begins download functions host starts with customer app name functions host logs empty host configuration file read functions host sees api warmup request dynamic cache claims run from package zip has finished downloading dwas claims customer site is specialized functions host claims it has started known workarounds turn off placeholders
0
226,183
17,971,122,495
IssuesEvent
2021-09-14 02:13:12
elastic/kibana
https://api.github.com/repos/elastic/kibana
opened
Failing test: Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional_basic/apps/ml/permissions/read_ml_access·ts - apps machine learning basic license permissions for user with read ML access (ft_ml_viewer) should display elements on Index Data Visualizer page correctly
failed-test
A test failed on a tracked branch ``` Error: retry.tryForTime timeout: Error: Expected total document count to be '287' (got '0') at Assertion.assert (/dev/shm/workspace/parallel/16/kibana/node_modules/@kbn/expect/expect.js:100:11) at Assertion.eql (/dev/shm/workspace/parallel/16/kibana/node_modules/@kbn/expect/expect.js:244:8) at /dev/shm/workspace/parallel/16/kibana/x-pack/test/functional/services/ml/data_visualizer_index_based.ts:28:29 at runMicrotasks (<anonymous>) at processTicksAndRejections (internal/process/task_queues.js:95:5) at runAttempt (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:27:15) at retryForSuccess (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:66:21) at RetryService.tryForTime (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry.ts:22:12) at Object.assertTotalDocumentCount (test/functional/services/ml/data_visualizer_index_based.ts:26:7) at Object.clickUseFullDataButton (test/functional/services/ml/data_visualizer_index_based.ts:37:7) at Context.<anonymous> (test/functional_basic/apps/ml/permissions/read_ml_access.ts:121:11) at Object.apply (/dev/shm/workspace/parallel/16/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) at onFailure (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:57:13) at RetryService.tryForTime (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry.ts:22:12) at Object.assertTotalDocumentCount (test/functional/services/ml/data_visualizer_index_based.ts:26:7) at Object.clickUseFullDataButton (test/functional/services/ml/data_visualizer_index_based.ts:37:7) at Context.<anonymous> (test/functional_basic/apps/ml/permissions/read_ml_access.ts:121:11) at Object.apply (/dev/shm/workspace/parallel/16/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/17006/) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional_basic/apps/ml/permissions/read_ml_access·ts","test.name":"apps machine learning basic license permissions for user with read ML access (ft_ml_viewer) should display elements on Index Data Visualizer page correctly","test.failCount":1}} -->
1.0
Failing test: Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional_basic/apps/ml/permissions/read_ml_access·ts - apps machine learning basic license permissions for user with read ML access (ft_ml_viewer) should display elements on Index Data Visualizer page correctly - A test failed on a tracked branch ``` Error: retry.tryForTime timeout: Error: Expected total document count to be '287' (got '0') at Assertion.assert (/dev/shm/workspace/parallel/16/kibana/node_modules/@kbn/expect/expect.js:100:11) at Assertion.eql (/dev/shm/workspace/parallel/16/kibana/node_modules/@kbn/expect/expect.js:244:8) at /dev/shm/workspace/parallel/16/kibana/x-pack/test/functional/services/ml/data_visualizer_index_based.ts:28:29 at runMicrotasks (<anonymous>) at processTicksAndRejections (internal/process/task_queues.js:95:5) at runAttempt (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:27:15) at retryForSuccess (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:66:21) at RetryService.tryForTime (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry.ts:22:12) at Object.assertTotalDocumentCount (test/functional/services/ml/data_visualizer_index_based.ts:26:7) at Object.clickUseFullDataButton (test/functional/services/ml/data_visualizer_index_based.ts:37:7) at Context.<anonymous> (test/functional_basic/apps/ml/permissions/read_ml_access.ts:121:11) at Object.apply (/dev/shm/workspace/parallel/16/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) at onFailure (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:17:9) at retryForSuccess (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry_for_success.ts:57:13) at RetryService.tryForTime (/dev/shm/workspace/parallel/16/kibana/test/common/services/retry/retry.ts:22:12) at Object.assertTotalDocumentCount (test/functional/services/ml/data_visualizer_index_based.ts:26:7) at Object.clickUseFullDataButton (test/functional/services/ml/data_visualizer_index_based.ts:37:7) at Context.<anonymous> (test/functional_basic/apps/ml/permissions/read_ml_access.ts:121:11) at Object.apply (/dev/shm/workspace/parallel/16/kibana/node_modules/@kbn/test/target_node/functional_test_runner/lib/mocha/wrap_function.js:87:16) ``` First failure: [Jenkins Build](https://kibana-ci.elastic.co/job/elastic+kibana+master/17006/) <!-- kibanaCiData = {"failed-test":{"test.class":"Chrome X-Pack UI Functional Tests Basic License.x-pack/test/functional_basic/apps/ml/permissions/read_ml_access·ts","test.name":"apps machine learning basic license permissions for user with read ML access (ft_ml_viewer) should display elements on Index Data Visualizer page correctly","test.failCount":1}} -->
test
failing test chrome x pack ui functional tests basic license x pack test functional basic apps ml permissions read ml access·ts apps machine learning basic license permissions for user with read ml access ft ml viewer should display elements on index data visualizer page correctly a test failed on a tracked branch error retry tryfortime timeout error expected total document count to be got at assertion assert dev shm workspace parallel kibana node modules kbn expect expect js at assertion eql dev shm workspace parallel kibana node modules kbn expect expect js at dev shm workspace parallel kibana x pack test functional services ml data visualizer index based ts at runmicrotasks at processticksandrejections internal process task queues js at runattempt dev shm workspace parallel kibana test common services retry retry for success ts at retryforsuccess dev shm workspace parallel kibana test common services retry retry for success ts at retryservice tryfortime dev shm workspace parallel kibana test common services retry retry ts at object asserttotaldocumentcount test functional services ml data visualizer index based ts at object clickusefulldatabutton test functional services ml data visualizer index based ts at context test functional basic apps ml permissions read ml access ts at object apply dev shm workspace parallel kibana node modules kbn test target node functional test runner lib mocha wrap function js at onfailure dev shm workspace parallel kibana test common services retry retry for success ts at retryforsuccess dev shm workspace parallel kibana test common services retry retry for success ts at retryservice tryfortime dev shm workspace parallel kibana test common services retry retry ts at object asserttotaldocumentcount test functional services ml data visualizer index based ts at object clickusefulldatabutton test functional services ml data visualizer index based ts at context test functional basic apps ml permissions read ml access ts at object apply dev shm workspace parallel kibana node modules kbn test target node functional test runner lib mocha wrap function js first failure
1
238,969
19,786,495,764
IssuesEvent
2022-01-18 07:32:43
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
kv/kvserver/batcheval: TestLeaseTransferWithPipelinedWrite failed
C-test-failure O-robot branch-master
kv/kvserver/batcheval.TestLeaseTransferWithPipelinedWrite [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4141083&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4141083&tab=artifacts#/) on master @ [365b4da8bd02c06ee59d2130a56dec74ffc9ce21](https://github.com/cockroachdb/cockroach/commits/365b4da8bd02c06ee59d2130a56dec74ffc9ce21): ``` /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:446 +0x1219 github.com/cockroachdb/cockroach/pkg/server.(*Server).PreStart() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1669 +0x310c github.com/cockroachdb/cockroach/pkg/server.(*Server).Start() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1251 +0x44 github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start() /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:500 +0x109 github.com/cockroachdb/cockroach/pkg/testutils/testcluster.(*TestCluster).startServer() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/testcluster/testcluster.go:481 +0x10a github.com/cockroachdb/cockroach/pkg/testutils/testcluster.(*TestCluster).Start() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/testcluster/testcluster.go:311 +0x4b5 github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartNewTestCluster() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_cluster_shim.go:218 +0x141 github.com/cockroachdb/cockroach/pkg/kv/kvserver/batcheval.TestLeaseTransferWithPipelinedWrite() /go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/batcheval/cmd_lease_test.go:46 +0x164 testing.tRunner() /usr/local/go/src/testing/testing.go:1259 +0x22f testing.(*T).Run·dwrap·21() /usr/local/go/src/testing/testing.go:1306 +0x47 Goroutine 48 (running) created at: github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx() /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:478 +0x615 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask() /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:354 +0x13d github.com/cockroachdb/cockroach/pkg/spanconfig/spanconfigkvsubscriber.(*KVSubscriber).Start() /go/src/github.com/cockroachdb/cockroach/pkg/spanconfig/spanconfigkvsubscriber/kvsubscriber.go:177 +0x26 github.com/cockroachdb/cockroach/pkg/server.(*Server).PreStart() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1801 +0x41fa github.com/cockroachdb/cockroach/pkg/server.(*Server).Start() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1251 +0x44 github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start() /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:500 +0x109 github.com/cockroachdb/cockroach/pkg/testutils/testcluster.(*TestCluster).startServer() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/testcluster/testcluster.go:481 +0x10a github.com/cockroachdb/cockroach/pkg/testutils/testcluster.(*TestCluster).Start() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/testcluster/testcluster.go:311 +0x4b5 github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartNewTestCluster() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_cluster_shim.go:218 +0x141 github.com/cockroachdb/cockroach/pkg/kv/kvserver/batcheval.TestLeaseTransferWithPipelinedWrite() /go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/batcheval/cmd_lease_test.go:46 +0x164 testing.tRunner() /usr/local/go/src/testing/testing.go:1259 +0x22f testing.(*T).Run·dwrap·21() /usr/local/go/src/testing/testing.go:1306 +0x47 ================== cmd_lease_test.go:115: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestLeaseTransferWithPipelinedWrite1544528607 testing.go:1152: race detected during execution of test --- FAIL: TestLeaseTransferWithPipelinedWrite (30.15s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - GOFLAGS=-race -parallel=4 </p> </details> /cc @cockroachdb/kv <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLeaseTransferWithPipelinedWrite.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
1.0
kv/kvserver/batcheval: TestLeaseTransferWithPipelinedWrite failed - kv/kvserver/batcheval.TestLeaseTransferWithPipelinedWrite [failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=4141083&tab=buildLog) with [artifacts](https://teamcity.cockroachdb.com/viewLog.html?buildId=4141083&tab=artifacts#/) on master @ [365b4da8bd02c06ee59d2130a56dec74ffc9ce21](https://github.com/cockroachdb/cockroach/commits/365b4da8bd02c06ee59d2130a56dec74ffc9ce21): ``` /go/src/github.com/cockroachdb/cockroach/pkg/server/node.go:446 +0x1219 github.com/cockroachdb/cockroach/pkg/server.(*Server).PreStart() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1669 +0x310c github.com/cockroachdb/cockroach/pkg/server.(*Server).Start() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1251 +0x44 github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start() /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:500 +0x109 github.com/cockroachdb/cockroach/pkg/testutils/testcluster.(*TestCluster).startServer() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/testcluster/testcluster.go:481 +0x10a github.com/cockroachdb/cockroach/pkg/testutils/testcluster.(*TestCluster).Start() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/testcluster/testcluster.go:311 +0x4b5 github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartNewTestCluster() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_cluster_shim.go:218 +0x141 github.com/cockroachdb/cockroach/pkg/kv/kvserver/batcheval.TestLeaseTransferWithPipelinedWrite() /go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/batcheval/cmd_lease_test.go:46 +0x164 testing.tRunner() /usr/local/go/src/testing/testing.go:1259 +0x22f testing.(*T).Run·dwrap·21() /usr/local/go/src/testing/testing.go:1306 +0x47 Goroutine 48 (running) created at: github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTaskEx() /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:478 +0x615 github.com/cockroachdb/cockroach/pkg/util/stop.(*Stopper).RunAsyncTask() /go/src/github.com/cockroachdb/cockroach/pkg/util/stop/stopper.go:354 +0x13d github.com/cockroachdb/cockroach/pkg/spanconfig/spanconfigkvsubscriber.(*KVSubscriber).Start() /go/src/github.com/cockroachdb/cockroach/pkg/spanconfig/spanconfigkvsubscriber/kvsubscriber.go:177 +0x26 github.com/cockroachdb/cockroach/pkg/server.(*Server).PreStart() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1801 +0x41fa github.com/cockroachdb/cockroach/pkg/server.(*Server).Start() /go/src/github.com/cockroachdb/cockroach/pkg/server/server.go:1251 +0x44 github.com/cockroachdb/cockroach/pkg/server.(*TestServer).Start() /go/src/github.com/cockroachdb/cockroach/pkg/server/testserver.go:500 +0x109 github.com/cockroachdb/cockroach/pkg/testutils/testcluster.(*TestCluster).startServer() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/testcluster/testcluster.go:481 +0x10a github.com/cockroachdb/cockroach/pkg/testutils/testcluster.(*TestCluster).Start() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/testcluster/testcluster.go:311 +0x4b5 github.com/cockroachdb/cockroach/pkg/testutils/serverutils.StartNewTestCluster() /go/src/github.com/cockroachdb/cockroach/pkg/testutils/serverutils/test_cluster_shim.go:218 +0x141 github.com/cockroachdb/cockroach/pkg/kv/kvserver/batcheval.TestLeaseTransferWithPipelinedWrite() /go/src/github.com/cockroachdb/cockroach/pkg/kv/kvserver/batcheval/cmd_lease_test.go:46 +0x164 testing.tRunner() /usr/local/go/src/testing/testing.go:1259 +0x22f testing.(*T).Run·dwrap·21() /usr/local/go/src/testing/testing.go:1306 +0x47 ================== cmd_lease_test.go:115: -- test log scope end -- test logs left over in: /go/src/github.com/cockroachdb/cockroach/artifacts/logTestLeaseTransferWithPipelinedWrite1544528607 testing.go:1152: race detected during execution of test --- FAIL: TestLeaseTransferWithPipelinedWrite (30.15s) ``` <details><summary>Help</summary> <p> See also: [How To Investigate a Go Test Failure \(internal\)](https://cockroachlabs.atlassian.net/l/c/HgfXfJgM) Parameters in this failure: - GOFLAGS=-race -parallel=4 </p> </details> /cc @cockroachdb/kv <sub> [This test on roachdash](https://roachdash.crdb.dev/?filter=status:open%20t:.*TestLeaseTransferWithPipelinedWrite.*&sort=title+created&display=lastcommented+project) | [Improve this report!](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues) </sub>
test
kv kvserver batcheval testleasetransferwithpipelinedwrite failed kv kvserver batcheval testleasetransferwithpipelinedwrite with on master go src github com cockroachdb cockroach pkg server node go github com cockroachdb cockroach pkg server server prestart go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg server testserver start go src github com cockroachdb cockroach pkg server testserver go github com cockroachdb cockroach pkg testutils testcluster testcluster startserver go src github com cockroachdb cockroach pkg testutils testcluster testcluster go github com cockroachdb cockroach pkg testutils testcluster testcluster start go src github com cockroachdb cockroach pkg testutils testcluster testcluster go github com cockroachdb cockroach pkg testutils serverutils startnewtestcluster go src github com cockroachdb cockroach pkg testutils serverutils test cluster shim go github com cockroachdb cockroach pkg kv kvserver batcheval testleasetransferwithpipelinedwrite go src github com cockroachdb cockroach pkg kv kvserver batcheval cmd lease test go testing trunner usr local go src testing testing go testing t run·dwrap· usr local go src testing testing go goroutine running created at github com cockroachdb cockroach pkg util stop stopper runasynctaskex go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg util stop stopper runasynctask go src github com cockroachdb cockroach pkg util stop stopper go github com cockroachdb cockroach pkg spanconfig spanconfigkvsubscriber kvsubscriber start go src github com cockroachdb cockroach pkg spanconfig spanconfigkvsubscriber kvsubscriber go github com cockroachdb cockroach pkg server server prestart go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg server server start go src github com cockroachdb cockroach pkg server server go github com cockroachdb cockroach pkg server testserver start go src github com cockroachdb cockroach pkg server testserver go github com cockroachdb cockroach pkg testutils testcluster testcluster startserver go src github com cockroachdb cockroach pkg testutils testcluster testcluster go github com cockroachdb cockroach pkg testutils testcluster testcluster start go src github com cockroachdb cockroach pkg testutils testcluster testcluster go github com cockroachdb cockroach pkg testutils serverutils startnewtestcluster go src github com cockroachdb cockroach pkg testutils serverutils test cluster shim go github com cockroachdb cockroach pkg kv kvserver batcheval testleasetransferwithpipelinedwrite go src github com cockroachdb cockroach pkg kv kvserver batcheval cmd lease test go testing trunner usr local go src testing testing go testing t run·dwrap· usr local go src testing testing go cmd lease test go test log scope end test logs left over in go src github com cockroachdb cockroach artifacts testing go race detected during execution of test fail testleasetransferwithpipelinedwrite help see also parameters in this failure goflags race parallel cc cockroachdb kv
1
250
2,518,281,703
IssuesEvent
2015-01-16 20:54:05
dart-lang/dartdoc-viewer
https://api.github.com/repos/dart-lang/dartdoc-viewer
closed
Dart APIDocs Back Button is broken.
bug imported-from-google-code triaged
_From [jrr9...@gmail.com](https://code.google.com/u/115262904355300685333/) on May 30, 2014 19:31:20_ If I enter an invalid URL on the Dart API docs page it will redirect to the closest valid page, but will still push the incorrect history state into my browser, breaking my ability to use the back button. Given that the structure of the docs appears to have changed recently and Google has cached the old results, this is a frequent and somewhat annoying problem. For instance if I Google "dart logging", this is the first result: api.dartlang.org/logging.html It then routes me to https://api.dartlang.org/apidocs/channels/stable/dartdoc-viewer/home I am then unable to navigate back to my search results by the back button.Not a huge issue, but irritating enough to be worth mentioning, so maybe worth fixing. Cheers. Jacob Read _Original issue: http://code.google.com/p/dart/issues/detail?id=19104_
1.0
Dart APIDocs Back Button is broken. - _From [jrr9...@gmail.com](https://code.google.com/u/115262904355300685333/) on May 30, 2014 19:31:20_ If I enter an invalid URL on the Dart API docs page it will redirect to the closest valid page, but will still push the incorrect history state into my browser, breaking my ability to use the back button. Given that the structure of the docs appears to have changed recently and Google has cached the old results, this is a frequent and somewhat annoying problem. For instance if I Google "dart logging", this is the first result: api.dartlang.org/logging.html It then routes me to https://api.dartlang.org/apidocs/channels/stable/dartdoc-viewer/home I am then unable to navigate back to my search results by the back button.Not a huge issue, but irritating enough to be worth mentioning, so maybe worth fixing. Cheers. Jacob Read _Original issue: http://code.google.com/p/dart/issues/detail?id=19104_
non_test
dart apidocs back button is broken from on may if i enter an invalid url on the dart api docs page it will redirect to the closest valid page but will still push the incorrect history state into my browser breaking my ability to use the back button given that the structure of the docs appears to have changed recently and google has cached the old results this is a frequent and somewhat annoying problem for instance if i google dart logging this is the first result api dartlang org logging html it then routes me to i am then unable to navigate back to my search results by the back button not a huge issue but irritating enough to be worth mentioning so maybe worth fixing cheers jacob read original issue
0
120,386
4,788,546,572
IssuesEvent
2016-10-30 16:34:20
CogRob/workspace
https://api.github.com/repos/CogRob/workspace
closed
Investigate the cause of unsuccessful dance demo
P1 - HIGH PRIORITY
(1) Investigate the mechanism for E-stop to work. Is it a rosnode? At which level it performs emergency stop? Which programs are affected? (2) Rule: Before doing a demo in front of public, we should always test it on ourselves. (3) We should have a procedure to reset the system to initial status. This could be hard given many ros packages are affected in the process, but we should figure it out. (4) Keep in mind the E-stop behavior when we design ROS program.
1.0
Investigate the cause of unsuccessful dance demo - (1) Investigate the mechanism for E-stop to work. Is it a rosnode? At which level it performs emergency stop? Which programs are affected? (2) Rule: Before doing a demo in front of public, we should always test it on ourselves. (3) We should have a procedure to reset the system to initial status. This could be hard given many ros packages are affected in the process, but we should figure it out. (4) Keep in mind the E-stop behavior when we design ROS program.
non_test
investigate the cause of unsuccessful dance demo investigate the mechanism for e stop to work is it a rosnode at which level it performs emergency stop which programs are affected rule before doing a demo in front of public we should always test it on ourselves we should have a procedure to reset the system to initial status this could be hard given many ros packages are affected in the process but we should figure it out keep in mind the e stop behavior when we design ros program
0
125,114
17,799,011,740
IssuesEvent
2021-09-01 04:13:33
deckhouse/deckhouse
https://api.github.com/repos/deckhouse/deckhouse
opened
[control-plane-manager] Add ability to encrypt etcd data
area/security type/feature-request
It is possible to use the `EncryptionConfiguration` file to make apiserver encrypting some Kubernetes objects, e.g., secrets, configmaps.
True
[control-plane-manager] Add ability to encrypt etcd data - It is possible to use the `EncryptionConfiguration` file to make apiserver encrypting some Kubernetes objects, e.g., secrets, configmaps.
non_test
add ability to encrypt etcd data it is possible to use the encryptionconfiguration file to make apiserver encrypting some kubernetes objects e g secrets configmaps
0
130,303
18,155,710,967
IssuesEvent
2021-09-27 01:05:50
gms-ws-sandbox/nibrs
https://api.github.com/repos/gms-ws-sandbox/nibrs
opened
WS-2016-7112 (Medium) detected in spring-context-3.2.16.RELEASE.jar
security vulnerability
## WS-2016-7112 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-context-3.2.16.RELEASE.jar</b></p></summary> <p>Spring Context</p> <p>Library home page: <a href="https://github.com/SpringSource/spring-framework">https://github.com/SpringSource/spring-framework</a></p> <p>Path to dependency file: nibrs/tools/nibrs-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-context/3.2.16.RELEASE/spring-context-3.2.16.RELEASE.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - uimafit-core-2.2.0.jar - :x: **spring-context-3.2.16.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 3.0.0.RELEASE through 3.2.17.RELEASE, 4.0.0.RELEASE through 4.2.7.RELEASE and 4.3.0.RELEASE through 4.3.1.RELEASE are vulnerable to Stack-based Buffer Overflow, which allows an authenticated attacker to crash the application when giving CronSequenceGenerator a reversed range in the “minutes” or “hours” fields. <p>Publish Date: 2021-09-23 <p>URL: <a href=https://github.com/spring-projects/spring-framework/commit/e431624e8472b3b53d1a0c4528bf736c612f1fd9>WS-2016-7112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/spring-projects/spring-framework/releases/tag/v5.0.0.M1">https://github.com/spring-projects/spring-framework/releases/tag/v5.0.0.M1</a></p> <p>Release Date: 2016-07-14</p> <p>Fix Resolution: org.springframework:spring-context:3.2.18.RELEASE,4.2.8.RELEASE,4.3.2.RELEASE,5.0.0.RELEASE</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-context","packageVersion":"3.2.16.RELEASE","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;org.apache.uima:uimafit-core:2.2.0;org.springframework:spring-context:3.2.16.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-context:3.2.18.RELEASE,4.2.8.RELEASE,4.3.2.RELEASE,5.0.0.RELEASE"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2016-7112","vulnerabilityDetails":"In Spring Framework, versions 3.0.0.RELEASE through 3.2.17.RELEASE, 4.0.0.RELEASE through 4.2.7.RELEASE and 4.3.0.RELEASE through 4.3.1.RELEASE are vulnerable to Stack-based Buffer Overflow, which allows an authenticated attacker to crash the application when giving CronSequenceGenerator a reversed range in the “minutes” or “hours” fields.","vulnerabilityUrl":"https://github.com/spring-projects/spring-framework/commit/e431624e8472b3b53d1a0c4528bf736c612f1fd9","cvss3Severity":"medium","cvss3Score":"4.9","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
True
WS-2016-7112 (Medium) detected in spring-context-3.2.16.RELEASE.jar - ## WS-2016-7112 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>spring-context-3.2.16.RELEASE.jar</b></p></summary> <p>Spring Context</p> <p>Library home page: <a href="https://github.com/SpringSource/spring-framework">https://github.com/SpringSource/spring-framework</a></p> <p>Path to dependency file: nibrs/tools/nibrs-common/pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/springframework/spring-context/3.2.16.RELEASE/spring-context-3.2.16.RELEASE.jar</p> <p> Dependency Hierarchy: - tika-parsers-1.18.jar (Root Library) - uimafit-core-2.2.0.jar - :x: **spring-context-3.2.16.RELEASE.jar** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In Spring Framework, versions 3.0.0.RELEASE through 3.2.17.RELEASE, 4.0.0.RELEASE through 4.2.7.RELEASE and 4.3.0.RELEASE through 4.3.1.RELEASE are vulnerable to Stack-based Buffer Overflow, which allows an authenticated attacker to crash the application when giving CronSequenceGenerator a reversed range in the “minutes” or “hours” fields. <p>Publish Date: 2021-09-23 <p>URL: <a href=https://github.com/spring-projects/spring-framework/commit/e431624e8472b3b53d1a0c4528bf736c612f1fd9>WS-2016-7112</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.9</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: N/A - Attack Complexity: N/A - Privileges Required: N/A - User Interaction: N/A - Scope: N/A - Impact Metrics: - Confidentiality Impact: N/A - Integrity Impact: N/A - Availability Impact: N/A </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/spring-projects/spring-framework/releases/tag/v5.0.0.M1">https://github.com/spring-projects/spring-framework/releases/tag/v5.0.0.M1</a></p> <p>Release Date: 2016-07-14</p> <p>Fix Resolution: org.springframework:spring-context:3.2.18.RELEASE,4.2.8.RELEASE,4.3.2.RELEASE,5.0.0.RELEASE</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.springframework","packageName":"spring-context","packageVersion":"3.2.16.RELEASE","packageFilePaths":["/tools/nibrs-common/pom.xml"],"isTransitiveDependency":true,"dependencyTree":"org.apache.tika:tika-parsers:1.18;org.apache.uima:uimafit-core:2.2.0;org.springframework:spring-context:3.2.16.RELEASE","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.springframework:spring-context:3.2.18.RELEASE,4.2.8.RELEASE,4.3.2.RELEASE,5.0.0.RELEASE"}],"baseBranches":["master"],"vulnerabilityIdentifier":"WS-2016-7112","vulnerabilityDetails":"In Spring Framework, versions 3.0.0.RELEASE through 3.2.17.RELEASE, 4.0.0.RELEASE through 4.2.7.RELEASE and 4.3.0.RELEASE through 4.3.1.RELEASE are vulnerable to Stack-based Buffer Overflow, which allows an authenticated attacker to crash the application when giving CronSequenceGenerator a reversed range in the “minutes” or “hours” fields.","vulnerabilityUrl":"https://github.com/spring-projects/spring-framework/commit/e431624e8472b3b53d1a0c4528bf736c612f1fd9","cvss3Severity":"medium","cvss3Score":"4.9","cvss3Metrics":{"A":"N/A","AC":"N/A","PR":"N/A","S":"N/A","C":"N/A","UI":"N/A","AV":"N/A","I":"N/A"},"extraData":{}}</REMEDIATE> -->
non_test
ws medium detected in spring context release jar ws medium severity vulnerability vulnerable library spring context release jar spring context library home page a href path to dependency file nibrs tools nibrs common pom xml path to vulnerable library home wss scanner repository org springframework spring context release spring context release jar dependency hierarchy tika parsers jar root library uimafit core jar x spring context release jar vulnerable library found in base branch master vulnerability details in spring framework versions release through release release through release and release through release are vulnerable to stack based buffer overflow which allows an authenticated attacker to crash the application when giving cronsequencegenerator a reversed range in the “minutes” or “hours” fields publish date url a href cvss score details base score metrics exploitability metrics attack vector n a attack complexity n a privileges required n a user interaction n a scope n a impact metrics confidentiality impact n a integrity impact n a availability impact n a for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org springframework spring context release release release release isopenpronvulnerability true ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org apache tika tika parsers org apache uima uimafit core org springframework spring context release isminimumfixversionavailable true minimumfixversion org springframework spring context release release release release basebranches vulnerabilityidentifier ws vulnerabilitydetails in spring framework versions release through release release through release and release through release are vulnerable to stack based buffer overflow which allows an authenticated attacker to crash the application when giving cronsequencegenerator a reversed range in the “minutes” or “hours” fields vulnerabilityurl
0
47,372
6,049,372,716
IssuesEvent
2017-06-12 18:36:42
cs340tabyu/Spring2017ImplementationTests
https://api.github.com/repos/cs340tabyu/Spring2017ImplementationTests
closed
The Game History doesn't include when someone claims a route
P4: Aesthetic or Design Flaw Team 3
There isn't a game history entry for when someone claims a route, so you just have to pay attention to the board really closely to notice where someone claimed a route.
1.0
The Game History doesn't include when someone claims a route - There isn't a game history entry for when someone claims a route, so you just have to pay attention to the board really closely to notice where someone claimed a route.
non_test
the game history doesn t include when someone claims a route there isn t a game history entry for when someone claims a route so you just have to pay attention to the board really closely to notice where someone claimed a route
0
51,990
10,741,654,349
IssuesEvent
2019-10-29 20:42:14
spyder-ide/spyder
https://api.github.com/repos/spyder-ide/spyder
closed
Decreasing automatic_completions_after_ms setting causes lag when typing
component:Code Completion type:Bug
<!--- **PLEASE READ:** When submitting here, please ensure you've completed the following checklist and checked the boxes to confirm. Issue reports without it may be closed. Thanks! ---> ## Problem Description Changing the setting `automatic_completions_after_ms` to a low but reasonable (from user's point of view) number causes noticeable lag when editing code. ### What steps reproduce the problem? 1. Change the `automatic_completions_after_ms` setting to 50ms: ![image](https://user-images.githubusercontent.com/3752225/67116621-a7d9ca00-f195-11e9-9a72-dad3d62562ed.png) 2. Try editing a Python file e.g. type out: ```python import numpy as np import matplotlib.pyplot as plt ``` ### What is the expected output? What do you see instead? __Expected__ As a user, I would expect minimal lag. __Actual__ There is noticeable lag in between typing a character and it being inserted into the buffer. This happens regardless of what completions engines are enabled and the value of the setting `automatic_completions_after_chars`. It appears as though something is blocking the UI thread in between keystroke → completion request → buffer update in UI. ## Versions <!--- You can get this information from Help > About Spyder... or (if Spyder won't launch) the "conda list" command from the Anaconda Prompt/Terminal/command line. ---> * Spyder version: 4.0.0.dev0 (Commit: c2b7673b1) * Python version: 3.7.4 64-bit * Qt version: Qt 5.9.6 * PyQt version: PyQt5 5.9.2 * Operating System name/version: Darwin 18.7.0 ### Dependencies <!--- Please go to the menu entry Help > Dependencies, press the Copy to clipboard button and paste below ---> ``` cloudpickle >=0.5.0 : 1.2.2 (OK) pygments >=2.0 : 2.4.2 (OK) qtconsole >=4.5.5 : 4.5.5 (OK) nbconvert >=4.0 : 5.6.0 (OK) sphinx >=0.6.6 : 2.2.0 (OK) pylint >=0.25 : 2.4.2 (OK) psutil >=0.3 : 5.6.3 (OK) qtawesome >=0.5.7 : 0.6.0 (OK) qtpy >=1.5.0 : 1.9.0 (OK) pickleshare >=0.4 : 0.7.5 (OK) zmq >=17 : 18.1.0 (OK) chardet >=2.0.0 : 3.0.4 (OK) numpydoc >=0.6.0 : 0.9.1 (OK) spyder_kernels >=1.5.0;<2.0.0: 1.6.0 (OK) qdarkstyle >=2.7 : 2.7 (OK) atomicwrites >=1.2.0 : 1.3.0 (OK) diff_match_patch >=20181111 : 20181111 (OK) watchdog : None (OK) keyring : None (OK) pexpect >=4.4.0 : 4.7.0 (OK) pympler : None (OK) sympy >=0.7.3 : None (NOK) cython >=0.21 : None (NOK) IPython >=4.0 : 7.8.0 (OK) matplotlib >=2.0.0 : None (NOK) pandas >=0.13.1 : None (NOK) numpy >=1.7 : None (NOK) scipy >=0.17.0 : None (NOK) pyls >=0.28.2;<0.29.0 : 0.28.3 (OK) rtree >=0.8.3 : 0.8.3 (OK) ```
1.0
Decreasing automatic_completions_after_ms setting causes lag when typing - <!--- **PLEASE READ:** When submitting here, please ensure you've completed the following checklist and checked the boxes to confirm. Issue reports without it may be closed. Thanks! ---> ## Problem Description Changing the setting `automatic_completions_after_ms` to a low but reasonable (from user's point of view) number causes noticeable lag when editing code. ### What steps reproduce the problem? 1. Change the `automatic_completions_after_ms` setting to 50ms: ![image](https://user-images.githubusercontent.com/3752225/67116621-a7d9ca00-f195-11e9-9a72-dad3d62562ed.png) 2. Try editing a Python file e.g. type out: ```python import numpy as np import matplotlib.pyplot as plt ``` ### What is the expected output? What do you see instead? __Expected__ As a user, I would expect minimal lag. __Actual__ There is noticeable lag in between typing a character and it being inserted into the buffer. This happens regardless of what completions engines are enabled and the value of the setting `automatic_completions_after_chars`. It appears as though something is blocking the UI thread in between keystroke → completion request → buffer update in UI. ## Versions <!--- You can get this information from Help > About Spyder... or (if Spyder won't launch) the "conda list" command from the Anaconda Prompt/Terminal/command line. ---> * Spyder version: 4.0.0.dev0 (Commit: c2b7673b1) * Python version: 3.7.4 64-bit * Qt version: Qt 5.9.6 * PyQt version: PyQt5 5.9.2 * Operating System name/version: Darwin 18.7.0 ### Dependencies <!--- Please go to the menu entry Help > Dependencies, press the Copy to clipboard button and paste below ---> ``` cloudpickle >=0.5.0 : 1.2.2 (OK) pygments >=2.0 : 2.4.2 (OK) qtconsole >=4.5.5 : 4.5.5 (OK) nbconvert >=4.0 : 5.6.0 (OK) sphinx >=0.6.6 : 2.2.0 (OK) pylint >=0.25 : 2.4.2 (OK) psutil >=0.3 : 5.6.3 (OK) qtawesome >=0.5.7 : 0.6.0 (OK) qtpy >=1.5.0 : 1.9.0 (OK) pickleshare >=0.4 : 0.7.5 (OK) zmq >=17 : 18.1.0 (OK) chardet >=2.0.0 : 3.0.4 (OK) numpydoc >=0.6.0 : 0.9.1 (OK) spyder_kernels >=1.5.0;<2.0.0: 1.6.0 (OK) qdarkstyle >=2.7 : 2.7 (OK) atomicwrites >=1.2.0 : 1.3.0 (OK) diff_match_patch >=20181111 : 20181111 (OK) watchdog : None (OK) keyring : None (OK) pexpect >=4.4.0 : 4.7.0 (OK) pympler : None (OK) sympy >=0.7.3 : None (NOK) cython >=0.21 : None (NOK) IPython >=4.0 : 7.8.0 (OK) matplotlib >=2.0.0 : None (NOK) pandas >=0.13.1 : None (NOK) numpy >=1.7 : None (NOK) scipy >=0.17.0 : None (NOK) pyls >=0.28.2;<0.29.0 : 0.28.3 (OK) rtree >=0.8.3 : 0.8.3 (OK) ```
non_test
decreasing automatic completions after ms setting causes lag when typing problem description changing the setting automatic completions after ms to a low but reasonable from user s point of view number causes noticeable lag when editing code what steps reproduce the problem change the automatic completions after ms setting to try editing a python file e g type out python import numpy as np import matplotlib pyplot as plt what is the expected output what do you see instead expected as a user i would expect minimal lag actual there is noticeable lag in between typing a character and it being inserted into the buffer this happens regardless of what completions engines are enabled and the value of the setting automatic completions after chars it appears as though something is blocking the ui thread in between keystroke → completion request → buffer update in ui versions about spyder or if spyder won t launch the conda list command from the anaconda prompt terminal command line spyder version commit python version bit qt version qt pyqt version operating system name version darwin dependencies dependencies press the copy to clipboard button and paste below cloudpickle ok pygments ok qtconsole ok nbconvert ok sphinx ok pylint ok psutil ok qtawesome ok qtpy ok pickleshare ok zmq ok chardet ok numpydoc ok spyder kernels ok qdarkstyle ok atomicwrites ok diff match patch ok watchdog none ok keyring none ok pexpect ok pympler none ok sympy none nok cython none nok ipython ok matplotlib none nok pandas none nok numpy none nok scipy none nok pyls ok rtree ok
0
89,917
8,216,922,263
IssuesEvent
2018-09-05 10:41:02
nodejs/node
https://api.github.com/repos/nodejs/node
closed
Investigate flaky test/async-hooks/test-callback-error
CI / flaky test async_hooks macos test
<!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> Running `test/async-hooks/test-callback-error.js` test on the master branch(no local changes) reports failures on async hooks tests. * **Version**: v9.0.0-pre * **Platform**: Darwin Kernel Version 15.5.0: Tue Apr 19 18:36:36 PDT 2016; root:xnu-3248.50.21~8/RELEASE_X86_64 x86_64 * **Subsystem**: async_hooks <!-- Enter your issue details below this comment. --> ``` $ ./node ./test/async-hooks/test-callback-error.js start case 1 end case 1: 85.844ms start case 2 end case 2: 86.863ms start case 3 end case 3: 4.478ms Error: test_callback_abort at ActivityCollector.initHooks.oninit.common.mustCall (/Users/efreiberg/Documents/node_demo_2/node/test/async-hooks/test-callback-error.js:36:45) at ActivityCollector.oninit (/Users/efreiberg/Documents/node_demo_2/node/test/common/index.js:517:15) at ActivityCollector._init (/Users/efreiberg/Documents/node_demo_2/node/test/async-hooks/init-hooks.js:182:10) at emitInitNative (async_hooks.js:466:43) at Object.emitInitScript [as emitInit] (async_hooks.js:369:3) at Object.<anonymous> (/Users/efreiberg/Documents/node_demo_2/node/test/async-hooks/test-callback-error.js:38:17) at Module._compile (module.js:600:30) at Object.Module._extensions..js (module.js:611:10) at Module.load (module.js:521:32) at tryModuleLoad (module.js:484:12) 1: node::Abort() [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 2: node::Chdir(v8::FunctionCallbackInfo<v8::Value> const&) [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 3: v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 4: v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 5: v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 6: 0x35a344a0463d ``` ``` $ sw_vers -productVersion 10.11.5 ```
2.0
Investigate flaky test/async-hooks/test-callback-error - <!-- Thank you for reporting an issue. This issue tracker is for bugs and issues found within Node.js core. If you require more general support please file an issue on our help repo. https://github.com/nodejs/help Please fill in as much of the template below as you're able. Version: output of `node -v` Platform: output of `uname -a` (UNIX), or version and 32 or 64-bit (Windows) Subsystem: if known, please specify affected core module name If possible, please provide code that demonstrates the problem, keeping it as simple and free of external dependencies as you are able. --> Running `test/async-hooks/test-callback-error.js` test on the master branch(no local changes) reports failures on async hooks tests. * **Version**: v9.0.0-pre * **Platform**: Darwin Kernel Version 15.5.0: Tue Apr 19 18:36:36 PDT 2016; root:xnu-3248.50.21~8/RELEASE_X86_64 x86_64 * **Subsystem**: async_hooks <!-- Enter your issue details below this comment. --> ``` $ ./node ./test/async-hooks/test-callback-error.js start case 1 end case 1: 85.844ms start case 2 end case 2: 86.863ms start case 3 end case 3: 4.478ms Error: test_callback_abort at ActivityCollector.initHooks.oninit.common.mustCall (/Users/efreiberg/Documents/node_demo_2/node/test/async-hooks/test-callback-error.js:36:45) at ActivityCollector.oninit (/Users/efreiberg/Documents/node_demo_2/node/test/common/index.js:517:15) at ActivityCollector._init (/Users/efreiberg/Documents/node_demo_2/node/test/async-hooks/init-hooks.js:182:10) at emitInitNative (async_hooks.js:466:43) at Object.emitInitScript [as emitInit] (async_hooks.js:369:3) at Object.<anonymous> (/Users/efreiberg/Documents/node_demo_2/node/test/async-hooks/test-callback-error.js:38:17) at Module._compile (module.js:600:30) at Object.Module._extensions..js (module.js:611:10) at Module.load (module.js:521:32) at tryModuleLoad (module.js:484:12) 1: node::Abort() [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 2: node::Chdir(v8::FunctionCallbackInfo<v8::Value> const&) [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 3: v8::internal::FunctionCallbackArguments::Call(void (*)(v8::FunctionCallbackInfo<v8::Value> const&)) [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 4: v8::internal::MaybeHandle<v8::internal::Object> v8::internal::(anonymous namespace)::HandleApiCallHelper<false>(v8::internal::Isolate*, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::HeapObject>, v8::internal::Handle<v8::internal::FunctionTemplateInfo>, v8::internal::Handle<v8::internal::Object>, v8::internal::BuiltinArguments) [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 5: v8::internal::Builtin_Impl_HandleApiCall(v8::internal::BuiltinArguments, v8::internal::Isolate*) [/Users/efreiberg/Documents/node_demo_2/node/out/Release/node] 6: 0x35a344a0463d ``` ``` $ sw_vers -productVersion 10.11.5 ```
test
investigate flaky test async hooks test callback error thank you for reporting an issue this issue tracker is for bugs and issues found within node js core if you require more general support please file an issue on our help repo please fill in as much of the template below as you re able version output of node v platform output of uname a unix or version and or bit windows subsystem if known please specify affected core module name if possible please provide code that demonstrates the problem keeping it as simple and free of external dependencies as you are able running test async hooks test callback error js test on the master branch no local changes reports failures on async hooks tests version pre platform darwin kernel version tue apr pdt root xnu release subsystem async hooks node test async hooks test callback error js start case end case start case end case start case end case error test callback abort at activitycollector inithooks oninit common mustcall users efreiberg documents node demo node test async hooks test callback error js at activitycollector oninit users efreiberg documents node demo node test common index js at activitycollector init users efreiberg documents node demo node test async hooks init hooks js at emitinitnative async hooks js at object emitinitscript async hooks js at object users efreiberg documents node demo node test async hooks test callback error js at module compile module js at object module extensions js module js at module load module js at trymoduleload module js node abort node chdir functioncallbackinfo const internal functioncallbackarguments call void functioncallbackinfo const internal maybehandle internal anonymous namespace handleapicallhelper internal isolate internal handle internal handle internal handle internal handle internal builtinarguments internal builtin impl handleapicall internal builtinarguments internal isolate sw vers productversion
1