Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
631,633
20,156,931,669
IssuesEvent
2022-02-09 17:17:56
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Stricter server-side validation of JSON entities
area/api area/apiserver sig/api-machinery kind/feature priority/important-longterm lifecycle/frozen
In issue #5850 we mentioned that we may consider a stricter validation of the JSON entities (explicitly failing on unknown fields instead of ignoring them).
1.0
Stricter server-side validation of JSON entities - In issue #5850 we mentioned that we may consider a stricter validation of the JSON entities (explicitly failing on unknown fields instead of ignoring them).
non_test
stricter server side validation of json entities in issue we mentioned that we may consider a stricter validation of the json entities explicitly failing on unknown fields instead of ignoring them
0
8,354
2,982,021,861
IssuesEvent
2015-07-17 08:03:40
rust-lang/rust
https://api.github.com/repos/rust-lang/rust
closed
ICE with bounded associated type in where clause of a trait
A-associated-items E-needstest I-ICE
`rustc 1.0.0-dev (3614e1de6 2015-01-14 04:42:01 +0000)` ```rust pub trait Trait where Self::Out: std::fmt::Show { type Out; } ``` ``` thread 'rustc' panicked at 'no entry found for key', /home/tom/src/rust/rust/src/libcore/option.rs:330 stack backtrace: 1: 0x7f4c3e42ca80 - sys::backtrace::write::ha0aae27c13f1e1e2wRt 2: 0x7f4c3e44e460 - failure::on_fail::h6467179db076ae40d5z 3: 0x7f4c3e3bd510 - rt::unwind::begin_unwind_inner::h60ee3cb21c325c186Jz 4: 0x7f4c3e3be040 - rt::unwind::begin_unwind_fmt::h374483fb441173d8DIz 5: 0x7f4c3e44e2c0 - rust_begin_unwind 6: 0x7f4c3e49fb80 - panicking::panic_fmt::h61e6123d33617eebvym 7: 0x7f4c3dc60570 - astconv::ast_ty_to_ty::unboxed_closure.29412 8: 0x7f4c3dc12b10 - astconv::ast_ty_to_ty::h1763fcd7e8e2744azSs 9: 0x7f4c3dc8c960 - collect::ty_generics::he250f27d9578b0d23Vu 10: 0x7f4c3dc73ac0 - collect::trait_def_of_item::h6774115495711e070wu 11: 0x7f4c3dc71690 - collect::CollectTraitDefVisitor<'a, 'tcx>.visit..Visitor<'v>::visit_item::he26bc84d696b35decKt 12: 0x7f4c3dcbc0b0 - check_crate::unboxed_closure.30645 13: 0x7f4c3dcba0f0 - check_crate::h022f16fd72b6c09dney 14: 0x7f4c3e986c70 - driver::phase_3_run_analysis_passes::h0ee4fa5e73acbe9cEEa 15: 0x7f4c3e974a50 - driver::compile_input::hc689346ac2e9c62fAba 16: 0x7f4c3ea36310 - run_compiler::he89321e2dd531b37e5b 17: 0x7f4c3ea34a80 - thunk::F.Invoke<A, R>::invoke::h16839135703482139551 18: 0x7f4c3ea339e0 - rt::unwind::try::try_fn::h363853857083200694 19: 0x7f4c3e4be130 - rust_try_inner 20: 0x7f4c3e4be120 - rust_try 21: 0x7f4c3ea33c90 - thunk::F.Invoke<A, R>::invoke::h16782837970792084443 22: 0x7f4c3e43c100 - sys::thread::thread_start::h5480a52fb17d8858jJw 23: 0x7f4c38b73250 - start_thread 24: 0x7f4c3e06d219 - clone 25: 0x0 - <unknown> ``` Thought this might be a dupe of #20220 #20772 but I don't think so.
1.0
ICE with bounded associated type in where clause of a trait - `rustc 1.0.0-dev (3614e1de6 2015-01-14 04:42:01 +0000)` ```rust pub trait Trait where Self::Out: std::fmt::Show { type Out; } ``` ``` thread 'rustc' panicked at 'no entry found for key', /home/tom/src/rust/rust/src/libcore/option.rs:330 stack backtrace: 1: 0x7f4c3e42ca80 - sys::backtrace::write::ha0aae27c13f1e1e2wRt 2: 0x7f4c3e44e460 - failure::on_fail::h6467179db076ae40d5z 3: 0x7f4c3e3bd510 - rt::unwind::begin_unwind_inner::h60ee3cb21c325c186Jz 4: 0x7f4c3e3be040 - rt::unwind::begin_unwind_fmt::h374483fb441173d8DIz 5: 0x7f4c3e44e2c0 - rust_begin_unwind 6: 0x7f4c3e49fb80 - panicking::panic_fmt::h61e6123d33617eebvym 7: 0x7f4c3dc60570 - astconv::ast_ty_to_ty::unboxed_closure.29412 8: 0x7f4c3dc12b10 - astconv::ast_ty_to_ty::h1763fcd7e8e2744azSs 9: 0x7f4c3dc8c960 - collect::ty_generics::he250f27d9578b0d23Vu 10: 0x7f4c3dc73ac0 - collect::trait_def_of_item::h6774115495711e070wu 11: 0x7f4c3dc71690 - collect::CollectTraitDefVisitor<'a, 'tcx>.visit..Visitor<'v>::visit_item::he26bc84d696b35decKt 12: 0x7f4c3dcbc0b0 - check_crate::unboxed_closure.30645 13: 0x7f4c3dcba0f0 - check_crate::h022f16fd72b6c09dney 14: 0x7f4c3e986c70 - driver::phase_3_run_analysis_passes::h0ee4fa5e73acbe9cEEa 15: 0x7f4c3e974a50 - driver::compile_input::hc689346ac2e9c62fAba 16: 0x7f4c3ea36310 - run_compiler::he89321e2dd531b37e5b 17: 0x7f4c3ea34a80 - thunk::F.Invoke<A, R>::invoke::h16839135703482139551 18: 0x7f4c3ea339e0 - rt::unwind::try::try_fn::h363853857083200694 19: 0x7f4c3e4be130 - rust_try_inner 20: 0x7f4c3e4be120 - rust_try 21: 0x7f4c3ea33c90 - thunk::F.Invoke<A, R>::invoke::h16782837970792084443 22: 0x7f4c3e43c100 - sys::thread::thread_start::h5480a52fb17d8858jJw 23: 0x7f4c38b73250 - start_thread 24: 0x7f4c3e06d219 - clone 25: 0x0 - <unknown> ``` Thought this might be a dupe of #20220 #20772 but I don't think so.
test
ice with bounded associated type in where clause of a trait rustc dev rust pub trait trait where self out std fmt show type out thread rustc panicked at no entry found for key home tom src rust rust src libcore option rs stack backtrace sys backtrace write failure on fail rt unwind begin unwind inner rt unwind begin unwind fmt rust begin unwind panicking panic fmt astconv ast ty to ty unboxed closure astconv ast ty to ty collect ty generics collect trait def of item collect collecttraitdefvisitor visit visitor visit item check crate unboxed closure check crate driver phase run analysis passes driver compile input run compiler thunk f invoke invoke rt unwind try try fn rust try inner rust try thunk f invoke invoke sys thread thread start start thread clone thought this might be a dupe of but i don t think so
1
243,421
20,386,268,737
IssuesEvent
2022-02-22 07:19:12
metaplex-foundation/metaplex
https://api.github.com/repos/metaplex-foundation/metaplex
opened
[Bug]: TypeError: Cannot read properties of undefined (reading 'program')
needs tests bug
### Which package is this bug report for? candy machine cli ### Issue description After running candy machine upload command I tried deleting and reimporting the original .cache file since the mint resulted in unrecognized tokens and a large loss of SOL... I am trying to verify the upload since everything seems to be working and can be viewed in arweave. Please help! I just want to mint a portion to be freely distributed and have another portion able to be minted by the user. ### Command ```shell ts-node js/packages/cli/src/candy-machine-v2-cli.ts verify_upload --keypair ~/.config/solana/mainnet.json ``` ### Relevant log output ```shell wallet public key: DHM1mtD6wSMwSwYGMYkZ7uMPa5jBG67HFqkBWNu83nFf C:\Users\Zgkra\OneDrive\Desktop\PROJECT FILES\metaplex-master\js\packages\cli\src\candy-machine-v2-cli.ts:471 ^ TypeError: Cannot read properties of undefined (reading 'program') at Command.<anonymous> (C:\Users\Zgkra\OneDrive\Desktop\PROJECT FILES\metaplex-master\js\packages\cli\src\candy-machine-v2-cli.ts:471:34) at processTicksAndRejections (node:internal/process/task_queues:96:5) ``` ### Operating system windows 10 ### Priority this issue should have High (immediate attention needed) ### Check the Docs First - [X] I have checked the docs and it didn't solve my issue
1.0
[Bug]: TypeError: Cannot read properties of undefined (reading 'program') - ### Which package is this bug report for? candy machine cli ### Issue description After running candy machine upload command I tried deleting and reimporting the original .cache file since the mint resulted in unrecognized tokens and a large loss of SOL... I am trying to verify the upload since everything seems to be working and can be viewed in arweave. Please help! I just want to mint a portion to be freely distributed and have another portion able to be minted by the user. ### Command ```shell ts-node js/packages/cli/src/candy-machine-v2-cli.ts verify_upload --keypair ~/.config/solana/mainnet.json ``` ### Relevant log output ```shell wallet public key: DHM1mtD6wSMwSwYGMYkZ7uMPa5jBG67HFqkBWNu83nFf C:\Users\Zgkra\OneDrive\Desktop\PROJECT FILES\metaplex-master\js\packages\cli\src\candy-machine-v2-cli.ts:471 ^ TypeError: Cannot read properties of undefined (reading 'program') at Command.<anonymous> (C:\Users\Zgkra\OneDrive\Desktop\PROJECT FILES\metaplex-master\js\packages\cli\src\candy-machine-v2-cli.ts:471:34) at processTicksAndRejections (node:internal/process/task_queues:96:5) ``` ### Operating system windows 10 ### Priority this issue should have High (immediate attention needed) ### Check the Docs First - [X] I have checked the docs and it didn't solve my issue
test
typeerror cannot read properties of undefined reading program which package is this bug report for candy machine cli issue description after running candy machine upload command i tried deleting and reimporting the original cache file since the mint resulted in unrecognized tokens and a large loss of sol i am trying to verify the upload since everything seems to be working and can be viewed in arweave please help i just want to mint a portion to be freely distributed and have another portion able to be minted by the user command shell ts node js packages cli src candy machine cli ts verify upload keypair config solana mainnet json relevant log output shell wallet public key c users zgkra onedrive desktop project files metaplex master js packages cli src candy machine cli ts typeerror cannot read properties of undefined reading program at command c users zgkra onedrive desktop project files metaplex master js packages cli src candy machine cli ts at processticksandrejections node internal process task queues operating system windows priority this issue should have high immediate attention needed check the docs first i have checked the docs and it didn t solve my issue
1
319,276
27,361,830,612
IssuesEvent
2023-02-27 16:20:23
epam/badgerdoc
https://api.github.com/repos/epam/badgerdoc
closed
Document-level labels are not visible during validation
enhancement priority front-end back-end testing
_Came from UAT_ **Preconditions:** At least 3 users exist (2 annotators and 1 validator). **Steps:** 1. Annotate a document with some document-level labels as both annotators. 2. Log in as validator, start the corresponding validation task. **Result:** Document-level labels assigned by annotators are not visible. The bug occurs under certain conditions. It appears that one possible condition is that validator is not annotator of the same document at the same time, or another possible condition is that validator is also annotator of the document and there are at least 2 other annotators.
1.0
Document-level labels are not visible during validation - _Came from UAT_ **Preconditions:** At least 3 users exist (2 annotators and 1 validator). **Steps:** 1. Annotate a document with some document-level labels as both annotators. 2. Log in as validator, start the corresponding validation task. **Result:** Document-level labels assigned by annotators are not visible. The bug occurs under certain conditions. It appears that one possible condition is that validator is not annotator of the same document at the same time, or another possible condition is that validator is also annotator of the document and there are at least 2 other annotators.
test
document level labels are not visible during validation came from uat preconditions at least users exist annotators and validator steps annotate a document with some document level labels as both annotators log in as validator start the corresponding validation task result document level labels assigned by annotators are not visible the bug occurs under certain conditions it appears that one possible condition is that validator is not annotator of the same document at the same time or another possible condition is that validator is also annotator of the document and there are at least other annotators
1
134,509
10,918,057,092
IssuesEvent
2019-11-21 16:12:47
reproducible-biomedical-modeling/CRBM-Viz
https://api.github.com/repos/reproducible-biomedical-modeling/CRBM-Viz
closed
The indiviudal components should be mocked
testing todo :spiral_notepad:
https://github.com/reproducible-biomedical-modeling/CRBM-Viz/blob/1b71baf50c90d234805e9c76d1c1347742c9816e/CRBM-Viz/src/app/Layout/navigation/navigation.component.spec.ts#L4-L9 --- ###### This issue was generated by [todo](https://todo.jasonet.co) based on a `TODO` comment in 1b71baf50c90d234805e9c76d1c1347742c9816e. It&#x27;s been assigned to @bilalshaikh42 because they committed the code.
1.0
The indiviudal components should be mocked - https://github.com/reproducible-biomedical-modeling/CRBM-Viz/blob/1b71baf50c90d234805e9c76d1c1347742c9816e/CRBM-Viz/src/app/Layout/navigation/navigation.component.spec.ts#L4-L9 --- ###### This issue was generated by [todo](https://todo.jasonet.co) based on a `TODO` comment in 1b71baf50c90d234805e9c76d1c1347742c9816e. It&#x27;s been assigned to @bilalshaikh42 because they committed the code.
test
the indiviudal components should be mocked this issue was generated by based on a todo comment in it s been assigned to because they committed the code
1
255,755
21,951,970,022
IssuesEvent
2022-05-24 08:43:39
stores-cedcommerce/Internal---Maana-Creation-StoreRedes--29-April-22
https://api.github.com/repos/stores-cedcommerce/Internal---Maana-Creation-StoreRedes--29-April-22
closed
Back to login page button missing on create account page
Functional / bug Open Ready to test Account pages Desktop
Bug - Back to login page button missing on create account. Exp - Add button for back to login page on create account. Ref Link - https://nimb.ws/4i2QTe
1.0
Back to login page button missing on create account page - Bug - Back to login page button missing on create account. Exp - Add button for back to login page on create account. Ref Link - https://nimb.ws/4i2QTe
test
back to login page button missing on create account page bug back to login page button missing on create account exp add button for back to login page on create account ref link
1
434,043
30,441,122,915
IssuesEvent
2023-07-15 04:23:19
DucTran999/fashion-shop
https://api.github.com/repos/DucTran999/fashion-shop
closed
Have struggled on installing the projects
documentation
The current project installation instruction is not specific enough so major users find it hard to do.
1.0
Have struggled on installing the projects - The current project installation instruction is not specific enough so major users find it hard to do.
non_test
have struggled on installing the projects the current project installation instruction is not specific enough so major users find it hard to do
0
234,078
19,093,803,151
IssuesEvent
2021-11-29 14:47:58
BCSDLab/KOALA_FRONT_END
https://api.github.com/repos/BCSDLab/KOALA_FRONT_END
closed
KOALA ๊ฐœ๋ฐœ ์ค€๋น„ ๋ฐ ์„ค๊ณ„
Docs Test Chore
--- type: Init Project about: ํ”„๋กœ์ ํŠธ ๊ฐœ๋ฐœ ์ค€๋น„ ๋ฐ ์„ค๊ณ„ title: "[FEATURE] " assignees: ''daepan" --- ## ์ง„ํ–‰ ์‚ฌํ•ญ - [x] ํด๋” ๊ตฌ์กฐ ์ •๋ฆฌ - [x] ์™ธ๋ถ€๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ ์„ธํŒ…
1.0
KOALA ๊ฐœ๋ฐœ ์ค€๋น„ ๋ฐ ์„ค๊ณ„ - --- type: Init Project about: ํ”„๋กœ์ ํŠธ ๊ฐœ๋ฐœ ์ค€๋น„ ๋ฐ ์„ค๊ณ„ title: "[FEATURE] " assignees: ''daepan" --- ## ์ง„ํ–‰ ์‚ฌํ•ญ - [x] ํด๋” ๊ตฌ์กฐ ์ •๋ฆฌ - [x] ์™ธ๋ถ€๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ ์„ธํŒ…
test
koala ๊ฐœ๋ฐœ ์ค€๋น„ ๋ฐ ์„ค๊ณ„ type init project about ํ”„๋กœ์ ํŠธ ๊ฐœ๋ฐœ ์ค€๋น„ ๋ฐ ์„ค๊ณ„ title assignees daepan ์ง„ํ–‰ ์‚ฌํ•ญ ํด๋” ๊ตฌ์กฐ ์ •๋ฆฌ ์™ธ๋ถ€๋ผ์ด๋ธŒ๋Ÿฌ๋ฆฌ ์„ธํŒ…
1
85,358
7,966,965,273
IssuesEvent
2018-07-15 07:38:53
coala/coala
https://api.github.com/repos/coala/coala
closed
CI build failure because of httpstat.us SSL certificate error
area/CI area/tests type/bug
https://travis-ci.org/virresh/coala/jobs/402970785#L1642 https://ci.appveyor.com/project/anctartica/coala https://circleci.com/gh/virresh/coala/6 difficulty/low importance/high
1.0
CI build failure because of httpstat.us SSL certificate error - https://travis-ci.org/virresh/coala/jobs/402970785#L1642 https://ci.appveyor.com/project/anctartica/coala https://circleci.com/gh/virresh/coala/6 difficulty/low importance/high
test
ci build failure because of httpstat us ssl certificate error difficulty low importance high
1
216,581
16,771,778,005
IssuesEvent
2021-06-14 15:34:25
CBIIT/ccdhwebportal
https://api.github.com/repos/CBIIT/ccdhwebportal
closed
Fix Navigation Menu
Ready-to-Test bug
Menu behavior is erratic for mobile and when the page loads. CSS requires some adjustments during load and at media breakpoints.
1.0
Fix Navigation Menu - Menu behavior is erratic for mobile and when the page loads. CSS requires some adjustments during load and at media breakpoints.
test
fix navigation menu menu behavior is erratic for mobile and when the page loads css requires some adjustments during load and at media breakpoints
1
44,483
13,055,409,054
IssuesEvent
2020-07-30 01:35:44
jgeraigery/moo-ui-select
https://api.github.com/repos/jgeraigery/moo-ui-select
opened
CVE-2020-11023 (Medium) detected in jquery-1.8.1.min.js, jquery-1.11.1.min.js
security vulnerability
## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.8.1.min.js</b>, <b>jquery-1.11.1.min.js</b></p></summary> <p> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/moo-ui-select/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: /moo-ui-select/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/moo-ui-select/node_modules/policyfile/doc/index.html</p> <p>Path to vulnerable library: /moo-ui-select/node_modules/policyfile/doc/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.1.min.js** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.8.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.11.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 3.5.0"}],"vulnerabilityIdentifier":"CVE-2020-11023","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing \u003coption\u003e elements from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
True
CVE-2020-11023 (Medium) detected in jquery-1.8.1.min.js, jquery-1.11.1.min.js - ## CVE-2020-11023 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>jquery-1.8.1.min.js</b>, <b>jquery-1.11.1.min.js</b></p></summary> <p> <details><summary><b>jquery-1.8.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.8.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/moo-ui-select/node_modules/redeyed/examples/browser/index.html</p> <p>Path to vulnerable library: /moo-ui-select/node_modules/redeyed/examples/browser/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.8.1.min.js** (Vulnerable Library) </details> <details><summary><b>jquery-1.11.1.min.js</b></p></summary> <p>JavaScript library for DOM operations</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js">https://cdnjs.cloudflare.com/ajax/libs/jquery/1.11.1/jquery.min.js</a></p> <p>Path to dependency file: /tmp/ws-scm/moo-ui-select/node_modules/policyfile/doc/index.html</p> <p>Path to vulnerable library: /moo-ui-select/node_modules/policyfile/doc/index.html</p> <p> Dependency Hierarchy: - :x: **jquery-1.11.1.min.js** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing <option> elements from untrusted sources - even after sanitizing it - to one of jQuery's DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0. <p>Publish Date: 2020-04-29 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023>CVE-2020-11023</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>6.1</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: Required - Scope: Changed - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: None </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2020-11023</a></p> <p>Release Date: 2020-04-29</p> <p>Fix Resolution: jquery - 3.5.0</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":true,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.8.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.8.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 3.5.0"},{"packageType":"JavaScript","packageName":"jquery","packageVersion":"1.11.1","isTransitiveDependency":false,"dependencyTree":"jquery:1.11.1","isMinimumFixVersionAvailable":true,"minimumFixVersion":"jquery - 3.5.0"}],"vulnerabilityIdentifier":"CVE-2020-11023","vulnerabilityDetails":"In jQuery versions greater than or equal to 1.0.3 and before 3.5.0, passing HTML containing \u003coption\u003e elements from untrusted sources - even after sanitizing it - to one of jQuery\u0027s DOM manipulation methods (i.e. .html(), .append(), and others) may execute untrusted code. This problem is patched in jQuery 3.5.0.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-11023","cvss3Severity":"medium","cvss3Score":"6.1","cvss3Metrics":{"A":"None","AC":"Low","PR":"None","S":"Changed","C":"Low","UI":"Required","AV":"Network","I":"Low"},"extraData":{}}</REMEDIATE> -->
non_test
cve medium detected in jquery min js jquery min js cve medium severity vulnerability vulnerable libraries jquery min js jquery min js jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm moo ui select node modules redeyed examples browser index html path to vulnerable library moo ui select node modules redeyed examples browser index html dependency hierarchy x jquery min js vulnerable library jquery min js javascript library for dom operations library home page a href path to dependency file tmp ws scm moo ui select node modules policyfile doc index html path to vulnerable library moo ui select node modules policyfile doc index html dependency hierarchy x jquery min js vulnerable library vulnerability details in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery s dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction required scope changed impact metrics confidentiality impact low integrity impact low availability impact none for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution jquery isopenpronvulnerability true ispackagebased true isdefaultbranch true packages vulnerabilityidentifier cve vulnerabilitydetails in jquery versions greater than or equal to and before passing html containing elements from untrusted sources even after sanitizing it to one of jquery dom manipulation methods i e html append and others may execute untrusted code this problem is patched in jquery vulnerabilityurl
0
12,165
3,256,129,989
IssuesEvent
2015-10-20 12:25:48
TEAMMATES/repo
https://api.github.com/repos/TEAMMATES/repo
closed
Bug in AppPage.verifyHtml causing non-equal htmls to pass
a-Testing d.Moderate p.Medium s.Accepted t.Bug
_From [arnold.k...@gmail.com](https://code.google.com/u/105779801956403207546/) on January 09, 2014 15:52:46_ I encountered a strange bug where verifyHtml would "confirm" that two htmls are equal even though the expected html is missing some parts. I attached the html that caused this. running actual.html against expected.html will pass the test (it shouldn't as expected.html is missing the parts shown in missing part.png). Adding a random line at the point where the missing part is supposed to appear will cause the test to fail. In expected-fail.html I added the line THISWILLFAILTHETEST at the point where the missign parts is supposed to be and the test now properly fails. **Attachment:** [actual.html expected.html expected-fail.html missing part.png](http://code.google.com/p/teammatespes/issues/detail?id=1469) _Original issue: http://code.google.com/p/teammatespes/issues/detail?id=1469_
1.0
Bug in AppPage.verifyHtml causing non-equal htmls to pass - _From [arnold.k...@gmail.com](https://code.google.com/u/105779801956403207546/) on January 09, 2014 15:52:46_ I encountered a strange bug where verifyHtml would "confirm" that two htmls are equal even though the expected html is missing some parts. I attached the html that caused this. running actual.html against expected.html will pass the test (it shouldn't as expected.html is missing the parts shown in missing part.png). Adding a random line at the point where the missing part is supposed to appear will cause the test to fail. In expected-fail.html I added the line THISWILLFAILTHETEST at the point where the missign parts is supposed to be and the test now properly fails. **Attachment:** [actual.html expected.html expected-fail.html missing part.png](http://code.google.com/p/teammatespes/issues/detail?id=1469) _Original issue: http://code.google.com/p/teammatespes/issues/detail?id=1469_
test
bug in apppage verifyhtml causing non equal htmls to pass from on january i encountered a strange bug where verifyhtml would confirm that two htmls are equal even though the expected html is missing some parts i attached the html that caused this running actual html against expected html will pass the test it shouldn t as expected html is missing the parts shown in missing part png adding a random line at the point where the missing part is supposed to appear will cause the test to fail in expected fail html i added the line thiswillfailthetest at the point where the missign parts is supposed to be and the test now properly fails attachment original issue
1
97,160
3,985,504,914
IssuesEvent
2016-05-07 22:53:21
Brickimedia/brickimedia
https://api.github.com/repos/Brickimedia/brickimedia
closed
Make Refreshed less brickimedia-specific
[browser] IE [improv] Enhancement [priority] Mid-high [skin] Refreshed [wiki] Global
The Refreshed skin should be less brickimedia specific so other wikis don't have to download unnecessary files they're not going to use (files should only be there if they're going to be used globally otherwise local files should be on our own local site), this means: - [ ] The `/images` folder shouldn't include brickimedia wiki logos, they should just include the icons required for mobile. - [x] Remove the `.png` wordmark fallbacks since we're officially not supporting IE 8 anymore. Also helps with performance since loading less files that aren't being used - [ ] Requires some rewriting of Refreshed header component - [ ] That means all these logos should turn into variables required to be entered into [brickimedia/LocalSettings](https://github.com/Brickimedia/LocalSettings), meaning this would require making changes across multiple repositories. (all of our local variable values would be from `image.brickimedia.org` and uploaded to Meta), and for other wikis they would require editing their own `LocalSettings.php` file - [ ] Since this breaks directory paths, this should be counted as a major patch instead of a minor patch. So instead of 3.1.1, this should be included in 4.1.1 and should be reflected on the [Refreshed.php](https://github.com/Brickimedia/Refreshed/blob/master/Refreshed.php) file (Note: I found a good way of versioning that I like in of these comments at https://github.com/alrra/browser-logos/issues/76) - [ ] Reflect these changes on mw docs https://www.mediawiki.org/wiki/Skin:Refreshed
1.0
Make Refreshed less brickimedia-specific - The Refreshed skin should be less brickimedia specific so other wikis don't have to download unnecessary files they're not going to use (files should only be there if they're going to be used globally otherwise local files should be on our own local site), this means: - [ ] The `/images` folder shouldn't include brickimedia wiki logos, they should just include the icons required for mobile. - [x] Remove the `.png` wordmark fallbacks since we're officially not supporting IE 8 anymore. Also helps with performance since loading less files that aren't being used - [ ] Requires some rewriting of Refreshed header component - [ ] That means all these logos should turn into variables required to be entered into [brickimedia/LocalSettings](https://github.com/Brickimedia/LocalSettings), meaning this would require making changes across multiple repositories. (all of our local variable values would be from `image.brickimedia.org` and uploaded to Meta), and for other wikis they would require editing their own `LocalSettings.php` file - [ ] Since this breaks directory paths, this should be counted as a major patch instead of a minor patch. So instead of 3.1.1, this should be included in 4.1.1 and should be reflected on the [Refreshed.php](https://github.com/Brickimedia/Refreshed/blob/master/Refreshed.php) file (Note: I found a good way of versioning that I like in of these comments at https://github.com/alrra/browser-logos/issues/76) - [ ] Reflect these changes on mw docs https://www.mediawiki.org/wiki/Skin:Refreshed
non_test
make refreshed less brickimedia specific the refreshed skin should be less brickimedia specific so other wikis don t have to download unnecessary files they re not going to use files should only be there if they re going to be used globally otherwise local files should be on our own local site this means the images folder shouldn t include brickimedia wiki logos they should just include the icons required for mobile remove the png wordmark fallbacks since we re officially not supporting ie anymore also helps with performance since loading less files that aren t being used requires some rewriting of refreshed header component that means all these logos should turn into variables required to be entered into meaning this would require making changes across multiple repositories all of our local variable values would be from image brickimedia org and uploaded to meta and for other wikis they would require editing their own localsettings php file since this breaks directory paths this should be counted as a major patch instead of a minor patch so instead of this should be included in and should be reflected on the file note i found a good way of versioning that i like in of these comments at reflect these changes on mw docs
0
243,618
20,509,875,931
IssuesEvent
2022-03-01 04:30:21
bhanupalagati/GatorMart
https://api.github.com/repos/bhanupalagati/GatorMart
closed
Cypress Test Case Set 1 for homepage and detailed view
Sprint2 TestCase
First set of test cases for the homepage and detailed view
1.0
Cypress Test Case Set 1 for homepage and detailed view - First set of test cases for the homepage and detailed view
test
cypress test case set for homepage and detailed view first set of test cases for the homepage and detailed view
1
127,769
10,481,961,914
IssuesEvent
2019-09-24 10:50:25
johntruckenbrodt/pyroSAR
https://api.github.com/repos/johntruckenbrodt/pyroSAR
closed
Travis CI testing with Python 3.5+
test
The current Travis OS is Ubuntu Trusty, which comes with Python 3.4. Testing with Python 3.5 or 3.6 is not possible because system site packages need to be enabled and this seems to only work with the OS-installed Python version. Is there a way to still test the other versions?
1.0
Travis CI testing with Python 3.5+ - The current Travis OS is Ubuntu Trusty, which comes with Python 3.4. Testing with Python 3.5 or 3.6 is not possible because system site packages need to be enabled and this seems to only work with the OS-installed Python version. Is there a way to still test the other versions?
test
travis ci testing with python the current travis os is ubuntu trusty which comes with python testing with python or is not possible because system site packages need to be enabled and this seems to only work with the os installed python version is there a way to still test the other versions
1
300,275
25,955,918,232
IssuesEvent
2022-12-18 08:06:34
MichaIng/DietPi
https://api.github.com/repos/MichaIng/DietPi
closed
Open Beta v8.12 | Please help testing and hardening our upcoming release
Information :information_source: Beta :test_tube:
RC version | v8.12.1 ---------- | ------- v8.12.0 PR | #5952 v8.12.1 PR | #5984 Raw changelog | https://github.com/MichaIng/DietPi/blob/beta/CHANGELOG.txt Code changes | https://github.com/MichaIng/DietPi/compare/master...beta Release date | **2022-11-17** ### How to apply: https://github.com/MichaIng/DietPi/blob/master/BRANCH_SYSTEM.md #### Related/solved issues: https://github.com/MichaIng/DietPi/issues?q=is%3Aissue+milestone%3Av8.12 ____ ### Beta v8.12.1 _(2022-12-17)_ #### New images - ROCK 5B | Support for Radxa's new flagship SBC has been added to DietPi with hardware ID 78. Many thanks to @docgalaxyblock for doing this request: https://github.com/MichaIng/DietPi/discussions/5247 - NanoPi R6S | Support for FriendlyELEC's new flagship router SBC has been added to DietPi with hardware ID 79. Many thanks to FriendlyELEC for sending us free developer samples. #### Enhancements - General | Informational kernel logs to console are now omitted, to avoid overlaps with login banner and informational logs, like regular network state changes caused by Docker. Many thanks to `@TRENT7` and `@vontainment` for reporting related inconveniences: https://dietpi.com/forum/t/unfinished-services-during-boot-at-login-prompt/15145, https://dietpi.com/forum/t/network-state-filling-up-terminal/14401 - Radxa Zero | USB OTG is now enabled by default, so that the Zero can be attached via USB2/PWR port in gadget mode as peripheral to a host PC. Many thanks to @kazuoteramoto for suggesting this enhancement: https://github.com/MichaIng/DietPi/issues/5931 - DietPi-Backup | The free space check before the backup runs can now be disabled. Many thanks to `@kriznik` for doing this suggestion: https://dietpi.com/forum/t/dietpi-backup-remove-dry-run-pls/13279 - DietPi-Backup | On NFS target drives, a fixed block size of 4096 bytes is now assumed. "stat -fc '%s' /dir" reports a very large network/protocol transfer-related size - DietPi-Software | Java 17 is now installed on ARMv7/8 Raspberry Pi hardware with ARMv6 (Raspbian) image. The binary packages are now provided by Raspbian, as well as via Raspberry Pi repository. - DietPi-Software | Koel: Updated dependencies, especially removed redundant PHP Composer call and dependency. - DietPi-Software | Grafana: For ARMv6 RPi models, RPi 1 and Zero (1), bumped version to latest stable v9.3.0, update via reinstall: dietpi-software reinstall 77 - DietPi-Software | myMPD: On fresh installs, SSL is now enabled by default on the same port 1333 which was previously used for plain HTTP. For plain HTTP, now port 1332 is used, which redirects to HTTPS automatically, but currently cannot be disabled completely. - DietPi-Software | RPi Cam Web Interface: Renamed to correct project name and removed obsolete motion dependency. - DietPi-Software | Node-RED: The service user is now added to the "dialout" group by default, to allow using serial console nodes. Many tanks to `@devifast` for doing this suggestion: https://dietpi.com/forum/t/cannot-connect-serial-port-uart-to-node-red-raspberry-pi/15215 - DietPi-Software | Shairport Sync: Updated to latest version 4.1.1. The update be applied via reinstall: dietpi-software reinstall 37 #### Bug fixes - ROCK Pi 4 | Resolved an issue where since latest kernel upgrade 3.5mm audio did not work. Many thanks to @jalsco for reporting this issue: https://github.com/MichaIng/DietPi/issues/5941 - Odroid C2 | Resolved an issue where USB devices were not detected. Many thanks to @esage85 for reporting this issue: https://github.com/MichaIng/DietPi/issues/5963 - DietPi-Software | WireGuard: It was incorrectly still offered for legacy Odroid systems with vendor kernel, while recent WireGuard does not support those old kernel versions anymore. We hence did now disable the WireGuard install options for those. Many thanks to `@brtravel` for reporting the failed install: https://dietpi.com/forum/t/wireguard-install-failed-now-cant-do-anything-else-with-dietpi-software/15173 - DietPi-Software | PaperMC: Resolved an issue where the installation failed on Raspberry Pi ARMv6 images (with ARMv7/8 hardware), since latest PaperMC cannot run on Java 11. Since Java 17 binary packages are finally available via Raspbian and Raspberry Pi repositories, those can be installed now, allowing PaperMC to run. Many thanks to @blueyshark for reporting this issue: https://github.com/MichaIng/DietPi/issues/5668 - DietPi-Software | Grafana: Resolved an issue where the uninstall failed when trying to remove the "grafana" user. Many thanks to @enviousjag for reporting this issue: https://github.com/MichaIng/DietPi/issues/5935 - DietPi-Software | myMPD: Resolved an issue where the installation as well as the service start after latest myMPD release failed. Many thanks to @jalsco and @sofad for reporting these issues: https://github.com/MichaIng/DietPi/issues/5936, https://github.com/MichaIng/DietPi/issues/5919 - DietPi-Software | PiVPN: Resolved an issue where the unintended reboot from within the dedicated installer was offered again, and that dhcpcd was unintentionally installed. Many thanks to `@dg3` and `@jetlagalex` for reporting dhcpcd related issues: https://dietpi.com/forum/t/static-ip-changed-but-retains-old-address/15175, https://dietpi.com/forum/t/dietpi-config-dns-settings-not-persisting/15204
1.0
Open Beta v8.12 | Please help testing and hardening our upcoming release - RC version | v8.12.1 ---------- | ------- v8.12.0 PR | #5952 v8.12.1 PR | #5984 Raw changelog | https://github.com/MichaIng/DietPi/blob/beta/CHANGELOG.txt Code changes | https://github.com/MichaIng/DietPi/compare/master...beta Release date | **2022-11-17** ### How to apply: https://github.com/MichaIng/DietPi/blob/master/BRANCH_SYSTEM.md #### Related/solved issues: https://github.com/MichaIng/DietPi/issues?q=is%3Aissue+milestone%3Av8.12 ____ ### Beta v8.12.1 _(2022-12-17)_ #### New images - ROCK 5B | Support for Radxa's new flagship SBC has been added to DietPi with hardware ID 78. Many thanks to @docgalaxyblock for doing this request: https://github.com/MichaIng/DietPi/discussions/5247 - NanoPi R6S | Support for FriendlyELEC's new flagship router SBC has been added to DietPi with hardware ID 79. Many thanks to FriendlyELEC for sending us free developer samples. #### Enhancements - General | Informational kernel logs to console are now omitted, to avoid overlaps with login banner and informational logs, like regular network state changes caused by Docker. Many thanks to `@TRENT7` and `@vontainment` for reporting related inconveniences: https://dietpi.com/forum/t/unfinished-services-during-boot-at-login-prompt/15145, https://dietpi.com/forum/t/network-state-filling-up-terminal/14401 - Radxa Zero | USB OTG is now enabled by default, so that the Zero can be attached via USB2/PWR port in gadget mode as peripheral to a host PC. Many thanks to @kazuoteramoto for suggesting this enhancement: https://github.com/MichaIng/DietPi/issues/5931 - DietPi-Backup | The free space check before the backup runs can now be disabled. Many thanks to `@kriznik` for doing this suggestion: https://dietpi.com/forum/t/dietpi-backup-remove-dry-run-pls/13279 - DietPi-Backup | On NFS target drives, a fixed block size of 4096 bytes is now assumed. "stat -fc '%s' /dir" reports a very large network/protocol transfer-related size - DietPi-Software | Java 17 is now installed on ARMv7/8 Raspberry Pi hardware with ARMv6 (Raspbian) image. The binary packages are now provided by Raspbian, as well as via Raspberry Pi repository. - DietPi-Software | Koel: Updated dependencies, especially removed redundant PHP Composer call and dependency. - DietPi-Software | Grafana: For ARMv6 RPi models, RPi 1 and Zero (1), bumped version to latest stable v9.3.0, update via reinstall: dietpi-software reinstall 77 - DietPi-Software | myMPD: On fresh installs, SSL is now enabled by default on the same port 1333 which was previously used for plain HTTP. For plain HTTP, now port 1332 is used, which redirects to HTTPS automatically, but currently cannot be disabled completely. - DietPi-Software | RPi Cam Web Interface: Renamed to correct project name and removed obsolete motion dependency. - DietPi-Software | Node-RED: The service user is now added to the "dialout" group by default, to allow using serial console nodes. Many tanks to `@devifast` for doing this suggestion: https://dietpi.com/forum/t/cannot-connect-serial-port-uart-to-node-red-raspberry-pi/15215 - DietPi-Software | Shairport Sync: Updated to latest version 4.1.1. The update be applied via reinstall: dietpi-software reinstall 37 #### Bug fixes - ROCK Pi 4 | Resolved an issue where since latest kernel upgrade 3.5mm audio did not work. Many thanks to @jalsco for reporting this issue: https://github.com/MichaIng/DietPi/issues/5941 - Odroid C2 | Resolved an issue where USB devices were not detected. Many thanks to @esage85 for reporting this issue: https://github.com/MichaIng/DietPi/issues/5963 - DietPi-Software | WireGuard: It was incorrectly still offered for legacy Odroid systems with vendor kernel, while recent WireGuard does not support those old kernel versions anymore. We hence did now disable the WireGuard install options for those. Many thanks to `@brtravel` for reporting the failed install: https://dietpi.com/forum/t/wireguard-install-failed-now-cant-do-anything-else-with-dietpi-software/15173 - DietPi-Software | PaperMC: Resolved an issue where the installation failed on Raspberry Pi ARMv6 images (with ARMv7/8 hardware), since latest PaperMC cannot run on Java 11. Since Java 17 binary packages are finally available via Raspbian and Raspberry Pi repositories, those can be installed now, allowing PaperMC to run. Many thanks to @blueyshark for reporting this issue: https://github.com/MichaIng/DietPi/issues/5668 - DietPi-Software | Grafana: Resolved an issue where the uninstall failed when trying to remove the "grafana" user. Many thanks to @enviousjag for reporting this issue: https://github.com/MichaIng/DietPi/issues/5935 - DietPi-Software | myMPD: Resolved an issue where the installation as well as the service start after latest myMPD release failed. Many thanks to @jalsco and @sofad for reporting these issues: https://github.com/MichaIng/DietPi/issues/5936, https://github.com/MichaIng/DietPi/issues/5919 - DietPi-Software | PiVPN: Resolved an issue where the unintended reboot from within the dedicated installer was offered again, and that dhcpcd was unintentionally installed. Many thanks to `@dg3` and `@jetlagalex` for reporting dhcpcd related issues: https://dietpi.com/forum/t/static-ip-changed-but-retains-old-address/15175, https://dietpi.com/forum/t/dietpi-config-dns-settings-not-persisting/15204
test
open beta please help testing and hardening our upcoming release rc version pr pr raw changelog code changes release date how to apply related solved issues beta new images rock support for radxa s new flagship sbc has been added to dietpi with hardware id many thanks to docgalaxyblock for doing this request nanopi support for friendlyelec s new flagship router sbc has been added to dietpi with hardware id many thanks to friendlyelec for sending us free developer samples enhancements general informational kernel logs to console are now omitted to avoid overlaps with login banner and informational logs like regular network state changes caused by docker many thanks to and vontainment for reporting related inconveniences radxa zero usb otg is now enabled by default so that the zero can be attached via pwr port in gadget mode as peripheral to a host pc many thanks to kazuoteramoto for suggesting this enhancement dietpi backup the free space check before the backup runs can now be disabled many thanks to kriznik for doing this suggestion dietpi backup on nfs target drives a fixed block size of bytes is now assumed stat fc s dir reports a very large network protocol transfer related size dietpi software java is now installed on raspberry pi hardware with raspbian image the binary packages are now provided by raspbian as well as via raspberry pi repository dietpi software koel updated dependencies especially removed redundant php composer call and dependency dietpi software grafana for rpi models rpi and zero bumped version to latest stable update via reinstall dietpi software reinstall dietpi software mympd on fresh installs ssl is now enabled by default on the same port which was previously used for plain http for plain http now port is used which redirects to https automatically but currently cannot be disabled completely dietpi software rpi cam web interface renamed to correct project name and removed obsolete motion dependency dietpi software node red the service user is now added to the dialout group by default to allow using serial console nodes many tanks to devifast for doing this suggestion dietpi software shairport sync updated to latest version the update be applied via reinstall dietpi software reinstall bug fixes rock pi resolved an issue where since latest kernel upgrade audio did not work many thanks to jalsco for reporting this issue odroid resolved an issue where usb devices were not detected many thanks to for reporting this issue dietpi software wireguard it was incorrectly still offered for legacy odroid systems with vendor kernel while recent wireguard does not support those old kernel versions anymore we hence did now disable the wireguard install options for those many thanks to brtravel for reporting the failed install dietpi software papermc resolved an issue where the installation failed on raspberry pi images with hardware since latest papermc cannot run on java since java binary packages are finally available via raspbian and raspberry pi repositories those can be installed now allowing papermc to run many thanks to blueyshark for reporting this issue dietpi software grafana resolved an issue where the uninstall failed when trying to remove the grafana user many thanks to enviousjag for reporting this issue dietpi software mympd resolved an issue where the installation as well as the service start after latest mympd release failed many thanks to jalsco and sofad for reporting these issues dietpi software pivpn resolved an issue where the unintended reboot from within the dedicated installer was offered again and that dhcpcd was unintentionally installed many thanks to and jetlagalex for reporting dhcpcd related issues
1
772,901
27,141,014,382
IssuesEvent
2023-02-16 16:20:42
keycloak/keycloak-ui
https://api.github.com/repos/keycloak/keycloak-ui
closed
Missing user info encryption config
kind/bug priority/critical area/admin/ui team/ui
### Before reporting an issue - [X] I have searched existing issues - [X] I have reproduced the issue with the latest release ### Area admin/ui ### Describe the bug In the config client in the section 'Fine grain OpenID Connect configuration' we don't have these two configs: 1- User Info Response Encryption Key Management Algorithm 2- User Info Response Encryption Content Encryption Algorithm ### Version 20.0.3 ### Expected behavior Able to view and edit the value of these two configs: 1- User Info Response Encryption Key Management Algorithm 2- User Info Response Encryption Content Encryption Algorithm ### Actual behavior Not able to view and edit the value of these two configs: 1- User Info Response Encryption Key Management Algorithm 2- User Info Response Encryption Content Encryption Algorithm ### How to Reproduce? Create a new client, open the tab 'Advanced', in the section 'Fine grain OpenID Connect configuration', the two configs are missed. ### Anything else? _No response_
1.0
Missing user info encryption config - ### Before reporting an issue - [X] I have searched existing issues - [X] I have reproduced the issue with the latest release ### Area admin/ui ### Describe the bug In the config client in the section 'Fine grain OpenID Connect configuration' we don't have these two configs: 1- User Info Response Encryption Key Management Algorithm 2- User Info Response Encryption Content Encryption Algorithm ### Version 20.0.3 ### Expected behavior Able to view and edit the value of these two configs: 1- User Info Response Encryption Key Management Algorithm 2- User Info Response Encryption Content Encryption Algorithm ### Actual behavior Not able to view and edit the value of these two configs: 1- User Info Response Encryption Key Management Algorithm 2- User Info Response Encryption Content Encryption Algorithm ### How to Reproduce? Create a new client, open the tab 'Advanced', in the section 'Fine grain OpenID Connect configuration', the two configs are missed. ### Anything else? _No response_
non_test
missing user info encryption config before reporting an issue i have searched existing issues i have reproduced the issue with the latest release area admin ui describe the bug in the config client in the section fine grain openid connect configuration we don t have these two configs user info response encryption key management algorithm user info response encryption content encryption algorithm version expected behavior able to view and edit the value of these two configs user info response encryption key management algorithm user info response encryption content encryption algorithm actual behavior not able to view and edit the value of these two configs user info response encryption key management algorithm user info response encryption content encryption algorithm how to reproduce create a new client open the tab advanced in the section fine grain openid connect configuration the two configs are missed anything else no response
0
54,099
13,391,261,481
IssuesEvent
2020-09-02 22:09:36
solex2006/SELIProject
https://api.github.com/repos/solex2006/SELIProject
opened
Image is being requested for all scenes
:bug:bug P3 - Normal S2 - Minor Defects env:next release module:STORING TELLING type:functional
I think this is a error ## Expected Behavior <!--- Tell us what should happen --> If a video have been already upload in scene, an image should not be required ## Current Behavior <!--- Tell us what happens instead of the expected behavior --> I can't publish a story while not all scenes has an image file. ## Environment <!--- Please tell us about your environment --> <!-- Browser: all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView --> https://vm2161.kaj.pouta.csc.fi/ ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- Provide a detailed description of the change or addition you are proposing --> - one image OR one video should be required for a scene ## Related issues <!--- Not obligatory, but reference related issues --> #199
1.0
Image is being requested for all scenes - I think this is a error ## Expected Behavior <!--- Tell us what should happen --> If a video have been already upload in scene, an image should not be required ## Current Behavior <!--- Tell us what happens instead of the expected behavior --> I can't publish a story while not all scenes has an image file. ## Environment <!--- Please tell us about your environment --> <!-- Browser: all | Chrome XX | Firefox XX | IE XX | Safari XX | Mobile Chrome XX | Android X.X Web Browser | iOS XX Safari | iOS XX UIWebView | iOS XX WKWebView --> https://vm2161.kaj.pouta.csc.fi/ ## Possible Solution <!--- Not obligatory, but suggest a fix/reason for the bug, --> <!--- Provide a detailed description of the change or addition you are proposing --> - one image OR one video should be required for a scene ## Related issues <!--- Not obligatory, but reference related issues --> #199
non_test
image is being requested for all scenes i think this is a error expected behavior if a video have been already upload in scene an image should not be required current behavior i can t publish a story while not all scenes has an image file environment possible solution one image or one video should be required for a scene related issues
0
57,022
13,965,321,159
IssuesEvent
2020-10-25 21:56:01
programming-the-iot/book-exercise-tasks
https://api.github.com/repos/programming-the-iot/book-exercise-tasks
opened
PIOT-CFG-08-000: Install and configure Californium Tools for your platform
build
**Description** - Install and configure Californium CoAP tools to support initial testing of your CoAP server on your platform. **Actions** - Review the README for the Californium CoAP tools source code and utilities located at the [Californium Tools GitHub page](https://github.com/eclipse/californium.tools). Follow the instructions to install the test client for publishing and subscribing from the command line. - NOTE: The graphical client tool depends on the JavaFX UI toolkit, which may not be available in your JRE / JDK installation for your system. If you're interested in using JavaFX, you can read more at [OpenJFX](https://openjfx.io/). This is out of scope for this exercise; as such, these instructions will focus exclusively on command-line usage. - Using the command line, execute the following to download and build the Californium tools: - NOTE: This may take a few to many minutes, depending on a variety of factors, which include your Internet connection speed, system capability, and of course server responsiveness. ``` git clone https://github.com/eclipse/californium.tools.git cd californium.tools mvn clean install ``` - Once completed, you should have a number of directories created, including `run`, `cf-client`, and `cf-server`. You can use `cf-client` - and more specifically the `cf-client/target/cf-client-2.5.0-SNAPSHOT.jar` to test your CoAP server once you've completed the exercises in this Lab Module. - You can run the client using the `--help` argument to see what parameters it accepts: ``` cd cf-client/target/ java -jar cf-client-2.5.0-SNAPSHOT.jar --help ``` **Estimate** - This process should take < 1 hour **Tests** - Small **Tests** - After downloading and building the Californium Tools, you can test them by executing the following: - Start the server ``` ``` - Start the client and run a few tests to verify all is working correctly ``` ```
1.0
PIOT-CFG-08-000: Install and configure Californium Tools for your platform - **Description** - Install and configure Californium CoAP tools to support initial testing of your CoAP server on your platform. **Actions** - Review the README for the Californium CoAP tools source code and utilities located at the [Californium Tools GitHub page](https://github.com/eclipse/californium.tools). Follow the instructions to install the test client for publishing and subscribing from the command line. - NOTE: The graphical client tool depends on the JavaFX UI toolkit, which may not be available in your JRE / JDK installation for your system. If you're interested in using JavaFX, you can read more at [OpenJFX](https://openjfx.io/). This is out of scope for this exercise; as such, these instructions will focus exclusively on command-line usage. - Using the command line, execute the following to download and build the Californium tools: - NOTE: This may take a few to many minutes, depending on a variety of factors, which include your Internet connection speed, system capability, and of course server responsiveness. ``` git clone https://github.com/eclipse/californium.tools.git cd californium.tools mvn clean install ``` - Once completed, you should have a number of directories created, including `run`, `cf-client`, and `cf-server`. You can use `cf-client` - and more specifically the `cf-client/target/cf-client-2.5.0-SNAPSHOT.jar` to test your CoAP server once you've completed the exercises in this Lab Module. - You can run the client using the `--help` argument to see what parameters it accepts: ``` cd cf-client/target/ java -jar cf-client-2.5.0-SNAPSHOT.jar --help ``` **Estimate** - This process should take < 1 hour **Tests** - Small **Tests** - After downloading and building the Californium Tools, you can test them by executing the following: - Start the server ``` ``` - Start the client and run a few tests to verify all is working correctly ``` ```
non_test
piot cfg install and configure californium tools for your platform description install and configure californium coap tools to support initial testing of your coap server on your platform actions review the readme for the californium coap tools source code and utilities located at the follow the instructions to install the test client for publishing and subscribing from the command line note the graphical client tool depends on the javafx ui toolkit which may not be available in your jre jdk installation for your system if you re interested in using javafx you can read more at this is out of scope for this exercise as such these instructions will focus exclusively on command line usage using the command line execute the following to download and build the californium tools note this may take a few to many minutes depending on a variety of factors which include your internet connection speed system capability and of course server responsiveness git clone cd californium tools mvn clean install once completed you should have a number of directories created including run cf client and cf server you can use cf client and more specifically the cf client target cf client snapshot jar to test your coap server once you ve completed the exercises in this lab module you can run the client using the help argument to see what parameters it accepts cd cf client target java jar cf client snapshot jar help estimate this process should take hour tests small tests after downloading and building the californium tools you can test them by executing the following start the server start the client and run a few tests to verify all is working correctly
0
78,679
3,512,872,687
IssuesEvent
2016-01-11 06:13:12
Apollo-Community/ApolloStation
https://api.github.com/repos/Apollo-Community/ApolloStation
closed
Beepsky can hit anyone in line of sight
easy to fix its a feature! oversight priority: medium
Doesn't even have to be standing near them, just as long as he can see them. Probably missing a range check.
1.0
Beepsky can hit anyone in line of sight - Doesn't even have to be standing near them, just as long as he can see them. Probably missing a range check.
non_test
beepsky can hit anyone in line of sight doesn t even have to be standing near them just as long as he can see them probably missing a range check
0
11,745
5,078,739,839
IssuesEvent
2016-12-28 16:37:19
opendatakit/build
https://api.github.com/repos/opendatakit/build
opened
clean up terminology in bottom bar
Build Priority-Low Type-Enhancement
<a href="https://github.com/mitchellsundt"><img src="https://avatars.githubusercontent.com/u/2281171?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mitchellsundt](https://github.com/mitchellsundt)** _Thursday Jul 09, 2015 at 18:30 GMT_ _Originally opened as https://github.com/opendatakit/opendatakit/issues/440 (0 comment(s))_ ---- Originally reported on Google Code with ID 439 ``` these are pretty nitpicky, but a recent training showed some stumbling blocks. i don't feel strongly about any of these. either make it "choose one" and "choose multiple" or "select one" and "select multiple". might also be clearer to say "radio button" or "checkboxes" numeric feels awkward. maybe number? media is vague. but understood that 'image/video/audio' is awkward. maybe split them into different options? ``` Reported by `yanokwa` on 2011-11-21 00:44:29
1.0
clean up terminology in bottom bar - <a href="https://github.com/mitchellsundt"><img src="https://avatars.githubusercontent.com/u/2281171?v=3" align="left" width="96" height="96" hspace="10"></img></a> **Issue by [mitchellsundt](https://github.com/mitchellsundt)** _Thursday Jul 09, 2015 at 18:30 GMT_ _Originally opened as https://github.com/opendatakit/opendatakit/issues/440 (0 comment(s))_ ---- Originally reported on Google Code with ID 439 ``` these are pretty nitpicky, but a recent training showed some stumbling blocks. i don't feel strongly about any of these. either make it "choose one" and "choose multiple" or "select one" and "select multiple". might also be clearer to say "radio button" or "checkboxes" numeric feels awkward. maybe number? media is vague. but understood that 'image/video/audio' is awkward. maybe split them into different options? ``` Reported by `yanokwa` on 2011-11-21 00:44:29
non_test
clean up terminology in bottom bar issue by thursday jul at gmt originally opened as comment s originally reported on google code with id these are pretty nitpicky but a recent training showed some stumbling blocks i don t feel strongly about any of these either make it choose one and choose multiple or select one and select multiple might also be clearer to say radio button or checkboxes numeric feels awkward maybe number media is vague but understood that image video audio is awkward maybe split them into different options reported by yanokwa on
0
259,808
22,552,953,488
IssuesEvent
2022-06-27 07:43:47
junu0516/GithubIssueManager
https://api.github.com/repos/junu0516/GithubIssueManager
closed
์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ๊ตฌํ˜„, ๋‹จ์œ„ํ…Œ์ŠคํŠธ
feature test
- [x] ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ํ™”๋ฉด ๊ตฌํ˜„ - [x] ์ด์Šˆ ์ถ”๊ฐ€ ํ™”๋ฉด ์›ํ˜• - [x] ์ถ”๊ฐ€์ •๋ณด ์„ ํƒ ํ™”๋ฉด - [x] ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ์š”์ฒญ/์‘๋‹ต ํ™•์ธ - [x] ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ๋‹จ์œ„ํ…Œ์ŠคํŠธ - [x] #20
1.0
์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ๊ตฌํ˜„, ๋‹จ์œ„ํ…Œ์ŠคํŠธ - - [x] ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ํ™”๋ฉด ๊ตฌํ˜„ - [x] ์ด์Šˆ ์ถ”๊ฐ€ ํ™”๋ฉด ์›ํ˜• - [x] ์ถ”๊ฐ€์ •๋ณด ์„ ํƒ ํ™”๋ฉด - [x] ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ์š”์ฒญ/์‘๋‹ต ํ™•์ธ - [x] ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ๋‹จ์œ„ํ…Œ์ŠคํŠธ - [x] #20
test
์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ๊ตฌํ˜„ ๋‹จ์œ„ํ…Œ์ŠคํŠธ ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ํ™”๋ฉด ๊ตฌํ˜„ ์ด์Šˆ ์ถ”๊ฐ€ ํ™”๋ฉด ์›ํ˜• ์ถ”๊ฐ€์ •๋ณด ์„ ํƒ ํ™”๋ฉด ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ์š”์ฒญ ์‘๋‹ต ํ™•์ธ ์ด์Šˆ ์ถ”๊ฐ€ํ•˜๊ธฐ ๋‹จ์œ„ํ…Œ์ŠคํŠธ
1
208,912
7,162,936,048
IssuesEvent
2018-01-29 03:57:59
openshift/origin
https://api.github.com/repos/openshift/origin
closed
Confusing(and broken?) deployment behavior
component/apps kind/bug priority/P1
https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/18278/test_pull_request_origin_extended_image_ecosystem/450/ specifically the `[image_ecosystem][ruby][Slow] hot deploy for openshift ruby image Rails example should work with hot deploy [Suite:openshift] 10m29s` test In this test we create a buildconfig and DC that is triggered by that build. The DC also has a config change trigger. We expect exactly 1 deployment to occur (when the build completes), but we appear to be getting two. This did not used to happen. Looking at the events, we see: the build starts at 23:09:25 ``` Jan 24 23:19:28.534: INFO: At 2018-01-24 23:09:25 +0000 UTC - event for rails-postgresql-example-1-build: {default-scheduler } Scheduled: Successfully assigned rails-postgresql-example-1-build to ip-172-18-11-23.ec2.internal Jan 24 23:19:28.534: INFO: At 2018-01-24 23:09:36 +0000 UTC - event for rails-postgresql-example-1: {build-controller } BuildStarted: Build extended-test-s2i-ruby-llfdg-mr96q/rails-postgresql-example-1 is now running ``` Roughly simultaneously, the build completes and a deployment 2 is created (not sure where the event for deployment 1 is??) which cancels deployment 1. Presumably deployment 2 is being triggered by the newly built/pushed image. ``` Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:47 +0000 UTC - event for rails-postgresql-example: {deploymentconfig-controller } DeploymentAwaitingCancellation: Deployment of version 2 awaiting cancellation of older running deployments Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:47 +0000 UTC - event for rails-postgresql-example: {deploymentconfig-controller } DeploymentCancelled: Cancelled deployment "rails-postgresql-example-1" superceded by version 2 Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:47 +0000 UTC - event for rails-postgresql-example-1-deploy: {default-scheduler } Scheduled: Successfully assigned rails-postgresql-example-1-deploy to ip-172-18-11-23.ec2.internal Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:48 +0000 UTC - event for rails-postgresql-example-1: {build-controller } BuildCompleted: Build extended-test-s2i-ruby-llfdg-mr96q/rails-postgresql-example-1 completed successfully Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:53 +0000 UTC - event for rails-postgresql-example: {deployer-controller } RolloutCancelled: Rollout for "extended-test-s2i-ruby-llfdg-mr96q/rails-postgresql-example-1" cancelled Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:53 +0000 UTC - event for rails-postgresql-example: {deploymentconfig-controller } DeploymentCreated: Created new replication controller "rails-postgresql-example-2" for version 2 ``` So my question is: Assuming deployment 2 was caused by the build completion, what caused deployment 1? Was it the config change trigger (despite the imagestreamtag not existing yet?) Should deployment 1 have been created? It didn't used to work that way. you can see the template w/ the DC in question here: https://github.com/openshift/rails-ex/blob/6a59aa15bf863fde71e0bbfa43c5344290eed8f6/openshift/templates/rails-postgresql.json#L146 And note that this test just ran, so it should have included the informer stale cache fix. My impression is that deployment 1 got created by the configchangetrigger, and then hung waiting for the imagestreamtag to resolve. When the build updates the imagestreamtag, deployment 1 starts to proceed, but gets canceled because deployment 2 is triggered. @tnozicka @mfojtik @kargakis (marking as bug because it appears to be a change in behavior. if it's expected we can live with it, but i need to know it's intentional)
1.0
Confusing(and broken?) deployment behavior - https://openshift-gce-devel.appspot.com/build/origin-ci-test/pr-logs/pull/18278/test_pull_request_origin_extended_image_ecosystem/450/ specifically the `[image_ecosystem][ruby][Slow] hot deploy for openshift ruby image Rails example should work with hot deploy [Suite:openshift] 10m29s` test In this test we create a buildconfig and DC that is triggered by that build. The DC also has a config change trigger. We expect exactly 1 deployment to occur (when the build completes), but we appear to be getting two. This did not used to happen. Looking at the events, we see: the build starts at 23:09:25 ``` Jan 24 23:19:28.534: INFO: At 2018-01-24 23:09:25 +0000 UTC - event for rails-postgresql-example-1-build: {default-scheduler } Scheduled: Successfully assigned rails-postgresql-example-1-build to ip-172-18-11-23.ec2.internal Jan 24 23:19:28.534: INFO: At 2018-01-24 23:09:36 +0000 UTC - event for rails-postgresql-example-1: {build-controller } BuildStarted: Build extended-test-s2i-ruby-llfdg-mr96q/rails-postgresql-example-1 is now running ``` Roughly simultaneously, the build completes and a deployment 2 is created (not sure where the event for deployment 1 is??) which cancels deployment 1. Presumably deployment 2 is being triggered by the newly built/pushed image. ``` Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:47 +0000 UTC - event for rails-postgresql-example: {deploymentconfig-controller } DeploymentAwaitingCancellation: Deployment of version 2 awaiting cancellation of older running deployments Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:47 +0000 UTC - event for rails-postgresql-example: {deploymentconfig-controller } DeploymentCancelled: Cancelled deployment "rails-postgresql-example-1" superceded by version 2 Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:47 +0000 UTC - event for rails-postgresql-example-1-deploy: {default-scheduler } Scheduled: Successfully assigned rails-postgresql-example-1-deploy to ip-172-18-11-23.ec2.internal Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:48 +0000 UTC - event for rails-postgresql-example-1: {build-controller } BuildCompleted: Build extended-test-s2i-ruby-llfdg-mr96q/rails-postgresql-example-1 completed successfully Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:53 +0000 UTC - event for rails-postgresql-example: {deployer-controller } RolloutCancelled: Rollout for "extended-test-s2i-ruby-llfdg-mr96q/rails-postgresql-example-1" cancelled Jan 24 23:19:28.534: INFO: At 2018-01-24 23:14:53 +0000 UTC - event for rails-postgresql-example: {deploymentconfig-controller } DeploymentCreated: Created new replication controller "rails-postgresql-example-2" for version 2 ``` So my question is: Assuming deployment 2 was caused by the build completion, what caused deployment 1? Was it the config change trigger (despite the imagestreamtag not existing yet?) Should deployment 1 have been created? It didn't used to work that way. you can see the template w/ the DC in question here: https://github.com/openshift/rails-ex/blob/6a59aa15bf863fde71e0bbfa43c5344290eed8f6/openshift/templates/rails-postgresql.json#L146 And note that this test just ran, so it should have included the informer stale cache fix. My impression is that deployment 1 got created by the configchangetrigger, and then hung waiting for the imagestreamtag to resolve. When the build updates the imagestreamtag, deployment 1 starts to proceed, but gets canceled because deployment 2 is triggered. @tnozicka @mfojtik @kargakis (marking as bug because it appears to be a change in behavior. if it's expected we can live with it, but i need to know it's intentional)
non_test
confusing and broken deployment behavior specifically the hot deploy for openshift ruby image rails example should work with hot deploy test in this test we create a buildconfig and dc that is triggered by that build the dc also has a config change trigger we expect exactly deployment to occur when the build completes but we appear to be getting two this did not used to happen looking at the events we see the build starts at jan info at utc event for rails postgresql example build default scheduler scheduled successfully assigned rails postgresql example build to ip internal jan info at utc event for rails postgresql example build controller buildstarted build extended test ruby llfdg rails postgresql example is now running roughly simultaneously the build completes and a deployment is created not sure where the event for deployment is which cancels deployment presumably deployment is being triggered by the newly built pushed image jan info at utc event for rails postgresql example deploymentconfig controller deploymentawaitingcancellation deployment of version awaiting cancellation of older running deployments jan info at utc event for rails postgresql example deploymentconfig controller deploymentcancelled cancelled deployment rails postgresql example superceded by version jan info at utc event for rails postgresql example deploy default scheduler scheduled successfully assigned rails postgresql example deploy to ip internal jan info at utc event for rails postgresql example build controller buildcompleted build extended test ruby llfdg rails postgresql example completed successfully jan info at utc event for rails postgresql example deployer controller rolloutcancelled rollout for extended test ruby llfdg rails postgresql example cancelled jan info at utc event for rails postgresql example deploymentconfig controller deploymentcreated created new replication controller rails postgresql example for version so my question is assuming deployment was caused by the build completion what caused deployment was it the config change trigger despite the imagestreamtag not existing yet should deployment have been created it didn t used to work that way you can see the template w the dc in question here and note that this test just ran so it should have included the informer stale cache fix my impression is that deployment got created by the configchangetrigger and then hung waiting for the imagestreamtag to resolve when the build updates the imagestreamtag deployment starts to proceed but gets canceled because deployment is triggered tnozicka mfojtik kargakis marking as bug because it appears to be a change in behavior if it s expected we can live with it but i need to know it s intentional
0
56,397
8,071,266,535
IssuesEvent
2018-08-06 12:40:05
draft-js-plugins/draft-js-plugins
https://api.github.com/repos/draft-js-plugins/draft-js-plugins
closed
Static toolbar example is missing stylesheets in documentation
documentation
I was following along on the static-toolbar documentation as seen here: https://www.draft-js-plugins.com/plugin/static-toolbar And I noticed the Themed Static Toolbar Example does not show two of the three stylesheets that it uses. Both buttonStyles and toolbarStyles are missing from the landing page content. It's easy to imply what the styling might be, but it seems both should be embedded like the way editorStyles.css is.
1.0
Static toolbar example is missing stylesheets in documentation - I was following along on the static-toolbar documentation as seen here: https://www.draft-js-plugins.com/plugin/static-toolbar And I noticed the Themed Static Toolbar Example does not show two of the three stylesheets that it uses. Both buttonStyles and toolbarStyles are missing from the landing page content. It's easy to imply what the styling might be, but it seems both should be embedded like the way editorStyles.css is.
non_test
static toolbar example is missing stylesheets in documentation i was following along on the static toolbar documentation as seen here and i noticed the themed static toolbar example does not show two of the three stylesheets that it uses both buttonstyles and toolbarstyles are missing from the landing page content it s easy to imply what the styling might be but it seems both should be embedded like the way editorstyles css is
0
327,181
28,046,120,427
IssuesEvent
2023-03-28 23:08:26
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
reopened
Fix miscellaneous.test_numpy_nan_to_num
NumPy Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4507827096/jobs/7935970679" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4547391872/jobs/8017216171" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4507827096/jobs/7935970679" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4507827096/jobs/7935970679" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_mathematical_functions/test_miscellaneous.py::test_numpy_nan_to_num[cpu-ivy.functional.backends.torch-False-False]</summary> 2023-03-28T20:54:25.9675031Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T20:54:25.9675461Z E Falsifying example: test_numpy_nan_to_num( 2023-03-28T20:54:25.9675952Z E dtype_and_x=(['float16'], [array(0., dtype=float16)]), 2023-03-28T20:54:25.9676296Z E posinf=None, 2023-03-28T20:54:25.9677051Z E neginf=None, 2023-03-28T20:54:25.9677363Z E nan=0.0, 2023-03-28T20:54:25.9677655Z E copy=False, 2023-03-28T20:54:25.9678127Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T20:54:25.9678724Z E num_positional_args=0, 2023-03-28T20:54:25.9679061Z E with_out=False, 2023-03-28T20:54:25.9679372Z E inplace=False, 2023-03-28T20:54:25.9679702Z E as_variable=[False], 2023-03-28T20:54:25.9680154Z E native_arrays=[False], 2023-03-28T20:54:25.9680510Z E generate_frontend_arrays=False, 2023-03-28T20:54:25.9680818Z E ), 2023-03-28T20:54:25.9681394Z E fn_tree='ivy.functional.frontends.numpy.nan_to_num', 2023-03-28T20:54:25.9681822Z E frontend='numpy', 2023-03-28T20:54:25.9682150Z E on_device='cpu', 2023-03-28T20:54:25.9682436Z E ) 2023-03-28T20:54:25.9682672Z E 2023-03-28T20:54:25.9683309Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2DACwAAHwAB') as a decorator on your test case </details>
1.0
Fix miscellaneous.test_numpy_nan_to_num - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4507827096/jobs/7935970679" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4547391872/jobs/8017216171" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4507827096/jobs/7935970679" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4507827096/jobs/7935970679" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-success-success></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_numpy/test_mathematical_functions/test_miscellaneous.py::test_numpy_nan_to_num[cpu-ivy.functional.backends.torch-False-False]</summary> 2023-03-28T20:54:25.9675031Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T20:54:25.9675461Z E Falsifying example: test_numpy_nan_to_num( 2023-03-28T20:54:25.9675952Z E dtype_and_x=(['float16'], [array(0., dtype=float16)]), 2023-03-28T20:54:25.9676296Z E posinf=None, 2023-03-28T20:54:25.9677051Z E neginf=None, 2023-03-28T20:54:25.9677363Z E nan=0.0, 2023-03-28T20:54:25.9677655Z E copy=False, 2023-03-28T20:54:25.9678127Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T20:54:25.9678724Z E num_positional_args=0, 2023-03-28T20:54:25.9679061Z E with_out=False, 2023-03-28T20:54:25.9679372Z E inplace=False, 2023-03-28T20:54:25.9679702Z E as_variable=[False], 2023-03-28T20:54:25.9680154Z E native_arrays=[False], 2023-03-28T20:54:25.9680510Z E generate_frontend_arrays=False, 2023-03-28T20:54:25.9680818Z E ), 2023-03-28T20:54:25.9681394Z E fn_tree='ivy.functional.frontends.numpy.nan_to_num', 2023-03-28T20:54:25.9681822Z E frontend='numpy', 2023-03-28T20:54:25.9682150Z E on_device='cpu', 2023-03-28T20:54:25.9682436Z E ) 2023-03-28T20:54:25.9682672Z E 2023-03-28T20:54:25.9683309Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2DACwAAHwAB') as a decorator on your test case </details>
test
fix miscellaneous test numpy nan to num tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test numpy test mathematical functions test miscellaneous py test numpy nan to num e typeerror to native got an unexpected keyword argument dtype e falsifying example test numpy nan to num e dtype and x e posinf none e neginf none e nan e copy false e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends numpy nan to num e frontend numpy e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
464,004
13,304,976,918
IssuesEvent
2020-08-25 17:47:32
cds-snc/report-a-cybercrime
https://api.github.com/repos/cds-snc/report-a-cybercrime
closed
Update Devices Form
high priority suggestion
Update DevicesForm.js to use Formik, Yup, and Bootstrap as documented in issue #2136. These changes have been made to HowDidItStartForm.js, you can them here - #2138. A new component will need to be created for the input element (if not already created). Documentation can be found [here](https://react-bootstrap.github.io/components/forms/#forms-controls).
1.0
Update Devices Form - Update DevicesForm.js to use Formik, Yup, and Bootstrap as documented in issue #2136. These changes have been made to HowDidItStartForm.js, you can them here - #2138. A new component will need to be created for the input element (if not already created). Documentation can be found [here](https://react-bootstrap.github.io/components/forms/#forms-controls).
non_test
update devices form update devicesform js to use formik yup and bootstrap as documented in issue these changes have been made to howdiditstartform js you can them here a new component will need to be created for the input element if not already created documentation can be found
0
208,719
15,930,706,075
IssuesEvent
2021-04-14 01:32:58
bean-pod/switchboard
https://api.github.com/repos/bean-pod/switchboard
closed
Acceptance test for #48 As a user, when an event occurs during a stream, I want to see an event message in the admin panel
acceptance test
Steps: - Start the frontend and backend - Start the sample sender and sample receiver - Register the sample sender and sample receiver in the system - Open the frontend - Navigate to the stream creation page - Create a stream between your sample sender and sample receiver - Close the sample sender before the stream is concluded - Go to the stream details page - In the stream logs component, a log should appear with events that happened during the stream
1.0
Acceptance test for #48 As a user, when an event occurs during a stream, I want to see an event message in the admin panel - Steps: - Start the frontend and backend - Start the sample sender and sample receiver - Register the sample sender and sample receiver in the system - Open the frontend - Navigate to the stream creation page - Create a stream between your sample sender and sample receiver - Close the sample sender before the stream is concluded - Go to the stream details page - In the stream logs component, a log should appear with events that happened during the stream
test
acceptance test for as a user when an event occurs during a stream i want to see an event message in the admin panel steps start the frontend and backend start the sample sender and sample receiver register the sample sender and sample receiver in the system open the frontend navigate to the stream creation page create a stream between your sample sender and sample receiver close the sample sender before the stream is concluded go to the stream details page in the stream logs component a log should appear with events that happened during the stream
1
329,249
24,211,964,478
IssuesEvent
2022-09-26 00:38:58
robert-altom/test
https://api.github.com/repos/robert-altom/test
closed
Fix link for download AUT Alpha package
documentation revert 1.7.0
Link for downloading the AUT Alpha package ( https://altom.com/apphttps://gitlab.com/altom/altunity/altunitytester/uploads/altUnityProAlpha/AltUnityTester.unitypackage) should be fixed. ![image](https://gitlab.com/altom/altunity/altunitytester/uploads/54565637da1ebc9155f4fce7c5479cde/image.png) --- <sub>You can find the original issue from GitLab [here](https://gitlab.com/altom/altunity/altunitytester/-/issues/697).</sub>
1.0
Fix link for download AUT Alpha package - Link for downloading the AUT Alpha package ( https://altom.com/apphttps://gitlab.com/altom/altunity/altunitytester/uploads/altUnityProAlpha/AltUnityTester.unitypackage) should be fixed. ![image](https://gitlab.com/altom/altunity/altunitytester/uploads/54565637da1ebc9155f4fce7c5479cde/image.png) --- <sub>You can find the original issue from GitLab [here](https://gitlab.com/altom/altunity/altunitytester/-/issues/697).</sub>
non_test
fix link for download aut alpha package link for downloading the aut alpha package should be fixed you can find the original issue from gitlab
0
127,900
17,377,012,071
IssuesEvent
2021-07-31 00:01:43
envoyproxy/envoy
https://api.github.com/repos/envoyproxy/envoy
closed
Seeking proposals for how to implement STREAMING body mode in ext_proc
area/ext_proc design proposal stale
*Title*: Help figure out how to implement STREAMING body mode for the ext_proc filter *Description*: The ext_proc protocol describes four ways to handle request and response bodies: 1) Ignore them. This is the default and works fine. 2) BUFFERED mode -- this is already implemented. The filter buffers the body in memory, returns an error if the buffer fills up before end of stream, sends the buffer to the external processor, and waits for a response which will optionally modify the buffer. 3) STREAMING mode -- not implemented yet. This is intended to support the use case of an external processor that wants to modify the body one chunk at a time, or otherwise handle a body of arbitrary size. 4) BUFFERED_PARTIAL -- not implemented yet. This is intended for processors that want to inspect only the start of the body, or perhaps cache bodies smaller than a certain maximum size. In order for STREAMING mode to be implemented, the filter would need to do something like this, for both the encoding and decoding paths: 1) en|decodeData is called with a chunk of data. 2) Save a pointer to the buffer that we just received. 3) Send a message on the gRPC stream that contains the new body chunk. 4) When the processor sends back a response, apply any modifications to the body chunk saved in step 2. 5) Continue processing, which will presumably loop back to step 1. This doesn't work consistently because there is no way for step 1 (en|decodeData) to respond with a status that tells the filter manager, "do not continue iteration and do not send me any more chunks." As a result, in some cases, we perform steps 1, 2, and 3, and then while waiting for the response, we end up back on step 1 even though we never got to step 4. As a result, we don't know how to apply modifications to the body in a way that will allow us to properly process them in order. In some cases, a combination of StopIterationAndWatermark, combined with calling "*AboveWriteBufferHighWatermark" seems to do the trick. But sometimes it doesn't. The result is that after en|decodeData returns, but before it can call "continue," the next chunk is delivered. Since this doesn't work, I can think of a few more possibilities: 1) The filter could try to keep track of all the chunks that it received, assuming that a StopIterationAndWatermark status means that although more chunks will be received, none will be forwarded to other filters. Then, apply responses to individual chunks as they come back. But the problem here is that when we finally call "continue," won't all the chunks be sent on down the filter chain? 2) The filter could try to somehow be notified when the watermark really causes reading of writing is paused. It could buffer everything to that point and not clear the watermark until it has all been processed. I don't see how to do that in both directions, however. 3) The filter could keep track of multiple buffers, but then clear each buffer so that when we "continue" it would just continue with empty buffers down the filter chain. Then when we actually know what data should be forwarded, we could call "injectEn|DecodedDataToFilterChain." This might work and I'll have to try it. 4) We could give up and make streaming mode read-only. We can probably adjust things so that the filter will still have the ability to send back an error in many cases after reading the body, but otherwise the body would really be read-only. This would mean that there are cases in which ext_proc is not useful. If you read all this, then thanks -- as you can see I could use some advice on how to make this whole thing work!
1.0
Seeking proposals for how to implement STREAMING body mode in ext_proc - *Title*: Help figure out how to implement STREAMING body mode for the ext_proc filter *Description*: The ext_proc protocol describes four ways to handle request and response bodies: 1) Ignore them. This is the default and works fine. 2) BUFFERED mode -- this is already implemented. The filter buffers the body in memory, returns an error if the buffer fills up before end of stream, sends the buffer to the external processor, and waits for a response which will optionally modify the buffer. 3) STREAMING mode -- not implemented yet. This is intended to support the use case of an external processor that wants to modify the body one chunk at a time, or otherwise handle a body of arbitrary size. 4) BUFFERED_PARTIAL -- not implemented yet. This is intended for processors that want to inspect only the start of the body, or perhaps cache bodies smaller than a certain maximum size. In order for STREAMING mode to be implemented, the filter would need to do something like this, for both the encoding and decoding paths: 1) en|decodeData is called with a chunk of data. 2) Save a pointer to the buffer that we just received. 3) Send a message on the gRPC stream that contains the new body chunk. 4) When the processor sends back a response, apply any modifications to the body chunk saved in step 2. 5) Continue processing, which will presumably loop back to step 1. This doesn't work consistently because there is no way for step 1 (en|decodeData) to respond with a status that tells the filter manager, "do not continue iteration and do not send me any more chunks." As a result, in some cases, we perform steps 1, 2, and 3, and then while waiting for the response, we end up back on step 1 even though we never got to step 4. As a result, we don't know how to apply modifications to the body in a way that will allow us to properly process them in order. In some cases, a combination of StopIterationAndWatermark, combined with calling "*AboveWriteBufferHighWatermark" seems to do the trick. But sometimes it doesn't. The result is that after en|decodeData returns, but before it can call "continue," the next chunk is delivered. Since this doesn't work, I can think of a few more possibilities: 1) The filter could try to keep track of all the chunks that it received, assuming that a StopIterationAndWatermark status means that although more chunks will be received, none will be forwarded to other filters. Then, apply responses to individual chunks as they come back. But the problem here is that when we finally call "continue," won't all the chunks be sent on down the filter chain? 2) The filter could try to somehow be notified when the watermark really causes reading of writing is paused. It could buffer everything to that point and not clear the watermark until it has all been processed. I don't see how to do that in both directions, however. 3) The filter could keep track of multiple buffers, but then clear each buffer so that when we "continue" it would just continue with empty buffers down the filter chain. Then when we actually know what data should be forwarded, we could call "injectEn|DecodedDataToFilterChain." This might work and I'll have to try it. 4) We could give up and make streaming mode read-only. We can probably adjust things so that the filter will still have the ability to send back an error in many cases after reading the body, but otherwise the body would really be read-only. This would mean that there are cases in which ext_proc is not useful. If you read all this, then thanks -- as you can see I could use some advice on how to make this whole thing work!
non_test
seeking proposals for how to implement streaming body mode in ext proc title help figure out how to implement streaming body mode for the ext proc filter description the ext proc protocol describes four ways to handle request and response bodies ignore them this is the default and works fine buffered mode this is already implemented the filter buffers the body in memory returns an error if the buffer fills up before end of stream sends the buffer to the external processor and waits for a response which will optionally modify the buffer streaming mode not implemented yet this is intended to support the use case of an external processor that wants to modify the body one chunk at a time or otherwise handle a body of arbitrary size buffered partial not implemented yet this is intended for processors that want to inspect only the start of the body or perhaps cache bodies smaller than a certain maximum size in order for streaming mode to be implemented the filter would need to do something like this for both the encoding and decoding paths en decodedata is called with a chunk of data save a pointer to the buffer that we just received send a message on the grpc stream that contains the new body chunk when the processor sends back a response apply any modifications to the body chunk saved in step continue processing which will presumably loop back to step this doesn t work consistently because there is no way for step en decodedata to respond with a status that tells the filter manager do not continue iteration and do not send me any more chunks as a result in some cases we perform steps and and then while waiting for the response we end up back on step even though we never got to step as a result we don t know how to apply modifications to the body in a way that will allow us to properly process them in order in some cases a combination of stopiterationandwatermark combined with calling abovewritebufferhighwatermark seems to do the trick but sometimes it doesn t the result is that after en decodedata returns but before it can call continue the next chunk is delivered since this doesn t work i can think of a few more possibilities the filter could try to keep track of all the chunks that it received assuming that a stopiterationandwatermark status means that although more chunks will be received none will be forwarded to other filters then apply responses to individual chunks as they come back but the problem here is that when we finally call continue won t all the chunks be sent on down the filter chain the filter could try to somehow be notified when the watermark really causes reading of writing is paused it could buffer everything to that point and not clear the watermark until it has all been processed i don t see how to do that in both directions however the filter could keep track of multiple buffers but then clear each buffer so that when we continue it would just continue with empty buffers down the filter chain then when we actually know what data should be forwarded we could call injecten decodeddatatofilterchain this might work and i ll have to try it we could give up and make streaming mode read only we can probably adjust things so that the filter will still have the ability to send back an error in many cases after reading the body but otherwise the body would really be read only this would mean that there are cases in which ext proc is not useful if you read all this then thanks as you can see i could use some advice on how to make this whole thing work
0
115,068
4,651,466,521
IssuesEvent
2016-10-03 10:13:36
dhis2/maintenance-app
https://api.github.com/repos/dhis2/maintenance-app
closed
[PushAnalysis] Add button to manually execute report
enhancement priority:medium
Push analysis reports can be triggered independently of the schedule by sending a request to the API: `POST /api/pushAnalysis/<uid>/run` When editing an existing push analysis report in the maintenance app, there should be an option to trigger the report by clicking a button. See http://dhis2.github.io/dhis2-docs/master/en/developer/html/ch01s37.html
1.0
[PushAnalysis] Add button to manually execute report - Push analysis reports can be triggered independently of the schedule by sending a request to the API: `POST /api/pushAnalysis/<uid>/run` When editing an existing push analysis report in the maintenance app, there should be an option to trigger the report by clicking a button. See http://dhis2.github.io/dhis2-docs/master/en/developer/html/ch01s37.html
non_test
add button to manually execute report push analysis reports can be triggered independently of the schedule by sending a request to the api post api pushanalysis run when editing an existing push analysis report in the maintenance app there should be an option to trigger the report by clicking a button see
0
44,016
11,357,230,093
IssuesEvent
2020-01-25 02:59:26
DynamoRIO/dynamorio
https://api.github.com/repos/DynamoRIO/dynamorio
opened
Build race with API headers on Appveyor
Component-Build Hotlist-Travis OpSys-Windows
Recently I've seen this a few times on Appveyor: https://ci.appveyor.com/project/DynamoRIO/dynamorio/builds/30344395 ``` 13686====> FAILURE in debug-internal-64 <==== 13687debug-internal-64: **** 1 build errors **** 13688 C:\projects\dynamorio\build\build_debug-internal-64\include\dr_api.h(47) : fatal error C1083: Cannot open include file: 'dr_defines.h': No such file or directory 13689final package: **** 85 build errors **** 13690 ntdll_imports_x64.def : error LNK2001: unresolved external symbol __isascii 13691 ntdll_imports_x64.def : error LNK2001: unresolved external symbol __iscsym 13692 ntdll_imports_x64.def : error LNK2001: unresolved external symbol __iscsymf 13693 ntdll_imports_x64.def : error LNK2001: unresolved external symbol __toascii 13694 ntdll_imports_x64.def : error LNK2001: unresolved external symbol _atoi64 ``` A re-build makes it go away. I think it's a race introduced recently -- and my theory is drdisas is missing a dependence. I will fix it.
1.0
Build race with API headers on Appveyor - Recently I've seen this a few times on Appveyor: https://ci.appveyor.com/project/DynamoRIO/dynamorio/builds/30344395 ``` 13686====> FAILURE in debug-internal-64 <==== 13687debug-internal-64: **** 1 build errors **** 13688 C:\projects\dynamorio\build\build_debug-internal-64\include\dr_api.h(47) : fatal error C1083: Cannot open include file: 'dr_defines.h': No such file or directory 13689final package: **** 85 build errors **** 13690 ntdll_imports_x64.def : error LNK2001: unresolved external symbol __isascii 13691 ntdll_imports_x64.def : error LNK2001: unresolved external symbol __iscsym 13692 ntdll_imports_x64.def : error LNK2001: unresolved external symbol __iscsymf 13693 ntdll_imports_x64.def : error LNK2001: unresolved external symbol __toascii 13694 ntdll_imports_x64.def : error LNK2001: unresolved external symbol _atoi64 ``` A re-build makes it go away. I think it's a race introduced recently -- and my theory is drdisas is missing a dependence. I will fix it.
non_test
build race with api headers on appveyor recently i ve seen this a few times on appveyor failure in debug internal internal build errors c projects dynamorio build build debug internal include dr api h fatal error cannot open include file dr defines h no such file or directory package build errors ntdll imports def error unresolved external symbol isascii ntdll imports def error unresolved external symbol iscsym ntdll imports def error unresolved external symbol iscsymf ntdll imports def error unresolved external symbol toascii ntdll imports def error unresolved external symbol a re build makes it go away i think it s a race introduced recently and my theory is drdisas is missing a dependence i will fix it
0
99,640
30,516,644,918
IssuesEvent
2023-07-19 03:59:36
dotnet/runtime
https://api.github.com/repos/dotnet/runtime
closed
Half.TryParse assertion failure on System.Globalization.Ordinal.EqualsIgnoreCaseUtf8_Scalar
area-System.Text.Json blocking-clean-ci untriaged in-pr Known Build Error
## Build Information Build: https://dev.azure.com/dnceng-public/cbb18261-c48f-4abb-8651-8cdcb5474649/_build/results?buildId=344043 Build error leg or test failing: System.Text.Json.Tests.WorkItemExecution Pull request: https://github.com/dotnet/runtime/pull/89036 ### Error Blob ```json { "ErrorMessage": "System.Globalization.Ordinal.EqualsIgnoreCaseUtf8_Scalar", "BuildRetry": false, "ExcludeConsoleLog": false } ``` ### Reproduction Steps While adding support for Half to System.Text.Json (https://github.com/dotnet/runtime/pull/88962), I noticed one test triggers the assertion failure consistently on macOS-12.4. Note: This didn't repro for me locally on macOS 13.4.1 arm64. The test attempts to deserialize floting-point literals (NaN, infinity, -Infinity) in multiple casings/variants that are invalid for S.T.Json. Error happens on line 883. https://github.com/dotnet/runtime/blob/43fe8a8fc7642a1268e80b0433aebb92b049ee09/src/libraries/System.Text.Json/tests/Common/NumberHandlingTests.cs#L848-L883 Helix log: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-88962-merge-9757e201f6374e20b5/System.Text.Json.Tests/1/console.78feb75a.log?helixlogtype=result Call stack: ``` /private/tmp/helix/working/9EC208C7/w/A9F108F7/e /private/tmp/helix/working/9EC208C7/w/A9F108F7/e Discovering: System.Text.Json.Tests (method display = ClassAndMethod, method display options = None) Discovered: System.Text.Json.Tests (found 7160 of 7231 test cases) Starting: System.Text.Json.Tests (parallel test collections = on, max threads = 6) Process terminated. Assertion failed. at System.Globalization.Ordinal.EqualsIgnoreCaseUtf8_Scalar(Byte& charA, Int32 lengthA, Byte& charB, Int32 lengthB) in /_/src/libraries/System.Private.CoreLib/src/System/Globalization/Ordinal.Utf8.cs:line 308 at System.Number.TryParseFloat[TChar,TFloat](ReadOnlySpan`1 value, NumberStyles styles, NumberFormatInfo info, TFloat& result) in /_/src/libraries/System.Private.CoreLib/src/System/Number.Parsing.cs:line 1229 at System.Half.TryParse(ReadOnlySpan`1 utf8Text, NumberStyles style, IFormatProvider provider, Half& result) in /_/src/libraries/System.Private.CoreLib/src/System/Half.cs:line 2238 at System.Text.Json.Serialization.Converters.HalfConverter.TryParse(ReadOnlySpan`1 buffer, Half& result) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Converters/Value/HalfConverter.cs:line 197 at System.Text.Json.Serialization.Converters.HalfConverter.ReadCore(Utf8JsonReader& reader) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Converters/Value/HalfConverter.cs:line 50 at System.Text.Json.Serialization.Converters.HalfConverter.ReadNumberWithCustomHandling(Utf8JsonReader& reader, JsonNumberHandling handling, JsonSerializerOptions options) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Converters/Value/HalfConverter.cs:line 115 at System.Text.Json.Serialization.JsonConverter`1.TryRead(Utf8JsonReader& reader, Type typeToConvert, JsonSerializerOptions options, ReadStack& state, T& value, Boolean& isPopulatedValue) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonConverterOfT.cs:line 193 at System.Text.Json.Serialization.Metadata.JsonPropertyInfo`1.ReadJsonAndSetMember(Object obj, ReadStack& state, Utf8JsonReader& reader) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Metadata/JsonPropertyInfoOfT.cs:line 308 at System.Text.Json.Serialization.Converters.ObjectDefaultConverter`1.OnTryRead(Utf8JsonReader& reader, Type typeToConvert, JsonSerializerOptions options, ReadStack& state, T& value) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Converters/Object/ObjectDefaultConverter.cs:line 49 at System.Text.Json.Serialization.JsonConverter`1.TryRead(Utf8JsonReader& reader, Type typeToConvert, JsonSerializerOptions options, ReadStack& state, T& value, Boolean& isPopulatedValue) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonConverterOfT.cs:line 258 at System.Text.Json.Serialization.JsonConverter`1.ReadCore(Utf8JsonReader& reader, JsonSerializerOptions options, ReadStack& state) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonConverterOfT.ReadCore.cs:line 51 at System.Text.Json.Serialization.Metadata.JsonTypeInfo`1.Deserialize(Utf8JsonReader& reader, ReadStack& state) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Metadata/JsonTypeInfoOfT.ReadHelper.cs:line 22 at System.Text.Json.JsonSerializer.ReadFromSpan[TValue](ReadOnlySpan`1 utf8Json, JsonTypeInfo`1 jsonTypeInfo, Nullable`1 actualByteCount) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonSerializer.Read.Span.cs:line 160 at System.Text.Json.JsonSerializer.ReadFromSpan[TValue](ReadOnlySpan`1 json, JsonTypeInfo`1 jsonTypeInfo) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonSerializer.Read.String.cs:line 443 at System.Text.Json.JsonSerializer.Deserialize[TValue](String json, JsonSerializerOptions options) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonSerializer.Read.String.cs:line 55 at System.Text.Json.Serialization.Tests.JsonSerializerWrapper.StringSerializerWrapper.DeserializeWrapper[T](String json, JsonSerializerOptions options) in /_/src/libraries/System.Text.Json/tests/System.Text.Json.Tests/Serialization/JsonSerializerWrapper.Reflection.cs:line 129 at System.Text.Json.Serialization.Tests.NumberHandlingTests.<>c__DisplayClass39_0.<<FloatingPointConstants_Fail>b__1>d.MoveNext() in /_/src/libraries/System.Text.Json/tests/Common/NumberHandlingTests.cs:line 883 at System.Runtime.CompilerServices.AsyncMethodBuilderCore.Start[TStateMachine](TStateMachine& stateMachine) in /_/src/libraries/System.Private.CoreLib/src/System/Runtime/CompilerServices/AsyncMethodBuilderCore.cs:line 38 at System.Text.Json.Serialization.Tests.NumberHandlingTests.<>c__DisplayClass39_0.<FloatingPointConstants_Fail>b__1() ``` <!-- Known issue validation start --> ### Known issue validation **Build: :mag_right:** https://dev.azure.com/dnceng-public/public/_build/results?buildId=344043 **Error message validated:** `System.Globalization.Ordinal.EqualsIgnoreCaseUtf8_Scalar` **Result validation: :white_check_mark:** Known issue matched with the provided build. **Validation performed at:** 7/18/2023 11:30:55 PM UTC <!-- Known issue validation end --> <!--Known issue error report start --> ### Report |Build|Definition|Test|Pull Request| |---|---|---|---| |[344444](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344444)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344444&view=ms.vss-test-web.build-test-results-tab&runId=7199754&resultId=191955)|dotnet/runtime#89153| |[344421](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344421)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344421&view=ms.vss-test-web.build-test-results-tab&runId=7199328&resultId=191955)|dotnet/runtime#89154| |[344427](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344427)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344427&view=ms.vss-test-web.build-test-results-tab&runId=7199326&resultId=191955)|dotnet/runtime#89155| |[344398](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344398)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344398&view=ms.vss-test-web.build-test-results-tab&runId=7199020&resultId=191955)|dotnet/runtime#89151| |[344340](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344340)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344340&view=ms.vss-test-web.build-test-results-tab&runId=7197162&resultId=191950)|dotnet/runtime#87438| |[344330](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344330)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344330&view=ms.vss-test-web.build-test-results-tab&runId=7196836&resultId=191950)|dotnet/runtime#89061| |[344321](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344321)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344321&view=ms.vss-test-web.build-test-results-tab&runId=7196454&resultId=191950)|dotnet/runtime#89147| |[344317](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344317)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344317&view=ms.vss-test-web.build-test-results-tab&runId=7196442&resultId=191945)|dotnet/runtime#89123| |[344290](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344290)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344290&view=ms.vss-test-web.build-test-results-tab&runId=7195808&resultId=191945)|dotnet/runtime#87847| |[344296](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344296)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344296&view=ms.vss-test-web.build-test-results-tab&runId=7195748&resultId=191954)|dotnet/runtime#89102| |[344284](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344284)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344284&view=ms.vss-test-web.build-test-results-tab&runId=7195206&resultId=191950)|dotnet/runtime#89144| |[344268](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344268)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344268&view=ms.vss-test-web.build-test-results-tab&runId=7195142&resultId=191936)|dotnet/runtime#89130| |[344261](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344261)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344261&view=ms.vss-test-web.build-test-results-tab&runId=7194786&resultId=191936)|dotnet/runtime#85328| |[344239](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344239)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344239&view=ms.vss-test-web.build-test-results-tab&runId=7194322&resultId=191936)|dotnet/runtime#86089| |[344207](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344207)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344207&view=ms.vss-test-web.build-test-results-tab&runId=7193170&resultId=191941)|dotnet/runtime#89129| |[344182](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344182)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344182&view=ms.vss-test-web.build-test-results-tab&runId=7193228&resultId=191936)|dotnet/runtime#89135| |[344136](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344136)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344136&view=ms.vss-test-web.build-test-results-tab&runId=7191764&resultId=191941)|dotnet/runtime#87656| |[344099](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344099)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344099&view=ms.vss-test-web.build-test-results-tab&runId=7190988&resultId=191936)|dotnet/runtime#89064| |[344092](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344092)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344092&view=ms.vss-test-web.build-test-results-tab&runId=7190894&resultId=191936)|dotnet/runtime#89116| |[344056](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344056)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344056&view=ms.vss-test-web.build-test-results-tab&runId=7189838&resultId=191923)|dotnet/runtime#89123| |[344043](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344043)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344043&view=ms.vss-test-web.build-test-results-tab&runId=7189660&resultId=191923)|dotnet/runtime#89036| |[344049](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344049)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344049&view=ms.vss-test-web.build-test-results-tab&runId=7187996&resultId=188823)|| |[343994](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343994)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343994&view=ms.vss-test-web.build-test-results-tab&runId=7188360&resultId=191923)|dotnet/runtime#87438| |[343982](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343982)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343982&view=ms.vss-test-web.build-test-results-tab&runId=7188200&resultId=191923)|dotnet/runtime#85328| |[343063](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343063)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343063&view=ms.vss-test-web.build-test-results-tab&runId=7166966&resultId=188762)|dotnet/runtime#88962| |[342911](https://dev.azure.com/dnceng-public/public/_build/results?buildId=342911)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=342911&view=ms.vss-test-web.build-test-results-tab&runId=7164930&resultId=188756)|dotnet/runtime#88962| #### Summary |24-Hour Hit Count|7-Day Hit Count|1-Month Count| |---|---|---| |25|26|26| <!--Known issue error report end -->
1.0
Half.TryParse assertion failure on System.Globalization.Ordinal.EqualsIgnoreCaseUtf8_Scalar - ## Build Information Build: https://dev.azure.com/dnceng-public/cbb18261-c48f-4abb-8651-8cdcb5474649/_build/results?buildId=344043 Build error leg or test failing: System.Text.Json.Tests.WorkItemExecution Pull request: https://github.com/dotnet/runtime/pull/89036 ### Error Blob ```json { "ErrorMessage": "System.Globalization.Ordinal.EqualsIgnoreCaseUtf8_Scalar", "BuildRetry": false, "ExcludeConsoleLog": false } ``` ### Reproduction Steps While adding support for Half to System.Text.Json (https://github.com/dotnet/runtime/pull/88962), I noticed one test triggers the assertion failure consistently on macOS-12.4. Note: This didn't repro for me locally on macOS 13.4.1 arm64. The test attempts to deserialize floting-point literals (NaN, infinity, -Infinity) in multiple casings/variants that are invalid for S.T.Json. Error happens on line 883. https://github.com/dotnet/runtime/blob/43fe8a8fc7642a1268e80b0433aebb92b049ee09/src/libraries/System.Text.Json/tests/Common/NumberHandlingTests.cs#L848-L883 Helix log: https://helixre107v0xdeko0k025g8.blob.core.windows.net/dotnet-runtime-refs-pull-88962-merge-9757e201f6374e20b5/System.Text.Json.Tests/1/console.78feb75a.log?helixlogtype=result Call stack: ``` /private/tmp/helix/working/9EC208C7/w/A9F108F7/e /private/tmp/helix/working/9EC208C7/w/A9F108F7/e Discovering: System.Text.Json.Tests (method display = ClassAndMethod, method display options = None) Discovered: System.Text.Json.Tests (found 7160 of 7231 test cases) Starting: System.Text.Json.Tests (parallel test collections = on, max threads = 6) Process terminated. Assertion failed. at System.Globalization.Ordinal.EqualsIgnoreCaseUtf8_Scalar(Byte& charA, Int32 lengthA, Byte& charB, Int32 lengthB) in /_/src/libraries/System.Private.CoreLib/src/System/Globalization/Ordinal.Utf8.cs:line 308 at System.Number.TryParseFloat[TChar,TFloat](ReadOnlySpan`1 value, NumberStyles styles, NumberFormatInfo info, TFloat& result) in /_/src/libraries/System.Private.CoreLib/src/System/Number.Parsing.cs:line 1229 at System.Half.TryParse(ReadOnlySpan`1 utf8Text, NumberStyles style, IFormatProvider provider, Half& result) in /_/src/libraries/System.Private.CoreLib/src/System/Half.cs:line 2238 at System.Text.Json.Serialization.Converters.HalfConverter.TryParse(ReadOnlySpan`1 buffer, Half& result) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Converters/Value/HalfConverter.cs:line 197 at System.Text.Json.Serialization.Converters.HalfConverter.ReadCore(Utf8JsonReader& reader) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Converters/Value/HalfConverter.cs:line 50 at System.Text.Json.Serialization.Converters.HalfConverter.ReadNumberWithCustomHandling(Utf8JsonReader& reader, JsonNumberHandling handling, JsonSerializerOptions options) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Converters/Value/HalfConverter.cs:line 115 at System.Text.Json.Serialization.JsonConverter`1.TryRead(Utf8JsonReader& reader, Type typeToConvert, JsonSerializerOptions options, ReadStack& state, T& value, Boolean& isPopulatedValue) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonConverterOfT.cs:line 193 at System.Text.Json.Serialization.Metadata.JsonPropertyInfo`1.ReadJsonAndSetMember(Object obj, ReadStack& state, Utf8JsonReader& reader) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Metadata/JsonPropertyInfoOfT.cs:line 308 at System.Text.Json.Serialization.Converters.ObjectDefaultConverter`1.OnTryRead(Utf8JsonReader& reader, Type typeToConvert, JsonSerializerOptions options, ReadStack& state, T& value) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Converters/Object/ObjectDefaultConverter.cs:line 49 at System.Text.Json.Serialization.JsonConverter`1.TryRead(Utf8JsonReader& reader, Type typeToConvert, JsonSerializerOptions options, ReadStack& state, T& value, Boolean& isPopulatedValue) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonConverterOfT.cs:line 258 at System.Text.Json.Serialization.JsonConverter`1.ReadCore(Utf8JsonReader& reader, JsonSerializerOptions options, ReadStack& state) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonConverterOfT.ReadCore.cs:line 51 at System.Text.Json.Serialization.Metadata.JsonTypeInfo`1.Deserialize(Utf8JsonReader& reader, ReadStack& state) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/Metadata/JsonTypeInfoOfT.ReadHelper.cs:line 22 at System.Text.Json.JsonSerializer.ReadFromSpan[TValue](ReadOnlySpan`1 utf8Json, JsonTypeInfo`1 jsonTypeInfo, Nullable`1 actualByteCount) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonSerializer.Read.Span.cs:line 160 at System.Text.Json.JsonSerializer.ReadFromSpan[TValue](ReadOnlySpan`1 json, JsonTypeInfo`1 jsonTypeInfo) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonSerializer.Read.String.cs:line 443 at System.Text.Json.JsonSerializer.Deserialize[TValue](String json, JsonSerializerOptions options) in /_/src/libraries/System.Text.Json/src/System/Text/Json/Serialization/JsonSerializer.Read.String.cs:line 55 at System.Text.Json.Serialization.Tests.JsonSerializerWrapper.StringSerializerWrapper.DeserializeWrapper[T](String json, JsonSerializerOptions options) in /_/src/libraries/System.Text.Json/tests/System.Text.Json.Tests/Serialization/JsonSerializerWrapper.Reflection.cs:line 129 at System.Text.Json.Serialization.Tests.NumberHandlingTests.<>c__DisplayClass39_0.<<FloatingPointConstants_Fail>b__1>d.MoveNext() in /_/src/libraries/System.Text.Json/tests/Common/NumberHandlingTests.cs:line 883 at System.Runtime.CompilerServices.AsyncMethodBuilderCore.Start[TStateMachine](TStateMachine& stateMachine) in /_/src/libraries/System.Private.CoreLib/src/System/Runtime/CompilerServices/AsyncMethodBuilderCore.cs:line 38 at System.Text.Json.Serialization.Tests.NumberHandlingTests.<>c__DisplayClass39_0.<FloatingPointConstants_Fail>b__1() ``` <!-- Known issue validation start --> ### Known issue validation **Build: :mag_right:** https://dev.azure.com/dnceng-public/public/_build/results?buildId=344043 **Error message validated:** `System.Globalization.Ordinal.EqualsIgnoreCaseUtf8_Scalar` **Result validation: :white_check_mark:** Known issue matched with the provided build. **Validation performed at:** 7/18/2023 11:30:55 PM UTC <!-- Known issue validation end --> <!--Known issue error report start --> ### Report |Build|Definition|Test|Pull Request| |---|---|---|---| |[344444](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344444)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344444&view=ms.vss-test-web.build-test-results-tab&runId=7199754&resultId=191955)|dotnet/runtime#89153| |[344421](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344421)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344421&view=ms.vss-test-web.build-test-results-tab&runId=7199328&resultId=191955)|dotnet/runtime#89154| |[344427](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344427)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344427&view=ms.vss-test-web.build-test-results-tab&runId=7199326&resultId=191955)|dotnet/runtime#89155| |[344398](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344398)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344398&view=ms.vss-test-web.build-test-results-tab&runId=7199020&resultId=191955)|dotnet/runtime#89151| |[344340](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344340)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344340&view=ms.vss-test-web.build-test-results-tab&runId=7197162&resultId=191950)|dotnet/runtime#87438| |[344330](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344330)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344330&view=ms.vss-test-web.build-test-results-tab&runId=7196836&resultId=191950)|dotnet/runtime#89061| |[344321](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344321)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344321&view=ms.vss-test-web.build-test-results-tab&runId=7196454&resultId=191950)|dotnet/runtime#89147| |[344317](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344317)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344317&view=ms.vss-test-web.build-test-results-tab&runId=7196442&resultId=191945)|dotnet/runtime#89123| |[344290](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344290)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344290&view=ms.vss-test-web.build-test-results-tab&runId=7195808&resultId=191945)|dotnet/runtime#87847| |[344296](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344296)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344296&view=ms.vss-test-web.build-test-results-tab&runId=7195748&resultId=191954)|dotnet/runtime#89102| |[344284](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344284)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344284&view=ms.vss-test-web.build-test-results-tab&runId=7195206&resultId=191950)|dotnet/runtime#89144| |[344268](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344268)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344268&view=ms.vss-test-web.build-test-results-tab&runId=7195142&resultId=191936)|dotnet/runtime#89130| |[344261](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344261)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344261&view=ms.vss-test-web.build-test-results-tab&runId=7194786&resultId=191936)|dotnet/runtime#85328| |[344239](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344239)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344239&view=ms.vss-test-web.build-test-results-tab&runId=7194322&resultId=191936)|dotnet/runtime#86089| |[344207](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344207)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344207&view=ms.vss-test-web.build-test-results-tab&runId=7193170&resultId=191941)|dotnet/runtime#89129| |[344182](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344182)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344182&view=ms.vss-test-web.build-test-results-tab&runId=7193228&resultId=191936)|dotnet/runtime#89135| |[344136](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344136)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344136&view=ms.vss-test-web.build-test-results-tab&runId=7191764&resultId=191941)|dotnet/runtime#87656| |[344099](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344099)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344099&view=ms.vss-test-web.build-test-results-tab&runId=7190988&resultId=191936)|dotnet/runtime#89064| |[344092](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344092)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344092&view=ms.vss-test-web.build-test-results-tab&runId=7190894&resultId=191936)|dotnet/runtime#89116| |[344056](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344056)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344056&view=ms.vss-test-web.build-test-results-tab&runId=7189838&resultId=191923)|dotnet/runtime#89123| |[344043](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344043)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344043&view=ms.vss-test-web.build-test-results-tab&runId=7189660&resultId=191923)|dotnet/runtime#89036| |[344049](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344049)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=344049&view=ms.vss-test-web.build-test-results-tab&runId=7187996&resultId=188823)|| |[343994](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343994)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343994&view=ms.vss-test-web.build-test-results-tab&runId=7188360&resultId=191923)|dotnet/runtime#87438| |[343982](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343982)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343982&view=ms.vss-test-web.build-test-results-tab&runId=7188200&resultId=191923)|dotnet/runtime#85328| |[343063](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343063)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=343063&view=ms.vss-test-web.build-test-results-tab&runId=7166966&resultId=188762)|dotnet/runtime#88962| |[342911](https://dev.azure.com/dnceng-public/public/_build/results?buildId=342911)|dotnet/runtime|[System.Text.Json.Tests.WorkItemExecution](https://dev.azure.com/dnceng-public/public/_build/results?buildId=342911&view=ms.vss-test-web.build-test-results-tab&runId=7164930&resultId=188756)|dotnet/runtime#88962| #### Summary |24-Hour Hit Count|7-Day Hit Count|1-Month Count| |---|---|---| |25|26|26| <!--Known issue error report end -->
non_test
half tryparse assertion failure on system globalization ordinal scalar build information build build error leg or test failing system text json tests workitemexecution pull request error blob json errormessage system globalization ordinal scalar buildretry false excludeconsolelog false reproduction steps while adding support for half to system text json i noticed one test triggers the assertion failure consistently on macos note this didn t repro for me locally on macos the test attempts to deserialize floting point literals nan infinity infinity in multiple casings variants that are invalid for s t json error happens on line helix log call stack private tmp helix working w e private tmp helix working w e discovering system text json tests method display classandmethod method display options none discovered system text json tests found of test cases starting system text json tests parallel test collections on max threads process terminated assertion failed at system globalization ordinal scalar byte chara lengtha byte charb lengthb in src libraries system private corelib src system globalization ordinal cs line at system number tryparsefloat readonlyspan value numberstyles styles numberformatinfo info tfloat result in src libraries system private corelib src system number parsing cs line at system half tryparse readonlyspan numberstyles style iformatprovider provider half result in src libraries system private corelib src system half cs line at system text json serialization converters halfconverter tryparse readonlyspan buffer half result in src libraries system text json src system text json serialization converters value halfconverter cs line at system text json serialization converters halfconverter readcore reader in src libraries system text json src system text json serialization converters value halfconverter cs line at system text json serialization converters halfconverter readnumberwithcustomhandling reader jsonnumberhandling handling jsonserializeroptions options in src libraries system text json src system text json serialization converters value halfconverter cs line at system text json serialization jsonconverter tryread reader type typetoconvert jsonserializeroptions options readstack state t value boolean ispopulatedvalue in src libraries system text json src system text json serialization jsonconverteroft cs line at system text json serialization metadata jsonpropertyinfo readjsonandsetmember object obj readstack state reader in src libraries system text json src system text json serialization metadata jsonpropertyinfooft cs line at system text json serialization converters objectdefaultconverter ontryread reader type typetoconvert jsonserializeroptions options readstack state t value in src libraries system text json src system text json serialization converters object objectdefaultconverter cs line at system text json serialization jsonconverter tryread reader type typetoconvert jsonserializeroptions options readstack state t value boolean ispopulatedvalue in src libraries system text json src system text json serialization jsonconverteroft cs line at system text json serialization jsonconverter readcore reader jsonserializeroptions options readstack state in src libraries system text json src system text json serialization jsonconverteroft readcore cs line at system text json serialization metadata jsontypeinfo deserialize reader readstack state in src libraries system text json src system text json serialization metadata jsontypeinfooft readhelper cs line at system text json jsonserializer readfromspan readonlyspan jsontypeinfo jsontypeinfo nullable actualbytecount in src libraries system text json src system text json serialization jsonserializer read span cs line at system text json jsonserializer readfromspan readonlyspan json jsontypeinfo jsontypeinfo in src libraries system text json src system text json serialization jsonserializer read string cs line at system text json jsonserializer deserialize string json jsonserializeroptions options in src libraries system text json src system text json serialization jsonserializer read string cs line at system text json serialization tests jsonserializerwrapper stringserializerwrapper deserializewrapper string json jsonserializeroptions options in src libraries system text json tests system text json tests serialization jsonserializerwrapper reflection cs line at system text json serialization tests numberhandlingtests c b d movenext in src libraries system text json tests common numberhandlingtests cs line at system runtime compilerservices asyncmethodbuildercore start tstatemachine statemachine in src libraries system private corelib src system runtime compilerservices asyncmethodbuildercore cs line at system text json serialization tests numberhandlingtests c b known issue validation build mag right error message validated system globalization ordinal scalar result validation white check mark known issue matched with the provided build validation performed at pm utc report build definition test pull request summary hour hit count day hit count month count
0
186,517
14,395,698,953
IssuesEvent
2020-12-03 04:26:44
ThatGuySam/doesitarm
https://api.github.com/repos/ThatGuySam/doesitarm
closed
PPSSPP
Needs M1 Testing Needs Tracker Link New App Request
**The official name of the app** PPSSPP **Is there a supported build available on a public/release channel?** No **Proposed App Status** Yes, works via Rosetta 2 with native support in development **Proposed App Category** Entertainment and Media Apps **Related Issue Tracker Link or discussion** Issue: https://github.com/hrydgard/ppsspp/issues/13708 **An Official App Download Page** https://build.ppsspp.org/?page/downloads#osx **Full Screenshot with the App and Activity Monitor(if self reporting) or [Silicon Info](https://github.com/billycastelli/Silicon-Info)** ![Screen Shot 2020-12-02 at 3 53 15 PM](https://user-images.githubusercontent.com/3460639/100844341-b75dc480-34b6-11eb-8d69-ca240f8e6c8e.png) ![Screen Shot 2020-12-02 at 3 50 13 PM](https://user-images.githubusercontent.com/3460639/100844358-bcbb0f00-34b6-11eb-8b48-db80d0d6f56a.png)
1.0
PPSSPP - **The official name of the app** PPSSPP **Is there a supported build available on a public/release channel?** No **Proposed App Status** Yes, works via Rosetta 2 with native support in development **Proposed App Category** Entertainment and Media Apps **Related Issue Tracker Link or discussion** Issue: https://github.com/hrydgard/ppsspp/issues/13708 **An Official App Download Page** https://build.ppsspp.org/?page/downloads#osx **Full Screenshot with the App and Activity Monitor(if self reporting) or [Silicon Info](https://github.com/billycastelli/Silicon-Info)** ![Screen Shot 2020-12-02 at 3 53 15 PM](https://user-images.githubusercontent.com/3460639/100844341-b75dc480-34b6-11eb-8d69-ca240f8e6c8e.png) ![Screen Shot 2020-12-02 at 3 50 13 PM](https://user-images.githubusercontent.com/3460639/100844358-bcbb0f00-34b6-11eb-8b48-db80d0d6f56a.png)
test
ppsspp the official name of the app ppsspp is there a supported build available on a public release channel no proposed app status yes works via rosetta with native support in development proposed app category entertainment and media apps related issue tracker link or discussion issue an official app download page full screenshot with the app and activity monitor if self reporting or
1
239,019
19,804,970,691
IssuesEvent
2022-01-19 05:03:00
istio/istio
https://api.github.com/repos/istio/istio
closed
TCP Source IP
area/test and release area/networking lifecycle/stale
### Bug Description Istio version 1.9.5 We have a few database pods which needs to be in cluster and we do have a headless service for them. When we enable istio proxy on them these pods cannot form a cluster as they cannot really see the source ip and they see it coming from 127.0.0.1 Is there an alternative to use where istio will pass the source ip to the database container ? the protocol is TCP here. ### Version ```prose ./istioctl version client version: 1.9.5 control plane version: 1.9.5 data plane version: 1.9.5 (2 proxies) Client Version: version.Info{Major:"1", Minor:"21", GitVersion:"v1.21.0", GitCommit:"cb303e613a121a29364f75cc67d3d580833a7479", GitTreeState:"clean", BuildDate:"2021-04-08T16:31:21Z", GoVersion:"go1.16.1", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"21", GitVersion:"v1.21.5", GitCommit:"aea7bbadd2fc0cd689de94a54e5b7b758869d691", GitTreeState:"clean", BuildDate:"2021-09-15T21:04:16Z", GoVersion:"go1.16.8", Compiler:"gc", Platform:"linux/amd64"} ``` ### Additional Information _No response_
1.0
TCP Source IP - ### Bug Description Istio version 1.9.5 We have a few database pods which needs to be in cluster and we do have a headless service for them. When we enable istio proxy on them these pods cannot form a cluster as they cannot really see the source ip and they see it coming from 127.0.0.1 Is there an alternative to use where istio will pass the source ip to the database container ? the protocol is TCP here. ### Version ```prose ./istioctl version client version: 1.9.5 control plane version: 1.9.5 data plane version: 1.9.5 (2 proxies) Client Version: version.Info{Major:"1", Minor:"21", GitVersion:"v1.21.0", GitCommit:"cb303e613a121a29364f75cc67d3d580833a7479", GitTreeState:"clean", BuildDate:"2021-04-08T16:31:21Z", GoVersion:"go1.16.1", Compiler:"gc", Platform:"linux/amd64"} Server Version: version.Info{Major:"1", Minor:"21", GitVersion:"v1.21.5", GitCommit:"aea7bbadd2fc0cd689de94a54e5b7b758869d691", GitTreeState:"clean", BuildDate:"2021-09-15T21:04:16Z", GoVersion:"go1.16.8", Compiler:"gc", Platform:"linux/amd64"} ``` ### Additional Information _No response_
test
tcp source ip bug description istio version we have a few database pods which needs to be in cluster and we do have a headless service for them when we enable istio proxy on them these pods cannot form a cluster as they cannot really see the source ip and they see it coming from is there an alternative to use where istio will pass the source ip to the database container the protocol is tcp here version prose istioctl version client version control plane version data plane version proxies client version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux server version version info major minor gitversion gitcommit gittreestate clean builddate goversion compiler gc platform linux additional information no response
1
326,010
27,973,353,745
IssuesEvent
2023-03-25 09:21:43
jj248/RealmsInExile
https://api.github.com/repos/jj248/RealmsInExile
closed
Black Breath curable by Elves and Elrosionath
priority: medium scripting Ready for Testing
Now that Elves are around, Black Breath should be curable, though only by Elves or Elrosionath characters. We should move it from a modifier to an illness trait
1.0
Black Breath curable by Elves and Elrosionath - Now that Elves are around, Black Breath should be curable, though only by Elves or Elrosionath characters. We should move it from a modifier to an illness trait
test
black breath curable by elves and elrosionath now that elves are around black breath should be curable though only by elves or elrosionath characters we should move it from a modifier to an illness trait
1
79,146
22,623,797,366
IssuesEvent
2022-06-30 08:53:33
Shopify/react-native-skia
https://api.github.com/repos/Shopify/react-native-skia
closed
Build failing for React-native 0.69.0
build android
Hello, i am upgraded my react native and started to get this error, not sure whats the cause, thanks. ```` PS C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0> npm run android > appname@0.0.9 android > react-native run-android info Running jetifier to migrate libraries to AndroidX. You can disable it using "--no-jetifier" flag. Jetifier found 2265 file(s) to forward-jetify. Using 8 workers... info JS server already running. info Installing the app... > Configure project :lottie-react-native WARNING:Software Components will not be created automatically for Maven publishing from Android Gradle Plugin 8.0. To opt-in to the future behavior, set the Gradle property android.disableAutomaticComponentCreation=true in the `gradle.properties` file or use the new publishing DSL. > Configure project :react-native-reanimated Native libs debug enabled: true Android gradle plugin: 7.1.1 Gradle: 7.3.3 building Reanimated2 WARNING:Software Components will not be created automatically for Maven publishing from Android Gradle Plugin 8.0. To opt-in to the future behavior, set the Gradle property android.disableAutomaticComponentCreation=true in the `gradle.properties` file or use the new publishing DSL. > Configure project :shopify_react-native-skia react-native-skia: node_modules/ found at: C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules react-native-skia: React Native version: 69 react-native-skia: Is Source build: false react-native-skia: Prebuilt dir: C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\@shopify\react-native-skia\android\build/react-native-0*/jni Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0. You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins. See https://docs.gradle.org/7.3.3/userguide/command_line_interface.html#sec:command_line_warnings 6 actionable tasks: 6 up-to-date FAILURE: Build failed with an exception. * Where: Build file 'C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\@shopify\react-native-skia\android\build.gradle' line: 178 * What went wrong: A problem occurred evaluating project ':shopify_react-native-skia'. > Expected directory 'C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules/react-native/android' to contain exactly one file, however, it contains more than one file. * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Run with --scan to get full insights. * Get more help at https://help.gradle.org BUILD FAILED in 22s error Failed to install the app. Make sure you have the Android development environment set up: https://reactnative.dev/docs/environment-setup. Error: Command failed: gradlew.bat app:installDebug -PreactNativeDevServerPort=8081 FAILURE: Build failed with an exception. * Where: Build file 'C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\@shopify\react-native-skia\android\build.gradle' line: 178 * What went wrong: A problem occurred evaluating project ':shopify_react-native-skia'. > Expected directory 'C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules/react-native/android' to contain exactly one file, however, it contains more than one file. * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Run with --scan to get full insights. * Get more help at https://help.gradle.org BUILD FAILED in 22s at makeError (C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\execa\index.js:174:9) at C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\execa\index.js:278:16 at process.processTicksAndRejections (node:internal/process/task_queues:95:5) at async runOnAllDevices (C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\react-native\node_modules\@react-native-community\cli-platform-android\build\commands\runAndroid\runOnAllDevices.js:109:5) at async Command.handleAction (C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\@react-native-community\cli\build\index.js:192:9) info Run CLI with --verbose flag for more details. ````
1.0
Build failing for React-native 0.69.0 - Hello, i am upgraded my react native and started to get this error, not sure whats the cause, thanks. ```` PS C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0> npm run android > appname@0.0.9 android > react-native run-android info Running jetifier to migrate libraries to AndroidX. You can disable it using "--no-jetifier" flag. Jetifier found 2265 file(s) to forward-jetify. Using 8 workers... info JS server already running. info Installing the app... > Configure project :lottie-react-native WARNING:Software Components will not be created automatically for Maven publishing from Android Gradle Plugin 8.0. To opt-in to the future behavior, set the Gradle property android.disableAutomaticComponentCreation=true in the `gradle.properties` file or use the new publishing DSL. > Configure project :react-native-reanimated Native libs debug enabled: true Android gradle plugin: 7.1.1 Gradle: 7.3.3 building Reanimated2 WARNING:Software Components will not be created automatically for Maven publishing from Android Gradle Plugin 8.0. To opt-in to the future behavior, set the Gradle property android.disableAutomaticComponentCreation=true in the `gradle.properties` file or use the new publishing DSL. > Configure project :shopify_react-native-skia react-native-skia: node_modules/ found at: C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules react-native-skia: React Native version: 69 react-native-skia: Is Source build: false react-native-skia: Prebuilt dir: C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\@shopify\react-native-skia\android\build/react-native-0*/jni Deprecated Gradle features were used in this build, making it incompatible with Gradle 8.0. You can use '--warning-mode all' to show the individual deprecation warnings and determine if they come from your own scripts or plugins. See https://docs.gradle.org/7.3.3/userguide/command_line_interface.html#sec:command_line_warnings 6 actionable tasks: 6 up-to-date FAILURE: Build failed with an exception. * Where: Build file 'C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\@shopify\react-native-skia\android\build.gradle' line: 178 * What went wrong: A problem occurred evaluating project ':shopify_react-native-skia'. > Expected directory 'C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules/react-native/android' to contain exactly one file, however, it contains more than one file. * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Run with --scan to get full insights. * Get more help at https://help.gradle.org BUILD FAILED in 22s error Failed to install the app. Make sure you have the Android development environment set up: https://reactnative.dev/docs/environment-setup. Error: Command failed: gradlew.bat app:installDebug -PreactNativeDevServerPort=8081 FAILURE: Build failed with an exception. * Where: Build file 'C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\@shopify\react-native-skia\android\build.gradle' line: 178 * What went wrong: A problem occurred evaluating project ':shopify_react-native-skia'. > Expected directory 'C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules/react-native/android' to contain exactly one file, however, it contains more than one file. * Try: > Run with --stacktrace option to get the stack trace. > Run with --info or --debug option to get more log output. > Run with --scan to get full insights. * Get more help at https://help.gradle.org BUILD FAILED in 22s at makeError (C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\execa\index.js:174:9) at C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\execa\index.js:278:16 at process.processTicksAndRejections (node:internal/process/task_queues:95:5) at async runOnAllDevices (C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\react-native\node_modules\@react-native-community\cli-platform-android\build\commands\runAndroid\runOnAllDevices.js:109:5) at async Command.handleAction (C:\Users\Abdullah\Desktop\Programming\company\apps\AppName\AppName-0.69.0\node_modules\@react-native-community\cli\build\index.js:192:9) info Run CLI with --verbose flag for more details. ````
non_test
build failing for react native hello i am upgraded my react native and started to get this error not sure whats the cause thanks ps c users abdullah desktop programming company apps appname appname npm run android appname android react native run android info running jetifier to migrate libraries to androidx you can disable it using no jetifier flag jetifier found file s to forward jetify using workers info js server already running info installing the app configure project lottie react native warning software components will not be created automatically for maven publishing from android gradle plugin to opt in to the future behavior set the gradle property android disableautomaticcomponentcreation true in the gradle properties file or use the new publishing dsl configure project react native reanimated native libs debug enabled true android gradle plugin gradle building warning software components will not be created automatically for maven publishing from android gradle plugin to opt in to the future behavior set the gradle property android disableautomaticcomponentcreation true in the gradle properties file or use the new publishing dsl configure project shopify react native skia react native skia node modules found at c users abdullah desktop programming company apps appname appname node modules react native skia react native version react native skia is source build false react native skia prebuilt dir c users abdullah desktop programming company apps appname appname node modules shopify react native skia android build react native jni deprecated gradle features were used in this build making it incompatible with gradle you can use warning mode all to show the individual deprecation warnings and determine if they come from your own scripts or plugins see actionable tasks up to date failure build failed with an exception where build file c users abdullah desktop programming company apps appname appname node modules shopify react native skia android build gradle line what went wrong a problem occurred evaluating project shopify react native skia expected directory c users abdullah desktop programming company apps appname appname node modules react native android to contain exactly one file however it contains more than one file try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights get more help at build failed in error failed to install the app make sure you have the android development environment set up error command failed gradlew bat app installdebug preactnativedevserverport failure build failed with an exception where build file c users abdullah desktop programming company apps appname appname node modules shopify react native skia android build gradle line what went wrong a problem occurred evaluating project shopify react native skia expected directory c users abdullah desktop programming company apps appname appname node modules react native android to contain exactly one file however it contains more than one file try run with stacktrace option to get the stack trace run with info or debug option to get more log output run with scan to get full insights get more help at build failed in at makeerror c users abdullah desktop programming company apps appname appname node modules execa index js at c users abdullah desktop programming company apps appname appname node modules execa index js at process processticksandrejections node internal process task queues at async runonalldevices c users abdullah desktop programming company apps appname appname node modules react native node modules react native community cli platform android build commands runandroid runonalldevices js at async command handleaction c users abdullah desktop programming company apps appname appname node modules react native community cli build index js info run cli with verbose flag for more details
0
37,763
8,358,894,963
IssuesEvent
2018-10-03 05:53:24
kreghek/Zilon
https://api.github.com/repos/kreghek/Zilon
closed
ะ˜ัะฟั€ะฐะฒะธั‚ัŒ ะทะฐะผะตั‡ะฐะฝะธะต ะฐะฝะฐะปะธะทะฐั‚ะพั€ะฐ: SectorEntityManager.Remove: Rename parameter 'entities' to 'actors' to match the interface declaration.
Hacktoberfest code improvement good first issue
ะ’ะตั€ะพัั‚ะฝะพ ะดะพะปะถะฝะพ ะฑั‹ั‚ัŒ ะฝะฐะพะฑะพั€ะพั‚ - ะฒ ะธะฝั‚ะตั€ั„ะตะนัะต ะธัะฟั€ะฐะฒะปะตะฝะพ ะฝะฐะทะฒะฐะฝะธะต ะฟะฐั€ะฐะผะตั‚ั€ะฐ.
1.0
ะ˜ัะฟั€ะฐะฒะธั‚ัŒ ะทะฐะผะตั‡ะฐะฝะธะต ะฐะฝะฐะปะธะทะฐั‚ะพั€ะฐ: SectorEntityManager.Remove: Rename parameter 'entities' to 'actors' to match the interface declaration. - ะ’ะตั€ะพัั‚ะฝะพ ะดะพะปะถะฝะพ ะฑั‹ั‚ัŒ ะฝะฐะพะฑะพั€ะพั‚ - ะฒ ะธะฝั‚ะตั€ั„ะตะนัะต ะธัะฟั€ะฐะฒะปะตะฝะพ ะฝะฐะทะฒะฐะฝะธะต ะฟะฐั€ะฐะผะตั‚ั€ะฐ.
non_test
ะธัะฟั€ะฐะฒะธั‚ัŒ ะทะฐะผะตั‡ะฐะฝะธะต ะฐะฝะฐะปะธะทะฐั‚ะพั€ะฐ sectorentitymanager remove rename parameter entities to actors to match the interface declaration ะฒะตั€ะพัั‚ะฝะพ ะดะพะปะถะฝะพ ะฑั‹ั‚ัŒ ะฝะฐะพะฑะพั€ะพั‚ ะฒ ะธะฝั‚ะตั€ั„ะตะนัะต ะธัะฟั€ะฐะฒะปะตะฝะพ ะฝะฐะทะฒะฐะฝะธะต ะฟะฐั€ะฐะผะตั‚ั€ะฐ
0
288,961
8,853,524,673
IssuesEvent
2019-01-08 21:38:29
visit-dav/issues-test
https://api.github.com/repos/visit-dav/issues-test
closed
Uninstaller on window needs safegaurds
bug likelihood medium priority reviewed severity high
Potential for removing files not related to VisIt install (eg entire 'Program Files' directory), due to $Rm /r $(InstDir)Can occur on versions prior to 2.3.0 if user changed default install location.With 2.3.0, 'LLNL\VisIt <version>' is always appended to user-chosen install path, but we should problably have the uninstaller only delete those files it added in the first place, rather than do a recursive remove on a directory. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 814 Status: Resolved Project: VisIt Tracker: Bug Priority: Urgent Subject: Uninstaller on window needs safegaurds Assigned to: Kathleen Biagas Category: - Target version: 2.3.2 Author: Kathleen Biagas Start: 08/15/2011 Due date: % Done: 0% Estimated time: Created: 08/15/2011 08:49 pm Updated: 08/29/2011 11:20 am Likelihood: 3 - Occasional Severity: 5 - Very Serious Found in version: 2.2.1 Impact: Expected Use: OS: Windows Support Group: Any Description: Potential for removing files not related to VisIt install (eg entire 'Program Files' directory), due to $Rm /r $(InstDir)Can occur on versions prior to 2.3.0 if user changed default install location.With 2.3.0, 'LLNL\VisIt <version>' is always appended to user-chosen install path, but we should problably have the uninstaller only delete those files it added in the first place, rather than do a recursive remove on a directory. Comments: Modified uninstaller to parse the install log and delete only installed files and directories.If visit_install.log not available when uninstaller is run, user is prompted to verify recursivedelete of installation directory (message shows full path that will be removed).M windowsbuild/installation/binaryinstallation.nsi
1.0
Uninstaller on window needs safegaurds - Potential for removing files not related to VisIt install (eg entire 'Program Files' directory), due to $Rm /r $(InstDir)Can occur on versions prior to 2.3.0 if user changed default install location.With 2.3.0, 'LLNL\VisIt <version>' is always appended to user-chosen install path, but we should problably have the uninstaller only delete those files it added in the first place, rather than do a recursive remove on a directory. -----------------------REDMINE MIGRATION----------------------- This ticket was migrated from Redmine. As such, not all information was able to be captured in the transition. Below is a complete record of the original redmine ticket. Ticket number: 814 Status: Resolved Project: VisIt Tracker: Bug Priority: Urgent Subject: Uninstaller on window needs safegaurds Assigned to: Kathleen Biagas Category: - Target version: 2.3.2 Author: Kathleen Biagas Start: 08/15/2011 Due date: % Done: 0% Estimated time: Created: 08/15/2011 08:49 pm Updated: 08/29/2011 11:20 am Likelihood: 3 - Occasional Severity: 5 - Very Serious Found in version: 2.2.1 Impact: Expected Use: OS: Windows Support Group: Any Description: Potential for removing files not related to VisIt install (eg entire 'Program Files' directory), due to $Rm /r $(InstDir)Can occur on versions prior to 2.3.0 if user changed default install location.With 2.3.0, 'LLNL\VisIt <version>' is always appended to user-chosen install path, but we should problably have the uninstaller only delete those files it added in the first place, rather than do a recursive remove on a directory. Comments: Modified uninstaller to parse the install log and delete only installed files and directories.If visit_install.log not available when uninstaller is run, user is prompted to verify recursivedelete of installation directory (message shows full path that will be removed).M windowsbuild/installation/binaryinstallation.nsi
non_test
uninstaller on window needs safegaurds potential for removing files not related to visit install eg entire program files directory due to rm r instdir can occur on versions prior to if user changed default install location with llnl visit is always appended to user chosen install path but we should problably have the uninstaller only delete those files it added in the first place rather than do a recursive remove on a directory redmine migration this ticket was migrated from redmine as such not all information was able to be captured in the transition below is a complete record of the original redmine ticket ticket number status resolved project visit tracker bug priority urgent subject uninstaller on window needs safegaurds assigned to kathleen biagas category target version author kathleen biagas start due date done estimated time created pm updated am likelihood occasional severity very serious found in version impact expected use os windows support group any description potential for removing files not related to visit install eg entire program files directory due to rm r instdir can occur on versions prior to if user changed default install location with llnl visit is always appended to user chosen install path but we should problably have the uninstaller only delete those files it added in the first place rather than do a recursive remove on a directory comments modified uninstaller to parse the install log and delete only installed files and directories if visit install log not available when uninstaller is run user is prompted to verify recursivedelete of installation directory message shows full path that will be removed m windowsbuild installation binaryinstallation nsi
0
106,903
11,501,838,527
IssuesEvent
2020-02-12 17:55:36
EricLacey/BadWeather
https://api.github.com/repos/EricLacey/BadWeather
closed
Proposal - Design Comps
Deliverable documentation
Two screens outlining the overall design direction and layout are to be created based on 2 screens from the wireframe document. The design comps should be polished and appear as though they are screengrabs from the completed project. Please include them as full-size images at the end of the PDF. (it is okay if the pages for the comps are larger in dimensions than the other pages, and landscape oriented, if necessary).
1.0
Proposal - Design Comps - Two screens outlining the overall design direction and layout are to be created based on 2 screens from the wireframe document. The design comps should be polished and appear as though they are screengrabs from the completed project. Please include them as full-size images at the end of the PDF. (it is okay if the pages for the comps are larger in dimensions than the other pages, and landscape oriented, if necessary).
non_test
proposal design comps two screens outlining the overall design direction and layout are to be created based on screens from the wireframe document the design comps should be polished and appear as though they are screengrabs from the completed project please include them as full size images at the end of the pdf it is okay if the pages for the comps are larger in dimensions than the other pages and landscape oriented if necessary
0
328,762
28,134,239,113
IssuesEvent
2023-04-01 07:12:09
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
closed
Fix metrics.test_tensorflow_poisson
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_metrics.py::test_tensorflow_poisson[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T14:33:05.3604181Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3614332Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3622297Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3623141Z E Falsifying example: test_tensorflow_poisson( 2023-03-28T14:33:05.3623699Z E dtype_and_x=(['float32', 'float32'], 2023-03-28T14:33:05.3624463Z E [array([-1.], dtype=float32), array([-1.], dtype=float32)]), 2023-03-28T14:33:05.3625625Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T14:33:05.3626079Z E num_positional_args=0, 2023-03-28T14:33:05.3626684Z E with_out=False, 2023-03-28T14:33:05.3627077Z E inplace=False, 2023-03-28T14:33:05.3627671Z E as_variable=[False], 2023-03-28T14:33:05.3628079Z E native_arrays=[False], 2023-03-28T14:33:05.3628702Z E generate_frontend_arrays=False, 2023-03-28T14:33:05.3629103Z E ), 2023-03-28T14:33:05.3629919Z E fn_tree='ivy.functional.frontends.tensorflow.keras.metrics.poisson', 2023-03-28T14:33:05.3630519Z E frontend='tensorflow', 2023-03-28T14:33:05.3631162Z E on_device='cpu', 2023-03-28T14:33:05.3631544Z E ) 2023-03-28T14:33:05.3632082Z E 2023-03-28T14:33:05.3632999Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2BkAAMoBaaR2QwMAACDAAY=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_metrics.py::test_tensorflow_poisson[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T14:33:05.3604181Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3614332Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3622297Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3623141Z E Falsifying example: test_tensorflow_poisson( 2023-03-28T14:33:05.3623699Z E dtype_and_x=(['float32', 'float32'], 2023-03-28T14:33:05.3624463Z E [array([-1.], dtype=float32), array([-1.], dtype=float32)]), 2023-03-28T14:33:05.3625625Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T14:33:05.3626079Z E num_positional_args=0, 2023-03-28T14:33:05.3626684Z E with_out=False, 2023-03-28T14:33:05.3627077Z E inplace=False, 2023-03-28T14:33:05.3627671Z E as_variable=[False], 2023-03-28T14:33:05.3628079Z E native_arrays=[False], 2023-03-28T14:33:05.3628702Z E generate_frontend_arrays=False, 2023-03-28T14:33:05.3629103Z E ), 2023-03-28T14:33:05.3629919Z E fn_tree='ivy.functional.frontends.tensorflow.keras.metrics.poisson', 2023-03-28T14:33:05.3630519Z E frontend='tensorflow', 2023-03-28T14:33:05.3631162Z E on_device='cpu', 2023-03-28T14:33:05.3631544Z E ) 2023-03-28T14:33:05.3632082Z E 2023-03-28T14:33:05.3632999Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2BkAAMoBaaR2QwMAACDAAY=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_metrics.py::test_tensorflow_poisson[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T14:33:05.3604181Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3614332Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3622297Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3623141Z E Falsifying example: test_tensorflow_poisson( 2023-03-28T14:33:05.3623699Z E dtype_and_x=(['float32', 'float32'], 2023-03-28T14:33:05.3624463Z E [array([-1.], dtype=float32), array([-1.], dtype=float32)]), 2023-03-28T14:33:05.3625625Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T14:33:05.3626079Z E num_positional_args=0, 2023-03-28T14:33:05.3626684Z E with_out=False, 2023-03-28T14:33:05.3627077Z E inplace=False, 2023-03-28T14:33:05.3627671Z E as_variable=[False], 2023-03-28T14:33:05.3628079Z E native_arrays=[False], 2023-03-28T14:33:05.3628702Z E generate_frontend_arrays=False, 2023-03-28T14:33:05.3629103Z E ), 2023-03-28T14:33:05.3629919Z E fn_tree='ivy.functional.frontends.tensorflow.keras.metrics.poisson', 2023-03-28T14:33:05.3630519Z E frontend='tensorflow', 2023-03-28T14:33:05.3631162Z E on_device='cpu', 2023-03-28T14:33:05.3631544Z E ) 2023-03-28T14:33:05.3632082Z E 2023-03-28T14:33:05.3632999Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2BkAAMoBaaR2QwMAACDAAY=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_metrics.py::test_tensorflow_poisson[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T14:33:05.3604181Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3614332Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3622297Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3623141Z E Falsifying example: test_tensorflow_poisson( 2023-03-28T14:33:05.3623699Z E dtype_and_x=(['float32', 'float32'], 2023-03-28T14:33:05.3624463Z E [array([-1.], dtype=float32), array([-1.], dtype=float32)]), 2023-03-28T14:33:05.3625625Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T14:33:05.3626079Z E num_positional_args=0, 2023-03-28T14:33:05.3626684Z E with_out=False, 2023-03-28T14:33:05.3627077Z E inplace=False, 2023-03-28T14:33:05.3627671Z E as_variable=[False], 2023-03-28T14:33:05.3628079Z E native_arrays=[False], 2023-03-28T14:33:05.3628702Z E generate_frontend_arrays=False, 2023-03-28T14:33:05.3629103Z E ), 2023-03-28T14:33:05.3629919Z E fn_tree='ivy.functional.frontends.tensorflow.keras.metrics.poisson', 2023-03-28T14:33:05.3630519Z E frontend='tensorflow', 2023-03-28T14:33:05.3631162Z E on_device='cpu', 2023-03-28T14:33:05.3631544Z E ) 2023-03-28T14:33:05.3632082Z E 2023-03-28T14:33:05.3632999Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2BkAAMoBaaR2QwMAACDAAY=') as a decorator on your test case </details>
1.0
Fix metrics.test_tensorflow_poisson - | | | |---|---| |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |torch|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/4543693013/jobs/8008709037" rel="noopener noreferrer" target="_blank"><img src=https://img.shields.io/badge/-failure-red></a> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_metrics.py::test_tensorflow_poisson[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T14:33:05.3604181Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3614332Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3622297Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3623141Z E Falsifying example: test_tensorflow_poisson( 2023-03-28T14:33:05.3623699Z E dtype_and_x=(['float32', 'float32'], 2023-03-28T14:33:05.3624463Z E [array([-1.], dtype=float32), array([-1.], dtype=float32)]), 2023-03-28T14:33:05.3625625Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T14:33:05.3626079Z E num_positional_args=0, 2023-03-28T14:33:05.3626684Z E with_out=False, 2023-03-28T14:33:05.3627077Z E inplace=False, 2023-03-28T14:33:05.3627671Z E as_variable=[False], 2023-03-28T14:33:05.3628079Z E native_arrays=[False], 2023-03-28T14:33:05.3628702Z E generate_frontend_arrays=False, 2023-03-28T14:33:05.3629103Z E ), 2023-03-28T14:33:05.3629919Z E fn_tree='ivy.functional.frontends.tensorflow.keras.metrics.poisson', 2023-03-28T14:33:05.3630519Z E frontend='tensorflow', 2023-03-28T14:33:05.3631162Z E on_device='cpu', 2023-03-28T14:33:05.3631544Z E ) 2023-03-28T14:33:05.3632082Z E 2023-03-28T14:33:05.3632999Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2BkAAMoBaaR2QwMAACDAAY=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_metrics.py::test_tensorflow_poisson[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T14:33:05.3604181Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3614332Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3622297Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3623141Z E Falsifying example: test_tensorflow_poisson( 2023-03-28T14:33:05.3623699Z E dtype_and_x=(['float32', 'float32'], 2023-03-28T14:33:05.3624463Z E [array([-1.], dtype=float32), array([-1.], dtype=float32)]), 2023-03-28T14:33:05.3625625Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T14:33:05.3626079Z E num_positional_args=0, 2023-03-28T14:33:05.3626684Z E with_out=False, 2023-03-28T14:33:05.3627077Z E inplace=False, 2023-03-28T14:33:05.3627671Z E as_variable=[False], 2023-03-28T14:33:05.3628079Z E native_arrays=[False], 2023-03-28T14:33:05.3628702Z E generate_frontend_arrays=False, 2023-03-28T14:33:05.3629103Z E ), 2023-03-28T14:33:05.3629919Z E fn_tree='ivy.functional.frontends.tensorflow.keras.metrics.poisson', 2023-03-28T14:33:05.3630519Z E frontend='tensorflow', 2023-03-28T14:33:05.3631162Z E on_device='cpu', 2023-03-28T14:33:05.3631544Z E ) 2023-03-28T14:33:05.3632082Z E 2023-03-28T14:33:05.3632999Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2BkAAMoBaaR2QwMAACDAAY=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_metrics.py::test_tensorflow_poisson[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T14:33:05.3604181Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3614332Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3622297Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3623141Z E Falsifying example: test_tensorflow_poisson( 2023-03-28T14:33:05.3623699Z E dtype_and_x=(['float32', 'float32'], 2023-03-28T14:33:05.3624463Z E [array([-1.], dtype=float32), array([-1.], dtype=float32)]), 2023-03-28T14:33:05.3625625Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T14:33:05.3626079Z E num_positional_args=0, 2023-03-28T14:33:05.3626684Z E with_out=False, 2023-03-28T14:33:05.3627077Z E inplace=False, 2023-03-28T14:33:05.3627671Z E as_variable=[False], 2023-03-28T14:33:05.3628079Z E native_arrays=[False], 2023-03-28T14:33:05.3628702Z E generate_frontend_arrays=False, 2023-03-28T14:33:05.3629103Z E ), 2023-03-28T14:33:05.3629919Z E fn_tree='ivy.functional.frontends.tensorflow.keras.metrics.poisson', 2023-03-28T14:33:05.3630519Z E frontend='tensorflow', 2023-03-28T14:33:05.3631162Z E on_device='cpu', 2023-03-28T14:33:05.3631544Z E ) 2023-03-28T14:33:05.3632082Z E 2023-03-28T14:33:05.3632999Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2BkAAMoBaaR2QwMAACDAAY=') as a decorator on your test case </details> <details> <summary>FAILED ivy_tests/test_ivy/test_frontends/test_tensorflow/test_metrics.py::test_tensorflow_poisson[cpu-ivy.functional.backends.jax-False-False]</summary> 2023-03-28T14:33:05.3604181Z E TypeError: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3614332Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3622297Z E ivy.utils.exceptions.IvyBackendException: tensorflow: nested_map: tensorflow: nested_map: to_native() got an unexpected keyword argument 'dtype' 2023-03-28T14:33:05.3623141Z E Falsifying example: test_tensorflow_poisson( 2023-03-28T14:33:05.3623699Z E dtype_and_x=(['float32', 'float32'], 2023-03-28T14:33:05.3624463Z E [array([-1.], dtype=float32), array([-1.], dtype=float32)]), 2023-03-28T14:33:05.3625625Z E test_flags=FrontendFunctionTestFlags( 2023-03-28T14:33:05.3626079Z E num_positional_args=0, 2023-03-28T14:33:05.3626684Z E with_out=False, 2023-03-28T14:33:05.3627077Z E inplace=False, 2023-03-28T14:33:05.3627671Z E as_variable=[False], 2023-03-28T14:33:05.3628079Z E native_arrays=[False], 2023-03-28T14:33:05.3628702Z E generate_frontend_arrays=False, 2023-03-28T14:33:05.3629103Z E ), 2023-03-28T14:33:05.3629919Z E fn_tree='ivy.functional.frontends.tensorflow.keras.metrics.poisson', 2023-03-28T14:33:05.3630519Z E frontend='tensorflow', 2023-03-28T14:33:05.3631162Z E on_device='cpu', 2023-03-28T14:33:05.3631544Z E ) 2023-03-28T14:33:05.3632082Z E 2023-03-28T14:33:05.3632999Z E You can reproduce this example by temporarily adding @reproduce_failure('6.70.1', b'AXicY2BkAAMoBaaR2QwMAACDAAY=') as a decorator on your test case </details>
test
fix metrics test tensorflow poisson tensorflow img src torch img src numpy img src jax img src failed ivy tests test ivy test frontends test tensorflow test metrics py test tensorflow poisson e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow poisson e dtype and x e dtype array dtype e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends tensorflow keras metrics poisson e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test metrics py test tensorflow poisson e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow poisson e dtype and x e dtype array dtype e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends tensorflow keras metrics poisson e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test metrics py test tensorflow poisson e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow poisson e dtype and x e dtype array dtype e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends tensorflow keras metrics poisson e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case failed ivy tests test ivy test frontends test tensorflow test metrics py test tensorflow poisson e typeerror to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map to native got an unexpected keyword argument dtype e ivy utils exceptions ivybackendexception tensorflow nested map tensorflow nested map to native got an unexpected keyword argument dtype e falsifying example test tensorflow poisson e dtype and x e dtype array dtype e test flags frontendfunctiontestflags e num positional args e with out false e inplace false e as variable e native arrays e generate frontend arrays false e e fn tree ivy functional frontends tensorflow keras metrics poisson e frontend tensorflow e on device cpu e e e you can reproduce this example by temporarily adding reproduce failure b as a decorator on your test case
1
147,852
11,810,060,004
IssuesEvent
2020-03-19 15:54:10
kowainik/autopack
https://api.github.com/repos/kowainik/autopack
opened
Create testing environment
tests
This could involve transformation to the multipackage project with the testing package check on different platforms at CI
1.0
Create testing environment - This could involve transformation to the multipackage project with the testing package check on different platforms at CI
test
create testing environment this could involve transformation to the multipackage project with the testing package check on different platforms at ci
1
16,476
3,534,764,057
IssuesEvent
2016-01-16 01:27:00
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
GCE L7 LoadBalancer Controller test failing in gke-flaky suite.
area/test priority/P2 team/cluster team/test-infra
`GCE L7 LoadBalancer Controller [Serial] [Slow] [Flaky] should create GCE L7 loadbalancers and verify Ingress` It probably shouldn't be running in the gke-flaky suite. Looks like #19021 might've added it, and it's failed every run since then.
2.0
GCE L7 LoadBalancer Controller test failing in gke-flaky suite. - `GCE L7 LoadBalancer Controller [Serial] [Slow] [Flaky] should create GCE L7 loadbalancers and verify Ingress` It probably shouldn't be running in the gke-flaky suite. Looks like #19021 might've added it, and it's failed every run since then.
test
gce loadbalancer controller test failing in gke flaky suite gce loadbalancer controller should create gce loadbalancers and verify ingress it probably shouldn t be running in the gke flaky suite looks like might ve added it and it s failed every run since then
1
237,422
19,623,090,247
IssuesEvent
2022-01-07 09:40:10
LimeChain/hedera-services
https://api.github.com/repos/LimeChain/hedera-services
closed
Implement Mint Precompile E2E Tests
p2 tests
- [x] HSCS-PREC-002 - [x] HSCS-PREC-003 - [x] HSCS-PREC-012 - [x] HSCS-PREC-019 - [x] HSCS-PREC-024 - [x] Make sure we are using the latest solidity library version (ping Danno afterwards)
1.0
Implement Mint Precompile E2E Tests - - [x] HSCS-PREC-002 - [x] HSCS-PREC-003 - [x] HSCS-PREC-012 - [x] HSCS-PREC-019 - [x] HSCS-PREC-024 - [x] Make sure we are using the latest solidity library version (ping Danno afterwards)
test
implement mint precompile tests hscs prec hscs prec hscs prec hscs prec hscs prec make sure we are using the latest solidity library version ping danno afterwards
1
33,173
14,008,725,626
IssuesEvent
2020-10-29 00:26:17
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Missing documentation of "Keep current state" in "Post Execution State"
Pri2 cognitive-services/svc cxp doc-bug speech-service/subsvc triaged
Post Execution State documentation mentions 3 options. There are now 4. The "Keep current state" option is not mentioned here: https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/custom-commands-references#post-execution-state --- #### Document Details โš  *Do not edit this section. It is required for docs.microsoft.com โžŸ GitHub issue linking.* * ID: d4e62f29-4423-9e2d-5be2-8471581743e2 * Version Independent ID: 3436a82a-ba1a-0b3a-7736-067fdbfe6258 * Content: [Custom Commands concepts and definitions - Speech service - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/custom-commands-references) * Content Source: [articles/cognitive-services/Speech-Service/custom-commands-references.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/cognitive-services/Speech-Service/custom-commands-references.md) * Service: **cognitive-services** * Sub-service: **speech-service** * GitHub Login: @singhsaumya * Microsoft Alias: **sausin**
2.0
Missing documentation of "Keep current state" in "Post Execution State" - Post Execution State documentation mentions 3 options. There are now 4. The "Keep current state" option is not mentioned here: https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/custom-commands-references#post-execution-state --- #### Document Details โš  *Do not edit this section. It is required for docs.microsoft.com โžŸ GitHub issue linking.* * ID: d4e62f29-4423-9e2d-5be2-8471581743e2 * Version Independent ID: 3436a82a-ba1a-0b3a-7736-067fdbfe6258 * Content: [Custom Commands concepts and definitions - Speech service - Azure Cognitive Services](https://docs.microsoft.com/en-us/azure/cognitive-services/speech-service/custom-commands-references) * Content Source: [articles/cognitive-services/Speech-Service/custom-commands-references.md](https://github.com/MicrosoftDocs/azure-docs/blob/master/articles/cognitive-services/Speech-Service/custom-commands-references.md) * Service: **cognitive-services** * Sub-service: **speech-service** * GitHub Login: @singhsaumya * Microsoft Alias: **sausin**
non_test
missing documentation of keep current state in post execution state post execution state documentation mentions options there are now the keep current state option is not mentioned here document details โš  do not edit this section it is required for docs microsoft com โžŸ github issue linking id version independent id content content source service cognitive services sub service speech service github login singhsaumya microsoft alias sausin
0
723,660
24,905,189,442
IssuesEvent
2022-10-29 06:24:30
bounswe/bounswe2022group7
https://api.github.com/repos/bounswe/bounswe2022group7
opened
Connecting the Register Form to the Backend
Status: In Progress Priority: High Difficulty: Medium Type: Implementation Target: Mobile
As discussed in the [Mobile Team Meeting 3](https://github.com/bounswe/bounswe2022group7/wiki/MOBILE-Meeting-Notes-%233), I was assigned to complete the register functionality by connecting to the back end register end point. * Here are the tasks that will be done to complete this issue: - [ ] Create input and output data transfer objects. - [ ] Take all inputs in the register form and put it into the input object. - [ ] Send input object to "/register" endpoint. - [ ] Register or show an error message according to response. **Deadline: 29/10/2022, 23:59** **Reviewer: @canatakan *
1.0
Connecting the Register Form to the Backend - As discussed in the [Mobile Team Meeting 3](https://github.com/bounswe/bounswe2022group7/wiki/MOBILE-Meeting-Notes-%233), I was assigned to complete the register functionality by connecting to the back end register end point. * Here are the tasks that will be done to complete this issue: - [ ] Create input and output data transfer objects. - [ ] Take all inputs in the register form and put it into the input object. - [ ] Send input object to "/register" endpoint. - [ ] Register or show an error message according to response. **Deadline: 29/10/2022, 23:59** **Reviewer: @canatakan *
non_test
connecting the register form to the backend as discussed in the i was assigned to complete the register functionality by connecting to the back end register end point here are the tasks that will be done to complete this issue create input and output data transfer objects take all inputs in the register form and put it into the input object send input object to register endpoint register or show an error message according to response deadline reviewer canatakan
0
25,755
25,834,877,332
IssuesEvent
2022-12-12 18:48:27
rokwire/illinois-app
https://api.github.com/repos/rokwire/illinois-app
closed
[USABILITY] Group privacy setting description
Type: Question Type: Usability
### Describe the usability issue Description text about the visibility of a group. ### To Reproduce Steps to reproduce the behavior: In the group visibility setting, the description text says: Private: ``` Only members can see group events. Anyone who uses the Illinois app can find this group. Only admins can see whose in the group. ``` Public: ``` Anyone can see group events. Anyone who uses the Illinois app can find this group. Only admins can see whose in the group. ``` ### Expected behavior 1. The second sentence in the description for both Private and Public groups is the same. Is this correct? 2. `Only admins can see whose in the group.` needs to be corrected to `Only admins can see who is in the group.` ### Screenshots If applicable, add screenshots to help explain your problem. ### Smartphone (please complete the following information): Version: 2.5.9 Additional context Add any other context about the problem here.
True
[USABILITY] Group privacy setting description - ### Describe the usability issue Description text about the visibility of a group. ### To Reproduce Steps to reproduce the behavior: In the group visibility setting, the description text says: Private: ``` Only members can see group events. Anyone who uses the Illinois app can find this group. Only admins can see whose in the group. ``` Public: ``` Anyone can see group events. Anyone who uses the Illinois app can find this group. Only admins can see whose in the group. ``` ### Expected behavior 1. The second sentence in the description for both Private and Public groups is the same. Is this correct? 2. `Only admins can see whose in the group.` needs to be corrected to `Only admins can see who is in the group.` ### Screenshots If applicable, add screenshots to help explain your problem. ### Smartphone (please complete the following information): Version: 2.5.9 Additional context Add any other context about the problem here.
non_test
group privacy setting description describe the usability issue description text about the visibility of a group to reproduce steps to reproduce the behavior in the group visibility setting the description text says private only members can see group events anyone who uses the illinois app can find this group only admins can see whose in the group public anyone can see group events anyone who uses the illinois app can find this group only admins can see whose in the group expected behavior the second sentence in the description for both private and public groups is the same is this correct only admins can see whose in the group needs to be corrected to only admins can see who is in the group screenshots if applicable add screenshots to help explain your problem smartphone please complete the following information version additional context add any other context about the problem here
0
174,868
21,300,518,115
IssuesEvent
2022-04-15 02:03:14
chiq2045/grunt-demo
https://api.github.com/repos/chiq2045/grunt-demo
opened
CVE-2019-10747 (High) detected in set-value-0.4.3.tgz, set-value-2.0.0.tgz
security vulnerability
## CVE-2019-10747 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-0.4.3.tgz</b>, <b>set-value-2.0.0.tgz</b></p></summary> <p> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: /grunt-demo/package.json</p> <p>Path to vulnerable library: /node_modules/union-value/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - gulp-cli-2.2.0.tgz (Root Library) - matchdep-2.0.0.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: /grunt-demo/package.json</p> <p>Path to vulnerable library: /node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - gulp-cli-2.2.0.tgz (Root Library) - matchdep-2.0.0.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads. <p>Publish Date: 2019-08-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747>CVE-2019-10747</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p> <p>Release Date: 2019-10-29</p> <p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (gulp-cli): 2.2.1</p><p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (gulp-cli): 2.2.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2019-10747 (High) detected in set-value-0.4.3.tgz, set-value-2.0.0.tgz - ## CVE-2019-10747 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>set-value-0.4.3.tgz</b>, <b>set-value-2.0.0.tgz</b></p></summary> <p> <details><summary><b>set-value-0.4.3.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz">https://registry.npmjs.org/set-value/-/set-value-0.4.3.tgz</a></p> <p>Path to dependency file: /grunt-demo/package.json</p> <p>Path to vulnerable library: /node_modules/union-value/node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - gulp-cli-2.2.0.tgz (Root Library) - matchdep-2.0.0.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - union-value-1.0.0.tgz - :x: **set-value-0.4.3.tgz** (Vulnerable Library) </details> <details><summary><b>set-value-2.0.0.tgz</b></p></summary> <p>Create nested values and any intermediaries using dot notation (`'a.b.c'`) paths.</p> <p>Library home page: <a href="https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz">https://registry.npmjs.org/set-value/-/set-value-2.0.0.tgz</a></p> <p>Path to dependency file: /grunt-demo/package.json</p> <p>Path to vulnerable library: /node_modules/set-value/package.json</p> <p> Dependency Hierarchy: - gulp-cli-2.2.0.tgz (Root Library) - matchdep-2.0.0.tgz - micromatch-3.1.10.tgz - snapdragon-0.8.2.tgz - base-0.11.2.tgz - cache-base-1.0.1.tgz - :x: **set-value-2.0.0.tgz** (Vulnerable Library) </details> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> set-value is vulnerable to Prototype Pollution in versions lower than 3.0.1. The function mixin-deep could be tricked into adding or modifying properties of Object.prototype using any of the constructor, prototype and _proto_ payloads. <p>Publish Date: 2019-08-23 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2019-10747>CVE-2019-10747</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f">https://github.com/jonschlinkert/set-value/commit/95e9d9923f8a8b4a01da1ea138fcc39ec7b6b15f</a></p> <p>Release Date: 2019-10-29</p> <p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (gulp-cli): 2.2.1</p><p>Fix Resolution (set-value): 2.0.1</p> <p>Direct dependency fix Resolution (gulp-cli): 2.2.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in set value tgz set value tgz cve high severity vulnerability vulnerable libraries set value tgz set value tgz set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file grunt demo package json path to vulnerable library node modules union value node modules set value package json dependency hierarchy gulp cli tgz root library matchdep tgz micromatch tgz snapdragon tgz base tgz cache base tgz union value tgz x set value tgz vulnerable library set value tgz create nested values and any intermediaries using dot notation a b c paths library home page a href path to dependency file grunt demo package json path to vulnerable library node modules set value package json dependency hierarchy gulp cli tgz root library matchdep tgz micromatch tgz snapdragon tgz base tgz cache base tgz x set value tgz vulnerable library vulnerability details set value is vulnerable to prototype pollution in versions lower than the function mixin deep could be tricked into adding or modifying properties of object prototype using any of the constructor prototype and proto payloads publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution set value direct dependency fix resolution gulp cli fix resolution set value direct dependency fix resolution gulp cli step up your open source security game with whitesource
0
385
2,499,902,393
IssuesEvent
2015-01-08 07:43:20
fossology/fossology
https://api.github.com/repos/fossology/fossology
opened
Add functional tests for Tagging
Category: Testing Component: Rank Component: Tester Priority: High Status: New Tracker: Support
--- Author Name: **Paul Holland** Original Redmine Issue: 2625, http://www.fossology.org/issues/2625 Original Date: 2012/06/15 Original Assignee: Bob Gobeille --- Fills in http://www.fossology.org/projects/fossology/wiki/FOSSology_tests
2.0
Add functional tests for Tagging - --- Author Name: **Paul Holland** Original Redmine Issue: 2625, http://www.fossology.org/issues/2625 Original Date: 2012/06/15 Original Assignee: Bob Gobeille --- Fills in http://www.fossology.org/projects/fossology/wiki/FOSSology_tests
test
add functional tests for tagging author name paul holland original redmine issue original date original assignee bob gobeille fills in
1
558,818
16,543,184,701
IssuesEvent
2021-05-27 19:41:21
ita-social-projects/TeachUA
https://api.github.com/repos/ita-social-projects/TeachUA
closed
[ะ ะพะทัˆะธั€ะตะฝะธะน ะฟะพัˆัƒะบ] The search is filtered by forbidden numbers
Priority: Medium bug
**Environment:** Windows 10, Google Chrome Version 90.0.4430.93 (Official Build) (64-bit) **Reproducible:** always **Build found:** https://speak-ukrainian.org.ua/dev/ Steps to reproduce 1. Go to 'https://speak-ukrainian.org.ua/dev/' 2. Open 'ะ“ัƒั€ั‚ะบะธ' page 3. Click on advanced search icon to open advanced search 4. Click on the age input and enter forbidden number(e.g. 184) **Actual result** The age input accepts numbers below and abow (2-18)interval and filter the search result by it ![image](https://user-images.githubusercontent.com/32505957/119124673-99225b80-ba39-11eb-911d-4998a8b7e7af.png) Expected result The age input accepts only positive integers from 2 to 18 User story and Bug report User story #274 Bug report #434 Labels to be added "Bug", Priority ("pri:Medium "), Severity ("severity:Minor"), Type ("Functional").
1.0
[ะ ะพะทัˆะธั€ะตะฝะธะน ะฟะพัˆัƒะบ] The search is filtered by forbidden numbers - **Environment:** Windows 10, Google Chrome Version 90.0.4430.93 (Official Build) (64-bit) **Reproducible:** always **Build found:** https://speak-ukrainian.org.ua/dev/ Steps to reproduce 1. Go to 'https://speak-ukrainian.org.ua/dev/' 2. Open 'ะ“ัƒั€ั‚ะบะธ' page 3. Click on advanced search icon to open advanced search 4. Click on the age input and enter forbidden number(e.g. 184) **Actual result** The age input accepts numbers below and abow (2-18)interval and filter the search result by it ![image](https://user-images.githubusercontent.com/32505957/119124673-99225b80-ba39-11eb-911d-4998a8b7e7af.png) Expected result The age input accepts only positive integers from 2 to 18 User story and Bug report User story #274 Bug report #434 Labels to be added "Bug", Priority ("pri:Medium "), Severity ("severity:Minor"), Type ("Functional").
non_test
the search is filtered by forbidden numbers environment windows google chrome version official build bit reproducible always build found steps to reproduce go to open ะณัƒั€ั‚ะบะธ page click on advanced search icon to open advanced search click on the age input and enter forbidden number e g actual result the age input accepts numbers below and abow interval and filter the search result by it expected result the age input accepts only positive integers from to user story and bug report user story bug report labels to be added bug priority pri medium severity severity minor type functional
0
39,553
5,240,830,913
IssuesEvent
2017-01-31 14:15:27
ValveSoftware/Source-1-Games
https://api.github.com/repos/ValveSoftware/Source-1-Games
closed
[CS:S] Disconnect freeze, To many particles, Material Error, Graphical Error
Counter-Strike: Source Need Retest Reviewed
Issue transferred from ValveSoftware/steam-for-linux#1517 @CommanderAlchemy posted at 2013-02-10T04:51:44Z: Disconnecting from a server freezes the client for some seconds, cannot remember the same thing happening on Windows. Also i get couple of complains in the console about particles: "Too many simultaneously active particles! Particles: maximum verts exceeded: 56328 verts, 84492 indexes" Material error: Error: Material "debug/debugluxels" uses unknown shader "DebugLuxels" Error: Material "models/humans/female/pupil_r" uses unknown shader "Eyeball" Error: Material "models/humans/female/pupil_l" uses unknown shader "Eyeball" Graphical error: Dust2, when smoke covers by the middle doors and you look into the lower tunnel you can see the lights shine through the smoke. Not sure if i should split this into 4 bug-reports but here I go. Best Regards Processor Information: Vendor: AuthenticAMD Speed: 3700 Mhz 4 logical processors 4 physical processors HyperThreading: Unsupported FCMOV: Supported SSE2: Supported SSE3: Supported SSSE3: Supported SSE4a: Supported SSE41: Unsupported SSE42: Unsupported Network Information: Network Speed: Operating System Version: "Arch Linux" (64 bit) Kernel Name: Linux Kernel Version: 3.7.6-1-ARCH X Server vendor: The X.Org Foundation X Server release: 11302000 Video Card: Driver: NVIDIA Corporation GeForce GTX 670/PCIe/SSE2 ``` Driver Version: 4.3.0 NVIDIA 313.18 Desktop Color Depth: 24 bits per pixel Monitor Refresh Rate: 60 Hz VendorID: 0x10de DeviceID: 0x1189 Number of Monitors: 3 Number of Logical Video Cards: 1 Primary Display Resolution: 1920 x 1080 Desktop Resolution: 5760 x 1080 Primary Display Size: 18.78" x 10.55" (21.54" diag) 47.7cm x 26.8cm (54.7cm diag) Primary Bus: PCI Express 16x Primary VRAM: 2048 MB Supported MSAA Modes: 2x 4x 8x 16x ``` Sound card: Audio device: AV200 Memory: RAM: 7988 Mb Miscellaneous: UI Language: English LANG: en_US.UTF-8 Microphone: Not set Total Hard Disk Space Available: 55436 Mb Largest Free Hard Disk Block: 3330 Mb Installed software: Recent Failure Reports: Sat Feb 9 19:03:41 2013 GMT: file ''/tmp/dumps/crash_20130209200338_1.dmp'', upload yes: ''CrashID=bp-53d1d340-9956-425a-98c1-ae5e92130209''
1.0
[CS:S] Disconnect freeze, To many particles, Material Error, Graphical Error - Issue transferred from ValveSoftware/steam-for-linux#1517 @CommanderAlchemy posted at 2013-02-10T04:51:44Z: Disconnecting from a server freezes the client for some seconds, cannot remember the same thing happening on Windows. Also i get couple of complains in the console about particles: "Too many simultaneously active particles! Particles: maximum verts exceeded: 56328 verts, 84492 indexes" Material error: Error: Material "debug/debugluxels" uses unknown shader "DebugLuxels" Error: Material "models/humans/female/pupil_r" uses unknown shader "Eyeball" Error: Material "models/humans/female/pupil_l" uses unknown shader "Eyeball" Graphical error: Dust2, when smoke covers by the middle doors and you look into the lower tunnel you can see the lights shine through the smoke. Not sure if i should split this into 4 bug-reports but here I go. Best Regards Processor Information: Vendor: AuthenticAMD Speed: 3700 Mhz 4 logical processors 4 physical processors HyperThreading: Unsupported FCMOV: Supported SSE2: Supported SSE3: Supported SSSE3: Supported SSE4a: Supported SSE41: Unsupported SSE42: Unsupported Network Information: Network Speed: Operating System Version: "Arch Linux" (64 bit) Kernel Name: Linux Kernel Version: 3.7.6-1-ARCH X Server vendor: The X.Org Foundation X Server release: 11302000 Video Card: Driver: NVIDIA Corporation GeForce GTX 670/PCIe/SSE2 ``` Driver Version: 4.3.0 NVIDIA 313.18 Desktop Color Depth: 24 bits per pixel Monitor Refresh Rate: 60 Hz VendorID: 0x10de DeviceID: 0x1189 Number of Monitors: 3 Number of Logical Video Cards: 1 Primary Display Resolution: 1920 x 1080 Desktop Resolution: 5760 x 1080 Primary Display Size: 18.78" x 10.55" (21.54" diag) 47.7cm x 26.8cm (54.7cm diag) Primary Bus: PCI Express 16x Primary VRAM: 2048 MB Supported MSAA Modes: 2x 4x 8x 16x ``` Sound card: Audio device: AV200 Memory: RAM: 7988 Mb Miscellaneous: UI Language: English LANG: en_US.UTF-8 Microphone: Not set Total Hard Disk Space Available: 55436 Mb Largest Free Hard Disk Block: 3330 Mb Installed software: Recent Failure Reports: Sat Feb 9 19:03:41 2013 GMT: file ''/tmp/dumps/crash_20130209200338_1.dmp'', upload yes: ''CrashID=bp-53d1d340-9956-425a-98c1-ae5e92130209''
test
disconnect freeze to many particles material error graphical error issue transferred from valvesoftware steam for linux commanderalchemy posted at disconnecting from a server freezes the client for some seconds cannot remember the same thing happening on windows also i get couple of complains in the console about particles too many simultaneously active particles particles maximum verts exceeded verts indexes material error error material debug debugluxels uses unknown shader debugluxels error material models humans female pupil r uses unknown shader eyeball error material models humans female pupil l uses unknown shader eyeball graphical error when smoke covers by the middle doors and you look into the lower tunnel you can see the lights shine through the smoke not sure if i should split this into bug reports but here i go best regards processor information vendor authenticamd speed mhz logical processors physical processors hyperthreading unsupported fcmov supported supported supported supported supported unsupported unsupported network information network speed operating system version arch linux bit kernel name linux kernel version arch x server vendor the x org foundation x server release video card driver nvidia corporation geforce gtx pcie driver version nvidia desktop color depth bits per pixel monitor refresh rate hz vendorid deviceid number of monitors number of logical video cards primary display resolution x desktop resolution x primary display size x diag x diag primary bus pci express primary vram mb supported msaa modes sound card audio device memory ram mb miscellaneous ui language english lang en us utf microphone not set total hard disk space available mb largest free hard disk block mb installed software recent failure reports sat feb gmt file tmp dumps crash dmp upload yes crashid bp
1
315,127
27,047,165,756
IssuesEvent
2023-02-13 10:35:05
IDgis/PlanoView2-bevindingen
https://api.github.com/repos/IDgis/PlanoView2-bevindingen
closed
zoekfunctie
getest met bevindingen
bij nieuw zoekopdracht zonder knop verwijder selectie te gebruiken wordt wel de nieuwe zoekopdracht gehighlight in inhoudsopgave maar de oude zoekopdracht blijft nog aangevinkt staan en gehighlight in de tekst. Bij nieuwe zoekopdracht automatisch oude zoekopdracht laten verwijderen ? ![image](https://user-images.githubusercontent.com/30464824/200583799-8975de1f-e27b-4abd-9946-098c57da1116.png) ![image](https://user-images.githubusercontent.com/30464824/200584411-d47b9597-9995-4192-9ca6-ff4a79ee45b5.png)
1.0
zoekfunctie - bij nieuw zoekopdracht zonder knop verwijder selectie te gebruiken wordt wel de nieuwe zoekopdracht gehighlight in inhoudsopgave maar de oude zoekopdracht blijft nog aangevinkt staan en gehighlight in de tekst. Bij nieuwe zoekopdracht automatisch oude zoekopdracht laten verwijderen ? ![image](https://user-images.githubusercontent.com/30464824/200583799-8975de1f-e27b-4abd-9946-098c57da1116.png) ![image](https://user-images.githubusercontent.com/30464824/200584411-d47b9597-9995-4192-9ca6-ff4a79ee45b5.png)
test
zoekfunctie bij nieuw zoekopdracht zonder knop verwijder selectie te gebruiken wordt wel de nieuwe zoekopdracht gehighlight in inhoudsopgave maar de oude zoekopdracht blijft nog aangevinkt staan en gehighlight in de tekst bij nieuwe zoekopdracht automatisch oude zoekopdracht laten verwijderen
1
50,319
6,077,796,876
IssuesEvent
2017-06-16 05:55:20
red/red
https://api.github.com/repos/red/red
closed
[macOS View] "Strike-through" font style not working on 'Base
GUI status.built status.tested type.bug
It works on button, but not works on base. ```Red f: make font! [ style: 'strike ] view [ base font f "OK" ] ``` This bug is only on macOS.
1.0
[macOS View] "Strike-through" font style not working on 'Base - It works on button, but not works on base. ```Red f: make font! [ style: 'strike ] view [ base font f "OK" ] ``` This bug is only on macOS.
test
strike through font style not working on base it works on button but not works on base red f make font view this bug is only on macos
1
51,007
10,577,235,577
IssuesEvent
2019-10-07 19:39:23
chatwoot/chatwoot
https://api.github.com/repos/chatwoot/chatwoot
opened
Fix "method_complexity" issue in app/models/conversation.rb
codeclimate hacktoberfest
Method `run_round_robin` has a Cognitive Complexity of 11 (exceeds 5 allowed). Consider refactoring. https://codeclimate.com/github/chatwoot/chatwoot/app/models/conversation.rb#issue_5d9b93b4d887000001000191
1.0
Fix "method_complexity" issue in app/models/conversation.rb - Method `run_round_robin` has a Cognitive Complexity of 11 (exceeds 5 allowed). Consider refactoring. https://codeclimate.com/github/chatwoot/chatwoot/app/models/conversation.rb#issue_5d9b93b4d887000001000191
non_test
fix method complexity issue in app models conversation rb method run round robin has a cognitive complexity of exceeds allowed consider refactoring
0
300,717
25,991,370,349
IssuesEvent
2022-12-20 07:53:30
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: X-Pack Alerting API Integration Tests.x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disableยทts - alerting api integration security and spaces enabled Alerts - Group 1 alerts bulkDisableRules superuser at space1 should not disable rule from another space
failed-test Team:ResponseOps
A test failed on a tracked branch ``` TypeError: Cannot read properties of undefined (reading 'id') at getDefaultRules (x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disable.ts:14:30) at getDefaultResponse (x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disable.ts:42:11) at Context.<anonymous> (x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disable.ts:437:44) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Object.apply (node_modules/@kbn/test/target_node/src/functional_test_runner/lib/mocha/wrap_function.js:78:16) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/23997#0184a3d2-9961-4e7a-bf4b-22a512f0d8ae) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Alerting API Integration Tests.x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disableยทts","test.name":"alerting api integration security and spaces enabled Alerts - Group 1 alerts bulkDisableRules superuser at space1 should not disable rule from another space","test.failCount":1}} -->
1.0
Failing test: X-Pack Alerting API Integration Tests.x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disableยทts - alerting api integration security and spaces enabled Alerts - Group 1 alerts bulkDisableRules superuser at space1 should not disable rule from another space - A test failed on a tracked branch ``` TypeError: Cannot read properties of undefined (reading 'id') at getDefaultRules (x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disable.ts:14:30) at getDefaultResponse (x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disable.ts:42:11) at Context.<anonymous> (x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disable.ts:437:44) at runMicrotasks (<anonymous>) at processTicksAndRejections (node:internal/process/task_queues:96:5) at Object.apply (node_modules/@kbn/test/target_node/src/functional_test_runner/lib/mocha/wrap_function.js:78:16) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/23997#0184a3d2-9961-4e7a-bf4b-22a512f0d8ae) <!-- kibanaCiData = {"failed-test":{"test.class":"X-Pack Alerting API Integration Tests.x-pack/test/alerting_api_integration/security_and_spaces/group1/tests/alerting/bulk_disableยทts","test.name":"alerting api integration security and spaces enabled Alerts - Group 1 alerts bulkDisableRules superuser at space1 should not disable rule from another space","test.failCount":1}} -->
test
failing test x pack alerting api integration tests x pack test alerting api integration security and spaces tests alerting bulk disableยทts alerting api integration security and spaces enabled alerts group alerts bulkdisablerules superuser at should not disable rule from another space a test failed on a tracked branch typeerror cannot read properties of undefined reading id at getdefaultrules x pack test alerting api integration security and spaces tests alerting bulk disable ts at getdefaultresponse x pack test alerting api integration security and spaces tests alerting bulk disable ts at context x pack test alerting api integration security and spaces tests alerting bulk disable ts at runmicrotasks at processticksandrejections node internal process task queues at object apply node modules kbn test target node src functional test runner lib mocha wrap function js first failure
1
116,233
9,829,137,508
IssuesEvent
2019-06-15 17:54:40
AlternateLife/Bugs
https://api.github.com/repos/AlternateLife/Bugs
opened
Bug Krรคuterhรคndlergarage
bug ingame stresstest
Beschreibe den Fehler Crash am Golfplatz an der gleichen Stelle, an der Staiy Probleme hat. Reproduktion des Fehlers Garage Krรคuterhรคndler Erwartetes Verhalten Kein Crash Screenshots [URL=https://www.bilder-upload.eu/bild-012cce-1560621127.jpg.html][IMG]https://www.bilder-upload.eu/thumb/012cce-1560621127.jpg[/IMG][/URL] System Betriebssystem: Windows 10 GTA Version: Steam
1.0
Bug Krรคuterhรคndlergarage - Beschreibe den Fehler Crash am Golfplatz an der gleichen Stelle, an der Staiy Probleme hat. Reproduktion des Fehlers Garage Krรคuterhรคndler Erwartetes Verhalten Kein Crash Screenshots [URL=https://www.bilder-upload.eu/bild-012cce-1560621127.jpg.html][IMG]https://www.bilder-upload.eu/thumb/012cce-1560621127.jpg[/IMG][/URL] System Betriebssystem: Windows 10 GTA Version: Steam
test
bug krรคuterhรคndlergarage beschreibe den fehler crash am golfplatz an der gleichen stelle an der staiy probleme hat reproduktion des fehlers garage krรคuterhรคndler erwartetes verhalten kein crash screenshots system betriebssystem windows gta version steam
1
215,149
16,652,743,192
IssuesEvent
2021-06-05 01:08:27
backend-br/vagas
https://api.github.com/repos/backend-br/vagas
closed
[Remoto] Desenvolvedor node.js na 3,2,1 Beauty
Alocado CI GraphQL JavaScript PJ PostgreSQL Presencial RabbitMQ Redis Remoto Scrum Stale TDD Testes automatizados TypeScript startup
### Sobre a empresa A 3,2,1 Beauty entrega serviรงos de beleza e bem-estar em grandes empresas, condomรญnios residenciais e em salรตes de rua, com agendamento e pagamento online.ย  Acreditamos que รฉ possรญvel conciliar o bem estar e a rotina agitada. Usamos a tecnologia para entregar serviรงos premium a preรงos acessรญveis, em tempo recorde, e com a melhor experiรชncia ao cliente. Desejamos causar um impacto positivo na nossa comunidade. Nascemos com a missรฃo empoderar profissionais de beleza a mudar de patamar de vida. Somos uma startup agile, de alto crescimento e com um modelo de negรณcios disruptivo. Hoje temos 10 lojas de rua, alรฉm de 60 localizaรงรตes corporativas. Queremos redefinir a relaรงรฃo das pessoas com a beleza e transformar o mercado de serviรงos de beleza no Brasil e no mundo. ### Descriรงรฃo da vaga Estamos buscando backend sรชnior para ajudar no desenvolvimento da nossa plataforma, facilitando ainda mais a vida dos nossos clientes e parceiros. Para se candidatar a essa vaga, vocรช deve ter uma grande experiรชnciaย comprovadaย com node.js, banco de dados relacional e nรฃo relacional, arquitetura de cรณdigo, alรฉm de conhecimentos em metodologias รกgeis como scrum e kanban. Ambiente descontraรญdo, com liberdade de horรกrios e muita autonomia. A vaga รฉ remota, mas pode ser presencial caso vocรช prefira e more na cidade de Sรฃo Paulo. ### Requisitos **Linguagens/tecnologias/frameworks:** - Javascript ES6 - Node (**mais de 4 anos de experiรชncia**) - CI/CD - Typescript - Go (Desejรกvel) **Banco de dados/Engines:** - Mongo - PostgreSQL - Redis - GraphQL - ElasticSearch (Desejรกvel) **Outros requisitos:** - Kafka/RabbitMQ (Desejรกvel) - Microservices - Firebase (Cloud Messaging, Remote Config) - Clean Architecture - Testes automatizados (DDD, TDD, ATDD) - Vue.js (Desejรกvel) ### Contrataรงรฃo PJ ### Como se candidatar Envie um e-mail contendo as seguintes informaรงรตes para **[tech@321beauty.com.br](mailto:tech@321beauty.com.br?Subject=Tenho%20experiรชncia%20flutter)** - Assunto: **Tenho experiรชncia em desenvolvimento com node.js;** - Conte a histรณria do projeto mais legal que vocรช jรก desenvolveu e qual foi sua participaรงรฃo no desenvolvimento dele; - Pretensรฃo salarial - Seu LinkedIn - Seu Github - Seu StackOverflow ### Labels Alocado ou Remoto Sรชnior PJ Salรกrio a combinar Ninja
1.0
[Remoto] Desenvolvedor node.js na 3,2,1 Beauty - ### Sobre a empresa A 3,2,1 Beauty entrega serviรงos de beleza e bem-estar em grandes empresas, condomรญnios residenciais e em salรตes de rua, com agendamento e pagamento online.ย  Acreditamos que รฉ possรญvel conciliar o bem estar e a rotina agitada. Usamos a tecnologia para entregar serviรงos premium a preรงos acessรญveis, em tempo recorde, e com a melhor experiรชncia ao cliente. Desejamos causar um impacto positivo na nossa comunidade. Nascemos com a missรฃo empoderar profissionais de beleza a mudar de patamar de vida. Somos uma startup agile, de alto crescimento e com um modelo de negรณcios disruptivo. Hoje temos 10 lojas de rua, alรฉm de 60 localizaรงรตes corporativas. Queremos redefinir a relaรงรฃo das pessoas com a beleza e transformar o mercado de serviรงos de beleza no Brasil e no mundo. ### Descriรงรฃo da vaga Estamos buscando backend sรชnior para ajudar no desenvolvimento da nossa plataforma, facilitando ainda mais a vida dos nossos clientes e parceiros. Para se candidatar a essa vaga, vocรช deve ter uma grande experiรชnciaย comprovadaย com node.js, banco de dados relacional e nรฃo relacional, arquitetura de cรณdigo, alรฉm de conhecimentos em metodologias รกgeis como scrum e kanban. Ambiente descontraรญdo, com liberdade de horรกrios e muita autonomia. A vaga รฉ remota, mas pode ser presencial caso vocรช prefira e more na cidade de Sรฃo Paulo. ### Requisitos **Linguagens/tecnologias/frameworks:** - Javascript ES6 - Node (**mais de 4 anos de experiรชncia**) - CI/CD - Typescript - Go (Desejรกvel) **Banco de dados/Engines:** - Mongo - PostgreSQL - Redis - GraphQL - ElasticSearch (Desejรกvel) **Outros requisitos:** - Kafka/RabbitMQ (Desejรกvel) - Microservices - Firebase (Cloud Messaging, Remote Config) - Clean Architecture - Testes automatizados (DDD, TDD, ATDD) - Vue.js (Desejรกvel) ### Contrataรงรฃo PJ ### Como se candidatar Envie um e-mail contendo as seguintes informaรงรตes para **[tech@321beauty.com.br](mailto:tech@321beauty.com.br?Subject=Tenho%20experiรชncia%20flutter)** - Assunto: **Tenho experiรชncia em desenvolvimento com node.js;** - Conte a histรณria do projeto mais legal que vocรช jรก desenvolveu e qual foi sua participaรงรฃo no desenvolvimento dele; - Pretensรฃo salarial - Seu LinkedIn - Seu Github - Seu StackOverflow ### Labels Alocado ou Remoto Sรชnior PJ Salรกrio a combinar Ninja
test
desenvolvedor node js na beauty sobre a empresa a beauty entrega serviรงos de beleza e bem estar em grandes empresas condomรญnios residenciais e em salรตes de rua com agendamento e pagamento online ย  acreditamos que รฉ possรญvel conciliar o bem estar e a rotina agitada usamos a tecnologia para entregar serviรงos premium a preรงos acessรญveis em tempo recorde e com a melhor experiรชncia ao cliente desejamos causar um impacto positivo na nossa comunidade nascemos com a missรฃo empoderar profissionais de beleza a mudar de patamar de vida somos uma startup agile de alto crescimento e com um modelo de negรณcios disruptivo hoje temos lojas de rua alรฉm de localizaรงรตes corporativas queremos redefinir a relaรงรฃo das pessoas com a beleza e transformar o mercado de serviรงos de beleza no brasil e no mundo descriรงรฃo da vaga estamos buscando backend sรชnior para ajudar no desenvolvimento da nossa plataforma facilitando ainda mais a vida dos nossos clientes e parceiros para se candidatar a essa vaga vocรช deve ter uma grande experiรชnciaย comprovadaย com node js banco de dados relacional e nรฃo relacional arquitetura de cรณdigo alรฉm de conhecimentos em metodologias รกgeis como scrum e kanban ambiente descontraรญdo com liberdade de horรกrios e muita autonomia a vaga รฉ remota mas pode ser presencial caso vocรช prefira e more na cidade de sรฃo paulo requisitos linguagens tecnologias frameworks javascript node mais de anos de experiรชncia ci cd typescript go desejรกvel banco de dados engines mongo postgresql redis graphql elasticsearch desejรกvel outros requisitos kafka rabbitmq desejรกvel microservices firebase cloud messaging remote config clean architecture testes automatizados ddd tdd atdd vue js desejรกvel contrataรงรฃo pj como se candidatar envie um e mail contendo as seguintes informaรงรตes para mailto tech com br subject tenho assunto tenho experiรชncia em desenvolvimento com node js conte a histรณria do projeto mais legal que vocรช jรก desenvolveu e qual foi sua participaรงรฃo no desenvolvimento dele pretensรฃo salarial seu linkedin seu github seu stackoverflow labels alocado ou remoto sรชnior pj salรกrio a combinar ninja
1
242,905
20,322,102,190
IssuesEvent
2022-02-18 00:03:54
gravitational/teleport
https://api.github.com/repos/gravitational/teleport
opened
Teleport 9.0 Test Plan
bug testplan
## Manual Testing Plan Below are the items that should be manually tested with each release of Teleport. These tests should be run on both a fresh install of the version to be released as well as an upgrade of the previous version of Teleport. - [ ] Adding nodes to a cluster - [ ] Adding Nodes via Valid Static Token - [ ] Adding Nodes via Valid Short-lived Tokens - [ ] Adding Nodes via Invalid Token Fails - [ ] Revoking Node Invitation - [ ] Labels - [ ] Static Labels - [ ] Dynamic Labels - [ ] Trusted Clusters - [ ] Adding Trusted Cluster Valid Static Token - [ ] Adding Trusted Cluster Valid Short-lived Token - [ ] Adding Trusted Cluster Invalid Token - [ ] Removing Trusted Cluster - [ ] RBAC Make sure that invalid and valid attempts are reflected in audit log. - [ ] Successfully connect to node with correct role - [ ] Unsuccessfully connect to a node in a role restricting access by label - [ ] Unsuccessfully connect to a node in a role restricting access by invalid SSH login - [ ] Allow/deny role option: SSH agent forwarding - [ ] Allow/deny role option: Port forwarding - [ ] Verify that custom PAM environment variables are available as expected. - [ ] Users With every user combination, try to login and signup with invalid second factor, invalid password to see how the system reacts. - [ ] Adding Users Password Only - [ ] Adding Users OTP - [ ] Adding Users U2F - [ ] Adding Users WebAuthn - [ ] Managing MFA devices - [ ] Add an OTP device with `tsh mfa add` - [ ] Add a U2F device with `tsh mfa add` - [ ] Verify that the U2F device works under WebAuthn - [ ] Add a WebAuthn device with `tsh mfa add` - [ ] List MFA devices with `tsh mfa ls` - [ ] Remove an OTP device with `tsh mfa rm` - [ ] Remove a U2F device with `tsh mfa rm` - [ ] Remove a WebAuthn device with `tsh mfa rm` - [ ] Attempt removing the last MFA device on the user - [ ] with `second_factor: on` in `auth_service`, should fail - [ ] with `second_factor: optional` in `auth_service`, should succeed - [ ] Login Password Only - [ ] Login with MFA - [ ] Add 2 OTP and 2 WebAuthn devices with `tsh mfa add` - [ ] Login via OTP - [ ] Login via WebAuthn - [ ] Login OIDC - [ ] Login SAML - [ ] Login GitHub - [ ] Deleting Users - [ ] Backends - [ ] Teleport runs with etcd - [ ] Teleport runs with dynamodb - [ ] Teleport runs with SQLite - [ ] Teleport runs with Firestore - [ ] Session Recording - [ ] Session recording can be disabled - [ ] Sessions can be recorded at the node - [ ] Sessions in remote clusters are recorded in remote clusters - [ ] Sessions can be recorded at the proxy - [ ] Sessions on remote clusters are recorded in the local cluster - [ ] Enable/disable host key checking. - [ ] Audit Log - [ ] Failed login attempts are recorded - [ ] Interactive sessions have the correct Server ID - [ ] Server ID is the ID of the node in "session_recording: node" mode - [ ] Server ID is the ID of the proxy in "session_recording: proxy" mode Node/Proxy ID may be found at `/var/lib/teleport/host_uuid` in the corresponding machine. Node IDs may also be queried via `tctl nodes ls`. - [ ] Exec commands are recorded - [ ] `scp` commands are recorded - [ ] Subsystem results are recorded Subsystem testing may be achieved using both [Recording Proxy mode]( https://goteleport.com/teleport/docs/architecture/proxy/#recording-proxy-mode) and [OpenSSH integration]( https://goteleport.com/docs/server-access/guides/openssh/). Assuming the proxy is `proxy.example.com:3023` and `node1` is a node running OpenSSH/sshd, you may use the following command to trigger a subsystem audit log: ```shell sftp -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" root@node1 ``` - [ ] Interact with a cluster using `tsh` These commands should ideally be tested for recording and non-recording modes as they are implemented in a different ways. - [ ] tsh ssh \<regular-node\> - [ ] tsh ssh \<node-remote-cluster\> - [ ] tsh ssh -A \<regular-node\> - [ ] tsh ssh -A \<node-remote-cluster\> - [ ] tsh ssh \<regular-node\> ls - [ ] tsh ssh \<node-remote-cluster\> ls - [ ] tsh join \<regular-node\> - [ ] tsh join \<node-remote-cluster\> - [ ] tsh play \<regular-node\> - [ ] tsh play \<node-remote-cluster\> - [ ] tsh scp \<regular-node\> - [ ] tsh scp \<node-remote-cluster\> - [ ] tsh ssh -L \<regular-node\> - [ ] tsh ssh -L \<node-remote-cluster\> - [ ] tsh ls - [ ] tsh clusters - [ ] Interact with a cluster using `ssh` Make sure to test both recording and regular proxy modes. - [ ] ssh \<regular-node\> - [ ] ssh \<node-remote-cluster\> - [ ] ssh -A \<regular-node\> - [ ] ssh -A \<node-remote-cluster\> - [ ] ssh \<regular-node\> ls - [ ] ssh \<node-remote-cluster\> ls - [ ] scp \<regular-node\> - [ ] scp \<node-remote-cluster\> - [ ] ssh -L \<regular-node\> - [ ] ssh -L \<node-remote-cluster\> - [ ] Interact with a cluster using the Web UI - [ ] Connect to a Teleport node - [ ] Connect to a OpenSSH node - [ ] Check agent forwarding is correct based on role and proxy mode. ### User accounting - [ ] Verify that active interactive sessions are tracked in `/var/run/utmp` on Linux. - [ ] Verify that interactive sessions are logged in `/var/log/wtmp` on Linux. ### Combinations For some manual testing, many combinations need to be tested. For example, for interactive sessions the 12 combinations are below. - [ ] Connect to a OpenSSH node in a local cluster using OpenSSH. - [ ] Connect to a OpenSSH node in a local cluster using Teleport. - [ ] Connect to a OpenSSH node in a local cluster using the Web UI. - [ ] Connect to a Teleport node in a local cluster using OpenSSH. - [ ] Connect to a Teleport node in a local cluster using Teleport. - [ ] Connect to a Teleport node in a local cluster using the Web UI. - [ ] Connect to a OpenSSH node in a remote cluster using OpenSSH. - [ ] Connect to a OpenSSH node in a remote cluster using Teleport. - [ ] Connect to a OpenSSH node in a remote cluster using the Web UI. - [ ] Connect to a Teleport node in a remote cluster using OpenSSH. - [ ] Connect to a Teleport node in a remote cluster using Teleport. - [ ] Connect to a Teleport node in a remote cluster using the Web UI. ### Teleport with EKS/GKE * [ ] Deploy Teleport on a single EKS cluster * [ ] Deploy Teleport on two EKS clusters and connect them via trusted cluster feature * [ ] Deploy Teleport Proxy outside of GKE cluster fronting connections to it (use [this script](https://github.com/gravitational/teleport/blob/master/examples/k8s-auth/get-kubeconfig.sh) to generate a kubeconfig) * [ ] Deploy Teleport Proxy outside of EKS cluster fronting connections to it (use [this script](https://github.com/gravitational/teleport/blob/master/examples/k8s-auth/get-kubeconfig.sh) to generate a kubeconfig) ### Teleport with multiple Kubernetes clusters Note: you can use GKE or EKS or minikube to run Kubernetes clusters. Minikube is the only caveat - it's not reachable publicly so don't run a proxy there. * [ ] Deploy combo auth/proxy/kubernetes_service outside of a Kubernetes cluster, using a kubeconfig * [ ] Login with `tsh login`, check that `tsh kube ls` has your cluster * [ ] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [ ] Verify that the audit log recorded the above request and session * [ ] Deploy combo auth/proxy/kubernetes_service inside of a Kubernetes cluster * [ ] Login with `tsh login`, check that `tsh kube ls` has your cluster * [ ] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [ ] Verify that the audit log recorded the above request and session * [ ] Deploy combo auth/proxy_service outside of the Kubernetes cluster and kubernetes_service inside of a Kubernetes cluster, connected over a reverse tunnel * [ ] Login with `tsh login`, check that `tsh kube ls` has your cluster * [ ] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [ ] Verify that the audit log recorded the above request and session * [ ] Deploy a second kubernetes_service inside of another Kubernetes cluster, connected over a reverse tunnel * [ ] Login with `tsh login`, check that `tsh kube ls` has both clusters * [ ] Switch to a second cluster using `tsh kube login` * [ ] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` on the new cluster * [ ] Verify that the audit log recorded the above request and session * [ ] Deploy combo auth/proxy/kubernetes_service outside of a Kubernetes cluster, using a kubeconfig with multiple clusters in it * [ ] Login with `tsh login`, check that `tsh kube ls` has all clusters * [ ] Test Kubernetes screen in the web UI (tab is located on left side nav on dashboard): * [ ] Verify that all kubes registered are shown with correct `name` and `labels` * [ ] Verify that clicking on a rows connect button renders a dialogue on manual instructions with `Step 2` login value matching the rows `name` column * [ ] Verify searching for `name` or `labels` in the search bar works * [ ] Verify you can sort by `name` colum ### Teleport with FIPS mode * [ ] Perform trusted clusters, Web and SSH sanity check with all teleport components deployed in FIPS mode. ### ACME - [ ] Teleport can fetch TLS certificate automatically using ACME protocol. ### Migrations * [ ] Migrate trusted clusters from 2.4.0 to 2.5.0 * [ ] Migrate auth server on main cluster, then rest of the servers on main cluster SSH should work for both main and old clusters * [ ] Migrate auth server on remote cluster, then rest of the remote cluster SSH should work ### Command Templates When interacting with a cluster, the following command templates are useful: #### OpenSSH ``` # when connecting to the recording proxy, `-o 'ForwardAgent yes'` is required. ssh -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" \ node.example.com # the above command only forwards the agent to the proxy, to forward the agent # to the target node, `-o 'ForwardAgent yes'` needs to be passed twice. ssh -o "ForwardAgent yes" \ -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" \ node.example.com # when connecting to a remote cluster using OpenSSH, the subsystem request is # updated with the name of the remote cluster. ssh -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p@foo.com" \ node.foo.com ``` #### Teleport ``` # when connecting to a OpenSSH node, remember `-p 22` needs to be passed. tsh --proxy=proxy.example.com --user=<username> --insecure ssh -p 22 node.example.com # an agent can be forwarded to the target node with `-A` tsh --proxy=proxy.example.com --user=<username> --insecure ssh -A -p 22 node.example.com # the --cluster flag is used to connect to a node in a remote cluster. tsh --proxy=proxy.example.com --user=<username> --insecure ssh --cluster=foo.com -p 22 node.foo.com ``` ### Teleport with SSO Providers - [ ] G Suite install instructions work - [ ] G Suite Screenshots are up to date - [ ] ActiveDirectory install instructions work - [ ] Active Directory Screenshots are up to date - [ ] Okta install instructions work - [ ] Okta Screenshots are up to date - [ ] OneLogin install instructions work - [ ] OneLogin Screenshots are up to date - [ ] OIDC install instructions work - [ ] OIDC Screenshots are up to date ### Teleport Plugins - [ ] Test receiving a message via Teleport Slackbot - [ ] Test receiving a new Jira Ticket via Teleport Jira ## WEB UI ## Main For main, test with a role that has access to all resources. #### Top Nav - [ ] Verify that cluster selector displays all (root + leaf) clusters - [ ] Verify that user name is displayed - [ ] Verify that user menu shows logout, help&support, and account settings (for local users) #### Side Nav - [ ] Verify that each item has an icon - [ ] Verify that Collapse/Expand works and collapsed has icon `>`, and expand has icon `v` - [ ] Verify that it automatically expands and highlights the item on page refresh #### Servers aka Nodes - [ ] Verify that "Servers" table shows all joined nodes - [ ] Verify that "Connect" button shows a list of available logins - [ ] Verify that "Hostname", "Address" and "Labels" columns show the current values - [ ] Verify that "Search" by hostname, address, labels works - [ ] Verify that terminal opens when clicking on one of the available logins - [ ] Verify that clicking on `Add Server` button renders dialogue set to `Automatically` view - [ ] Verify clicking on `Regenerate Script` regenerates token value in the bash command - [ ] Verify using the bash command successfully adds the server (refresh server list) - [ ] Verify that clicking on `Manually` tab renders manual steps - [ ] Verify that clicking back to `Automatically` tab renders bash command #### Applications - [ ] Verify that clicking on `Add Application` button renders dialogue - [ ] Verify input validation (prevent empty value and invalid url) - [ ] Verify after input and clicking on `Generate Script`, bash command is rendered - [ ] Verify clicking on `Regenerate` button regenerates token value in bash command #### Databases - [ ] Verify that clicking on `Add Database` button renders dialogue for manual instructions: - [ ] Verify selecting different options on `Step 4` changes `Step 5` commands #### Active Sessions - [ ] Verify that "empty" state is handled - [ ] Verify that it displays the session when session is active - [ ] Verify that "Description", "Session ID", "Users", "Nodes" and "Duration" columns show correct values - [ ] Verify that "OPTIONS" button allows to join a session #### Audit log - [ ] Verify that time range button is shown and works - [ ] Verify that clicking on `Session Ended` event icon, takes user to session player - [ ] Verify event detail dialogue renders when clicking on events `details` button - [ ] Verify searching by type, description, created works #### Users - [ ] Verify that users are shown - [ ] Verify that creating a new user works - [ ] Verify that editing user roles works - [ ] Verify that removing a user works - [ ] Verify resetting a user's password works - [ ] Verify search by username, roles, and type works #### Auth Connectors - [ ] Verify when there are no connectors, empty state renders - [ ] Verify that creating OIDC/SAML/GITHUB connectors works - [ ] Verify that editing OIDC/SAML/GITHUB connectors works - [ ] Verify that error is shown when saving an invalid YAML - [ ] Verify that correct hint text is shown on the right side - [ ] Verify that encrypted SAML assertions work with an identity provider that supports it (Azure). - [ ] Verify that created github, saml, oidc card has their icons #### Roles - [ ] Verify that roles are shown - [ ] Verify that "Create New Role" dialog works - [ ] Verify that deleting and editing works - [ ] Verify that error is shown when saving an invalid YAML - [ ] Verify that correct hint text is shown on the right side #### Managed Clusters - [ ] Verify that it displays a list of clusters (root + leaf) - [ ] Verify that every menu item works: nodes, apps, audit events, session recordings, etc. #### Help & Support - [ ] Verify that all URLs work and correct (no 404) ## Access Requests ### Creating Access Requests 1. Create a role with limited permissions (defined below as `allow-roles`). This role allows you to see the Role screen and ssh into all nodes. 1. Create another role with limited permissions (defined below as `allow-users`). This role session expires in 4 minutes, allows you to see Users screen, and denies access to all nodes. 1. Create another role with no permissions other than being able to create requests (defined below as `default`) 1. Create a user with role `default` assigned 1. Create a few requests under this user to test pending/approved/denied state. ``` kind: role metadata: name: allow-roles spec: allow: logins: - root node_labels: '*': '*' rules: - resources: - role verbs: - list - read options: max_session_ttl: 8h0m0s version: v3 ``` ``` kind: role metadata: name: allow-users-short-ttl spec: allow: rules: - resources: - user verbs: - list - read deny: node_labels: '*': '*' options: max_session_ttl: 4m0s version: v3 ``` ``` kind: role metadata: name: default spec: allow: request: roles: - allow-roles - allow-users suggested_reviewers: - random-user-1 - random-user-2 options: max_session_ttl: 8h0m0s version: v3 ``` - [ ] Verify that under requestable roles, only `allow-roles` and `allow-users` are listed - [ ] Verify input validation requires at least one role to be selected - [ ] Verify you can select/input/modify reviewers - [ ] Verify after creating a request, requests are listed in pending states - [ ] Verify you can't review own requests ### Viewing & Approving/Denying Requests Create a user with the role `reviewer` that allows you to review all requests, and delete them. ``` kind: role version: v3 metadata: name: reviewer spec: allow: review_requests: roles: ['*'] ``` - [ ] Verify you can view access request from request list - [ ] Verify there is list of reviewers you selected (empty list if none selected AND suggested_reviewers wasn't defined) - [ ] Verify threshold name is there (it will be `default` if thresholds weren't defined in role, or blank if not named) - [ ] Verify you can approve a request with message, and immediately see updated state with your review stamp (green checkmark) and message box - [ ] Verify you can deny a request, and immediately see updated state with your review stamp (red cross) - [ ] Verify deleting the denied request is removed from list ### Assuming Approved Requests - [ ] Verify assume buttons are only present for approved request and for logged in user - [ ] Verify that assuming `allow-roles` allows you to see roles screen and ssh into nodes - [ ] Verify that after clicking on the assume button, it is disabled in both the list and in viewing - [ ] After assuming `allow-roles`, verify that assuming `allow-users-short-ttl` allows you to see users screen, and denies access to nodes - [ ] Verify a switchback banner is rendered with roles assumed, and count down of when it expires - [ ] Verify `switching back` goes back to your default static role - [ ] Verify after re-assuming `allow-users-short-ttl` role, the user is automatically logged out after the expiry is met (4 minutes) - [ ] Verify that after logging out (or getting logged out automatically) and relogging in, permissions are reset to `default`, and requests that are not expired and are approved are assumable again ## Access Request Waiting Room #### Strategy Reason Create the following role: ``` kind: role metadata: name: waiting-room spec: allow: request: roles: - <some other role to assign user after approval> options: max_session_ttl: 8h0m0s request_access: reason request_prompt: <some custom prompt to show in reason dialogue> version: v3 ``` - [ ] Verify after login, reason dialogue is rendered with prompt set to `request_prompt` setting - [ ] Verify after clicking `send request`, pending dialogue renders - [ ] Verify after approving a request, dashboard is rendered - [ ] Verify the correct role was assigned #### Strategy Always With the previous role you created from `Strategy Reason`, change `request_access` to `always`: - [ ] Verify after login, pending dialogue is auto rendered - [ ] Verify after approving a request, dashboard is rendered - [ ] Verify after denying a request, access denied dialogue is rendered - [ ] Verify a switchback banner is rendered with roles assumed, and count down of when it expires - [ ] Verify switchback button says `Logout` and clicking goes back to the login screen #### Strategy Optional With the previous role you created from `Strategy Reason`, change `request_access` to `optional`: - [ ] Verify after login, dashboard is rendered as normal ## Terminal - [ ] Verify that top nav has a user menu (Main and Logout) - [ ] Verify that switching between tabs works on alt+[1...9] #### Node List Tab - [ ] Verify that Cluster selector works (URL should change too) - [ ] Verify that Quick launcher input works - [ ] Verify that Quick launcher input handles input errors - [ ] Verify that "Connect" button shows a list of available logins - [ ] Verify that "Hostname", "Address" and "Labels" columns show the current values - [ ] Verify that "Search" by hostname, address, labels work - [ ] Verify that new tab is created when starting a session #### Session Tab - [ ] Verify that session and browser tabs both show the title with login and node name - [ ] Verify that terminal resize works - Install midnight commander on the node you ssh into: `$ sudo apt-get install mc` - Run the program: `$ mc` - Resize the terminal to see if panels resize with it - [ ] Verify that session tab shows/updates number of participants when a new user joins the session - [ ] Verify that tab automatically closes on "$ exit" command - [ ] Verify that SCP Upload works - [ ] Verify that SCP Upload handles invalid paths and network errors - [ ] Verify that SCP Download works - [ ] Verify that SCP Download handles invalid paths and network errors ## Session Player - [ ] Verify that it can replay a session - [ ] Verify that when playing, scroller auto scrolls to bottom most content - [ ] Verify when resizing player to a small screen, scroller appears and is working - [ ] Verify that error message is displayed (enter an invalid SID in the URL) ## Invite and Reset Form - [ ] Verify that input validates - [ ] Verify that invite works with 2FA disabled - [ ] Verify that invite works with OTP enabled - [ ] Verify that invite works with U2F enabled - [ ] Verify that invite works with WebAuthn enabled - [ ] Verify that error message is shown if an invite is expired/invalid ## Login Form and Change Password - [ ] Verify that input validates - [ ] Verify that login works with 2FA disabled - [ ] Verify that changing passwords works for 2FA disabled - [ ] Verify that login works with OTP enabled - [ ] Verify that changing passwords works for OTP enabled - [ ] Verify that login works with U2F enabled - [ ] Verify that changing passwords works for U2F enabled - [ ] Verify that login works with WebAuthn enabled - [ ] Verify that changing passwords works for WebAuthn enabled - [ ] Verify that login works for Github/SAML/OIDC - [ ] Verify that redirect to original URL works after successful login - [ ] Verify that account is locked after several unsuccessful login attempts - [ ] Verify that account is locked after several unsuccessful change password attempts ## Multi-factor Authentication (mfa) Create/modify `teleport.yaml` and set the following authentication settings under `auth_service` ```yaml authentication: type: local second_factor: optional require_session_mfa: yes webauthn: rp_id: example.com ``` #### MFA invite, login, password reset, change password - [ ] Verify during invite/reset, second factor list all auth types: none, hardware key, and authenticator app - [ ] Verify registration works with all option types - [ ] Verify login with all option types - [ ] Verify changing password with all option types - [ ] Change `second_factor` type to `on` and verify that mfa is required (no option `none` in dropdown) #### MFA require auth Go to `Account Settings` > `Two-Factor Devices` and register a new device Using the same user as above: - [ ] Verify logging in with registered WebAuthn key works - [ ] Verify connecting to a ssh node prompts you to tap your registered WebAuthn key - [ ] Verify in the web terminal, you can scp upload/download files #### MFA Management - [ ] Verify adding first device works without requiring re-authentication - [ ] Verify re-authenticating with a WebAuthn device works - [ ] Verify re-authenticating with a U2F device works - [ ] Verify re-authenticating with a OTP device works - [ ] Verify adding a WebAuthn device works - [ ] Verify adding a U2F device works - [ ] Verify adding an OTP device works - [ ] Verify removing a device works - [ ] Verify `second_factor` set to `off` disables adding devices ## Cloud From your cloud staging account, change the field `teleportVersion` to the test version. ``` $ kubectl -n <namespace> edit tenant ``` #### Recovery Code Management - [ ] Verify generating recovery codes for local accounts with email usernames works - [ ] Verify local accounts with non-email usernames are not able to generate recovery codes - [ ] Verify SSO accounts are not able to generate recovery codes #### Invite/Reset - [ ] Verify email as usernames, renders recovery codes dialog - [ ] Verify non email usernames, does not render recovery codes dialog #### Recovery Flow: Add new mfa device - [ ] Verify recovering (adding) a new hardware key device with password - [ ] Verify recovering (adding) a new otp device with password - [ ] Verify viewing and deleting any old device (but not the one just added) - [ ] Verify new recovery codes are rendered at the end of flow #### Recovery Flow: Change password - [ ] Verify recovering password with any mfa device - [ ] Verify new recovery codes are rendered at the end of flow #### Recovery Email - [ ] Verify receiving email for link to start recovery - [ ] Verify receiving email for successfully recovering - [ ] Verify email link is invalid after successful recovery - [ ] Verify receiving email for locked account when max attempts reached ## RBAC Create a role, with no `allow.rules` defined: ``` kind: role metadata: name: rbac spec: allow: app_labels: '*': '*' logins: - root node_labels: '*': '*' options: max_session_ttl: 8h0m0s version: v3 ``` - [ ] Verify that a user has access only to: "Servers", "Applications", "Databases", "Kubernetes", "Active Sessions", "Access Requests" and "Manage Clusters" - [ ] Verify there is no `Add Server, Application, Databases, Kubernetes` button in each respective view - [ ] Verify only `Servers`, `Apps`, `Databases`, and `Kubernetes` are listed under `options` button in `Manage Clusters` Note: User has read/create access_request access to their own requests, despite resource settings Add the following under `spec.allow.rules` to enable read access to the audit log: ``` - resources: - event verbs: - list ``` - [ ] Verify that the `Audit Log` and `Session Recordings` is accessible - [ ] Verify that playing a recorded session is denied Add the following to enable read access to recorded sessions ``` - resources: - session verbs: - read ``` - [ ] Verify that a user can re-play a session (session.end) Add the following to enable read access to the roles ``` - resources: - role verbs: - list - read ``` - [ ] Verify that a user can see the roles - [ ] Verify that a user cannot create/delete/update a role Add the following to enable read access to the auth connectors ``` - resources: - auth_connector verbs: - list - read ``` - [ ] Verify that a user can see the list of auth connectors. - [ ] Verify that a user cannot create/delete/update the connectors Add the following to enable read access to users ``` - resources: - user verbs: - list - read ``` - [ ] Verify that a user can access the "Users" screen - [ ] Verify that a user cannot reset password and create/delete/update a user Add the following to enable read access to trusted clusters ``` - resources: - trusted_cluster verbs: - list - read ``` - [ ] Verify that a user can access the "Trust" screen - [ ] Verify that a user cannot create/delete/update a trusted cluster. ## Performance/Soak Test Using `tsh bench` tool, perform the soak tests and benchmark tests on the following configurations: * Cluster with 10K nodes in normal (non-IOT) node mode with ETCD * Cluster with 10K nodes in normal (non-IOT) mode with DynamoDB * Cluster with 1K IOT nodes with ETCD * Cluster with 1K IOT nodes with DynamoDB * Cluster with 500 trusted clusters with ETCD * Cluster with 500 trusted clusters with DynamoDB **Soak Tests** Run 4hour soak test with a mix of interactive/non-interactive sessions: ``` tsh bench --duration=4h user@teleport-monster-6757d7b487-x226b ls tsh bench -i --duration=4h user@teleport-monster-6757d7b487-x226b ps uax ``` Observe prometheus metrics for goroutines, open files, RAM, CPU, Timers and make sure there are no leaks - [ ] Verify that prometheus metrics are accurate. **Breaking load tests** Load system with tsh bench to the capacity and publish maximum numbers of concurrent sessions with interactive and non interactive tsh bench loads. ## Teleport with Cloud Providers ### AWS - [ ] Deploy Teleport to AWS. Using DynamoDB & S3 - [ ] Deploy Teleport Enterprise to AWS. Using HA Setup https://gravitational.com/teleport/docs/aws-terraform-guide/ ### GCP - [ ] Deploy Teleport to GCP. Using Cloud Firestore & Cloud Storage - [ ] Deploy Teleport to GKE. Google Kubernetes engine. - [ ] Deploy Teleport Enterprise to GCP. ### IBM - [ ] Deploy Teleport to IBM Cloud. Using IBM Database for etcd & IBM Object Store - [ ] Deploy Teleport to IBM Cloud Kubernetes. - [ ] Deploy Teleport Enterprise to IBM Cloud. ## Application Access - [ ] Run an application within local cluster. - [ ] Verify the debug application `debug_app: true` works. - [ ] Verify an application can be configured with command line flags. - [ ] Verify an application can be configured from file configuration. - [ ] Verify that applications are available at auto-generated addresses `name.rootProxyPublicAddr` and well as `publicAddr`. - [ ] Run an application within a trusted cluster. - [ ] Verify that applications are available at auto-generated addresses `name.rootProxyPublicAddr`. - [ ] Verify Audit Records. - [ ] `app.session.start` and `app.session.chunk` events are created in the Audit Log. - [ ] `app.session.chunk` points to a 5 minute session archive with multiple `app.session.request` events inside. - [ ] `tsh play <chunk-id>` can fetch and print a session chunk archive. - [ ] Verify JWT using [verify-jwt.go](https://github.com/gravitational/teleport/blob/master/examples/jwt/verify-jwt.go). - [ ] Verify RBAC. - [ ] Verify [CLI access](https://goteleport.com/docs/application-access/guides/api-access/) with `tsh app login`. - [ ] Verify AWS console access. - [ ] Can log into AWS web console through the web UI. - [ ] Can interact with AWS using `tsh aws` commands. - [ ] Verify dynamic registration. - [ ] Can register a new app using `tctl create`. - [ ] Can update registered app using `tctl create -f`. - [ ] Can delete registered app using `tctl rm`. - [ ] Test Applications screen in the web UI (tab is located on left side nav on dashboard): - [ ] Verify that all apps registered are shown - [ ] Verify that clicking on the app icon takes you to another tab - [ ] Verify using the bash command produced from `Add Application` dialogue works (refresh app screen to see it registered) ## Database Access - [ ] Connect to a database within a local cluster. - [ ] Self-hosted Postgres. - [ ] Self-hosted MySQL. - [ ] Self-hosted MariaDB. - [ ] Self-hosted MongoDB. - [ ] Self-hosted CockroachDB. - [ ] AWS Aurora Postgres. - [ ] AWS Aurora MySQL. - [ ] AWS Redshift. - [ ] GCP Cloud SQL Postgres. - [ ] GCP Cloud SQL MySQL. - [ ] Connect to a database within a remote cluster via a trusted cluster. - [ ] Self-hosted Postgres. - [ ] Self-hosted MySQL. - [ ] Self-hosted MariaDB. - [ ] Self-hosted MongoDB. - [ ] Self-hosted CockroachDB. - [ ] AWS Aurora Postgres. - [ ] AWS Aurora MySQL. - [ ] AWS Redshift. - [ ] GCP Cloud SQL Postgres. - [ ] GCP Cloud SQL MySQL. - [ ] Verify audit events. - [ ] `db.session.start` is emitted when you connect. - [ ] `db.session.end` is emitted when you disconnect. - [ ] `db.session.query` is emitted when you execute a SQL query. - [ ] Verify RBAC. - [ ] `tsh db ls` shows only databases matching role's `db_labels`. - [ ] Can only connect as users from `db_users`. - [ ] _(Postgres only)_ Can only connect to databases from `db_names`. - [ ] `db.session.start` is emitted when connection attempt is denied. - [ ] _(MongoDB only)_ Can only execute commands in databases from `db_names`. - [ ] `db.session.query` is emitted when command fails due to permissions. - [ ] Can configure per-session MFA. - [ ] MFA tap is required on each `tsh db connect`. - [ ] Verify dynamic registration. - [ ] Can register a new database using `tctl create`. - [ ] Can update registered database using `tctl create -f`. - [ ] Can delete registered database using `tctl rm`. - [ ] Verify discovery. - [ ] Can detect and register RDS instances. - [ ] Can detect and register Aurora clusters, and their reader and custom endpoints. - [ ] Can detect and register Redshift clusters. - [ ] Test Databases screen in the web UI (tab is located on left side nav on dashboard): - [ ] Verify that all dbs registered are shown with correct `name`, `description`, `type`, and `labels` - [ ] Verify that clicking on a rows connect button renders a dialogue on manual instructions with `Step 2` login value matching the rows `name` column - [ ] Verify searching for all columns in the search bar works - [ ] Verify you can sort by all columns except `labels` ## TLS Routing - [ ] Verify that teleport proxy `v2` configuration starts only a single listener. ``` version: v2 teleport: proxy_service: enabled: "yes" public_addr: ['root.example.com'] web_listen_addr: 0.0.0.0:3080 ``` - [ ] Run Teleport Proxy in `multiplex` mode `auth_service.proxy_listener_mode: "multiplex"` - [ ] Trusted cluster - [ ] Setup trusted clusters using single port setup `web_proxy_addr == tunnel_addr` ``` kind: trusted_cluster spec: ... web_proxy_addr: root.example.com:443 tunnel_addr: root.example.com:443 ... ``` - [ ] Database Access - [ ] Verify that `tsh db connect` works through proxy running in `multiplex` mode - [ ] Postgres - [ ] MySQL - [ ] MariaDB - [ ] MongoDB - [ ] CockroachDB - [ ] Verify connecting to a database through TLS ALPN SNI local proxy `tsh db proxy` with a GUI client. - [ ] Application Access - [ ] Verify app access through proxy running in `multiplex` mode - [ ] SSH Access - [ ] Connect to a OpenSSH server through a local ssh proxy `ssh -o "ForwardAgent yes" -o "ProxyCommand tsh proxy ssh" user@host.example.com` - [ ] Connect to a OpenSSH server on leaf-cluster through a local ssh proxy`ssh -o "ForwardAgent yes" -o "ProxyCommand tsh proxy ssh --user=%r --cluster=leaf-cluster %h:%p" user@node.foo.com` - [ ] Verify `tsh ssh` access through proxy running in multiplex mode - [ ] Kubernetes access: - [ ] Verify kubernetes access through proxy running in `multiplex` mode ## Desktop Access - [ ] Can connect to desktop defined in static `hosts` section. - [ ] Can connect to desktop discovered via LDAP - [ ] Download [Keyboard Key Info](https://dennisbabkin.com/kbdkeyinfo/) and verify all keys are processed correctly in each supported browser. Known issues: F11 cannot be captured by the browser without [special configuration](https://social.technet.microsoft.com/Forums/en-US/784b2bbe-353f-412e-ac9a-193d81f306b6/remote-desktop-for-mac-f11-key-not-working-on-macbook-pro-touchbar?forum=winRDc) on MacOS. - [ ] Left click and right click register as Windows clicks. (Right click on the desktop should show a Windows menu, not a browser context menu) - [ ] Vertical and horizontal scroll work. [Horizontal Scroll Test](https://codepen.io/jaemskyle/pen/inbmB) - [ ] All desktops have `teleport.dev/origin` label. - [ ] Dynamic desktops have additional `teleport.dev` labels for OS, OS Version, DNS hostname. - [ ] Verify that placing a user lock terminates an active desktop session. - [ ] Verify desktop session start/end audit events. - [ ] Regexp-based host labeling applies across all desktops, regardless of origin. - [ ] RBAC denies access to a Windows desktop due to labels - [ ] RBAC denies access to a Windows desktop with the wrong OS-login. - [ ] Multiple sessions as different users on the same desktop are allowed. - [ ] Connect multiple `windows_desktop_service`s to the same Teleport cluster, verify that connections to desktops on different AD domains works.
1.0
Teleport 9.0 Test Plan - ## Manual Testing Plan Below are the items that should be manually tested with each release of Teleport. These tests should be run on both a fresh install of the version to be released as well as an upgrade of the previous version of Teleport. - [ ] Adding nodes to a cluster - [ ] Adding Nodes via Valid Static Token - [ ] Adding Nodes via Valid Short-lived Tokens - [ ] Adding Nodes via Invalid Token Fails - [ ] Revoking Node Invitation - [ ] Labels - [ ] Static Labels - [ ] Dynamic Labels - [ ] Trusted Clusters - [ ] Adding Trusted Cluster Valid Static Token - [ ] Adding Trusted Cluster Valid Short-lived Token - [ ] Adding Trusted Cluster Invalid Token - [ ] Removing Trusted Cluster - [ ] RBAC Make sure that invalid and valid attempts are reflected in audit log. - [ ] Successfully connect to node with correct role - [ ] Unsuccessfully connect to a node in a role restricting access by label - [ ] Unsuccessfully connect to a node in a role restricting access by invalid SSH login - [ ] Allow/deny role option: SSH agent forwarding - [ ] Allow/deny role option: Port forwarding - [ ] Verify that custom PAM environment variables are available as expected. - [ ] Users With every user combination, try to login and signup with invalid second factor, invalid password to see how the system reacts. - [ ] Adding Users Password Only - [ ] Adding Users OTP - [ ] Adding Users U2F - [ ] Adding Users WebAuthn - [ ] Managing MFA devices - [ ] Add an OTP device with `tsh mfa add` - [ ] Add a U2F device with `tsh mfa add` - [ ] Verify that the U2F device works under WebAuthn - [ ] Add a WebAuthn device with `tsh mfa add` - [ ] List MFA devices with `tsh mfa ls` - [ ] Remove an OTP device with `tsh mfa rm` - [ ] Remove a U2F device with `tsh mfa rm` - [ ] Remove a WebAuthn device with `tsh mfa rm` - [ ] Attempt removing the last MFA device on the user - [ ] with `second_factor: on` in `auth_service`, should fail - [ ] with `second_factor: optional` in `auth_service`, should succeed - [ ] Login Password Only - [ ] Login with MFA - [ ] Add 2 OTP and 2 WebAuthn devices with `tsh mfa add` - [ ] Login via OTP - [ ] Login via WebAuthn - [ ] Login OIDC - [ ] Login SAML - [ ] Login GitHub - [ ] Deleting Users - [ ] Backends - [ ] Teleport runs with etcd - [ ] Teleport runs with dynamodb - [ ] Teleport runs with SQLite - [ ] Teleport runs with Firestore - [ ] Session Recording - [ ] Session recording can be disabled - [ ] Sessions can be recorded at the node - [ ] Sessions in remote clusters are recorded in remote clusters - [ ] Sessions can be recorded at the proxy - [ ] Sessions on remote clusters are recorded in the local cluster - [ ] Enable/disable host key checking. - [ ] Audit Log - [ ] Failed login attempts are recorded - [ ] Interactive sessions have the correct Server ID - [ ] Server ID is the ID of the node in "session_recording: node" mode - [ ] Server ID is the ID of the proxy in "session_recording: proxy" mode Node/Proxy ID may be found at `/var/lib/teleport/host_uuid` in the corresponding machine. Node IDs may also be queried via `tctl nodes ls`. - [ ] Exec commands are recorded - [ ] `scp` commands are recorded - [ ] Subsystem results are recorded Subsystem testing may be achieved using both [Recording Proxy mode]( https://goteleport.com/teleport/docs/architecture/proxy/#recording-proxy-mode) and [OpenSSH integration]( https://goteleport.com/docs/server-access/guides/openssh/). Assuming the proxy is `proxy.example.com:3023` and `node1` is a node running OpenSSH/sshd, you may use the following command to trigger a subsystem audit log: ```shell sftp -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" root@node1 ``` - [ ] Interact with a cluster using `tsh` These commands should ideally be tested for recording and non-recording modes as they are implemented in a different ways. - [ ] tsh ssh \<regular-node\> - [ ] tsh ssh \<node-remote-cluster\> - [ ] tsh ssh -A \<regular-node\> - [ ] tsh ssh -A \<node-remote-cluster\> - [ ] tsh ssh \<regular-node\> ls - [ ] tsh ssh \<node-remote-cluster\> ls - [ ] tsh join \<regular-node\> - [ ] tsh join \<node-remote-cluster\> - [ ] tsh play \<regular-node\> - [ ] tsh play \<node-remote-cluster\> - [ ] tsh scp \<regular-node\> - [ ] tsh scp \<node-remote-cluster\> - [ ] tsh ssh -L \<regular-node\> - [ ] tsh ssh -L \<node-remote-cluster\> - [ ] tsh ls - [ ] tsh clusters - [ ] Interact with a cluster using `ssh` Make sure to test both recording and regular proxy modes. - [ ] ssh \<regular-node\> - [ ] ssh \<node-remote-cluster\> - [ ] ssh -A \<regular-node\> - [ ] ssh -A \<node-remote-cluster\> - [ ] ssh \<regular-node\> ls - [ ] ssh \<node-remote-cluster\> ls - [ ] scp \<regular-node\> - [ ] scp \<node-remote-cluster\> - [ ] ssh -L \<regular-node\> - [ ] ssh -L \<node-remote-cluster\> - [ ] Interact with a cluster using the Web UI - [ ] Connect to a Teleport node - [ ] Connect to a OpenSSH node - [ ] Check agent forwarding is correct based on role and proxy mode. ### User accounting - [ ] Verify that active interactive sessions are tracked in `/var/run/utmp` on Linux. - [ ] Verify that interactive sessions are logged in `/var/log/wtmp` on Linux. ### Combinations For some manual testing, many combinations need to be tested. For example, for interactive sessions the 12 combinations are below. - [ ] Connect to a OpenSSH node in a local cluster using OpenSSH. - [ ] Connect to a OpenSSH node in a local cluster using Teleport. - [ ] Connect to a OpenSSH node in a local cluster using the Web UI. - [ ] Connect to a Teleport node in a local cluster using OpenSSH. - [ ] Connect to a Teleport node in a local cluster using Teleport. - [ ] Connect to a Teleport node in a local cluster using the Web UI. - [ ] Connect to a OpenSSH node in a remote cluster using OpenSSH. - [ ] Connect to a OpenSSH node in a remote cluster using Teleport. - [ ] Connect to a OpenSSH node in a remote cluster using the Web UI. - [ ] Connect to a Teleport node in a remote cluster using OpenSSH. - [ ] Connect to a Teleport node in a remote cluster using Teleport. - [ ] Connect to a Teleport node in a remote cluster using the Web UI. ### Teleport with EKS/GKE * [ ] Deploy Teleport on a single EKS cluster * [ ] Deploy Teleport on two EKS clusters and connect them via trusted cluster feature * [ ] Deploy Teleport Proxy outside of GKE cluster fronting connections to it (use [this script](https://github.com/gravitational/teleport/blob/master/examples/k8s-auth/get-kubeconfig.sh) to generate a kubeconfig) * [ ] Deploy Teleport Proxy outside of EKS cluster fronting connections to it (use [this script](https://github.com/gravitational/teleport/blob/master/examples/k8s-auth/get-kubeconfig.sh) to generate a kubeconfig) ### Teleport with multiple Kubernetes clusters Note: you can use GKE or EKS or minikube to run Kubernetes clusters. Minikube is the only caveat - it's not reachable publicly so don't run a proxy there. * [ ] Deploy combo auth/proxy/kubernetes_service outside of a Kubernetes cluster, using a kubeconfig * [ ] Login with `tsh login`, check that `tsh kube ls` has your cluster * [ ] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [ ] Verify that the audit log recorded the above request and session * [ ] Deploy combo auth/proxy/kubernetes_service inside of a Kubernetes cluster * [ ] Login with `tsh login`, check that `tsh kube ls` has your cluster * [ ] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [ ] Verify that the audit log recorded the above request and session * [ ] Deploy combo auth/proxy_service outside of the Kubernetes cluster and kubernetes_service inside of a Kubernetes cluster, connected over a reverse tunnel * [ ] Login with `tsh login`, check that `tsh kube ls` has your cluster * [ ] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [ ] Verify that the audit log recorded the above request and session * [ ] Deploy a second kubernetes_service inside of another Kubernetes cluster, connected over a reverse tunnel * [ ] Login with `tsh login`, check that `tsh kube ls` has both clusters * [ ] Switch to a second cluster using `tsh kube login` * [ ] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` on the new cluster * [ ] Verify that the audit log recorded the above request and session * [ ] Deploy combo auth/proxy/kubernetes_service outside of a Kubernetes cluster, using a kubeconfig with multiple clusters in it * [ ] Login with `tsh login`, check that `tsh kube ls` has all clusters * [ ] Test Kubernetes screen in the web UI (tab is located on left side nav on dashboard): * [ ] Verify that all kubes registered are shown with correct `name` and `labels` * [ ] Verify that clicking on a rows connect button renders a dialogue on manual instructions with `Step 2` login value matching the rows `name` column * [ ] Verify searching for `name` or `labels` in the search bar works * [ ] Verify you can sort by `name` colum ### Teleport with FIPS mode * [ ] Perform trusted clusters, Web and SSH sanity check with all teleport components deployed in FIPS mode. ### ACME - [ ] Teleport can fetch TLS certificate automatically using ACME protocol. ### Migrations * [ ] Migrate trusted clusters from 2.4.0 to 2.5.0 * [ ] Migrate auth server on main cluster, then rest of the servers on main cluster SSH should work for both main and old clusters * [ ] Migrate auth server on remote cluster, then rest of the remote cluster SSH should work ### Command Templates When interacting with a cluster, the following command templates are useful: #### OpenSSH ``` # when connecting to the recording proxy, `-o 'ForwardAgent yes'` is required. ssh -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" \ node.example.com # the above command only forwards the agent to the proxy, to forward the agent # to the target node, `-o 'ForwardAgent yes'` needs to be passed twice. ssh -o "ForwardAgent yes" \ -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" \ node.example.com # when connecting to a remote cluster using OpenSSH, the subsystem request is # updated with the name of the remote cluster. ssh -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p@foo.com" \ node.foo.com ``` #### Teleport ``` # when connecting to a OpenSSH node, remember `-p 22` needs to be passed. tsh --proxy=proxy.example.com --user=<username> --insecure ssh -p 22 node.example.com # an agent can be forwarded to the target node with `-A` tsh --proxy=proxy.example.com --user=<username> --insecure ssh -A -p 22 node.example.com # the --cluster flag is used to connect to a node in a remote cluster. tsh --proxy=proxy.example.com --user=<username> --insecure ssh --cluster=foo.com -p 22 node.foo.com ``` ### Teleport with SSO Providers - [ ] G Suite install instructions work - [ ] G Suite Screenshots are up to date - [ ] ActiveDirectory install instructions work - [ ] Active Directory Screenshots are up to date - [ ] Okta install instructions work - [ ] Okta Screenshots are up to date - [ ] OneLogin install instructions work - [ ] OneLogin Screenshots are up to date - [ ] OIDC install instructions work - [ ] OIDC Screenshots are up to date ### Teleport Plugins - [ ] Test receiving a message via Teleport Slackbot - [ ] Test receiving a new Jira Ticket via Teleport Jira ## WEB UI ## Main For main, test with a role that has access to all resources. #### Top Nav - [ ] Verify that cluster selector displays all (root + leaf) clusters - [ ] Verify that user name is displayed - [ ] Verify that user menu shows logout, help&support, and account settings (for local users) #### Side Nav - [ ] Verify that each item has an icon - [ ] Verify that Collapse/Expand works and collapsed has icon `>`, and expand has icon `v` - [ ] Verify that it automatically expands and highlights the item on page refresh #### Servers aka Nodes - [ ] Verify that "Servers" table shows all joined nodes - [ ] Verify that "Connect" button shows a list of available logins - [ ] Verify that "Hostname", "Address" and "Labels" columns show the current values - [ ] Verify that "Search" by hostname, address, labels works - [ ] Verify that terminal opens when clicking on one of the available logins - [ ] Verify that clicking on `Add Server` button renders dialogue set to `Automatically` view - [ ] Verify clicking on `Regenerate Script` regenerates token value in the bash command - [ ] Verify using the bash command successfully adds the server (refresh server list) - [ ] Verify that clicking on `Manually` tab renders manual steps - [ ] Verify that clicking back to `Automatically` tab renders bash command #### Applications - [ ] Verify that clicking on `Add Application` button renders dialogue - [ ] Verify input validation (prevent empty value and invalid url) - [ ] Verify after input and clicking on `Generate Script`, bash command is rendered - [ ] Verify clicking on `Regenerate` button regenerates token value in bash command #### Databases - [ ] Verify that clicking on `Add Database` button renders dialogue for manual instructions: - [ ] Verify selecting different options on `Step 4` changes `Step 5` commands #### Active Sessions - [ ] Verify that "empty" state is handled - [ ] Verify that it displays the session when session is active - [ ] Verify that "Description", "Session ID", "Users", "Nodes" and "Duration" columns show correct values - [ ] Verify that "OPTIONS" button allows to join a session #### Audit log - [ ] Verify that time range button is shown and works - [ ] Verify that clicking on `Session Ended` event icon, takes user to session player - [ ] Verify event detail dialogue renders when clicking on events `details` button - [ ] Verify searching by type, description, created works #### Users - [ ] Verify that users are shown - [ ] Verify that creating a new user works - [ ] Verify that editing user roles works - [ ] Verify that removing a user works - [ ] Verify resetting a user's password works - [ ] Verify search by username, roles, and type works #### Auth Connectors - [ ] Verify when there are no connectors, empty state renders - [ ] Verify that creating OIDC/SAML/GITHUB connectors works - [ ] Verify that editing OIDC/SAML/GITHUB connectors works - [ ] Verify that error is shown when saving an invalid YAML - [ ] Verify that correct hint text is shown on the right side - [ ] Verify that encrypted SAML assertions work with an identity provider that supports it (Azure). - [ ] Verify that created github, saml, oidc card has their icons #### Roles - [ ] Verify that roles are shown - [ ] Verify that "Create New Role" dialog works - [ ] Verify that deleting and editing works - [ ] Verify that error is shown when saving an invalid YAML - [ ] Verify that correct hint text is shown on the right side #### Managed Clusters - [ ] Verify that it displays a list of clusters (root + leaf) - [ ] Verify that every menu item works: nodes, apps, audit events, session recordings, etc. #### Help & Support - [ ] Verify that all URLs work and correct (no 404) ## Access Requests ### Creating Access Requests 1. Create a role with limited permissions (defined below as `allow-roles`). This role allows you to see the Role screen and ssh into all nodes. 1. Create another role with limited permissions (defined below as `allow-users`). This role session expires in 4 minutes, allows you to see Users screen, and denies access to all nodes. 1. Create another role with no permissions other than being able to create requests (defined below as `default`) 1. Create a user with role `default` assigned 1. Create a few requests under this user to test pending/approved/denied state. ``` kind: role metadata: name: allow-roles spec: allow: logins: - root node_labels: '*': '*' rules: - resources: - role verbs: - list - read options: max_session_ttl: 8h0m0s version: v3 ``` ``` kind: role metadata: name: allow-users-short-ttl spec: allow: rules: - resources: - user verbs: - list - read deny: node_labels: '*': '*' options: max_session_ttl: 4m0s version: v3 ``` ``` kind: role metadata: name: default spec: allow: request: roles: - allow-roles - allow-users suggested_reviewers: - random-user-1 - random-user-2 options: max_session_ttl: 8h0m0s version: v3 ``` - [ ] Verify that under requestable roles, only `allow-roles` and `allow-users` are listed - [ ] Verify input validation requires at least one role to be selected - [ ] Verify you can select/input/modify reviewers - [ ] Verify after creating a request, requests are listed in pending states - [ ] Verify you can't review own requests ### Viewing & Approving/Denying Requests Create a user with the role `reviewer` that allows you to review all requests, and delete them. ``` kind: role version: v3 metadata: name: reviewer spec: allow: review_requests: roles: ['*'] ``` - [ ] Verify you can view access request from request list - [ ] Verify there is list of reviewers you selected (empty list if none selected AND suggested_reviewers wasn't defined) - [ ] Verify threshold name is there (it will be `default` if thresholds weren't defined in role, or blank if not named) - [ ] Verify you can approve a request with message, and immediately see updated state with your review stamp (green checkmark) and message box - [ ] Verify you can deny a request, and immediately see updated state with your review stamp (red cross) - [ ] Verify deleting the denied request is removed from list ### Assuming Approved Requests - [ ] Verify assume buttons are only present for approved request and for logged in user - [ ] Verify that assuming `allow-roles` allows you to see roles screen and ssh into nodes - [ ] Verify that after clicking on the assume button, it is disabled in both the list and in viewing - [ ] After assuming `allow-roles`, verify that assuming `allow-users-short-ttl` allows you to see users screen, and denies access to nodes - [ ] Verify a switchback banner is rendered with roles assumed, and count down of when it expires - [ ] Verify `switching back` goes back to your default static role - [ ] Verify after re-assuming `allow-users-short-ttl` role, the user is automatically logged out after the expiry is met (4 minutes) - [ ] Verify that after logging out (or getting logged out automatically) and relogging in, permissions are reset to `default`, and requests that are not expired and are approved are assumable again ## Access Request Waiting Room #### Strategy Reason Create the following role: ``` kind: role metadata: name: waiting-room spec: allow: request: roles: - <some other role to assign user after approval> options: max_session_ttl: 8h0m0s request_access: reason request_prompt: <some custom prompt to show in reason dialogue> version: v3 ``` - [ ] Verify after login, reason dialogue is rendered with prompt set to `request_prompt` setting - [ ] Verify after clicking `send request`, pending dialogue renders - [ ] Verify after approving a request, dashboard is rendered - [ ] Verify the correct role was assigned #### Strategy Always With the previous role you created from `Strategy Reason`, change `request_access` to `always`: - [ ] Verify after login, pending dialogue is auto rendered - [ ] Verify after approving a request, dashboard is rendered - [ ] Verify after denying a request, access denied dialogue is rendered - [ ] Verify a switchback banner is rendered with roles assumed, and count down of when it expires - [ ] Verify switchback button says `Logout` and clicking goes back to the login screen #### Strategy Optional With the previous role you created from `Strategy Reason`, change `request_access` to `optional`: - [ ] Verify after login, dashboard is rendered as normal ## Terminal - [ ] Verify that top nav has a user menu (Main and Logout) - [ ] Verify that switching between tabs works on alt+[1...9] #### Node List Tab - [ ] Verify that Cluster selector works (URL should change too) - [ ] Verify that Quick launcher input works - [ ] Verify that Quick launcher input handles input errors - [ ] Verify that "Connect" button shows a list of available logins - [ ] Verify that "Hostname", "Address" and "Labels" columns show the current values - [ ] Verify that "Search" by hostname, address, labels work - [ ] Verify that new tab is created when starting a session #### Session Tab - [ ] Verify that session and browser tabs both show the title with login and node name - [ ] Verify that terminal resize works - Install midnight commander on the node you ssh into: `$ sudo apt-get install mc` - Run the program: `$ mc` - Resize the terminal to see if panels resize with it - [ ] Verify that session tab shows/updates number of participants when a new user joins the session - [ ] Verify that tab automatically closes on "$ exit" command - [ ] Verify that SCP Upload works - [ ] Verify that SCP Upload handles invalid paths and network errors - [ ] Verify that SCP Download works - [ ] Verify that SCP Download handles invalid paths and network errors ## Session Player - [ ] Verify that it can replay a session - [ ] Verify that when playing, scroller auto scrolls to bottom most content - [ ] Verify when resizing player to a small screen, scroller appears and is working - [ ] Verify that error message is displayed (enter an invalid SID in the URL) ## Invite and Reset Form - [ ] Verify that input validates - [ ] Verify that invite works with 2FA disabled - [ ] Verify that invite works with OTP enabled - [ ] Verify that invite works with U2F enabled - [ ] Verify that invite works with WebAuthn enabled - [ ] Verify that error message is shown if an invite is expired/invalid ## Login Form and Change Password - [ ] Verify that input validates - [ ] Verify that login works with 2FA disabled - [ ] Verify that changing passwords works for 2FA disabled - [ ] Verify that login works with OTP enabled - [ ] Verify that changing passwords works for OTP enabled - [ ] Verify that login works with U2F enabled - [ ] Verify that changing passwords works for U2F enabled - [ ] Verify that login works with WebAuthn enabled - [ ] Verify that changing passwords works for WebAuthn enabled - [ ] Verify that login works for Github/SAML/OIDC - [ ] Verify that redirect to original URL works after successful login - [ ] Verify that account is locked after several unsuccessful login attempts - [ ] Verify that account is locked after several unsuccessful change password attempts ## Multi-factor Authentication (mfa) Create/modify `teleport.yaml` and set the following authentication settings under `auth_service` ```yaml authentication: type: local second_factor: optional require_session_mfa: yes webauthn: rp_id: example.com ``` #### MFA invite, login, password reset, change password - [ ] Verify during invite/reset, second factor list all auth types: none, hardware key, and authenticator app - [ ] Verify registration works with all option types - [ ] Verify login with all option types - [ ] Verify changing password with all option types - [ ] Change `second_factor` type to `on` and verify that mfa is required (no option `none` in dropdown) #### MFA require auth Go to `Account Settings` > `Two-Factor Devices` and register a new device Using the same user as above: - [ ] Verify logging in with registered WebAuthn key works - [ ] Verify connecting to a ssh node prompts you to tap your registered WebAuthn key - [ ] Verify in the web terminal, you can scp upload/download files #### MFA Management - [ ] Verify adding first device works without requiring re-authentication - [ ] Verify re-authenticating with a WebAuthn device works - [ ] Verify re-authenticating with a U2F device works - [ ] Verify re-authenticating with a OTP device works - [ ] Verify adding a WebAuthn device works - [ ] Verify adding a U2F device works - [ ] Verify adding an OTP device works - [ ] Verify removing a device works - [ ] Verify `second_factor` set to `off` disables adding devices ## Cloud From your cloud staging account, change the field `teleportVersion` to the test version. ``` $ kubectl -n <namespace> edit tenant ``` #### Recovery Code Management - [ ] Verify generating recovery codes for local accounts with email usernames works - [ ] Verify local accounts with non-email usernames are not able to generate recovery codes - [ ] Verify SSO accounts are not able to generate recovery codes #### Invite/Reset - [ ] Verify email as usernames, renders recovery codes dialog - [ ] Verify non email usernames, does not render recovery codes dialog #### Recovery Flow: Add new mfa device - [ ] Verify recovering (adding) a new hardware key device with password - [ ] Verify recovering (adding) a new otp device with password - [ ] Verify viewing and deleting any old device (but not the one just added) - [ ] Verify new recovery codes are rendered at the end of flow #### Recovery Flow: Change password - [ ] Verify recovering password with any mfa device - [ ] Verify new recovery codes are rendered at the end of flow #### Recovery Email - [ ] Verify receiving email for link to start recovery - [ ] Verify receiving email for successfully recovering - [ ] Verify email link is invalid after successful recovery - [ ] Verify receiving email for locked account when max attempts reached ## RBAC Create a role, with no `allow.rules` defined: ``` kind: role metadata: name: rbac spec: allow: app_labels: '*': '*' logins: - root node_labels: '*': '*' options: max_session_ttl: 8h0m0s version: v3 ``` - [ ] Verify that a user has access only to: "Servers", "Applications", "Databases", "Kubernetes", "Active Sessions", "Access Requests" and "Manage Clusters" - [ ] Verify there is no `Add Server, Application, Databases, Kubernetes` button in each respective view - [ ] Verify only `Servers`, `Apps`, `Databases`, and `Kubernetes` are listed under `options` button in `Manage Clusters` Note: User has read/create access_request access to their own requests, despite resource settings Add the following under `spec.allow.rules` to enable read access to the audit log: ``` - resources: - event verbs: - list ``` - [ ] Verify that the `Audit Log` and `Session Recordings` is accessible - [ ] Verify that playing a recorded session is denied Add the following to enable read access to recorded sessions ``` - resources: - session verbs: - read ``` - [ ] Verify that a user can re-play a session (session.end) Add the following to enable read access to the roles ``` - resources: - role verbs: - list - read ``` - [ ] Verify that a user can see the roles - [ ] Verify that a user cannot create/delete/update a role Add the following to enable read access to the auth connectors ``` - resources: - auth_connector verbs: - list - read ``` - [ ] Verify that a user can see the list of auth connectors. - [ ] Verify that a user cannot create/delete/update the connectors Add the following to enable read access to users ``` - resources: - user verbs: - list - read ``` - [ ] Verify that a user can access the "Users" screen - [ ] Verify that a user cannot reset password and create/delete/update a user Add the following to enable read access to trusted clusters ``` - resources: - trusted_cluster verbs: - list - read ``` - [ ] Verify that a user can access the "Trust" screen - [ ] Verify that a user cannot create/delete/update a trusted cluster. ## Performance/Soak Test Using `tsh bench` tool, perform the soak tests and benchmark tests on the following configurations: * Cluster with 10K nodes in normal (non-IOT) node mode with ETCD * Cluster with 10K nodes in normal (non-IOT) mode with DynamoDB * Cluster with 1K IOT nodes with ETCD * Cluster with 1K IOT nodes with DynamoDB * Cluster with 500 trusted clusters with ETCD * Cluster with 500 trusted clusters with DynamoDB **Soak Tests** Run 4hour soak test with a mix of interactive/non-interactive sessions: ``` tsh bench --duration=4h user@teleport-monster-6757d7b487-x226b ls tsh bench -i --duration=4h user@teleport-monster-6757d7b487-x226b ps uax ``` Observe prometheus metrics for goroutines, open files, RAM, CPU, Timers and make sure there are no leaks - [ ] Verify that prometheus metrics are accurate. **Breaking load tests** Load system with tsh bench to the capacity and publish maximum numbers of concurrent sessions with interactive and non interactive tsh bench loads. ## Teleport with Cloud Providers ### AWS - [ ] Deploy Teleport to AWS. Using DynamoDB & S3 - [ ] Deploy Teleport Enterprise to AWS. Using HA Setup https://gravitational.com/teleport/docs/aws-terraform-guide/ ### GCP - [ ] Deploy Teleport to GCP. Using Cloud Firestore & Cloud Storage - [ ] Deploy Teleport to GKE. Google Kubernetes engine. - [ ] Deploy Teleport Enterprise to GCP. ### IBM - [ ] Deploy Teleport to IBM Cloud. Using IBM Database for etcd & IBM Object Store - [ ] Deploy Teleport to IBM Cloud Kubernetes. - [ ] Deploy Teleport Enterprise to IBM Cloud. ## Application Access - [ ] Run an application within local cluster. - [ ] Verify the debug application `debug_app: true` works. - [ ] Verify an application can be configured with command line flags. - [ ] Verify an application can be configured from file configuration. - [ ] Verify that applications are available at auto-generated addresses `name.rootProxyPublicAddr` and well as `publicAddr`. - [ ] Run an application within a trusted cluster. - [ ] Verify that applications are available at auto-generated addresses `name.rootProxyPublicAddr`. - [ ] Verify Audit Records. - [ ] `app.session.start` and `app.session.chunk` events are created in the Audit Log. - [ ] `app.session.chunk` points to a 5 minute session archive with multiple `app.session.request` events inside. - [ ] `tsh play <chunk-id>` can fetch and print a session chunk archive. - [ ] Verify JWT using [verify-jwt.go](https://github.com/gravitational/teleport/blob/master/examples/jwt/verify-jwt.go). - [ ] Verify RBAC. - [ ] Verify [CLI access](https://goteleport.com/docs/application-access/guides/api-access/) with `tsh app login`. - [ ] Verify AWS console access. - [ ] Can log into AWS web console through the web UI. - [ ] Can interact with AWS using `tsh aws` commands. - [ ] Verify dynamic registration. - [ ] Can register a new app using `tctl create`. - [ ] Can update registered app using `tctl create -f`. - [ ] Can delete registered app using `tctl rm`. - [ ] Test Applications screen in the web UI (tab is located on left side nav on dashboard): - [ ] Verify that all apps registered are shown - [ ] Verify that clicking on the app icon takes you to another tab - [ ] Verify using the bash command produced from `Add Application` dialogue works (refresh app screen to see it registered) ## Database Access - [ ] Connect to a database within a local cluster. - [ ] Self-hosted Postgres. - [ ] Self-hosted MySQL. - [ ] Self-hosted MariaDB. - [ ] Self-hosted MongoDB. - [ ] Self-hosted CockroachDB. - [ ] AWS Aurora Postgres. - [ ] AWS Aurora MySQL. - [ ] AWS Redshift. - [ ] GCP Cloud SQL Postgres. - [ ] GCP Cloud SQL MySQL. - [ ] Connect to a database within a remote cluster via a trusted cluster. - [ ] Self-hosted Postgres. - [ ] Self-hosted MySQL. - [ ] Self-hosted MariaDB. - [ ] Self-hosted MongoDB. - [ ] Self-hosted CockroachDB. - [ ] AWS Aurora Postgres. - [ ] AWS Aurora MySQL. - [ ] AWS Redshift. - [ ] GCP Cloud SQL Postgres. - [ ] GCP Cloud SQL MySQL. - [ ] Verify audit events. - [ ] `db.session.start` is emitted when you connect. - [ ] `db.session.end` is emitted when you disconnect. - [ ] `db.session.query` is emitted when you execute a SQL query. - [ ] Verify RBAC. - [ ] `tsh db ls` shows only databases matching role's `db_labels`. - [ ] Can only connect as users from `db_users`. - [ ] _(Postgres only)_ Can only connect to databases from `db_names`. - [ ] `db.session.start` is emitted when connection attempt is denied. - [ ] _(MongoDB only)_ Can only execute commands in databases from `db_names`. - [ ] `db.session.query` is emitted when command fails due to permissions. - [ ] Can configure per-session MFA. - [ ] MFA tap is required on each `tsh db connect`. - [ ] Verify dynamic registration. - [ ] Can register a new database using `tctl create`. - [ ] Can update registered database using `tctl create -f`. - [ ] Can delete registered database using `tctl rm`. - [ ] Verify discovery. - [ ] Can detect and register RDS instances. - [ ] Can detect and register Aurora clusters, and their reader and custom endpoints. - [ ] Can detect and register Redshift clusters. - [ ] Test Databases screen in the web UI (tab is located on left side nav on dashboard): - [ ] Verify that all dbs registered are shown with correct `name`, `description`, `type`, and `labels` - [ ] Verify that clicking on a rows connect button renders a dialogue on manual instructions with `Step 2` login value matching the rows `name` column - [ ] Verify searching for all columns in the search bar works - [ ] Verify you can sort by all columns except `labels` ## TLS Routing - [ ] Verify that teleport proxy `v2` configuration starts only a single listener. ``` version: v2 teleport: proxy_service: enabled: "yes" public_addr: ['root.example.com'] web_listen_addr: 0.0.0.0:3080 ``` - [ ] Run Teleport Proxy in `multiplex` mode `auth_service.proxy_listener_mode: "multiplex"` - [ ] Trusted cluster - [ ] Setup trusted clusters using single port setup `web_proxy_addr == tunnel_addr` ``` kind: trusted_cluster spec: ... web_proxy_addr: root.example.com:443 tunnel_addr: root.example.com:443 ... ``` - [ ] Database Access - [ ] Verify that `tsh db connect` works through proxy running in `multiplex` mode - [ ] Postgres - [ ] MySQL - [ ] MariaDB - [ ] MongoDB - [ ] CockroachDB - [ ] Verify connecting to a database through TLS ALPN SNI local proxy `tsh db proxy` with a GUI client. - [ ] Application Access - [ ] Verify app access through proxy running in `multiplex` mode - [ ] SSH Access - [ ] Connect to a OpenSSH server through a local ssh proxy `ssh -o "ForwardAgent yes" -o "ProxyCommand tsh proxy ssh" user@host.example.com` - [ ] Connect to a OpenSSH server on leaf-cluster through a local ssh proxy`ssh -o "ForwardAgent yes" -o "ProxyCommand tsh proxy ssh --user=%r --cluster=leaf-cluster %h:%p" user@node.foo.com` - [ ] Verify `tsh ssh` access through proxy running in multiplex mode - [ ] Kubernetes access: - [ ] Verify kubernetes access through proxy running in `multiplex` mode ## Desktop Access - [ ] Can connect to desktop defined in static `hosts` section. - [ ] Can connect to desktop discovered via LDAP - [ ] Download [Keyboard Key Info](https://dennisbabkin.com/kbdkeyinfo/) and verify all keys are processed correctly in each supported browser. Known issues: F11 cannot be captured by the browser without [special configuration](https://social.technet.microsoft.com/Forums/en-US/784b2bbe-353f-412e-ac9a-193d81f306b6/remote-desktop-for-mac-f11-key-not-working-on-macbook-pro-touchbar?forum=winRDc) on MacOS. - [ ] Left click and right click register as Windows clicks. (Right click on the desktop should show a Windows menu, not a browser context menu) - [ ] Vertical and horizontal scroll work. [Horizontal Scroll Test](https://codepen.io/jaemskyle/pen/inbmB) - [ ] All desktops have `teleport.dev/origin` label. - [ ] Dynamic desktops have additional `teleport.dev` labels for OS, OS Version, DNS hostname. - [ ] Verify that placing a user lock terminates an active desktop session. - [ ] Verify desktop session start/end audit events. - [ ] Regexp-based host labeling applies across all desktops, regardless of origin. - [ ] RBAC denies access to a Windows desktop due to labels - [ ] RBAC denies access to a Windows desktop with the wrong OS-login. - [ ] Multiple sessions as different users on the same desktop are allowed. - [ ] Connect multiple `windows_desktop_service`s to the same Teleport cluster, verify that connections to desktops on different AD domains works.
test
teleport test plan manual testing plan below are the items that should be manually tested with each release of teleport these tests should be run on both a fresh install of the version to be released as well as an upgrade of the previous version of teleport adding nodes to a cluster adding nodes via valid static token adding nodes via valid short lived tokens adding nodes via invalid token fails revoking node invitation labels static labels dynamic labels trusted clusters adding trusted cluster valid static token adding trusted cluster valid short lived token adding trusted cluster invalid token removing trusted cluster rbac make sure that invalid and valid attempts are reflected in audit log successfully connect to node with correct role unsuccessfully connect to a node in a role restricting access by label unsuccessfully connect to a node in a role restricting access by invalid ssh login allow deny role option ssh agent forwarding allow deny role option port forwarding verify that custom pam environment variables are available as expected users with every user combination try to login and signup with invalid second factor invalid password to see how the system reacts adding users password only adding users otp adding users adding users webauthn managing mfa devices add an otp device with tsh mfa add add a device with tsh mfa add verify that the device works under webauthn add a webauthn device with tsh mfa add list mfa devices with tsh mfa ls remove an otp device with tsh mfa rm remove a device with tsh mfa rm remove a webauthn device with tsh mfa rm attempt removing the last mfa device on the user with second factor on in auth service should fail with second factor optional in auth service should succeed login password only login with mfa add otp and webauthn devices with tsh mfa add login via otp login via webauthn login oidc login saml login github deleting users backends teleport runs with etcd teleport runs with dynamodb teleport runs with sqlite teleport runs with firestore session recording session recording can be disabled sessions can be recorded at the node sessions in remote clusters are recorded in remote clusters sessions can be recorded at the proxy sessions on remote clusters are recorded in the local cluster enable disable host key checking audit log failed login attempts are recorded interactive sessions have the correct server id server id is the id of the node in session recording node mode server id is the id of the proxy in session recording proxy mode node proxy id may be found at var lib teleport host uuid in the corresponding machine node ids may also be queried via tctl nodes ls exec commands are recorded scp commands are recorded subsystem results are recorded subsystem testing may be achieved using both and assuming the proxy is proxy example com and is a node running openssh sshd you may use the following command to trigger a subsystem audit log shell sftp o proxycommand ssh o forwardagent yes p r proxy example com s proxy h p root interact with a cluster using tsh these commands should ideally be tested for recording and non recording modes as they are implemented in a different ways tsh ssh tsh ssh tsh ssh a tsh ssh a tsh ssh ls tsh ssh ls tsh join tsh join tsh play tsh play tsh scp tsh scp tsh ssh l tsh ssh l tsh ls tsh clusters interact with a cluster using ssh make sure to test both recording and regular proxy modes ssh ssh ssh a ssh a ssh ls ssh ls scp scp ssh l ssh l interact with a cluster using the web ui connect to a teleport node connect to a openssh node check agent forwarding is correct based on role and proxy mode user accounting verify that active interactive sessions are tracked in var run utmp on linux verify that interactive sessions are logged in var log wtmp on linux combinations for some manual testing many combinations need to be tested for example for interactive sessions the combinations are below connect to a openssh node in a local cluster using openssh connect to a openssh node in a local cluster using teleport connect to a openssh node in a local cluster using the web ui connect to a teleport node in a local cluster using openssh connect to a teleport node in a local cluster using teleport connect to a teleport node in a local cluster using the web ui connect to a openssh node in a remote cluster using openssh connect to a openssh node in a remote cluster using teleport connect to a openssh node in a remote cluster using the web ui connect to a teleport node in a remote cluster using openssh connect to a teleport node in a remote cluster using teleport connect to a teleport node in a remote cluster using the web ui teleport with eks gke deploy teleport on a single eks cluster deploy teleport on two eks clusters and connect them via trusted cluster feature deploy teleport proxy outside of gke cluster fronting connections to it use to generate a kubeconfig deploy teleport proxy outside of eks cluster fronting connections to it use to generate a kubeconfig teleport with multiple kubernetes clusters note you can use gke or eks or minikube to run kubernetes clusters minikube is the only caveat it s not reachable publicly so don t run a proxy there deploy combo auth proxy kubernetes service outside of a kubernetes cluster using a kubeconfig login with tsh login check that tsh kube ls has your cluster run kubectl get nodes kubectl exec it some pod sh verify that the audit log recorded the above request and session deploy combo auth proxy kubernetes service inside of a kubernetes cluster login with tsh login check that tsh kube ls has your cluster run kubectl get nodes kubectl exec it some pod sh verify that the audit log recorded the above request and session deploy combo auth proxy service outside of the kubernetes cluster and kubernetes service inside of a kubernetes cluster connected over a reverse tunnel login with tsh login check that tsh kube ls has your cluster run kubectl get nodes kubectl exec it some pod sh verify that the audit log recorded the above request and session deploy a second kubernetes service inside of another kubernetes cluster connected over a reverse tunnel login with tsh login check that tsh kube ls has both clusters switch to a second cluster using tsh kube login run kubectl get nodes kubectl exec it some pod sh on the new cluster verify that the audit log recorded the above request and session deploy combo auth proxy kubernetes service outside of a kubernetes cluster using a kubeconfig with multiple clusters in it login with tsh login check that tsh kube ls has all clusters test kubernetes screen in the web ui tab is located on left side nav on dashboard verify that all kubes registered are shown with correct name and labels verify that clicking on a rows connect button renders a dialogue on manual instructions with step login value matching the rows name column verify searching for name or labels in the search bar works verify you can sort by name colum teleport with fips mode perform trusted clusters web and ssh sanity check with all teleport components deployed in fips mode acme teleport can fetch tls certificate automatically using acme protocol migrations migrate trusted clusters from to migrate auth server on main cluster then rest of the servers on main cluster ssh should work for both main and old clusters migrate auth server on remote cluster then rest of the remote cluster ssh should work command templates when interacting with a cluster the following command templates are useful openssh when connecting to the recording proxy o forwardagent yes is required ssh o proxycommand ssh o forwardagent yes p r proxy example com s proxy h p node example com the above command only forwards the agent to the proxy to forward the agent to the target node o forwardagent yes needs to be passed twice ssh o forwardagent yes o proxycommand ssh o forwardagent yes p r proxy example com s proxy h p node example com when connecting to a remote cluster using openssh the subsystem request is updated with the name of the remote cluster ssh o proxycommand ssh o forwardagent yes p r proxy example com s proxy h p foo com node foo com teleport when connecting to a openssh node remember p needs to be passed tsh proxy proxy example com user insecure ssh p node example com an agent can be forwarded to the target node with a tsh proxy proxy example com user insecure ssh a p node example com the cluster flag is used to connect to a node in a remote cluster tsh proxy proxy example com user insecure ssh cluster foo com p node foo com teleport with sso providers g suite install instructions work g suite screenshots are up to date activedirectory install instructions work active directory screenshots are up to date okta install instructions work okta screenshots are up to date onelogin install instructions work onelogin screenshots are up to date oidc install instructions work oidc screenshots are up to date teleport plugins test receiving a message via teleport slackbot test receiving a new jira ticket via teleport jira web ui main for main test with a role that has access to all resources top nav verify that cluster selector displays all root leaf clusters verify that user name is displayed verify that user menu shows logout help support and account settings for local users side nav verify that each item has an icon verify that collapse expand works and collapsed has icon and expand has icon v verify that it automatically expands and highlights the item on page refresh servers aka nodes verify that servers table shows all joined nodes verify that connect button shows a list of available logins verify that hostname address and labels columns show the current values verify that search by hostname address labels works verify that terminal opens when clicking on one of the available logins verify that clicking on add server button renders dialogue set to automatically view verify clicking on regenerate script regenerates token value in the bash command verify using the bash command successfully adds the server refresh server list verify that clicking on manually tab renders manual steps verify that clicking back to automatically tab renders bash command applications verify that clicking on add application button renders dialogue verify input validation prevent empty value and invalid url verify after input and clicking on generate script bash command is rendered verify clicking on regenerate button regenerates token value in bash command databases verify that clicking on add database button renders dialogue for manual instructions verify selecting different options on step changes step commands active sessions verify that empty state is handled verify that it displays the session when session is active verify that description session id users nodes and duration columns show correct values verify that options button allows to join a session audit log verify that time range button is shown and works verify that clicking on session ended event icon takes user to session player verify event detail dialogue renders when clicking on events details button verify searching by type description created works users verify that users are shown verify that creating a new user works verify that editing user roles works verify that removing a user works verify resetting a user s password works verify search by username roles and type works auth connectors verify when there are no connectors empty state renders verify that creating oidc saml github connectors works verify that editing oidc saml github connectors works verify that error is shown when saving an invalid yaml verify that correct hint text is shown on the right side verify that encrypted saml assertions work with an identity provider that supports it azure verify that created github saml oidc card has their icons roles verify that roles are shown verify that create new role dialog works verify that deleting and editing works verify that error is shown when saving an invalid yaml verify that correct hint text is shown on the right side managed clusters verify that it displays a list of clusters root leaf verify that every menu item works nodes apps audit events session recordings etc help support verify that all urls work and correct no access requests creating access requests create a role with limited permissions defined below as allow roles this role allows you to see the role screen and ssh into all nodes create another role with limited permissions defined below as allow users this role session expires in minutes allows you to see users screen and denies access to all nodes create another role with no permissions other than being able to create requests defined below as default create a user with role default assigned create a few requests under this user to test pending approved denied state kind role metadata name allow roles spec allow logins root node labels rules resources role verbs list read options max session ttl version kind role metadata name allow users short ttl spec allow rules resources user verbs list read deny node labels options max session ttl version kind role metadata name default spec allow request roles allow roles allow users suggested reviewers random user random user options max session ttl version verify that under requestable roles only allow roles and allow users are listed verify input validation requires at least one role to be selected verify you can select input modify reviewers verify after creating a request requests are listed in pending states verify you can t review own requests viewing approving denying requests create a user with the role reviewer that allows you to review all requests and delete them kind role version metadata name reviewer spec allow review requests roles verify you can view access request from request list verify there is list of reviewers you selected empty list if none selected and suggested reviewers wasn t defined verify threshold name is there it will be default if thresholds weren t defined in role or blank if not named verify you can approve a request with message and immediately see updated state with your review stamp green checkmark and message box verify you can deny a request and immediately see updated state with your review stamp red cross verify deleting the denied request is removed from list assuming approved requests verify assume buttons are only present for approved request and for logged in user verify that assuming allow roles allows you to see roles screen and ssh into nodes verify that after clicking on the assume button it is disabled in both the list and in viewing after assuming allow roles verify that assuming allow users short ttl allows you to see users screen and denies access to nodes verify a switchback banner is rendered with roles assumed and count down of when it expires verify switching back goes back to your default static role verify after re assuming allow users short ttl role the user is automatically logged out after the expiry is met minutes verify that after logging out or getting logged out automatically and relogging in permissions are reset to default and requests that are not expired and are approved are assumable again access request waiting room strategy reason create the following role kind role metadata name waiting room spec allow request roles options max session ttl request access reason request prompt version verify after login reason dialogue is rendered with prompt set to request prompt setting verify after clicking send request pending dialogue renders verify after approving a request dashboard is rendered verify the correct role was assigned strategy always with the previous role you created from strategy reason change request access to always verify after login pending dialogue is auto rendered verify after approving a request dashboard is rendered verify after denying a request access denied dialogue is rendered verify a switchback banner is rendered with roles assumed and count down of when it expires verify switchback button says logout and clicking goes back to the login screen strategy optional with the previous role you created from strategy reason change request access to optional verify after login dashboard is rendered as normal terminal verify that top nav has a user menu main and logout verify that switching between tabs works on alt node list tab verify that cluster selector works url should change too verify that quick launcher input works verify that quick launcher input handles input errors verify that connect button shows a list of available logins verify that hostname address and labels columns show the current values verify that search by hostname address labels work verify that new tab is created when starting a session session tab verify that session and browser tabs both show the title with login and node name verify that terminal resize works install midnight commander on the node you ssh into sudo apt get install mc run the program mc resize the terminal to see if panels resize with it verify that session tab shows updates number of participants when a new user joins the session verify that tab automatically closes on exit command verify that scp upload works verify that scp upload handles invalid paths and network errors verify that scp download works verify that scp download handles invalid paths and network errors session player verify that it can replay a session verify that when playing scroller auto scrolls to bottom most content verify when resizing player to a small screen scroller appears and is working verify that error message is displayed enter an invalid sid in the url invite and reset form verify that input validates verify that invite works with disabled verify that invite works with otp enabled verify that invite works with enabled verify that invite works with webauthn enabled verify that error message is shown if an invite is expired invalid login form and change password verify that input validates verify that login works with disabled verify that changing passwords works for disabled verify that login works with otp enabled verify that changing passwords works for otp enabled verify that login works with enabled verify that changing passwords works for enabled verify that login works with webauthn enabled verify that changing passwords works for webauthn enabled verify that login works for github saml oidc verify that redirect to original url works after successful login verify that account is locked after several unsuccessful login attempts verify that account is locked after several unsuccessful change password attempts multi factor authentication mfa create modify teleport yaml and set the following authentication settings under auth service yaml authentication type local second factor optional require session mfa yes webauthn rp id example com mfa invite login password reset change password verify during invite reset second factor list all auth types none hardware key and authenticator app verify registration works with all option types verify login with all option types verify changing password with all option types change second factor type to on and verify that mfa is required no option none in dropdown mfa require auth go to account settings two factor devices and register a new device using the same user as above verify logging in with registered webauthn key works verify connecting to a ssh node prompts you to tap your registered webauthn key verify in the web terminal you can scp upload download files mfa management verify adding first device works without requiring re authentication verify re authenticating with a webauthn device works verify re authenticating with a device works verify re authenticating with a otp device works verify adding a webauthn device works verify adding a device works verify adding an otp device works verify removing a device works verify second factor set to off disables adding devices cloud from your cloud staging account change the field teleportversion to the test version kubectl n edit tenant recovery code management verify generating recovery codes for local accounts with email usernames works verify local accounts with non email usernames are not able to generate recovery codes verify sso accounts are not able to generate recovery codes invite reset verify email as usernames renders recovery codes dialog verify non email usernames does not render recovery codes dialog recovery flow add new mfa device verify recovering adding a new hardware key device with password verify recovering adding a new otp device with password verify viewing and deleting any old device but not the one just added verify new recovery codes are rendered at the end of flow recovery flow change password verify recovering password with any mfa device verify new recovery codes are rendered at the end of flow recovery email verify receiving email for link to start recovery verify receiving email for successfully recovering verify email link is invalid after successful recovery verify receiving email for locked account when max attempts reached rbac create a role with no allow rules defined kind role metadata name rbac spec allow app labels logins root node labels options max session ttl version verify that a user has access only to servers applications databases kubernetes active sessions access requests and manage clusters verify there is no add server application databases kubernetes button in each respective view verify only servers apps databases and kubernetes are listed under options button in manage clusters note user has read create access request access to their own requests despite resource settings add the following under spec allow rules to enable read access to the audit log resources event verbs list verify that the audit log and session recordings is accessible verify that playing a recorded session is denied add the following to enable read access to recorded sessions resources session verbs read verify that a user can re play a session session end add the following to enable read access to the roles resources role verbs list read verify that a user can see the roles verify that a user cannot create delete update a role add the following to enable read access to the auth connectors resources auth connector verbs list read verify that a user can see the list of auth connectors verify that a user cannot create delete update the connectors add the following to enable read access to users resources user verbs list read verify that a user can access the users screen verify that a user cannot reset password and create delete update a user add the following to enable read access to trusted clusters resources trusted cluster verbs list read verify that a user can access the trust screen verify that a user cannot create delete update a trusted cluster performance soak test using tsh bench tool perform the soak tests and benchmark tests on the following configurations cluster with nodes in normal non iot node mode with etcd cluster with nodes in normal non iot mode with dynamodb cluster with iot nodes with etcd cluster with iot nodes with dynamodb cluster with trusted clusters with etcd cluster with trusted clusters with dynamodb soak tests run soak test with a mix of interactive non interactive sessions tsh bench duration user teleport monster ls tsh bench i duration user teleport monster ps uax observe prometheus metrics for goroutines open files ram cpu timers and make sure there are no leaks verify that prometheus metrics are accurate breaking load tests load system with tsh bench to the capacity and publish maximum numbers of concurrent sessions with interactive and non interactive tsh bench loads teleport with cloud providers aws deploy teleport to aws using dynamodb deploy teleport enterprise to aws using ha setup gcp deploy teleport to gcp using cloud firestore cloud storage deploy teleport to gke google kubernetes engine deploy teleport enterprise to gcp ibm deploy teleport to ibm cloud using ibm database for etcd ibm object store deploy teleport to ibm cloud kubernetes deploy teleport enterprise to ibm cloud application access run an application within local cluster verify the debug application debug app true works verify an application can be configured with command line flags verify an application can be configured from file configuration verify that applications are available at auto generated addresses name rootproxypublicaddr and well as publicaddr run an application within a trusted cluster verify that applications are available at auto generated addresses name rootproxypublicaddr verify audit records app session start and app session chunk events are created in the audit log app session chunk points to a minute session archive with multiple app session request events inside tsh play can fetch and print a session chunk archive verify jwt using verify rbac verify with tsh app login verify aws console access can log into aws web console through the web ui can interact with aws using tsh aws commands verify dynamic registration can register a new app using tctl create can update registered app using tctl create f can delete registered app using tctl rm test applications screen in the web ui tab is located on left side nav on dashboard verify that all apps registered are shown verify that clicking on the app icon takes you to another tab verify using the bash command produced from add application dialogue works refresh app screen to see it registered database access connect to a database within a local cluster self hosted postgres self hosted mysql self hosted mariadb self hosted mongodb self hosted cockroachdb aws aurora postgres aws aurora mysql aws redshift gcp cloud sql postgres gcp cloud sql mysql connect to a database within a remote cluster via a trusted cluster self hosted postgres self hosted mysql self hosted mariadb self hosted mongodb self hosted cockroachdb aws aurora postgres aws aurora mysql aws redshift gcp cloud sql postgres gcp cloud sql mysql verify audit events db session start is emitted when you connect db session end is emitted when you disconnect db session query is emitted when you execute a sql query verify rbac tsh db ls shows only databases matching role s db labels can only connect as users from db users postgres only can only connect to databases from db names db session start is emitted when connection attempt is denied mongodb only can only execute commands in databases from db names db session query is emitted when command fails due to permissions can configure per session mfa mfa tap is required on each tsh db connect verify dynamic registration can register a new database using tctl create can update registered database using tctl create f can delete registered database using tctl rm verify discovery can detect and register rds instances can detect and register aurora clusters and their reader and custom endpoints can detect and register redshift clusters test databases screen in the web ui tab is located on left side nav on dashboard verify that all dbs registered are shown with correct name description type and labels verify that clicking on a rows connect button renders a dialogue on manual instructions with step login value matching the rows name column verify searching for all columns in the search bar works verify you can sort by all columns except labels tls routing verify that teleport proxy configuration starts only a single listener version teleport proxy service enabled yes public addr web listen addr run teleport proxy in multiplex mode auth service proxy listener mode multiplex trusted cluster setup trusted clusters using single port setup web proxy addr tunnel addr kind trusted cluster spec web proxy addr root example com tunnel addr root example com database access verify that tsh db connect works through proxy running in multiplex mode postgres mysql mariadb mongodb cockroachdb verify connecting to a database through tls alpn sni local proxy tsh db proxy with a gui client application access verify app access through proxy running in multiplex mode ssh access connect to a openssh server through a local ssh proxy ssh o forwardagent yes o proxycommand tsh proxy ssh user host example com connect to a openssh server on leaf cluster through a local ssh proxy ssh o forwardagent yes o proxycommand tsh proxy ssh user r cluster leaf cluster h p user node foo com verify tsh ssh access through proxy running in multiplex mode kubernetes access verify kubernetes access through proxy running in multiplex mode desktop access can connect to desktop defined in static hosts section can connect to desktop discovered via ldap download and verify all keys are processed correctly in each supported browser known issues cannot be captured by the browser without on macos left click and right click register as windows clicks right click on the desktop should show a windows menu not a browser context menu vertical and horizontal scroll work all desktops have teleport dev origin label dynamic desktops have additional teleport dev labels for os os version dns hostname verify that placing a user lock terminates an active desktop session verify desktop session start end audit events regexp based host labeling applies across all desktops regardless of origin rbac denies access to a windows desktop due to labels rbac denies access to a windows desktop with the wrong os login multiple sessions as different users on the same desktop are allowed connect multiple windows desktop service s to the same teleport cluster verify that connections to desktops on different ad domains works
1
683,405
23,380,498,624
IssuesEvent
2022-08-11 09:00:42
blindnet-io/product-management
https://api.github.com/repos/blindnet-io/product-management
closed
Bug on download PDFs in ACCESS privacy request
type: bug priority: 1 (urgent)
As a submitter, when I make an ACCESS privacy request, when I try to download the PDF of a form, I sometimes can't and get the following "error server" message I notice that when I view the form, and come back to download it it works : ![Kapture 2022-08-01 at 18 14 44](https://user-images.githubusercontent.com/89908145/182194658-e26c132d-5eb0-4bee-a987-38ed472c12d9.gif)
1.0
Bug on download PDFs in ACCESS privacy request - As a submitter, when I make an ACCESS privacy request, when I try to download the PDF of a form, I sometimes can't and get the following "error server" message I notice that when I view the form, and come back to download it it works : ![Kapture 2022-08-01 at 18 14 44](https://user-images.githubusercontent.com/89908145/182194658-e26c132d-5eb0-4bee-a987-38ed472c12d9.gif)
non_test
bug on download pdfs in access privacy request as a submitter when i make an access privacy request when i try to download the pdf of a form i sometimes can t and get the following error server message i notice that when i view the form and come back to download it it works
0
379,713
26,383,085,042
IssuesEvent
2023-01-12 10:09:28
umbraco/UmbracoDocs
https://api.github.com/repos/umbraco/UmbracoDocs
closed
Docs to mention cache refresher notifications
category/missing-documentation category/umbraco-cms
***This is the page with issues: https://our.umbraco.com/documentation/Reference/Notifications/ContentService-Notifications/ There are several notifications mentioned here, but I couldn't find anything about these cache refresher notifications: - CacheRefresherNotification - ContentCacheRefresherNotification - MediaCacheRefresherNotification - MemberCacheRefresherNotification - UserCacheRefresherNotification Not sure if these belong under e.g. [ContentService Notifications](https://our.umbraco.com/documentation/Reference/Notifications/ContentService-Notifications/) and [MediaService Notifications](https://our.umbraco.com/documentation/Reference/Notifications/MediaService-Notifications/) or if there should be a separate page for cache refresher notification? Also see this old issue as reference in v8 https://github.com/umbraco/Umbraco-CMS/issues/10169 where I used these `ContentCacheRefresherUpdated` event.
1.0
Docs to mention cache refresher notifications - ***This is the page with issues: https://our.umbraco.com/documentation/Reference/Notifications/ContentService-Notifications/ There are several notifications mentioned here, but I couldn't find anything about these cache refresher notifications: - CacheRefresherNotification - ContentCacheRefresherNotification - MediaCacheRefresherNotification - MemberCacheRefresherNotification - UserCacheRefresherNotification Not sure if these belong under e.g. [ContentService Notifications](https://our.umbraco.com/documentation/Reference/Notifications/ContentService-Notifications/) and [MediaService Notifications](https://our.umbraco.com/documentation/Reference/Notifications/MediaService-Notifications/) or if there should be a separate page for cache refresher notification? Also see this old issue as reference in v8 https://github.com/umbraco/Umbraco-CMS/issues/10169 where I used these `ContentCacheRefresherUpdated` event.
non_test
docs to mention cache refresher notifications this is the page with issues there are several notifications mentioned here but i couldn t find anything about these cache refresher notifications cacherefreshernotification contentcacherefreshernotification mediacacherefreshernotification membercacherefreshernotification usercacherefreshernotification not sure if these belong under e g and or if there should be a separate page for cache refresher notification also see this old issue as reference in where i used these contentcacherefresherupdated event
0
80,716
23,287,420,041
IssuesEvent
2022-08-05 18:07:52
hyperledger/caliper
https://api.github.com/repos/hyperledger/caliper
closed
Code coverage reporting integrated into the build
enhancement build
### Please share the technical limitation of Caliper that you encountered. No code coverage report is available for PRs ### Please detail your feature idea that could alleviate the limitation. IT would be good for the build to produce a code coverage report. This would allow maintainers to review the PR more easily wereas at the moment a maintainer has to checkout the PR and run the tests to get a report. Also it could be useful for submitters as well to review. ### Please share some details about your use case if possible, and how the new feature would make Caliper a better performance benchmarking framework. _No response_ ### Please share any suggestions about the new feature's code/configuration API (using formatted YAML segments or pseudo-code). _No response_
1.0
Code coverage reporting integrated into the build - ### Please share the technical limitation of Caliper that you encountered. No code coverage report is available for PRs ### Please detail your feature idea that could alleviate the limitation. IT would be good for the build to produce a code coverage report. This would allow maintainers to review the PR more easily wereas at the moment a maintainer has to checkout the PR and run the tests to get a report. Also it could be useful for submitters as well to review. ### Please share some details about your use case if possible, and how the new feature would make Caliper a better performance benchmarking framework. _No response_ ### Please share any suggestions about the new feature's code/configuration API (using formatted YAML segments or pseudo-code). _No response_
non_test
code coverage reporting integrated into the build please share the technical limitation of caliper that you encountered no code coverage report is available for prs please detail your feature idea that could alleviate the limitation it would be good for the build to produce a code coverage report this would allow maintainers to review the pr more easily wereas at the moment a maintainer has to checkout the pr and run the tests to get a report also it could be useful for submitters as well to review please share some details about your use case if possible and how the new feature would make caliper a better performance benchmarking framework no response please share any suggestions about the new feature s code configuration api using formatted yaml segments or pseudo code no response
0
123,125
17,772,175,925
IssuesEvent
2021-08-30 14:49:26
kapseliboi/compiler-explorer
https://api.github.com/repos/kapseliboi/compiler-explorer
opened
WS-2019-0103 (Medium) detected in handlebars-4.0.5.min.js
security vulnerability
## WS-2019-0103 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.min.js</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.min.js</a></p> <p>Path to dependency file: compiler-explorer/node_modules/vis/examples/timeline/styling/itemTemplates.html</p> <p>Path to vulnerable library: /node_modules/vis/examples/timeline/styling/itemTemplates.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.0.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/compiler-explorer/commit/e37a5b97c4d98423bce7e1c48beb642adb05b447">e37a5b97c4d98423bce7e1c48beb642adb05b447</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars.js before 4.1.0 has Remote Code Execution (RCE) <p>Publish Date: 2019-01-30 <p>URL: <a href=https://github.com/wycats/handlebars.js/issues/1267#issue-187151586>WS-2019-0103</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57">https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57</a></p> <p>Release Date: 2019-05-30</p> <p>Fix Resolution: 4.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
WS-2019-0103 (Medium) detected in handlebars-4.0.5.min.js - ## WS-2019-0103 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>handlebars-4.0.5.min.js</b></p></summary> <p>Handlebars provides the power necessary to let you build semantic templates effectively with no frustration</p> <p>Library home page: <a href="https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.min.js">https://cdnjs.cloudflare.com/ajax/libs/handlebars.js/4.0.5/handlebars.min.js</a></p> <p>Path to dependency file: compiler-explorer/node_modules/vis/examples/timeline/styling/itemTemplates.html</p> <p>Path to vulnerable library: /node_modules/vis/examples/timeline/styling/itemTemplates.html</p> <p> Dependency Hierarchy: - :x: **handlebars-4.0.5.min.js** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/kapseliboi/compiler-explorer/commit/e37a5b97c4d98423bce7e1c48beb642adb05b447">e37a5b97c4d98423bce7e1c48beb642adb05b447</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Handlebars.js before 4.1.0 has Remote Code Execution (RCE) <p>Publish Date: 2019-01-30 <p>URL: <a href=https://github.com/wycats/handlebars.js/issues/1267#issue-187151586>WS-2019-0103</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57">https://github.com/wycats/handlebars.js/commit/edc6220d51139b32c28e51641fadad59a543ae57</a></p> <p>Release Date: 2019-05-30</p> <p>Fix Resolution: 4.1.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
ws medium detected in handlebars min js ws medium severity vulnerability vulnerable library handlebars min js handlebars provides the power necessary to let you build semantic templates effectively with no frustration library home page a href path to dependency file compiler explorer node modules vis examples timeline styling itemtemplates html path to vulnerable library node modules vis examples timeline styling itemtemplates html dependency hierarchy x handlebars min js vulnerable library found in head commit a href found in base branch main vulnerability details handlebars js before has remote code execution rce publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with whitesource
0
335,021
30,004,918,821
IssuesEvent
2023-06-26 11:50:17
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
closed
[CI] AutoFollowIT testAutoFollowPatterns failing
>test-failure :Distributed/CCR Team:Distributed
**Build scan:** https://gradle-enterprise.elastic.co/s/x57fzeiwekdve/tests/:x-pack:plugin:ccr:qa:multi-cluster:follow-cluster/org.elasticsearch.xpack.ccr.AutoFollowIT/testAutoFollowPatterns **Reproduction line:** ``` ./gradlew ':x-pack:plugin:ccr:qa:multi-cluster:follow-cluster' --tests "org.elasticsearch.xpack.ccr.AutoFollowIT.testAutoFollowPatterns" -Dtests.seed=50A0B430B1BC44DE -Dtests.locale=sr-BA -Dtests.timezone=Pacific/Auckland -Druntime.java=8 ``` **Applicable branches:** 7.17 **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.ccr.AutoFollowIT&tests.test=testAutoFollowPatterns **Failure excerpt:** ``` java.lang.AssertionError: Unexpected number of followed indices [{"took":10,"timed_out":false,"_shards":{"total":2,"successful":2,"skipped":0,"failed":0},"hits":{"total":{"value":34,"relation":"eq"},"max_score":2.75246,"hits":[{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"cVLbv4QBur_2UWL7Sj95","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:50.533Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:50.357Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"eVLbv4QBur_2UWL7cT-p","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:00.547Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:00.358Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"gVLbv4QBur_2UWL7mD-9","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:10.563Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:10.358Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"oFLcv4QBur_2UWL7DT_d","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:40.538Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:40.359Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"i1Lbv4QBur_2UWL7vz-6","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:20.535Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:20.358Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"YVLav4QBur_2UWL7_D9Y","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:30.505Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:30.353Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"aVLbv4QBur_2UWL7Iz9S","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:40.503Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:40.354Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"l1Lbv4QBur_2UWL75j_l","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:30.503Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:30.359Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"WVLav4QBur_2UWL71j8T","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:20.659Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:20.353Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"UVLav4QBur_2UWL7uj-y","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:11.054Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:10.353Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}}]}}] Expected: a value equal to or greater than <1> but: <0> was less than <1> at __randomizedtesting.SeedInfo.seed([50A0B430B1BC44DE:F9070A1B7952D55A]:0) at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) at org.junit.Assert.assertThat(Assert.java:956) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.verifyAutoFollowMonitoring(ESCCRRestTestCase.java:247) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1141) at org.elasticsearch.xpack.ccr.AutoFollowIT.assertLongBusy(AutoFollowIT.java:905) at org.elasticsearch.xpack.ccr.AutoFollowIT.testAutoFollowPatterns(AutoFollowIT.java:176) at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:748) ```
1.0
[CI] AutoFollowIT testAutoFollowPatterns failing - **Build scan:** https://gradle-enterprise.elastic.co/s/x57fzeiwekdve/tests/:x-pack:plugin:ccr:qa:multi-cluster:follow-cluster/org.elasticsearch.xpack.ccr.AutoFollowIT/testAutoFollowPatterns **Reproduction line:** ``` ./gradlew ':x-pack:plugin:ccr:qa:multi-cluster:follow-cluster' --tests "org.elasticsearch.xpack.ccr.AutoFollowIT.testAutoFollowPatterns" -Dtests.seed=50A0B430B1BC44DE -Dtests.locale=sr-BA -Dtests.timezone=Pacific/Auckland -Druntime.java=8 ``` **Applicable branches:** 7.17 **Reproduces locally?:** No **Failure history:** https://gradle-enterprise.elastic.co/scans/tests?tests.container=org.elasticsearch.xpack.ccr.AutoFollowIT&tests.test=testAutoFollowPatterns **Failure excerpt:** ``` java.lang.AssertionError: Unexpected number of followed indices [{"took":10,"timed_out":false,"_shards":{"total":2,"successful":2,"skipped":0,"failed":0},"hits":{"total":{"value":34,"relation":"eq"},"max_score":2.75246,"hits":[{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"cVLbv4QBur_2UWL7Sj95","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:50.533Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:50.357Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"eVLbv4QBur_2UWL7cT-p","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:00.547Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:00.358Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"gVLbv4QBur_2UWL7mD-9","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:10.563Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:10.358Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"oFLcv4QBur_2UWL7DT_d","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:40.538Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:40.359Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"i1Lbv4QBur_2UWL7vz-6","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:20.535Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:20.358Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"YVLav4QBur_2UWL7_D9Y","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:30.505Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:30.353Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"aVLbv4QBur_2UWL7Iz9S","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:40.503Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:40.354Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"l1Lbv4QBur_2UWL75j_l","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:09:30.503Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:09:30.359Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"WVLav4QBur_2UWL71j8T","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:20.659Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:20.353Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}},{"_index":".monitoring-es-7-2022.11.28","_type":"_doc","_id":"UVLav4QBur_2UWL7uj-y","_score":2.75246,"_source":{"cluster_uuid":"MJcQUR1jTqqfK7SYlC2tbw","timestamp":"2022-11-28T20:08:11.054Z","interval_ms":10000,"type":"ccr_auto_follow_stats","source_node":{"uuid":"Ql0gc-E9Si-p7FWiiH_TWw","host":"127.0.0.1","transport_address":"127.0.0.1:43959","ip":"127.0.0.1","name":"follow-cluster-0","timestamp":"2022-11-28T20:08:10.353Z"},"ccr_auto_follow_stats":{"number_of_failed_follow_indices":0,"number_of_failed_remote_cluster_state_requests":0,"number_of_successful_follow_indices":0,"recent_auto_follow_errors":[],"auto_followed_clusters":[]}}}]}}] Expected: a value equal to or greater than <1> but: <0> was less than <1> at __randomizedtesting.SeedInfo.seed([50A0B430B1BC44DE:F9070A1B7952D55A]:0) at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:18) at org.junit.Assert.assertThat(Assert.java:956) at org.elasticsearch.xpack.ccr.ESCCRRestTestCase.verifyAutoFollowMonitoring(ESCCRRestTestCase.java:247) at org.elasticsearch.test.ESTestCase.assertBusy(ESTestCase.java:1141) at org.elasticsearch.xpack.ccr.AutoFollowIT.assertLongBusy(AutoFollowIT.java:905) at org.elasticsearch.xpack.ccr.AutoFollowIT.testAutoFollowPatterns(AutoFollowIT.java:176) at sun.reflect.NativeMethodAccessorImpl.invoke0(NativeMethodAccessorImpl.java:-2) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at com.carrotsearch.randomizedtesting.RandomizedRunner.invoke(RandomizedRunner.java:1758) at com.carrotsearch.randomizedtesting.RandomizedRunner$8.evaluate(RandomizedRunner.java:946) at com.carrotsearch.randomizedtesting.RandomizedRunner$9.evaluate(RandomizedRunner.java:982) at com.carrotsearch.randomizedtesting.RandomizedRunner$10.evaluate(RandomizedRunner.java:996) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleSetupTeardownChained$1.evaluate(TestRuleSetupTeardownChained.java:49) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at org.apache.lucene.util.TestRuleThreadAndTestName$1.evaluate(TestRuleThreadAndTestName.java:48) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.forkTimeoutingTask(ThreadLeakControl.java:843) at com.carrotsearch.randomizedtesting.ThreadLeakControl$3.evaluate(ThreadLeakControl.java:490) at com.carrotsearch.randomizedtesting.RandomizedRunner.runSingleTest(RandomizedRunner.java:955) at com.carrotsearch.randomizedtesting.RandomizedRunner$5.evaluate(RandomizedRunner.java:840) at com.carrotsearch.randomizedtesting.RandomizedRunner$6.evaluate(RandomizedRunner.java:891) at com.carrotsearch.randomizedtesting.RandomizedRunner$7.evaluate(RandomizedRunner.java:902) at org.apache.lucene.util.AbstractBeforeAfterRule$1.evaluate(AbstractBeforeAfterRule.java:45) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleStoreClassName$1.evaluate(TestRuleStoreClassName.java:41) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.NoShadowingOrOverridesOnMethodsRule$1.evaluate(NoShadowingOrOverridesOnMethodsRule.java:40) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at org.apache.lucene.util.TestRuleAssertionsRequired$1.evaluate(TestRuleAssertionsRequired.java:53) at org.apache.lucene.util.TestRuleMarkFailure$1.evaluate(TestRuleMarkFailure.java:47) at org.apache.lucene.util.TestRuleIgnoreAfterMaxFailures$1.evaluate(TestRuleIgnoreAfterMaxFailures.java:64) at org.apache.lucene.util.TestRuleIgnoreTestSuites$1.evaluate(TestRuleIgnoreTestSuites.java:54) at com.carrotsearch.randomizedtesting.rules.StatementAdapter.evaluate(StatementAdapter.java:36) at com.carrotsearch.randomizedtesting.ThreadLeakControl$StatementRunner.run(ThreadLeakControl.java:390) at com.carrotsearch.randomizedtesting.ThreadLeakControl.lambda$forkTimeoutingTask$0(ThreadLeakControl.java:850) at java.lang.Thread.run(Thread.java:748) ```
test
autofollowit testautofollowpatterns failing build scan reproduction line gradlew x pack plugin ccr qa multi cluster follow cluster tests org elasticsearch xpack ccr autofollowit testautofollowpatterns dtests seed dtests locale sr ba dtests timezone pacific auckland druntime java applicable branches reproduces locally no failure history failure excerpt java lang assertionerror unexpected number of followed indices auto followed clusters index monitoring es type doc id p score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters index monitoring es type doc id score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters index monitoring es type doc id d score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters index monitoring es type doc id score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters index monitoring es type doc id score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters index monitoring es type doc id score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters index monitoring es type doc id l score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters index monitoring es type doc id score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters index monitoring es type doc id y score source cluster uuid timestamp interval ms type ccr auto follow stats source node uuid tww host transport address ip name follow cluster timestamp ccr auto follow stats number of failed follow indices number of failed remote cluster state requests number of successful follow indices recent auto follow errors auto followed clusters expected a value equal to or greater than but was less than at randomizedtesting seedinfo seed at org hamcrest matcherassert assertthat matcherassert java at org junit assert assertthat assert java at org elasticsearch xpack ccr esccrresttestcase verifyautofollowmonitoring esccrresttestcase java at org elasticsearch test estestcase assertbusy estestcase java at org elasticsearch xpack ccr autofollowit assertlongbusy autofollowit java at org elasticsearch xpack ccr autofollowit testautofollowpatterns autofollowit java at sun reflect nativemethodaccessorimpl nativemethodaccessorimpl java at sun reflect nativemethodaccessorimpl invoke nativemethodaccessorimpl java at sun reflect delegatingmethodaccessorimpl invoke delegatingmethodaccessorimpl java at java lang reflect method invoke method java at com carrotsearch randomizedtesting randomizedrunner invoke randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulesetupteardownchained evaluate testrulesetupteardownchained java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at org apache lucene util testrulethreadandtestname evaluate testrulethreadandtestname java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol forktimeoutingtask threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol evaluate threadleakcontrol java at com carrotsearch randomizedtesting randomizedrunner runsingletest randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at com carrotsearch randomizedtesting randomizedrunner evaluate randomizedrunner java at org apache lucene util abstractbeforeafterrule evaluate abstractbeforeafterrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testrulestoreclassname evaluate testrulestoreclassname java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules noshadowingoroverridesonmethodsrule evaluate noshadowingoroverridesonmethodsrule java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at org apache lucene util testruleassertionsrequired evaluate testruleassertionsrequired java at org apache lucene util testrulemarkfailure evaluate testrulemarkfailure java at org apache lucene util testruleignoreaftermaxfailures evaluate testruleignoreaftermaxfailures java at org apache lucene util testruleignoretestsuites evaluate testruleignoretestsuites java at com carrotsearch randomizedtesting rules statementadapter evaluate statementadapter java at com carrotsearch randomizedtesting threadleakcontrol statementrunner run threadleakcontrol java at com carrotsearch randomizedtesting threadleakcontrol lambda forktimeoutingtask threadleakcontrol java at java lang thread run thread java
1
184,653
14,289,809,596
IssuesEvent
2020-11-23 19:51:46
github-vet/rangeclosure-findings
https://api.github.com/repos/github-vet/rangeclosure-findings
closed
matt-deboer/mpp: vendor/github.com/matt-deboer/go-marathon/application_test.go; 29 LoC
fresh small test
Found a possible issue in [matt-deboer/mpp](https://www.github.com/matt-deboer/mpp) at [vendor/github.com/matt-deboer/go-marathon/application_test.go](https://github.com/matt-deboer/mpp/blob/5903cbc2f71cace128623f2bb3c28fa6dcb28333/vendor/github.com/matt-deboer/go-marathon/application_test.go#L581-L609) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/matt-deboer/mpp/blob/5903cbc2f71cace128623f2bb3c28fa6dcb28333/vendor/github.com/matt-deboer/go-marathon/application_test.go#L581-L609) <details> <summary>Click here to show the 29 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range tests { defaultConfig := NewDefaultConfig() defaultConfig.PollingWaitTime = waitTime configs := &configContainer{ client: &defaultConfig, server: &serverConfig{ scope: test.testScope, }, } endpoint := newFakeMarathonEndpoint(t, configs) defer endpoint.Close() errCh := make(chan error) go func() { errCh <- endpoint.Client.WaitOnApplication(test.appName, test.timeout) }() select { case <-time.After(400 * time.Millisecond): assert.Fail(t, fmt.Sprintf("%s: WaitOnApplication did not complete in time", test.desc)) case err := <-errCh: if test.shouldSucceed { assert.NoError(t, err, test.desc) } else { assert.IsType(t, err, ErrTimeoutError, test.desc) } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 5903cbc2f71cace128623f2bb3c28fa6dcb28333
1.0
matt-deboer/mpp: vendor/github.com/matt-deboer/go-marathon/application_test.go; 29 LoC - Found a possible issue in [matt-deboer/mpp](https://www.github.com/matt-deboer/mpp) at [vendor/github.com/matt-deboer/go-marathon/application_test.go](https://github.com/matt-deboer/mpp/blob/5903cbc2f71cace128623f2bb3c28fa6dcb28333/vendor/github.com/matt-deboer/go-marathon/application_test.go#L581-L609) The below snippet of Go code triggered static analysis which searches for goroutines and/or defer statements which capture loop variables. [Click here to see the code in its original context.](https://github.com/matt-deboer/mpp/blob/5903cbc2f71cace128623f2bb3c28fa6dcb28333/vendor/github.com/matt-deboer/go-marathon/application_test.go#L581-L609) <details> <summary>Click here to show the 29 line(s) of Go which triggered the analyzer.</summary> ```go for _, test := range tests { defaultConfig := NewDefaultConfig() defaultConfig.PollingWaitTime = waitTime configs := &configContainer{ client: &defaultConfig, server: &serverConfig{ scope: test.testScope, }, } endpoint := newFakeMarathonEndpoint(t, configs) defer endpoint.Close() errCh := make(chan error) go func() { errCh <- endpoint.Client.WaitOnApplication(test.appName, test.timeout) }() select { case <-time.After(400 * time.Millisecond): assert.Fail(t, fmt.Sprintf("%s: WaitOnApplication did not complete in time", test.desc)) case err := <-errCh: if test.shouldSucceed { assert.NoError(t, err, test.desc) } else { assert.IsType(t, err, ErrTimeoutError, test.desc) } } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 5903cbc2f71cace128623f2bb3c28fa6dcb28333
test
matt deboer mpp vendor github com matt deboer go marathon application test go loc found a possible issue in at the below snippet of go code triggered static analysis which searches for goroutines and or defer statements which capture loop variables click here to show the line s of go which triggered the analyzer go for test range tests defaultconfig newdefaultconfig defaultconfig pollingwaittime waittime configs configcontainer client defaultconfig server serverconfig scope test testscope endpoint newfakemarathonendpoint t configs defer endpoint close errch make chan error go func errch endpoint client waitonapplication test appname test timeout select case time after time millisecond assert fail t fmt sprintf s waitonapplication did not complete in time test desc case err errch if test shouldsucceed assert noerror t err test desc else assert istype t err errtimeouterror test desc leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
57,003
6,536,253,091
IssuesEvent
2017-08-31 17:25:11
pydata/xarray
https://api.github.com/repos/pydata/xarray
closed
rolling requires pandas >= 0.18
pandas-like testing
We need pandas >= 0.18 because [dataframe.rolling](https://pandas.pydata.org/pandas-docs/stable/generated/pandas.DataFrame.rolling.html) is supported after 0.18. But `requirements` in our setup.py says we need pandas >= 0.15. Additionally, I noticed that in travis's CONDA_ENV=py27-min setup, our unit tests run with pandas == 0.20, though it might be intended to run with pandas == 0.15. By `conda remove scipy`, pandas.0.15 is removed. (Here is the travis log) ``` if [[ "$CONDA_ENV" == "py27-min" ]]; then conda remove scipy; fi Fetching package metadata ......... Solving package specifications: . Package plan for package removal in environment /home/travis/miniconda/envs/test_env: The following packages will be REMOVED: pandas: 0.15.0-np19py27_0 defaults scipy: 0.17.1-np19py27_1 defaults ``` then in `python setup.py install`, pandas==0.20.3 is installed. ``` Searching for pandas>=0.15.0 Reading https://pypi.python.org/simple/pandas/ Downloading https://pypi.python.org/packages/ee/aa/90c06f249cf4408fa75135ad0df7d64c09cf74c9870733862491ed5f3a50/pandas-0.20.3.tar.gz#md5=4df858f28b4bf4fa07d9fbb7f2568173 Best match: pandas 0.20.3 ```
1.0
rolling requires pandas >= 0.18 - We need pandas >= 0.18 because [dataframe.rolling](https://pandas.pydata.org/pandas-docs/stable/generated/pandas.DataFrame.rolling.html) is supported after 0.18. But `requirements` in our setup.py says we need pandas >= 0.15. Additionally, I noticed that in travis's CONDA_ENV=py27-min setup, our unit tests run with pandas == 0.20, though it might be intended to run with pandas == 0.15. By `conda remove scipy`, pandas.0.15 is removed. (Here is the travis log) ``` if [[ "$CONDA_ENV" == "py27-min" ]]; then conda remove scipy; fi Fetching package metadata ......... Solving package specifications: . Package plan for package removal in environment /home/travis/miniconda/envs/test_env: The following packages will be REMOVED: pandas: 0.15.0-np19py27_0 defaults scipy: 0.17.1-np19py27_1 defaults ``` then in `python setup.py install`, pandas==0.20.3 is installed. ``` Searching for pandas>=0.15.0 Reading https://pypi.python.org/simple/pandas/ Downloading https://pypi.python.org/packages/ee/aa/90c06f249cf4408fa75135ad0df7d64c09cf74c9870733862491ed5f3a50/pandas-0.20.3.tar.gz#md5=4df858f28b4bf4fa07d9fbb7f2568173 Best match: pandas 0.20.3 ```
test
rolling requires pandas we need pandas because is supported after but requirements in our setup py says we need pandas additionally i noticed that in travis s conda env min setup our unit tests run with pandas though it might be intended to run with pandas by conda remove scipy pandas is removed here is the travis log if then conda remove scipy fi fetching package metadata solving package specifications package plan for package removal in environment home travis miniconda envs test env the following packages will be removed pandas defaults scipy defaults then in python setup py install pandas is installed searching for pandas reading downloading best match pandas
1
154,935
19,765,559,560
IssuesEvent
2022-01-17 01:27:52
Vali00/serverless-full-stack-apps-azure-sql
https://api.github.com/repos/Vali00/serverless-full-stack-apps-azure-sql
opened
CVE-2022-0155 (High) detected in follow-redirects-1.13.3.tgz, follow-redirects-1.13.2.tgz
security vulnerability
## CVE-2022-0155 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>follow-redirects-1.13.3.tgz</b>, <b>follow-redirects-1.13.2.tgz</b></p></summary> <p> <details><summary><b>follow-redirects-1.13.3.tgz</b></p></summary> <p>HTTP and HTTPS modules that follow redirects.</p> <p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.13.3.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.13.3.tgz</a></p> <p>Path to dependency file: /azure-static-web-app/api/node/package.json</p> <p>Path to vulnerable library: /azure-static-web-app/api/node/node_modules/follow-redirects/package.json</p> <p> Dependency Hierarchy: - mssql-6.3.1.tgz (Root Library) - tedious-6.7.0.tgz - ms-rest-nodeauth-2.0.2.tgz - ms-rest-js-1.11.2.tgz - axios-0.21.1.tgz - :x: **follow-redirects-1.13.3.tgz** (Vulnerable Library) </details> <details><summary><b>follow-redirects-1.13.2.tgz</b></p></summary> <p>HTTP and HTTPS modules that follow redirects.</p> <p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.13.2.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.13.2.tgz</a></p> <p>Path to dependency file: /azure-function/node/package.json</p> <p>Path to vulnerable library: /azure-function/node/node_modules/follow-redirects/package.json</p> <p> Dependency Hierarchy: - mssql-6.3.1.tgz (Root Library) - tedious-6.7.0.tgz - ms-rest-nodeauth-2.0.2.tgz - ms-rest-js-1.11.2.tgz - axios-0.21.1.tgz - :x: **follow-redirects-1.13.2.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> follow-redirects is vulnerable to Exposure of Private Personal Information to an Unauthorized Actor <p>Publish Date: 2022-01-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155>CVE-2022-0155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/">https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: follow-redirects - v1.14.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-0155 (High) detected in follow-redirects-1.13.3.tgz, follow-redirects-1.13.2.tgz - ## CVE-2022-0155 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>follow-redirects-1.13.3.tgz</b>, <b>follow-redirects-1.13.2.tgz</b></p></summary> <p> <details><summary><b>follow-redirects-1.13.3.tgz</b></p></summary> <p>HTTP and HTTPS modules that follow redirects.</p> <p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.13.3.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.13.3.tgz</a></p> <p>Path to dependency file: /azure-static-web-app/api/node/package.json</p> <p>Path to vulnerable library: /azure-static-web-app/api/node/node_modules/follow-redirects/package.json</p> <p> Dependency Hierarchy: - mssql-6.3.1.tgz (Root Library) - tedious-6.7.0.tgz - ms-rest-nodeauth-2.0.2.tgz - ms-rest-js-1.11.2.tgz - axios-0.21.1.tgz - :x: **follow-redirects-1.13.3.tgz** (Vulnerable Library) </details> <details><summary><b>follow-redirects-1.13.2.tgz</b></p></summary> <p>HTTP and HTTPS modules that follow redirects.</p> <p>Library home page: <a href="https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.13.2.tgz">https://registry.npmjs.org/follow-redirects/-/follow-redirects-1.13.2.tgz</a></p> <p>Path to dependency file: /azure-function/node/package.json</p> <p>Path to vulnerable library: /azure-function/node/node_modules/follow-redirects/package.json</p> <p> Dependency Hierarchy: - mssql-6.3.1.tgz (Root Library) - tedious-6.7.0.tgz - ms-rest-nodeauth-2.0.2.tgz - ms-rest-js-1.11.2.tgz - axios-0.21.1.tgz - :x: **follow-redirects-1.13.2.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> follow-redirects is vulnerable to Exposure of Private Personal Information to an Unauthorized Actor <p>Publish Date: 2022-01-10 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-0155>CVE-2022-0155</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>8.0</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/">https://huntr.dev/bounties/fc524e4b-ebb6-427d-ab67-a64181020406/</a></p> <p>Release Date: 2022-01-10</p> <p>Fix Resolution: follow-redirects - v1.14.7</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in follow redirects tgz follow redirects tgz cve high severity vulnerability vulnerable libraries follow redirects tgz follow redirects tgz follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file azure static web app api node package json path to vulnerable library azure static web app api node node modules follow redirects package json dependency hierarchy mssql tgz root library tedious tgz ms rest nodeauth tgz ms rest js tgz axios tgz x follow redirects tgz vulnerable library follow redirects tgz http and https modules that follow redirects library home page a href path to dependency file azure function node package json path to vulnerable library azure function node node modules follow redirects package json dependency hierarchy mssql tgz root library tedious tgz ms rest nodeauth tgz ms rest js tgz axios tgz x follow redirects tgz vulnerable library found in base branch main vulnerability details follow redirects is vulnerable to exposure of private personal information to an unauthorized actor publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required low user interaction required scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution follow redirects step up your open source security game with whitesource
0
149,233
11,886,448,961
IssuesEvent
2020-03-27 21:58:01
FlightControl-Master/MOOSE
https://api.github.com/repos/FlightControl-Master/MOOSE
closed
ATIS: Disable subtitles
enhancement implemented ready for testing
`ATIS:SetSubtitleDuration(0)` should completely disable the subtitles.
1.0
ATIS: Disable subtitles - `ATIS:SetSubtitleDuration(0)` should completely disable the subtitles.
test
atis disable subtitles atis setsubtitleduration should completely disable the subtitles
1
123,730
10,281,172,769
IssuesEvent
2019-08-26 07:50:04
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
closed
Question about local E2E test
sig/docs sig/testing
Recently I am trying to do e2e test in my local server and I followed this guidance, https://github.com/kubernetes/community/blob/master/contributors/devel/sig-testing/e2e-tests.md#local-clusters However, I found local e2e test had been removed from source code. I tracked the change log, but did not find any information on why remove it. Here are two pr for removing the code: https://github.com/kubernetes/kubernetes/pull/49148 https://github.com/kubernetes/kubernetes/pull/76711 So here is my question, is it still possible to do local e2e test now and how? If not, is there any plan to enable local e2e test in the future? Thanks a lot for your answer, I have trapped here really a long time. /sig Testing /sig Docs
1.0
Question about local E2E test - Recently I am trying to do e2e test in my local server and I followed this guidance, https://github.com/kubernetes/community/blob/master/contributors/devel/sig-testing/e2e-tests.md#local-clusters However, I found local e2e test had been removed from source code. I tracked the change log, but did not find any information on why remove it. Here are two pr for removing the code: https://github.com/kubernetes/kubernetes/pull/49148 https://github.com/kubernetes/kubernetes/pull/76711 So here is my question, is it still possible to do local e2e test now and how? If not, is there any plan to enable local e2e test in the future? Thanks a lot for your answer, I have trapped here really a long time. /sig Testing /sig Docs
test
question about local test recently i am trying to do test in my local server and i followed this guidance however i found local test had been removed from source code i tracked the change log but did not find any information on why remove it here are two pr for removing the code so here is my question is it still possible to do local test now and how if not is there any plan to enable local test in the future thanks a lot for your answer i have trapped here really a long time sig testing sig docs
1
323,063
27,664,906,746
IssuesEvent
2023-03-12 23:06:41
systemd/systemd
https://api.github.com/repos/systemd/systemd
closed
Still test-time-util is unstable
bug ๐Ÿ› util-lib tests
See https://download.copr.fedorainfracloud.org/results/packit/systemd-systemd-26678/fedora-rawhide-i386/05594021-systemd/builder-live.log.gz ``` /* test_format_timestamp_with_tz_one(Asia/Dili) */ Assertion 'parse_timestamp(xx, &y) >= 0' failed at src/test/test-time-util.c:398, function test_format_timestamp_impl(). Aborting. ```
1.0
Still test-time-util is unstable - See https://download.copr.fedorainfracloud.org/results/packit/systemd-systemd-26678/fedora-rawhide-i386/05594021-systemd/builder-live.log.gz ``` /* test_format_timestamp_with_tz_one(Asia/Dili) */ Assertion 'parse_timestamp(xx, &y) >= 0' failed at src/test/test-time-util.c:398, function test_format_timestamp_impl(). Aborting. ```
test
still test time util is unstable see test format timestamp with tz one asia dili assertion parse timestamp xx y failed at src test test time util c function test format timestamp impl aborting
1
114,955
14,667,597,966
IssuesEvent
2020-12-29 19:02:17
newrelic/docs-website
https://api.github.com/repos/newrelic/docs-website
closed
User sessions: UX feedback punch list
design product work in progress
This issue will contain a list of all feedback we get in our user preview sessions so we can consider what changes we want to make. ### Feedback - [x] The Sign Up buttons to create an account should be more explicit relating to what the user is doing on our DECO sites on the Docs site: "Sign Up" makes sense but on the Dev site and OSS Site the button may be "Start Now". Decide what approach to take an implement . issue: https://github.com/newrelic/gatsby-theme-newrelic/issues/182 - [x] We should add an intro paragraph to all our list index pages, to help with SEO. ex: https://docs-preview.newrelic.com/docs/agents ex: https://docs-preview.newrelic.com/docs/accounts issue: https://github.com/newrelic/docs-website/issues/491 - [ ] Does a button that leads you into the product need an indicator that will tell a user we are taking them out of the site exp? ex: Add Java Data found on this page: https://docs-preview.newrelic.com/docs/agents/java-agent/getting-started/introduction-new-relic-java - [x ] Do we need more clarity around what the create an issue / edit this page means to a user in the right rail? `A: no, it's pretty explicit` - [x] should we move the `For more help` headers to the right rail as a dedicated resource link? `A: no, as we are putting related content in the right rail. ` - [x] Instrument the right rail to determine if a user clicks on the right rail auto-gen link vs manually added links. Probably needs Tessen integration. `A: added to MMF11 for instrumentation` - [x] In the Right rail, should we treat the non-numbered headers slightly different so a user understands that the step process is over and the rest of the information is just for deeper Knowledge or examples, etc. `A: fixed` - [x] Related Resources could auto generate related content like blogs, and what is the best way to weight that and ensure it makes it on a page? How do we surface the right blog content and add them as related resources? Can this be a manual process at first, and then something we automate? `A: this is part of our ongoing work to better manage swiftype indexing and content labels in swiftype, any url can be added as a static resource to any page. ` ### Contributor Feedback - [x] **look into how GH Codespace**s be used To help contributors get more comfortable with using the VSCode editor, Codespaces lets you have the power of an IDE all within the browser, so you donโ€™t need to worry about local dev environments shenanigans. There eventually is a cost associated with the user, that weโ€™d have to be ok paying for on this repo. `A: Individuals can use codespaces on public or private repos, but enterprise repos won't have this functionality initially. each person is charged for their own usage, and GH isn't offering this for teams yet. eventually it may be useful for NR teams, but it's not yet available for enterprises and is charged to the individuals GH account` - [x] **Create Getting Started Scripts** `A: this has been added to MMF11 for considerations` - Allowing a user to run a script to get their local environment setup quickly? Maybe this setups the following: - Installs Vscode and all plugins to edit - Installs Gatsby CLI, and all dependencies to work with Gatsby, correct node version, yarn, etc - Clones the repo - Setups github SSH authentication for the user so they can contribute directly to repo - Runs yarn install to grab all the dependencies - Runs yarn start - Loads the site in their default browser on local-host: 8000 - Now have fun! `A: this has been added to MMF11 for considerations` - [x] **Leverage Github Actions;** Is there a way to better showcase where to find the amplify preview link? Netlify does a good job of making that very obvious in the GitHub action section of the PR. What else could we surface in a Github action to aid a contributor in feeling comfortable they didnโ€™t break anything? Some linter tests, other build checks, etc. essentially anything that would let a contributor feel confident in what they are about to contribute.
1.0
User sessions: UX feedback punch list - This issue will contain a list of all feedback we get in our user preview sessions so we can consider what changes we want to make. ### Feedback - [x] The Sign Up buttons to create an account should be more explicit relating to what the user is doing on our DECO sites on the Docs site: "Sign Up" makes sense but on the Dev site and OSS Site the button may be "Start Now". Decide what approach to take an implement . issue: https://github.com/newrelic/gatsby-theme-newrelic/issues/182 - [x] We should add an intro paragraph to all our list index pages, to help with SEO. ex: https://docs-preview.newrelic.com/docs/agents ex: https://docs-preview.newrelic.com/docs/accounts issue: https://github.com/newrelic/docs-website/issues/491 - [ ] Does a button that leads you into the product need an indicator that will tell a user we are taking them out of the site exp? ex: Add Java Data found on this page: https://docs-preview.newrelic.com/docs/agents/java-agent/getting-started/introduction-new-relic-java - [x ] Do we need more clarity around what the create an issue / edit this page means to a user in the right rail? `A: no, it's pretty explicit` - [x] should we move the `For more help` headers to the right rail as a dedicated resource link? `A: no, as we are putting related content in the right rail. ` - [x] Instrument the right rail to determine if a user clicks on the right rail auto-gen link vs manually added links. Probably needs Tessen integration. `A: added to MMF11 for instrumentation` - [x] In the Right rail, should we treat the non-numbered headers slightly different so a user understands that the step process is over and the rest of the information is just for deeper Knowledge or examples, etc. `A: fixed` - [x] Related Resources could auto generate related content like blogs, and what is the best way to weight that and ensure it makes it on a page? How do we surface the right blog content and add them as related resources? Can this be a manual process at first, and then something we automate? `A: this is part of our ongoing work to better manage swiftype indexing and content labels in swiftype, any url can be added as a static resource to any page. ` ### Contributor Feedback - [x] **look into how GH Codespace**s be used To help contributors get more comfortable with using the VSCode editor, Codespaces lets you have the power of an IDE all within the browser, so you donโ€™t need to worry about local dev environments shenanigans. There eventually is a cost associated with the user, that weโ€™d have to be ok paying for on this repo. `A: Individuals can use codespaces on public or private repos, but enterprise repos won't have this functionality initially. each person is charged for their own usage, and GH isn't offering this for teams yet. eventually it may be useful for NR teams, but it's not yet available for enterprises and is charged to the individuals GH account` - [x] **Create Getting Started Scripts** `A: this has been added to MMF11 for considerations` - Allowing a user to run a script to get their local environment setup quickly? Maybe this setups the following: - Installs Vscode and all plugins to edit - Installs Gatsby CLI, and all dependencies to work with Gatsby, correct node version, yarn, etc - Clones the repo - Setups github SSH authentication for the user so they can contribute directly to repo - Runs yarn install to grab all the dependencies - Runs yarn start - Loads the site in their default browser on local-host: 8000 - Now have fun! `A: this has been added to MMF11 for considerations` - [x] **Leverage Github Actions;** Is there a way to better showcase where to find the amplify preview link? Netlify does a good job of making that very obvious in the GitHub action section of the PR. What else could we surface in a Github action to aid a contributor in feeling comfortable they didnโ€™t break anything? Some linter tests, other build checks, etc. essentially anything that would let a contributor feel confident in what they are about to contribute.
non_test
user sessions ux feedback punch list this issue will contain a list of all feedback we get in our user preview sessions so we can consider what changes we want to make feedback the sign up buttons to create an account should be more explicit relating to what the user is doing on our deco sites on the docs site sign up makes sense but on the dev site and oss site the button may be start now decide what approach to take an implement issue we should add an intro paragraph to all our list index pages to help with seo ex ex issue does a button that leads you into the product need an indicator that will tell a user we are taking them out of the site exp ex add java data found on this page do we need more clarity around what the create an issue edit this page means to a user in the right rail a no it s pretty explicit should we move the for more help headers to the right rail as a dedicated resource link a no as we are putting related content in the right rail instrument the right rail to determine if a user clicks on the right rail auto gen link vs manually added links probably needs tessen integration a added to for instrumentation in the right rail should we treat the non numbered headers slightly different so a user understands that the step process is over and the rest of the information is just for deeper knowledge or examples etc a fixed related resources could auto generate related content like blogs and what is the best way to weight that and ensure it makes it on a page how do we surface the right blog content and add them as related resources can this be a manual process at first and then something we automate a this is part of our ongoing work to better manage swiftype indexing and content labels in swiftype any url can be added as a static resource to any page contributor feedback look into how gh codespace s be used to help contributors get more comfortable with using the vscode editor codespaces lets you have the power of an ide all within the browser so you donโ€™t need to worry about local dev environments shenanigans there eventually is a cost associated with the user that weโ€™d have to be ok paying for on this repo a individuals can use codespaces on public or private repos but enterprise repos won t have this functionality initially each person is charged for their own usage and gh isn t offering this for teams yet eventually it may be useful for nr teams but it s not yet available for enterprises and is charged to the individuals gh account create getting started scripts a this has been added to for considerations allowing a user to run a script to get their local environment setup quickly maybe this setups the following installs vscode and all plugins to edit installs gatsby cli and all dependencies to work with gatsby correct node version yarn etc clones the repo setups github ssh authentication for the user so they can contribute directly to repo runs yarn install to grab all the dependencies runs yarn start loads the site in their default browser on local host now have fun a this has been added to for considerations leverage github actions is there a way to better showcase where to find the amplify preview link netlify does a good job of making that very obvious in the github action section of the pr what else could we surface in a github action to aid a contributor in feeling comfortable they didnโ€™t break anything some linter tests other build checks etc essentially anything that would let a contributor feel confident in what they are about to contribute
0
187,032
14,426,916,111
IssuesEvent
2020-12-06 00:45:42
kalexmills/github-vet-tests-dec2020
https://api.github.com/repos/kalexmills/github-vet-tests-dec2020
closed
futurewei-cloud/global-scheduler: staging/src/k8s.io/apiextensions-apiserver/test/integration/registration_test.go; 5 LoC
fresh test tiny
Found a possible issue in [futurewei-cloud/global-scheduler](https://www.github.com/futurewei-cloud/global-scheduler) at [staging/src/k8s.io/apiextensions-apiserver/test/integration/registration_test.go](https://github.com/futurewei-cloud/global-scheduler/blob/b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e/staging/src/k8s.io/apiextensions-apiserver/test/integration/registration_test.go#L127-L131) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to a at line 128 may start a goroutine [Click here to see the code in its original context.](https://github.com/futurewei-cloud/global-scheduler/blob/b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e/staging/src/k8s.io/apiextensions-apiserver/test/integration/registration_test.go#L127-L131) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, a := range listWithItem.Items { if e := instances[a.GetName()].Instance; !reflect.DeepEqual(e, &a) { t.Errorf("expected %v, got %v", e, a) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e
1.0
futurewei-cloud/global-scheduler: staging/src/k8s.io/apiextensions-apiserver/test/integration/registration_test.go; 5 LoC - Found a possible issue in [futurewei-cloud/global-scheduler](https://www.github.com/futurewei-cloud/global-scheduler) at [staging/src/k8s.io/apiextensions-apiserver/test/integration/registration_test.go](https://github.com/futurewei-cloud/global-scheduler/blob/b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e/staging/src/k8s.io/apiextensions-apiserver/test/integration/registration_test.go#L127-L131) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > function call which takes a reference to a at line 128 may start a goroutine [Click here to see the code in its original context.](https://github.com/futurewei-cloud/global-scheduler/blob/b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e/staging/src/k8s.io/apiextensions-apiserver/test/integration/registration_test.go#L127-L131) <details> <summary>Click here to show the 5 line(s) of Go which triggered the analyzer.</summary> ```go for _, a := range listWithItem.Items { if e := instances[a.GetName()].Instance; !reflect.DeepEqual(e, &a) { t.Errorf("expected %v, got %v", e, a) } } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: b9329a9fbcd5ca571c8acc498b9d42f5e8b9c19e
test
futurewei cloud global scheduler staging src io apiextensions apiserver test integration registration test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message function call which takes a reference to a at line may start a goroutine click here to show the line s of go which triggered the analyzer go for a range listwithitem items if e instances instance reflect deepequal e a t errorf expected v got v e a leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
41,451
2,869,007,449
IssuesEvent
2015-06-05 22:32:29
dart-lang/http_server
https://api.github.com/repos/dart-lang/http_server
closed
http_server depends on pkg path but does not specify it in pubspec.yaml
bug Fixed Priority-Medium
<a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)** _Originally opened as dart-lang/sdk#14499_ ---- simply declaring http_server as a dependency is not enough. workaround for now: also depend on path.
1.0
http_server depends on pkg path but does not specify it in pubspec.yaml - <a href="https://github.com/sethladd"><img src="https://avatars.githubusercontent.com/u/5479?v=3" align="left" width="96" height="96"hspace="10"></img></a> **Issue by [sethladd](https://github.com/sethladd)** _Originally opened as dart-lang/sdk#14499_ ---- simply declaring http_server as a dependency is not enough. workaround for now: also depend on path.
non_test
http server depends on pkg path but does not specify it in pubspec yaml issue by originally opened as dart lang sdk simply declaring http server as a dependency is not enough workaround for now also depend on path
0
79,289
7,703,855,151
IssuesEvent
2018-05-21 09:57:22
skycoin/skycoin
https://api.github.com/repos/skycoin/skycoin
closed
Consider dropping go1.9 support from travis
tests
Reasons: * json has DisallowUnknownFields now, we could possibly use this for integration tests as an extra verification against unexpected API response data [or maybe our response structs do not match up] * rand.Shuffle
1.0
Consider dropping go1.9 support from travis - Reasons: * json has DisallowUnknownFields now, we could possibly use this for integration tests as an extra verification against unexpected API response data [or maybe our response structs do not match up] * rand.Shuffle
test
consider dropping support from travis reasons json has disallowunknownfields now we could possibly use this for integration tests as an extra verification against unexpected api response data rand shuffle
1
218,850
16,772,649,158
IssuesEvent
2021-06-14 16:34:36
thomas-lehmann-private/hyperion
https://api.github.com/repos/thomas-lehmann-private/hyperion
closed
Provide cababilities command for system depending features
documentation enhancement priority:low
Depending on your system you can / can not use things: - Powershell on Windows (if given you also want to know to know the version) - Batch on Windows - Groovy (embedded - you want to know the version) - JShell (embedded - you want to know the version) - Docker (not yet implemented, all platforms, if given you want to know the version) - Kotlin (not yet implemented, embedded, you want to know the version)
1.0
Provide cababilities command for system depending features - Depending on your system you can / can not use things: - Powershell on Windows (if given you also want to know to know the version) - Batch on Windows - Groovy (embedded - you want to know the version) - JShell (embedded - you want to know the version) - Docker (not yet implemented, all platforms, if given you want to know the version) - Kotlin (not yet implemented, embedded, you want to know the version)
non_test
provide cababilities command for system depending features depending on your system you can can not use things powershell on windows if given you also want to know to know the version batch on windows groovy embedded you want to know the version jshell embedded you want to know the version docker not yet implemented all platforms if given you want to know the version kotlin not yet implemented embedded you want to know the version
0
2,328
24,711,085,152
IssuesEvent
2022-10-20 00:51:59
ppy/osu
https://api.github.com/repos/ppy/osu
closed
Game crashing after map loaded from the editor.
type:online type:reliability
### Type Crash to desktop ### Bug description I was using osu! **without an internet connection** and wanted to play some maps, I got to the song selector and loaded a map, but **it was taking a lot of time**, so I canceled the map and reloaded another one. The same thing happened, so I loaded another map in the editor to see if there was something wrong, but there wasn't. So I played the map from the editor this time, everything was going well, but after 3/4 seconds of playing, **the game froze**, but the audio was still playing. 5 seconds later **the screen went black**, but the audio was still playing. 10 seconds after that, **the audio stopped playing**, and the usual _"osu!lazer has stopped working."_ popped up, but I was waiting, then **the game played audio again**, but the screen was still black. **The game crashed** after like **3 seconds**. This stopped happening after I wrote this and I still don't know what caused this reactive behavior. ### Screenshots or videos I don't have any photos/videos that show this in action. ### Version osu! 2022.1008.2-lazer ### Logs I'm going to send them all, but I found that runtime is most interesting... [database.log](https://github.com/ppy/osu/files/9785324/database.log) [input.log](https://github.com/ppy/osu/files/9785326/input.log) [network.log](https://github.com/ppy/osu/files/9785327/network.log) [performance.log](https://github.com/ppy/osu/files/9785328/performance.log) [runtime.log](https://github.com/ppy/osu/files/9785329/runtime.log) [updater.log](https://github.com/ppy/osu/files/9785330/updater.log)
True
Game crashing after map loaded from the editor. - ### Type Crash to desktop ### Bug description I was using osu! **without an internet connection** and wanted to play some maps, I got to the song selector and loaded a map, but **it was taking a lot of time**, so I canceled the map and reloaded another one. The same thing happened, so I loaded another map in the editor to see if there was something wrong, but there wasn't. So I played the map from the editor this time, everything was going well, but after 3/4 seconds of playing, **the game froze**, but the audio was still playing. 5 seconds later **the screen went black**, but the audio was still playing. 10 seconds after that, **the audio stopped playing**, and the usual _"osu!lazer has stopped working."_ popped up, but I was waiting, then **the game played audio again**, but the screen was still black. **The game crashed** after like **3 seconds**. This stopped happening after I wrote this and I still don't know what caused this reactive behavior. ### Screenshots or videos I don't have any photos/videos that show this in action. ### Version osu! 2022.1008.2-lazer ### Logs I'm going to send them all, but I found that runtime is most interesting... [database.log](https://github.com/ppy/osu/files/9785324/database.log) [input.log](https://github.com/ppy/osu/files/9785326/input.log) [network.log](https://github.com/ppy/osu/files/9785327/network.log) [performance.log](https://github.com/ppy/osu/files/9785328/performance.log) [runtime.log](https://github.com/ppy/osu/files/9785329/runtime.log) [updater.log](https://github.com/ppy/osu/files/9785330/updater.log)
non_test
game crashing after map loaded from the editor type crash to desktop bug description i was using osu without an internet connection and wanted to play some maps i got to the song selector and loaded a map but it was taking a lot of time so i canceled the map and reloaded another one the same thing happened so i loaded another map in the editor to see if there was something wrong but there wasn t so i played the map from the editor this time everything was going well but after seconds of playing the game froze but the audio was still playing seconds later the screen went black but the audio was still playing seconds after that the audio stopped playing and the usual osu lazer has stopped working popped up but i was waiting then the game played audio again but the screen was still black the game crashed after like seconds this stopped happening after i wrote this and i still don t know what caused this reactive behavior screenshots or videos i don t have any photos videos that show this in action version osu lazer logs i m going to send them all but i found that runtime is most interesting
0
310,355
26,711,721,485
IssuesEvent
2023-01-28 01:28:32
opentibiabr/canary
https://api.github.com/repos/opentibiabr/canary
closed
server error when trying to open/close fence gate/door (data-canary)
Type: Bug Priority: Medium Status: Pending Test
### Priority Medium ### Area - [X] Datapack - [ ] Source - [ ] Map - [ ] Other ### What happened? datapack em questรฃo.: data-canary Ao colocar portas e portas de cerca no mapa (custom) e tentar abrir com um char o server retorna um erro id porta aberta 6893 (abri pelo editor) `scriptInterface: [Scripts Interface] scriptId: [C:\Users\jrcar\Desktop\NoName\data-canary\scripts\actions\door\key_door.lua:callback] timerEvent: [] callbackId:[] function: [] error [...top\NoName\data-canary\scripts\actions\door\key_door.lua:47: attempt to call field 'isInsideDoor' (a nil value) stack traceback: [C]: in function 'isInsideDoor' ...top\NoName\data-canary\scripts\actions\door\key_door.lua:47: in function <...top\NoName\data-canary\scripts\actions\door\key_door.lua:31>]` e portรฃo de cerquinha id da cerquinha aberta 2180 `scriptInterface: [Scripts Interface] scriptId: [C:\Users\jrcar\Desktop\NoName\data-canary\scripts\actions\door\custom_door.lua:callback] timerEvent: [] callbackId:[] function: [] error [...\NoName\data-canary\scripts\actions\door\custom_door.lua:15: attempt to call field 'isInsideDoor' (a nil value) stack traceback: [C]: in function 'isInsideDoor' ...\NoName\data-canary\scripts\actions\door\custom_door.lua:15: in function <...\NoName\data-canary\scripts\actions\door\custom_door.lua:14>]` Build para windows usando a versรฃo atual (de hoje no caso) Testado tambรฉm pelo Majesty para confirmaรงรฃo no erro do data-canary Caso hajam mais problemas advindos do uso do data-canary estarei comentando aqui para evitar ficar abrindo outros issues. Qualquer dรบvida ou pedido de teste especรญfico basta estar deixando aqui. Grato ### What OS are you seeing the problem on? Windows ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
1.0
server error when trying to open/close fence gate/door (data-canary) - ### Priority Medium ### Area - [X] Datapack - [ ] Source - [ ] Map - [ ] Other ### What happened? datapack em questรฃo.: data-canary Ao colocar portas e portas de cerca no mapa (custom) e tentar abrir com um char o server retorna um erro id porta aberta 6893 (abri pelo editor) `scriptInterface: [Scripts Interface] scriptId: [C:\Users\jrcar\Desktop\NoName\data-canary\scripts\actions\door\key_door.lua:callback] timerEvent: [] callbackId:[] function: [] error [...top\NoName\data-canary\scripts\actions\door\key_door.lua:47: attempt to call field 'isInsideDoor' (a nil value) stack traceback: [C]: in function 'isInsideDoor' ...top\NoName\data-canary\scripts\actions\door\key_door.lua:47: in function <...top\NoName\data-canary\scripts\actions\door\key_door.lua:31>]` e portรฃo de cerquinha id da cerquinha aberta 2180 `scriptInterface: [Scripts Interface] scriptId: [C:\Users\jrcar\Desktop\NoName\data-canary\scripts\actions\door\custom_door.lua:callback] timerEvent: [] callbackId:[] function: [] error [...\NoName\data-canary\scripts\actions\door\custom_door.lua:15: attempt to call field 'isInsideDoor' (a nil value) stack traceback: [C]: in function 'isInsideDoor' ...\NoName\data-canary\scripts\actions\door\custom_door.lua:15: in function <...\NoName\data-canary\scripts\actions\door\custom_door.lua:14>]` Build para windows usando a versรฃo atual (de hoje no caso) Testado tambรฉm pelo Majesty para confirmaรงรฃo no erro do data-canary Caso hajam mais problemas advindos do uso do data-canary estarei comentando aqui para evitar ficar abrindo outros issues. Qualquer dรบvida ou pedido de teste especรญfico basta estar deixando aqui. Grato ### What OS are you seeing the problem on? Windows ### Code of Conduct - [X] I agree to follow this project's Code of Conduct
test
server error when trying to open close fence gate door data canary priority medium area datapack source map other what happened datapack em questรฃo data canary ao colocar portas e portas de cerca no mapa custom e tentar abrir com um char o server retorna um erro id porta aberta abri pelo editor scriptinterface scriptid timerevent callbackid function error top noname data canary scripts actions door key door lua attempt to call field isinsidedoor a nil value stack traceback in function isinsidedoor top noname data canary scripts actions door key door lua in function e portรฃo de cerquinha id da cerquinha aberta scriptinterface scriptid timerevent callbackid function error noname data canary scripts actions door custom door lua attempt to call field isinsidedoor a nil value stack traceback in function isinsidedoor noname data canary scripts actions door custom door lua in function build para windows usando a versรฃo atual de hoje no caso testado tambรฉm pelo majesty para confirmaรงรฃo no erro do data canary caso hajam mais problemas advindos do uso do data canary estarei comentando aqui para evitar ficar abrindo outros issues qualquer dรบvida ou pedido de teste especรญfico basta estar deixando aqui grato what os are you seeing the problem on windows code of conduct i agree to follow this project s code of conduct
1
2,188
4,939,386,445
IssuesEvent
2016-11-29 14:15:13
metarhia/Impress
https://api.github.com/repos/metarhia/Impress
closed
Use the standalone JSTP library
compatibility enhancement
A new, more robust and flexible, implementation of JSTP is being actively developed in [its own repo](https://github.com/metarhia/JSTP). As soon as it is ready it will replace the current implementation as a dependency. The library will provide hooks and events for servers like Impress to add their own custom logic for logging, authentication, API integration etc.
True
Use the standalone JSTP library - A new, more robust and flexible, implementation of JSTP is being actively developed in [its own repo](https://github.com/metarhia/JSTP). As soon as it is ready it will replace the current implementation as a dependency. The library will provide hooks and events for servers like Impress to add their own custom logic for logging, authentication, API integration etc.
non_test
use the standalone jstp library a new more robust and flexible implementation of jstp is being actively developed in as soon as it is ready it will replace the current implementation as a dependency the library will provide hooks and events for servers like impress to add their own custom logic for logging authentication api integration etc
0
586,197
17,572,406,410
IssuesEvent
2021-08-15 00:41:41
RootGC/JethJS
https://api.github.com/repos/RootGC/JethJS
closed
Alguns comandos estรฃo sem suporte para usuรกrios com GIF de perfil mostrando apenas imagem estรกtica no embed
Status: On Hold Priority: Low Module: Jeth (Discord) โ€โ€โœ… accepted
Precisa-se verificar os arquivos do bot e encontrar os comandos que ainda nรฃo suportam imagens gifs dos usuรกrios em .avatarURL() {EMBED}
1.0
Alguns comandos estรฃo sem suporte para usuรกrios com GIF de perfil mostrando apenas imagem estรกtica no embed - Precisa-se verificar os arquivos do bot e encontrar os comandos que ainda nรฃo suportam imagens gifs dos usuรกrios em .avatarURL() {EMBED}
non_test
alguns comandos estรฃo sem suporte para usuรกrios com gif de perfil mostrando apenas imagem estรกtica no embed precisa se verificar os arquivos do bot e encontrar os comandos que ainda nรฃo suportam imagens gifs dos usuรกrios em avatarurl embed
0
103,662
8,925,312,764
IssuesEvent
2019-01-21 22:08:54
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
closed
roachtest: import/tpcc/warehouses=1000/nodes=32 failed
C-test-failure O-robot
SHA: https://github.com/cockroachdb/cockroach/commits/89ce71e9b733df3855b370aa20bb809db3d4362c Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=import/tpcc/warehouses=1000/nodes=32 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1103150&tab=buildLog ``` The test failed on provisional_201901161553_v2.1.4: test.go:727,cluster.go:997,import.go:34,import.go:66: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod put teamcity-1103150-import-tpcc-warehouses-1000-nodes-32 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/workload ./workload returned: stderr: stdout: teamcity-1103150-import-tpcc-warehouses-1000-nodes-32: putting (dist) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/workload ./workload ....... 1: done 2: done 3: done 4: done 5: done 6: done 7: done 8: done 9: done 10: done 11: done 12: ~ scp -r -C -o StrictHostKeyChecking=no -i /root/.ssh/id_rsa -i /root/.ssh/google_compute_engine root@104.197.211.92:./workload root@35.193.225.59:./workload ssh_exchange_identification: Connection closed by remote host : exit status 1 13: done 14: done 15: done 16: done 17: done 18: done 19: done 20: done 21: done 22: done 23: done 24: done 25: done 26: done 27: done 28: done 29: done 30: done 31: done 32: done I190121 17:18:20.288179 1 cluster_synced.go:846 put /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/workload failed : exit status 1 ```
1.0
roachtest: import/tpcc/warehouses=1000/nodes=32 failed - SHA: https://github.com/cockroachdb/cockroach/commits/89ce71e9b733df3855b370aa20bb809db3d4362c Parameters: To repro, try: ``` # Don't forget to check out a clean suitable branch and experiment with the # stress invocation until the desired results present themselves. For example, # using stress instead of stressrace and passing the '-p' stressflag which # controls concurrency. ./scripts/gceworker.sh start && ./scripts/gceworker.sh mosh cd ~/go/src/github.com/cockroachdb/cockroach && \ stdbuf -oL -eL \ make stressrace TESTS=import/tpcc/warehouses=1000/nodes=32 PKG=roachtest TESTTIMEOUT=5m STRESSFLAGS='-maxtime 20m -timeout 10m' 2>&1 | tee /tmp/stress.log ``` Failed test: https://teamcity.cockroachdb.com/viewLog.html?buildId=1103150&tab=buildLog ``` The test failed on provisional_201901161553_v2.1.4: test.go:727,cluster.go:997,import.go:34,import.go:66: /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod put teamcity-1103150-import-tpcc-warehouses-1000-nodes-32 /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/workload ./workload returned: stderr: stdout: teamcity-1103150-import-tpcc-warehouses-1000-nodes-32: putting (dist) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/workload ./workload ....... 1: done 2: done 3: done 4: done 5: done 6: done 7: done 8: done 9: done 10: done 11: done 12: ~ scp -r -C -o StrictHostKeyChecking=no -i /root/.ssh/id_rsa -i /root/.ssh/google_compute_engine root@104.197.211.92:./workload root@35.193.225.59:./workload ssh_exchange_identification: Connection closed by remote host : exit status 1 13: done 14: done 15: done 16: done 17: done 18: done 19: done 20: done 21: done 22: done 23: done 24: done 25: done 26: done 27: done 28: done 29: done 30: done 31: done 32: done I190121 17:18:20.288179 1 cluster_synced.go:846 put /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/workload failed : exit status 1 ```
test
roachtest import tpcc warehouses nodes failed sha parameters to repro try don t forget to check out a clean suitable branch and experiment with the stress invocation until the desired results present themselves for example using stress instead of stressrace and passing the p stressflag which controls concurrency scripts gceworker sh start scripts gceworker sh mosh cd go src github com cockroachdb cockroach stdbuf ol el make stressrace tests import tpcc warehouses nodes pkg roachtest testtimeout stressflags maxtime timeout tee tmp stress log failed test the test failed on provisional test go cluster go import go import go home agent work go src github com cockroachdb cockroach bin roachprod put teamcity import tpcc warehouses nodes home agent work go src github com cockroachdb cockroach bin workload workload returned stderr stdout teamcity import tpcc warehouses nodes putting dist home agent work go src github com cockroachdb cockroach bin workload workload done done done done done done done done done done done scp r c o stricthostkeychecking no i root ssh id rsa i root ssh google compute engine root workload root workload ssh exchange identification connection closed by remote host exit status done done done done done done done done done done done done done done done done done done done done cluster synced go put home agent work go src github com cockroachdb cockroach bin workload failed exit status
1
306,591
9,396,993,687
IssuesEvent
2019-04-08 08:41:26
Fabian-Sommer/HeroesLounge
https://api.github.com/repos/Fabian-Sommer/HeroesLounge
opened
Speed up match saving in matchmaking
high priority
Currently, saving matches is done one by one, causing it to take multiple minutes and scratching php timelimit. Bulk insertion could help speed this up.
1.0
Speed up match saving in matchmaking - Currently, saving matches is done one by one, causing it to take multiple minutes and scratching php timelimit. Bulk insertion could help speed this up.
non_test
speed up match saving in matchmaking currently saving matches is done one by one causing it to take multiple minutes and scratching php timelimit bulk insertion could help speed this up
0
132,087
5,169,494,821
IssuesEvent
2017-01-18 01:11:01
kubernetes/kubernetes
https://api.github.com/repos/kubernetes/kubernetes
reopened
Document scheduler jargon documentation (from design/podaffinity.md) to (docs/devel/scheduler*)
priority/important-soon sig/scheduling team/control-plane (deprecated - do not use)
_Problem_ There is alot of jargon in the scheduler code, some examples: - `hardPodAffinityWeight` == RequiredDuringSchedulingScheduling - `softPod...` == "PreferredDuringScheduling pod affinity." - Symmetry, a major part of the affinity algorithm implementation, is also undocumented in docs/devel/scheduler. Reading podaffinity along side the scheduler docs + the code gives a good picture. _Solution_ Since the affinity, symmetry, hard/soft pod stuff is now in the production code, lets migrate out of the proposal (or maybe duplicate if proposals are meant to live forever), and Lets converge podaffinity proposal into the first class docs/devel/scheduler\* area. cc @kubernetes/sig-scheduling
1.0
Document scheduler jargon documentation (from design/podaffinity.md) to (docs/devel/scheduler*) - _Problem_ There is alot of jargon in the scheduler code, some examples: - `hardPodAffinityWeight` == RequiredDuringSchedulingScheduling - `softPod...` == "PreferredDuringScheduling pod affinity." - Symmetry, a major part of the affinity algorithm implementation, is also undocumented in docs/devel/scheduler. Reading podaffinity along side the scheduler docs + the code gives a good picture. _Solution_ Since the affinity, symmetry, hard/soft pod stuff is now in the production code, lets migrate out of the proposal (or maybe duplicate if proposals are meant to live forever), and Lets converge podaffinity proposal into the first class docs/devel/scheduler\* area. cc @kubernetes/sig-scheduling
non_test
document scheduler jargon documentation from design podaffinity md to docs devel scheduler problem there is alot of jargon in the scheduler code some examples hardpodaffinityweight requiredduringschedulingscheduling softpod preferredduringscheduling pod affinity symmetry a major part of the affinity algorithm implementation is also undocumented in docs devel scheduler reading podaffinity along side the scheduler docs the code gives a good picture solution since the affinity symmetry hard soft pod stuff is now in the production code lets migrate out of the proposal or maybe duplicate if proposals are meant to live forever and lets converge podaffinity proposal into the first class docs devel scheduler area cc kubernetes sig scheduling
0
25,714
4,417,991,577
IssuesEvent
2016-08-15 08:55:32
OpenMS/OpenMS
https://api.github.com/repos/OpenMS/OpenMS
closed
AASequence: parsing prefers C-terminal mod. over residue mod.
defect
When parsing amino acid sequences with modifications using `AASequence::fromString`, there may be ambiguities in the interpretation of a mod. at the end of the sequence: Is it a mod. of the last residue or of the C-terminus? (See also #1792.) OpenMS currently prefers the second interpretation, i.e. it will attribute a mod. to the C-terminus if there is a possible C-term specificity. I believe this choice is suboptimal; it would be better to attribute the mod. to the residue, if the residue can accept that mod., as this seems the more likely scenario. Furthermore, OpenMS currently doesn't check whether there is a matching residue specificity for a C-terminal mod. For example, in the peptide sequence "PEPM(Oxidation)", the "Oxidation" will be considered as C-terminal, but the only matching mod. is "Oxidation (C-term G)", which doesn't match the "M" in the sequence. Nevertheless, this mod. is assigned, as revealed by `ResidueModification::getFullId`.
1.0
AASequence: parsing prefers C-terminal mod. over residue mod. - When parsing amino acid sequences with modifications using `AASequence::fromString`, there may be ambiguities in the interpretation of a mod. at the end of the sequence: Is it a mod. of the last residue or of the C-terminus? (See also #1792.) OpenMS currently prefers the second interpretation, i.e. it will attribute a mod. to the C-terminus if there is a possible C-term specificity. I believe this choice is suboptimal; it would be better to attribute the mod. to the residue, if the residue can accept that mod., as this seems the more likely scenario. Furthermore, OpenMS currently doesn't check whether there is a matching residue specificity for a C-terminal mod. For example, in the peptide sequence "PEPM(Oxidation)", the "Oxidation" will be considered as C-terminal, but the only matching mod. is "Oxidation (C-term G)", which doesn't match the "M" in the sequence. Nevertheless, this mod. is assigned, as revealed by `ResidueModification::getFullId`.
non_test
aasequence parsing prefers c terminal mod over residue mod when parsing amino acid sequences with modifications using aasequence fromstring there may be ambiguities in the interpretation of a mod at the end of the sequence is it a mod of the last residue or of the c terminus see also openms currently prefers the second interpretation i e it will attribute a mod to the c terminus if there is a possible c term specificity i believe this choice is suboptimal it would be better to attribute the mod to the residue if the residue can accept that mod as this seems the more likely scenario furthermore openms currently doesn t check whether there is a matching residue specificity for a c terminal mod for example in the peptide sequence pepm oxidation the oxidation will be considered as c terminal but the only matching mod is oxidation c term g which doesn t match the m in the sequence nevertheless this mod is assigned as revealed by residuemodification getfullid
0
195,953
14,788,420,179
IssuesEvent
2021-01-12 09:11:37
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96 failed
C-test-failure O-roachtest O-robot branch-release-20.1 release-blocker
[(roachtest).overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2574842&tab=buildLog) on [release-20.1@e395c0c7c48a279334f0e94dfb7030a3eafa093f](https://github.com/cockroachdb/cockroach/commits/e395c0c7c48a279334f0e94dfb7030a3eafa093f): ``` | main.(*monitor).Go.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2606 | github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (2) 2 safe details enclosed Wraps: (3) output in run_091106.426_n4_workload_run_querybench Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2574842-1610435721-87-n4cpu8:4 -- ./workload run querybench --db tpcc --tolerate-errors=t --concurrency=96 --query-file queries.sql --histograms=perf/stats.json --ramp=2m0s --duration=3m0s {pgurl:1-3} returned | stderr: | bash: line 1: 4604 Illegal instruction (core dumped) bash -c "./workload run querybench --db tpcc --tolerate-errors=t --concurrency=96 --query-file queries.sql --histograms=perf/stats.json --ramp=2m0s --duration=3m0s 'postgres://root@10.128.0.122:26257?sslmode=disable' 'postgres://root@10.128.0.113:26257?sslmode=disable' 'postgres://root@10.128.0.101:26257?sslmode=disable'" | Error: COMMAND_PROBLEM: exit status 132 | (1) COMMAND_PROBLEM | Wraps: (2) Node 4. Command with error: | | ``` | | ./workload run querybench --db tpcc --tolerate-errors=t --concurrency=96 --query-file queries.sql --histograms=perf/stats.json --ramp=2m0s --duration=3m0s {pgurl:1-3} | | ``` | Wraps: (3) exit status 132 | Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError | | stdout: Wraps: (5) exit status 20 Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError cluster.go:2628,overload_tpcc_olap.go:71,test_runner.go:749: monitor failure: monitor task failed: t.Fatal() was called (1) attached stack trace | main.(*monitor).WaitE | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2616 | main.(*monitor).Wait | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2624 | main.tpccOLAPSpec.run | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/overload_tpcc_olap.go:71 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749 Wraps: (2) monitor failure Wraps: (3) attached stack trace | main.(*monitor).wait.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2672 Wraps: (4) monitor task failed Wraps: (5) attached stack trace | main.init | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2586 | runtime.doInit | /usr/local/go/src/runtime/proc.go:5652 | runtime.main | /usr/local/go/src/runtime/proc.go:191 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (6) t.Fatal() was called Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96](https://teamcity.cockroachdb.com/viewLog.html?buildId=2574842&tab=artifacts#/overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96) Related: - #56359 roachtest: overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aoverload%2Ftpcc_olap%2Fnodes%3D3%2Fcpu%3D8%2Fw%3D50%2Fc%3D96.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96 failed - [(roachtest).overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96 failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2574842&tab=buildLog) on [release-20.1@e395c0c7c48a279334f0e94dfb7030a3eafa093f](https://github.com/cockroachdb/cockroach/commits/e395c0c7c48a279334f0e94dfb7030a3eafa093f): ``` | main.(*monitor).Go.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2606 | github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup.(*Group).Go.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/golang.org/x/sync/errgroup/errgroup.go:57 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (2) 2 safe details enclosed Wraps: (3) output in run_091106.426_n4_workload_run_querybench Wraps: (4) /home/agent/work/.go/src/github.com/cockroachdb/cockroach/bin/roachprod run teamcity-2574842-1610435721-87-n4cpu8:4 -- ./workload run querybench --db tpcc --tolerate-errors=t --concurrency=96 --query-file queries.sql --histograms=perf/stats.json --ramp=2m0s --duration=3m0s {pgurl:1-3} returned | stderr: | bash: line 1: 4604 Illegal instruction (core dumped) bash -c "./workload run querybench --db tpcc --tolerate-errors=t --concurrency=96 --query-file queries.sql --histograms=perf/stats.json --ramp=2m0s --duration=3m0s 'postgres://root@10.128.0.122:26257?sslmode=disable' 'postgres://root@10.128.0.113:26257?sslmode=disable' 'postgres://root@10.128.0.101:26257?sslmode=disable'" | Error: COMMAND_PROBLEM: exit status 132 | (1) COMMAND_PROBLEM | Wraps: (2) Node 4. Command with error: | | ``` | | ./workload run querybench --db tpcc --tolerate-errors=t --concurrency=96 --query-file queries.sql --histograms=perf/stats.json --ramp=2m0s --duration=3m0s {pgurl:1-3} | | ``` | Wraps: (3) exit status 132 | Error types: (1) errors.Cmd (2) *hintdetail.withDetail (3) *exec.ExitError | | stdout: Wraps: (5) exit status 20 Error types: (1) *withstack.withStack (2) *safedetails.withSafeDetails (3) *errutil.withMessage (4) *main.withCommandDetails (5) *exec.ExitError cluster.go:2628,overload_tpcc_olap.go:71,test_runner.go:749: monitor failure: monitor task failed: t.Fatal() was called (1) attached stack trace | main.(*monitor).WaitE | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2616 | main.(*monitor).Wait | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2624 | main.tpccOLAPSpec.run | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/overload_tpcc_olap.go:71 | main.(*testRunner).runTest.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/test_runner.go:749 Wraps: (2) monitor failure Wraps: (3) attached stack trace | main.(*monitor).wait.func2 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2672 Wraps: (4) monitor task failed Wraps: (5) attached stack trace | main.init | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachtest/cluster.go:2586 | runtime.doInit | /usr/local/go/src/runtime/proc.go:5652 | runtime.main | /usr/local/go/src/runtime/proc.go:191 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (6) t.Fatal() was called Error types: (1) *withstack.withStack (2) *errutil.withMessage (3) *withstack.withStack (4) *errutil.withMessage (5) *withstack.withStack (6) *errors.errorString ``` <details><summary>More</summary><p> Artifacts: [/overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96](https://teamcity.cockroachdb.com/viewLog.html?buildId=2574842&tab=artifacts#/overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96) Related: - #56359 roachtest: overload/tpcc_olap/nodes=3/cpu=8/w=50/c=96 failed [C-test-failure](https://api.github.com/repos/cockroachdb/cockroach/labels/C-test-failure) [O-roachtest](https://api.github.com/repos/cockroachdb/cockroach/labels/O-roachtest) [O-robot](https://api.github.com/repos/cockroachdb/cockroach/labels/O-robot) [branch-master](https://api.github.com/repos/cockroachdb/cockroach/labels/branch-master) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Aoverload%2Ftpcc_olap%2Fnodes%3D3%2Fcpu%3D8%2Fw%3D50%2Fc%3D96.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest overload tpcc olap nodes cpu w c failed on main monitor go home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go github com cockroachdb cockroach vendor golang org x sync errgroup group go home agent work go src github com cockroachdb cockroach vendor golang org x sync errgroup errgroup go runtime goexit usr local go src runtime asm s wraps safe details enclosed wraps output in run workload run querybench wraps home agent work go src github com cockroachdb cockroach bin roachprod run teamcity workload run querybench db tpcc tolerate errors t concurrency query file queries sql histograms perf stats json ramp duration pgurl returned stderr bash line illegal instruction core dumped bash c workload run querybench db tpcc tolerate errors t concurrency query file queries sql histograms perf stats json ramp duration postgres root sslmode disable postgres root sslmode disable postgres root sslmode disable error command problem exit status command problem wraps node command with error workload run querybench db tpcc tolerate errors t concurrency query file queries sql histograms perf stats json ramp duration pgurl wraps exit status error types errors cmd hintdetail withdetail exec exiterror stdout wraps exit status error types withstack withstack safedetails withsafedetails errutil withmessage main withcommanddetails exec exiterror cluster go overload tpcc olap go test runner go monitor failure monitor task failed t fatal was called attached stack trace main monitor waite home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go main tpccolapspec run home agent work go src github com cockroachdb cockroach pkg cmd roachtest overload tpcc olap go main testrunner runtest home agent work go src github com cockroachdb cockroach pkg cmd roachtest test runner go wraps monitor failure wraps attached stack trace main monitor wait home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go wraps monitor task failed wraps attached stack trace main init home agent work go src github com cockroachdb cockroach pkg cmd roachtest cluster go runtime doinit usr local go src runtime proc go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps t fatal was called error types withstack withstack errutil withmessage withstack withstack errutil withmessage withstack withstack errors errorstring more artifacts related roachtest overload tpcc olap nodes cpu w c failed powered by
1
745,219
25,975,122,466
IssuesEvent
2022-12-19 14:19:37
GoogleCloudPlatform/python-docs-samples
https://api.github.com/repos/GoogleCloudPlatform/python-docs-samples
closed
people-and-planet-ai.geospatial-classification.e2e_test: test_predict failed
priority: p1 type: bug samples flakybot: issue
Note: #8732 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 112506ea590242a3e46461c4622e111da02cdb54 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e64079e3-607a-4f48-b6bb-4b326781ced4), [Sponge](http://sponge2/e64079e3-607a-4f48-b6bb-4b326781ced4) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.RESOURCE_EXHAUSTED details = "Quota exceeded for quota metric 'Job or LRO submission requests' and limit 'Job or LRO submission requests per minute per region' of service 'aiplatform.googleapis.com' for consumer 'project_number:1012616486416'." debug_error_string = "UNKNOWN:Error received from peer ipv4:173.194.203.95:443 {created_time:"2022-12-18T14:09:31.358582056+00:00", grpc_status:8, grpc_message:"Quota exceeded for quota metric \'Job or LRO submission requests\' and limit \'Job or LRO submission requests per minute per region\' of service \'aiplatform.googleapis.com\' for consumer \'project_number:1012616486416\'."}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/people-and-planet-ai/geospatial-classification/e2e_test.py", line 293, in train_model job.run( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform/training_jobs.py", line 2114, in run return self._run( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform/base.py", line 750, in wrapper return method(*args, **kwargs) File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform/training_jobs.py", line 2384, in _run model = self._run_job( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform/training_jobs.py", line 749, in _run_job training_pipeline = self.api_client.create_training_pipeline( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform_v1/services/pipeline_service/client.py", line 734, in create_training_pipeline response = rpc( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/gapic_v1/method.py", line 154, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.ResourceExhausted: 429 Quota exceeded for quota metric 'Job or LRO submission requests' and limit 'Job or LRO submission requests per minute per region' of service 'aiplatform.googleapis.com' for consumer 'project_number:1012616486416'. [reason: "RATE_LIMIT_EXCEEDED" domain: "googleapis.com" metadata { key: "consumer" value: "projects/1012616486416" } metadata { key: "quota_limit" value: "JobOrLroSubmissionRequestsPerMinutePerProjectPerRegion" } metadata { key: "quota_limit_value" value: "60" } metadata { key: "quota_location" value: "us-central1" } metadata { key: "quota_metric" value: "aiplatform.googleapis.com/job_or_lro_submission_requests" } metadata { key: "service" value: "aiplatform.googleapis.com" } , links { description: "Request a higher quota limit." url: "https://cloud.google.com/docs/quota#requesting_higher_quota" } ]</pre></details>
1.0
people-and-planet-ai.geospatial-classification.e2e_test: test_predict failed - Note: #8732 was also for this test, but it was closed more than 10 days ago. So, I didn't mark it flaky. ---- commit: 112506ea590242a3e46461c4622e111da02cdb54 buildURL: [Build Status](https://source.cloud.google.com/results/invocations/e64079e3-607a-4f48-b6bb-4b326781ced4), [Sponge](http://sponge2/e64079e3-607a-4f48-b6bb-4b326781ced4) status: failed <details><summary>Test output</summary><br><pre>Traceback (most recent call last): File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 72, in error_remapped_callable return callable_(*args, **kwargs) File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 946, in __call__ return _end_unary_response_blocking(state, call, False, None) File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/grpc/_channel.py", line 849, in _end_unary_response_blocking raise _InactiveRpcError(state) grpc._channel._InactiveRpcError: <_InactiveRpcError of RPC that terminated with: status = StatusCode.RESOURCE_EXHAUSTED details = "Quota exceeded for quota metric 'Job or LRO submission requests' and limit 'Job or LRO submission requests per minute per region' of service 'aiplatform.googleapis.com' for consumer 'project_number:1012616486416'." debug_error_string = "UNKNOWN:Error received from peer ipv4:173.194.203.95:443 {created_time:"2022-12-18T14:09:31.358582056+00:00", grpc_status:8, grpc_message:"Quota exceeded for quota metric \'Job or LRO submission requests\' and limit \'Job or LRO submission requests per minute per region\' of service \'aiplatform.googleapis.com\' for consumer \'project_number:1012616486416\'."}" > The above exception was the direct cause of the following exception: Traceback (most recent call last): File "/workspace/people-and-planet-ai/geospatial-classification/e2e_test.py", line 293, in train_model job.run( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform/training_jobs.py", line 2114, in run return self._run( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform/base.py", line 750, in wrapper return method(*args, **kwargs) File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform/training_jobs.py", line 2384, in _run model = self._run_job( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform/training_jobs.py", line 749, in _run_job training_pipeline = self.api_client.create_training_pipeline( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/cloud/aiplatform_v1/services/pipeline_service/client.py", line 734, in create_training_pipeline response = rpc( File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/gapic_v1/method.py", line 154, in __call__ return wrapped_func(*args, **kwargs) File "/workspace/people-and-planet-ai/geospatial-classification/.nox/py-3-9/lib/python3.9/site-packages/google/api_core/grpc_helpers.py", line 74, in error_remapped_callable raise exceptions.from_grpc_error(exc) from exc google.api_core.exceptions.ResourceExhausted: 429 Quota exceeded for quota metric 'Job or LRO submission requests' and limit 'Job or LRO submission requests per minute per region' of service 'aiplatform.googleapis.com' for consumer 'project_number:1012616486416'. [reason: "RATE_LIMIT_EXCEEDED" domain: "googleapis.com" metadata { key: "consumer" value: "projects/1012616486416" } metadata { key: "quota_limit" value: "JobOrLroSubmissionRequestsPerMinutePerProjectPerRegion" } metadata { key: "quota_limit_value" value: "60" } metadata { key: "quota_location" value: "us-central1" } metadata { key: "quota_metric" value: "aiplatform.googleapis.com/job_or_lro_submission_requests" } metadata { key: "service" value: "aiplatform.googleapis.com" } , links { description: "Request a higher quota limit." url: "https://cloud.google.com/docs/quota#requesting_higher_quota" } ]</pre></details>
non_test
people and planet ai geospatial classification test test predict failed note was also for this test but it was closed more than days ago so i didn t mark it flaky commit buildurl status failed test output traceback most recent call last file workspace people and planet ai geospatial classification nox py lib site packages google api core grpc helpers py line in error remapped callable return callable args kwargs file workspace people and planet ai geospatial classification nox py lib site packages grpc channel py line in call return end unary response blocking state call false none file workspace people and planet ai geospatial classification nox py lib site packages grpc channel py line in end unary response blocking raise inactiverpcerror state grpc channel inactiverpcerror inactiverpcerror of rpc that terminated with status statuscode resource exhausted details quota exceeded for quota metric job or lro submission requests and limit job or lro submission requests per minute per region of service aiplatform googleapis com for consumer project number debug error string unknown error received from peer created time grpc status grpc message quota exceeded for quota metric job or lro submission requests and limit job or lro submission requests per minute per region of service aiplatform googleapis com for consumer project number the above exception was the direct cause of the following exception traceback most recent call last file workspace people and planet ai geospatial classification test py line in train model job run file workspace people and planet ai geospatial classification nox py lib site packages google cloud aiplatform training jobs py line in run return self run file workspace people and planet ai geospatial classification nox py lib site packages google cloud aiplatform base py line in wrapper return method args kwargs file workspace people and planet ai geospatial classification nox py lib site packages google cloud aiplatform training jobs py line in run model self run job file workspace people and planet ai geospatial classification nox py lib site packages google cloud aiplatform training jobs py line in run job training pipeline self api client create training pipeline file workspace people and planet ai geospatial classification nox py lib site packages google cloud aiplatform services pipeline service client py line in create training pipeline response rpc file workspace people and planet ai geospatial classification nox py lib site packages google api core gapic method py line in call return wrapped func args kwargs file workspace people and planet ai geospatial classification nox py lib site packages google api core grpc helpers py line in error remapped callable raise exceptions from grpc error exc from exc google api core exceptions resourceexhausted quota exceeded for quota metric job or lro submission requests and limit job or lro submission requests per minute per region of service aiplatform googleapis com for consumer project number reason rate limit exceeded domain googleapis com metadata key consumer value projects metadata key quota limit value joborlrosubmissionrequestsperminuteperprojectperregion metadata key quota limit value value metadata key quota location value us metadata key quota metric value aiplatform googleapis com job or lro submission requests metadata key service value aiplatform googleapis com links description request a higher quota limit url
0
10,551
8,101,219,347
IssuesEvent
2018-08-12 11:01:00
gravit0/Launcher
https://api.github.com/repos/gravit0/Launcher
closed
ะ˜ะดะตะธ ะฟะพ gradle + native
feature security
ะฏ ะผะพะณัƒ ะฟะพะผะพั‡ัŒ ั ะธะฝั‚ะตะณั€ะฐั†ะธะตะน ั gradle (pr) + ั ะฝะฐั‚ะธะฒะฝะพะน ะฑะธะฑะปะธะพั‚ะตะบะพะน ะฟะพะด win32/64, ะฟะพะด linux ะฟั€ะตะดะปะฐะณะฐัŽ ะฝะฐะฟะธัะฐั‚ัŒ ะฝะตั‡ั‚ะพ ะฒั€ะพะดะต ะทะฐั‰ะธั‚ั‹ ะพั‚ ะธะฝัŠะตะบั†ะธะน (ะธะฑะพ ะฟั€ะพะณั€ะฐะผะผะธัั‚ะฐ ัั‚ะพ ะฝะต ะพัั‚ะฐะฝะพะฒะธั‚ ะฐ ัˆะบะพะปัŒะฝะธะบะฐ ะดะฐ). ะšัั‚ะฐั‚ะธ ะตั‰ั‘ ะฟั€ะพัˆัƒ ะฒั‹ัะปะฐั‚ัŒ ะฝะฐ email ะบะพะฝั‚ะฐะบั‚ะฝั‹ะต ะดะฐะฝะฝั‹ะต (ะฒัั‘ ั‡ั‚ะพ ัƒะณะพะดะฝะพ ะฝะพ ะฝะต github). Discord: ะ—ะฐั…ะฐั€#7395
True
ะ˜ะดะตะธ ะฟะพ gradle + native - ะฏ ะผะพะณัƒ ะฟะพะผะพั‡ัŒ ั ะธะฝั‚ะตะณั€ะฐั†ะธะตะน ั gradle (pr) + ั ะฝะฐั‚ะธะฒะฝะพะน ะฑะธะฑะปะธะพั‚ะตะบะพะน ะฟะพะด win32/64, ะฟะพะด linux ะฟั€ะตะดะปะฐะณะฐัŽ ะฝะฐะฟะธัะฐั‚ัŒ ะฝะตั‡ั‚ะพ ะฒั€ะพะดะต ะทะฐั‰ะธั‚ั‹ ะพั‚ ะธะฝัŠะตะบั†ะธะน (ะธะฑะพ ะฟั€ะพะณั€ะฐะผะผะธัั‚ะฐ ัั‚ะพ ะฝะต ะพัั‚ะฐะฝะพะฒะธั‚ ะฐ ัˆะบะพะปัŒะฝะธะบะฐ ะดะฐ). ะšัั‚ะฐั‚ะธ ะตั‰ั‘ ะฟั€ะพัˆัƒ ะฒั‹ัะปะฐั‚ัŒ ะฝะฐ email ะบะพะฝั‚ะฐะบั‚ะฝั‹ะต ะดะฐะฝะฝั‹ะต (ะฒัั‘ ั‡ั‚ะพ ัƒะณะพะดะฝะพ ะฝะพ ะฝะต github). Discord: ะ—ะฐั…ะฐั€#7395
non_test
ะธะดะตะธ ะฟะพ gradle native ั ะผะพะณัƒ ะฟะพะผะพั‡ัŒ ั ะธะฝั‚ะตะณั€ะฐั†ะธะตะน ั gradle pr ั ะฝะฐั‚ะธะฒะฝะพะน ะฑะธะฑะปะธะพั‚ะตะบะพะน ะฟะพะด ะฟะพะด linux ะฟั€ะตะดะปะฐะณะฐัŽ ะฝะฐะฟะธัะฐั‚ัŒ ะฝะตั‡ั‚ะพ ะฒั€ะพะดะต ะทะฐั‰ะธั‚ั‹ ะพั‚ ะธะฝัŠะตะบั†ะธะน ะธะฑะพ ะฟั€ะพะณั€ะฐะผะผะธัั‚ะฐ ัั‚ะพ ะฝะต ะพัั‚ะฐะฝะพะฒะธั‚ ะฐ ัˆะบะพะปัŒะฝะธะบะฐ ะดะฐ ะบัั‚ะฐั‚ะธ ะตั‰ั‘ ะฟั€ะพัˆัƒ ะฒั‹ัะปะฐั‚ัŒ ะฝะฐ email ะบะพะฝั‚ะฐะบั‚ะฝั‹ะต ะดะฐะฝะฝั‹ะต ะฒัั‘ ั‡ั‚ะพ ัƒะณะพะดะฝะพ ะฝะพ ะฝะต github discord ะทะฐั…ะฐั€
0
348,360
24,912,356,295
IssuesEvent
2022-10-30 01:40:32
AY2223S1-CS2103-F13-3/tp
https://api.github.com/repos/AY2223S1-CS2103-F13-3/tp
closed
[PE-D][Tester A] Add person to meeting syntax unclear
DocumentationBug In the works
Even with the added }} between the names I'm still unable to delete both of the contacts from the first meeting. The UG states that I am able to delete a list of contacts with the syntax [NAMES] so I'm unsure how the command functions. An example could be given, as well as the constraints on how [NAMES] is input should be elaborated ![image.png](https://raw.githubusercontent.com/JordanChua/ped/main/files/b5796de4-1efb-4e0a-a9e8-d8b9ce8ba904.png) <!--session: 1666944084337-fb4076ef-3095-4db5-aabc-323dc432badd--> <!--Version: Web v3.4.4--> ------------- Labels: `severity.High` `type.FunctionalityBug` original: JordanChua/ped#11
1.0
[PE-D][Tester A] Add person to meeting syntax unclear - Even with the added }} between the names I'm still unable to delete both of the contacts from the first meeting. The UG states that I am able to delete a list of contacts with the syntax [NAMES] so I'm unsure how the command functions. An example could be given, as well as the constraints on how [NAMES] is input should be elaborated ![image.png](https://raw.githubusercontent.com/JordanChua/ped/main/files/b5796de4-1efb-4e0a-a9e8-d8b9ce8ba904.png) <!--session: 1666944084337-fb4076ef-3095-4db5-aabc-323dc432badd--> <!--Version: Web v3.4.4--> ------------- Labels: `severity.High` `type.FunctionalityBug` original: JordanChua/ped#11
non_test
add person to meeting syntax unclear even with the added between the names i m still unable to delete both of the contacts from the first meeting the ug states that i am able to delete a list of contacts with the syntax so i m unsure how the command functions an example could be given as well as the constraints on how is input should be elaborated labels severity high type functionalitybug original jordanchua ped
0
206,924
7,122,909,815
IssuesEvent
2018-01-19 13:41:30
imjuangarcia/personal-website
https://api.github.com/repos/imjuangarcia/personal-website
opened
Change iframe video from the course detail page
Content Frontend Priority: High
[Wake up](https://www.youtube.com/watch?v=HBB37gsHJmQ) is a great song, but here should be my introductory video, as soon as it's finished โœ… * [ ] Change it on the atf modal; * [ ] Change it on the information below
1.0
Change iframe video from the course detail page - [Wake up](https://www.youtube.com/watch?v=HBB37gsHJmQ) is a great song, but here should be my introductory video, as soon as it's finished โœ… * [ ] Change it on the atf modal; * [ ] Change it on the information below
non_test
change iframe video from the course detail page is a great song but here should be my introductory video as soon as it s finished โœ… change it on the atf modal change it on the information below
0
66,902
7,025,636,519
IssuesEvent
2017-12-23 13:50:03
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
[CI] SharedClusterSnapshotRestoreIT.testAbortedSnapshotDuringInitDoesNotStart() fails
:Snapshot/Restore jenkins test
Added in #27931, this test fails regularly on CI.
1.0
[CI] SharedClusterSnapshotRestoreIT.testAbortedSnapshotDuringInitDoesNotStart() fails - Added in #27931, this test fails regularly on CI.
test
sharedclustersnapshotrestoreit testabortedsnapshotduringinitdoesnotstart fails added in this test fails regularly on ci
1
260,334
8,208,592,299
IssuesEvent
2018-09-04 02:46:54
kubernetes/kubeadm
https://api.github.com/repos/kubernetes/kubeadm
closed
Remotely callable `kubeadm upgrade plan`
area/upgrades kind/feature lifecycle/rotten priority/backlog
[Feature Request] We would like to show the tables that tells user about potential upgrade options from `kubeadm upgrade plan` command in a webpage. Currently this command needs (i guess) to be run from the master machine for the local dir checks. This means the api server backing the webpage needs to do ssh into the master. Can a lighter version of `upgrade plan` command be supported that skips the local dir checks and only do the parts that checks against kube api server and remote dl-stable urls? This will allow calling this without needing to ssh into master instance.
1.0
Remotely callable `kubeadm upgrade plan` - [Feature Request] We would like to show the tables that tells user about potential upgrade options from `kubeadm upgrade plan` command in a webpage. Currently this command needs (i guess) to be run from the master machine for the local dir checks. This means the api server backing the webpage needs to do ssh into the master. Can a lighter version of `upgrade plan` command be supported that skips the local dir checks and only do the parts that checks against kube api server and remote dl-stable urls? This will allow calling this without needing to ssh into master instance.
non_test
remotely callable kubeadm upgrade plan we would like to show the tables that tells user about potential upgrade options from kubeadm upgrade plan command in a webpage currently this command needs i guess to be run from the master machine for the local dir checks this means the api server backing the webpage needs to do ssh into the master can a lighter version of upgrade plan command be supported that skips the local dir checks and only do the parts that checks against kube api server and remote dl stable urls this will allow calling this without needing to ssh into master instance
0
595,840
18,075,782,582
IssuesEvent
2021-09-21 09:43:20
kubernetes-sigs/cluster-api-provider-aws
https://api.github.com/repos/kubernetes-sigs/cluster-api-provider-aws
opened
v1beta1 Tracking Issue
priority/important-soon area/api
/kind feature **Describe the solution you'd like** * Bump API version to v1beta1 to match Cluster API Core **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] Rough task list - [ ] Bump Controller Runtime to v0.10 - [ ] Refactor v1alpha3 conversion webhooks to move to v1beta1 (do not delete them as in previous releases) - [ ] Add any needed conversion webhooks for v1alpha4 - [ ] Set clusterctl metadata Related Issues: * https://github.com/kubernetes-sigs/cluster-api/issues/5260 * https://github.com/kubernetes-sigs/cluster-api/issues/5262 * https://github.com/kubernetes-sigs/cluster-api/issues/5269 * https://github.com/kubernetes-sigs/cluster-api/issues/5277 **Environment:** - Cluster-api-provider-aws version: - Kubernetes version: (use `kubectl version`): - OS (e.g. from `/etc/os-release`):
1.0
v1beta1 Tracking Issue - /kind feature **Describe the solution you'd like** * Bump API version to v1beta1 to match Cluster API Core **Anything else you would like to add:** [Miscellaneous information that will assist in solving the issue.] Rough task list - [ ] Bump Controller Runtime to v0.10 - [ ] Refactor v1alpha3 conversion webhooks to move to v1beta1 (do not delete them as in previous releases) - [ ] Add any needed conversion webhooks for v1alpha4 - [ ] Set clusterctl metadata Related Issues: * https://github.com/kubernetes-sigs/cluster-api/issues/5260 * https://github.com/kubernetes-sigs/cluster-api/issues/5262 * https://github.com/kubernetes-sigs/cluster-api/issues/5269 * https://github.com/kubernetes-sigs/cluster-api/issues/5277 **Environment:** - Cluster-api-provider-aws version: - Kubernetes version: (use `kubectl version`): - OS (e.g. from `/etc/os-release`):
non_test
tracking issue kind feature describe the solution you d like bump api version to to match cluster api core anything else you would like to add rough task list bump controller runtime to refactor conversion webhooks to move to do not delete them as in previous releases add any needed conversion webhooks for set clusterctl metadata related issues environment cluster api provider aws version kubernetes version use kubectl version os e g from etc os release
0
9,967
4,690,649,576
IssuesEvent
2016-10-11 07:42:59
LLNL/spack
https://api.github.com/repos/LLNL/spack
closed
trilinos recipe: spec['blas'].prefix.lib is wrong for mkl
blas-lapack-scalapack build-error external-packages
I can no longer build `trilinos ^mkl` because `spec['blas'].prefix.lib` does not point to the correct directory for MKL. At least for my installation, the MKL libraries are located at `$MKLROOT/lib/intel64` and not at `$MKLROOT/lib`. I'm not sure how to fix this because I don't know if there is significant variation in directory names for MKL installations. It is simple to modify the trilinos recipe to do something special for MKL. Should I submit a PR with these changes? Does anyone else have experience building trilinos against MKL? Are there other configurations that I should be aware of?
1.0
trilinos recipe: spec['blas'].prefix.lib is wrong for mkl - I can no longer build `trilinos ^mkl` because `spec['blas'].prefix.lib` does not point to the correct directory for MKL. At least for my installation, the MKL libraries are located at `$MKLROOT/lib/intel64` and not at `$MKLROOT/lib`. I'm not sure how to fix this because I don't know if there is significant variation in directory names for MKL installations. It is simple to modify the trilinos recipe to do something special for MKL. Should I submit a PR with these changes? Does anyone else have experience building trilinos against MKL? Are there other configurations that I should be aware of?
non_test
trilinos recipe spec prefix lib is wrong for mkl i can no longer build trilinos mkl because spec prefix lib does not point to the correct directory for mkl at least for my installation the mkl libraries are located at mklroot lib and not at mklroot lib i m not sure how to fix this because i don t know if there is significant variation in directory names for mkl installations it is simple to modify the trilinos recipe to do something special for mkl should i submit a pr with these changes does anyone else have experience building trilinos against mkl are there other configurations that i should be aware of
0
216,095
16,628,960,019
IssuesEvent
2021-06-03 13:21:27
thomas-lehmann-private/hyperion
https://api.github.com/repos/thomas-lehmann-private/hyperion
closed
Provide simple Docker container support
complexity:moderate documentation enhancement priority:normal
* task type: **docker-container** * choosing the image (required) * supporting mapping ports (optional) * supporting mapping of paths (optional) * cleanup of container after execution (default) * possible to keep container (optional) * possible to run in background (optional)
1.0
Provide simple Docker container support - * task type: **docker-container** * choosing the image (required) * supporting mapping ports (optional) * supporting mapping of paths (optional) * cleanup of container after execution (default) * possible to keep container (optional) * possible to run in background (optional)
non_test
provide simple docker container support task type docker container choosing the image required supporting mapping ports optional supporting mapping of paths optional cleanup of container after execution default possible to keep container optional possible to run in background optional
0
349,651
31,818,491,220
IssuesEvent
2023-09-13 22:55:28
rollerozxa/mt-uptime-status
https://api.github.com/repos/rollerozxa/mt-uptime-status
closed
๐Ÿ›‘ wiki.minetest.land (unofficial mirror) is down
status wiki-minetest-land-unofficial-mirror
In [`ed9c6e1`](https://github.com/rollerozxa/mt-uptime-status/commit/ed9c6e135eea4c8f8d28f12fd2d41ead4dbd4fe9 ), wiki.minetest.land (unofficial mirror) (https://wiki.minetest.land/Main_Page) was **down**: - HTTP code: 0 - Response time: 0 ms
1.0
๐Ÿ›‘ wiki.minetest.land (unofficial mirror) is down - In [`ed9c6e1`](https://github.com/rollerozxa/mt-uptime-status/commit/ed9c6e135eea4c8f8d28f12fd2d41ead4dbd4fe9 ), wiki.minetest.land (unofficial mirror) (https://wiki.minetest.land/Main_Page) was **down**: - HTTP code: 0 - Response time: 0 ms
test
๐Ÿ›‘ wiki minetest land unofficial mirror is down in wiki minetest land unofficial mirror was down http code response time ms
1
801,999
28,565,012,150
IssuesEvent
2023-04-21 00:45:40
microsoft/rushstack
https://api.github.com/repos/microsoft/rushstack
closed
[rush] >O(n^2) performance in `rush version`
repro confirmed priority
## Summary When running `rush version --bump` in a monorepo with 877 projects, the `semver.satisfies` check in `PublishUtilties._updateDownstreamDependency` was invoked 80450952 times. ## Repro steps In a large monorepo with a few hundred pending change files, run `rush version --bump`. ## Details From inspection, the version bumping algorithm performs the following steps in a loop: 1) For each package with changes, update the consuming package's `dependencies`, `devDependencies`, and `peerDependencies` fields. Record as a change on said consuming package and recurse to its consumers. For some reason the use of `workspace:*` in a dependency field automatically recurses even if the newly added change does not alter the current package (e.g. because the same significance of change is already present). 1) For each package with changes, if it is part of a lockstep policy, apply the lockstep version change. At no point in this process are the changes memoized. I propose the following alternative process for applying changes: 1) Read all change JSON files and group by package name 1) Filter out any groups that apply to non-existent projects (and optionally delete the change files for said non-existent projects, maybe behind a flag) 1) Build an augmented project graph in which any lockstepped version policy replaces all projects that are part of the policy with a single composite node which corresponds to the policy and to which all change calculations are applied 1) For each node in the augmented graph, perform a memoized depth-first search to determine the final changeType for the node 1) Apply all changes. ## Standard questions Please answer these questions to help us investigate your issue more quickly: | Question | Answer | | -------- | -------- | | `@microsoft/rush` globally installed version? | 5.97.0 | | `rushVersion` from rush.json? | 5.97.0 | | `useWorkspaces` from rush.json? | true| | Operating system? | Linux | | Would you consider contributing a PR? | Yes | | Node.js version (`node -v`)? | 16.19.1 |
1.0
[rush] >O(n^2) performance in `rush version` - ## Summary When running `rush version --bump` in a monorepo with 877 projects, the `semver.satisfies` check in `PublishUtilties._updateDownstreamDependency` was invoked 80450952 times. ## Repro steps In a large monorepo with a few hundred pending change files, run `rush version --bump`. ## Details From inspection, the version bumping algorithm performs the following steps in a loop: 1) For each package with changes, update the consuming package's `dependencies`, `devDependencies`, and `peerDependencies` fields. Record as a change on said consuming package and recurse to its consumers. For some reason the use of `workspace:*` in a dependency field automatically recurses even if the newly added change does not alter the current package (e.g. because the same significance of change is already present). 1) For each package with changes, if it is part of a lockstep policy, apply the lockstep version change. At no point in this process are the changes memoized. I propose the following alternative process for applying changes: 1) Read all change JSON files and group by package name 1) Filter out any groups that apply to non-existent projects (and optionally delete the change files for said non-existent projects, maybe behind a flag) 1) Build an augmented project graph in which any lockstepped version policy replaces all projects that are part of the policy with a single composite node which corresponds to the policy and to which all change calculations are applied 1) For each node in the augmented graph, perform a memoized depth-first search to determine the final changeType for the node 1) Apply all changes. ## Standard questions Please answer these questions to help us investigate your issue more quickly: | Question | Answer | | -------- | -------- | | `@microsoft/rush` globally installed version? | 5.97.0 | | `rushVersion` from rush.json? | 5.97.0 | | `useWorkspaces` from rush.json? | true| | Operating system? | Linux | | Would you consider contributing a PR? | Yes | | Node.js version (`node -v`)? | 16.19.1 |
non_test
o n performance in rush version summary when running rush version bump in a monorepo with projects the semver satisfies check in publishutilties updatedownstreamdependency was invoked times repro steps in a large monorepo with a few hundred pending change files run rush version bump details from inspection the version bumping algorithm performs the following steps in a loop for each package with changes update the consuming package s dependencies devdependencies and peerdependencies fields record as a change on said consuming package and recurse to its consumers for some reason the use of workspace in a dependency field automatically recurses even if the newly added change does not alter the current package e g because the same significance of change is already present for each package with changes if it is part of a lockstep policy apply the lockstep version change at no point in this process are the changes memoized i propose the following alternative process for applying changes read all change json files and group by package name filter out any groups that apply to non existent projects and optionally delete the change files for said non existent projects maybe behind a flag build an augmented project graph in which any lockstepped version policy replaces all projects that are part of the policy with a single composite node which corresponds to the policy and to which all change calculations are applied for each node in the augmented graph perform a memoized depth first search to determine the final changetype for the node apply all changes standard questions please answer these questions to help us investigate your issue more quickly question answer microsoft rush globally installed version rushversion from rush json useworkspaces from rush json true operating system linux would you consider contributing a pr yes node js version node v
0
735,317
25,389,196,057
IssuesEvent
2022-11-22 01:40:15
SlimeVR/SlimeVR-Server
https://api.github.com/repos/SlimeVR/SlimeVR-Server
closed
Interactive skeleton settings and calibration setup
Type: Feature Request Difficulty: Math Heavy Area: Skeletal Model Priority: High
There can be a lot of things to improve calibration (reset) and skeleton settings that will be more precise than pressing reset while standing still, or asking users to input their measurements or align trackers to their body. Here will be a list of ideas. We will need a framework to implement them in the future, and add more. * Bend leg in the knee and raise it up to configure forward axis. The yaw 0 will be on the plane that goes through the angle made between the start and the end of the bend, parallel to the floor. This will eliminate the need to put trackers on exactly facing one of the cardinal sides. * Bend forward to configure the yaw 0 for chest and waist trackers, similar to the leg. * Move foot up and down to configure the yaw 0 for the foot.
1.0
Interactive skeleton settings and calibration setup - There can be a lot of things to improve calibration (reset) and skeleton settings that will be more precise than pressing reset while standing still, or asking users to input their measurements or align trackers to their body. Here will be a list of ideas. We will need a framework to implement them in the future, and add more. * Bend leg in the knee and raise it up to configure forward axis. The yaw 0 will be on the plane that goes through the angle made between the start and the end of the bend, parallel to the floor. This will eliminate the need to put trackers on exactly facing one of the cardinal sides. * Bend forward to configure the yaw 0 for chest and waist trackers, similar to the leg. * Move foot up and down to configure the yaw 0 for the foot.
non_test
interactive skeleton settings and calibration setup there can be a lot of things to improve calibration reset and skeleton settings that will be more precise than pressing reset while standing still or asking users to input their measurements or align trackers to their body here will be a list of ideas we will need a framework to implement them in the future and add more bend leg in the knee and raise it up to configure forward axis the yaw will be on the plane that goes through the angle made between the start and the end of the bend parallel to the floor this will eliminate the need to put trackers on exactly facing one of the cardinal sides bend forward to configure the yaw for chest and waist trackers similar to the leg move foot up and down to configure the yaw for the foot
0
199,926
15,082,946,632
IssuesEvent
2021-02-05 15:13:00
odrotbohm/moduliths
https://api.github.com/repos/odrotbohm/moduliths
closed
Assure Spring Boot 2.4 compatibility
in: test support
Spring Boot 2.4 slightly [tunes the arrangement](https://github.com/spring-projects/spring-boot/commit/2af550ef73b3d9d1506888ba9a58026743cedcc0) of how the auto-configuration packages are handled. We need to adapt our way of customizing those to that.
1.0
Assure Spring Boot 2.4 compatibility - Spring Boot 2.4 slightly [tunes the arrangement](https://github.com/spring-projects/spring-boot/commit/2af550ef73b3d9d1506888ba9a58026743cedcc0) of how the auto-configuration packages are handled. We need to adapt our way of customizing those to that.
test
assure spring boot compatibility spring boot slightly of how the auto configuration packages are handled we need to adapt our way of customizing those to that
1
368,641
25,801,279,445
IssuesEvent
2022-12-11 02:03:27
external-secrets/external-secrets
https://api.github.com/repos/external-secrets/external-secrets
closed
Documentation does not state that CreationPolicy=Owner setting ownerReference field
good first issue area/documentation Stale
**Describe the solution you'd like** Update the creationPolicy=Merge (or create a new policy) that also sets the `ownerReference` on the managed secret. **What is the added value?** This would be useful for bootstrapping the `Secret` used by a `SecretStore`, and having external-secrets automatically syncronise changes to the secret after the initial secret creation. This is possible with the Merge strategy already, however my GitOps tooling (argocd) is attempting to self-heal (converge on declared state) and clean up the secret as without the `ownerReference` metadata it does not understand the `Secret`s relationship to the `ExternalSecret` resource. **Give us examples of the outcome** I think this could be achieved by extending the check in the `mutationFunc` in `pkg/controllers/externalsecret/externalsecret_controller.go` to also set apply owner fields for the merge creationPolicy. **Observations (Constraints, Context, etc):** The initial secret is being created directly with `kubectl create secret generic ...`.
1.0
Documentation does not state that CreationPolicy=Owner setting ownerReference field - **Describe the solution you'd like** Update the creationPolicy=Merge (or create a new policy) that also sets the `ownerReference` on the managed secret. **What is the added value?** This would be useful for bootstrapping the `Secret` used by a `SecretStore`, and having external-secrets automatically syncronise changes to the secret after the initial secret creation. This is possible with the Merge strategy already, however my GitOps tooling (argocd) is attempting to self-heal (converge on declared state) and clean up the secret as without the `ownerReference` metadata it does not understand the `Secret`s relationship to the `ExternalSecret` resource. **Give us examples of the outcome** I think this could be achieved by extending the check in the `mutationFunc` in `pkg/controllers/externalsecret/externalsecret_controller.go` to also set apply owner fields for the merge creationPolicy. **Observations (Constraints, Context, etc):** The initial secret is being created directly with `kubectl create secret generic ...`.
non_test
documentation does not state that creationpolicy owner setting ownerreference field describe the solution you d like update the creationpolicy merge or create a new policy that also sets the ownerreference on the managed secret what is the added value this would be useful for bootstrapping the secret used by a secretstore and having external secrets automatically syncronise changes to the secret after the initial secret creation this is possible with the merge strategy already however my gitops tooling argocd is attempting to self heal converge on declared state and clean up the secret as without the ownerreference metadata it does not understand the secret s relationship to the externalsecret resource give us examples of the outcome i think this could be achieved by extending the check in the mutationfunc in pkg controllers externalsecret externalsecret controller go to also set apply owner fields for the merge creationpolicy observations constraints context etc the initial secret is being created directly with kubectl create secret generic
0
122,904
10,240,447,548
IssuesEvent
2019-08-19 20:49:59
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Default value mismatch between API and Rancher UI for int type fields
[zube]: To Test kind/bug-qa priority/2 team/az
**What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** While creating node templates for the providers - Amazon EC2, Azure, Digital Ocean and VMware vSphere, there is value mismatch between the default values for `int` type fields in API vs Rancher UI. <img width="446" alt="Screen Shot 2019-03-19 at 11 13 35 AM" src="https://user-images.githubusercontent.com/26032343/54643755-e919c800-4a54-11e9-9d07-a4ae730bb382.png"> **Other details that may be helpful:** **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): rancher:master (image id: `022651b0ec13`)
1.0
Default value mismatch between API and Rancher UI for int type fields - **What kind of request is this (question/bug/enhancement/feature request):** bug **Steps to reproduce (least amount of steps as possible):** While creating node templates for the providers - Amazon EC2, Azure, Digital Ocean and VMware vSphere, there is value mismatch between the default values for `int` type fields in API vs Rancher UI. <img width="446" alt="Screen Shot 2019-03-19 at 11 13 35 AM" src="https://user-images.githubusercontent.com/26032343/54643755-e919c800-4a54-11e9-9d07-a4ae730bb382.png"> **Other details that may be helpful:** **Environment information** - Rancher version (`rancher/rancher`/`rancher/server` image tag or shown bottom left in the UI): rancher:master (image id: `022651b0ec13`)
test
default value mismatch between api and rancher ui for int type fields what kind of request is this question bug enhancement feature request bug steps to reproduce least amount of steps as possible while creating node templates for the providers amazon azure digital ocean and vmware vsphere there is value mismatch between the default values for int type fields in api vs rancher ui img width alt screen shot at am src other details that may be helpful environment information rancher version rancher rancher rancher server image tag or shown bottom left in the ui rancher master image id
1
100,873
4,103,829,595
IssuesEvent
2016-06-04 23:30:20
afresh1/xdeadly
https://api.github.com/repos/afresh1/xdeadly
opened
Element - submission
Priority: Low Status: Available Type: Enhancement
Probably just needs to become the correct method on the article I think it referrs to the submission date, but I could be wrong.
1.0
Element - submission - Probably just needs to become the correct method on the article I think it referrs to the submission date, but I could be wrong.
non_test
element submission probably just needs to become the correct method on the article i think it referrs to the submission date but i could be wrong
0
102,028
21,893,470,606
IssuesEvent
2022-05-20 05:58:34
DataTalksClub/data-engineering-zoomcamp
https://api.github.com/repos/DataTalksClub/data-engineering-zoomcamp
opened
Timecodes for "DE Zoomcamp 1.3.2 - Creating GCP Infrastructure with Terraform"
timecodes
Youtube video: https://www.youtube.com/watch?v=dNkEgO-CExg
1.0
Timecodes for "DE Zoomcamp 1.3.2 - Creating GCP Infrastructure with Terraform" - Youtube video: https://www.youtube.com/watch?v=dNkEgO-CExg
non_test
timecodes for de zoomcamp creating gcp infrastructure with terraform youtube video
0
36,503
15,014,499,747
IssuesEvent
2021-02-01 06:44:01
Azure/azure-sdk-for-js
https://api.github.com/repos/Azure/azure-sdk-for-js
closed
[Service Bus (potentially EH)] Check if we're always passing in the abortSignal to retry<> and the underlying operation
Client Service Bus
Check if we missed passing abortSignal in the core-amqp retry logic and add if it is required. https://github.com/Azure/azure-sdk-for-js/blob/309a14b6ecce12cfcfb66e88028ccd64af52377c/sdk/core/core-amqp/src/retry.ts#L185 @richardpark-msft @chradek @ramya-rao-a
1.0
[Service Bus (potentially EH)] Check if we're always passing in the abortSignal to retry<> and the underlying operation - Check if we missed passing abortSignal in the core-amqp retry logic and add if it is required. https://github.com/Azure/azure-sdk-for-js/blob/309a14b6ecce12cfcfb66e88028ccd64af52377c/sdk/core/core-amqp/src/retry.ts#L185 @richardpark-msft @chradek @ramya-rao-a
non_test
check if we re always passing in the abortsignal to retry and the underlying operation check if we missed passing abortsignal in the core amqp retry logic and add if it is required richardpark msft chradek ramya rao a
0