Unnamed: 0
int64
0
832k
id
float64
2.49B
32.1B
type
stringclasses
1 value
created_at
stringlengths
19
19
repo
stringlengths
4
112
repo_url
stringlengths
33
141
action
stringclasses
3 values
title
stringlengths
1
1.02k
labels
stringlengths
4
1.54k
body
stringlengths
1
262k
index
stringclasses
17 values
text_combine
stringlengths
95
262k
label
stringclasses
2 values
text
stringlengths
96
252k
binary_label
int64
0
1
160,353
12,508,500,850
IssuesEvent
2020-06-02 15:38:54
learn-able/learnable-be
https://api.github.com/repos/learn-able/learnable-be
closed
Testing Spike
BE_Testing Spike
- [x] Check out best practices for testing HTTP requests - [x] Add shelled testing for http - [ ] Expand testing to cover more handlers - [x] See if Gin offers anything specific for testing it's own processes - [x] Get an idea what it will be to test other files that may be missed from coverage
1.0
Testing Spike - - [x] Check out best practices for testing HTTP requests - [x] Add shelled testing for http - [ ] Expand testing to cover more handlers - [x] See if Gin offers anything specific for testing it's own processes - [x] Get an idea what it will be to test other files that may be missed from coverage
test
testing spike check out best practices for testing http requests add shelled testing for http expand testing to cover more handlers see if gin offers anything specific for testing it s own processes get an idea what it will be to test other files that may be missed from coverage
1
12,249
3,263,689,022
IssuesEvent
2015-10-22 07:15:16
JuliaLang/julia
https://api.github.com/repos/JuliaLang/julia
opened
Intermittent segfault in libdl test on linux64 buildbots
32-bit arm linux test
See recent failures at http://buildbot.e.ip.saba.us:8010/builders/build_ubuntu12.04-x64?numbuilds=250 and http://buildbot.e.ip.saba.us:8010/builders/build_centos7.1-x64?numbuilds=250 ``` From worker 3: * enums in 1.33 seconds, maxrss 1165.46 MB From worker 2: * misc in 24.35 seconds, maxrss 1406.89 MB From worker 2: * i18n in 0.03 seconds, maxrss 1406.89 MB From worker 2: * workspace in 0.35 seconds, maxrss 1406.89 MB signal (11): Segmentation fault _IO_feof at /lib64/libc.so.6 (unknown line) unknown function (ip: 0x7f8cf8b107dd) anonymous at essentials.jl:116 cd at file.jl:22 jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) unknown function (ip: 0x7f8cf8b0d7a3) unknown function (ip: 0x7f8cf8b0cbe1) unknown function (ip: 0x7f8cf8b0e038) unknown function (ip: 0x7f8cf8b0e1ee) unknown function (ip: 0x7f8cf8b219df) unknown function (ip: 0x7f8cf8b222a9) jl_load_file_string at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) include_string at essentials.jl:116 jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) include_from_node1 at ./loading.jl:387 jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) runtests at util.jl:179 jlcall_runtests_22758 at (unknown line) jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) anonymous at /home/centos/buildbot/slave/build_centos7_1-x64/build/test/runtests.jl:36 jl_f_apply at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) anonymous at multi.jl:898 run_work_thunk at multi.jl:651 jlcall_run_work_thunk_22669 at (unknown line) jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) anonymous at multi.jl:898 unknown function (ip: 0x7f8cf8b13dc8) unknown function (ip: (nil)) From worker 2: * libdl Worker 2 terminated. ERROR (unhandled task failure): EOFError: read end of file From worker 4: * int in 1.33 seconds, maxrss 156.14 MB From worker 4: * intset in 0.75 seconds, maxrss 156.14 MB From worker 4: * floatfuncs in 1.66 seconds, maxrss 156.14 MB From worker 3: * cmdlineargs in 28.07 seconds, maxrss 1175.49 MB From worker 3: * fft in 24.96 seconds, maxrss 1226.80 MB From worker 4: * parallel From worker 4: in 35.81 seconds, maxrss 156.14 MB From worker 3: * dsp in 15.16 seconds, maxrss 1229.05 MB From worker 4: * examples in 22.73 seconds, maxrss 191.95 MB Exception running test libdl : ProcessExitedException() ERROR: LoadError: Some tests exited with errors. [inlined code] from error.jl:21 in anonymous at /home/centos/buildbot/slave/build_centos7_1-x64/build/test/runtests.jl:64 in cd at file.jl:22 in include at ./boot.jl:261 in include_from_node1 at ./loading.jl:384 [inlined code] from ./operators.jl:313 in process_options at ./client.jl:277 in _start at ./client.jl:377 while loading /home/centos/buildbot/slave/build_centos7_1-x64/build/test/runtests.jl, in expression starting on line 13 ```
1.0
Intermittent segfault in libdl test on linux64 buildbots - See recent failures at http://buildbot.e.ip.saba.us:8010/builders/build_ubuntu12.04-x64?numbuilds=250 and http://buildbot.e.ip.saba.us:8010/builders/build_centos7.1-x64?numbuilds=250 ``` From worker 3: * enums in 1.33 seconds, maxrss 1165.46 MB From worker 2: * misc in 24.35 seconds, maxrss 1406.89 MB From worker 2: * i18n in 0.03 seconds, maxrss 1406.89 MB From worker 2: * workspace in 0.35 seconds, maxrss 1406.89 MB signal (11): Segmentation fault _IO_feof at /lib64/libc.so.6 (unknown line) unknown function (ip: 0x7f8cf8b107dd) anonymous at essentials.jl:116 cd at file.jl:22 jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) unknown function (ip: 0x7f8cf8b0d7a3) unknown function (ip: 0x7f8cf8b0cbe1) unknown function (ip: 0x7f8cf8b0e038) unknown function (ip: 0x7f8cf8b0e1ee) unknown function (ip: 0x7f8cf8b219df) unknown function (ip: 0x7f8cf8b222a9) jl_load_file_string at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) include_string at essentials.jl:116 jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) include_from_node1 at ./loading.jl:387 jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) runtests at util.jl:179 jlcall_runtests_22758 at (unknown line) jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) anonymous at /home/centos/buildbot/slave/build_centos7_1-x64/build/test/runtests.jl:36 jl_f_apply at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) anonymous at multi.jl:898 run_work_thunk at multi.jl:651 jlcall_run_work_thunk_22669 at (unknown line) jl_apply_generic at /home/centos/buildbot/slave/build_centos7_1-x64/build/usr/bin/../lib/libjulia.so (unknown line) anonymous at multi.jl:898 unknown function (ip: 0x7f8cf8b13dc8) unknown function (ip: (nil)) From worker 2: * libdl Worker 2 terminated. ERROR (unhandled task failure): EOFError: read end of file From worker 4: * int in 1.33 seconds, maxrss 156.14 MB From worker 4: * intset in 0.75 seconds, maxrss 156.14 MB From worker 4: * floatfuncs in 1.66 seconds, maxrss 156.14 MB From worker 3: * cmdlineargs in 28.07 seconds, maxrss 1175.49 MB From worker 3: * fft in 24.96 seconds, maxrss 1226.80 MB From worker 4: * parallel From worker 4: in 35.81 seconds, maxrss 156.14 MB From worker 3: * dsp in 15.16 seconds, maxrss 1229.05 MB From worker 4: * examples in 22.73 seconds, maxrss 191.95 MB Exception running test libdl : ProcessExitedException() ERROR: LoadError: Some tests exited with errors. [inlined code] from error.jl:21 in anonymous at /home/centos/buildbot/slave/build_centos7_1-x64/build/test/runtests.jl:64 in cd at file.jl:22 in include at ./boot.jl:261 in include_from_node1 at ./loading.jl:384 [inlined code] from ./operators.jl:313 in process_options at ./client.jl:277 in _start at ./client.jl:377 while loading /home/centos/buildbot/slave/build_centos7_1-x64/build/test/runtests.jl, in expression starting on line 13 ```
test
intermittent segfault in libdl test on buildbots see recent failures at and from worker enums in seconds maxrss mb from worker misc in seconds maxrss mb from worker in seconds maxrss mb from worker workspace in seconds maxrss mb signal segmentation fault io feof at libc so unknown line unknown function ip anonymous at essentials jl cd at file jl jl apply generic at home centos buildbot slave build build usr bin lib libjulia so unknown line unknown function ip unknown function ip unknown function ip unknown function ip unknown function ip unknown function ip jl load file string at home centos buildbot slave build build usr bin lib libjulia so unknown line include string at essentials jl jl apply generic at home centos buildbot slave build build usr bin lib libjulia so unknown line include from at loading jl jl apply generic at home centos buildbot slave build build usr bin lib libjulia so unknown line runtests at util jl jlcall runtests at unknown line jl apply generic at home centos buildbot slave build build usr bin lib libjulia so unknown line anonymous at home centos buildbot slave build build test runtests jl jl f apply at home centos buildbot slave build build usr bin lib libjulia so unknown line anonymous at multi jl run work thunk at multi jl jlcall run work thunk at unknown line jl apply generic at home centos buildbot slave build build usr bin lib libjulia so unknown line anonymous at multi jl unknown function ip unknown function ip nil from worker libdl worker terminated error unhandled task failure eoferror read end of file from worker int in seconds maxrss mb from worker intset in seconds maxrss mb from worker floatfuncs in seconds maxrss mb from worker cmdlineargs in seconds maxrss mb from worker fft in seconds maxrss mb from worker parallel from worker in seconds maxrss mb from worker dsp in seconds maxrss mb from worker examples in seconds maxrss mb exception running test libdl processexitedexception error loaderror some tests exited with errors from error jl in anonymous at home centos buildbot slave build build test runtests jl in cd at file jl in include at boot jl in include from at loading jl from operators jl in process options at client jl in start at client jl while loading home centos buildbot slave build build test runtests jl in expression starting on line
1
138,114
20,329,758,425
IssuesEvent
2022-02-18 09:34:36
TravelMockPlanner/TravelManagerPlanner
https://api.github.com/repos/TravelMockPlanner/TravelManagerPlanner
closed
About Design
design
## 기능 설명 마지막 Design 관련해서 작성해보자 ## 완료 조건 - [x] TMP의 로고를 만들어보자 - [x] 로고를 기준으로 LaunchScreen을 작성해보자
1.0
About Design - ## 기능 설명 마지막 Design 관련해서 작성해보자 ## 완료 조건 - [x] TMP의 로고를 만들어보자 - [x] 로고를 기준으로 LaunchScreen을 작성해보자
non_test
about design 기능 설명 마지막 design 관련해서 작성해보자 완료 조건 tmp의 로고를 만들어보자 로고를 기준으로 launchscreen을 작성해보자
0
34,070
4,889,939,962
IssuesEvent
2016-11-18 12:02:25
jadu/pulsar
https://api.github.com/repos/jadu/pulsar
closed
Test core Pulsar layouts and helpers using WAVE tool
awaiting review tests
First steps in confirming WAVE compliance in Continuum Products. Lexicon can be tested using browser plugins. http://wave.webaim.org
1.0
Test core Pulsar layouts and helpers using WAVE tool - First steps in confirming WAVE compliance in Continuum Products. Lexicon can be tested using browser plugins. http://wave.webaim.org
test
test core pulsar layouts and helpers using wave tool first steps in confirming wave compliance in continuum products lexicon can be tested using browser plugins
1
10,050
3,351,839,330
IssuesEvent
2015-11-17 20:12:13
bevanhunt/meteor-leaflet
https://api.github.com/repos/bevanhunt/meteor-leaflet
closed
issue map div is hidden
documentation
I had the issue, that without adding the style: ```css #map { min-height: 350px; min-width: 100%; } ``` my ```<div id='map' class="map"></div>``` would not show up. Maybe its good to add this to the documentation?
1.0
issue map div is hidden - I had the issue, that without adding the style: ```css #map { min-height: 350px; min-width: 100%; } ``` my ```<div id='map' class="map"></div>``` would not show up. Maybe its good to add this to the documentation?
non_test
issue map div is hidden i had the issue that without adding the style css map min height min width my would not show up maybe its good to add this to the documentation
0
131,382
10,693,780,015
IssuesEvent
2019-10-23 09:30:41
I2PC/scipion
https://api.github.com/repos/I2PC/scipion
closed
Flag new protocols/obsolete ones
betatesting enhancement gui
Figure out a way to flag new protocols as new and obsolete ones too
1.0
Flag new protocols/obsolete ones - Figure out a way to flag new protocols as new and obsolete ones too
test
flag new protocols obsolete ones figure out a way to flag new protocols as new and obsolete ones too
1
733,653
25,316,020,621
IssuesEvent
2022-11-17 21:43:04
space-wizards/space-station-14
https://api.github.com/repos/space-wizards/space-station-14
closed
Artifacts should choose prototypes on mapinit rather than componentinit
Priority: 2-Before Release Size: Very Small Difficulty: 1-Easy
This is how all of our other RNG stuff does it and would make salvage more variable.
1.0
Artifacts should choose prototypes on mapinit rather than componentinit - This is how all of our other RNG stuff does it and would make salvage more variable.
non_test
artifacts should choose prototypes on mapinit rather than componentinit this is how all of our other rng stuff does it and would make salvage more variable
0
347,089
31,117,588,746
IssuesEvent
2023-08-15 02:02:43
Almondia/meca_front
https://api.github.com/repos/Almondia/meca_front
opened
Storybook 관련개선
:frog: FE :test_tube: test
## Summary Storybook 관련개선 <br> ## TODO - [ ] msw handler 사용방식을 스토리 컴포넌트 decorator에 선언하여 사용하도록 통일 - [ ] 정확한 UI확인을 위한 불필요한 padding 설정 제거 - [ ] 불필요하게 정적 파일 등이 msw로 캡쳐링되는 문제 해결 - [ ] Component Subtitle 제거 ![image](https://github.com/Almondia/meca_front/assets/76927397/c5621c5d-a1f7-4bf0-b50b-27292652d258)
2.0
Storybook 관련개선 - ## Summary Storybook 관련개선 <br> ## TODO - [ ] msw handler 사용방식을 스토리 컴포넌트 decorator에 선언하여 사용하도록 통일 - [ ] 정확한 UI확인을 위한 불필요한 padding 설정 제거 - [ ] 불필요하게 정적 파일 등이 msw로 캡쳐링되는 문제 해결 - [ ] Component Subtitle 제거 ![image](https://github.com/Almondia/meca_front/assets/76927397/c5621c5d-a1f7-4bf0-b50b-27292652d258)
test
storybook 관련개선 summary storybook 관련개선 todo msw handler 사용방식을 스토리 컴포넌트 decorator에 선언하여 사용하도록 통일 정확한 ui확인을 위한 불필요한 padding 설정 제거 불필요하게 정적 파일 등이 msw로 캡쳐링되는 문제 해결 component subtitle 제거
1
12,387
3,268,473,621
IssuesEvent
2015-10-23 11:43:54
ova13lastar/src_dipa
https://api.github.com/repos/ova13lastar/src_dipa
closed
Faire un test sur Phie MBG
04_done app:dipa comp:action-detect type:test
La patientèle semble faible : 59xxx009 <!--- @huboard:{"order":40.0,"milestone_order":55,"custom_state":""} -->
1.0
Faire un test sur Phie MBG - La patientèle semble faible : 59xxx009 <!--- @huboard:{"order":40.0,"milestone_order":55,"custom_state":""} -->
test
faire un test sur phie mbg la patientèle semble faible huboard order milestone order custom state
1
145,357
11,686,538,851
IssuesEvent
2020-03-05 11:03:07
mratsim/Arraymancer
https://api.github.com/repos/mratsim/Arraymancer
closed
Refactor and optimize test suite
optimization tests
The test suite is quite slow, testament (Nim compiler test framework) could be a scalable alternative to unittest. See benches in #359
1.0
Refactor and optimize test suite - The test suite is quite slow, testament (Nim compiler test framework) could be a scalable alternative to unittest. See benches in #359
test
refactor and optimize test suite the test suite is quite slow testament nim compiler test framework could be a scalable alternative to unittest see benches in
1
136,522
12,717,512,206
IssuesEvent
2020-06-24 05:23:18
intel/dffml
https://api.github.com/repos/intel/dffml
opened
docs: model: scikit: Mention daal4py auto patching if installed gives speedup
documentation
In the `model/scikit/dffml_model_scikit/__init__.py` we need to mention how we [use](https://github.com/intel/dffml/commit/df5a07de9d28e916a9843650f082e411c21aaa21) daal4py if it's installed to give users [noticeable performance improvements](https://intelpython.github.io/daal4py/sklearn.html)
1.0
docs: model: scikit: Mention daal4py auto patching if installed gives speedup - In the `model/scikit/dffml_model_scikit/__init__.py` we need to mention how we [use](https://github.com/intel/dffml/commit/df5a07de9d28e916a9843650f082e411c21aaa21) daal4py if it's installed to give users [noticeable performance improvements](https://intelpython.github.io/daal4py/sklearn.html)
non_test
docs model scikit mention auto patching if installed gives speedup in the model scikit dffml model scikit init py we need to mention how we if it s installed to give users
0
404,659
27,492,662,371
IssuesEvent
2023-03-04 20:20:23
jrnl-org/jrnl
https://api.github.com/repos/jrnl-org/jrnl
closed
Update console examples on jrnl.sh front page
documentation :pushpin:
### Affected Page(s) https://jrnl.sh ### What Could Be Better? It's been years since this text has changed. It would be good to review it against jrnl's current feature set. I think some style changes could be nice too: maybe we don't need to be "like a boss" to use jrnl. And I'd rather not taunt a powerful intelligence agency on our front page. ### Other Information _No response_
1.0
Update console examples on jrnl.sh front page - ### Affected Page(s) https://jrnl.sh ### What Could Be Better? It's been years since this text has changed. It would be good to review it against jrnl's current feature set. I think some style changes could be nice too: maybe we don't need to be "like a boss" to use jrnl. And I'd rather not taunt a powerful intelligence agency on our front page. ### Other Information _No response_
non_test
update console examples on jrnl sh front page affected page s what could be better it s been years since this text has changed it would be good to review it against jrnl s current feature set i think some style changes could be nice too maybe we don t need to be like a boss to use jrnl and i d rather not taunt a powerful intelligence agency on our front page other information no response
0
196,559
6,935,124,158
IssuesEvent
2017-12-03 03:59:36
prettier/prettier
https://api.github.com/repos/prettier/prettier
closed
Typescript: decorator + readonly + comment leads to un-compilable code
area:comments lang:flow lang:typescript priority:high status:has pr type:bug
<!-- BUGGY OR UGLY? Please use this template. Tip! Don't write this stuff manually. 1. Go to https://prettier.io/playground 2. Paste your code and set options 3. Press the "Report issue" button in the lower right --> **Prettier 1.8.2** ```sh # Options (if any): ``` **Input:** ```ts class Something { @Annotateme() //TODO will it break readonly property: Array<string> } ``` **Output:** ```ts class Something { @Annotateme() readonly //TODO will it break property: Array<string>; } ``` **Expected behavior:** Unchanged code. Current version leads to code that Typescript 2.4.1 refuses to compile.
1.0
Typescript: decorator + readonly + comment leads to un-compilable code - <!-- BUGGY OR UGLY? Please use this template. Tip! Don't write this stuff manually. 1. Go to https://prettier.io/playground 2. Paste your code and set options 3. Press the "Report issue" button in the lower right --> **Prettier 1.8.2** ```sh # Options (if any): ``` **Input:** ```ts class Something { @Annotateme() //TODO will it break readonly property: Array<string> } ``` **Output:** ```ts class Something { @Annotateme() readonly //TODO will it break property: Array<string>; } ``` **Expected behavior:** Unchanged code. Current version leads to code that Typescript 2.4.1 refuses to compile.
non_test
typescript decorator readonly comment leads to un compilable code buggy or ugly please use this template tip don t write this stuff manually go to paste your code and set options press the report issue button in the lower right prettier sh options if any input ts class something annotateme todo will it break readonly property array output ts class something annotateme readonly todo will it break property array expected behavior unchanged code current version leads to code that typescript refuses to compile
0
335,021
10,147,907,061
IssuesEvent
2019-08-05 11:51:41
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
media.pearsoncmg.com - video or audio doesn't play
browser-firefox engine-gecko priority-normal
<!-- @browser: Firefox 64.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 --> <!-- @reported_with: desktop-reporter --> **URL**: http://media.pearsoncmg.com/bc/bc_masteringaandp/IP/cardcycl/topic1.html?41&D **Browser / Version**: Firefox 64.0 **Operating System**: Windows 10 **Tested Another Browser**: Unknown **Problem type**: Video or audio doesn't play **Description**: plug in not supported no video **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/7/125a6fe0-af7f-4572-8195-3960a1063e2c-thumb.jpeg)](https://webcompat.com/uploads/2019/7/125a6fe0-af7f-4572-8195-3960a1063e2c.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181205213546</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: release</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
media.pearsoncmg.com - video or audio doesn't play - <!-- @browser: Firefox 64.0 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 --> <!-- @reported_with: desktop-reporter --> **URL**: http://media.pearsoncmg.com/bc/bc_masteringaandp/IP/cardcycl/topic1.html?41&D **Browser / Version**: Firefox 64.0 **Operating System**: Windows 10 **Tested Another Browser**: Unknown **Problem type**: Video or audio doesn't play **Description**: plug in not supported no video **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/7/125a6fe0-af7f-4572-8195-3960a1063e2c-thumb.jpeg)](https://webcompat.com/uploads/2019/7/125a6fe0-af7f-4572-8195-3960a1063e2c.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20181205213546</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: release</li> </ul> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
media pearsoncmg com video or audio doesn t play url browser version firefox operating system windows tested another browser unknown problem type video or audio doesn t play description plug in not supported no video steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel release from with ❤️
0
231,191
25,494,265,688
IssuesEvent
2022-11-27 13:29:18
deckhouse/deckhouse
https://api.github.com/repos/deckhouse/deckhouse
closed
[admission-policy-engine] Maintain the list of constraints to enforce policies
area/security source/external-client area/compliance
### Preflight Checklist - [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success. ### Use case. Why is this important? As of today, our gatekeeper installations validate all namespaced resources not under Deckhouse control and all cluster resources. ```yaml namespaceSelector: matchExpressions: - key: heritage operator: NotIn values: - deckhouse objectSelector: {} rules: - apiGroups: - '*' apiVersions: - '*' operations: - CREATE - UPDATE resources: - '*' ``` ### No available replicas If gatekeeper is unavailable, all requests take 3+ seconds. Some odd examples of problematic requests. Example 1 - External authorization request (used by all kube-rbac-proxy deployments): ``` POST https://127.0.0.1:6445/apis/authorization.k8s.io/v1/subjectaccessreviews Response Status: 201 Created in 3019 milliseconds ``` Example 2 - Patching NodeGroup status: ``` POST https://127.0.0.1:6445/apis/deckhouse.io/v1/nodegroups/status Response Status: 200 Created in 3007 milliseconds ``` In the kube-apiserver logs, there are entries like the following: ``` I1104 15:39:35.048214 1 trace.go:205] Trace[93865027]: "Call validating webhook" configuration:d8-admission-policy-engine-config,webhook:admission-policy-engine.deckhouse.io,resource:coordination.k8s.io/v1, Resource=leases,subresource:,operation:UPDATE,UID:84dc390f-1587-4785-bb1d-5b2bd27f9a5d (04-Nov-2022 15:39:32.047) (total time: 3000ms): Trace[93865027]: [3.000731799s] [3.000731799s] END ``` ### Service network is not working In this case, all validations will fail. See https://github.com/deckhouse/deckhouse/issues/2886#issuecomment-1292719882 for more details. ### Proposed Solution 1. Provide a default list of resources to apply policies. Parse created user policies and add resources to the validating webhook list accordingly. Is it possible? 2. Hardcode some resources to avoid policy enforcement, e.g., subjectaccessreview. ### Additional Information The way that should be avoided is maintaining the list of resources by users.
True
[admission-policy-engine] Maintain the list of constraints to enforce policies - ### Preflight Checklist - [X] I agree to follow the [Code of Conduct](https://github.com/deckhouse/deckhouse/blob/main/CODE_OF_CONDUCT.md) that this project adheres to. - [X] I have searched the [issue tracker](https://github.com/deckhouse/deckhouse/issues) for an issue that matches the one I want to file, without success. ### Use case. Why is this important? As of today, our gatekeeper installations validate all namespaced resources not under Deckhouse control and all cluster resources. ```yaml namespaceSelector: matchExpressions: - key: heritage operator: NotIn values: - deckhouse objectSelector: {} rules: - apiGroups: - '*' apiVersions: - '*' operations: - CREATE - UPDATE resources: - '*' ``` ### No available replicas If gatekeeper is unavailable, all requests take 3+ seconds. Some odd examples of problematic requests. Example 1 - External authorization request (used by all kube-rbac-proxy deployments): ``` POST https://127.0.0.1:6445/apis/authorization.k8s.io/v1/subjectaccessreviews Response Status: 201 Created in 3019 milliseconds ``` Example 2 - Patching NodeGroup status: ``` POST https://127.0.0.1:6445/apis/deckhouse.io/v1/nodegroups/status Response Status: 200 Created in 3007 milliseconds ``` In the kube-apiserver logs, there are entries like the following: ``` I1104 15:39:35.048214 1 trace.go:205] Trace[93865027]: "Call validating webhook" configuration:d8-admission-policy-engine-config,webhook:admission-policy-engine.deckhouse.io,resource:coordination.k8s.io/v1, Resource=leases,subresource:,operation:UPDATE,UID:84dc390f-1587-4785-bb1d-5b2bd27f9a5d (04-Nov-2022 15:39:32.047) (total time: 3000ms): Trace[93865027]: [3.000731799s] [3.000731799s] END ``` ### Service network is not working In this case, all validations will fail. See https://github.com/deckhouse/deckhouse/issues/2886#issuecomment-1292719882 for more details. ### Proposed Solution 1. Provide a default list of resources to apply policies. Parse created user policies and add resources to the validating webhook list accordingly. Is it possible? 2. Hardcode some resources to avoid policy enforcement, e.g., subjectaccessreview. ### Additional Information The way that should be avoided is maintaining the list of resources by users.
non_test
maintain the list of constraints to enforce policies preflight checklist i agree to follow the that this project adheres to i have searched the for an issue that matches the one i want to file without success use case why is this important as of today our gatekeeper installations validate all namespaced resources not under deckhouse control and all cluster resources yaml namespaceselector matchexpressions key heritage operator notin values deckhouse objectselector rules apigroups apiversions operations create update resources no available replicas if gatekeeper is unavailable all requests take seconds some odd examples of problematic requests example external authorization request used by all kube rbac proxy deployments post response status created in milliseconds example patching nodegroup status post response status created in milliseconds in the kube apiserver logs there are entries like the following trace go trace call validating webhook configuration admission policy engine config webhook admission policy engine deckhouse io resource coordination io resource leases subresource operation update uid nov total time trace end service network is not working in this case all validations will fail see for more details proposed solution provide a default list of resources to apply policies parse created user policies and add resources to the validating webhook list accordingly is it possible hardcode some resources to avoid policy enforcement e g subjectaccessreview additional information the way that should be avoided is maintaining the list of resources by users
0
165,272
12,835,820,800
IssuesEvent
2020-07-07 13:28:55
softmatterlab/Braph-2.0-Matlab
https://api.github.com/repos/softmatterlab/Braph-2.0-Matlab
closed
ComparisonfMRI
analysis test
**Branch from and merge to gv-analysis-comparison** - [ ] ComparisonfMRI - [ ] test_ComparisonfMRI Use as reference ComparisonMRI Double-check with @annacanal, see issue #526
1.0
ComparisonfMRI - **Branch from and merge to gv-analysis-comparison** - [ ] ComparisonfMRI - [ ] test_ComparisonfMRI Use as reference ComparisonMRI Double-check with @annacanal, see issue #526
test
comparisonfmri branch from and merge to gv analysis comparison comparisonfmri test comparisonfmri use as reference comparisonmri double check with annacanal see issue
1
702,981
24,143,720,619
IssuesEvent
2022-09-21 16:48:26
KinsonDigital/CICD
https://api.github.com/repos/KinsonDigital/CICD
closed
🚧Fix project dotnet tool setup
high priority preview ✨new feature
### Complete The Item Below - [X] I have updated the title without removing the 🚧 emoji. ### Description Fix an issue where the project is not setup as a dotnet tool. This needs to be done in the csproj file and it simply not setup correctly. Add the 2 XML elements below to the csproj file to fix the issue ```xml <PackAsTool>true</PackAsTool> <ToolCommandName>cicd</ToolCommandName> ``` ### Acceptance Criteria - [x] csproj setup correctly to behave as a dotnet tool ### ToDo Items - [X] Change type labels added to this issue. Refer to the _**Change Type Labels**_ section below. - [X] Priority label added to this issue. Refer to the _**Priority Type Labels**_ section below. - [X] Issue linked to the correct project _(if applicable)_. - [x] Issue linked to the correct milestone _(if applicable)_. - [x] Draft pull request created and linked to this issue _(only required with code changes)_. ### Issue Dependencies _No response_ ### Related Work _No response_ ### Additional Information: **_<details closed><summary>Change Type Labels</summary>_** | Change Type | Label | |---------------------|----------------------| | Bug Fixes | `🐛bug` | | Breaking Changes | `🧨breaking changes` | | New Feature | `✨new feature` | | Workflow Changes | `workflow` | | Code Doc Changes | `🗒️documentation/code` | | Product Doc Changes | `📝documentation/product` | </details> **_<details closed><summary>Priority Type Labels</summary>_** | Priority Type | Label | |---------------------|-------------------| | Low Priority | `low priority` | | Medium Priority | `medium priority` | | High Priority | `high priority` | </details> ### Code of Conduct - [X] I agree to follow this project's Code of Conduct.
1.0
🚧Fix project dotnet tool setup - ### Complete The Item Below - [X] I have updated the title without removing the 🚧 emoji. ### Description Fix an issue where the project is not setup as a dotnet tool. This needs to be done in the csproj file and it simply not setup correctly. Add the 2 XML elements below to the csproj file to fix the issue ```xml <PackAsTool>true</PackAsTool> <ToolCommandName>cicd</ToolCommandName> ``` ### Acceptance Criteria - [x] csproj setup correctly to behave as a dotnet tool ### ToDo Items - [X] Change type labels added to this issue. Refer to the _**Change Type Labels**_ section below. - [X] Priority label added to this issue. Refer to the _**Priority Type Labels**_ section below. - [X] Issue linked to the correct project _(if applicable)_. - [x] Issue linked to the correct milestone _(if applicable)_. - [x] Draft pull request created and linked to this issue _(only required with code changes)_. ### Issue Dependencies _No response_ ### Related Work _No response_ ### Additional Information: **_<details closed><summary>Change Type Labels</summary>_** | Change Type | Label | |---------------------|----------------------| | Bug Fixes | `🐛bug` | | Breaking Changes | `🧨breaking changes` | | New Feature | `✨new feature` | | Workflow Changes | `workflow` | | Code Doc Changes | `🗒️documentation/code` | | Product Doc Changes | `📝documentation/product` | </details> **_<details closed><summary>Priority Type Labels</summary>_** | Priority Type | Label | |---------------------|-------------------| | Low Priority | `low priority` | | Medium Priority | `medium priority` | | High Priority | `high priority` | </details> ### Code of Conduct - [X] I agree to follow this project's Code of Conduct.
non_test
🚧fix project dotnet tool setup complete the item below i have updated the title without removing the 🚧 emoji description fix an issue where the project is not setup as a dotnet tool this needs to be done in the csproj file and it simply not setup correctly add the xml elements below to the csproj file to fix the issue xml true cicd acceptance criteria csproj setup correctly to behave as a dotnet tool todo items change type labels added to this issue refer to the change type labels section below priority label added to this issue refer to the priority type labels section below issue linked to the correct project if applicable issue linked to the correct milestone if applicable draft pull request created and linked to this issue only required with code changes issue dependencies no response related work no response additional information change type labels change type label bug fixes 🐛bug breaking changes 🧨breaking changes new feature ✨new feature workflow changes workflow code doc changes 🗒️documentation code product doc changes 📝documentation product priority type labels priority type label low priority low priority medium priority medium priority high priority high priority code of conduct i agree to follow this project s code of conduct
0
133,699
10,856,084,731
IssuesEvent
2019-11-13 19:48:25
microcks/microcks
https://api.github.com/repos/microcks/microcks
closed
How to add my SSL certificate to Microcks for integration tests
component/tests kind/enhancement
Hi, I am trying Microcks 0.7.2 on our own Openshift PaaS. I used the Charts provided in the official install docs. Is there a way to add my certificates to call my endpoints for the integration tests without rebuilding the official Microcks image? I can see Secrets are only handling authentication tokens and not certificates for self signed or unkown certification authority. Thanks, Vina.
1.0
How to add my SSL certificate to Microcks for integration tests - Hi, I am trying Microcks 0.7.2 on our own Openshift PaaS. I used the Charts provided in the official install docs. Is there a way to add my certificates to call my endpoints for the integration tests without rebuilding the official Microcks image? I can see Secrets are only handling authentication tokens and not certificates for self signed or unkown certification authority. Thanks, Vina.
test
how to add my ssl certificate to microcks for integration tests hi i am trying microcks on our own openshift paas i used the charts provided in the official install docs is there a way to add my certificates to call my endpoints for the integration tests without rebuilding the official microcks image i can see secrets are only handling authentication tokens and not certificates for self signed or unkown certification authority thanks vina
1
342,847
30,641,039,030
IssuesEvent
2023-07-24 22:02:11
unifyai/ivy
https://api.github.com/repos/unifyai/ivy
opened
Fix raw_ops.test_tensorflow_Relu
TensorFlow Frontend Sub Task Failing Test
| | | |---|---| |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5649119338"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5648325047"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5648325047"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5649119338"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5649119338"><img src=https://img.shields.io/badge/-failure-red></a>
1.0
Fix raw_ops.test_tensorflow_Relu - | | | |---|---| |torch|<a href="https://github.com/unifyai/ivy/actions/runs/5649119338"><img src=https://img.shields.io/badge/-success-success></a> |numpy|<a href="https://github.com/unifyai/ivy/actions/runs/5648325047"><img src=https://img.shields.io/badge/-success-success></a> |jax|<a href="https://github.com/unifyai/ivy/actions/runs/5648325047"><img src=https://img.shields.io/badge/-success-success></a> |tensorflow|<a href="https://github.com/unifyai/ivy/actions/runs/5649119338"><img src=https://img.shields.io/badge/-success-success></a> |paddle|<a href="https://github.com/unifyai/ivy/actions/runs/5649119338"><img src=https://img.shields.io/badge/-failure-red></a>
test
fix raw ops test tensorflow relu torch a href src numpy a href src jax a href src tensorflow a href src paddle a href src
1
205,595
15,649,039,452
IssuesEvent
2021-03-23 06:54:46
cockroachdb/cockroach
https://api.github.com/repos/cockroachdb/cockroach
opened
roachtest: sqlsmith/setup=tpch-sf1/setting=no-ddl failed
C-test-failure O-roachtest O-robot branch-master release-blocker
[(roachtest).sqlsmith/setup=tpch-sf1/setting=no-ddl failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2802936&tab=buildLog) on [master@53bf501e233c337b9863755914d9c00010517329](https://github.com/cockroachdb/cockroach/commits/53bf501e233c337b9863755914d9c00010517329): ``` | | main.main | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1852 | | runtime.main | | /usr/local/go/src/runtime/proc.go:204 | | runtime.goexit | | /usr/local/go/src/runtime/asm_amd64.s:1374 | Wraps: (2) 4: dead | Error types: (1) *withstack.withStack (2) *errutil.leafError Wraps: (3) secondary error attachment | 2: dead | (1) attached stack trace | -- stack trace: | | main.glob..func14 | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1147 | | main.wrap.func1 | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:271 | | github.com/spf13/cobra.(*Command).execute | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:830 | | github.com/spf13/cobra.(*Command).ExecuteC | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:914 | | github.com/spf13/cobra.(*Command).Execute | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:864 | | main.main | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1852 | | runtime.main | | /usr/local/go/src/runtime/proc.go:204 | | runtime.goexit | | /usr/local/go/src/runtime/asm_amd64.s:1374 | Wraps: (2) 2: dead | Error types: (1) *withstack.withStack (2) *errutil.leafError Wraps: (4) attached stack trace -- stack trace: | main.glob..func14 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1147 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:271 | github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:830 | github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:914 | github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:864 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1852 | runtime.main | /usr/local/go/src/runtime/proc.go:204 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (5) 3: dead Error types: (1) errors.Unclassified (2) *secondary.withSecondaryError (3) *secondary.withSecondaryError (4) *withstack.withStack (5) *errutil.leafError ``` <details><summary>More</summary><p> Artifacts: [/sqlsmith/setup=tpch-sf1/setting=no-ddl](https://teamcity.cockroachdb.com/viewLog.html?buildId=2802936&tab=artifacts#/sqlsmith/setup=tpch-sf1/setting=no-ddl) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Asqlsmith%2Fsetup%3Dtpch-sf1%2Fsetting%3Dno-ddl.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
2.0
roachtest: sqlsmith/setup=tpch-sf1/setting=no-ddl failed - [(roachtest).sqlsmith/setup=tpch-sf1/setting=no-ddl failed](https://teamcity.cockroachdb.com/viewLog.html?buildId=2802936&tab=buildLog) on [master@53bf501e233c337b9863755914d9c00010517329](https://github.com/cockroachdb/cockroach/commits/53bf501e233c337b9863755914d9c00010517329): ``` | | main.main | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1852 | | runtime.main | | /usr/local/go/src/runtime/proc.go:204 | | runtime.goexit | | /usr/local/go/src/runtime/asm_amd64.s:1374 | Wraps: (2) 4: dead | Error types: (1) *withstack.withStack (2) *errutil.leafError Wraps: (3) secondary error attachment | 2: dead | (1) attached stack trace | -- stack trace: | | main.glob..func14 | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1147 | | main.wrap.func1 | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:271 | | github.com/spf13/cobra.(*Command).execute | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:830 | | github.com/spf13/cobra.(*Command).ExecuteC | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:914 | | github.com/spf13/cobra.(*Command).Execute | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:864 | | main.main | | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1852 | | runtime.main | | /usr/local/go/src/runtime/proc.go:204 | | runtime.goexit | | /usr/local/go/src/runtime/asm_amd64.s:1374 | Wraps: (2) 2: dead | Error types: (1) *withstack.withStack (2) *errutil.leafError Wraps: (4) attached stack trace -- stack trace: | main.glob..func14 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1147 | main.wrap.func1 | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:271 | github.com/spf13/cobra.(*Command).execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:830 | github.com/spf13/cobra.(*Command).ExecuteC | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:914 | github.com/spf13/cobra.(*Command).Execute | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/vendor/github.com/spf13/cobra/command.go:864 | main.main | /home/agent/work/.go/src/github.com/cockroachdb/cockroach/pkg/cmd/roachprod/main.go:1852 | runtime.main | /usr/local/go/src/runtime/proc.go:204 | runtime.goexit | /usr/local/go/src/runtime/asm_amd64.s:1374 Wraps: (5) 3: dead Error types: (1) errors.Unclassified (2) *secondary.withSecondaryError (3) *secondary.withSecondaryError (4) *withstack.withStack (5) *errutil.leafError ``` <details><summary>More</summary><p> Artifacts: [/sqlsmith/setup=tpch-sf1/setting=no-ddl](https://teamcity.cockroachdb.com/viewLog.html?buildId=2802936&tab=artifacts#/sqlsmith/setup=tpch-sf1/setting=no-ddl) [See this test on roachdash](https://roachdash.crdb.dev/?filter=status%3Aopen+t%3A.%2Asqlsmith%2Fsetup%3Dtpch-sf1%2Fsetting%3Dno-ddl.%2A&sort=title&restgroup=false&display=lastcommented+project) <sub>powered by [pkg/cmd/internal/issues](https://github.com/cockroachdb/cockroach/tree/master/pkg/cmd/internal/issues)</sub></p></details>
test
roachtest sqlsmith setup tpch setting no ddl failed on main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types withstack withstack errutil leaferror wraps secondary error attachment dead attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types withstack withstack errutil leaferror wraps attached stack trace stack trace main glob home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go main wrap home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command executec home agent work go src github com cockroachdb cockroach vendor github com cobra command go github com cobra command execute home agent work go src github com cockroachdb cockroach vendor github com cobra command go main main home agent work go src github com cockroachdb cockroach pkg cmd roachprod main go runtime main usr local go src runtime proc go runtime goexit usr local go src runtime asm s wraps dead error types errors unclassified secondary withsecondaryerror secondary withsecondaryerror withstack withstack errutil leaferror more artifacts powered by
1
54,912
13,942,790,681
IssuesEvent
2020-10-22 21:39:51
Whizkevina/uchi-sidebar-clone
https://api.github.com/repos/Whizkevina/uchi-sidebar-clone
opened
CVE-2018-11499 (High) detected in node-sassv4.13.1
security vulnerability
## CVE-2018-11499 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.13.1</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Whizkevina/uchi-sidebar-clone/commit/5405eeecb088ab7acf45ef51e052988d72c3fe7f">5405eeecb088ab7acf45ef51e052988d72c3fe7f</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>uchi-sidebar-clone/node_modules/node-sass/src/libsass/src/parser.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free vulnerability exists in handle_error() in sass_context.cpp in LibSass 3.4.x and 3.5.x through 3.5.4 that could be leveraged to cause a denial of service (application crash) or possibly unspecified other impact. <p>Publish Date: 2018-05-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11499>CVE-2018-11499</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11499">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11499</a></p> <p>Release Date: 2018-05-26</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2018-11499 (High) detected in node-sassv4.13.1 - ## CVE-2018-11499 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>node-sassv4.13.1</b></p></summary> <p> <p>:rainbow: Node.js bindings to libsass</p> <p>Library home page: <a href=https://github.com/sass/node-sass.git>https://github.com/sass/node-sass.git</a></p> <p>Found in HEAD commit: <a href="https://github.com/Whizkevina/uchi-sidebar-clone/commit/5405eeecb088ab7acf45ef51e052988d72c3fe7f">5405eeecb088ab7acf45ef51e052988d72c3fe7f</a></p> <p>Found in base branch: <b>main</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (1)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>uchi-sidebar-clone/node_modules/node-sass/src/libsass/src/parser.cpp</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> A use-after-free vulnerability exists in handle_error() in sass_context.cpp in LibSass 3.4.x and 3.5.x through 3.5.4 that could be leveraged to cause a denial of service (application crash) or possibly unspecified other impact. <p>Publish Date: 2018-05-26 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2018-11499>CVE-2018-11499</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>9.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11499">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2018-11499</a></p> <p>Release Date: 2018-05-26</p> <p>Fix Resolution: LibSass - 3.6.0</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in node cve high severity vulnerability vulnerable library node rainbow node js bindings to libsass library home page a href found in head commit a href found in base branch main vulnerable source files uchi sidebar clone node modules node sass src libsass src parser cpp vulnerability details a use after free vulnerability exists in handle error in sass context cpp in libsass x and x through that could be leveraged to cause a denial of service application crash or possibly unspecified other impact publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution libsass step up your open source security game with whitesource
0
45,018
9,665,761,619
IssuesEvent
2019-05-21 09:14:57
feelpp/feelpp
https://api.github.com/repos/feelpp/feelpp
closed
Warning: explicitly defaulted copy assignment operator is implicitly deleted
type:clean-code type:refactoring
We need to be careful with defaulted constructors as they might be implictly deleted due to members which are reference ``` make -j30 feelpp_qs_hdg_laplacian_3d/fast Building CXX object feelpp/quickstart/CMakeFiles/feelpp_qs_hdg_laplacian_3d.dir/qs_hdg_laplacian.cpp.o In file included from /home/prudhomm/Devel/feelpp.hdg/feelpp/quickstart/qs_hdg_laplacian.cpp:24: In file included from /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feel.hpp:47: In file included from /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/pde.hpp:27: /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/boundaryconditions.hpp:187:25: warning: explicitly defaulted copy assignment operator is implicitly deleted [-Wdefaulted-function-deleted] BoundaryConditions& operator=( BoundaryConditions const& bc ) = default; ^ /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/boundaryconditions.hpp:379:22: note: copy assignment operator of 'BoundaryConditions' is implicitly deleted because field 'M_worldComm' is of reference type 'const Feel::WorldComm &' WorldComm const& M_worldComm; ^ /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/boundaryconditions.hpp:188:25: warning: explicitly defaulted move assignment operator is implicitly deleted [-Wdefaulted-function-deleted] BoundaryConditions& operator=( BoundaryConditions && bc ) = default; ^ /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/boundaryconditions.hpp:379:22: note: move assignment operator of 'BoundaryConditions' is implicitly deleted because field 'M_worldComm' is of reference type 'const Feel::WorldComm &' WorldComm const& M_worldComm; ``` @vincentchabannes we get this message using clang-8, we need to clean this up for 107 I think and avoid implicitly deleted constructors In the case above I suggest doing whqt I did with CommObject and use eg `std::shared_ptr`
1.0
Warning: explicitly defaulted copy assignment operator is implicitly deleted - We need to be careful with defaulted constructors as they might be implictly deleted due to members which are reference ``` make -j30 feelpp_qs_hdg_laplacian_3d/fast Building CXX object feelpp/quickstart/CMakeFiles/feelpp_qs_hdg_laplacian_3d.dir/qs_hdg_laplacian.cpp.o In file included from /home/prudhomm/Devel/feelpp.hdg/feelpp/quickstart/qs_hdg_laplacian.cpp:24: In file included from /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feel.hpp:47: In file included from /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/pde.hpp:27: /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/boundaryconditions.hpp:187:25: warning: explicitly defaulted copy assignment operator is implicitly deleted [-Wdefaulted-function-deleted] BoundaryConditions& operator=( BoundaryConditions const& bc ) = default; ^ /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/boundaryconditions.hpp:379:22: note: copy assignment operator of 'BoundaryConditions' is implicitly deleted because field 'M_worldComm' is of reference type 'const Feel::WorldComm &' WorldComm const& M_worldComm; ^ /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/boundaryconditions.hpp:188:25: warning: explicitly defaulted move assignment operator is implicitly deleted [-Wdefaulted-function-deleted] BoundaryConditions& operator=( BoundaryConditions && bc ) = default; ^ /home/prudhomm/Devel/feelpp.hdg/feelpp/feel/../feel/feelpde/boundaryconditions.hpp:379:22: note: move assignment operator of 'BoundaryConditions' is implicitly deleted because field 'M_worldComm' is of reference type 'const Feel::WorldComm &' WorldComm const& M_worldComm; ``` @vincentchabannes we get this message using clang-8, we need to clean this up for 107 I think and avoid implicitly deleted constructors In the case above I suggest doing whqt I did with CommObject and use eg `std::shared_ptr`
non_test
warning explicitly defaulted copy assignment operator is implicitly deleted we need to be careful with defaulted constructors as they might be implictly deleted due to members which are reference make feelpp qs hdg laplacian fast building cxx object feelpp quickstart cmakefiles feelpp qs hdg laplacian dir qs hdg laplacian cpp o in file included from home prudhomm devel feelpp hdg feelpp quickstart qs hdg laplacian cpp in file included from home prudhomm devel feelpp hdg feelpp feel feel feel hpp in file included from home prudhomm devel feelpp hdg feelpp feel feel feelpde pde hpp home prudhomm devel feelpp hdg feelpp feel feel feelpde boundaryconditions hpp warning explicitly defaulted copy assignment operator is implicitly deleted boundaryconditions operator boundaryconditions const bc default home prudhomm devel feelpp hdg feelpp feel feel feelpde boundaryconditions hpp note copy assignment operator of boundaryconditions is implicitly deleted because field m worldcomm is of reference type const feel worldcomm worldcomm const m worldcomm home prudhomm devel feelpp hdg feelpp feel feel feelpde boundaryconditions hpp warning explicitly defaulted move assignment operator is implicitly deleted boundaryconditions operator boundaryconditions bc default home prudhomm devel feelpp hdg feelpp feel feel feelpde boundaryconditions hpp note move assignment operator of boundaryconditions is implicitly deleted because field m worldcomm is of reference type const feel worldcomm worldcomm const m worldcomm vincentchabannes we get this message using clang we need to clean this up for i think and avoid implicitly deleted constructors in the case above i suggest doing whqt i did with commobject and use eg std shared ptr
0
170,593
13,194,198,137
IssuesEvent
2020-08-13 16:24:42
bitcoin/bitcoin
https://api.github.com/repos/bitcoin/bitcoin
closed
example_test.py doesn't work with Python 3.5
Bug Tests
<!-- This issue tracker is only for technical issues related to Bitcoin Core. General bitcoin questions and/or support requests are best directed to the Bitcoin StackExchange at https://bitcoin.stackexchange.com. For reporting security issues, please read instructions at https://bitcoincore.org/en/contact/. If the node is "stuck" during sync or giving "block checksum mismatch" errors, please ensure your hardware is stable by running memtest and observe CPU temperature with a load-test tool such as linpack before creating an issue! --> <!-- Describe the issue --> **Expected behavior** The `example_test.py` functional test should run without errors with Python 3.5.0 since this is the minimum supported version. <!--- What behavior did you expect? --> **Actual behavior** Test fails with `AttributeError: '_SelectorSocketTransport' object has no attribute 'is_closing'` <!--- What was the actual behavior (provide screenshots if the issue is GUI-related)? --> **To reproduce** Run `example_test.py` with Python 3.5.0. <!--- How reliably can you reproduce the issue, what are the steps to do so? --> **System information** Latest version of Bitcoin Core compiled on macOS 10.15.5. <!-- What version of Bitcoin Core are you using, where did you get it (website, self-compiled, etc)? --> <!-- What type of machine are you observing the error on (OS/CPU and disk type)? --> <!-- GUI-related issue? What is your operating system and its version? If Linux, what is your desktop environment and graphical shell? --> <!-- Any extra information that might be useful in the debugging process. --> <!--- This is normally the contents of a `debug.log` or `config.log` file. Raw text or a link to a pastebin type site are preferred. --> PR #13747 fixed the issue but it seems like the Python documentation incorrectly says that `is_closing()` was added to `BaseTransport` in 3.4.4 ([here](https://docs.python.org/3.4/library/asyncio-protocol.html#asyncio.BaseTransport.is_closing)), but then on another page on the Python docs ([here](https://docs.python.org/3.5/library/asyncio-protocol.html#asyncio.BaseTransport.is_closing)) it says it was included in Python 3.5.1. The previous fix was reverted when moving to 3.5, but it seems like that may have reintroduced the problem. I'm not sure how soon the minimum Python version will be increased, so not sure if fixing this is important.
1.0
example_test.py doesn't work with Python 3.5 - <!-- This issue tracker is only for technical issues related to Bitcoin Core. General bitcoin questions and/or support requests are best directed to the Bitcoin StackExchange at https://bitcoin.stackexchange.com. For reporting security issues, please read instructions at https://bitcoincore.org/en/contact/. If the node is "stuck" during sync or giving "block checksum mismatch" errors, please ensure your hardware is stable by running memtest and observe CPU temperature with a load-test tool such as linpack before creating an issue! --> <!-- Describe the issue --> **Expected behavior** The `example_test.py` functional test should run without errors with Python 3.5.0 since this is the minimum supported version. <!--- What behavior did you expect? --> **Actual behavior** Test fails with `AttributeError: '_SelectorSocketTransport' object has no attribute 'is_closing'` <!--- What was the actual behavior (provide screenshots if the issue is GUI-related)? --> **To reproduce** Run `example_test.py` with Python 3.5.0. <!--- How reliably can you reproduce the issue, what are the steps to do so? --> **System information** Latest version of Bitcoin Core compiled on macOS 10.15.5. <!-- What version of Bitcoin Core are you using, where did you get it (website, self-compiled, etc)? --> <!-- What type of machine are you observing the error on (OS/CPU and disk type)? --> <!-- GUI-related issue? What is your operating system and its version? If Linux, what is your desktop environment and graphical shell? --> <!-- Any extra information that might be useful in the debugging process. --> <!--- This is normally the contents of a `debug.log` or `config.log` file. Raw text or a link to a pastebin type site are preferred. --> PR #13747 fixed the issue but it seems like the Python documentation incorrectly says that `is_closing()` was added to `BaseTransport` in 3.4.4 ([here](https://docs.python.org/3.4/library/asyncio-protocol.html#asyncio.BaseTransport.is_closing)), but then on another page on the Python docs ([here](https://docs.python.org/3.5/library/asyncio-protocol.html#asyncio.BaseTransport.is_closing)) it says it was included in Python 3.5.1. The previous fix was reverted when moving to 3.5, but it seems like that may have reintroduced the problem. I'm not sure how soon the minimum Python version will be increased, so not sure if fixing this is important.
test
example test py doesn t work with python this issue tracker is only for technical issues related to bitcoin core general bitcoin questions and or support requests are best directed to the bitcoin stackexchange at for reporting security issues please read instructions at if the node is stuck during sync or giving block checksum mismatch errors please ensure your hardware is stable by running memtest and observe cpu temperature with a load test tool such as linpack before creating an issue expected behavior the example test py functional test should run without errors with python since this is the minimum supported version actual behavior test fails with attributeerror selectorsockettransport object has no attribute is closing to reproduce run example test py with python system information latest version of bitcoin core compiled on macos pr fixed the issue but it seems like the python documentation incorrectly says that is closing was added to basetransport in but then on another page on the python docs it says it was included in python the previous fix was reverted when moving to but it seems like that may have reintroduced the problem i m not sure how soon the minimum python version will be increased so not sure if fixing this is important
1
51,725
12,798,926,494
IssuesEvent
2020-07-02 14:38:33
SJSU-Dev2/SJSU-Dev2
https://api.github.com/repos/SJSU-Dev2/SJSU-Dev2
closed
Rewrite build system makefile
Build
# Feature Request Build system should double down on the idea of unit builds. Build system should be as simple as possible.
1.0
Rewrite build system makefile - # Feature Request Build system should double down on the idea of unit builds. Build system should be as simple as possible.
non_test
rewrite build system makefile feature request build system should double down on the idea of unit builds build system should be as simple as possible
0
50,021
13,187,308,930
IssuesEvent
2020-08-13 03:00:16
icecube-trac/tix3
https://api.github.com/repos/icecube-trac/tix3
closed
glshovel docs and plugin examples (Trac #15)
Migrated from Trac defect glshovel
<details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/15 , reported by troy and owned by troy_</summary> <p> ```json { "status": "closed", "changetime": "2007-11-11T03:51:18", "description": "\n", "reporter": "troy", "cc": "", "resolution": "duplicate", "_ts": "1194753078000000", "component": "glshovel", "summary": "glshovel docs and plugin examples", "priority": "normal", "keywords": "", "time": "2007-06-03T16:33:00", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
1.0
glshovel docs and plugin examples (Trac #15) - <details> <summary>_Migrated from https://code.icecube.wisc.edu/ticket/15 , reported by troy and owned by troy_</summary> <p> ```json { "status": "closed", "changetime": "2007-11-11T03:51:18", "description": "\n", "reporter": "troy", "cc": "", "resolution": "duplicate", "_ts": "1194753078000000", "component": "glshovel", "summary": "glshovel docs and plugin examples", "priority": "normal", "keywords": "", "time": "2007-06-03T16:33:00", "milestone": "", "owner": "troy", "type": "defect" } ``` </p> </details>
non_test
glshovel docs and plugin examples trac migrated from reported by troy and owned by troy json status closed changetime description n reporter troy cc resolution duplicate ts component glshovel summary glshovel docs and plugin examples priority normal keywords time milestone owner troy type defect
0
237,749
19,671,320,610
IssuesEvent
2022-01-11 07:39:46
jhipster/generator-jhipster
https://api.github.com/repos/jhipster/generator-jhipster
reopened
Unit test fails on Windows 10
area: needs-reproduction theme: windows theme: tests area: stale
<!-- - Please follow the issue template below for bug reports. - If you have a support request rather than a bug, please use [Stack Overflow](http://stackoverflow.com/questions/tagged/jhipster) with the JHipster tag. - For bug reports it is mandatory to run the command `jhipster info` in your project's root folder, and paste the result here. - Tickets opened without any of these pieces of information will be **closed** without any explanation. --> ##### **Overview of the issue** The following command fail on my environment due an error with the unit test ``` mvnw -ntp -Pprod verify jib:dockerBuild [ERROR] Failures: [ERROR] WebConfigurerTest.testCustomizeServletContainer:83 Expecting: <C:\jhipster\demo\target\classes\static> to be equal to: <c:\jhipster\demo\target\classes\static> but was not. ``` Note the "C" on upper case at the begin. ##### **Related issues** Similar to #5805 see also #5860 ##### **Suggest a Fix** I have fixed the test by hand but I'm not sure if is good enough to send a PR. ```java assertThat(container.getDocumentRoot().getAbsolutePath()).isEqualToIgnoringCase(new File("target/classes/static/").getAbsolutePath()); ``` ##### **JHipster Version(s)** 6.10.5 ##### **Browsers and Operating System** <!-- What OS are you on? is this a problem with all browsers or only IE8? --> - [ x] Checking this box is mandatory (this is just to show you read everything) <!-- Love JHipster? Please consider supporting our collective: 👉 https://opencollective.com/generator-jhipster/donate -->
1.0
Unit test fails on Windows 10 - <!-- - Please follow the issue template below for bug reports. - If you have a support request rather than a bug, please use [Stack Overflow](http://stackoverflow.com/questions/tagged/jhipster) with the JHipster tag. - For bug reports it is mandatory to run the command `jhipster info` in your project's root folder, and paste the result here. - Tickets opened without any of these pieces of information will be **closed** without any explanation. --> ##### **Overview of the issue** The following command fail on my environment due an error with the unit test ``` mvnw -ntp -Pprod verify jib:dockerBuild [ERROR] Failures: [ERROR] WebConfigurerTest.testCustomizeServletContainer:83 Expecting: <C:\jhipster\demo\target\classes\static> to be equal to: <c:\jhipster\demo\target\classes\static> but was not. ``` Note the "C" on upper case at the begin. ##### **Related issues** Similar to #5805 see also #5860 ##### **Suggest a Fix** I have fixed the test by hand but I'm not sure if is good enough to send a PR. ```java assertThat(container.getDocumentRoot().getAbsolutePath()).isEqualToIgnoringCase(new File("target/classes/static/").getAbsolutePath()); ``` ##### **JHipster Version(s)** 6.10.5 ##### **Browsers and Operating System** <!-- What OS are you on? is this a problem with all browsers or only IE8? --> - [ x] Checking this box is mandatory (this is just to show you read everything) <!-- Love JHipster? Please consider supporting our collective: 👉 https://opencollective.com/generator-jhipster/donate -->
test
unit test fails on windows please follow the issue template below for bug reports if you have a support request rather than a bug please use with the jhipster tag for bug reports it is mandatory to run the command jhipster info in your project s root folder and paste the result here tickets opened without any of these pieces of information will be closed without any explanation overview of the issue the following command fail on my environment due an error with the unit test mvnw ntp pprod verify jib dockerbuild failures webconfigurertest testcustomizeservletcontainer expecting to be equal to but was not note the c on upper case at the begin related issues similar to see also suggest a fix i have fixed the test by hand but i m not sure if is good enough to send a pr java assertthat container getdocumentroot getabsolutepath isequaltoignoringcase new file target classes static getabsolutepath jhipster version s browsers and operating system checking this box is mandatory this is just to show you read everything love jhipster please consider supporting our collective 👉
1
823,897
31,073,167,695
IssuesEvent
2023-08-12 06:29:21
python/mypy
https://api.github.com/repos/python/mypy
closed
yield from inside list comprehension gives wrong type
bug priority-2-low topic-runtime-semantics
**Bug Report** runtime and mypy types don't match, possibly related to #6457 **To Reproduce** ``` from typing import Generator def f() -> Generator[int, None, str]: yield 1 return "lol" def g() -> Generator[int, None, None]: x = [(yield from f()) for lel in range(5)] print(x) # <generator object g.<locals>.<listcomp> at 0x7fba92696a20> reveal_type(x) # List[str] g() ``` **Expected Behavior** `reveal_type` outputting `Generator[something]` to match runtime **Actual Behavior** `List[str]` **Your Environment** - Mypy version used: 0.812 - Mypy command-line flags: none - Mypy configuration options from `mypy.ini` (and other config files): none - Python version used: 3.7.3
1.0
yield from inside list comprehension gives wrong type - **Bug Report** runtime and mypy types don't match, possibly related to #6457 **To Reproduce** ``` from typing import Generator def f() -> Generator[int, None, str]: yield 1 return "lol" def g() -> Generator[int, None, None]: x = [(yield from f()) for lel in range(5)] print(x) # <generator object g.<locals>.<listcomp> at 0x7fba92696a20> reveal_type(x) # List[str] g() ``` **Expected Behavior** `reveal_type` outputting `Generator[something]` to match runtime **Actual Behavior** `List[str]` **Your Environment** - Mypy version used: 0.812 - Mypy command-line flags: none - Mypy configuration options from `mypy.ini` (and other config files): none - Python version used: 3.7.3
non_test
yield from inside list comprehension gives wrong type bug report runtime and mypy types don t match possibly related to to reproduce from typing import generator def f generator yield return lol def g generator x print x at reveal type x list g expected behavior reveal type outputting generator to match runtime actual behavior list your environment mypy version used mypy command line flags none mypy configuration options from mypy ini and other config files none python version used
0
802,813
29,046,780,650
IssuesEvent
2023-05-13 17:10:57
aiyu-ayaan/BIT-App
https://api.github.com/repos/aiyu-ayaan/BIT-App
reopened
JDK Version issue in GitHub Action
bug priority high
```gradle Caused by: com.android.builder.errors.EvalIssueException: Android Gradle plugin requires Java 17 to run. You are currently using Java 11. Your current JDK is located in /opt/hostedtoolcache/jdk/11.0.19/x64 You can try some of the following options: - changing the IDE settings. - changing the JAVA_HOME environment variable. - changing `org.gradle.java.home` in `gradle.properties`. at com.android.builder.errors.IssueReporter.reportError(IssueReporter.kt:114) at com.android.builder.errors.IssueReporter.reportError$default(IssueReporter.kt:110) at com.android.build.gradle.internal.plugins.AndroidPluginBaseServices.checkMinJvmVersion(AndroidPluginBaseServices.kt:[278](https://github.com/aiyu-ayaan/BIT-App/actions/runs/4950968295/jobs/8855304716#step:8:278)) at com.android.build.gradle.internal.plugins.AndroidPluginBaseServices.basePluginApply(AndroidPluginBaseServices.kt:121) at com.android.build.gradle.internal.plugins.BasePlugin$apply$1.run(BasePlugin.kt:350) at com.android.build.gradle.internal.crash.CrashReporting.runAction(crash_reporting.kt:27) at com.android.build.gradle.internal.plugins.BasePlugin.apply(BasePlugin.kt:349) at com.android.build.gradle.internal.plugins.BasePlugin.apply(BasePlugin.kt:129) at org.gradle.api.internal.plugins.ImperativeOnlyPluginTarget.applyImperative(ImperativeOnlyPluginTarget.java:43) at org.gradle.api.internal.plugins.RuleBasedPluginTarget.applyImperative(RuleBasedPluginTarget.java:51) at org.gradle.api.internal.plugins.DefaultPluginManager.addPlugin(DefaultPluginManager.java:187) at org.gradle.api.internal.plugins.DefaultPluginManager.access$100(DefaultPluginManager.java:52) at org.gradle.api.internal.plugins.DefaultPluginManager$AddPluginBuildOperation.run(DefaultPluginManager.java:[282](https://github.com/aiyu-ayaan/BIT-App/actions/runs/4950968295/jobs/8855304716#step:8:282)) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.api.internal.plugins.DefaultPluginManager.lambda$doApply$0(DefaultPluginManager.java:167) at org.gradle.configuration.internal.DefaultUserCodeApplicationContext.apply(DefaultUserCodeApplicationContext.java:44) at org.gradle.api.internal.plugins.DefaultPluginManager.doApply(DefaultPluginManager.java:166) ... 201 more ```
1.0
JDK Version issue in GitHub Action - ```gradle Caused by: com.android.builder.errors.EvalIssueException: Android Gradle plugin requires Java 17 to run. You are currently using Java 11. Your current JDK is located in /opt/hostedtoolcache/jdk/11.0.19/x64 You can try some of the following options: - changing the IDE settings. - changing the JAVA_HOME environment variable. - changing `org.gradle.java.home` in `gradle.properties`. at com.android.builder.errors.IssueReporter.reportError(IssueReporter.kt:114) at com.android.builder.errors.IssueReporter.reportError$default(IssueReporter.kt:110) at com.android.build.gradle.internal.plugins.AndroidPluginBaseServices.checkMinJvmVersion(AndroidPluginBaseServices.kt:[278](https://github.com/aiyu-ayaan/BIT-App/actions/runs/4950968295/jobs/8855304716#step:8:278)) at com.android.build.gradle.internal.plugins.AndroidPluginBaseServices.basePluginApply(AndroidPluginBaseServices.kt:121) at com.android.build.gradle.internal.plugins.BasePlugin$apply$1.run(BasePlugin.kt:350) at com.android.build.gradle.internal.crash.CrashReporting.runAction(crash_reporting.kt:27) at com.android.build.gradle.internal.plugins.BasePlugin.apply(BasePlugin.kt:349) at com.android.build.gradle.internal.plugins.BasePlugin.apply(BasePlugin.kt:129) at org.gradle.api.internal.plugins.ImperativeOnlyPluginTarget.applyImperative(ImperativeOnlyPluginTarget.java:43) at org.gradle.api.internal.plugins.RuleBasedPluginTarget.applyImperative(RuleBasedPluginTarget.java:51) at org.gradle.api.internal.plugins.DefaultPluginManager.addPlugin(DefaultPluginManager.java:187) at org.gradle.api.internal.plugins.DefaultPluginManager.access$100(DefaultPluginManager.java:52) at org.gradle.api.internal.plugins.DefaultPluginManager$AddPluginBuildOperation.run(DefaultPluginManager.java:[282](https://github.com/aiyu-ayaan/BIT-App/actions/runs/4950968295/jobs/8855304716#step:8:282)) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:29) at org.gradle.internal.operations.DefaultBuildOperationRunner$1.execute(DefaultBuildOperationRunner.java:26) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:66) at org.gradle.internal.operations.DefaultBuildOperationRunner$2.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:157) at org.gradle.internal.operations.DefaultBuildOperationRunner.execute(DefaultBuildOperationRunner.java:59) at org.gradle.internal.operations.DefaultBuildOperationRunner.run(DefaultBuildOperationRunner.java:47) at org.gradle.internal.operations.DefaultBuildOperationExecutor.run(DefaultBuildOperationExecutor.java:68) at org.gradle.api.internal.plugins.DefaultPluginManager.lambda$doApply$0(DefaultPluginManager.java:167) at org.gradle.configuration.internal.DefaultUserCodeApplicationContext.apply(DefaultUserCodeApplicationContext.java:44) at org.gradle.api.internal.plugins.DefaultPluginManager.doApply(DefaultPluginManager.java:166) ... 201 more ```
non_test
jdk version issue in github action gradle caused by com android builder errors evalissueexception android gradle plugin requires java to run you are currently using java your current jdk is located in opt hostedtoolcache jdk you can try some of the following options changing the ide settings changing the java home environment variable changing org gradle java home in gradle properties at com android builder errors issuereporter reporterror issuereporter kt at com android builder errors issuereporter reporterror default issuereporter kt at com android build gradle internal plugins androidpluginbaseservices checkminjvmversion androidpluginbaseservices kt at com android build gradle internal plugins androidpluginbaseservices basepluginapply androidpluginbaseservices kt at com android build gradle internal plugins baseplugin apply run baseplugin kt at com android build gradle internal crash crashreporting runaction crash reporting kt at com android build gradle internal plugins baseplugin apply baseplugin kt at com android build gradle internal plugins baseplugin apply baseplugin kt at org gradle api internal plugins imperativeonlyplugintarget applyimperative imperativeonlyplugintarget java at org gradle api internal plugins rulebasedplugintarget applyimperative rulebasedplugintarget java at org gradle api internal plugins defaultpluginmanager addplugin defaultpluginmanager java at org gradle api internal plugins defaultpluginmanager access defaultpluginmanager java at org gradle api internal plugins defaultpluginmanager addpluginbuildoperation run defaultpluginmanager java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner execute defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationrunner run defaultbuildoperationrunner java at org gradle internal operations defaultbuildoperationexecutor run defaultbuildoperationexecutor java at org gradle api internal plugins defaultpluginmanager lambda doapply defaultpluginmanager java at org gradle configuration internal defaultusercodeapplicationcontext apply defaultusercodeapplicationcontext java at org gradle api internal plugins defaultpluginmanager doapply defaultpluginmanager java more
0
5,688
7,307,620,952
IssuesEvent
2018-02-28 03:48:26
MicrosoftDocs/azure-docs
https://api.github.com/repos/MicrosoftDocs/azure-docs
closed
Typo: NGIX should be NGINX
bug container-service cxp triaged
The sentence "This document walks through a sample deployment of the NGIX ingress controller in an Azure Container Service (AKS) cluster" at the top of the page contains a typo. It should read: "This document walks through a sample deployment of the NGINX ingress controller in an Azure Container Service (AKS) cluster" --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ada58160-dc50-b977-d3bc-65f92406e16d * Version Independent ID: 5c596e04-fddb-effe-3b72-575f4df42d4e * [Content](https://docs.microsoft.com/en-us/azure/aks/ingress) * [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/ingress.md) * Service: container-service
1.0
Typo: NGIX should be NGINX - The sentence "This document walks through a sample deployment of the NGIX ingress controller in an Azure Container Service (AKS) cluster" at the top of the page contains a typo. It should read: "This document walks through a sample deployment of the NGINX ingress controller in an Azure Container Service (AKS) cluster" --- #### Document Details ⚠ *Do not edit this section. It is required for docs.microsoft.com ➟ GitHub issue linking.* * ID: ada58160-dc50-b977-d3bc-65f92406e16d * Version Independent ID: 5c596e04-fddb-effe-3b72-575f4df42d4e * [Content](https://docs.microsoft.com/en-us/azure/aks/ingress) * [Content Source](https://github.com/Microsoft/azure-docs/blob/master/articles/aks/ingress.md) * Service: container-service
non_test
typo ngix should be nginx the sentence this document walks through a sample deployment of the ngix ingress controller in an azure container service aks cluster at the top of the page contains a typo it should read this document walks through a sample deployment of the nginx ingress controller in an azure container service aks cluster document details ⚠ do not edit this section it is required for docs microsoft com ➟ github issue linking id version independent id fddb effe service container service
0
347,136
31,142,289,274
IssuesEvent
2023-08-16 01:44:31
dotnet/winforms
https://api.github.com/repos/dotnet/winforms
closed
We need tests for calling COM interface methods from c++
test-enhancement
We need tests for various interfaces implemented by managed classes (e.g. IOleControlSite etc) The best way to test this would be to implement tests in c++ and PInvoke them passing the managed class as an interface as an a IntPtr to unmanaged code. We’d then call functions from c++ and test things like HRESULTS etc I have a prototype but not sure how to integrate a test c++ dll into the product infrastructure
1.0
We need tests for calling COM interface methods from c++ - We need tests for various interfaces implemented by managed classes (e.g. IOleControlSite etc) The best way to test this would be to implement tests in c++ and PInvoke them passing the managed class as an interface as an a IntPtr to unmanaged code. We’d then call functions from c++ and test things like HRESULTS etc I have a prototype but not sure how to integrate a test c++ dll into the product infrastructure
test
we need tests for calling com interface methods from c we need tests for various interfaces implemented by managed classes e g iolecontrolsite etc the best way to test this would be to implement tests in c and pinvoke them passing the managed class as an interface as an a intptr to unmanaged code we’d then call functions from c and test things like hresults etc i have a prototype but not sure how to integrate a test c dll into the product infrastructure
1
667,659
22,495,582,416
IssuesEvent
2022-06-23 07:15:45
projectdiscovery/interactsh
https://api.github.com/repos/projectdiscovery/interactsh
closed
Issue with Interactsh (dns) server in Cloud VMs with no public IP
Priority: High Status: Completed Type: Bug
### Interactsh version: dev,latest ### Current Behavior: ### Expected Behavior: From: https://github.com/projectdiscovery/interactsh/pull/181 if a `listen ip` wasn't specified the public IP looked up from a **what is my IP** API is used. In many cases cloud VMs do not have their **public** IP bound to an interface (such as Amazon EC2, GCP), causing startup to fail. ### Steps To Reproduce: - Create EC2/GCP box and run interactsh server with configured domain ```bash $ ./interactsh-server -domain example.com [FTL] Could not listen for udp DNS on 203.0.113.42:53 (listen udp 203.0.113.42:53: bind: cannot assign requested address) ``` ### Anything else: <img width="906" alt="image" src="https://user-images.githubusercontent.com/8293321/174589046-bf18c20d-b387-471e-a00d-3d64d1557350.png"> Attempted fix: https://github.com/projectdiscovery/interactsh/pull/181 which causes issues with the regular environment, so while looking for a fix, following points need to be considered: - DNS server works in regular environment (publicly available IP mapped with interface) - DNS server works with Cloud VMs **where no public IP mapped to interface**
1.0
Issue with Interactsh (dns) server in Cloud VMs with no public IP - ### Interactsh version: dev,latest ### Current Behavior: ### Expected Behavior: From: https://github.com/projectdiscovery/interactsh/pull/181 if a `listen ip` wasn't specified the public IP looked up from a **what is my IP** API is used. In many cases cloud VMs do not have their **public** IP bound to an interface (such as Amazon EC2, GCP), causing startup to fail. ### Steps To Reproduce: - Create EC2/GCP box and run interactsh server with configured domain ```bash $ ./interactsh-server -domain example.com [FTL] Could not listen for udp DNS on 203.0.113.42:53 (listen udp 203.0.113.42:53: bind: cannot assign requested address) ``` ### Anything else: <img width="906" alt="image" src="https://user-images.githubusercontent.com/8293321/174589046-bf18c20d-b387-471e-a00d-3d64d1557350.png"> Attempted fix: https://github.com/projectdiscovery/interactsh/pull/181 which causes issues with the regular environment, so while looking for a fix, following points need to be considered: - DNS server works in regular environment (publicly available IP mapped with interface) - DNS server works with Cloud VMs **where no public IP mapped to interface**
non_test
issue with interactsh dns server in cloud vms with no public ip interactsh version dev latest current behavior expected behavior from if a listen ip wasn t specified the public ip looked up from a what is my ip api is used in many cases cloud vms do not have their public ip bound to an interface such as amazon gcp causing startup to fail steps to reproduce create gcp box and run interactsh server with configured domain bash interactsh server domain example com could not listen for udp dns on listen udp bind cannot assign requested address anything else img width alt image src attempted fix which causes issues with the regular environment so while looking for a fix following points need to be considered dns server works in regular environment publicly available ip mapped with interface dns server works with cloud vms where no public ip mapped to interface
0
178,237
6,601,523,189
IssuesEvent
2017-09-18 01:35:29
sussol/mobile
https://api.github.com/repos/sussol/mobile
opened
No icons on main menu page
Bug Priority: Low
Build Number: 2.0.0 Description: There are no icons on on main menu ![screenshot_20170918-133144](https://user-images.githubusercontent.com/7684221/30526853-ded09b78-9c75-11e7-84ed-0fb45916d13f.png) Reproducible: Reproduction Steps: 1. Tab A 2016 2. Install 2.0.0 3. Sync a site with 200k records 4. No icons? Comments: App functionality seems fine, maybe a bit slower than usual. Though, it is working on a rather large realm database.
1.0
No icons on main menu page - Build Number: 2.0.0 Description: There are no icons on on main menu ![screenshot_20170918-133144](https://user-images.githubusercontent.com/7684221/30526853-ded09b78-9c75-11e7-84ed-0fb45916d13f.png) Reproducible: Reproduction Steps: 1. Tab A 2016 2. Install 2.0.0 3. Sync a site with 200k records 4. No icons? Comments: App functionality seems fine, maybe a bit slower than usual. Though, it is working on a rather large realm database.
non_test
no icons on main menu page build number description there are no icons on on main menu reproducible reproduction steps tab a install sync a site with records no icons comments app functionality seems fine maybe a bit slower than usual though it is working on a rather large realm database
0
207,816
15,837,446,474
IssuesEvent
2021-04-06 20:48:01
ansible/awx
https://api.github.com/repos/ansible/awx
closed
Enable html in custom login text
component:ui priority:high state:needs_test type:feature
### summary see: https://github.com/ansible/awx/pull/7585 --> https://github.com/ansible/awx/issues/7600 - Could try to use `__dangerouslySetInnerHTML` along with `html-entities` to sanitize (like we do for job events: https://github.com/ansible/awx/blob/devel/awx/ui_next/src/screens/Job/JobOutput/JobEvent.jsx#L55) - If that doesn't work we'd probably need to first put in a request to pf-react to explicitly enable passing a `React.ReactNode` to the LoginPage component for the `textContent` prop like they do for the `helperText` prop: https://patternfly-react.surge.sh/documentation/react/components/loginpage
1.0
Enable html in custom login text - ### summary see: https://github.com/ansible/awx/pull/7585 --> https://github.com/ansible/awx/issues/7600 - Could try to use `__dangerouslySetInnerHTML` along with `html-entities` to sanitize (like we do for job events: https://github.com/ansible/awx/blob/devel/awx/ui_next/src/screens/Job/JobOutput/JobEvent.jsx#L55) - If that doesn't work we'd probably need to first put in a request to pf-react to explicitly enable passing a `React.ReactNode` to the LoginPage component for the `textContent` prop like they do for the `helperText` prop: https://patternfly-react.surge.sh/documentation/react/components/loginpage
test
enable html in custom login text summary see could try to use dangerouslysetinnerhtml along with html entities to sanitize like we do for job events if that doesn t work we d probably need to first put in a request to pf react to explicitly enable passing a react reactnode to the loginpage component for the textcontent prop like they do for the helpertext prop
1
167,615
20,726,265,282
IssuesEvent
2022-03-14 02:30:29
kapseliboi/Datascope
https://api.github.com/repos/kapseliboi/Datascope
opened
CVE-2020-8203 (High) detected in multiple libraries
security vulnerability
## CVE-2020-8203 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-3.10.1.tgz</b>, <b>lodash-0.9.2.tgz</b>, <b>lodash-1.0.2.tgz</b>, <b>lodash-2.4.2.tgz</b>, <b>lodash-4.17.11.tgz</b></p></summary> <p> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/grunt-browserify/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-browserify-3.8.0.tgz (Root Library) - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-0.9.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-0.9.2.tgz">https://registry.npmjs.org/lodash/-/lodash-0.9.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/grunt-legacy-util/node_modules/lodash/package.json,/node_modules/grunt/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-0.4.5.tgz (Root Library) - :x: **lodash-0.9.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-1.0.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/globule/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-watch-0.6.1.tgz (Root Library) - gaze-0.5.2.tgz - globule-0.1.0.tgz - :x: **lodash-1.0.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/grunt-legacy-log-utils/node_modules/lodash/package.json,/node_modules/grunt-legacy-log/node_modules/lodash/package.json,/node_modules/grunt-contrib-watch/node_modules/lodash/package.json,/node_modules/findup-sync/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-watch-0.6.1.tgz (Root Library) - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - babel-core-6.26.3.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (grunt-browserify): 5.1.0</p><p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (grunt): 1.0.3</p><p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (grunt-contrib-watch): 1.0.1</p><p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (grunt-contrib-watch): 1.0.1</p><p>Fix Resolution (lodash): 4.17.19</p> <p>Direct dependency fix Resolution (babel-core): 7.0.0-alpha.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2020-8203 (High) detected in multiple libraries - ## CVE-2020-8203 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Libraries - <b>lodash-3.10.1.tgz</b>, <b>lodash-0.9.2.tgz</b>, <b>lodash-1.0.2.tgz</b>, <b>lodash-2.4.2.tgz</b>, <b>lodash-4.17.11.tgz</b></p></summary> <p> <details><summary><b>lodash-3.10.1.tgz</b></p></summary> <p>The modern build of lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz">https://registry.npmjs.org/lodash/-/lodash-3.10.1.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/grunt-browserify/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-browserify-3.8.0.tgz (Root Library) - :x: **lodash-3.10.1.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-0.9.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-0.9.2.tgz">https://registry.npmjs.org/lodash/-/lodash-0.9.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/grunt-legacy-util/node_modules/lodash/package.json,/node_modules/grunt/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-0.4.5.tgz (Root Library) - :x: **lodash-0.9.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-1.0.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, and extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz">https://registry.npmjs.org/lodash/-/lodash-1.0.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/globule/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-watch-0.6.1.tgz (Root Library) - gaze-0.5.2.tgz - globule-0.1.0.tgz - :x: **lodash-1.0.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-2.4.2.tgz</b></p></summary> <p>A utility library delivering consistency, customization, performance, & extras.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz">https://registry.npmjs.org/lodash/-/lodash-2.4.2.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/grunt-legacy-log-utils/node_modules/lodash/package.json,/node_modules/grunt-legacy-log/node_modules/lodash/package.json,/node_modules/grunt-contrib-watch/node_modules/lodash/package.json,/node_modules/findup-sync/node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - grunt-contrib-watch-0.6.1.tgz (Root Library) - :x: **lodash-2.4.2.tgz** (Vulnerable Library) </details> <details><summary><b>lodash-4.17.11.tgz</b></p></summary> <p>Lodash modular utilities.</p> <p>Library home page: <a href="https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz">https://registry.npmjs.org/lodash/-/lodash-4.17.11.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/lodash/package.json</p> <p> Dependency Hierarchy: - babel-core-6.26.3.tgz (Root Library) - :x: **lodash-4.17.11.tgz** (Vulnerable Library) </details> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Prototype pollution attack when using _.zipObjectDeep in lodash before 4.17.20. <p>Publish Date: 2020-07-15 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2020-8203>CVE-2020-8203</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.4</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.npmjs.com/advisories/1523">https://www.npmjs.com/advisories/1523</a></p> <p>Release Date: 2020-07-15</p> <p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (grunt-browserify): 5.1.0</p><p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (grunt): 1.0.3</p><p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (grunt-contrib-watch): 1.0.1</p><p>Fix Resolution (lodash): 4.17.9</p> <p>Direct dependency fix Resolution (grunt-contrib-watch): 1.0.1</p><p>Fix Resolution (lodash): 4.17.19</p> <p>Direct dependency fix Resolution (babel-core): 7.0.0-alpha.1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with WhiteSource [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in multiple libraries cve high severity vulnerability vulnerable libraries lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz lodash tgz the modern build of lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules grunt browserify node modules lodash package json dependency hierarchy grunt browserify tgz root library x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file package json path to vulnerable library node modules grunt legacy util node modules lodash package json node modules grunt node modules lodash package json dependency hierarchy grunt tgz root library x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance and extras library home page a href path to dependency file package json path to vulnerable library node modules globule node modules lodash package json dependency hierarchy grunt contrib watch tgz root library gaze tgz globule tgz x lodash tgz vulnerable library lodash tgz a utility library delivering consistency customization performance extras library home page a href path to dependency file package json path to vulnerable library node modules grunt legacy log utils node modules lodash package json node modules grunt legacy log node modules lodash package json node modules grunt contrib watch node modules lodash package json node modules findup sync node modules lodash package json dependency hierarchy grunt contrib watch tgz root library x lodash tgz vulnerable library lodash tgz lodash modular utilities library home page a href path to dependency file package json path to vulnerable library node modules lodash package json dependency hierarchy babel core tgz root library x lodash tgz vulnerable library found in base branch master vulnerability details prototype pollution attack when using zipobjectdeep in lodash before publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution lodash direct dependency fix resolution grunt browserify fix resolution lodash direct dependency fix resolution grunt fix resolution lodash direct dependency fix resolution grunt contrib watch fix resolution lodash direct dependency fix resolution grunt contrib watch fix resolution lodash direct dependency fix resolution babel core alpha step up your open source security game with whitesource
0
98,879
8,685,911,944
IssuesEvent
2018-12-03 09:21:15
humera987/FXLabs-Test-Automation
https://api.github.com/repos/humera987/FXLabs-Test-Automation
reopened
FX Testing 3 : ApiV1DashboardIssuetrackerIdIssuetrackerSavingsGetPathParamIdNullValue
FX Testing 3
Project : FX Testing 3 Job : UAT Env : UAT Region : US_WEST Result : fail Status Code : 404 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=OTE4YTE2N2QtODAxZS00ZmJiLWE2YTctZTdkMGU4NWMyM2Mx; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 03 Dec 2018 07:44:45 GMT]} Endpoint : http://13.56.210.25/api/v1/api/v1/dashboard/issuetracker/null/issuetracker-savings Request : Response : { "timestamp" : "2018-12-03T07:44:45.434+0000", "status" : 404, "error" : "Not Found", "message" : "No message available", "path" : "/api/v1/api/v1/dashboard/issuetracker/null/issuetracker-savings" } Logs : Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 500] resolved-to [404 != 500] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]Assertion [@StatusCode != 200] resolved-to [404 != 200] result [Passed] --- FX Bot ---
1.0
FX Testing 3 : ApiV1DashboardIssuetrackerIdIssuetrackerSavingsGetPathParamIdNullValue - Project : FX Testing 3 Job : UAT Env : UAT Region : US_WEST Result : fail Status Code : 404 Headers : {X-Content-Type-Options=[nosniff], X-XSS-Protection=[1; mode=block], Cache-Control=[no-cache, no-store, max-age=0, must-revalidate], Pragma=[no-cache], Expires=[0], X-Frame-Options=[DENY], Set-Cookie=[SESSION=OTE4YTE2N2QtODAxZS00ZmJiLWE2YTctZTdkMGU4NWMyM2Mx; Path=/; HttpOnly], Content-Type=[application/json;charset=UTF-8], Transfer-Encoding=[chunked], Date=[Mon, 03 Dec 2018 07:44:45 GMT]} Endpoint : http://13.56.210.25/api/v1/api/v1/dashboard/issuetracker/null/issuetracker-savings Request : Response : { "timestamp" : "2018-12-03T07:44:45.434+0000", "status" : 404, "error" : "Not Found", "message" : "No message available", "path" : "/api/v1/api/v1/dashboard/issuetracker/null/issuetracker-savings" } Logs : Assertion [@StatusCode != 401] resolved-to [404 != 401] result [Passed]Assertion [@StatusCode != 500] resolved-to [404 != 500] result [Passed]Assertion [@StatusCode != 404] resolved-to [404 != 404] result [Failed]Assertion [@StatusCode != 200] resolved-to [404 != 200] result [Passed] --- FX Bot ---
test
fx testing project fx testing job uat env uat region us west result fail status code headers x content type options x xss protection cache control pragma expires x frame options set cookie content type transfer encoding date endpoint request response timestamp status error not found message no message available path api api dashboard issuetracker null issuetracker savings logs assertion resolved to result assertion resolved to result assertion resolved to result assertion resolved to result fx bot
1
764,288
26,793,541,760
IssuesEvent
2023-02-01 10:13:34
python/mypy
https://api.github.com/repos/python/mypy
closed
Better error message for rejected protocol subtyping due to mutability
priority-1-normal topic-usability topic-protocols topic-error-reporting
It appears that Union types for Protocol attributes aren't recognized. Repro example: ``` from typing import Type, Union from typing_extensions import Protocol class HasFoo(Protocol): # foo: str # works foo: Union[str, int] # fails def dec(cls: Type[HasFoo]) -> Type[HasFoo]: return cls @dec class Thing: foo = 'abc' ``` Failing output: ```Argument 1 to "dec" has incompatible type "Type[Thing]"; expected "Type[HasFoo]"``` Misc info: Tested with both mypy 0.630 and mypy master (e12be3ba9e6be3e9242e3a81ffaa9b3a136876fe), run with `--strict`
1.0
Better error message for rejected protocol subtyping due to mutability - It appears that Union types for Protocol attributes aren't recognized. Repro example: ``` from typing import Type, Union from typing_extensions import Protocol class HasFoo(Protocol): # foo: str # works foo: Union[str, int] # fails def dec(cls: Type[HasFoo]) -> Type[HasFoo]: return cls @dec class Thing: foo = 'abc' ``` Failing output: ```Argument 1 to "dec" has incompatible type "Type[Thing]"; expected "Type[HasFoo]"``` Misc info: Tested with both mypy 0.630 and mypy master (e12be3ba9e6be3e9242e3a81ffaa9b3a136876fe), run with `--strict`
non_test
better error message for rejected protocol subtyping due to mutability it appears that union types for protocol attributes aren t recognized repro example from typing import type union from typing extensions import protocol class hasfoo protocol foo str works foo union fails def dec cls type type return cls dec class thing foo abc failing output argument to dec has incompatible type type expected type misc info tested with both mypy and mypy master run with strict
0
95,382
27,491,687,431
IssuesEvent
2023-03-04 17:50:20
jozefizso/generator-license
https://api.github.com/repos/jozefizso/generator-license
closed
Configure coveralls
test build maintanance
Configure `jest` to run coverage report and send it to https://coveralls.io/github/jozefizso/generator-license service. ## Note There is bug in `jest` and the `lcov.info` will have incorrect path to the `app/index.js` source code. We must modify the `lcov.info` file to have correct path so the `coverallsapp/github-action` will report all data to coveralls.io service. See https://github.com/facebook/jest/issues/9773 and https://github.com/coverallsapp/github-action/issues/153
1.0
Configure coveralls - Configure `jest` to run coverage report and send it to https://coveralls.io/github/jozefizso/generator-license service. ## Note There is bug in `jest` and the `lcov.info` will have incorrect path to the `app/index.js` source code. We must modify the `lcov.info` file to have correct path so the `coverallsapp/github-action` will report all data to coveralls.io service. See https://github.com/facebook/jest/issues/9773 and https://github.com/coverallsapp/github-action/issues/153
non_test
configure coveralls configure jest to run coverage report and send it to service note there is bug in jest and the lcov info will have incorrect path to the app index js source code we must modify the lcov info file to have correct path so the coverallsapp github action will report all data to coveralls io service see and
0
304,338
26,269,825,389
IssuesEvent
2023-01-06 15:56:26
Joystream/pioneer
https://api.github.com/repos/Joystream/pioneer
closed
`My votes` shows wrong numbers for `My Stake`
bug scope:council qa-task community-dev qa-tested-ready-for-prod bounty-backlog SP:2 release:1.1.0
## Scope: - [ ] Display stake that was added by voting via logged in member in my votes NB: "My Stake" that was candidate stake[ is addressed here> ](https://github.com/Joystream/pioneer/issues/2795) ## Details https://dao.joystream.org/#/election -> My Votes `My Stake` shows the candidate's own stake. As voter i am only interested how much i voted.
1.0
`My votes` shows wrong numbers for `My Stake` - ## Scope: - [ ] Display stake that was added by voting via logged in member in my votes NB: "My Stake" that was candidate stake[ is addressed here> ](https://github.com/Joystream/pioneer/issues/2795) ## Details https://dao.joystream.org/#/election -> My Votes `My Stake` shows the candidate's own stake. As voter i am only interested how much i voted.
test
my votes shows wrong numbers for my stake scope display stake that was added by voting via logged in member in my votes nb my stake that was candidate stake details my votes my stake shows the candidate s own stake as voter i am only interested how much i voted
1
7,812
6,229,513,475
IssuesEvent
2017-07-11 04:18:04
phetsims/molecule-polarity
https://api.github.com/repos/phetsims/molecule-polarity
opened
Some garbage collection (?) make motion a bit jerky at times.
type:performance
This is not a learning goal of the sim, and in fact the rotation dynamics is not physically correct, but I do notice garbage collection making the rotation animation halt/glitch at times when I play with it. Low priority in my view, but though I would report it.
True
Some garbage collection (?) make motion a bit jerky at times. - This is not a learning goal of the sim, and in fact the rotation dynamics is not physically correct, but I do notice garbage collection making the rotation animation halt/glitch at times when I play with it. Low priority in my view, but though I would report it.
non_test
some garbage collection make motion a bit jerky at times this is not a learning goal of the sim and in fact the rotation dynamics is not physically correct but i do notice garbage collection making the rotation animation halt glitch at times when i play with it low priority in my view but though i would report it
0
328,709
28,131,487,500
IssuesEvent
2023-04-01 00:03:16
WordPress/gutenberg
https://api.github.com/repos/WordPress/gutenberg
closed
[Flaky Test] last-inserted block should be given and keep the focus (large viewport)
[Status] Stale [Type] Flaky Test
<!-- __META_DATA__:{} --> **Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.** ## Test title last-inserted block should be given and keep the focus (large viewport) ## Test path `specs/editor/various/inserting-blocks.test.js` ## Errors <!-- __TEST_RESULTS_LIST__ --> <!-- __TEST_RESULT__ --><time datetime="2023-02-26T21:41:49.759Z"><code>[2023-02-26T21:41:49.759Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4276949696"><code>try/attrs</code></a>.<!-- /__TEST_RESULT__ --> <br/> <!-- __TEST_RESULT__ --><details> <summary> <time datetime="2023-02-27T22:16:03.202Z"><code>[2023-02-27T22:16:03.202Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4287086701"><code>try/attrs</code></a>. </summary> ``` ● Inserting blocks › last-inserted block should be given and keep the focus (large viewport) expect(jest.fn()).not.toHaveErrored(expected) Expected mock function not to be called but it was called with: ["TypeError: Cannot read properties of undefined (reading 'perPage') at b_ (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/block-library/index.min.js?ver=091eafc10324dc9e60c7:118:9562) at wt (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:47623) at vr (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:56438) at $s (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:121605) at wl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88644) at bl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88572) at yl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88435) at il (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:85259) at ol (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:83816) at v (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react.min.js?ver=18:1:3385)"],["TypeError: Cannot read properties of undefined (reading 'perPage') at b_ (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/block-library/index.min.js?ver=091eafc10324dc9e60c7:118:9562) at wt (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:47623) at vr (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:56438) at $s (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:121605) at wl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88644) at bl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88572) at yl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88435) at il (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:85259) at ol (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:83816) at v (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react.min.js?ver=18:1:3385)"] at runMicrotasks (<anonymous>) ``` </details><!-- /__TEST_RESULT__ --> <!-- /__TEST_RESULTS_LIST__ -->
1.0
[Flaky Test] last-inserted block should be given and keep the focus (large viewport) - <!-- __META_DATA__:{} --> **Flaky test detected. This is an auto-generated issue by GitHub Actions. Please do NOT edit this manually.** ## Test title last-inserted block should be given and keep the focus (large viewport) ## Test path `specs/editor/various/inserting-blocks.test.js` ## Errors <!-- __TEST_RESULTS_LIST__ --> <!-- __TEST_RESULT__ --><time datetime="2023-02-26T21:41:49.759Z"><code>[2023-02-26T21:41:49.759Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4276949696"><code>try/attrs</code></a>.<!-- /__TEST_RESULT__ --> <br/> <!-- __TEST_RESULT__ --><details> <summary> <time datetime="2023-02-27T22:16:03.202Z"><code>[2023-02-27T22:16:03.202Z]</code></time> Test passed after 1 failed attempt on <a href="https://github.com/WordPress/gutenberg/actions/runs/4287086701"><code>try/attrs</code></a>. </summary> ``` ● Inserting blocks › last-inserted block should be given and keep the focus (large viewport) expect(jest.fn()).not.toHaveErrored(expected) Expected mock function not to be called but it was called with: ["TypeError: Cannot read properties of undefined (reading 'perPage') at b_ (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/block-library/index.min.js?ver=091eafc10324dc9e60c7:118:9562) at wt (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:47623) at vr (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:56438) at $s (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:121605) at wl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88644) at bl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88572) at yl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88435) at il (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:85259) at ol (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:83816) at v (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react.min.js?ver=18:1:3385)"],["TypeError: Cannot read properties of undefined (reading 'perPage') at b_ (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/block-library/index.min.js?ver=091eafc10324dc9e60c7:118:9562) at wt (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:47623) at vr (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:56438) at $s (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:121605) at wl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88644) at bl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88572) at yl (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:88435) at il (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:85259) at ol (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react-dom.min.js?ver=18:1:83816) at v (../../http:/localhost:8889/wp-content/plugins/gutenberg/build/vendors/react.min.js?ver=18:1:3385)"] at runMicrotasks (<anonymous>) ``` </details><!-- /__TEST_RESULT__ --> <!-- /__TEST_RESULTS_LIST__ -->
test
last inserted block should be given and keep the focus large viewport flaky test detected this is an auto generated issue by github actions please do not edit this manually test title last inserted block should be given and keep the focus large viewport test path specs editor various inserting blocks test js errors test passed after failed attempt on test passed after failed attempt on a href ● inserting blocks › last inserted block should be given and keep the focus large viewport expect jest fn not tohaveerrored expected expected mock function not to be called but it was called with typeerror cannot read properties of undefined reading perpage at b http localhost wp content plugins gutenberg build block library index min js ver at wt http localhost wp content plugins gutenberg build vendors react dom min js ver at vr http localhost wp content plugins gutenberg build vendors react dom min js ver at s http localhost wp content plugins gutenberg build vendors react dom min js ver at wl http localhost wp content plugins gutenberg build vendors react dom min js ver at bl http localhost wp content plugins gutenberg build vendors react dom min js ver at yl http localhost wp content plugins gutenberg build vendors react dom min js ver at il http localhost wp content plugins gutenberg build vendors react dom min js ver at ol http localhost wp content plugins gutenberg build vendors react dom min js ver at v http localhost wp content plugins gutenberg build vendors react min js ver typeerror cannot read properties of undefined reading perpage at b http localhost wp content plugins gutenberg build block library index min js ver at wt http localhost wp content plugins gutenberg build vendors react dom min js ver at vr http localhost wp content plugins gutenberg build vendors react dom min js ver at s http localhost wp content plugins gutenberg build vendors react dom min js ver at wl http localhost wp content plugins gutenberg build vendors react dom min js ver at bl http localhost wp content plugins gutenberg build vendors react dom min js ver at yl http localhost wp content plugins gutenberg build vendors react dom min js ver at il http localhost wp content plugins gutenberg build vendors react dom min js ver at ol http localhost wp content plugins gutenberg build vendors react dom min js ver at v http localhost wp content plugins gutenberg build vendors react min js ver at runmicrotasks
1
114,498
4,635,133,239
IssuesEvent
2016-09-29 05:30:39
yairodriguez/Twitter
https://api.github.com/repos/yairodriguez/Twitter
opened
Testing React Components
[priority] low [status] accepted [type] enhancement
### Description --- ### Issue Checklist All issues in milestone: [0.4.0 milestone](https://github.com/yairodriguez/Twitter/milestone/4) --- ### Assignees - [ ] Final assign @yairodriguez
1.0
Testing React Components - ### Description --- ### Issue Checklist All issues in milestone: [0.4.0 milestone](https://github.com/yairodriguez/Twitter/milestone/4) --- ### Assignees - [ ] Final assign @yairodriguez
non_test
testing react components description issue checklist all issues in milestone assignees final assign yairodriguez
0
253,792
8,066,023,476
IssuesEvent
2018-08-04 10:10:58
AeneasPlatform/Aeneas
https://api.github.com/repos/AeneasPlatform/Aeneas
closed
Pre-mining setup fail handling
Aeneas:Core Priority 1 : High bug
**It is not any fail handling of mining timeout error.** **To Reproduce** Steps to reproduce the behavior: 1. Launch node at the first time. 2. Wait till pre-mining setup will end. 3. If setup value is bigger than it should be, you will see an error if mining timeout happens. **Expected behavior** It should get new block from the network, apply it to history and start new iteration. **Desktop (please complete the following information):** - OS: OS X El Capitan - Device: Macbook Pro 2015 Early - Version : commit № [839c3ed45c32687d14fc8af8178eb1138bda1ca1](https://github.com/AeneasPlatform/Aeneas/commit/839c3ed45c32687d14fc8af8178eb1138bda1ca1)
1.0
Pre-mining setup fail handling - **It is not any fail handling of mining timeout error.** **To Reproduce** Steps to reproduce the behavior: 1. Launch node at the first time. 2. Wait till pre-mining setup will end. 3. If setup value is bigger than it should be, you will see an error if mining timeout happens. **Expected behavior** It should get new block from the network, apply it to history and start new iteration. **Desktop (please complete the following information):** - OS: OS X El Capitan - Device: Macbook Pro 2015 Early - Version : commit № [839c3ed45c32687d14fc8af8178eb1138bda1ca1](https://github.com/AeneasPlatform/Aeneas/commit/839c3ed45c32687d14fc8af8178eb1138bda1ca1)
non_test
pre mining setup fail handling it is not any fail handling of mining timeout error to reproduce steps to reproduce the behavior launch node at the first time wait till pre mining setup will end if setup value is bigger than it should be you will see an error if mining timeout happens expected behavior it should get new block from the network apply it to history and start new iteration desktop please complete the following information os os x el capitan device macbook pro early version commit №
0
209,576
16,236,388,300
IssuesEvent
2021-05-07 01:37:05
SystemTruck/SystemTruck
https://api.github.com/repos/SystemTruck/SystemTruck
closed
NF_About_and_Readme.md
documentation enhancement
# Update Readme.md file for newcommers <!--- Choose a title depending on the issue. --> <!--- Include: BC for Bug Fix or Change, NF for New Feature, TE for TESTING. --> ## Description <!--- Please write a detailed summary about the issue and root cause. --> In the about section we need to give a little explation of what is this project. The readme.md file is our cover about this project. It explains what this projects is about, may have images, diagrams, tables and clear explanations for the github community what this project is about. It can be held also the roles in this projects and motivations. ## Checklist - [x] Select asignee(s) - [x] Select labels (optional) - [x] Associate to the project ## Include feature - [x] Create a Branch - [x] Implement the feature - [x] Explain how the feature works - [x] Create a Pull Request - [ ] Link the Pull request to the issue
1.0
NF_About_and_Readme.md - # Update Readme.md file for newcommers <!--- Choose a title depending on the issue. --> <!--- Include: BC for Bug Fix or Change, NF for New Feature, TE for TESTING. --> ## Description <!--- Please write a detailed summary about the issue and root cause. --> In the about section we need to give a little explation of what is this project. The readme.md file is our cover about this project. It explains what this projects is about, may have images, diagrams, tables and clear explanations for the github community what this project is about. It can be held also the roles in this projects and motivations. ## Checklist - [x] Select asignee(s) - [x] Select labels (optional) - [x] Associate to the project ## Include feature - [x] Create a Branch - [x] Implement the feature - [x] Explain how the feature works - [x] Create a Pull Request - [ ] Link the Pull request to the issue
non_test
nf about and readme md update readme md file for newcommers description in the about section we need to give a little explation of what is this project the readme md file is our cover about this project it explains what this projects is about may have images diagrams tables and clear explanations for the github community what this project is about it can be held also the roles in this projects and motivations checklist select asignee s select labels optional associate to the project include feature create a branch implement the feature explain how the feature works create a pull request link the pull request to the issue
0
94,603
8,506,552,614
IssuesEvent
2018-10-30 16:50:19
elastic/elasticsearch
https://api.github.com/repos/elastic/elasticsearch
opened
SQL: [tests] remove AbstractBuilderTestCase to increase speed
:Search/SQL >test v6.5.1 v6.6.0 v7.0.0
Remove `AbstractBuilderTestCase` from tests as this increases test execution time
1.0
SQL: [tests] remove AbstractBuilderTestCase to increase speed - Remove `AbstractBuilderTestCase` from tests as this increases test execution time
test
sql remove abstractbuildertestcase to increase speed remove abstractbuildertestcase from tests as this increases test execution time
1
29,693
4,533,830,789
IssuesEvent
2016-09-08 12:59:41
ipfs/go-ipfs
https://api.github.com/repos/ipfs/go-ipfs
opened
TestRepublish fails randomly
test failure
in namesys/republisher, TestRepublish fails randomly with: 'failed to find any peer in table' https://travis-ci.org/ipfs/go-ipfs/jobs/158430466
1.0
TestRepublish fails randomly - in namesys/republisher, TestRepublish fails randomly with: 'failed to find any peer in table' https://travis-ci.org/ipfs/go-ipfs/jobs/158430466
test
testrepublish fails randomly in namesys republisher testrepublish fails randomly with failed to find any peer in table
1
322,679
9,821,522,641
IssuesEvent
2019-06-14 07:28:26
webcompat/web-bugs
https://api.github.com/repos/webcompat/web-bugs
closed
www.facebook.com - see bug description
browser-fixme priority-critical type-search-hijacking
<!-- @browser: Dragon 64.0.2 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 IceDragon/64.0.2 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.facebook.com/settings?ref=mb&drop **Browser / Version**: Dragon 64.0.2 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: SOME ONE PLAYING WITH MY PAGE ON HER **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/6/ebea0d6f-a4c6-46dd-a249-f6c24041aa93-thumb.jpeg)](https://webcompat.com/uploads/2019/6/ebea0d6f-a4c6-46dd-a249-f6c24041aa93.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190116153700</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: default</li> </ul> <p>Console Messages:</p> <pre> [u'[console.timeStamp(t_start) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_domcontent) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_tti) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_prehooks) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_hooks) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.log(\n\n\n \n .d8888b. 888 888 \nd88P Y88b 888 888 \nY88b. 888 888 This is a browser feature intended for \n "Y888b. 888888 .d88b. 88888b. 888 developers. If someone told you to copy-paste \n "Y88b. 888 d88""88b 888 "88b 888 something here to enable a Facebook feature \n "888 888 888 888 888 888 Y8P or "hack" someone\'s account, it is a \nY88b d88P Y88b. Y88..88P 888 d88P scam and will give them access to your \n "Y8888P" "Y888 "Y88P" 88888P" 888 Facebook account.\n 888 \n 888 \n 888 \n\nSee https://www.facebook.com/selfxss for more information.\n) https://static.xx.fbcdn.net/rsrc.php/v3iqES4/y5/l/en_US/ISnRwO575wo.js:7:1478]', u'[console.timeStamp(t_layout) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(perf_trace {"name": "e2e", "parent": "PageEvents.BIGPIPE_ONLOAD"}) https://static.xx.fbcdn.net/rsrc.php/v3/yB/r/WOS8wdnBc-8.js:181:2966]', u'[console.timeStamp(t_paint) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_domcontent) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(perf_trace {"name": "e2e", "parent": "PageEvents.AJAXPIPE_ONLOAD"}) https://static.xx.fbcdn.net/rsrc.php/v3/yi/r/x4F9XJqLOCR.js:10:3152]', u'[console.timeStamp(t_hooks) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_layout) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_onload) https://www.facebook.com/settings?ref=mb&drop:3:1207]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
1.0
www.facebook.com - see bug description - <!-- @browser: Dragon 64.0.2 --> <!-- @ua_header: Mozilla/5.0 (Windows NT 10.0; Win64; x64; rv:64.0) Gecko/20100101 Firefox/64.0 IceDragon/64.0.2 --> <!-- @reported_with: desktop-reporter --> **URL**: https://www.facebook.com/settings?ref=mb&drop **Browser / Version**: Dragon 64.0.2 **Operating System**: Windows 10 **Tested Another Browser**: Yes **Problem type**: Something else **Description**: SOME ONE PLAYING WITH MY PAGE ON HER **Steps to Reproduce**: [![Screenshot Description](https://webcompat.com/uploads/2019/6/ebea0d6f-a4c6-46dd-a249-f6c24041aa93-thumb.jpeg)](https://webcompat.com/uploads/2019/6/ebea0d6f-a4c6-46dd-a249-f6c24041aa93.jpeg) <details> <summary>Browser Configuration</summary> <ul> <li>mixed active content blocked: false</li><li>image.mem.shared: true</li><li>buildID: 20190116153700</li><li>tracking content blocked: false</li><li>gfx.webrender.blob-images: true</li><li>hasTouchScreen: false</li><li>mixed passive content blocked: false</li><li>gfx.webrender.enabled: false</li><li>gfx.webrender.all: false</li><li>channel: default</li> </ul> <p>Console Messages:</p> <pre> [u'[console.timeStamp(t_start) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_domcontent) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_tti) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_prehooks) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_hooks) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.log(\n\n\n \n .d8888b. 888 888 \nd88P Y88b 888 888 \nY88b. 888 888 This is a browser feature intended for \n "Y888b. 888888 .d88b. 88888b. 888 developers. If someone told you to copy-paste \n "Y88b. 888 d88""88b 888 "88b 888 something here to enable a Facebook feature \n "888 888 888 888 888 888 Y8P or "hack" someone\'s account, it is a \nY88b d88P Y88b. Y88..88P 888 d88P scam and will give them access to your \n "Y8888P" "Y888 "Y88P" 88888P" 888 Facebook account.\n 888 \n 888 \n 888 \n\nSee https://www.facebook.com/selfxss for more information.\n) https://static.xx.fbcdn.net/rsrc.php/v3iqES4/y5/l/en_US/ISnRwO575wo.js:7:1478]', u'[console.timeStamp(t_layout) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(perf_trace {"name": "e2e", "parent": "PageEvents.BIGPIPE_ONLOAD"}) https://static.xx.fbcdn.net/rsrc.php/v3/yB/r/WOS8wdnBc-8.js:181:2966]', u'[console.timeStamp(t_paint) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_domcontent) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(perf_trace {"name": "e2e", "parent": "PageEvents.AJAXPIPE_ONLOAD"}) https://static.xx.fbcdn.net/rsrc.php/v3/yi/r/x4F9XJqLOCR.js:10:3152]', u'[console.timeStamp(t_hooks) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_layout) https://www.facebook.com/settings?ref=mb&drop:3:1207]', u'[console.timeStamp(t_onload) https://www.facebook.com/settings?ref=mb&drop:3:1207]'] </pre> </details> _From [webcompat.com](https://webcompat.com/) with ❤️_
non_test
see bug description url browser version dragon operating system windows tested another browser yes problem type something else description some one playing with my page on her steps to reproduce browser configuration mixed active content blocked false image mem shared true buildid tracking content blocked false gfx webrender blob images true hastouchscreen false mixed passive content blocked false gfx webrender enabled false gfx webrender all false channel default console messages u u u u u u u u u u u u u from with ❤️
0
169,243
6,397,177,408
IssuesEvent
2017-08-04 17:20:59
robertgarrigos/ubercart
https://api.github.com/repos/robertgarrigos/ubercart
opened
default cart form pane no set on new install
priority - low type - bug
In a fresh install of core modules, in cart basic settings, default cart form pane should be checked.
1.0
default cart form pane no set on new install - In a fresh install of core modules, in cart basic settings, default cart form pane should be checked.
non_test
default cart form pane no set on new install in a fresh install of core modules in cart basic settings default cart form pane should be checked
0
339,127
30,346,300,226
IssuesEvent
2023-07-11 15:38:15
SWM-Cupid/jikting-backend
https://api.github.com/repos/SWM-Cupid/jikting-backend
opened
[SP-124] 채팅방 입장 API 명세서 작성
feature docs test
## Jira Issue [SP-124](https://soma-cupid.atlassian.net/browse/SP-124?atlOrigin=eyJpIjoiMDA2ZjdhMjE5MmQ2NDMyZmIzY2U1NjMzZmViNmE2MWQiLCJwIjoiaiJ9) ## 요구사항 - [ ] 채팅방 입장 성공 API 작성 - [ ] 채팅방 입장 실패 API 작성
1.0
[SP-124] 채팅방 입장 API 명세서 작성 - ## Jira Issue [SP-124](https://soma-cupid.atlassian.net/browse/SP-124?atlOrigin=eyJpIjoiMDA2ZjdhMjE5MmQ2NDMyZmIzY2U1NjMzZmViNmE2MWQiLCJwIjoiaiJ9) ## 요구사항 - [ ] 채팅방 입장 성공 API 작성 - [ ] 채팅방 입장 실패 API 작성
test
채팅방 입장 api 명세서 작성 jira issue 요구사항 채팅방 입장 성공 api 작성 채팅방 입장 실패 api 작성
1
2,809
2,639,758,424
IssuesEvent
2015-03-11 05:57:42
aspnet/Mvc
https://api.github.com/repos/aspnet/Mvc
closed
MVC Functional Tests: Remove BuildFallbackServiceProvider code in TestServices/RazorPrecompilation
3 - Working Add Tests task
Once there's a better story for how functional tests should work, this code should be removed, it was added to preserve functionality after fallback was removed from DI <!--- @huboard:{"custom_state":"ready"} -->
1.0
MVC Functional Tests: Remove BuildFallbackServiceProvider code in TestServices/RazorPrecompilation - Once there's a better story for how functional tests should work, this code should be removed, it was added to preserve functionality after fallback was removed from DI <!--- @huboard:{"custom_state":"ready"} -->
test
mvc functional tests remove buildfallbackserviceprovider code in testservices razorprecompilation once there s a better story for how functional tests should work this code should be removed it was added to preserve functionality after fallback was removed from di huboard custom state ready
1
95,415
10,879,680,454
IssuesEvent
2019-11-17 04:05:28
aragon/aragon-court
https://api.github.com/repos/aragon/aragon-court
opened
Fix comment regarding ERC-165 compliance
documentation
https://github.com/aragon/aragon-apps/blob/master/apps/agent/contracts/Agent.sol#L243-L247 Per EIP-165, contracts that support it must also return true when asked about conformance to the EIP-165 interface, but the comment in `IArbitrable` specifies that `true` must only be returned if asked for the Arbitrable interface id
1.0
Fix comment regarding ERC-165 compliance - https://github.com/aragon/aragon-apps/blob/master/apps/agent/contracts/Agent.sol#L243-L247 Per EIP-165, contracts that support it must also return true when asked about conformance to the EIP-165 interface, but the comment in `IArbitrable` specifies that `true` must only be returned if asked for the Arbitrable interface id
non_test
fix comment regarding erc compliance per eip contracts that support it must also return true when asked about conformance to the eip interface but the comment in iarbitrable specifies that true must only be returned if asked for the arbitrable interface id
0
67,239
27,759,185,455
IssuesEvent
2023-03-16 06:39:03
elastic/integrations
https://api.github.com/repos/elastic/integrations
closed
[O11y][Memcached] Lens Migration
Integration:Memcached v8.2.0 Team:Service-Integrations
Manually migrate Memcached visualizations to lens in current Kibana version `8.2.0` itself. ## Migration stats <html> <body> <!--StartFragment--><google-sheets-html-origin>   | Maps | Search | Lens | Visualization | Tag Cloud -- | -- | -- | -- | -- | -- [Metrics Memcached] Overview | 0 | 0 | 7 | 0 | 0   | 0 | 0 | 7 | 0 | 0 <!--EndFragment--> </body> </html> All visualizations are present in lens.
1.0
[O11y][Memcached] Lens Migration - Manually migrate Memcached visualizations to lens in current Kibana version `8.2.0` itself. ## Migration stats <html> <body> <!--StartFragment--><google-sheets-html-origin>   | Maps | Search | Lens | Visualization | Tag Cloud -- | -- | -- | -- | -- | -- [Metrics Memcached] Overview | 0 | 0 | 7 | 0 | 0   | 0 | 0 | 7 | 0 | 0 <!--EndFragment--> </body> </html> All visualizations are present in lens.
non_test
lens migration manually migrate memcached visualizations to lens in current kibana version itself migration stats   maps search lens visualization tag cloud overview   all visualizations are present in lens
0
115,789
14,887,900,180
IssuesEvent
2021-01-20 19:01:31
MetaMask/metamask-extension
https://api.github.com/repos/MetaMask/metamask-extension
opened
[transaction confirmation] Improve "insufficient funds" error. Add link to Buy ETH
N00-needsDesign
As a user, I want to have more specific “insufficient funds” messages (insufficient token balance or insufficient ETH for the gas fee) AND have a link to Buy ETH and be directed to onramp flow
1.0
[transaction confirmation] Improve "insufficient funds" error. Add link to Buy ETH - As a user, I want to have more specific “insufficient funds” messages (insufficient token balance or insufficient ETH for the gas fee) AND have a link to Buy ETH and be directed to onramp flow
non_test
improve insufficient funds error add link to buy eth as a user i want to have more specific “insufficient funds” messages insufficient token balance or insufficient eth for the gas fee and have a link to buy eth and be directed to onramp flow
0
71,355
9,506,597,467
IssuesEvent
2019-04-25 01:05:15
Naoghuman/lib-i18n
https://api.github.com/repos/Naoghuman/lib-i18n
opened
[doc] Switch the two sections 'Examples' and 'Features.
documentation refactoring
[doc] Switch the two sections 'Examples' and 'Features.
1.0
[doc] Switch the two sections 'Examples' and 'Features. - [doc] Switch the two sections 'Examples' and 'Features.
non_test
switch the two sections examples and features switch the two sections examples and features
0
163,158
12,706,147,981
IssuesEvent
2020-06-23 06:35:17
DrewScoggins/performance-2
https://api.github.com/repos/DrewScoggins/performance-2
closed
[Perf -18%] System.Memory.Span<Char>.IndexOfValue
Baseline Outlier Test Test Test Test Test Windows 10.0.18362 refs/heads/master x64
### Run Information Architecture | x64 -- | -- OS | Windows 10.0.18362 Changes | [diff](https://github.com/dotnet/runtime/compare/518df4f20d20f233727cba2fcc42603a31f750df...8e3b55435b3ed75cffb6687916ae1edcf7253408) ### Regressions in System.Memory.Span<Char> Benchmark | Baseline | Test | Test/Base | Modality | Baseline Outlier -- | -- | -- | -- | -- | -- IndexOfValue | 13.61 ns | 16.07 ns | 1.18 | | True ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_12_2020/refs/heads/master_x64_Windows%2010.0.18362_Test%20Test%20Test%20Test%20Test/System.Memory.Span(Char)_1.png>) [Historical Data in Reporting System](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_12_2020/refs/heads/master_x64_Windows%2010.0.18362_Test%20Test%20Test%20Test%20Test/System.Memory.Span(Char).html>) ### Repro ```cmd git clone https://github.com/dotnet/performance.git py .\performance\scripts\benchmarks_ci.py -f netcoreapp5.0 --filter 'System.Memory.Span<Char>*'; ``` <details> ### Histogram #### System.Memory.Span<Char>.IndexOfValue(Size: 512) ```log [13.481 ; 14.241) | @@@@ [14.241 ; 14.636) | [14.636 ; 15.401) | @@@@ [15.401 ; 16.257) | @@@@@@@@@@@@@@@@ [16.257 ; 17.018) | @@@@@ ``` ### Docs [Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md) [Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md) </details>
5.0
[Perf -18%] System.Memory.Span<Char>.IndexOfValue - ### Run Information Architecture | x64 -- | -- OS | Windows 10.0.18362 Changes | [diff](https://github.com/dotnet/runtime/compare/518df4f20d20f233727cba2fcc42603a31f750df...8e3b55435b3ed75cffb6687916ae1edcf7253408) ### Regressions in System.Memory.Span<Char> Benchmark | Baseline | Test | Test/Base | Modality | Baseline Outlier -- | -- | -- | -- | -- | -- IndexOfValue | 13.61 ns | 16.07 ns | 1.18 | | True ![graph](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_12_2020/refs/heads/master_x64_Windows%2010.0.18362_Test%20Test%20Test%20Test%20Test/System.Memory.Span(Char)_1.png>) [Historical Data in Reporting System](<https://pvscmdupload.blob.core.windows.net/autofilereport/autofilereports/06_12_2020/refs/heads/master_x64_Windows%2010.0.18362_Test%20Test%20Test%20Test%20Test/System.Memory.Span(Char).html>) ### Repro ```cmd git clone https://github.com/dotnet/performance.git py .\performance\scripts\benchmarks_ci.py -f netcoreapp5.0 --filter 'System.Memory.Span<Char>*'; ``` <details> ### Histogram #### System.Memory.Span<Char>.IndexOfValue(Size: 512) ```log [13.481 ; 14.241) | @@@@ [14.241 ; 14.636) | [14.636 ; 15.401) | @@@@ [15.401 ; 16.257) | @@@@@@@@@@@@@@@@ [16.257 ; 17.018) | @@@@@ ``` ### Docs [Profiling workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/profiling-workflow-dotnet-runtime.md) [Benchmarking workflow for dotnet/runtime repository](https://github.com/dotnet/performance/blob/master/docs/benchmarking-workflow-dotnet-runtime.md) </details>
test
system memory span indexofvalue run information architecture os windows changes regressions in system memory span benchmark baseline test test base modality baseline outlier indexofvalue ns ns true repro cmd git clone py performance scripts benchmarks ci py f filter system memory span histogram system memory span indexofvalue size log docs
1
38,939
5,205,480,235
IssuesEvent
2017-01-24 18:03:14
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
OpenLDAP: Expose existing ldap config attributes in UI
area/access-control area/ui priority/-1 status/resolved status/to-test
**Rancher Versions:** Server: 1.3.x **Steps to Reproduce:** 1. Setup Rancher and go to Admin ->Access Control 2. Choose OpenLDAP 3. The config page should provide a way to configure following 2 ldapconfig fields: 'groupMemberMappingAttribute' 'userMemberAttribute' 4. Currently the POST request to set the config, uses default values for these two: Request URL:/v2-beta/openldapconfig Request Method:POST Request Payload: {"id":null,"type":"openldapconfig","baseType":"openldapconfig","accessMode":"unrestricted","connectionTimeout":1000,"domain":"dc=us-west-2,dc=compute,dc=internal","enabled":false,"groupDNField":"entryDN",**"groupMemberMappingAttribute":"memberUid"**,"groupMemberUserAttribute":"entryDN","groupNameField":"cn","groupObjectClass":"groupofNames","groupSearchField":"cn","loginDomain":null,"name":"openldapconfig","port":389,"server":"52.41.23.7","serviceAccountPassword":"cattle","serviceAccountUsername":"cn=admin,dc=us-west-2,dc=compute,dc=internal","tls":false,"userDisabledBitMask":0,"userEnabledAttribute":null,"userLoginField":"uid",**"userMemberAttribute":"memberOf"**,"userNameField":"givenName","userObjectClass":"inetOrgPerson","userSearchField":"uid"} **Results:** Please expose them in UI so that they can be configured easily since every ldap setup can have custom values.
1.0
OpenLDAP: Expose existing ldap config attributes in UI - **Rancher Versions:** Server: 1.3.x **Steps to Reproduce:** 1. Setup Rancher and go to Admin ->Access Control 2. Choose OpenLDAP 3. The config page should provide a way to configure following 2 ldapconfig fields: 'groupMemberMappingAttribute' 'userMemberAttribute' 4. Currently the POST request to set the config, uses default values for these two: Request URL:/v2-beta/openldapconfig Request Method:POST Request Payload: {"id":null,"type":"openldapconfig","baseType":"openldapconfig","accessMode":"unrestricted","connectionTimeout":1000,"domain":"dc=us-west-2,dc=compute,dc=internal","enabled":false,"groupDNField":"entryDN",**"groupMemberMappingAttribute":"memberUid"**,"groupMemberUserAttribute":"entryDN","groupNameField":"cn","groupObjectClass":"groupofNames","groupSearchField":"cn","loginDomain":null,"name":"openldapconfig","port":389,"server":"52.41.23.7","serviceAccountPassword":"cattle","serviceAccountUsername":"cn=admin,dc=us-west-2,dc=compute,dc=internal","tls":false,"userDisabledBitMask":0,"userEnabledAttribute":null,"userLoginField":"uid",**"userMemberAttribute":"memberOf"**,"userNameField":"givenName","userObjectClass":"inetOrgPerson","userSearchField":"uid"} **Results:** Please expose them in UI so that they can be configured easily since every ldap setup can have custom values.
test
openldap expose existing ldap config attributes in ui rancher versions server x steps to reproduce setup rancher and go to admin access control choose openldap the config page should provide a way to configure following ldapconfig fields groupmembermappingattribute usermemberattribute currently the post request to set the config uses default values for these two request url beta openldapconfig request method post request payload id null type openldapconfig basetype openldapconfig accessmode unrestricted connectiontimeout domain dc us west dc compute dc internal enabled false groupdnfield entrydn groupmembermappingattribute memberuid groupmemberuserattribute entrydn groupnamefield cn groupobjectclass groupofnames groupsearchfield cn logindomain null name openldapconfig port server serviceaccountpassword cattle serviceaccountusername cn admin dc us west dc compute dc internal tls false userdisabledbitmask userenabledattribute null userloginfield uid usermemberattribute memberof usernamefield givenname userobjectclass inetorgperson usersearchfield uid results please expose them in ui so that they can be configured easily since every ldap setup can have custom values
1
271,819
20,718,864,539
IssuesEvent
2022-03-13 03:33:09
wagumi/community-bot
https://api.github.com/repos/wagumi/community-bot
closed
プロジェクトノート用 Wiki ページを作成する
documentation
# 概要 Story Issue: # プロジェクトごとの良かったこと、留意事項、次回以降の PJ への申し送りなどを一覧できる Wiki ページを作成します。 ## タスク - [ ] 下記のプロジェクトノートを追記 - [ベース環境構築](https://github.com/wagumi/community-bot/milestone/1) - [Bot 開発用のベース機能を追加](https://github.com/wagumi/community-bot/milestone/3) - [ ] プロジェクトノート用の文書テンプレートを作成 ## 参考 - TBD ## QA 項目 - TBD ## 作業用スレッド - TBD
1.0
プロジェクトノート用 Wiki ページを作成する - # 概要 Story Issue: # プロジェクトごとの良かったこと、留意事項、次回以降の PJ への申し送りなどを一覧できる Wiki ページを作成します。 ## タスク - [ ] 下記のプロジェクトノートを追記 - [ベース環境構築](https://github.com/wagumi/community-bot/milestone/1) - [Bot 開発用のベース機能を追加](https://github.com/wagumi/community-bot/milestone/3) - [ ] プロジェクトノート用の文書テンプレートを作成 ## 参考 - TBD ## QA 項目 - TBD ## 作業用スレッド - TBD
non_test
プロジェクトノート用 wiki ページを作成する 概要 story issue プロジェクトごとの良かったこと、留意事項、次回以降の pj への申し送りなどを一覧できる wiki ページを作成します。 タスク 下記のプロジェクトノートを追記 プロジェクトノート用の文書テンプレートを作成 参考 tbd qa 項目 tbd 作業用スレッド tbd
0
234,360
19,144,988,102
IssuesEvent
2021-12-02 06:19:18
ita-social-projects/EventsExpress
https://api.github.com/repos/ita-social-projects/EventsExpress
closed
Verified when the user press the "Join EventsExpress" burton the login pop-up form is displayed [Login] [Registration]
test case
**Date Time** 01-12-2021 **Priority** Medium **Description** When a User clicks the "Join EventExpress"button , the system redirects him to the registration form. **Pre-requisites** Move to: https://eventsexpress-test.azurewebsites.net/ **Test Cases** | S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated | |:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:| | 1. <br> | Click the "Join EventsExpress" button from landing page<br> | <br> | Displayed the pop-up registration form <br> | <br> | <br> | <br> | | <br> | <br> | <br> | <br> | <br> | <br> | <br> | **Environment:** - OS: Windows10 - Browser: Chrome version 96.0.4664.45 - Version [e.g. 22] **Additional context** User Story #708
1.0
Verified when the user press the "Join EventsExpress" burton the login pop-up form is displayed [Login] [Registration] - **Date Time** 01-12-2021 **Priority** Medium **Description** When a User clicks the "Join EventExpress"button , the system redirects him to the registration form. **Pre-requisites** Move to: https://eventsexpress-test.azurewebsites.net/ **Test Cases** | S# | Action | Test Data | Expected Result | Actual Result | P/F | Automated | |:-----------:|:-----------:|:-----------:|:---------------:|:-------------:|:-----------:|:-----------:| | 1. <br> | Click the "Join EventsExpress" button from landing page<br> | <br> | Displayed the pop-up registration form <br> | <br> | <br> | <br> | | <br> | <br> | <br> | <br> | <br> | <br> | <br> | **Environment:** - OS: Windows10 - Browser: Chrome version 96.0.4664.45 - Version [e.g. 22] **Additional context** User Story #708
test
verified when the user press the join eventsexpress burton the login pop up form is displayed date time priority medium description when a user clicks the join eventexpress button the system redirects him to the registration form pre requisites move to test cases s action test data expected result actual result p f automated click the join eventsexpress button from landing page displayed the pop up registration form environment os browser chrome version version additional context user story
1
319,154
27,353,063,426
IssuesEvent
2023-02-27 10:56:14
airbytehq/airbyte
https://api.github.com/repos/airbytehq/airbyte
opened
E2E stream table tests - update exisitng stream table
area/platform team/platform-move area/frontend ui/connection ui/tests e2e-testing-tool platform-move/requires-grooming
## Tell us about the problem you're trying to solve Add E2E tests for stream functionality in the existing connection stream table ### Tasks - [ ] When a new stream is added to the table: after refreshing the source schema or just turning on the disabled sync stream. - [ ] ...will be updated...
2.0
E2E stream table tests - update exisitng stream table - ## Tell us about the problem you're trying to solve Add E2E tests for stream functionality in the existing connection stream table ### Tasks - [ ] When a new stream is added to the table: after refreshing the source schema or just turning on the disabled sync stream. - [ ] ...will be updated...
test
stream table tests update exisitng stream table tell us about the problem you re trying to solve add tests for stream functionality in the existing connection stream table tasks when a new stream is added to the table after refreshing the source schema or just turning on the disabled sync stream will be updated
1
46,002
18,923,534,209
IssuesEvent
2021-11-17 06:35:13
redhat-developer/app-services-cli
https://api.github.com/repos/redhat-developer/app-services-cli
opened
Print URL to view artifact in browser while getting the metadata
enhancement feature/service-registry priority/important-soon triage/accepted
### Feature or problem description `rhoas service-registry artifact metadata-get` should print the URL to UI to view the artifact as well. ### Proposed solution The artifact url in the ui looks something like this - `https://console.redhat.com/beta/application-services/service-registry/t/d5e6012f-ce90-4aaf-85b3-5f6e21a97e60/artifacts/default/rama-5/versions/latest` URL format: `<registry-browser-url>/artifacts/<group>/<id>/versions/<version>` Similar to #1321
1.0
Print URL to view artifact in browser while getting the metadata - ### Feature or problem description `rhoas service-registry artifact metadata-get` should print the URL to UI to view the artifact as well. ### Proposed solution The artifact url in the ui looks something like this - `https://console.redhat.com/beta/application-services/service-registry/t/d5e6012f-ce90-4aaf-85b3-5f6e21a97e60/artifacts/default/rama-5/versions/latest` URL format: `<registry-browser-url>/artifacts/<group>/<id>/versions/<version>` Similar to #1321
non_test
print url to view artifact in browser while getting the metadata feature or problem description rhoas service registry artifact metadata get should print the url to ui to view the artifact as well proposed solution the artifact url in the ui looks something like this url format artifacts versions similar to
0
248,146
26,784,832,530
IssuesEvent
2023-02-01 01:18:16
ElCattivo13/ECqr
https://api.github.com/repos/ElCattivo13/ECqr
opened
cli-13.2.6.tgz: 1 vulnerabilities (highest severity is: 5.3)
security vulnerability
<details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cli-13.2.6.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/http-cache-semantics/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (cli version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-25881](https://www.mend.io/vulnerability-database/CVE-2022-25881) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | http-cache-semantics-4.1.0.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-25881</summary> ### Vulnerable Library - <b>http-cache-semantics-4.1.0.tgz</b></p> <p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p> <p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/http-cache-semantics/package.json</p> <p> Dependency Hierarchy: - cli-13.2.6.tgz (Root Library) - pacote-12.0.3.tgz - npm-registry-fetch-12.0.2.tgz - make-fetch-happen-10.0.5.tgz - :x: **http-cache-semantics-4.1.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library. <p>Publish Date: 2023-01-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p> <p>Release Date: 2023-01-31</p> <p>Fix Resolution: http-cache-semantics - 4.1.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
True
cli-13.2.6.tgz: 1 vulnerabilities (highest severity is: 5.3) - <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>cli-13.2.6.tgz</b></p></summary> <p></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/http-cache-semantics/package.json</p> <p> </details> ## Vulnerabilities | CVE | Severity | <img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS | Dependency | Type | Fixed in (cli version) | Remediation Available | | ------------- | ------------- | ----- | ----- | ----- | ------------- | --- | | [CVE-2022-25881](https://www.mend.io/vulnerability-database/CVE-2022-25881) | <img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Medium | 5.3 | http-cache-semantics-4.1.0.tgz | Transitive | N/A* | &#10060; | <p>*For some transitive vulnerabilities, there is no version of direct dependency with a fix. Check the section "Details" below to see if there is a version of transitive dependency where vulnerability is fixed.</p> ## Details <details> <summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> CVE-2022-25881</summary> ### Vulnerable Library - <b>http-cache-semantics-4.1.0.tgz</b></p> <p>Parses Cache-Control and other headers. Helps building correct HTTP caches and proxies</p> <p>Library home page: <a href="https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz">https://registry.npmjs.org/http-cache-semantics/-/http-cache-semantics-4.1.0.tgz</a></p> <p>Path to dependency file: /package.json</p> <p>Path to vulnerable library: /node_modules/http-cache-semantics/package.json</p> <p> Dependency Hierarchy: - cli-13.2.6.tgz (Root Library) - pacote-12.0.3.tgz - npm-registry-fetch-12.0.2.tgz - make-fetch-happen-10.0.5.tgz - :x: **http-cache-semantics-4.1.0.tgz** (Vulnerable Library) <p>Found in base branch: <b>master</b></p> </p> <p></p> ### Vulnerability Details <p> This affects versions of the package http-cache-semantics before 4.1.1. The issue can be exploited via malicious request header values sent to a server, when that server reads the cache policy from the request using this library. <p>Publish Date: 2023-01-31 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2022-25881>CVE-2022-25881</a></p> </p> <p></p> ### CVSS 3 Score Details (<b>5.3</b>) <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> <p></p> ### Suggested Fix <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.cve.org/CVERecord?id=CVE-2022-25881">https://www.cve.org/CVERecord?id=CVE-2022-25881</a></p> <p>Release Date: 2023-01-31</p> <p>Fix Resolution: http-cache-semantics - 4.1.1</p> </p> <p></p> Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github) </details>
non_test
cli tgz vulnerabilities highest severity is vulnerable library cli tgz path to dependency file package json path to vulnerable library node modules http cache semantics package json vulnerabilities cve severity cvss dependency type fixed in cli version remediation available medium http cache semantics tgz transitive n a for some transitive vulnerabilities there is no version of direct dependency with a fix check the section details below to see if there is a version of transitive dependency where vulnerability is fixed details cve vulnerable library http cache semantics tgz parses cache control and other headers helps building correct http caches and proxies library home page a href path to dependency file package json path to vulnerable library node modules http cache semantics package json dependency hierarchy cli tgz root library pacote tgz npm registry fetch tgz make fetch happen tgz x http cache semantics tgz vulnerable library found in base branch master vulnerability details this affects versions of the package http cache semantics before the issue can be exploited via malicious request header values sent to a server when that server reads the cache policy from the request using this library publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution http cache semantics step up your open source security game with mend
0
772,151
27,108,564,313
IssuesEvent
2023-02-15 13:52:54
zephyrproject-rtos/zephyr
https://api.github.com/repos/zephyrproject-rtos/zephyr
closed
The handle of att indication violates the spec
bug priority: low area: Bluetooth Host
According to spec 5.3, 3.4.7.2 ATT_HANDLE_VALUE_IND, _"If the attribute value is longer than (ATT_MTU-3) octets, then only the first (ATT_MTU - 3) octets of this attributes value can be sent in an indication. Note: For a client to get a long attribute, it must use the ATT_READ_BLOB_REQ PDU (see Section 3.4.4.5) following the receipt of this indication."_ Both the handle of sending att_indication and receving att_indication violate the above statement. the ATT_MTU isn't considered.
1.0
The handle of att indication violates the spec - According to spec 5.3, 3.4.7.2 ATT_HANDLE_VALUE_IND, _"If the attribute value is longer than (ATT_MTU-3) octets, then only the first (ATT_MTU - 3) octets of this attributes value can be sent in an indication. Note: For a client to get a long attribute, it must use the ATT_READ_BLOB_REQ PDU (see Section 3.4.4.5) following the receipt of this indication."_ Both the handle of sending att_indication and receving att_indication violate the above statement. the ATT_MTU isn't considered.
non_test
the handle of att indication violates the spec according to spec att handle value ind if the attribute value is longer than att mtu octets then only the first att mtu octets of this attributes value can be sent in an indication note for a client to get a long attribute it must use the att read blob req pdu see section following the receipt of this indication both the handle of sending att indication and receving att indication violate the above statement the att mtu isn t considered
0
174,129
13,458,016,687
IssuesEvent
2020-09-09 09:59:28
S7NetPlus/s7netplus
https://api.github.com/repos/S7NetPlus/s7netplus
reopened
getting exceptions when reading bytes with single request
retest
Since I am using the version 0.6.0 I am getting execptions when reading values from a DB in a S7-1500. The expection is as follows: >S7.Net.PlcException: Received 15 bytes: '32-03-00-00-00-0A-00-02-00-01-00-00-05-01-FF', expected 19 bytes. ---> S7.Net.PlcException: Received 15 bytes: '32-03-00-00-00-0A-00-02-00-01-00-00-05-01-FF', expected 19 bytes. bei S7.Net.Plc.AssertReadResponse(Byte[] s7Data, Int32 dataLength) bei S7.Net.Plc.ReadBytesWithSingleRequest(DataType dataType, Int32 db, Int32 startByteAdr, Int32 count) My internal log-file adresses often the same line of code as follows: ```csharp try { if (plcKonti2.IsConnected) <------- this line is adressed in the exception { #region Konti2ISA // K2 Rakel Laminierstation var statusRakelLaminierstation = plcKonti2.ReadBytes(DataType.DataBlock, 153, 0, 1); K2RakelLaminiereinheitFreigabe = statusRakelLaminierstation[0].SelectBit(0); } } ``` The connection to the plc is opened before I call the method above. The code was running for nearly 1 year without any exceptions at this line of code. I will check if this also happens with version 0.5.0.
1.0
getting exceptions when reading bytes with single request - Since I am using the version 0.6.0 I am getting execptions when reading values from a DB in a S7-1500. The expection is as follows: >S7.Net.PlcException: Received 15 bytes: '32-03-00-00-00-0A-00-02-00-01-00-00-05-01-FF', expected 19 bytes. ---> S7.Net.PlcException: Received 15 bytes: '32-03-00-00-00-0A-00-02-00-01-00-00-05-01-FF', expected 19 bytes. bei S7.Net.Plc.AssertReadResponse(Byte[] s7Data, Int32 dataLength) bei S7.Net.Plc.ReadBytesWithSingleRequest(DataType dataType, Int32 db, Int32 startByteAdr, Int32 count) My internal log-file adresses often the same line of code as follows: ```csharp try { if (plcKonti2.IsConnected) <------- this line is adressed in the exception { #region Konti2ISA // K2 Rakel Laminierstation var statusRakelLaminierstation = plcKonti2.ReadBytes(DataType.DataBlock, 153, 0, 1); K2RakelLaminiereinheitFreigabe = statusRakelLaminierstation[0].SelectBit(0); } } ``` The connection to the plc is opened before I call the method above. The code was running for nearly 1 year without any exceptions at this line of code. I will check if this also happens with version 0.5.0.
test
getting exceptions when reading bytes with single request since i am using the version i am getting execptions when reading values from a db in a the expection is as follows net plcexception received bytes ff expected bytes net plcexception received bytes ff expected bytes bei net plc assertreadresponse byte datalength bei net plc readbyteswithsinglerequest datatype datatype db startbyteadr count my internal log file adresses often the same line of code as follows csharp try if isconnected this line is adressed in the exception region rakel laminierstation var statusrakellaminierstation readbytes datatype datablock statusrakellaminierstation selectbit the connection to the plc is opened before i call the method above the code was running for nearly year without any exceptions at this line of code i will check if this also happens with version
1
335,986
30,111,944,496
IssuesEvent
2023-06-30 08:28:47
elastic/kibana
https://api.github.com/repos/elastic/kibana
closed
Failing test: Jest Tests.x-pack/plugins/cases/public/components/case_view - CaseViewPage Tabs description renders the descriptions user correctly
failed-test skipped-test Team:ResponseOps Feature:Cases
A test failed on a tracked branch ``` TestingLibraryElementError: Unable to find an element by: [data-test-subj="description-action"] Ignored nodes: comments, script, style <body> <div /> </body> at Object.getElementError (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/config.js:40:19) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:90:38 at /var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:62:17 at /var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:111:19 at Object.getByTestId (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/case_view/case_view_page.test.tsx:591:43) at Promise.then.completed (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:289:28) at new Promise (<anonymous>) at callAsyncCircusFn (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:222:10) at _callCircusTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:248:40) at _runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:184:3) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:86:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at run (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:26:3) at runAndTransformResultsToJestFormat (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:120:21) at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:79:19) at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:367:16) at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:444:34) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/28165#018700b9-e60d-4992-8f25-cb96c43b1925) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/cases/public/components/case_view","test.name":"CaseViewPage Tabs description renders the descriptions user correctly","test.failCount":2}} -->
2.0
Failing test: Jest Tests.x-pack/plugins/cases/public/components/case_view - CaseViewPage Tabs description renders the descriptions user correctly - A test failed on a tracked branch ``` TestingLibraryElementError: Unable to find an element by: [data-test-subj="description-action"] Ignored nodes: comments, script, style <body> <div /> </body> at Object.getElementError (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/config.js:40:19) at /var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:90:38 at /var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:62:17 at /var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/@testing-library/dom/dist/query-helpers.js:111:19 at Object.getByTestId (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/x-pack/plugins/cases/public/components/case_view/case_view_page.test.tsx:591:43) at Promise.then.completed (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:289:28) at new Promise (<anonymous>) at callAsyncCircusFn (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/utils.js:222:10) at _callCircusTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:248:40) at _runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:184:3) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:86:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at _runTestsForDescribeBlock (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:81:9) at run (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/run.js:26:3) at runAndTransformResultsToJestFormat (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapterInit.js:120:21) at jestAdapter (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-circus/build/legacy-code-todo-rewrite/jestAdapter.js:79:19) at runTestInternal (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:367:16) at runTest (/var/lib/buildkite-agent/builds/kb-n2-4-spot-42644c847ce73e72/elastic/kibana-on-merge/kibana/node_modules/jest-runner/build/runTest.js:444:34) ``` First failure: [CI Build - main](https://buildkite.com/elastic/kibana-on-merge/builds/28165#018700b9-e60d-4992-8f25-cb96c43b1925) <!-- kibanaCiData = {"failed-test":{"test.class":"Jest Tests.x-pack/plugins/cases/public/components/case_view","test.name":"CaseViewPage Tabs description renders the descriptions user correctly","test.failCount":2}} -->
test
failing test jest tests x pack plugins cases public components case view caseviewpage tabs description renders the descriptions user correctly a test failed on a tracked branch testinglibraryelementerror unable to find an element by ignored nodes comments script style at object getelementerror var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist config js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules testing library dom dist query helpers js at object getbytestid var lib buildkite agent builds kb spot elastic kibana on merge kibana x pack plugins cases public components case view case view page test tsx at promise then completed var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build utils js at new promise at callasynccircusfn var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build utils js at callcircustest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runtestsfordescribeblock var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at run var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build run js at runandtransformresultstojestformat var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapterinit js at jestadapter var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest circus build legacy code todo rewrite jestadapter js at runtestinternal var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js at runtest var lib buildkite agent builds kb spot elastic kibana on merge kibana node modules jest runner build runtest js first failure
1
340,804
30,543,851,071
IssuesEvent
2023-07-20 01:01:27
dotnet/maui
https://api.github.com/repos/dotnet/maui
closed
.NET MAUI Shell app page navigation memory leaks
t/bug area/shell 🐢 s/verified s/try-latest-version memory-leak 💦
### Description In a .NET MAUI Shell app running on Windows it looks like pages are held in memory after navigating to and then away from the page. This results in an increase in memory over time as pages are opened and closed. We noticed this issue as we have some particularly memory intensive pages that result in steep memory jumps as they are opened and closed. ### Steps to Reproduce Setup 1. Create new Maui app. File -> New Solution -> Maui App 2. Create two new MAUI ContentPages `SecondaryPage` and `TertiaryPage` 3. Add `SecondaryPage` to the Shell Flyout Items 4. Add a Button to the `SecondaryPage` with a clicked event `await Navigation.PushAsync(new TertiaryPage());` Testing 1. Run the app targeting Windows 2. Navigate to the `SecondaryPage` via the Flyout 3. Navigate to the `TertiaryPage` via the Button on `SecondaryPage` 4. Click the back button 5. Take a snapshot of the memory usage 6. Navigate to the `TertiaryPage` via the Button on `SecondaryPage` 7. Click the back button 8. Take a snapshot of the memory usage 6. Repeat as many times as desired Result: Looking at the memory usage snapshots the number of instances of `TertiaryPage` in memory increases with the number of times the page has been opened. ![image](https://github.com/dotnet/maui/assets/99325835/ebdf2796-133c-4f44-83c8-15e4848c57fa) ![image](https://github.com/dotnet/maui/assets/99325835/8ca3e191-aa72-495a-8fd1-cffadf087631) Expected: The page should no longer exist in memory after navigating away, the number of instances in memory should not increase as the page is repeatedly opened. ### Link to public reproduction project repository https://github.com/duffh/MauiMemoryLeakTestApp ### Version with bug 8.0.0-preview.5.8529 ### Last version that worked well Unknown/Other ### Affected platforms Windows, I was *not* able test on other platforms ### Affected platform versions Windows SDK 10.0.19041.0 ### Did you find any workaround? _No response_ ### Relevant log output _No response_
1.0
.NET MAUI Shell app page navigation memory leaks - ### Description In a .NET MAUI Shell app running on Windows it looks like pages are held in memory after navigating to and then away from the page. This results in an increase in memory over time as pages are opened and closed. We noticed this issue as we have some particularly memory intensive pages that result in steep memory jumps as they are opened and closed. ### Steps to Reproduce Setup 1. Create new Maui app. File -> New Solution -> Maui App 2. Create two new MAUI ContentPages `SecondaryPage` and `TertiaryPage` 3. Add `SecondaryPage` to the Shell Flyout Items 4. Add a Button to the `SecondaryPage` with a clicked event `await Navigation.PushAsync(new TertiaryPage());` Testing 1. Run the app targeting Windows 2. Navigate to the `SecondaryPage` via the Flyout 3. Navigate to the `TertiaryPage` via the Button on `SecondaryPage` 4. Click the back button 5. Take a snapshot of the memory usage 6. Navigate to the `TertiaryPage` via the Button on `SecondaryPage` 7. Click the back button 8. Take a snapshot of the memory usage 6. Repeat as many times as desired Result: Looking at the memory usage snapshots the number of instances of `TertiaryPage` in memory increases with the number of times the page has been opened. ![image](https://github.com/dotnet/maui/assets/99325835/ebdf2796-133c-4f44-83c8-15e4848c57fa) ![image](https://github.com/dotnet/maui/assets/99325835/8ca3e191-aa72-495a-8fd1-cffadf087631) Expected: The page should no longer exist in memory after navigating away, the number of instances in memory should not increase as the page is repeatedly opened. ### Link to public reproduction project repository https://github.com/duffh/MauiMemoryLeakTestApp ### Version with bug 8.0.0-preview.5.8529 ### Last version that worked well Unknown/Other ### Affected platforms Windows, I was *not* able test on other platforms ### Affected platform versions Windows SDK 10.0.19041.0 ### Did you find any workaround? _No response_ ### Relevant log output _No response_
test
net maui shell app page navigation memory leaks description in a net maui shell app running on windows it looks like pages are held in memory after navigating to and then away from the page this results in an increase in memory over time as pages are opened and closed we noticed this issue as we have some particularly memory intensive pages that result in steep memory jumps as they are opened and closed steps to reproduce setup create new maui app file new solution maui app create two new maui contentpages secondarypage and tertiarypage add secondarypage to the shell flyout items add a button to the secondarypage with a clicked event await navigation pushasync new tertiarypage testing run the app targeting windows navigate to the secondarypage via the flyout navigate to the tertiarypage via the button on secondarypage click the back button take a snapshot of the memory usage navigate to the tertiarypage via the button on secondarypage click the back button take a snapshot of the memory usage repeat as many times as desired result looking at the memory usage snapshots the number of instances of tertiarypage in memory increases with the number of times the page has been opened expected the page should no longer exist in memory after navigating away the number of instances in memory should not increase as the page is repeatedly opened link to public reproduction project repository version with bug preview last version that worked well unknown other affected platforms windows i was not able test on other platforms affected platform versions windows sdk did you find any workaround no response relevant log output no response
1
289,519
31,933,076,511
IssuesEvent
2023-09-19 08:43:56
Trinadh465/linux-4.1.15_CVE-2023-4128
https://api.github.com/repos/Trinadh465/linux-4.1.15_CVE-2023-4128
opened
CVE-2021-0447 (Medium) detected in linuxlinux-4.6
Mend: dependency security vulnerability
## CVE-2021-0447 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/l2tp/l2tp_ppp.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/l2tp/l2tp_ppp.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A security vulnerability was found in Linux Kernel before 3.2.99, 3.16.54, 4.4.225, 4.9.225 and 4.14.182. Pppol2tp_session_create() registers sessions that can't have their corresponding socket initialised. <p>Publish Date: 2020-11-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0447>CVE-2021-0447</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-0447">https://www.linuxkernelcves.com/cves/CVE-2021-0447</a></p> <p>Release Date: 2020-11-07</p> <p>Fix Resolution: v3.2.99,v3.16.54,v4.4.225,v4.9.225,v4.14.182,v4.15-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2021-0447 (Medium) detected in linuxlinux-4.6 - ## CVE-2021-0447 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>linuxlinux-4.6</b></p></summary> <p> <p>The Linux Kernel</p> <p>Library home page: <a href=https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux>https://mirrors.edge.kernel.org/pub/linux/kernel/v4.x/?wsslib=linux</a></p> <p>Found in HEAD commit: <a href="https://github.com/Trinadh465/linux-4.1.15_CVE-2023-4128/commit/0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8">0c6c8d8c809f697cd5fc581c6c08e9ad646c55a8</a></p> <p>Found in base branch: <b>master</b></p></p> </details> </p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Source Files (2)</summary> <p></p> <p> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/l2tp/l2tp_ppp.c</b> <img src='https://s3.amazonaws.com/wss-public/bitbucketImages/xRedImage.png' width=19 height=20> <b>/net/l2tp/l2tp_ppp.c</b> </p> </details> <p></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png?' width=19 height=20> Vulnerability Details</summary> <p> A security vulnerability was found in Linux Kernel before 3.2.99, 3.16.54, 4.4.225, 4.9.225 and 4.14.182. Pppol2tp_session_create() registers sessions that can't have their corresponding socket initialised. <p>Publish Date: 2020-11-07 <p>URL: <a href=https://www.mend.io/vulnerability-database/CVE-2021-0447>CVE-2021-0447</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>4.6</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: High - Privileges Required: Low - User Interaction: Required - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: Low - Integrity Impact: Low - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://www.linuxkernelcves.com/cves/CVE-2021-0447">https://www.linuxkernelcves.com/cves/CVE-2021-0447</a></p> <p>Release Date: 2020-11-07</p> <p>Fix Resolution: v3.2.99,v3.16.54,v4.4.225,v4.9.225,v4.14.182,v4.15-rc1</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve medium detected in linuxlinux cve medium severity vulnerability vulnerable library linuxlinux the linux kernel library home page a href found in head commit a href found in base branch master vulnerable source files net ppp c net ppp c vulnerability details a security vulnerability was found in linux kernel before and session create registers sessions that can t have their corresponding socket initialised publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity high privileges required low user interaction required scope unchanged impact metrics confidentiality impact low integrity impact low availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution step up your open source security game with mend
0
213,761
16,535,851,975
IssuesEvent
2021-05-27 11:46:27
tracim/tracim
https://api.github.com/repos/tracim/tracim
closed
Bug: Recent activites can display less than 15 activities when loaded
activity-feed add to changelog frontend manually tested
## Description and expectations Even if there are many more activities (clicking on the `See more` button displays them). <!-- If you know which component may be affected, feel free to mention it. --> ### How to reproduce <!-- Detail steps to reproduce the bug: --> 1. Create a space 2. Add many members to it (15+) 3. Remove those members 4. Display the space recent activities ### Version information <!-- Optional - describe the environment which reveals the bug: --> - Tracim version: 3.8.1 ## Screenshots <!-- Help us understand the problem with screenshots. You can use [peek](https://github.com/phw/peek) to generate animated GIFs. --> ________________________________________________ <!-- *The following sections are for the maintainers* --> ## Diagnostic The code filtering the activities to display is independent from the code fetching and constructing the activities. Filtering should occur during the fetching/creation of activities (`ActivityList.jsx`/`withActivity.js`). ## Prerequisites <!-- Optional - list the issues that must be solved or what needs to be done before handling this issue. --> <!-- ## Required sections, if relevant ## --> <!-- - To be discussed before development --> <!-- - Interface --> <!-- - Translations --> <!-- - Workaround --> <!-- - Extra information --> <!-- - Implemented solution -->
1.0
Bug: Recent activites can display less than 15 activities when loaded - ## Description and expectations Even if there are many more activities (clicking on the `See more` button displays them). <!-- If you know which component may be affected, feel free to mention it. --> ### How to reproduce <!-- Detail steps to reproduce the bug: --> 1. Create a space 2. Add many members to it (15+) 3. Remove those members 4. Display the space recent activities ### Version information <!-- Optional - describe the environment which reveals the bug: --> - Tracim version: 3.8.1 ## Screenshots <!-- Help us understand the problem with screenshots. You can use [peek](https://github.com/phw/peek) to generate animated GIFs. --> ________________________________________________ <!-- *The following sections are for the maintainers* --> ## Diagnostic The code filtering the activities to display is independent from the code fetching and constructing the activities. Filtering should occur during the fetching/creation of activities (`ActivityList.jsx`/`withActivity.js`). ## Prerequisites <!-- Optional - list the issues that must be solved or what needs to be done before handling this issue. --> <!-- ## Required sections, if relevant ## --> <!-- - To be discussed before development --> <!-- - Interface --> <!-- - Translations --> <!-- - Workaround --> <!-- - Extra information --> <!-- - Implemented solution -->
test
bug recent activites can display less than activities when loaded description and expectations even if there are many more activities clicking on the see more button displays them how to reproduce create a space add many members to it remove those members display the space recent activities version information tracim version screenshots diagnostic the code filtering the activities to display is independent from the code fetching and constructing the activities filtering should occur during the fetching creation of activities activitylist jsx withactivity js prerequisites
1
221,045
17,288,360,323
IssuesEvent
2021-07-24 07:08:39
WoWManiaUK/Redemption
https://api.github.com/repos/WoWManiaUK/Redemption
closed
[Item/Mount] Reindeer's Reins
Fixed on PTR - Tester Confirmed
**Links:** https://www.wow-mania.com/armory/?item=21044 https://www.wow-mania.com/armory/?spell=25858 **What is Happening:** [Reindeer's Reins](https://www.wow-mania.com/armory/?spell=25858 ) is an item that is not obtainable by normal means, but has been awarded before in events as it is a pretty festive mount that looks quite inoffensive at first glance. However, this item allows you to mount a 60% speed mount with no cast time. Obviously this can be very powerful and is currently being abused in Wintergrasp (could also be done in battlegrounds). Any spell that takes you out of combat such as Feign Death, Vanish or whatever, allow for an instant mount-up seemingly mid-combat. Otherwise a fleeing opponent can mount this festive partner instantly as soon as they exit combat, giving them an advantage over whoever might be chasing them. **What Should happen:** This item should either not be available to regular players, or should be changed to require the standard 1.5s cast in order to mount up.
1.0
[Item/Mount] Reindeer's Reins - **Links:** https://www.wow-mania.com/armory/?item=21044 https://www.wow-mania.com/armory/?spell=25858 **What is Happening:** [Reindeer's Reins](https://www.wow-mania.com/armory/?spell=25858 ) is an item that is not obtainable by normal means, but has been awarded before in events as it is a pretty festive mount that looks quite inoffensive at first glance. However, this item allows you to mount a 60% speed mount with no cast time. Obviously this can be very powerful and is currently being abused in Wintergrasp (could also be done in battlegrounds). Any spell that takes you out of combat such as Feign Death, Vanish or whatever, allow for an instant mount-up seemingly mid-combat. Otherwise a fleeing opponent can mount this festive partner instantly as soon as they exit combat, giving them an advantage over whoever might be chasing them. **What Should happen:** This item should either not be available to regular players, or should be changed to require the standard 1.5s cast in order to mount up.
test
reindeer s reins links what is happening is an item that is not obtainable by normal means but has been awarded before in events as it is a pretty festive mount that looks quite inoffensive at first glance however this item allows you to mount a speed mount with no cast time obviously this can be very powerful and is currently being abused in wintergrasp could also be done in battlegrounds any spell that takes you out of combat such as feign death vanish or whatever allow for an instant mount up seemingly mid combat otherwise a fleeing opponent can mount this festive partner instantly as soon as they exit combat giving them an advantage over whoever might be chasing them what should happen this item should either not be available to regular players or should be changed to require the standard cast in order to mount up
1
790,635
27,830,950,291
IssuesEvent
2023-03-20 04:49:49
AY2223S2-CS2103T-W14-1/tp
https://api.github.com/repos/AY2223S2-CS2103T-W14-1/tp
closed
Refactor UI package to be more extensible
priority.med type.refactor severity.Medium
UI package is a messy platter of java files with little to no organisation.
1.0
Refactor UI package to be more extensible - UI package is a messy platter of java files with little to no organisation.
non_test
refactor ui package to be more extensible ui package is a messy platter of java files with little to no organisation
0
757,048
26,494,853,051
IssuesEvent
2023-01-18 04:04:06
pulsejet/memories
https://api.github.com/repos/pulsejet/memories
closed
Incorrect behavior of on this day on year boundaries
bug priority: high
Shows up as multiple sections with the same title. Just club these blindly.
1.0
Incorrect behavior of on this day on year boundaries - Shows up as multiple sections with the same title. Just club these blindly.
non_test
incorrect behavior of on this day on year boundaries shows up as multiple sections with the same title just club these blindly
0
221,596
24,650,961,532
IssuesEvent
2022-10-17 18:36:19
RG4421/featuretools
https://api.github.com/repos/RG4421/featuretools
opened
CVE-2022-42969 (Medium) detected in py-1.10.0-py2.py3-none-any.whl
security vulnerability
## CVE-2022-42969 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>py-1.10.0-py2.py3-none-any.whl</b></p></summary> <p>library with cross-python path, ini-parsing, io, code, log facilities</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/67/32/6fe01cfc3d1a27c92fdbcdfc3f67856da8cbadf0dd9f2e18055202b2dc62/py-1.10.0-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/67/32/6fe01cfc3d1a27c92fdbcdfc3f67856da8cbadf0dd9f2e18055202b2dc62/py-1.10.0-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /test-requirements.txt</p> <p>Path to vulnerable library: /test-requirements.txt</p> <p> Dependency Hierarchy: - pytest_xdist-1.26.1-py2.py3-none-any.whl (Root Library) - pytest-4.6.11-py2.py3-none-any.whl - :x: **py-1.10.0-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The py library through 1.11.0 for Python allows remote attackers to conduct a ReDoS (Regular expression Denial of Service) attack via a Subversion repository with crafted info data, because the InfoSvnCommand argument is mishandled. <p>Publish Date: 2022-10-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42969>CVE-2022-42969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-42969">https://nvd.nist.gov/vuln/detail/CVE-2022-42969</a></p> <p>Release Date: 2022-10-16</p> <p>Fix Resolution: py - 1.5.0</p> </p> </details> <p></p>
True
CVE-2022-42969 (Medium) detected in py-1.10.0-py2.py3-none-any.whl - ## CVE-2022-42969 - Medium Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>py-1.10.0-py2.py3-none-any.whl</b></p></summary> <p>library with cross-python path, ini-parsing, io, code, log facilities</p> <p>Library home page: <a href="https://files.pythonhosted.org/packages/67/32/6fe01cfc3d1a27c92fdbcdfc3f67856da8cbadf0dd9f2e18055202b2dc62/py-1.10.0-py2.py3-none-any.whl">https://files.pythonhosted.org/packages/67/32/6fe01cfc3d1a27c92fdbcdfc3f67856da8cbadf0dd9f2e18055202b2dc62/py-1.10.0-py2.py3-none-any.whl</a></p> <p>Path to dependency file: /test-requirements.txt</p> <p>Path to vulnerable library: /test-requirements.txt</p> <p> Dependency Hierarchy: - pytest_xdist-1.26.1-py2.py3-none-any.whl (Root Library) - pytest-4.6.11-py2.py3-none-any.whl - :x: **py-1.10.0-py2.py3-none-any.whl** (Vulnerable Library) <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/medium_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The py library through 1.11.0 for Python allows remote attackers to conduct a ReDoS (Regular expression Denial of Service) attack via a Subversion repository with crafted info data, because the InfoSvnCommand argument is mishandled. <p>Publish Date: 2022-10-16 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-42969>CVE-2022-42969</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>5.3</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: Low </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://nvd.nist.gov/vuln/detail/CVE-2022-42969">https://nvd.nist.gov/vuln/detail/CVE-2022-42969</a></p> <p>Release Date: 2022-10-16</p> <p>Fix Resolution: py - 1.5.0</p> </p> </details> <p></p>
non_test
cve medium detected in py none any whl cve medium severity vulnerability vulnerable library py none any whl library with cross python path ini parsing io code log facilities library home page a href path to dependency file test requirements txt path to vulnerable library test requirements txt dependency hierarchy pytest xdist none any whl root library pytest none any whl x py none any whl vulnerable library found in base branch main vulnerability details the py library through for python allows remote attackers to conduct a redos regular expression denial of service attack via a subversion repository with crafted info data because the infosvncommand argument is mishandled publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact low for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution py
0
56,946
3,081,220,921
IssuesEvent
2015-08-22 14:06:12
bitfighter/bitfighter
https://api.github.com/repos/bitfighter/bitfighter
closed
Shaky commander's map when other player ships move out of bounds
019 bug imported Priority-Medium
_From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on November 09, 2012 16:44:08_ I believe this is caused when an object suddenly changes the size of the map, such as when shooting off the edge of the map and bullets die. Could be solved by detecting changes of world size and easing a transition from one size to the next. Could involve shrinking or growing. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=173_
1.0
Shaky commander's map when other player ships move out of bounds - _From [watusim...@bitfighter.org](https://code.google.com/u/105427273526970468779/) on November 09, 2012 16:44:08_ I believe this is caused when an object suddenly changes the size of the map, such as when shooting off the edge of the map and bullets die. Could be solved by detecting changes of world size and easing a transition from one size to the next. Could involve shrinking or growing. _Original issue: http://code.google.com/p/bitfighter/issues/detail?id=173_
non_test
shaky commander s map when other player ships move out of bounds from on november i believe this is caused when an object suddenly changes the size of the map such as when shooting off the edge of the map and bullets die could be solved by detecting changes of world size and easing a transition from one size to the next could involve shrinking or growing original issue
0
176,776
13,652,833,710
IssuesEvent
2020-09-27 09:36:05
coloredcow-admin/glific-website
https://api.github.com/repos/coloredcow-admin/glific-website
closed
Clarification on footer links
status : ready to test
@SunnaMalhotra @abhi1203 please provide details on the pages on these links in footer <img width="1494" alt="Screenshot 2020-09-23 at 2 10 11 PM" src="https://user-images.githubusercontent.com/12884297/93988881-dd743380-fda6-11ea-9794-9726693b65f7.png">
1.0
Clarification on footer links - @SunnaMalhotra @abhi1203 please provide details on the pages on these links in footer <img width="1494" alt="Screenshot 2020-09-23 at 2 10 11 PM" src="https://user-images.githubusercontent.com/12884297/93988881-dd743380-fda6-11ea-9794-9726693b65f7.png">
test
clarification on footer links sunnamalhotra please provide details on the pages on these links in footer img width alt screenshot at pm src
1
489,814
14,112,492,653
IssuesEvent
2020-11-07 05:41:29
kubeflow/kubeflow
https://api.github.com/repos/kubeflow/kubeflow
closed
Kubefow+ LetsEncrypt+ Istio Dex 1.0.2 yaml
area/istio kind/bug lifecycle/stale platform/onprem priority/p2
/kind bug I am unable to receive a working certificate thru letsencrypt prod/staging API: ![image](https://user-images.githubusercontent.com/27989098/88207986-2f48f200-cc6e-11ea-87ca-e84c5d42cef6.png) The current setup includes below approach K8s cluster setup with 1.0.2 Istio Dex Yaml, mentioned in kubeflow.org documentation, which comes with istio 1.3.1 and cert-manager by default. As my major task is to complete this with kubeflow enabled with istio-ingressgateway along with HTTPS Lets-Encrypt ACME SSL Certificate generation. Kindly let me know a possible solution. **Environment:** - Kubeflow version: - K8s API version 1.14 - kfctl version: 1.0.2 - Kubernetes platform: GKE - Kubernetes version: (kubectl version --short Client Version: v1.17.0 Server Version: v1.14.10-gke.36) - OS: default Istio -1.3.1 Cert-Manager: the same which comes with default manifest.
1.0
Kubefow+ LetsEncrypt+ Istio Dex 1.0.2 yaml - /kind bug I am unable to receive a working certificate thru letsencrypt prod/staging API: ![image](https://user-images.githubusercontent.com/27989098/88207986-2f48f200-cc6e-11ea-87ca-e84c5d42cef6.png) The current setup includes below approach K8s cluster setup with 1.0.2 Istio Dex Yaml, mentioned in kubeflow.org documentation, which comes with istio 1.3.1 and cert-manager by default. As my major task is to complete this with kubeflow enabled with istio-ingressgateway along with HTTPS Lets-Encrypt ACME SSL Certificate generation. Kindly let me know a possible solution. **Environment:** - Kubeflow version: - K8s API version 1.14 - kfctl version: 1.0.2 - Kubernetes platform: GKE - Kubernetes version: (kubectl version --short Client Version: v1.17.0 Server Version: v1.14.10-gke.36) - OS: default Istio -1.3.1 Cert-Manager: the same which comes with default manifest.
non_test
kubefow letsencrypt istio dex yaml kind bug i am unable to receive a working certificate thru letsencrypt prod staging api the current setup includes below approach cluster setup with istio dex yaml mentioned in kubeflow org documentation which comes with istio and cert manager by default as my major task is to complete this with kubeflow enabled with istio ingressgateway along with https lets encrypt acme ssl certificate generation kindly let me know a possible solution environment kubeflow version api version kfctl version kubernetes platform gke kubernetes version kubectl version short client version server version gke os default istio cert manager the same which comes with default manifest
0
107,120
9,202,174,348
IssuesEvent
2019-03-07 21:39:33
brave/brave-browser
https://api.github.com/repos/brave/brave-browser
opened
Manual test run on Linux for 0.61.x - Release Candidate
OS/Linux QA/Yes release-notes/exclude tests
### Installer - [ ] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window ### Data(Upgrade from previous release) - [ ] Make sure that data from the last version appears in the new version OK - [ ] With data from the last version, verify that - [ ] bookmarks on the bookmark toolbar and bookmark folders can be opened - [ ] cookies are preserved - [ ] installed extensions are retained and work correctly - [ ] opened tabs can be reloaded - [ ] stored passwords are preserved - [ ] unpinned tabs can be pinned ## Extensions/Plugins tests - [ ] Verify one item from Brave Update server is installable (Example: Ad-block DAT file on fresh extension) - [ ] Verify one item from Google Update server is installable (Example: Extensions from CWS) - [ ] Verify PDFJS, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled - [ ] Verify older version of an extension gets updated to new version via Google server - [ ] Verify older version of an extension gets updated to new version via Brave server - [ ] Verify magnet links and .torrent files loads Torrent viewer page and able to download torrent ### CWS - [ ] Verify installing ABP from CWS shows warning message `NOT A RECOMMENDED BRAVE EXTENSION!`but still allows to install the extension - [ ] Verify installing LastPass from CWS doesn't show any warning message - [ ] Verify installing an extension that is not vetted by Brave gets blocked ### PDF - [ ] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf - [ ] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf ### Widevine - [ ] Verify `Widevine Notification` is shown when you visit Netflix for the first time - [ ] Test that you can stream on Netflix on a fresh profile after installing Widevine ### Bravery settings - [ ] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/ - [ ] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/ - [ ] Verify that toggling `Ads and trackers blocked` works as expected - [ ] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items - [ ] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly - [ ] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked ### Fingerprint Tests - [ ] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00` - [ ] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on - [ ] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/ - [ ] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on ### Rewards - [ ] Verify wallet is auto created after enabling rewards - [ ] Verify account balance shows correct BAT and USD value - [ ] Verify you are able to restore a wallet - [ ] Verify wallet address matches the QR code that is generated under `Add funds` - [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel - [ ] Verify adding funds via any of the currencies flows into wallet after specified amount of time - [ ] Verify adding funds to an existing wallet with amount, adjusts the BAT value appropriately - [ ] Verify monthly budget shows correct BAT and USD value - [ ] Verify you are able to exclude a publisher from the auto-contribute table by clicking on the `x` in auto-contribute table and popup list of sites - [ ] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel - [ ] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %) - [ ] Verify when you click on `Send a tip`, the custom tip banner displays - [ ] Verify you are able to make one-time tip and they display in tips panel - [ ] Verify you are able to make recurring tip and they display in tips panel - [ ] Verify you can tip a verified publisher - [ ] Verify you can tip a verified YouTube creator - [ ] Verify tip panel shows a verified checkmark for a verified publisher/verified YouTube creator - [ ] Verify tip panel shows a message about unverified publisher - [ ] Verify BR panel shows message about an unverified publisher - [ ] Verify you are able to perform a contribution - [ ] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators - [ ] Verify that disabling Rewards and enabling it again does not lose state - [ ] Verify that disabling auto-contribute and enabling it again does not lose state - [ ] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings - [ ] Upgrade from older version - [ ] Verify the wallet balance is retained and wallet backup code isn't corrupted - [ ] Verify auto-contribute list is not lost after upgrade - [ ] Verify tips list is not lost after upgrade - [ ] Verify wallet panel transactions list is not lost after upgrade ### Tor Tabs - [ ] Visit https://check.torproject.org in a Tor window, ensure its shows success message for using a Tor exit node - [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads - [ ] Visit https://protonirockerxow.onion/ in a Tor window, ensure login page is shown - [ ] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown ### Session storage - [ ] Temporarily move away your browser profile and test that a new profile is created when browser is launched - macOS - `~/Library/Application\ Support/BraveSoftware/` - Windows - `%userprofile%\appdata\Local\BraveSoftware\` - Linux(Ubuntu) - `~/.config/BraveSoftware/` - [ ] Test that windows and tabs restore when closed, including active tab - [ ] Ensure that the tabs in the above session are being lazy loaded when the session is restored ## Update tests - [ ] Verify visiting `brave://settings/help` triggers update check - [ ] Verify once update is downloaded, prompts to `Relaunch` to install update ## Chromium upgrade tests - [ ] Verify `brave://gpu` on Brave and `chrome://gpu` on Chrome are similar for the same Chromium version on both browsers #### Adblock - [ ] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/` #### Components - [ ] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components
1.0
Manual test run on Linux for 0.61.x - Release Candidate - ### Installer - [ ] Check signature: If OS Run `spctl --assess --verbose /Applications/Brave-Browser-Beta.app/` and make sure it returns `accepted`. If Windows right click on the `brave_installer-x64.exe` and go to Properties, go to the Digital Signatures tab and double click on the signature. Make sure it says "The digital signature is OK" in the popup window ### Data(Upgrade from previous release) - [ ] Make sure that data from the last version appears in the new version OK - [ ] With data from the last version, verify that - [ ] bookmarks on the bookmark toolbar and bookmark folders can be opened - [ ] cookies are preserved - [ ] installed extensions are retained and work correctly - [ ] opened tabs can be reloaded - [ ] stored passwords are preserved - [ ] unpinned tabs can be pinned ## Extensions/Plugins tests - [ ] Verify one item from Brave Update server is installable (Example: Ad-block DAT file on fresh extension) - [ ] Verify one item from Google Update server is installable (Example: Extensions from CWS) - [ ] Verify PDFJS, Torrent viewer extensions are installed automatically on fresh profile and cannot be disabled - [ ] Verify older version of an extension gets updated to new version via Google server - [ ] Verify older version of an extension gets updated to new version via Brave server - [ ] Verify magnet links and .torrent files loads Torrent viewer page and able to download torrent ### CWS - [ ] Verify installing ABP from CWS shows warning message `NOT A RECOMMENDED BRAVE EXTENSION!`but still allows to install the extension - [ ] Verify installing LastPass from CWS doesn't show any warning message - [ ] Verify installing an extension that is not vetted by Brave gets blocked ### PDF - [ ] Test that PDF is loaded over HTTPS at https://basicattentiontoken.org/BasicAttentionTokenWhitePaper-4.pdf - [ ] Test that PDF is loaded over HTTP at http://www.pdf995.com/samples/pdf.pdf ### Widevine - [ ] Verify `Widevine Notification` is shown when you visit Netflix for the first time - [ ] Test that you can stream on Netflix on a fresh profile after installing Widevine ### Bravery settings - [ ] Verify that HTTPS Everywhere works by loading http://https-everywhere.badssl.com/ - [ ] Turning HTTPS Everywhere off and shields off both disable the redirect to https://https-everywhere.badssl.com/ - [ ] Verify that toggling `Ads and trackers blocked` works as expected - [ ] Visit https://testsafebrowsing.appspot.com/s/phishing.html, verify that Safe Browsing (via our Proxy) works for all the listed items - [ ] Visit https://brianbondy.com/ and then turn on script blocking, page should not load. Allow it from the script blocking UI in the URL bar and it should load the page correctly - [ ] Test that 3rd party storage results are blank at https://jsfiddle.net/7ke9r14a/9/ when 3rd party cookies are blocked and not blank when 3rd party cookies are unblocked ### Fingerprint Tests - [ ] Visit https://jsfiddle.net/bkf50r8v/13/, ensure 3 blocked items are listed in shields. Result window should show `got canvas fingerprint 0` and `got webgl fingerprint 00` - [ ] Test that audio fingerprint is blocked at https://audiofingerprint.openwpm.com/ only when `Block all fingerprinting protection` is on - [ ] Test that Brave browser isn't detected on https://extensions.inrialpes.fr/brave/ - [ ] Test that https://diafygi.github.io/webrtc-ips/ doesn't leak IP address when `Block all fingerprinting protection` is on ### Rewards - [ ] Verify wallet is auto created after enabling rewards - [ ] Verify account balance shows correct BAT and USD value - [ ] Verify you are able to restore a wallet - [ ] Verify wallet address matches the QR code that is generated under `Add funds` - [ ] Verify actions taken (claiming grant, tipping, auto-contribute) display in wallet panel - [ ] Verify adding funds via any of the currencies flows into wallet after specified amount of time - [ ] Verify adding funds to an existing wallet with amount, adjusts the BAT value appropriately - [ ] Verify monthly budget shows correct BAT and USD value - [ ] Verify you are able to exclude a publisher from the auto-contribute table by clicking on the `x` in auto-contribute table and popup list of sites - [ ] Verify you are able to exclude a publisher by using the toggle on the Rewards Panel - [ ] Verify when you click on the BR panel while on a site, the panel displays site specific information (site favicon, domain, attention %) - [ ] Verify when you click on `Send a tip`, the custom tip banner displays - [ ] Verify you are able to make one-time tip and they display in tips panel - [ ] Verify you are able to make recurring tip and they display in tips panel - [ ] Verify you can tip a verified publisher - [ ] Verify you can tip a verified YouTube creator - [ ] Verify tip panel shows a verified checkmark for a verified publisher/verified YouTube creator - [ ] Verify tip panel shows a message about unverified publisher - [ ] Verify BR panel shows message about an unverified publisher - [ ] Verify you are able to perform a contribution - [ ] Verify if you disable auto-contribute you are still able to tip regular sites and YouTube creators - [ ] Verify that disabling Rewards and enabling it again does not lose state - [ ] Verify that disabling auto-contribute and enabling it again does not lose state - [ ] Adjust min visit/time in settings. Visit some sites and YouTube channels to verify they are added to the table after the specified settings - [ ] Upgrade from older version - [ ] Verify the wallet balance is retained and wallet backup code isn't corrupted - [ ] Verify auto-contribute list is not lost after upgrade - [ ] Verify tips list is not lost after upgrade - [ ] Verify wallet panel transactions list is not lost after upgrade ### Tor Tabs - [ ] Visit https://check.torproject.org in a Tor window, ensure its shows success message for using a Tor exit node - [ ] Visit https://check.torproject.org in a Tor window, note down exit node IP address. Do a hard refresh (Ctrl+Shift+R/Cmd+Shift+R), ensure exit IP changes after page reloads - [ ] Visit https://protonirockerxow.onion/ in a Tor window, ensure login page is shown - [ ] Visit https://browserleaks.com/geo in a Tor window, ensure location isn't shown ### Session storage - [ ] Temporarily move away your browser profile and test that a new profile is created when browser is launched - macOS - `~/Library/Application\ Support/BraveSoftware/` - Windows - `%userprofile%\appdata\Local\BraveSoftware\` - Linux(Ubuntu) - `~/.config/BraveSoftware/` - [ ] Test that windows and tabs restore when closed, including active tab - [ ] Ensure that the tabs in the above session are being lazy loaded when the session is restored ## Update tests - [ ] Verify visiting `brave://settings/help` triggers update check - [ ] Verify once update is downloaded, prompts to `Relaunch` to install update ## Chromium upgrade tests - [ ] Verify `brave://gpu` on Brave and `chrome://gpu` on Chrome are similar for the same Chromium version on both browsers #### Adblock - [ ] Verify referrer blocking works properly for TLD+1. Visit `https://technology.slashdot.org/` and verify adblock works properly similar to `https://slashdot.org/` #### Components - [ ] Delete Adblock folder from browser profile and restart browser. Visit `brave://components` and verify `Brave Ad Block Updater` downloads and update the component. Repeat for all Brave components
test
manual test run on linux for x release candidate installer check signature if os run spctl assess verbose applications brave browser beta app and make sure it returns accepted if windows right click on the brave installer exe and go to properties go to the digital signatures tab and double click on the signature make sure it says the digital signature is ok in the popup window data upgrade from previous release make sure that data from the last version appears in the new version ok with data from the last version verify that bookmarks on the bookmark toolbar and bookmark folders can be opened cookies are preserved installed extensions are retained and work correctly opened tabs can be reloaded stored passwords are preserved unpinned tabs can be pinned extensions plugins tests verify one item from brave update server is installable example ad block dat file on fresh extension verify one item from google update server is installable example extensions from cws verify pdfjs torrent viewer extensions are installed automatically on fresh profile and cannot be disabled verify older version of an extension gets updated to new version via google server verify older version of an extension gets updated to new version via brave server verify magnet links and torrent files loads torrent viewer page and able to download torrent cws verify installing abp from cws shows warning message not a recommended brave extension but still allows to install the extension verify installing lastpass from cws doesn t show any warning message verify installing an extension that is not vetted by brave gets blocked pdf test that pdf is loaded over https at test that pdf is loaded over http at widevine verify widevine notification is shown when you visit netflix for the first time test that you can stream on netflix on a fresh profile after installing widevine bravery settings verify that https everywhere works by loading turning https everywhere off and shields off both disable the redirect to verify that toggling ads and trackers blocked works as expected visit verify that safe browsing via our proxy works for all the listed items visit and then turn on script blocking page should not load allow it from the script blocking ui in the url bar and it should load the page correctly test that party storage results are blank at when party cookies are blocked and not blank when party cookies are unblocked fingerprint tests visit ensure blocked items are listed in shields result window should show got canvas fingerprint and got webgl fingerprint test that audio fingerprint is blocked at only when block all fingerprinting protection is on test that brave browser isn t detected on test that doesn t leak ip address when block all fingerprinting protection is on rewards verify wallet is auto created after enabling rewards verify account balance shows correct bat and usd value verify you are able to restore a wallet verify wallet address matches the qr code that is generated under add funds verify actions taken claiming grant tipping auto contribute display in wallet panel verify adding funds via any of the currencies flows into wallet after specified amount of time verify adding funds to an existing wallet with amount adjusts the bat value appropriately verify monthly budget shows correct bat and usd value verify you are able to exclude a publisher from the auto contribute table by clicking on the x in auto contribute table and popup list of sites verify you are able to exclude a publisher by using the toggle on the rewards panel verify when you click on the br panel while on a site the panel displays site specific information site favicon domain attention verify when you click on send a tip the custom tip banner displays verify you are able to make one time tip and they display in tips panel verify you are able to make recurring tip and they display in tips panel verify you can tip a verified publisher verify you can tip a verified youtube creator verify tip panel shows a verified checkmark for a verified publisher verified youtube creator verify tip panel shows a message about unverified publisher verify br panel shows message about an unverified publisher verify you are able to perform a contribution verify if you disable auto contribute you are still able to tip regular sites and youtube creators verify that disabling rewards and enabling it again does not lose state verify that disabling auto contribute and enabling it again does not lose state adjust min visit time in settings visit some sites and youtube channels to verify they are added to the table after the specified settings upgrade from older version verify the wallet balance is retained and wallet backup code isn t corrupted verify auto contribute list is not lost after upgrade verify tips list is not lost after upgrade verify wallet panel transactions list is not lost after upgrade tor tabs visit in a tor window ensure its shows success message for using a tor exit node visit in a tor window note down exit node ip address do a hard refresh ctrl shift r cmd shift r ensure exit ip changes after page reloads visit in a tor window ensure login page is shown visit in a tor window ensure location isn t shown session storage temporarily move away your browser profile and test that a new profile is created when browser is launched macos library application support bravesoftware windows userprofile appdata local bravesoftware linux ubuntu config bravesoftware test that windows and tabs restore when closed including active tab ensure that the tabs in the above session are being lazy loaded when the session is restored update tests verify visiting brave settings help triggers update check verify once update is downloaded prompts to relaunch to install update chromium upgrade tests verify brave gpu on brave and chrome gpu on chrome are similar for the same chromium version on both browsers adblock verify referrer blocking works properly for tld visit and verify adblock works properly similar to components delete adblock folder from browser profile and restart browser visit brave components and verify brave ad block updater downloads and update the component repeat for all brave components
1
732,356
25,256,753,384
IssuesEvent
2022-11-15 18:47:38
kytos-ng/coloring
https://api.github.com/repos/kytos-ng/coloring
closed
coloring should use a cookie_prefix like mef_eline and of_lldp
enhancement priority_medium 2022.3
Coloring push its own FlowMods, we have a cookie_prefix pattern similarly to how mef_eline and of_lldp https://github.com/kytos-ng/kytos-ng.github.io/issues/13, this helps to identify ownership, plus e2e testability since they can be filtered out accordingly in the output later on. Once the prefix is defined, it should also be documented in the linked issue just so it becomes reserved for this NApp
1.0
coloring should use a cookie_prefix like mef_eline and of_lldp - Coloring push its own FlowMods, we have a cookie_prefix pattern similarly to how mef_eline and of_lldp https://github.com/kytos-ng/kytos-ng.github.io/issues/13, this helps to identify ownership, plus e2e testability since they can be filtered out accordingly in the output later on. Once the prefix is defined, it should also be documented in the linked issue just so it becomes reserved for this NApp
non_test
coloring should use a cookie prefix like mef eline and of lldp coloring push its own flowmods we have a cookie prefix pattern similarly to how mef eline and of lldp this helps to identify ownership plus testability since they can be filtered out accordingly in the output later on once the prefix is defined it should also be documented in the linked issue just so it becomes reserved for this napp
0
195,123
14,704,016,632
IssuesEvent
2021-01-04 15:52:48
github-vet/rangeloop-pointer-findings
https://api.github.com/repos/github-vet/rangeloop-pointer-findings
closed
radondb/radon: src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/client/client_test.go; 8 LoC
fresh test tiny
Found a possible issue in [radondb/radon](https://www.github.com/radondb/radon) at [src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/client/client_test.go](https://github.com/radondb/radon/blob/3885e5acfe135e1b414dabc6c86e78f3bac4b5eb/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/client/client_test.go#L93-L100) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable s used in defer or goroutine at line 97 [Click here to see the code in its original context.](https://github.com/radondb/radon/blob/3885e5acfe135e1b414dabc6c86e78f3bac4b5eb/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/client/client_test.go#L93-L100) <details> <summary>Click here to show the 8 line(s) of Go which triggered the analyzer.</summary> ```go for _, s := range cs { wg.Add(1) go func() { defer wg.Done() err := s.c.Ping() assert.Nil(t, err) }() } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 3885e5acfe135e1b414dabc6c86e78f3bac4b5eb
1.0
radondb/radon: src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/client/client_test.go; 8 LoC - Found a possible issue in [radondb/radon](https://www.github.com/radondb/radon) at [src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/client/client_test.go](https://github.com/radondb/radon/blob/3885e5acfe135e1b414dabc6c86e78f3bac4b5eb/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/client/client_test.go#L93-L100) Below is the message reported by the analyzer for this snippet of code. Beware that the analyzer only reports the first issue it finds, so please do not limit your consideration to the contents of the below message. > range-loop variable s used in defer or goroutine at line 97 [Click here to see the code in its original context.](https://github.com/radondb/radon/blob/3885e5acfe135e1b414dabc6c86e78f3bac4b5eb/src/vendor/github.com/radondb/shift/vendor/github.com/siddontang/go-mysql/client/client_test.go#L93-L100) <details> <summary>Click here to show the 8 line(s) of Go which triggered the analyzer.</summary> ```go for _, s := range cs { wg.Add(1) go func() { defer wg.Done() err := s.c.Ping() assert.Nil(t, err) }() } ``` </details> Leave a reaction on this issue to contribute to the project by classifying this instance as a **Bug** :-1:, **Mitigated** :+1:, or **Desirable Behavior** :rocket: See the descriptions of the classifications [here](https://github.com/github-vet/rangeclosure-findings#how-can-i-help) for more information. commit ID: 3885e5acfe135e1b414dabc6c86e78f3bac4b5eb
test
radondb radon src vendor github com radondb shift vendor github com siddontang go mysql client client test go loc found a possible issue in at below is the message reported by the analyzer for this snippet of code beware that the analyzer only reports the first issue it finds so please do not limit your consideration to the contents of the below message range loop variable s used in defer or goroutine at line click here to show the line s of go which triggered the analyzer go for s range cs wg add go func defer wg done err s c ping assert nil t err leave a reaction on this issue to contribute to the project by classifying this instance as a bug mitigated or desirable behavior rocket see the descriptions of the classifications for more information commit id
1
746,604
26,037,322,320
IssuesEvent
2022-12-22 06:58:19
evo-lua/evo-luvi
https://api.github.com/repos/evo-lua/evo-luvi
closed
Integrate (semi-)automatic changelog generation for GitHub Releases
Complexity: Moderate Type: Task Status: Needs Review Priority: Normal Scope: GitHub Actions
Goals: - [x] Change logs can be generated from git commits automatically - [x] They are appended to GitHub Releaes automatically (example [here](https://github.com/softprops/action-gh-release#-external-release-notes)) Roadmap: - [x] Implement changelog generation via ``C_BuildTools`` (for now) - [x] API to discover the SECOND latest git version (``git tag --sort=-creatordate | head -2 | tail -1``) - [x] API to discover all PRs between two releases (``git log v3.0.0...v2.18.0 --merges --pretty="format:* \[[`%h`](https://github.com/evo-lua/evo-luvi/commit/%H)] - %b"`` - [x] API to create a Markdown rendition of a given changelog description - [x] API to insert the PRs, contributors, and human-readable notes (optional) - [x] API to create a log for the latest tag and save it to ``CHANGELOG.MD`` - [x] For human-readable summaries, need to store a Lua table/JSON DB somewhere (and update manually) - [x] Add scripts to generate a ``CHANGELOG.MD`` for the current (latest) release tag - [x] Ignore this file via ``.gitignore`` (no need to commit it as it's ephemeral) - [x] Add an extra step to the release workflows that generates a changelog for the last two versions (vOLD..vNEW) - [x] Update all CI release workflows to automatically set this file's contents as the release body (it's idempotent) Formats (TBD): * git commits only (see luvit releases) * https://github.com/nodejs/node/blob/1bbd14eac20684a733250629c7896262c0ca4576/doc/changelogs/CHANGELOG_V19.md * Manually-added notes (see existing changelogger tool)
1.0
Integrate (semi-)automatic changelog generation for GitHub Releases - Goals: - [x] Change logs can be generated from git commits automatically - [x] They are appended to GitHub Releaes automatically (example [here](https://github.com/softprops/action-gh-release#-external-release-notes)) Roadmap: - [x] Implement changelog generation via ``C_BuildTools`` (for now) - [x] API to discover the SECOND latest git version (``git tag --sort=-creatordate | head -2 | tail -1``) - [x] API to discover all PRs between two releases (``git log v3.0.0...v2.18.0 --merges --pretty="format:* \[[`%h`](https://github.com/evo-lua/evo-luvi/commit/%H)] - %b"`` - [x] API to create a Markdown rendition of a given changelog description - [x] API to insert the PRs, contributors, and human-readable notes (optional) - [x] API to create a log for the latest tag and save it to ``CHANGELOG.MD`` - [x] For human-readable summaries, need to store a Lua table/JSON DB somewhere (and update manually) - [x] Add scripts to generate a ``CHANGELOG.MD`` for the current (latest) release tag - [x] Ignore this file via ``.gitignore`` (no need to commit it as it's ephemeral) - [x] Add an extra step to the release workflows that generates a changelog for the last two versions (vOLD..vNEW) - [x] Update all CI release workflows to automatically set this file's contents as the release body (it's idempotent) Formats (TBD): * git commits only (see luvit releases) * https://github.com/nodejs/node/blob/1bbd14eac20684a733250629c7896262c0ca4576/doc/changelogs/CHANGELOG_V19.md * Manually-added notes (see existing changelogger tool)
non_test
integrate semi automatic changelog generation for github releases goals change logs can be generated from git commits automatically they are appended to github releaes automatically example roadmap implement changelog generation via c buildtools for now api to discover the second latest git version git tag sort creatordate head tail api to discover all prs between two releases git log merges pretty format b api to create a markdown rendition of a given changelog description api to insert the prs contributors and human readable notes optional api to create a log for the latest tag and save it to changelog md for human readable summaries need to store a lua table json db somewhere and update manually add scripts to generate a changelog md for the current latest release tag ignore this file via gitignore no need to commit it as it s ephemeral add an extra step to the release workflows that generates a changelog for the last two versions vold vnew update all ci release workflows to automatically set this file s contents as the release body it s idempotent formats tbd git commits only see luvit releases manually added notes see existing changelogger tool
0
47,755
5,914,516,135
IssuesEvent
2017-05-22 03:18:26
arvoelke/nengolib
https://api.github.com/repos/arvoelke/nengolib
closed
test_check_sys_equal fails on scipy==0.19.0
bug testing
``` if np.all(den == 0): > raise ValueError("Denominator must have at least on nonzero element.") E ValueError: Denominator must have at least on nonzero element. ```
1.0
test_check_sys_equal fails on scipy==0.19.0 - ``` if np.all(den == 0): > raise ValueError("Denominator must have at least on nonzero element.") E ValueError: Denominator must have at least on nonzero element. ```
test
test check sys equal fails on scipy if np all den raise valueerror denominator must have at least on nonzero element e valueerror denominator must have at least on nonzero element
1
145,671
22,757,667,907
IssuesEvent
2022-07-07 18:01:53
Opentrons/opentrons
https://api.github.com/repos/Opentrons/opentrons
closed
bug: Run log does not track steps of a JSONv6 protocol
bug protocol designer robot-svcs robot server
## Overview When running a Python Protocol API v3 protocol or a JSON v6 protocol, the Opentrons App's run log does not update itself to show which command is currently running. ## Steps to reproduce Run this JSONv6 protocol: [JSON v6 Test Protocol.json.zip](https://github.com/Opentrons/opentrons/files/8961467/JSON.v6.Test.Protocol.json.zip) Tested on 6.0.0-beta.0 and `edge` (as of 2022-06-22) ## Current behavior The app shows steps from the protocol's analysis, but does not show the steps actually executed in the run. ![Screen Shot 2022-06-16 at 12 17 11 PM](https://user-images.githubusercontent.com/3236864/174121074-25a97d21-3863-4369-a0f8-6a789c8aaaca.png) Pausing, playing, or letting the run finish don't appear to have any effect. ## Expected behavior As with Python Protocol API v2 protocols: * The current step should be highlighted in blue. * The start and finish time of each step should be shown. ## Implemenatation notes - Protocol Designer should attach unique `key` fields to each command - `key` is already present in JSON Protocol Schema v6 - The ProtocolEngine should accept an optional `key` parameter as part of the `CommandCreate` request object - If the `CommandCreate` object has a `key`, it should be used and returned by the HTTP API ## Additional notes This is also true of PAPIv3 protocols, but more-or-less expected because we don't support command keys in native ProtocolEngine commands yet. However, behavior of the run log UI in a PAPIv3 context is still problematic, because the run log does not update even as the `Run` resource tells the app that commands are executing. [Dumps of HTTP responses during a PAPIv3 run](https://gist.github.com/SyntaxColoring/9df466c9bf0135ceabb02496e6b21f57). Note that some commands are missing because this was taken midway through the protocol, when the protocol was paused.
1.0
bug: Run log does not track steps of a JSONv6 protocol - ## Overview When running a Python Protocol API v3 protocol or a JSON v6 protocol, the Opentrons App's run log does not update itself to show which command is currently running. ## Steps to reproduce Run this JSONv6 protocol: [JSON v6 Test Protocol.json.zip](https://github.com/Opentrons/opentrons/files/8961467/JSON.v6.Test.Protocol.json.zip) Tested on 6.0.0-beta.0 and `edge` (as of 2022-06-22) ## Current behavior The app shows steps from the protocol's analysis, but does not show the steps actually executed in the run. ![Screen Shot 2022-06-16 at 12 17 11 PM](https://user-images.githubusercontent.com/3236864/174121074-25a97d21-3863-4369-a0f8-6a789c8aaaca.png) Pausing, playing, or letting the run finish don't appear to have any effect. ## Expected behavior As with Python Protocol API v2 protocols: * The current step should be highlighted in blue. * The start and finish time of each step should be shown. ## Implemenatation notes - Protocol Designer should attach unique `key` fields to each command - `key` is already present in JSON Protocol Schema v6 - The ProtocolEngine should accept an optional `key` parameter as part of the `CommandCreate` request object - If the `CommandCreate` object has a `key`, it should be used and returned by the HTTP API ## Additional notes This is also true of PAPIv3 protocols, but more-or-less expected because we don't support command keys in native ProtocolEngine commands yet. However, behavior of the run log UI in a PAPIv3 context is still problematic, because the run log does not update even as the `Run` resource tells the app that commands are executing. [Dumps of HTTP responses during a PAPIv3 run](https://gist.github.com/SyntaxColoring/9df466c9bf0135ceabb02496e6b21f57). Note that some commands are missing because this was taken midway through the protocol, when the protocol was paused.
non_test
bug run log does not track steps of a protocol overview when running a python protocol api protocol or a json protocol the opentrons app s run log does not update itself to show which command is currently running steps to reproduce run this protocol tested on beta and edge as of current behavior the app shows steps from the protocol s analysis but does not show the steps actually executed in the run pausing playing or letting the run finish don t appear to have any effect expected behavior as with python protocol api protocols the current step should be highlighted in blue the start and finish time of each step should be shown implemenatation notes protocol designer should attach unique key fields to each command key is already present in json protocol schema the protocolengine should accept an optional key parameter as part of the commandcreate request object if the commandcreate object has a key it should be used and returned by the http api additional notes this is also true of protocols but more or less expected because we don t support command keys in native protocolengine commands yet however behavior of the run log ui in a context is still problematic because the run log does not update even as the run resource tells the app that commands are executing note that some commands are missing because this was taken midway through the protocol when the protocol was paused
0
218,672
24,390,301,196
IssuesEvent
2022-10-04 14:46:12
MidnightBSD/security-advisory
https://api.github.com/repos/MidnightBSD/security-advisory
closed
CVE-2022-25857 (High) detected in snakeyaml-1.28.jar - autoclosed
security vulnerability
## CVE-2022-25857 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.28.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-2.5.14.jar (Root Library) - :x: **snakeyaml-1.28.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MidnightBSD/security-advisory/commit/ddbf982f54a01dcec86cab13425f8047dcb250f3">ddbf982f54a01dcec86cab13425f8047dcb250f3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package org.yaml:snakeyaml from 0 and before 1.31 are vulnerable to Denial of Service (DoS) due missing to nested depth limitation for collections. <p>Publish Date: 2022-08-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25857>CVE-2022-25857</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25857">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25857</a></p> <p>Release Date: 2022-08-30</p> <p>Fix Resolution (org.yaml:snakeyaml): 1.31</p> <p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter): 2.6.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
True
CVE-2022-25857 (High) detected in snakeyaml-1.28.jar - autoclosed - ## CVE-2022-25857 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>snakeyaml-1.28.jar</b></p></summary> <p>YAML 1.1 parser and emitter for Java</p> <p>Library home page: <a href="http://www.snakeyaml.org">http://www.snakeyaml.org</a></p> <p>Path to dependency file: /pom.xml</p> <p>Path to vulnerable library: /home/wss-scanner/.m2/repository/org/yaml/snakeyaml/1.28/snakeyaml-1.28.jar</p> <p> Dependency Hierarchy: - spring-boot-starter-2.5.14.jar (Root Library) - :x: **snakeyaml-1.28.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://github.com/MidnightBSD/security-advisory/commit/ddbf982f54a01dcec86cab13425f8047dcb250f3">ddbf982f54a01dcec86cab13425f8047dcb250f3</a></p> <p>Found in base branch: <b>master</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> The package org.yaml:snakeyaml from 0 and before 1.31 are vulnerable to Denial of Service (DoS) due missing to nested depth limitation for collections. <p>Publish Date: 2022-08-30 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2022-25857>CVE-2022-25857</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.5</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Network - Attack Complexity: Low - Privileges Required: None - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: None - Integrity Impact: None - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25857">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2022-25857</a></p> <p>Release Date: 2022-08-30</p> <p>Fix Resolution (org.yaml:snakeyaml): 1.31</p> <p>Direct dependency fix Resolution (org.springframework.boot:spring-boot-starter): 2.6.9</p> </p> </details> <p></p> *** Step up your Open Source Security Game with Mend [here](https://www.whitesourcesoftware.com/full_solution_bolt_github)
non_test
cve high detected in snakeyaml jar autoclosed cve high severity vulnerability vulnerable library snakeyaml jar yaml parser and emitter for java library home page a href path to dependency file pom xml path to vulnerable library home wss scanner repository org yaml snakeyaml snakeyaml jar dependency hierarchy spring boot starter jar root library x snakeyaml jar vulnerable library found in head commit a href found in base branch master vulnerability details the package org yaml snakeyaml from and before are vulnerable to denial of service dos due missing to nested depth limitation for collections publish date url a href cvss score details base score metrics exploitability metrics attack vector network attack complexity low privileges required none user interaction none scope unchanged impact metrics confidentiality impact none integrity impact none availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org yaml snakeyaml direct dependency fix resolution org springframework boot spring boot starter step up your open source security game with mend
0
153,346
12,141,151,566
IssuesEvent
2020-04-23 21:52:52
rancher/rancher
https://api.github.com/repos/rancher/rancher
closed
Support for Custom Image Name or ID on Azure Node Templates
[zube]: To Test internal kind/feature
Is there a workaround to beb able to use Custom Images names or IDs for Azure Node Templates? Currently Rancher Node Template Image field only support Image with URN format publisher:offer:sku:version Best regards, Edgar. --- | Useful | Info | | :-- | :-- | |Versions|Rancher `v2.2.7` UI: `v2.2.93` | |Route|`global-admin.clusters.index`|
1.0
Support for Custom Image Name or ID on Azure Node Templates - Is there a workaround to beb able to use Custom Images names or IDs for Azure Node Templates? Currently Rancher Node Template Image field only support Image with URN format publisher:offer:sku:version Best regards, Edgar. --- | Useful | Info | | :-- | :-- | |Versions|Rancher `v2.2.7` UI: `v2.2.93` | |Route|`global-admin.clusters.index`|
test
support for custom image name or id on azure node templates is there a workaround to beb able to use custom images names or ids for azure node templates currently rancher node template image field only support image with urn format publisher offer sku version best regards edgar useful info versions rancher ui route global admin clusters index
1
135,153
19,519,039,486
IssuesEvent
2021-12-29 15:08:12
department-of-veterans-affairs/vets-design-system-documentation
https://api.github.com/repos/department-of-veterans-affairs/vets-design-system-documentation
opened
Audit current use of Tabs and create Sketch symbols for VA Pattern Library
vsp-design-system-team
# Feature Request - [x] I’ve searched for any related issues and avoided creating a duplicate issue. ## This update is for: - [ ] Content styleguide - [x] Component - [ ] Pattern - [ ] Utility - [ ] Other ## What is the name Tabs ## What is the nature of this update? <!-- Please provide all that apply. Not all are required. --> - [x] How to build this component/pattern - [ ] When to use this component/pattern - [ ] When to use something else - [ ] Usability guidance - [ ] Accessibility - [ ] Implementation - [ ] Research insights - [ ] Package information ## Additional Context There aren't too many examples of tabs in use currently, but we should do an audit and make Sketch symbols of the optimal tabs design.
1.0
Audit current use of Tabs and create Sketch symbols for VA Pattern Library - # Feature Request - [x] I’ve searched for any related issues and avoided creating a duplicate issue. ## This update is for: - [ ] Content styleguide - [x] Component - [ ] Pattern - [ ] Utility - [ ] Other ## What is the name Tabs ## What is the nature of this update? <!-- Please provide all that apply. Not all are required. --> - [x] How to build this component/pattern - [ ] When to use this component/pattern - [ ] When to use something else - [ ] Usability guidance - [ ] Accessibility - [ ] Implementation - [ ] Research insights - [ ] Package information ## Additional Context There aren't too many examples of tabs in use currently, but we should do an audit and make Sketch symbols of the optimal tabs design.
non_test
audit current use of tabs and create sketch symbols for va pattern library feature request i’ve searched for any related issues and avoided creating a duplicate issue this update is for content styleguide component pattern utility other what is the name tabs what is the nature of this update please provide all that apply not all are required how to build this component pattern when to use this component pattern when to use something else usability guidance accessibility implementation research insights package information additional context there aren t too many examples of tabs in use currently but we should do an audit and make sketch symbols of the optimal tabs design
0
317,484
9,665,947,998
IssuesEvent
2019-05-21 09:39:34
conan-io/conan
https://api.github.com/repos/conan-io/conan
closed
Retry upload for error '500' if the user is requesting it (concurrent uploading)
complex: low component: artifactory component: bintray priority: medium stage: review type: feature
To help us debug your issue please explain: - [ x] I've read the [CONTRIBUTING guide](https://github.com/conan-io/conan/blob/develop/.github/CONTRIBUTING.md). - [ x] I've specified the Conan version, operating system version and any tool that can be relevant. - [ x] I've explained the steps to reproduce the error or the motivation/use case of the question/suggestion. I try to upload a package up to our Artifactory server. We are building for multiple compiler on multiple slaves. Sometimes it happens, that they upload nearly at the same time slot. I thought, that with "--force --retry 5 --retry-wait 10" it would try again to upload again, when this is happening. Unluckily it stops with the first exception happening. [hil_drv_cifx] $ cmd.exe /C "conan upload "*" --all --force --retry 5 --retry-wait 10 -r 9d380644-252b-44ff-b4af-feded185890d --confirm && exit %%ERRORLEVEL%%" ERROR: Error uploading file: conanmanifest.txt, 'b'{\r\n "errors" : [ {\r\n "status" : 500,\r\n "message" : "Failed to persist file with sha1: 3093033974f36098596e33acf477de5a9e8ddd87"\r\n } ]\r\n}'' I am using Conan 1.13.1 on multiple Windows & Linux Jenkins Slaves (all python3.6). We are using the Artifactory plugin for Jenkins (newest version, just some days old)
1.0
Retry upload for error '500' if the user is requesting it (concurrent uploading) - To help us debug your issue please explain: - [ x] I've read the [CONTRIBUTING guide](https://github.com/conan-io/conan/blob/develop/.github/CONTRIBUTING.md). - [ x] I've specified the Conan version, operating system version and any tool that can be relevant. - [ x] I've explained the steps to reproduce the error or the motivation/use case of the question/suggestion. I try to upload a package up to our Artifactory server. We are building for multiple compiler on multiple slaves. Sometimes it happens, that they upload nearly at the same time slot. I thought, that with "--force --retry 5 --retry-wait 10" it would try again to upload again, when this is happening. Unluckily it stops with the first exception happening. [hil_drv_cifx] $ cmd.exe /C "conan upload "*" --all --force --retry 5 --retry-wait 10 -r 9d380644-252b-44ff-b4af-feded185890d --confirm && exit %%ERRORLEVEL%%" ERROR: Error uploading file: conanmanifest.txt, 'b'{\r\n "errors" : [ {\r\n "status" : 500,\r\n "message" : "Failed to persist file with sha1: 3093033974f36098596e33acf477de5a9e8ddd87"\r\n } ]\r\n}'' I am using Conan 1.13.1 on multiple Windows & Linux Jenkins Slaves (all python3.6). We are using the Artifactory plugin for Jenkins (newest version, just some days old)
non_test
retry upload for error if the user is requesting it concurrent uploading to help us debug your issue please explain i ve read the i ve specified the conan version operating system version and any tool that can be relevant i ve explained the steps to reproduce the error or the motivation use case of the question suggestion i try to upload a package up to our artifactory server we are building for multiple compiler on multiple slaves sometimes it happens that they upload nearly at the same time slot i thought that with force retry retry wait it would try again to upload again when this is happening unluckily it stops with the first exception happening cmd exe c conan upload all force retry retry wait r confirm exit errorlevel error error uploading file conanmanifest txt b r n errors r n i am using conan on multiple windows linux jenkins slaves all we are using the artifactory plugin for jenkins newest version just some days old
0
194,944
14,693,566,466
IssuesEvent
2021-01-03 09:42:50
blockstack/stacks-blockchain
https://api.github.com/repos/blockstack/stacks-blockchain
closed
Process abort due to thread panic
bug testnet
I'm trying to start stacks-node with custom Xenon configuration. I have downloaded Xenon release from github. When I try to run `stacks-node start --conf=~/.stacks-node/testnet-miner-conf.toml` command I immediately receive the error below: ``` Process abort due to thread panic 0: stacks_node::main::{{closure}} at /src/testnet/stacks-node/src/main.rs:57 1: std::panicking::rust_panic_with_hook at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:597 2: std::panicking::begin_panic_handler::{{closure}} at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:499 3: std::sys_common::backtrace::__rust_end_short_backtrace at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/sys_common/backtrace.rs:141 4: rust_begin_unwind at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:495 5: core::panicking::panic_fmt at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/panicking.rs:92 6: core::option::expect_none_failed at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/option.rs:1268 7: core::result::Result<T,E>::unwrap at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/result.rs:973 stacks_node::main at /src/testnet/stacks-node/src/main.rs:110 8: core::ops::function::FnOnce::call_once at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/ops/function.rs:227 std::sys_common::backtrace::__rust_begin_short_backtrace at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/sys_common/backtrace.rs:125 9: std::rt::lang_start::{{closure}} at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/rt.rs:66 10: core::ops::function::impls::<impl core::ops::function::FnOnce<A> for &F>::call_once at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/ops/function.rs:259 std::panicking::try::do_call at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:381 std::panicking::try at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:345 std::panic::catch_unwind at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panic.rs:396 std::rt::lang_start_internal at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/rt.rs:51 11: main 12: __libc_start_main 13: _start Dumping core for pid 144531 Quit ``` My Bitcoin testnet node config: ``` server=1 daemon=1 rpcuser=my_username rpcpassword=my_pass testnet=1 txindex=0 listen=1 rpcserialversion=0 maxorphantx=1 banscore=1 [test] bind=0.0.0.0:18333 rpcbind=0.0.0.0:18332 rpcport=18332 ``` My stacks-node configuration: ```[node] rpc_bind = "0.0.0.0:20443" p2p_bind = "0.0.0.0:20444" seed = "my_private_key" local_peer_seed = "my_private_key" miner = true bootstrap_node = "047435c194e9b01b3d7f7a2802d6684a3af68d05bbf4ec8f17021980d777691f1d51651f7f1d566532c804da506c117bbf79ad62eea81213ba58f8808b4d9504ad@xenon.blockstack.org:20444" working_dir = "~/.stacks-node" [burnchain] chain = "bitcoin" mode = "xenon" peer_host = "127.0.0.1" rpc_port = 18332 peer_port = 18333 username = "my_bitcoin_rpc_username" password = "my_bitcoin_rpc_password" [[ustx_balance]] address = "STB44HYPYAT2BB2QE513NSP81HTMYWBJP02HPGK6" amount = 10000000000000000 [[ustx_balance]] address = "ST11NJTTKGVT6D1HY4NJRVQWMQM7TVAR091EJ8P2Y" amount = 10000000000000000 [[ustx_balance]] address = "ST1HB1T8WRNBYB0Y3T7WXZS38NKKPTBR3EG9EPJKR" amount = 10000000000000000 [[ustx_balance]] address = "STRYYQQ9M8KAF4NS7WNZQYY59X93XEKR31JP64CP" amount = 10000000000000000 ``` I'm running Ubuntu 20.04.1 with 8GB RAM and 4 cpus Any suggestions? What am I doing wrong?
1.0
Process abort due to thread panic - I'm trying to start stacks-node with custom Xenon configuration. I have downloaded Xenon release from github. When I try to run `stacks-node start --conf=~/.stacks-node/testnet-miner-conf.toml` command I immediately receive the error below: ``` Process abort due to thread panic 0: stacks_node::main::{{closure}} at /src/testnet/stacks-node/src/main.rs:57 1: std::panicking::rust_panic_with_hook at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:597 2: std::panicking::begin_panic_handler::{{closure}} at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:499 3: std::sys_common::backtrace::__rust_end_short_backtrace at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/sys_common/backtrace.rs:141 4: rust_begin_unwind at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:495 5: core::panicking::panic_fmt at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/panicking.rs:92 6: core::option::expect_none_failed at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/option.rs:1268 7: core::result::Result<T,E>::unwrap at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/result.rs:973 stacks_node::main at /src/testnet/stacks-node/src/main.rs:110 8: core::ops::function::FnOnce::call_once at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/ops/function.rs:227 std::sys_common::backtrace::__rust_begin_short_backtrace at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/sys_common/backtrace.rs:125 9: std::rt::lang_start::{{closure}} at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/rt.rs:66 10: core::ops::function::impls::<impl core::ops::function::FnOnce<A> for &F>::call_once at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/core/src/ops/function.rs:259 std::panicking::try::do_call at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:381 std::panicking::try at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panicking.rs:345 std::panic::catch_unwind at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/panic.rs:396 std::rt::lang_start_internal at /rustc/e1884a8e3c3e813aada8254edfa120e85bf5ffca/library/std/src/rt.rs:51 11: main 12: __libc_start_main 13: _start Dumping core for pid 144531 Quit ``` My Bitcoin testnet node config: ``` server=1 daemon=1 rpcuser=my_username rpcpassword=my_pass testnet=1 txindex=0 listen=1 rpcserialversion=0 maxorphantx=1 banscore=1 [test] bind=0.0.0.0:18333 rpcbind=0.0.0.0:18332 rpcport=18332 ``` My stacks-node configuration: ```[node] rpc_bind = "0.0.0.0:20443" p2p_bind = "0.0.0.0:20444" seed = "my_private_key" local_peer_seed = "my_private_key" miner = true bootstrap_node = "047435c194e9b01b3d7f7a2802d6684a3af68d05bbf4ec8f17021980d777691f1d51651f7f1d566532c804da506c117bbf79ad62eea81213ba58f8808b4d9504ad@xenon.blockstack.org:20444" working_dir = "~/.stacks-node" [burnchain] chain = "bitcoin" mode = "xenon" peer_host = "127.0.0.1" rpc_port = 18332 peer_port = 18333 username = "my_bitcoin_rpc_username" password = "my_bitcoin_rpc_password" [[ustx_balance]] address = "STB44HYPYAT2BB2QE513NSP81HTMYWBJP02HPGK6" amount = 10000000000000000 [[ustx_balance]] address = "ST11NJTTKGVT6D1HY4NJRVQWMQM7TVAR091EJ8P2Y" amount = 10000000000000000 [[ustx_balance]] address = "ST1HB1T8WRNBYB0Y3T7WXZS38NKKPTBR3EG9EPJKR" amount = 10000000000000000 [[ustx_balance]] address = "STRYYQQ9M8KAF4NS7WNZQYY59X93XEKR31JP64CP" amount = 10000000000000000 ``` I'm running Ubuntu 20.04.1 with 8GB RAM and 4 cpus Any suggestions? What am I doing wrong?
test
process abort due to thread panic i m trying to start stacks node with custom xenon configuration i have downloaded xenon release from github when i try to run stacks node start conf stacks node testnet miner conf toml command i immediately receive the error below process abort due to thread panic stacks node main closure at src testnet stacks node src main rs std panicking rust panic with hook at rustc library std src panicking rs std panicking begin panic handler closure at rustc library std src panicking rs std sys common backtrace rust end short backtrace at rustc library std src sys common backtrace rs rust begin unwind at rustc library std src panicking rs core panicking panic fmt at rustc library core src panicking rs core option expect none failed at rustc library core src option rs core result result unwrap at rustc library core src result rs stacks node main at src testnet stacks node src main rs core ops function fnonce call once at rustc library core src ops function rs std sys common backtrace rust begin short backtrace at rustc library std src sys common backtrace rs std rt lang start closure at rustc library std src rt rs core ops function impls for f call once at rustc library core src ops function rs std panicking try do call at rustc library std src panicking rs std panicking try at rustc library std src panicking rs std panic catch unwind at rustc library std src panic rs std rt lang start internal at rustc library std src rt rs main libc start main start dumping core for pid quit my bitcoin testnet node config server daemon rpcuser my username rpcpassword my pass testnet txindex listen rpcserialversion maxorphantx banscore bind rpcbind rpcport my stacks node configuration rpc bind bind seed my private key local peer seed my private key miner true bootstrap node xenon blockstack org working dir stacks node chain bitcoin mode xenon peer host rpc port peer port username my bitcoin rpc username password my bitcoin rpc password address amount address amount address amount address amount i m running ubuntu with ram and cpus any suggestions what am i doing wrong
1
295,843
9,101,449,522
IssuesEvent
2019-02-20 11:07:32
jenkins-x/jx
https://api.github.com/repos/jenkins-x/jx
closed
nexus password not updated when running jx install on existing installation
area/install kind/bug priority/important-soon
### Summary Nexus admin password does not get updated after running "jx install" again after the initial install ### Steps to reproduce the behavior jx install ### Jx version 1.3.488 The output of `jx version` is: ``` NAME VERSION jx 1.3.488 jenkins x platform 0.0.2806 Kubernetes cluster v1.11.3 kubectl v1.12.1 helm client v2.11.0+g2e55dbe helm server v2.11.0+g2e55dbe git git version 1.8.3.1 ``` ### Kubernetes cluster What kind of Kubernetes cluster are you using & how did you create it? ### Operating system / Environment ### Expected behavior nexus password is updated to the one that is displayed after jx installs successfully ### Actual behavior it's using the old password from the initial install of jx
1.0
nexus password not updated when running jx install on existing installation - ### Summary Nexus admin password does not get updated after running "jx install" again after the initial install ### Steps to reproduce the behavior jx install ### Jx version 1.3.488 The output of `jx version` is: ``` NAME VERSION jx 1.3.488 jenkins x platform 0.0.2806 Kubernetes cluster v1.11.3 kubectl v1.12.1 helm client v2.11.0+g2e55dbe helm server v2.11.0+g2e55dbe git git version 1.8.3.1 ``` ### Kubernetes cluster What kind of Kubernetes cluster are you using & how did you create it? ### Operating system / Environment ### Expected behavior nexus password is updated to the one that is displayed after jx installs successfully ### Actual behavior it's using the old password from the initial install of jx
non_test
nexus password not updated when running jx install on existing installation summary nexus admin password does not get updated after running jx install again after the initial install steps to reproduce the behavior jx install jx version the output of jx version is name version jx jenkins x platform kubernetes cluster kubectl helm client helm server git git version kubernetes cluster what kind of kubernetes cluster are you using how did you create it operating system environment expected behavior nexus password is updated to the one that is displayed after jx installs successfully actual behavior it s using the old password from the initial install of jx
0
109,772
16,891,982,962
IssuesEvent
2021-06-23 10:20:42
epam/TimeBase
https://api.github.com/repos/epam/TimeBase
closed
CVE-2016-2175 (High) detected in pdfbox-1.8.10.jar - autoclosed
security vulnerability
## CVE-2016-2175 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>pdfbox-1.8.10.jar</b></p></summary> <p>The Apache PDFBox library is an open source Java tool for working with PDF documents.</p> <p>Path to dependency file: TimeBase/java/installer/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.pdfbox/pdfbox/1.8.10/bc5d1254495be36d0a3b3d6c35f88d05200b9311/pdfbox-1.8.10.jar</p> <p> Dependency Hierarchy: - izpack-compiler-5.1.3.jar (Root Library) - tika-parsers-1.9.jar - :x: **pdfbox-1.8.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/epam/TimeBase/commits/98f6880b361c00a247f77e79a787646e9664cadd">98f6880b361c00a247f77e79a787646e9664cadd</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache PDFBox before 1.8.12 and 2.x before 2.0.1 does not properly initialize the XML parsers, which allows context-dependent attackers to conduct XML External Entity (XXE) attacks via a crafted PDF. <p>Publish Date: 2016-06-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2175>CVE-2016-2175</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-2175">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-2175</a></p> <p>Release Date: 2016-06-01</p> <p>Fix Resolution: org.apache.pdfbox:pdfbox:2.0.1,1.8.12,org.apache.pdfbox:pdfbox-app:2.0.1,1.8.12,org.apache.pdfbox:debugger-app:2.0.1,1.8.12,org.apache.pdfbox:preflight-app:2.0.1,1.8.12</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.pdfbox","packageName":"pdfbox","packageVersion":"1.8.10","packageFilePaths":["/java/installer/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.codehaus.izpack:izpack-compiler:5.1.3;org.apache.tika:tika-parsers:1.9;org.apache.pdfbox:pdfbox:1.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.pdfbox:pdfbox:2.0.1,1.8.12,org.apache.pdfbox:pdfbox-app:2.0.1,1.8.12,org.apache.pdfbox:debugger-app:2.0.1,1.8.12,org.apache.pdfbox:preflight-app:2.0.1,1.8.12"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-2175","vulnerabilityDetails":"Apache PDFBox before 1.8.12 and 2.x before 2.0.1 does not properly initialize the XML parsers, which allows context-dependent attackers to conduct XML External Entity (XXE) attacks via a crafted PDF.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2175","cvss3Severity":"high","cvss3Score":"7.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> -->
True
CVE-2016-2175 (High) detected in pdfbox-1.8.10.jar - autoclosed - ## CVE-2016-2175 - High Severity Vulnerability <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/vulnerability_details.png' width=19 height=20> Vulnerable Library - <b>pdfbox-1.8.10.jar</b></p></summary> <p>The Apache PDFBox library is an open source Java tool for working with PDF documents.</p> <p>Path to dependency file: TimeBase/java/installer/build.gradle</p> <p>Path to vulnerable library: /home/wss-scanner/.gradle/caches/modules-2/files-2.1/org.apache.pdfbox/pdfbox/1.8.10/bc5d1254495be36d0a3b3d6c35f88d05200b9311/pdfbox-1.8.10.jar</p> <p> Dependency Hierarchy: - izpack-compiler-5.1.3.jar (Root Library) - tika-parsers-1.9.jar - :x: **pdfbox-1.8.10.jar** (Vulnerable Library) <p>Found in HEAD commit: <a href="https://api.github.com/repos/epam/TimeBase/commits/98f6880b361c00a247f77e79a787646e9664cadd">98f6880b361c00a247f77e79a787646e9664cadd</a></p> <p>Found in base branch: <b>main</b></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/high_vul.png' width=19 height=20> Vulnerability Details</summary> <p> Apache PDFBox before 1.8.12 and 2.x before 2.0.1 does not properly initialize the XML parsers, which allows context-dependent attackers to conduct XML External Entity (XXE) attacks via a crafted PDF. <p>Publish Date: 2016-06-01 <p>URL: <a href=https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2175>CVE-2016-2175</a></p> </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/cvss3.png' width=19 height=20> CVSS 3 Score Details (<b>7.8</b>)</summary> <p> Base Score Metrics: - Exploitability Metrics: - Attack Vector: Local - Attack Complexity: Low - Privileges Required: Low - User Interaction: None - Scope: Unchanged - Impact Metrics: - Confidentiality Impact: High - Integrity Impact: High - Availability Impact: High </p> For more information on CVSS3 Scores, click <a href="https://www.first.org/cvss/calculator/3.0">here</a>. </p> </details> <p></p> <details><summary><img src='https://whitesource-resources.whitesourcesoftware.com/suggested_fix.png' width=19 height=20> Suggested Fix</summary> <p> <p>Type: Upgrade version</p> <p>Origin: <a href="https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-2175">https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2016-2175</a></p> <p>Release Date: 2016-06-01</p> <p>Fix Resolution: org.apache.pdfbox:pdfbox:2.0.1,1.8.12,org.apache.pdfbox:pdfbox-app:2.0.1,1.8.12,org.apache.pdfbox:debugger-app:2.0.1,1.8.12,org.apache.pdfbox:preflight-app:2.0.1,1.8.12</p> </p> </details> <p></p> <!-- <REMEDIATE>{"isOpenPROnVulnerability":false,"isPackageBased":true,"isDefaultBranch":true,"packages":[{"packageType":"Java","groupId":"org.apache.pdfbox","packageName":"pdfbox","packageVersion":"1.8.10","packageFilePaths":["/java/installer/build.gradle"],"isTransitiveDependency":true,"dependencyTree":"org.codehaus.izpack:izpack-compiler:5.1.3;org.apache.tika:tika-parsers:1.9;org.apache.pdfbox:pdfbox:1.8.10","isMinimumFixVersionAvailable":true,"minimumFixVersion":"org.apache.pdfbox:pdfbox:2.0.1,1.8.12,org.apache.pdfbox:pdfbox-app:2.0.1,1.8.12,org.apache.pdfbox:debugger-app:2.0.1,1.8.12,org.apache.pdfbox:preflight-app:2.0.1,1.8.12"}],"baseBranches":["main"],"vulnerabilityIdentifier":"CVE-2016-2175","vulnerabilityDetails":"Apache PDFBox before 1.8.12 and 2.x before 2.0.1 does not properly initialize the XML parsers, which allows context-dependent attackers to conduct XML External Entity (XXE) attacks via a crafted PDF.","vulnerabilityUrl":"https://vuln.whitesourcesoftware.com/vulnerability/CVE-2016-2175","cvss3Severity":"high","cvss3Score":"7.8","cvss3Metrics":{"A":"High","AC":"Low","PR":"Low","S":"Unchanged","C":"High","UI":"None","AV":"Local","I":"High"},"extraData":{}}</REMEDIATE> -->
non_test
cve high detected in pdfbox jar autoclosed cve high severity vulnerability vulnerable library pdfbox jar the apache pdfbox library is an open source java tool for working with pdf documents path to dependency file timebase java installer build gradle path to vulnerable library home wss scanner gradle caches modules files org apache pdfbox pdfbox pdfbox jar dependency hierarchy izpack compiler jar root library tika parsers jar x pdfbox jar vulnerable library found in head commit a href found in base branch main vulnerability details apache pdfbox before and x before does not properly initialize the xml parsers which allows context dependent attackers to conduct xml external entity xxe attacks via a crafted pdf publish date url a href cvss score details base score metrics exploitability metrics attack vector local attack complexity low privileges required low user interaction none scope unchanged impact metrics confidentiality impact high integrity impact high availability impact high for more information on scores click a href suggested fix type upgrade version origin a href release date fix resolution org apache pdfbox pdfbox org apache pdfbox pdfbox app org apache pdfbox debugger app org apache pdfbox preflight app isopenpronvulnerability false ispackagebased true isdefaultbranch true packages istransitivedependency true dependencytree org codehaus izpack izpack compiler org apache tika tika parsers org apache pdfbox pdfbox isminimumfixversionavailable true minimumfixversion org apache pdfbox pdfbox org apache pdfbox pdfbox app org apache pdfbox debugger app org apache pdfbox preflight app basebranches vulnerabilityidentifier cve vulnerabilitydetails apache pdfbox before and x before does not properly initialize the xml parsers which allows context dependent attackers to conduct xml external entity xxe attacks via a crafted pdf vulnerabilityurl
0
18,584
3,697,747,154
IssuesEvent
2016-02-27 21:45:23
ethcore/parity
https://api.github.com/repos/ethcore/parity
opened
Comprehensive tests for blockchain forks.
F4-tests
Especially to check if all caches are updated properly.
1.0
Comprehensive tests for blockchain forks. - Especially to check if all caches are updated properly.
test
comprehensive tests for blockchain forks especially to check if all caches are updated properly
1
246,249
20,830,896,500
IssuesEvent
2022-03-19 12:21:18
ActivityWatch/activitywatch
https://api.github.com/repos/ActivityWatch/activitywatch
opened
Epic: Tracking accuracy
area: watcher area: analysis improves: testing/qa !pinned epic
This issue is meant to track accuracy issues present in ActivityWatch. Inaccuracies can arise in two places: ## Collection Bugs in how events are sent, queued, and merged can affect the data and lead to inconsistencies like unexpected gaps and overlaps. Inaccuracies introduced by bugs during collection are taken very seriously, as they may lead to unrecoverable loss of quality tracking data. However, spurious behavior and difficulty in testing well presents difficulties in ensuring things are bug-free. Some inaccuracies arise from assumptions made during collection, such as setting constants for polling intervals, AFK timeouts, etc. These should be intelligently set to minimize inaccuracies and make a good tradeoff between data detail and space/compute requirements. ### Known issues - There are occurrences where events are duplicated, often then keep getting heartbeats such that there are two events with the same start and data but different durations. See here: https://github.com/ActivityWatch/activitywatch/issues/239#issuecomment-821085836 - May downstream lead to - Sometimes, there are weird occurrences of very long-stretching events passing multiple others, such as in this screenshot: https://github.com/ActivityWatch/activitywatch/issues/602#issuecomment-970240728 - Cause unknown ## Analysis Bugs in transforms and queries may lead to analysis issues. However, these bugs are generally considered less critical due to the underlying data still being correct, such that an analysis can always be corrected later. Transforms and queries are also easy to test, and the dual-implementation of transforms in both Python and Rust make it a suitable target for cross-implementation testing to ensure robust analysis methods. # Known issues - #239 - #602 - #722
1.0
Epic: Tracking accuracy - This issue is meant to track accuracy issues present in ActivityWatch. Inaccuracies can arise in two places: ## Collection Bugs in how events are sent, queued, and merged can affect the data and lead to inconsistencies like unexpected gaps and overlaps. Inaccuracies introduced by bugs during collection are taken very seriously, as they may lead to unrecoverable loss of quality tracking data. However, spurious behavior and difficulty in testing well presents difficulties in ensuring things are bug-free. Some inaccuracies arise from assumptions made during collection, such as setting constants for polling intervals, AFK timeouts, etc. These should be intelligently set to minimize inaccuracies and make a good tradeoff between data detail and space/compute requirements. ### Known issues - There are occurrences where events are duplicated, often then keep getting heartbeats such that there are two events with the same start and data but different durations. See here: https://github.com/ActivityWatch/activitywatch/issues/239#issuecomment-821085836 - May downstream lead to - Sometimes, there are weird occurrences of very long-stretching events passing multiple others, such as in this screenshot: https://github.com/ActivityWatch/activitywatch/issues/602#issuecomment-970240728 - Cause unknown ## Analysis Bugs in transforms and queries may lead to analysis issues. However, these bugs are generally considered less critical due to the underlying data still being correct, such that an analysis can always be corrected later. Transforms and queries are also easy to test, and the dual-implementation of transforms in both Python and Rust make it a suitable target for cross-implementation testing to ensure robust analysis methods. # Known issues - #239 - #602 - #722
test
epic tracking accuracy this issue is meant to track accuracy issues present in activitywatch inaccuracies can arise in two places collection bugs in how events are sent queued and merged can affect the data and lead to inconsistencies like unexpected gaps and overlaps inaccuracies introduced by bugs during collection are taken very seriously as they may lead to unrecoverable loss of quality tracking data however spurious behavior and difficulty in testing well presents difficulties in ensuring things are bug free some inaccuracies arise from assumptions made during collection such as setting constants for polling intervals afk timeouts etc these should be intelligently set to minimize inaccuracies and make a good tradeoff between data detail and space compute requirements known issues there are occurrences where events are duplicated often then keep getting heartbeats such that there are two events with the same start and data but different durations see here may downstream lead to sometimes there are weird occurrences of very long stretching events passing multiple others such as in this screenshot cause unknown analysis bugs in transforms and queries may lead to analysis issues however these bugs are generally considered less critical due to the underlying data still being correct such that an analysis can always be corrected later transforms and queries are also easy to test and the dual implementation of transforms in both python and rust make it a suitable target for cross implementation testing to ensure robust analysis methods known issues
1
189,800
14,523,232,483
IssuesEvent
2020-12-14 09:51:32
microsoft/vscode
https://api.github.com/repos/microsoft/vscode
closed
Flaky test: NativeStorageService Migrate Data
unit-test-failure windows
See https://github.com/microsoft/vscode/runs/1549049837?check_suite_focus=true#step:9:12878 ``` 6391 passing (2m) 42 pending 1 failing 1) NativeStorageService Migrate Data: Error: timeout of 5000ms exceeded. Ensure the done() callback is being called in this test. ```
1.0
Flaky test: NativeStorageService Migrate Data - See https://github.com/microsoft/vscode/runs/1549049837?check_suite_focus=true#step:9:12878 ``` 6391 passing (2m) 42 pending 1 failing 1) NativeStorageService Migrate Data: Error: timeout of 5000ms exceeded. Ensure the done() callback is being called in this test. ```
test
flaky test nativestorageservice migrate data see passing pending failing nativestorageservice migrate data error timeout of exceeded ensure the done callback is being called in this test
1
218,686
17,015,290,474
IssuesEvent
2021-07-02 11:06:45
woo00oo/goGetterProjectServer
https://api.github.com/repos/woo00oo/goGetterProjectServer
opened
Book Report 도메인 계층 재 구현
🏗️ Type : Test 🏷️Sever 👷 Type : Feature / Function 💥 Priority : Critical
Book에 대한 Table을 삭제하고, Book Report에 Book Name의 컬럼을 추가하여 작업하기로 작업사항이 수정되어 이에 대한 Book Report 도메인 계층을 재 구현
1.0
Book Report 도메인 계층 재 구현 - Book에 대한 Table을 삭제하고, Book Report에 Book Name의 컬럼을 추가하여 작업하기로 작업사항이 수정되어 이에 대한 Book Report 도메인 계층을 재 구현
test
book report 도메인 계층 재 구현 book에 대한 table을 삭제하고 book report에 book name의 컬럼을 추가하여 작업하기로 작업사항이 수정되어 이에 대한 book report 도메인 계층을 재 구현
1
102,624
8,851,025,953
IssuesEvent
2019-01-08 14:48:06
NativeScript/nativescript-angular
https://api.github.com/repos/NativeScript/nativescript-angular
closed
Rendering issue when using *ngFor inside a TabView, NS 4.0
backlog bug os: ios ready for test tab-view
**Platform** iOS **Description** Given a template that uses *ngFor inside a tab content: The first rendering works fine, but if the underlying datasource is replaced rendering the items of the new source doesn't seem to happen. **Steps to reproduce** 1. Download the playground app linked below 2. tns run ios 3. click the refresh button and the list is empty Worked fine in NS 3.4 https://play.nativescript.org/?template=play-ng&id=u8JfoL&v=2
1.0
Rendering issue when using *ngFor inside a TabView, NS 4.0 - **Platform** iOS **Description** Given a template that uses *ngFor inside a tab content: The first rendering works fine, but if the underlying datasource is replaced rendering the items of the new source doesn't seem to happen. **Steps to reproduce** 1. Download the playground app linked below 2. tns run ios 3. click the refresh button and the list is empty Worked fine in NS 3.4 https://play.nativescript.org/?template=play-ng&id=u8JfoL&v=2
test
rendering issue when using ngfor inside a tabview ns platform ios description given a template that uses ngfor inside a tab content the first rendering works fine but if the underlying datasource is replaced rendering the items of the new source doesn t seem to happen steps to reproduce download the playground app linked below tns run ios click the refresh button and the list is empty worked fine in ns
1
296,820
22,325,861,272
IssuesEvent
2022-06-14 10:32:31
forem/forem
https://api.github.com/repos/forem/forem
opened
Update deprecated Storybook code
ready for dev area: documentation tech: frontend
In the upgrade to storybook 6.5.x there are a few deprecations that we need to non-urgently address. These can be tackled in separate PRs depending how large the scope is: - [ ] [Importing from '@storybook/addon-docs/blocks' is deprecated, import directly from '@storybook/addon-docs' instead](https://github.com/storybookjs/storybook/blob/next/MIGRATION.md#deprecated-scoped-blocks-imports) - [ ] [CSF .story annotations deprecated; annotate story functions directly](https://github.com/storybookjs/storybook/blob/next/MIGRATION.md#hoisted-csf-annotations) - [ ] ['control.options' is deprecated and will be removed in Storybook 7.0. Define 'options' directly on the argType instead, and use 'control.labels' for custom labels.](https://github.com/storybookjs/storybook/blob/next/MIGRATION.md#deprecated-controloptions) These tasks are taken directly from warning output running `yarn storybook`, and may be added to as further storybook packages are bumped
1.0
Update deprecated Storybook code - In the upgrade to storybook 6.5.x there are a few deprecations that we need to non-urgently address. These can be tackled in separate PRs depending how large the scope is: - [ ] [Importing from '@storybook/addon-docs/blocks' is deprecated, import directly from '@storybook/addon-docs' instead](https://github.com/storybookjs/storybook/blob/next/MIGRATION.md#deprecated-scoped-blocks-imports) - [ ] [CSF .story annotations deprecated; annotate story functions directly](https://github.com/storybookjs/storybook/blob/next/MIGRATION.md#hoisted-csf-annotations) - [ ] ['control.options' is deprecated and will be removed in Storybook 7.0. Define 'options' directly on the argType instead, and use 'control.labels' for custom labels.](https://github.com/storybookjs/storybook/blob/next/MIGRATION.md#deprecated-controloptions) These tasks are taken directly from warning output running `yarn storybook`, and may be added to as further storybook packages are bumped
non_test
update deprecated storybook code in the upgrade to storybook x there are a few deprecations that we need to non urgently address these can be tackled in separate prs depending how large the scope is these tasks are taken directly from warning output running yarn storybook and may be added to as further storybook packages are bumped
0
3,051
3,068,563,636
IssuesEvent
2015-08-18 16:10:58
idiap/bob
https://api.github.com/repos/idiap/bob
closed
bob nosetest fails on Mac OS 10.8.5
Build System Problems
I am running into an error with the ``` make nosetests ``` step of the compiling from source process. I am using a Mac OS Mountain Lion 10.8.5 (64bits) and installing bob with homebrew. The following is installed: Homebrew HOMEBREW_VERSION: 0.9.5 CPU: 8-core 64-bit ivybridge OS X: 10.8.5-x86_64 Xcode: 5.0.2 CLT: 5.0.1.0.1.1377666378 X11: 2.7.5 Python: 2.7.6 Nose: 1.3.0 VLFEAT is not installed and ``` -DWITH_VLFEAT=OFF ``` is specified I wrote a homebrew formula to install bob and it seems to work for now without the nosetests step. [bob.q formula here](https://gist.github.com/tcgeophysics/8156717) To repeat the problem 1 - install homebrew, 2 - Create a bob formula: ```brew create http://www.idiap.ch/software/bob/packages/bob-1.2.2.tar.gz``` 3 - Edit the bob formula: ```brew edit bob``` delete everything and paste the contents of the [bob.q formula here](https://gist.github.com/tcgeophysics/8156717) 4 - Run ```brew install -v bob``` to install in debug mode The install process does not complete with nosetests and exit with the following message: ``` ====================================================================== ERROR: bob.io.test.test_video.test_frameskip_format_mov_codec_wmv1 ---------------------------------------------------------------------- Traceback (most recent call last): File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/nose/case.py", line 197, in runTest self.test(*self.arg) File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/test/utils.py", line 108, in wrapper return test(*args, **kwargs) File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/io/test/test_video.py", line 71, in check_format_codec orig, framerate, encoded = function(shape, framerate, format, codec, fname) File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/io/utils.py", line 145, in frameskip_detection fontsize = estimate_fontsize(height, width, text_format) File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/io/utils.py", line 32, in estimate_fontsize font = ImageFont.truetype(DEFAULT_FONT, best_size) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/PIL/ImageFont.py", line 218, in truetype return FreeTypeFont(filename, size, index, encoding) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/PIL/ImageFont.py", line 134, in __init__ self.font = core.getfont(file, size, index, encoding) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/PIL/ImageFont.py", line 34, in __getattr__ raise ImportError("The _imagingft C module is not installed") ImportError: The _imagingft C module is not installed ====================================================================== FAIL: test01_plda_EM_vs_Python (bob.trainer.test.test_plda.PLDATrainerTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/trainer/test/test_plda.py", line 391, in test01_plda_EM_vs_Python self.assertTrue(numpy.allclose(m.f, m_py.f)) AssertionError: False is not true -------------------- >> begin captured logging << -------------------- bob.c++: INFO: # EMTrainer: bob.c++: INFO: # Iteration 1 bob.c++: INFO: # Iteration 2 bob.c++: INFO: # Iteration 3 bob.c++: INFO: # Iteration 4 bob.c++: INFO: # Iteration 5 bob.c++: INFO: # Iteration 6 bob.c++: INFO: # Iteration 7 bob.c++: INFO: # Iteration 8 bob.c++: INFO: # Iteration 9 bob.c++: INFO: # Iteration 10 bob.c++: INFO: # EM terminated: maximum number of iterations reached. --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- Ran 467 tests in 133.735s FAILED (SKIP=4, errors=100, failures=1) make[3]: *** [python/CMakeFiles/nosetests] Error 1 make[2]: *** [python/CMakeFiles/nosetests.dir/all] Error 2 make[1]: *** [python/CMakeFiles/nosetests.dir/rule] Error 2 make: *** [nosetests] Error 2 ``` The full nosetests output can be found on this gist: https://gist.github.com/tcgeophysics/8286521 Any advices on how to solves the issues pointed out by the nose output? Are they critical for bob? It seems there is a problem with PIL which might not be maintained anymore. Could PILLOW work instead?
1.0
bob nosetest fails on Mac OS 10.8.5 - I am running into an error with the ``` make nosetests ``` step of the compiling from source process. I am using a Mac OS Mountain Lion 10.8.5 (64bits) and installing bob with homebrew. The following is installed: Homebrew HOMEBREW_VERSION: 0.9.5 CPU: 8-core 64-bit ivybridge OS X: 10.8.5-x86_64 Xcode: 5.0.2 CLT: 5.0.1.0.1.1377666378 X11: 2.7.5 Python: 2.7.6 Nose: 1.3.0 VLFEAT is not installed and ``` -DWITH_VLFEAT=OFF ``` is specified I wrote a homebrew formula to install bob and it seems to work for now without the nosetests step. [bob.q formula here](https://gist.github.com/tcgeophysics/8156717) To repeat the problem 1 - install homebrew, 2 - Create a bob formula: ```brew create http://www.idiap.ch/software/bob/packages/bob-1.2.2.tar.gz``` 3 - Edit the bob formula: ```brew edit bob``` delete everything and paste the contents of the [bob.q formula here](https://gist.github.com/tcgeophysics/8156717) 4 - Run ```brew install -v bob``` to install in debug mode The install process does not complete with nosetests and exit with the following message: ``` ====================================================================== ERROR: bob.io.test.test_video.test_frameskip_format_mov_codec_wmv1 ---------------------------------------------------------------------- Traceback (most recent call last): File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/nose/case.py", line 197, in runTest self.test(*self.arg) File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/test/utils.py", line 108, in wrapper return test(*args, **kwargs) File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/io/test/test_video.py", line 71, in check_format_codec orig, framerate, encoded = function(shape, framerate, format, codec, fname) File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/io/utils.py", line 145, in frameskip_detection fontsize = estimate_fontsize(height, width, text_format) File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/io/utils.py", line 32, in estimate_fontsize font = ImageFont.truetype(DEFAULT_FONT, best_size) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/PIL/ImageFont.py", line 218, in truetype return FreeTypeFont(filename, size, index, encoding) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/PIL/ImageFont.py", line 134, in __init__ self.font = core.getfont(file, size, index, encoding) File "/usr/local/Cellar/python/2.7.6/Frameworks/Python.framework/Versions/2.7/lib/python2.7/site-packages/PIL/ImageFont.py", line 34, in __getattr__ raise ImportError("The _imagingft C module is not installed") ImportError: The _imagingft C module is not installed ====================================================================== FAIL: test01_plda_EM_vs_Python (bob.trainer.test.test_plda.PLDATrainerTest) ---------------------------------------------------------------------- Traceback (most recent call last): File "/tmp/bob-OifB/bob-1.2.2/build/lib/python2.7/site-packages/bob/trainer/test/test_plda.py", line 391, in test01_plda_EM_vs_Python self.assertTrue(numpy.allclose(m.f, m_py.f)) AssertionError: False is not true -------------------- >> begin captured logging << -------------------- bob.c++: INFO: # EMTrainer: bob.c++: INFO: # Iteration 1 bob.c++: INFO: # Iteration 2 bob.c++: INFO: # Iteration 3 bob.c++: INFO: # Iteration 4 bob.c++: INFO: # Iteration 5 bob.c++: INFO: # Iteration 6 bob.c++: INFO: # Iteration 7 bob.c++: INFO: # Iteration 8 bob.c++: INFO: # Iteration 9 bob.c++: INFO: # Iteration 10 bob.c++: INFO: # EM terminated: maximum number of iterations reached. --------------------- >> end captured logging << --------------------- ---------------------------------------------------------------------- Ran 467 tests in 133.735s FAILED (SKIP=4, errors=100, failures=1) make[3]: *** [python/CMakeFiles/nosetests] Error 1 make[2]: *** [python/CMakeFiles/nosetests.dir/all] Error 2 make[1]: *** [python/CMakeFiles/nosetests.dir/rule] Error 2 make: *** [nosetests] Error 2 ``` The full nosetests output can be found on this gist: https://gist.github.com/tcgeophysics/8286521 Any advices on how to solves the issues pointed out by the nose output? Are they critical for bob? It seems there is a problem with PIL which might not be maintained anymore. Could PILLOW work instead?
non_test
bob nosetest fails on mac os i am running into an error with the make nosetests step of the compiling from source process i am using a mac os mountain lion and installing bob with homebrew the following is installed homebrew homebrew version cpu core bit ivybridge os x xcode clt python nose vlfeat is not installed and dwith vlfeat off is specified i wrote a homebrew formula to install bob and it seems to work for now without the nosetests step to repeat the problem install homebrew create a bob formula brew create edit the bob formula brew edit bob delete everything and paste the contents of the run brew install v bob to install in debug mode the install process does not complete with nosetests and exit with the following message error bob io test test video test frameskip format mov codec traceback most recent call last file usr local cellar python frameworks python framework versions lib site packages nose case py line in runtest self test self arg file tmp bob oifb bob build lib site packages bob test utils py line in wrapper return test args kwargs file tmp bob oifb bob build lib site packages bob io test test video py line in check format codec orig framerate encoded function shape framerate format codec fname file tmp bob oifb bob build lib site packages bob io utils py line in frameskip detection fontsize estimate fontsize height width text format file tmp bob oifb bob build lib site packages bob io utils py line in estimate fontsize font imagefont truetype default font best size file usr local cellar python frameworks python framework versions lib site packages pil imagefont py line in truetype return freetypefont filename size index encoding file usr local cellar python frameworks python framework versions lib site packages pil imagefont py line in init self font core getfont file size index encoding file usr local cellar python frameworks python framework versions lib site packages pil imagefont py line in getattr raise importerror the imagingft c module is not installed importerror the imagingft c module is not installed fail plda em vs python bob trainer test test plda pldatrainertest traceback most recent call last file tmp bob oifb bob build lib site packages bob trainer test test plda py line in plda em vs python self asserttrue numpy allclose m f m py f assertionerror false is not true begin captured logging bob c info emtrainer bob c info iteration bob c info iteration bob c info iteration bob c info iteration bob c info iteration bob c info iteration bob c info iteration bob c info iteration bob c info iteration bob c info iteration bob c info em terminated maximum number of iterations reached end captured logging ran tests in failed skip errors failures make error make error make error make error the full nosetests output can be found on this gist any advices on how to solves the issues pointed out by the nose output are they critical for bob it seems there is a problem with pil which might not be maintained anymore could pillow work instead
0
384,498
26,588,267,653
IssuesEvent
2023-01-23 05:15:06
Soontao/cds-mysql
https://api.github.com/repos/Soontao/cds-mysql
opened
[FEATURE REQUEST] Load Testing
documentation enhancement
**Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] Load Test for - [ ] 500 tables and xxx views - [ ] 1000 tables and 1000 views - [ ] 5000 tables and 5000 views - [ ] 100k CSV data provision - [ ] 1m CSV data provision **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
1.0
[FEATURE REQUEST] Load Testing - **Is your feature request related to a problem? Please describe.** A clear and concise description of what the problem is. Ex. I'm always frustrated when [...] Load Test for - [ ] 500 tables and xxx views - [ ] 1000 tables and 1000 views - [ ] 5000 tables and 5000 views - [ ] 100k CSV data provision - [ ] 1m CSV data provision **Describe the solution you'd like** A clear and concise description of what you want to happen. **Describe alternatives you've considered** A clear and concise description of any alternative solutions or features you've considered. **Additional context** Add any other context or screenshots about the feature request here.
non_test
load testing is your feature request related to a problem please describe a clear and concise description of what the problem is ex i m always frustrated when load test for tables and xxx views tables and views tables and views csv data provision csv data provision describe the solution you d like a clear and concise description of what you want to happen describe alternatives you ve considered a clear and concise description of any alternative solutions or features you ve considered additional context add any other context or screenshots about the feature request here
0
287,713
24,853,928,990
IssuesEvent
2022-10-26 23:10:58
gravitational/teleport
https://api.github.com/repos/gravitational/teleport
closed
Teleport 11 Test Plan
testplan
## Manual Testing Plan Below are the items that should be manually tested with each release of Teleport. These tests should be run on both a fresh install of the version to be released as well as an upgrade of the previous version of Teleport. - [x] Adding nodes to a cluster @EdwardDowling - [x] Adding Nodes via Valid Static Token - [x] Adding Nodes via Valid Short-lived Tokens - [x] Adding Nodes via Invalid Token Fails - [x] Revoking Node Invitation - [x] Labels @EdwardDowling - [x] Static Labels - [x] Dynamic Labels - [x] Trusted Clusters @lxea - [x] Adding Trusted Cluster Valid Static Token - [x] Adding Trusted Cluster Valid Short-lived Token - [x] Adding Trusted Cluster Invalid Token - [x] Removing Trusted Cluster - [x] RBAC @atburke Make sure that invalid and valid attempts are reflected in audit log. - [x] Successfully connect to node with correct role - [x] Unsuccessfully connect to a node in a role restricting access by label - [x] Unsuccessfully connect to a node in a role restricting access by invalid SSH login - [x] Allow/deny role option: SSH agent forwarding - [x] Allow/deny role option: Port forwarding - [x] Verify that custom PAM environment variables are available as expected. @jakule - [x] Users @codingllama With every user combination, try to login and signup with invalid second factor, invalid password to see how the system reacts. WebAuthn in the release `tsh` binary is implemented using libfido2 for linux/macOS. Ask for a statically built pre-release binary for realistic tests. (`tsh fido2 diag` should work in our binary.) Webauthn in Windows build is implemented using `webauthn.dll`. (`tsh webauthn diag` with security key selected in dialog should work.) Touch ID requires a signed `tsh`, ask for a signed pre-release binary so you may run the tests. Windows Webauthn requires Windows 10 19H1 and device capable of Windows Hello. - [x] Adding Users Password Only - [x] Adding Users OTP - [x] Adding Users WebAuthn - [x] macOS/Linux - [x] Windows - [x] Adding Users via platform authenticator - [x] Touch ID - [x] Windows Hello - [x] Managing MFA devices - [x] Add an OTP device with `tsh mfa add` - [x] Add a WebAuthn device with `tsh mfa add` - [x] macOS/Linux - [x] Windows - [x] Add platform authenticator device with `tsh mfa add` - [x] Touch ID - [x] Windows Hello - [x] List MFA devices with `tsh mfa ls` - [x] Remove an OTP device with `tsh mfa rm` - [x] Remove a WebAuthn device with `tsh mfa rm` - [x] Attempt removing the last MFA device on the user - [x] with `second_factor: on` in `auth_service`, should fail - [x] with `second_factor: optional` in `auth_service`, should succeed - [x] Login Password Only - [x] Login with MFA - [x] Add an OTP, a WebAuthn and a Touch ID device with `tsh mfa add` - [x] Login via OTP - [x] Login via WebAuthn - [x] macOS/Linux - [x] Windows - [x] Login via platform authenticator - [x] Touch ID - [x] Windows Hello - [x] Login via WebAuthn using an U2F device U2F devices must be registered in a previous version of Teleport. Using Teleport v9, set `auth_service.authentication.second_factor = u2f`, restart the server and then register an U2F device (`tsh mfa add`). Upgrade the install to the current Teleport version (one major at a time) and try to login using the U2F device as your second factor - it should work. - [x] Deleting Users - [x] SSO @camscale - [x] Login OIDC - [x] Login SAML - [x] Login GitHub - [x] Backends @Joerger - [x] Teleport runs with etcd - [x] Teleport runs with dynamodb - [x] Teleport runs with SQLite - [x] Teleport runs with Firestore - [x] Session Recording @strideynet - [x] Session recording can be disabled - [x] Sessions can be recorded at the node - [x] Sessions in remote clusters are recorded in remote clusters - [x] Sessions can be recorded at the proxy - [x] Sessions on remote clusters are recorded in the local cluster - [x] Enable/disable host key checking. - [x] Audit Log @capnspacehook - [x] Failed login attempts are recorded - [x] Interactive sessions have the correct Server ID - [x] Server ID is the ID of the node in "session_recording: node" mode - [x] Server ID is the ID of the proxy in "session_recording: proxy" mode Node/Proxy ID may be found at `/var/lib/teleport/host_uuid` in the corresponding machine. Node IDs may also be queried via `tctl nodes ls`. - [x] Exec commands are recorded - [x] `scp` commands are recorded - [x] Subsystem results are recorded Subsystem testing may be achieved using both [Recording Proxy mode]( https://goteleport.com/teleport/docs/architecture/proxy/#recording-proxy-mode) and [OpenSSH integration]( https://goteleport.com/docs/server-access/guides/openssh/). Assuming the proxy is `proxy.example.com:3023` and `node1` is a node running OpenSSH/sshd, you may use the following command to trigger a subsystem audit log: ```shell sftp -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" root@node1 ``` - [x] Interact with a cluster using `tsh` @mdwn These commands should ideally be tested for recording and non-recording modes as they are implemented in a different ways. - [x] tsh ssh \<regular-node\> - [x] tsh ssh \<node-remote-cluster\> - [x] tsh ssh -A \<regular-node\> - [x] tsh ssh -A \<node-remote-cluster\> - [x] tsh ssh \<regular-node\> ls - [x] tsh ssh \<node-remote-cluster\> ls - [x] tsh join \<regular-node\> - [x] tsh join \<node-remote-cluster\> - [x] tsh play \<regular-node\> - [x] tsh play \<node-remote-cluster\> - [x] tsh scp \<regular-node\> - [x] tsh scp \<node-remote-cluster\> - [x] tsh ssh -L \<regular-node\> - [x] tsh ssh -L \<node-remote-cluster\> - [x] tsh ls - [x] tsh clusters - [x] Interact with a cluster using `ssh` @tobiaszheller Make sure to test both recording and regular proxy modes. - [x] ssh \<regular-node\> - [x] ssh \<node-remote-cluster\> - [x] ssh -A \<regular-node\> - [x] ssh -A \<node-remote-cluster\> - [x] ssh \<regular-node\> ls - [x] ssh \<node-remote-cluster\> ls - [x] scp \<regular-node\> - [x] scp \<node-remote-cluster\> - [x] ssh -L \<regular-node\> - [x] ssh -L \<node-remote-cluster\> - [x] Verify proxy jump functionality @Joerger Log into leaf cluster via root, shut down the root proxy and verify proxy jump works. - [x] tls routing disabled - [x] tsh ssh -J \<leaf.proxy.example.com:3023\> - [x] ssh -J \<leaf.proxy.example.com:3023\> - [x] tls routing enabled - [x] tsh ssh -J \<leaf.proxy.example.com:3080\> - [x] tsh proxy ssh -J \<leaf.proxy.example.com:3080\> - [x] Interact with a cluster using the Web UI @capnspacehook - [x] Connect to a Teleport node - [x] Connect to a OpenSSH node - [x] Check agent forwarding is correct based on role and proxy mode. ### User accounting @jakule - [x] Verify that active interactive sessions are tracked in `/var/run/utmp` on Linux. - [ ] Verify that interactive sessions are logged in `/var/log/wtmp` on Linux. ### Combinations @nklaassen For some manual testing, many combinations need to be tested. For example, for interactive sessions the 12 combinations are below. - [x] Connect to a OpenSSH node in a local cluster using OpenSSH. - [x] Connect to a OpenSSH node in a local cluster using Teleport. - [x] Connect to a OpenSSH node in a local cluster using the Web UI. - [x] Connect to a Teleport node in a local cluster using OpenSSH. - [x] Connect to a Teleport node in a local cluster using Teleport. - [x] Connect to a Teleport node in a local cluster using the Web UI. - [x] Connect to a OpenSSH node in a remote cluster using OpenSSH. - [x] Connect to a OpenSSH node in a remote cluster using Teleport. - [x] Connect to a OpenSSH node in a remote cluster using the Web UI. - [x] Connect to a Teleport node in a remote cluster using OpenSSH. - [x] Connect to a Teleport node in a remote cluster using Teleport. - [x] Connect to a Teleport node in a remote cluster using the Web UI. ### Teleport with EKS/GKE @tigrato * [x] Deploy Teleport on a single EKS cluster * [x] Deploy Teleport on two EKS clusters and connect them via trusted cluster feature * [x] Deploy Teleport Proxy outside of GKE cluster fronting connections to it (use [this script](https://github.com/gravitational/teleport/blob/master/examples/k8s-auth/get-kubeconfig.sh) to generate a kubeconfig) * [x] Deploy Teleport Proxy outside of EKS cluster fronting connections to it (use [this script](https://github.com/gravitational/teleport/blob/master/examples/k8s-auth/get-kubeconfig.sh) to generate a kubeconfig) ### Teleport with multiple Kubernetes clusters @AntonAM Note: you can use GKE or EKS or minikube to run Kubernetes clusters. Minikube is the only caveat - it's not reachable publicly so don't run a proxy there. * [x] Deploy combo auth/proxy/kubernetes_service outside of a Kubernetes cluster, using a kubeconfig * [x] Login with `tsh login`, check that `tsh kube ls` has your cluster * [x] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [x] Verify that the audit log recorded the above request and session * [x] Deploy combo auth/proxy/kubernetes_service inside of a Kubernetes cluster * [x] Login with `tsh login`, check that `tsh kube ls` has your cluster * [x] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [x] Verify that the audit log recorded the above request and session * [x] Deploy combo auth/proxy_service outside of the Kubernetes cluster and kubernetes_service inside of a Kubernetes cluster, connected over a reverse tunnel * [x] Login with `tsh login`, check that `tsh kube ls` has your cluster * [x] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [x] Verify that the audit log recorded the above request and session * [x] Deploy a second kubernetes_service inside of another Kubernetes cluster, connected over a reverse tunnel * [x] Login with `tsh login`, check that `tsh kube ls` has both clusters * [x] Switch to a second cluster using `tsh kube login` * [x] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` on the new cluster * [x] Verify that the audit log recorded the above request and session * [x] Deploy combo auth/proxy/kubernetes_service outside of a Kubernetes cluster, using a kubeconfig with multiple clusters in it * [x] Login with `tsh login`, check that `tsh kube ls` has all clusters * [x] Test Kubernetes screen in the web UI (tab is located on left side nav on dashboard): * [x] Verify that all kubes registered are shown with correct `name` and `labels` * [x] Verify that clicking on a rows connect button renders a dialogue on manual instructions with `Step 2` login value matching the rows `name` column * [x] Verify searching for `name` or `labels` in the search bar works * [x] Verify you can sort by `name` column ### Kubernetes auto-discovery @tigrato * [x] Test Kubernetes auto-discovery: * [x] Verify that Azure AKS clusters are discovered and enrolled for different Azure Auth configs: * [x] Local Accounts only * [x] Azure AD * [x] Azure RBAC * [x] Verify that AWS EKS clusters are discovered and enrolled ### Kubernetes Secret Storage @tigrato * [x] Kubernetes Secret storage for Agent's Identity * [x] Install Teleport 11 agent with a short-lived token * [x] Validate if the Teleport is installed as a Kubernetes `Statefulset` * [x] Restart the agent after token TTL expires to see if it reuses the same identity. (#17474) * [x] Upgrade from Teleport 10 agent with storage * [x] Validate if agent identity was read from storage and stored in the secret without generating a new one * [x] Validate if the Teleport is still running as a `Statefulset` resource and if it contains the new ENV variables * [x] Upgrade from Teleport 10 agent without storage (Failing due to #17437) * [x] Validate if the agent identity is created and stored in the secret using the long-lived token. * [x] Validate if the Teleport Kubernetes `Deployment` was correctly converted into a Statefulset and if the old `Deployment` object was removed after a successful upgrade * [x] Force cluster CA rotation * [x] Test Kubernetes exec via websockets - [client](https://github.com/kubernetes-client/javascript/blob/45b68c98e62b6cc4152189b9fd4a27ad32781bc4/examples/typescript/exec/exec-example.ts) ### Teleport with FIPS mode @alistanis * [x] Perform trusted clusters, Web and SSH sanity check with all teleport components deployed in FIPS mode. ### ACME @alistanis - [x] Teleport can fetch TLS certificate automatically using ACME protocol. ### Migrations @jakule * [ ] Migrate trusted clusters from 10 to 11 * [ ] Migrate auth server on main cluster, then rest of the servers on main cluster SSH should work for both main and old clusters * [ ] Migrate auth server on remote cluster, then rest of the remote cluster SSH should work ### Command Templates When interacting with a cluster, the following command templates are useful: #### OpenSSH ``` # when connecting to the recording proxy, `-o 'ForwardAgent yes'` is required. ssh -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" \ node.example.com # the above command only forwards the agent to the proxy, to forward the agent # to the target node, `-o 'ForwardAgent yes'` needs to be passed twice. ssh -o "ForwardAgent yes" \ -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" \ node.example.com # when connecting to a remote cluster using OpenSSH, the subsystem request is # updated with the name of the remote cluster. ssh -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p@foo.com" \ node.foo.com ``` #### Teleport ``` # when connecting to a OpenSSH node, remember `-p 22` needs to be passed. tsh --proxy=proxy.example.com --user=<username> --insecure ssh -p 22 node.example.com # an agent can be forwarded to the target node with `-A` tsh --proxy=proxy.example.com --user=<username> --insecure ssh -A -p 22 node.example.com # the --cluster flag is used to connect to a node in a remote cluster. tsh --proxy=proxy.example.com --user=<username> --insecure ssh --cluster=foo.com -p 22 node.foo.com ``` ### Teleport with SSO Providers - [x] G Suite install instructions work @AntonAM - [x] G Suite Screenshots are up to date - [x] Azure Active Directory (AD) install instructions work @alistanis - [x] Azure Active Directory (AD) Screenshots are up to date - [x] ActiveDirectory (ADFS) install instructions work @alistanis - [x] Active Directory (ADFS) Screenshots are up to date - [ ] Okta install instructions work @camscale - [ ] Okta Screenshots are up to date - [x] OneLogin install instructions work @hugoShaka - [ ] OneLogin Screenshots are up to date: https://github.com/gravitational/teleport/issues/17485 - [x] GitLab install instructions work @capnspacehook - [x] GitLab Screenshots are up to date - [ ] OIDC install instructions work @camscale - [ ] OIDC Screenshots are up to date - [ ] All providers with guides in docs are covered in this test plan ### `tctl sso` family of commands @Tener For help with setting up sso connectors, check out the [Quick GitHub/SAML/OIDC Setup Tips] `tctl sso configure` helps to construct a valid connector definition: - [x] `tctl sso configure github ...` creates valid connector definitions - [x] `tctl sso configure oidc ...` creates valid connector definitions - [x] `tctl sso configure saml ...` creates valid connector definitions `tctl sso test` test a provided connector definition, which can be loaded from file or piped in with `tctl sso configure` or `tctl get --with-secrets`. Valid connectors are accepted, invalid are rejected with sensible error messages. - [x] Connectors can be tested with `tctl sso test`. - [x] GitHub - [x] SAML - [x] OIDC - [x] Google Workspace - [x] Non-Google IdP ### Teleport Plugins @hugoShaka - [x] Test receiving a message via Teleport Slackbot - [x] Test receiving a new Jira Ticket via Teleport Jira ### AWS Node Joining @nklaassen [Docs](https://goteleport.com/docs/setup/guides/joining-nodes-aws/) - [x] On EC2 instance with `ec2:DescribeInstances` permissions for local account: `TELEPORT_TEST_EC2=1 go test ./integration -run TestEC2NodeJoin` - [x] On EC2 instance with any attached role: `TELEPORT_TEST_EC2=1 go test ./integration -run TestIAMNodeJoin` - [x] EC2 Join method in IoT mode with node and auth in different AWS accounts - [x] IAM Join method in IoT mode with node and auth in different AWS accounts ### Passwordless @codingllama Passwordless requires `tsh` compiled with libfido2 for most operations (apart from Touch ID). Ask for a statically-built `tsh` binary for realistic tests. Touch ID requires a properly built and signed `tsh` binary. Ask for a pre-release binary, so you may run the tests. This sections complements "Users -> Managing MFA devices". `tsh` binaries for each operating system (Linux, macOS and Windows) must be tested separately for FIDO2 items. - [x] Diagnostics Commands should pass all tests. - [x] `tsh fido2 diag` (macOS/Linux) - [x] `tsh touchid diag` (macOS only) - [x] `tsh webauthnwin diag` (Windows only) - [x] Registration - [x] Register a passworldess FIDO2 key (`tsh mfa add`, choose WEBAUTHN and passwordless) - [x] macOS/Linux - [x] Windows - [x] Register a platform authenticator - [x] Touch ID credential (`tsh mfa add`, choose TOUCHID) - [x] Windows hello credential (`tsh mfa add`, choose WEBAUTHN and passwordless) - [x] Login - [x] Passwordless login using FIDO2 (`tsh login --auth=passwordless`) - [x] macOS/Linux - [x] Windows - [x] Passwordless login using platform authenticator (`tsh login --auth=passwordless`) - [x] Touch ID - [x] Windows Hello - [x] `tsh login --auth=passwordless --mfa-mode=cross-platform` uses FIDO2 - [x] macOS/Linux - [x] Windows - [x] `tsh login --auth=passwordless --mfa-mode=platform` uses platform authenticator - [x] Touch ID - [x] Windows Hello - [x] `tsh login --auth=passwordless --mfa-mode=auto` prefers platform authenticator - [x] Touch ID - [x] Windows Hello - [x] Passwordless disable switch works (`auth_service.authentication.passwordless = false`) - [x] Cluster in passwordless mode defaults to passwordless (`auth_service.authentication.connector_name = passwordless`) - [x] Cluster in passwordless mode allows MFA login (`tsh login --auth=local`) - [x] Touch ID support commands - [x] `tsh touchid ls` works - [x] `tsh touchid rm` works (careful, may lock you out!) ### Hardware Key Support @Joerger Hardware Key Support is an Enterprise feature and is not available for OSS. You will need a YubiKey 4.3+ to test this feature. This feature has additional build requirements, so it should be tested with a pre-release build from Drone (eg: `https://get.gravitational.com/teleport-ent-v11.0.0-alpha.2-linux-amd64-bin.tar.gz`). These tests should be carried out sequentially. `tsh` tests should be carried out on Linux, MacOS, and Windows. - [x] `tsh login` as user with [Webauthn](https://goteleport.com/docs/access-controls/guides/webauthn/) login and no hardware key requirement. - [x] Request a role with `role.role_options.require_session_mfa: hardware_key` - `tsh login --request-roles=hardware_key_required` - [x] Assuming the role should force automatic re-login with yubikey - [x] `tsh ssh` - [x] Requires yubikey to be connected for re-login - [x] Prompts for per-session MFA - [x] Request a role with `role.role_options.require_session_mfa: hardware_key_touch` - `tsh login --request-roles=hardware_key_touch_required` - [x] Assuming the role should force automatic re-login with yubikey - [x] Prompts for touch if not cached (last touch within 15 seconds) - [x] `tsh ssh` - [x] Requires yubikey to be connected for re-login - [x] Prompts for touch if not cached - [x] `tsh logout` and `tsh login` as the user with no hardware key requirement. - [x] Upgrade auth settings to `auth_service.authentication.require_session_mfa: hardware_key` - [x] Using the existing login session (`tsh ls`) should force automatic re-login with yubikey - [x] `tsh ssh` - [x] Requires yubikey to be connected for re-login - [x] Prompts for per-session MFA - [x] Upgrade auth settings to `auth_service.authentication.require_session_mfa: hardware_key_touch` - [x] Using the existing login session (`tsh ls`) should force automatic re-login with yubikey - [x] Prompts for touch if not cached - [x] `tsh ssh` - [x] Requires yubikey to be connected for re-login - [x] Prompts for touch if not cached ## Performance @rosstimothy @fspmarshall Perform all tests on the following configurations: - [x] With default networking configuration - [x] With Proxy Peering Enabled - [x] With TLS Routing Enabled * Cluster with 10K direct dial nodes: - [x] etcd - [x] DynamoDB - [ ] Firestore * Cluster with 10K reverse tunnel nodes: - [x] etcd - [x] DynamoDB - [ ] Firestore * Cluster with 500 trusted clusters: - [x] etcd - [x] DynamoDB - [ ] Firestore ### Soak Test Run 30 minute soak test with a mix of interactive/non-interactive sessions for both direct and reverse tunnel nodes: ```shell tsh bench --duration=30m user@direct-dial-node ls tsh bench -i --duration=30m user@direct-dial-node ps uax tsh bench --duration=30m user@reverse-tunnel-node ls tsh bench -i --duration=30m user@reverse-tunnel-node ps uax ``` Observe prometheus metrics for goroutines, open files, RAM, CPU, Timers and make sure there are no leaks - [x] Verify that prometheus metrics are accurate. ### Concurrent Session Test * Cluster with 1k reverse tunnel nodes Run a concurrent session test that will spawn 5 interactive sessions per node in the cluster: ```shell tsh bench sessions --max=5000 user ls tsh bench sessions --max=5000 --web user ls ``` - [x] Verify that all 5000 sessions are able to be established. - [ ] Verify that tsh and the web UI are still functional. ## Teleport with Cloud Providers ### AWS @hugoShaka - [x] Deploy Teleport to AWS. Using DynamoDB & S3 - [x] Deploy Teleport Enterprise to AWS. Using HA Setup https://gravitational.com/teleport/docs/aws-terraform-guide/ ### GCP @AntonAM - [x] Deploy Teleport to GCP. Using Cloud Firestore & Cloud Storage - [x] Deploy Teleport to GKE. Google Kubernetes engine. - [x] Deploy Teleport Enterprise to GCP. ### IBM @atburke - [x] Deploy Teleport to IBM Cloud. Using IBM Database for etcd & IBM Object Store - [x] Deploy Teleport to IBM Cloud Kubernetes. - [x] Deploy Teleport Enterprise to IBM Cloud. ## Application Access @mdwn - [x] Run an application within local cluster. - [x] Verify the debug application `debug_app: true` works. - [x] Verify an application can be configured with command line flags. - [x] Verify an application can be configured from file configuration. - [x] Verify that applications are available at auto-generated addresses `name.rootProxyPublicAddr` and well as `publicAddr`. - [x] Run an application within a trusted cluster. - [x] Verify that applications are available at auto-generated addresses `name.rootProxyPublicAddr`. - [x] Verify Audit Records. - [x] `app.session.start` and `app.session.chunk` events are created in the Audit Log. - [x] `app.session.chunk` points to a 5 minute session archive with multiple `app.session.request` events inside. - [x] `tsh play <chunk-id>` can fetch and print a session chunk archive. - [x] Verify JWT using [verify-jwt.go](https://github.com/gravitational/teleport/blob/master/examples/jwt/verify-jwt.go). - [x] Verify RBAC. - [x] Verify [CLI access](https://goteleport.com/docs/application-access/guides/api-access/) with `tsh app login`. - [x] Verify AWS console access. - [x] Can log into AWS web console through the web UI. - [x] Can interact with AWS using `tsh aws` commands. - [x] Verify dynamic registration. - [x] Can register a new app using `tctl create`. - [x] Can update registered app using `tctl create -f`. - [x] Can delete registered app using `tctl rm`. - [x] Test Applications screen in the web UI (tab is located on left side nav on dashboard): - [x] Verify that all apps registered are shown - [x] Verify that clicking on the app icon takes you to another tab - [x] Verify using the bash command produced from `Add Application` dialogue works (refresh app screen to see it registered) ## Database Access @smallinsky + db access team - [x] Connect to a database within a local cluster. - [x] Self-hosted Postgres. @Tener - [x] Self-hosted MySQL. @GavinFrazar - [x] Self-hosted MariaDB. @GavinFrazar - [x] Self-hosted MongoDB. @Tener - [x] Self-hosted CockroachDB. @GavinFrazar - [x] Self-hosted Redis. @greedy52 - [x] Self-hosted Redis Cluster. @greedy52 - [x] Self-hosted MSSQL. @smallinsky - [x] AWS Aurora Postgres. @gabrielcorado - [x] AWS Aurora MySQL. @Tener - [x] AWS Redshift. @gabrielcorado - [x] AWS ElastiCache. @greedy52 - [x] AWS MemoryDB. @greedy52 - [x] GCP Cloud SQL Postgres. @gabrielcorado - [x] GCP Cloud SQL MySQL. @Tener - [x] Snowflake. @smallinsky - [x] Azure Cache for Redis. @smallinsky - [x] Elasticsearch. @greedy52 - [x] Cassandra/ScyllaDB. @gabrielcorado - [x] Connect to a database within a remote cluster via a trusted cluster. - [x] Self-hosted Postgres. @Tener - [x] Self-hosted MySQL. @GavinFrazar - [x] Self-hosted MariaDB. @GavinFrazar - [x] Self-hosted MongoDB. @Tener - [x] Self-hosted CockroachDB. @GavinFrazar - [x] Self-hosted Redis. @greedy52 - [x] Self-hosted Redis Cluster. @greedy52 - [x] Self-hosted MSSQL. @smallinsky - [x] AWS Aurora Postgres. @gabrielcorado - [x] AWS Aurora MySQL. @Tener - [x] AWS Redshift. @gabrielcorado - [x] AWS ElastiCache. @greedy52 - [x] AWS MemoryDB. @greedy52 - [x] GCP Cloud SQL Postgres. @gabrielcorado - [x] GCP Cloud SQL MySQL. @Tener - [x] Snowflake. @smallinsky - [x] Azure Cache for Redis. @smallinsky - [x] Elasticsearch. @greedy52 - [x] Cassandra/ScyllaDB. @gabrielcorado - [x] Verify audit events. @GavinFrazar - [x] `db.session.start` is emitted when you connect. - [x] `db.session.end` is emitted when you disconnect. - [x] `db.session.query` is emitted when you execute a SQL query. - [x] Verify RBAC. - [x] `tsh db ls` shows only databases matching role's `db_labels`. @gabrielcorado - [x] Can only connect as users from `db_users`. @gabrielcorado - [x] _(Postgres only)_ Can only connect to databases from `db_names`. @gabrielcorado - [x] `db.session.start` is emitted when connection attempt is denied. - [x] _(MongoDB only)_ Can only execute commands in databases from `db_names`. @gabrielcorado - [x] `db.session.query` is emitted when command fails due to permissions. - [x] Can configure per-session MFA. @GavinFrazar - [x] MFA tap is required on each `tsh db connect`. - [x] Verify dynamic registration. @GavinFrazar - [x] Can register a new database using `tctl create`. - [x] Can update registered database using `tctl create -f`. - [x] Can delete registered database using `tctl rm`. - [x] Verify discovery. - [x] AWS - [x] Can detect and register RDS instances. @GavinFrazar - [x] Can detect and register Aurora clusters, and their reader and custom endpoints. @gabrielcorado - [x] Can detect and register Redshift clusters. @smallinsky - [x] Can detect and register ElastiCache Redis clusters. @greedy52 - [x] Can detect and register MemoryDB clusters. @greedy52 - [x] Azure - [x] Can detect and register MySQL and Postgres instances. @GavinFrazar - [x] Can detect and register Azure Cache for Redis servers. @smallinsky - [x] Verify Teleport managed users (password rotation, auto 'auth' on connection, etc.). @greedy52 - [x] Can detect and manage ElastiCache users - [x] Can detect and manage MemoryDB users - [x] Test Databases screen in the web UI (tab is located on left side nav on dashboard): @Tener - [x] Verify that all dbs registered are shown with correct `name`, `description`, `type`, and `labels` - [x] Verify that clicking on a rows connect button renders a dialogue on manual instructions with `Step 2` login value matching the rows `name` column - [x] Verify searching for all columns in the search bar works - [x] Verify you can sort by all columns except `labels` - [x] Other - [x] MySQL server version reported by Teleport is correct. @smallinsky ## TLS Routing @smallinsky - [x] Verify that teleport proxy `v2` configuration starts only a single listener. @smallinsky ``` version: v2 teleport: proxy_service: enabled: "yes" public_addr: ['root.example.com'] web_listen_addr: 0.0.0.0:3080 ``` - [x] Run Teleport Proxy in `multiplex` mode `auth_service.proxy_listener_mode: "multiplex"` @smallinsky - [x] Trusted cluster - [x] Setup trusted clusters using single port setup `web_proxy_addr == tunnel_addr` ``` kind: trusted_cluster spec: ... web_proxy_addr: root.example.com:443 tunnel_addr: root.example.com:443 ... ``` - [ ] Database Access - [x] Verify that `tsh db connect` works through proxy running in `multiplex` mode - [x] Postgres @Tener - [x] MySQL @GavinFrazar - [x] MariaDB @GavinFrazar - [x] MongoDB @Tener - [x] CockroachDB @GavinFrazar - [x] Redis @greedy52 - [x] MSSQL @smallinsky - [x] Snowflake @smallinsky - [x] Elasticsearch. @greedy52 - [x] Cassandra/ScyllaDB. @gabrielcorado - [ ] Verify connecting to a database through TLS ALPN SNI local proxy `tsh proxy db` with a GUI client. @smallinsky @GavinFrazar @greedy52 @Tener @gabrielcorado - [x] Verify tsh proxy db with teleport proxy behind ALB. @smallinsky - [x] Application Access @GavinFrazar - [x] Verify app access through proxy running in `multiplex` mode - [x] SSH Access @gabrielcorado - [x] Connect to a OpenSSH server through a local ssh proxy `ssh -o "ForwardAgent yes" -o "ProxyCommand tsh proxy ssh" user@host.example.com` - [x] Connect to a OpenSSH server on leaf-cluster through a local ssh proxy`ssh -o "ForwardAgent yes" -o "ProxyCommand tsh proxy ssh --user=%r --cluster=leaf-cluster %h:%p" user@node.foo.com` - [x] Verify `tsh ssh` access through proxy running in multiplex mode - [x] Kubernetes access: @GavinFrazar - [x] Verify kubernetes access through proxy running in `multiplex` mode ## Desktop Access @ibeckermayer @probakowski @LKozlowski - Direct mode (set `listen_addr`): - [x] Can connect to desktop defined in static `hosts` section. - [x] Can connect to desktop discovered via LDAP - IoT mode (reverse tunnel through proxy): - [x] Can connect to desktop defined in static `hosts` section. - [x] Can connect to desktop discovered via LDAP - [x] Connect multiple `windows_desktop_service`s to the same Teleport cluster, verify that connections to desktops on different AD domains works. (Attempt to connect several times to verify that you are routed to the correct `windows_desktop_service`) - Verify user input - [x] Download [Keyboard Key Info](https://dennisbabkin.com/kbdkeyinfo/) and verify all keys are processed correctly in each supported browser. Known issues: F11 cannot be captured by the browser without [special configuration](https://social.technet.microsoft.com/Forums/en-US/784b2bbe-353f-412e-ac9a-193d81f306b6/remote-desktop-for-mac-f11-key-not-working-on-macbook-pro-touchbar?forum=winRDc) on MacOS. - [x] Left click and right click register as Windows clicks. (Right click on the desktop should show a Windows menu, not a browser context menu) - [x] Vertical and horizontal scroll work. [Horizontal Scroll Test](https://codepen.io/jaemskyle/pen/inbmB) - Locking - [x] Verify that placing a user lock terminates an active desktop session. - [x] Verify that placing a desktop lock terminates an active desktop session. - [x] Verify that placing a role lock terminates an active desktop session. - Labeling - [x] Set `client_idle_timeout` to a small value and verify that idle sessions are terminated (the session should end and an audit event will confirm it was due to idle connection) - [x] All desktops have `teleport.dev/origin` label. - [x] Dynamic desktops have additional `teleport.dev` labels for OS, OS Version, DNS hostname. - [x] Regexp-based host labeling applies across all desktops, regardless of origin. - RBAC - [x] RBAC denies access to a Windows desktop due to labels - [x] RBAC denies access to a Windows desktop with the wrong OS-login. - Clipboard Support - When a user has a role with clipboard sharing enabled and is using a chromium based browser - [x] Going to a desktop when clipboard permissions are in "Ask" mode (aka "prompt") causes the browser to show a prompt while the UI shows a spinner - [x] X-ing out of the prompt (causing the clipboard permission to remain in "Ask" mode) causes the prompt to show up again - [x] Denying clipboard permissions brings up a relevant error alert (with "Clipboard Sharing Disabled" in the top bar) - [x] Allowing clipboard permissions allows you to see the desktop session, with "Clipboard Sharing Enabled" highlighted in the top bar - [x] Copy text from local workstation, paste into remote desktop - [x] Copy text from remote desktop, paste into local workstation - When a user has a role with clipboard sharing enabled and is *not* using a chromium based browser - [x] The UI shows a relevant alert and "Clipboard Sharing Disabled" is highlighted in the top bar - When a user has a role with clipboard sharing *disabled* and is using a chromium and non-chromium based browser (confirm both) - [x] The live session should show disabled in the top bar and copy/paste should not work between your workstation and the remote desktop. - Directory Sharing - On supported, non-chromium based browsers (Firefox/Safari) - [x] Attempting to share directory shows a dismissible "Unsupported Action" dialog - On supported, chromium based browsers (Chrome/Edge) - Begin sharing works - [x] The shared directory icon in the top right of the screen is highlighted when directory sharing is initiated - [x] The shared directory appears as a network drive named "<directory_name> on teleport" - [x] The share directory menu option dissapears from the menu - Navigation - [x] The folders of the shared directory are navigable (move up and down the directory tree) - CRUD - [x] A new text file can be created - [x] The text file can be written to (saved) - [x] The text file can be read (close it, check that its saved on the local machine, then open it again on the remote) - [x] The text file can be deleted - File/Folder movement - In to out (make at least one of these from a non-top-level-directory) - [x] A file from inside the shared directory can be drag-and-dropped outside the shared directory - [x] A folder from inside the shared directory can be drag-and-dropped outside the shared directory (and its contents retained) - [x] A file from inside the shared directory can be cut-pasted outside the shared directory - [x] A folder from inside the shared directory can be cut-pasted outside the shared directory - [x] A file from inside the shared directory can be copy-pasted outside the shared directory - [x] A folder from inside the shared directory can be copy-pasted outside the shared directory - Out to in (make at least one of these overwrite an existing file, and one go into a non-top-level directory) - [x] A file from outside the shared directory can be drag-and-dropped into the shared directory - [x] A folder from outside the shared directory can be drag-and-dropped into the shared directory (and its contents retained) - [x] A file from outside the shared directory can be cut-pasted into the shared directory - [x] A folder from outside the shared directory can be cut-pasted into the shared directory - [x] A file from outside the shared directory can be copy-pasted into the shared directory - [x] A folder from outside the shared directory can be copy-pasted into the shared directory - Within - [x] A file from inside the shared directory cannot be drag-and-dropped to another folder inside the shared directory: a dismissible "Unsupported Action" dialog is shown - [x] A folder from inside the shared directory cannot be drag-and-dropped to another folder inside the shared directory: a dismissible "Unsupported Action" dialog is shown - [x] A file from inside the shared directory cannot be cut-pasted to another folder inside the shared directory: a dismissible "Unsupported Action" dialog is shown - [x] A folder from inside the shared directory cannot be cut-pasted to another folder inside the shared directory: a dismissible "Unsupported Action" dialog is shown - [x] A file from inside the shared directory can be copy-pasted to another folder inside the shared directory - [x] A folder from inside the shared directory can be copy-pasted to another folder inside shared directory (and its contents retained) - RBAC - [x] Give the user one role that explicitly disables directory sharing (`desktop_directory_sharing: false`) and confirm that the option to share a directory doesn't appear in the menu - Per-Session MFA (try webauthn on each of Chrome, Safari, and Firefox; u2f only works with Firefox) - [x] Attempting to start a session no keys registered shows an error message - [x] ~Attempting to start a session with a u2f key registered shows an error message~ (N/A now that u2f support has been removed) - [x] Attempting to start a session with a webauthn registered pops up the "Verify Your Identity" dialog - [x] Hitting "Cancel" shows an error message - [x] Hitting "Verify" causes your browser to prompt you for MFA - [x] Cancelling that browser MFA prompt shows an error - [x] Successful MFA verification allows you to connect - Session Recording - [x] Verify sessions are not recorded if *all* of a user's roles disable recording - [x] Verify sync recording (`mode: node-sync` or `mode: proy-sync`) - [x] Verify async recording (`mode: node` or `mode: proxy`) - [x] Sessions show up in session recordings UI with desktop icon - [x] Sessions can be played back, including play/pause functionality - [x] A session that ends with a TDP error message can be played back, ends by displaying the error message, and the progress bar progresses to the end. - [x] Attempting to play back a session that doesn't exist (i.e. by entering a non-existing session id in the url) shows a relevant error message. - [x] RBAC for sessions: ensure users can only see their own recordings when using the RBAC rule from our [docs](../../docs/pages/access-controls/reference.mdx#rbac-for-sessions) - Audit Events (check these after performing the above tests) - [x] `windows.desktop.session.start` (`TDP00I`) emitted on start - [x] `windows.desktop.session.start` (`TDP00W`) emitted when session fails to start (due to RBAC, for example) - [x] `windows.desktop.session.end` (`TDP01I`) emitted on end - [x] `desktop.clipboard.send` (`TDP02I`) emitted for local copy -> remote paste - [x] `desktop.clipboard.receive` (`TDP03I`) emitted for remote copy -> local paste ## Binaries compatibility @fheinecke - Verify that teleport/tsh/tctl/tbot run on: - [ ] CentOS 7 - [ ] CentOS 8 - [ ] Ubuntu 18.04 - [ ] Ubuntu 20.04 - [ ] Debian 9 - Verify tsh runs on: - [ ] Windows 10 - [ ] MacOS ## Machine ID @timothyb89 ### SSH With a default Teleport instance configured with a SSH node: - [x] Verify you are able to create a new bot user with `tctl bots add robot --roles=access`. Follow the instructions provided in the output to start `tbot` - [x] Verify you are able to connect to the SSH node using openssh with the generated `ssh_config` in the destination directory - [x] Verify that after the renewal period (default 20m, but this can be reduced via configuration), that newly generated certificates are placed in the destination directory - [x] Verify that sending both `SIGUSR1` and `SIGHUP` to a running tbot process causes a renewal and new certificates to be generated - [x] Verify that you are able to make a connection to the SSH node using the `ssh_config` provided by `tbot` after each phase of a manual CA rotation. Ensure the above tests are completed for both: - [x] Directly connecting to the auth server - [x] Connecting to the auth server via the proxy reverse tunnel ### DB Access With a default Postgres DB instance, a Teleport instance configured with DB access and a bot user configured: - [x] Verify you are able to connect to and interact with a database using `tbot db` while `tbot start` is running ## Host users creation @lxea [Host users creation docs](https://github.com/gravitational/teleport/pull/13056) [Host users creation RFD](https://github.com/gravitational/teleport/pull/11077) <!--- TODO(lxea): replace links with actual docs once merged [Host users creation docs](../../docs/pages/server-access/guides/host-user-creation.mdx) [Host users creation RFD](../../rfd/0057-automatic-user-provisioning.md) --> - Verify host users creation functionality - [x] non-existing users are created automatically - [x] users are added to groups - [x] non existing configured groups are created - [x] created users are added to the `teleport-system` group - [x] users are cleaned up after their session ends - [x] cleanup occurs if a program was left running after session ends - [x] sudoers file creation is successful - [x] Invalid sudoers files are _not_ created - [x] existing host users are not modified - [x] setting `disable_create_host_user: true` stops user creation from occurring ## CA rotations @espadolini - Verify the CA rotation functionality itself (by checking in the backend or with `tctl get cert_authority`) - [x] `standby` phase: only `active_keys`, no `additional_trusted_keys` - [x] `init` phase: `active_keys` and `additional_trusted_keys` - [x] `update_clients` and `update_servers` phases: the certs from the `init` phase are swapped - [x] `standby` phase: only the new certs remain in `active_keys`, nothing in `additional_trusted_keys` - [x] `rollback` phase (second pass, after completing a regular rotation): same content as in the `init` phase - [x] `standby` phase after `rollback`: same content as in the previous `standby` phase - Verify functionality in all phases (clients might have to log in again in lieu of waiting for credentials to expire between phases) - [x] SSH session in tsh from a previous phase - [x] SSH session in web UI from a previous phase - [x] New SSH session with tsh - [x] New SSH session with web UI - [x] New SSH session in a child cluster on the same major version - [x] New SSH session in a child cluster on the previous major version - [x] New SSH session from a parent cluster - [x] Application access through a browser - [x] Application access through curl with `tsh app login` - [x] `kubectl get po` after `tsh kube login` - [ ] Database access (no configuration change should be necessary if the database CA isn't rotated, other Teleport functionality should not be affected if only the database CA is rotated) ## EC2 Discovery @lxea [EC2 Discovery docs](https://goteleport.com/docs/ver/11.0/server-access/guides/ec2-discovery/) - Verify EC2 instance discovery - [x] Only EC2 instances matching given AWS tags have the installer executed on them - [x] Only the IAM permissions mentioned in the discovery docs are required for operation - [x] Custom scripts specified in different matchers are executed - [x] Custom SSM documents specified in different matchers are executed - [x] New EC2 instances with matching AWS tags are discovered and added to the teleport cluster - [x] Large numbers of EC2 instances (51+) are all successfully added to the cluster - [x] Nodes that have been discovered do not have the install script run on the node multiple times ## Resources [Quick GitHub/SAML/OIDC Setup Tips] <!--- reference style links --> [Quick GitHub/SAML/OIDC Setup Tips]: https://gravitational.slab.com/posts/quick-git-hub-saml-oidc-setup-6dfp292a
1.0
Teleport 11 Test Plan - ## Manual Testing Plan Below are the items that should be manually tested with each release of Teleport. These tests should be run on both a fresh install of the version to be released as well as an upgrade of the previous version of Teleport. - [x] Adding nodes to a cluster @EdwardDowling - [x] Adding Nodes via Valid Static Token - [x] Adding Nodes via Valid Short-lived Tokens - [x] Adding Nodes via Invalid Token Fails - [x] Revoking Node Invitation - [x] Labels @EdwardDowling - [x] Static Labels - [x] Dynamic Labels - [x] Trusted Clusters @lxea - [x] Adding Trusted Cluster Valid Static Token - [x] Adding Trusted Cluster Valid Short-lived Token - [x] Adding Trusted Cluster Invalid Token - [x] Removing Trusted Cluster - [x] RBAC @atburke Make sure that invalid and valid attempts are reflected in audit log. - [x] Successfully connect to node with correct role - [x] Unsuccessfully connect to a node in a role restricting access by label - [x] Unsuccessfully connect to a node in a role restricting access by invalid SSH login - [x] Allow/deny role option: SSH agent forwarding - [x] Allow/deny role option: Port forwarding - [x] Verify that custom PAM environment variables are available as expected. @jakule - [x] Users @codingllama With every user combination, try to login and signup with invalid second factor, invalid password to see how the system reacts. WebAuthn in the release `tsh` binary is implemented using libfido2 for linux/macOS. Ask for a statically built pre-release binary for realistic tests. (`tsh fido2 diag` should work in our binary.) Webauthn in Windows build is implemented using `webauthn.dll`. (`tsh webauthn diag` with security key selected in dialog should work.) Touch ID requires a signed `tsh`, ask for a signed pre-release binary so you may run the tests. Windows Webauthn requires Windows 10 19H1 and device capable of Windows Hello. - [x] Adding Users Password Only - [x] Adding Users OTP - [x] Adding Users WebAuthn - [x] macOS/Linux - [x] Windows - [x] Adding Users via platform authenticator - [x] Touch ID - [x] Windows Hello - [x] Managing MFA devices - [x] Add an OTP device with `tsh mfa add` - [x] Add a WebAuthn device with `tsh mfa add` - [x] macOS/Linux - [x] Windows - [x] Add platform authenticator device with `tsh mfa add` - [x] Touch ID - [x] Windows Hello - [x] List MFA devices with `tsh mfa ls` - [x] Remove an OTP device with `tsh mfa rm` - [x] Remove a WebAuthn device with `tsh mfa rm` - [x] Attempt removing the last MFA device on the user - [x] with `second_factor: on` in `auth_service`, should fail - [x] with `second_factor: optional` in `auth_service`, should succeed - [x] Login Password Only - [x] Login with MFA - [x] Add an OTP, a WebAuthn and a Touch ID device with `tsh mfa add` - [x] Login via OTP - [x] Login via WebAuthn - [x] macOS/Linux - [x] Windows - [x] Login via platform authenticator - [x] Touch ID - [x] Windows Hello - [x] Login via WebAuthn using an U2F device U2F devices must be registered in a previous version of Teleport. Using Teleport v9, set `auth_service.authentication.second_factor = u2f`, restart the server and then register an U2F device (`tsh mfa add`). Upgrade the install to the current Teleport version (one major at a time) and try to login using the U2F device as your second factor - it should work. - [x] Deleting Users - [x] SSO @camscale - [x] Login OIDC - [x] Login SAML - [x] Login GitHub - [x] Backends @Joerger - [x] Teleport runs with etcd - [x] Teleport runs with dynamodb - [x] Teleport runs with SQLite - [x] Teleport runs with Firestore - [x] Session Recording @strideynet - [x] Session recording can be disabled - [x] Sessions can be recorded at the node - [x] Sessions in remote clusters are recorded in remote clusters - [x] Sessions can be recorded at the proxy - [x] Sessions on remote clusters are recorded in the local cluster - [x] Enable/disable host key checking. - [x] Audit Log @capnspacehook - [x] Failed login attempts are recorded - [x] Interactive sessions have the correct Server ID - [x] Server ID is the ID of the node in "session_recording: node" mode - [x] Server ID is the ID of the proxy in "session_recording: proxy" mode Node/Proxy ID may be found at `/var/lib/teleport/host_uuid` in the corresponding machine. Node IDs may also be queried via `tctl nodes ls`. - [x] Exec commands are recorded - [x] `scp` commands are recorded - [x] Subsystem results are recorded Subsystem testing may be achieved using both [Recording Proxy mode]( https://goteleport.com/teleport/docs/architecture/proxy/#recording-proxy-mode) and [OpenSSH integration]( https://goteleport.com/docs/server-access/guides/openssh/). Assuming the proxy is `proxy.example.com:3023` and `node1` is a node running OpenSSH/sshd, you may use the following command to trigger a subsystem audit log: ```shell sftp -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" root@node1 ``` - [x] Interact with a cluster using `tsh` @mdwn These commands should ideally be tested for recording and non-recording modes as they are implemented in a different ways. - [x] tsh ssh \<regular-node\> - [x] tsh ssh \<node-remote-cluster\> - [x] tsh ssh -A \<regular-node\> - [x] tsh ssh -A \<node-remote-cluster\> - [x] tsh ssh \<regular-node\> ls - [x] tsh ssh \<node-remote-cluster\> ls - [x] tsh join \<regular-node\> - [x] tsh join \<node-remote-cluster\> - [x] tsh play \<regular-node\> - [x] tsh play \<node-remote-cluster\> - [x] tsh scp \<regular-node\> - [x] tsh scp \<node-remote-cluster\> - [x] tsh ssh -L \<regular-node\> - [x] tsh ssh -L \<node-remote-cluster\> - [x] tsh ls - [x] tsh clusters - [x] Interact with a cluster using `ssh` @tobiaszheller Make sure to test both recording and regular proxy modes. - [x] ssh \<regular-node\> - [x] ssh \<node-remote-cluster\> - [x] ssh -A \<regular-node\> - [x] ssh -A \<node-remote-cluster\> - [x] ssh \<regular-node\> ls - [x] ssh \<node-remote-cluster\> ls - [x] scp \<regular-node\> - [x] scp \<node-remote-cluster\> - [x] ssh -L \<regular-node\> - [x] ssh -L \<node-remote-cluster\> - [x] Verify proxy jump functionality @Joerger Log into leaf cluster via root, shut down the root proxy and verify proxy jump works. - [x] tls routing disabled - [x] tsh ssh -J \<leaf.proxy.example.com:3023\> - [x] ssh -J \<leaf.proxy.example.com:3023\> - [x] tls routing enabled - [x] tsh ssh -J \<leaf.proxy.example.com:3080\> - [x] tsh proxy ssh -J \<leaf.proxy.example.com:3080\> - [x] Interact with a cluster using the Web UI @capnspacehook - [x] Connect to a Teleport node - [x] Connect to a OpenSSH node - [x] Check agent forwarding is correct based on role and proxy mode. ### User accounting @jakule - [x] Verify that active interactive sessions are tracked in `/var/run/utmp` on Linux. - [ ] Verify that interactive sessions are logged in `/var/log/wtmp` on Linux. ### Combinations @nklaassen For some manual testing, many combinations need to be tested. For example, for interactive sessions the 12 combinations are below. - [x] Connect to a OpenSSH node in a local cluster using OpenSSH. - [x] Connect to a OpenSSH node in a local cluster using Teleport. - [x] Connect to a OpenSSH node in a local cluster using the Web UI. - [x] Connect to a Teleport node in a local cluster using OpenSSH. - [x] Connect to a Teleport node in a local cluster using Teleport. - [x] Connect to a Teleport node in a local cluster using the Web UI. - [x] Connect to a OpenSSH node in a remote cluster using OpenSSH. - [x] Connect to a OpenSSH node in a remote cluster using Teleport. - [x] Connect to a OpenSSH node in a remote cluster using the Web UI. - [x] Connect to a Teleport node in a remote cluster using OpenSSH. - [x] Connect to a Teleport node in a remote cluster using Teleport. - [x] Connect to a Teleport node in a remote cluster using the Web UI. ### Teleport with EKS/GKE @tigrato * [x] Deploy Teleport on a single EKS cluster * [x] Deploy Teleport on two EKS clusters and connect them via trusted cluster feature * [x] Deploy Teleport Proxy outside of GKE cluster fronting connections to it (use [this script](https://github.com/gravitational/teleport/blob/master/examples/k8s-auth/get-kubeconfig.sh) to generate a kubeconfig) * [x] Deploy Teleport Proxy outside of EKS cluster fronting connections to it (use [this script](https://github.com/gravitational/teleport/blob/master/examples/k8s-auth/get-kubeconfig.sh) to generate a kubeconfig) ### Teleport with multiple Kubernetes clusters @AntonAM Note: you can use GKE or EKS or minikube to run Kubernetes clusters. Minikube is the only caveat - it's not reachable publicly so don't run a proxy there. * [x] Deploy combo auth/proxy/kubernetes_service outside of a Kubernetes cluster, using a kubeconfig * [x] Login with `tsh login`, check that `tsh kube ls` has your cluster * [x] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [x] Verify that the audit log recorded the above request and session * [x] Deploy combo auth/proxy/kubernetes_service inside of a Kubernetes cluster * [x] Login with `tsh login`, check that `tsh kube ls` has your cluster * [x] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [x] Verify that the audit log recorded the above request and session * [x] Deploy combo auth/proxy_service outside of the Kubernetes cluster and kubernetes_service inside of a Kubernetes cluster, connected over a reverse tunnel * [x] Login with `tsh login`, check that `tsh kube ls` has your cluster * [x] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` * [x] Verify that the audit log recorded the above request and session * [x] Deploy a second kubernetes_service inside of another Kubernetes cluster, connected over a reverse tunnel * [x] Login with `tsh login`, check that `tsh kube ls` has both clusters * [x] Switch to a second cluster using `tsh kube login` * [x] Run `kubectl get nodes`, `kubectl exec -it $SOME_POD -- sh` on the new cluster * [x] Verify that the audit log recorded the above request and session * [x] Deploy combo auth/proxy/kubernetes_service outside of a Kubernetes cluster, using a kubeconfig with multiple clusters in it * [x] Login with `tsh login`, check that `tsh kube ls` has all clusters * [x] Test Kubernetes screen in the web UI (tab is located on left side nav on dashboard): * [x] Verify that all kubes registered are shown with correct `name` and `labels` * [x] Verify that clicking on a rows connect button renders a dialogue on manual instructions with `Step 2` login value matching the rows `name` column * [x] Verify searching for `name` or `labels` in the search bar works * [x] Verify you can sort by `name` column ### Kubernetes auto-discovery @tigrato * [x] Test Kubernetes auto-discovery: * [x] Verify that Azure AKS clusters are discovered and enrolled for different Azure Auth configs: * [x] Local Accounts only * [x] Azure AD * [x] Azure RBAC * [x] Verify that AWS EKS clusters are discovered and enrolled ### Kubernetes Secret Storage @tigrato * [x] Kubernetes Secret storage for Agent's Identity * [x] Install Teleport 11 agent with a short-lived token * [x] Validate if the Teleport is installed as a Kubernetes `Statefulset` * [x] Restart the agent after token TTL expires to see if it reuses the same identity. (#17474) * [x] Upgrade from Teleport 10 agent with storage * [x] Validate if agent identity was read from storage and stored in the secret without generating a new one * [x] Validate if the Teleport is still running as a `Statefulset` resource and if it contains the new ENV variables * [x] Upgrade from Teleport 10 agent without storage (Failing due to #17437) * [x] Validate if the agent identity is created and stored in the secret using the long-lived token. * [x] Validate if the Teleport Kubernetes `Deployment` was correctly converted into a Statefulset and if the old `Deployment` object was removed after a successful upgrade * [x] Force cluster CA rotation * [x] Test Kubernetes exec via websockets - [client](https://github.com/kubernetes-client/javascript/blob/45b68c98e62b6cc4152189b9fd4a27ad32781bc4/examples/typescript/exec/exec-example.ts) ### Teleport with FIPS mode @alistanis * [x] Perform trusted clusters, Web and SSH sanity check with all teleport components deployed in FIPS mode. ### ACME @alistanis - [x] Teleport can fetch TLS certificate automatically using ACME protocol. ### Migrations @jakule * [ ] Migrate trusted clusters from 10 to 11 * [ ] Migrate auth server on main cluster, then rest of the servers on main cluster SSH should work for both main and old clusters * [ ] Migrate auth server on remote cluster, then rest of the remote cluster SSH should work ### Command Templates When interacting with a cluster, the following command templates are useful: #### OpenSSH ``` # when connecting to the recording proxy, `-o 'ForwardAgent yes'` is required. ssh -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" \ node.example.com # the above command only forwards the agent to the proxy, to forward the agent # to the target node, `-o 'ForwardAgent yes'` needs to be passed twice. ssh -o "ForwardAgent yes" \ -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p" \ node.example.com # when connecting to a remote cluster using OpenSSH, the subsystem request is # updated with the name of the remote cluster. ssh -o "ProxyCommand ssh -o 'ForwardAgent yes' -p 3023 %r@proxy.example.com -s proxy:%h:%p@foo.com" \ node.foo.com ``` #### Teleport ``` # when connecting to a OpenSSH node, remember `-p 22` needs to be passed. tsh --proxy=proxy.example.com --user=<username> --insecure ssh -p 22 node.example.com # an agent can be forwarded to the target node with `-A` tsh --proxy=proxy.example.com --user=<username> --insecure ssh -A -p 22 node.example.com # the --cluster flag is used to connect to a node in a remote cluster. tsh --proxy=proxy.example.com --user=<username> --insecure ssh --cluster=foo.com -p 22 node.foo.com ``` ### Teleport with SSO Providers - [x] G Suite install instructions work @AntonAM - [x] G Suite Screenshots are up to date - [x] Azure Active Directory (AD) install instructions work @alistanis - [x] Azure Active Directory (AD) Screenshots are up to date - [x] ActiveDirectory (ADFS) install instructions work @alistanis - [x] Active Directory (ADFS) Screenshots are up to date - [ ] Okta install instructions work @camscale - [ ] Okta Screenshots are up to date - [x] OneLogin install instructions work @hugoShaka - [ ] OneLogin Screenshots are up to date: https://github.com/gravitational/teleport/issues/17485 - [x] GitLab install instructions work @capnspacehook - [x] GitLab Screenshots are up to date - [ ] OIDC install instructions work @camscale - [ ] OIDC Screenshots are up to date - [ ] All providers with guides in docs are covered in this test plan ### `tctl sso` family of commands @Tener For help with setting up sso connectors, check out the [Quick GitHub/SAML/OIDC Setup Tips] `tctl sso configure` helps to construct a valid connector definition: - [x] `tctl sso configure github ...` creates valid connector definitions - [x] `tctl sso configure oidc ...` creates valid connector definitions - [x] `tctl sso configure saml ...` creates valid connector definitions `tctl sso test` test a provided connector definition, which can be loaded from file or piped in with `tctl sso configure` or `tctl get --with-secrets`. Valid connectors are accepted, invalid are rejected with sensible error messages. - [x] Connectors can be tested with `tctl sso test`. - [x] GitHub - [x] SAML - [x] OIDC - [x] Google Workspace - [x] Non-Google IdP ### Teleport Plugins @hugoShaka - [x] Test receiving a message via Teleport Slackbot - [x] Test receiving a new Jira Ticket via Teleport Jira ### AWS Node Joining @nklaassen [Docs](https://goteleport.com/docs/setup/guides/joining-nodes-aws/) - [x] On EC2 instance with `ec2:DescribeInstances` permissions for local account: `TELEPORT_TEST_EC2=1 go test ./integration -run TestEC2NodeJoin` - [x] On EC2 instance with any attached role: `TELEPORT_TEST_EC2=1 go test ./integration -run TestIAMNodeJoin` - [x] EC2 Join method in IoT mode with node and auth in different AWS accounts - [x] IAM Join method in IoT mode with node and auth in different AWS accounts ### Passwordless @codingllama Passwordless requires `tsh` compiled with libfido2 for most operations (apart from Touch ID). Ask for a statically-built `tsh` binary for realistic tests. Touch ID requires a properly built and signed `tsh` binary. Ask for a pre-release binary, so you may run the tests. This sections complements "Users -> Managing MFA devices". `tsh` binaries for each operating system (Linux, macOS and Windows) must be tested separately for FIDO2 items. - [x] Diagnostics Commands should pass all tests. - [x] `tsh fido2 diag` (macOS/Linux) - [x] `tsh touchid diag` (macOS only) - [x] `tsh webauthnwin diag` (Windows only) - [x] Registration - [x] Register a passworldess FIDO2 key (`tsh mfa add`, choose WEBAUTHN and passwordless) - [x] macOS/Linux - [x] Windows - [x] Register a platform authenticator - [x] Touch ID credential (`tsh mfa add`, choose TOUCHID) - [x] Windows hello credential (`tsh mfa add`, choose WEBAUTHN and passwordless) - [x] Login - [x] Passwordless login using FIDO2 (`tsh login --auth=passwordless`) - [x] macOS/Linux - [x] Windows - [x] Passwordless login using platform authenticator (`tsh login --auth=passwordless`) - [x] Touch ID - [x] Windows Hello - [x] `tsh login --auth=passwordless --mfa-mode=cross-platform` uses FIDO2 - [x] macOS/Linux - [x] Windows - [x] `tsh login --auth=passwordless --mfa-mode=platform` uses platform authenticator - [x] Touch ID - [x] Windows Hello - [x] `tsh login --auth=passwordless --mfa-mode=auto` prefers platform authenticator - [x] Touch ID - [x] Windows Hello - [x] Passwordless disable switch works (`auth_service.authentication.passwordless = false`) - [x] Cluster in passwordless mode defaults to passwordless (`auth_service.authentication.connector_name = passwordless`) - [x] Cluster in passwordless mode allows MFA login (`tsh login --auth=local`) - [x] Touch ID support commands - [x] `tsh touchid ls` works - [x] `tsh touchid rm` works (careful, may lock you out!) ### Hardware Key Support @Joerger Hardware Key Support is an Enterprise feature and is not available for OSS. You will need a YubiKey 4.3+ to test this feature. This feature has additional build requirements, so it should be tested with a pre-release build from Drone (eg: `https://get.gravitational.com/teleport-ent-v11.0.0-alpha.2-linux-amd64-bin.tar.gz`). These tests should be carried out sequentially. `tsh` tests should be carried out on Linux, MacOS, and Windows. - [x] `tsh login` as user with [Webauthn](https://goteleport.com/docs/access-controls/guides/webauthn/) login and no hardware key requirement. - [x] Request a role with `role.role_options.require_session_mfa: hardware_key` - `tsh login --request-roles=hardware_key_required` - [x] Assuming the role should force automatic re-login with yubikey - [x] `tsh ssh` - [x] Requires yubikey to be connected for re-login - [x] Prompts for per-session MFA - [x] Request a role with `role.role_options.require_session_mfa: hardware_key_touch` - `tsh login --request-roles=hardware_key_touch_required` - [x] Assuming the role should force automatic re-login with yubikey - [x] Prompts for touch if not cached (last touch within 15 seconds) - [x] `tsh ssh` - [x] Requires yubikey to be connected for re-login - [x] Prompts for touch if not cached - [x] `tsh logout` and `tsh login` as the user with no hardware key requirement. - [x] Upgrade auth settings to `auth_service.authentication.require_session_mfa: hardware_key` - [x] Using the existing login session (`tsh ls`) should force automatic re-login with yubikey - [x] `tsh ssh` - [x] Requires yubikey to be connected for re-login - [x] Prompts for per-session MFA - [x] Upgrade auth settings to `auth_service.authentication.require_session_mfa: hardware_key_touch` - [x] Using the existing login session (`tsh ls`) should force automatic re-login with yubikey - [x] Prompts for touch if not cached - [x] `tsh ssh` - [x] Requires yubikey to be connected for re-login - [x] Prompts for touch if not cached ## Performance @rosstimothy @fspmarshall Perform all tests on the following configurations: - [x] With default networking configuration - [x] With Proxy Peering Enabled - [x] With TLS Routing Enabled * Cluster with 10K direct dial nodes: - [x] etcd - [x] DynamoDB - [ ] Firestore * Cluster with 10K reverse tunnel nodes: - [x] etcd - [x] DynamoDB - [ ] Firestore * Cluster with 500 trusted clusters: - [x] etcd - [x] DynamoDB - [ ] Firestore ### Soak Test Run 30 minute soak test with a mix of interactive/non-interactive sessions for both direct and reverse tunnel nodes: ```shell tsh bench --duration=30m user@direct-dial-node ls tsh bench -i --duration=30m user@direct-dial-node ps uax tsh bench --duration=30m user@reverse-tunnel-node ls tsh bench -i --duration=30m user@reverse-tunnel-node ps uax ``` Observe prometheus metrics for goroutines, open files, RAM, CPU, Timers and make sure there are no leaks - [x] Verify that prometheus metrics are accurate. ### Concurrent Session Test * Cluster with 1k reverse tunnel nodes Run a concurrent session test that will spawn 5 interactive sessions per node in the cluster: ```shell tsh bench sessions --max=5000 user ls tsh bench sessions --max=5000 --web user ls ``` - [x] Verify that all 5000 sessions are able to be established. - [ ] Verify that tsh and the web UI are still functional. ## Teleport with Cloud Providers ### AWS @hugoShaka - [x] Deploy Teleport to AWS. Using DynamoDB & S3 - [x] Deploy Teleport Enterprise to AWS. Using HA Setup https://gravitational.com/teleport/docs/aws-terraform-guide/ ### GCP @AntonAM - [x] Deploy Teleport to GCP. Using Cloud Firestore & Cloud Storage - [x] Deploy Teleport to GKE. Google Kubernetes engine. - [x] Deploy Teleport Enterprise to GCP. ### IBM @atburke - [x] Deploy Teleport to IBM Cloud. Using IBM Database for etcd & IBM Object Store - [x] Deploy Teleport to IBM Cloud Kubernetes. - [x] Deploy Teleport Enterprise to IBM Cloud. ## Application Access @mdwn - [x] Run an application within local cluster. - [x] Verify the debug application `debug_app: true` works. - [x] Verify an application can be configured with command line flags. - [x] Verify an application can be configured from file configuration. - [x] Verify that applications are available at auto-generated addresses `name.rootProxyPublicAddr` and well as `publicAddr`. - [x] Run an application within a trusted cluster. - [x] Verify that applications are available at auto-generated addresses `name.rootProxyPublicAddr`. - [x] Verify Audit Records. - [x] `app.session.start` and `app.session.chunk` events are created in the Audit Log. - [x] `app.session.chunk` points to a 5 minute session archive with multiple `app.session.request` events inside. - [x] `tsh play <chunk-id>` can fetch and print a session chunk archive. - [x] Verify JWT using [verify-jwt.go](https://github.com/gravitational/teleport/blob/master/examples/jwt/verify-jwt.go). - [x] Verify RBAC. - [x] Verify [CLI access](https://goteleport.com/docs/application-access/guides/api-access/) with `tsh app login`. - [x] Verify AWS console access. - [x] Can log into AWS web console through the web UI. - [x] Can interact with AWS using `tsh aws` commands. - [x] Verify dynamic registration. - [x] Can register a new app using `tctl create`. - [x] Can update registered app using `tctl create -f`. - [x] Can delete registered app using `tctl rm`. - [x] Test Applications screen in the web UI (tab is located on left side nav on dashboard): - [x] Verify that all apps registered are shown - [x] Verify that clicking on the app icon takes you to another tab - [x] Verify using the bash command produced from `Add Application` dialogue works (refresh app screen to see it registered) ## Database Access @smallinsky + db access team - [x] Connect to a database within a local cluster. - [x] Self-hosted Postgres. @Tener - [x] Self-hosted MySQL. @GavinFrazar - [x] Self-hosted MariaDB. @GavinFrazar - [x] Self-hosted MongoDB. @Tener - [x] Self-hosted CockroachDB. @GavinFrazar - [x] Self-hosted Redis. @greedy52 - [x] Self-hosted Redis Cluster. @greedy52 - [x] Self-hosted MSSQL. @smallinsky - [x] AWS Aurora Postgres. @gabrielcorado - [x] AWS Aurora MySQL. @Tener - [x] AWS Redshift. @gabrielcorado - [x] AWS ElastiCache. @greedy52 - [x] AWS MemoryDB. @greedy52 - [x] GCP Cloud SQL Postgres. @gabrielcorado - [x] GCP Cloud SQL MySQL. @Tener - [x] Snowflake. @smallinsky - [x] Azure Cache for Redis. @smallinsky - [x] Elasticsearch. @greedy52 - [x] Cassandra/ScyllaDB. @gabrielcorado - [x] Connect to a database within a remote cluster via a trusted cluster. - [x] Self-hosted Postgres. @Tener - [x] Self-hosted MySQL. @GavinFrazar - [x] Self-hosted MariaDB. @GavinFrazar - [x] Self-hosted MongoDB. @Tener - [x] Self-hosted CockroachDB. @GavinFrazar - [x] Self-hosted Redis. @greedy52 - [x] Self-hosted Redis Cluster. @greedy52 - [x] Self-hosted MSSQL. @smallinsky - [x] AWS Aurora Postgres. @gabrielcorado - [x] AWS Aurora MySQL. @Tener - [x] AWS Redshift. @gabrielcorado - [x] AWS ElastiCache. @greedy52 - [x] AWS MemoryDB. @greedy52 - [x] GCP Cloud SQL Postgres. @gabrielcorado - [x] GCP Cloud SQL MySQL. @Tener - [x] Snowflake. @smallinsky - [x] Azure Cache for Redis. @smallinsky - [x] Elasticsearch. @greedy52 - [x] Cassandra/ScyllaDB. @gabrielcorado - [x] Verify audit events. @GavinFrazar - [x] `db.session.start` is emitted when you connect. - [x] `db.session.end` is emitted when you disconnect. - [x] `db.session.query` is emitted when you execute a SQL query. - [x] Verify RBAC. - [x] `tsh db ls` shows only databases matching role's `db_labels`. @gabrielcorado - [x] Can only connect as users from `db_users`. @gabrielcorado - [x] _(Postgres only)_ Can only connect to databases from `db_names`. @gabrielcorado - [x] `db.session.start` is emitted when connection attempt is denied. - [x] _(MongoDB only)_ Can only execute commands in databases from `db_names`. @gabrielcorado - [x] `db.session.query` is emitted when command fails due to permissions. - [x] Can configure per-session MFA. @GavinFrazar - [x] MFA tap is required on each `tsh db connect`. - [x] Verify dynamic registration. @GavinFrazar - [x] Can register a new database using `tctl create`. - [x] Can update registered database using `tctl create -f`. - [x] Can delete registered database using `tctl rm`. - [x] Verify discovery. - [x] AWS - [x] Can detect and register RDS instances. @GavinFrazar - [x] Can detect and register Aurora clusters, and their reader and custom endpoints. @gabrielcorado - [x] Can detect and register Redshift clusters. @smallinsky - [x] Can detect and register ElastiCache Redis clusters. @greedy52 - [x] Can detect and register MemoryDB clusters. @greedy52 - [x] Azure - [x] Can detect and register MySQL and Postgres instances. @GavinFrazar - [x] Can detect and register Azure Cache for Redis servers. @smallinsky - [x] Verify Teleport managed users (password rotation, auto 'auth' on connection, etc.). @greedy52 - [x] Can detect and manage ElastiCache users - [x] Can detect and manage MemoryDB users - [x] Test Databases screen in the web UI (tab is located on left side nav on dashboard): @Tener - [x] Verify that all dbs registered are shown with correct `name`, `description`, `type`, and `labels` - [x] Verify that clicking on a rows connect button renders a dialogue on manual instructions with `Step 2` login value matching the rows `name` column - [x] Verify searching for all columns in the search bar works - [x] Verify you can sort by all columns except `labels` - [x] Other - [x] MySQL server version reported by Teleport is correct. @smallinsky ## TLS Routing @smallinsky - [x] Verify that teleport proxy `v2` configuration starts only a single listener. @smallinsky ``` version: v2 teleport: proxy_service: enabled: "yes" public_addr: ['root.example.com'] web_listen_addr: 0.0.0.0:3080 ``` - [x] Run Teleport Proxy in `multiplex` mode `auth_service.proxy_listener_mode: "multiplex"` @smallinsky - [x] Trusted cluster - [x] Setup trusted clusters using single port setup `web_proxy_addr == tunnel_addr` ``` kind: trusted_cluster spec: ... web_proxy_addr: root.example.com:443 tunnel_addr: root.example.com:443 ... ``` - [ ] Database Access - [x] Verify that `tsh db connect` works through proxy running in `multiplex` mode - [x] Postgres @Tener - [x] MySQL @GavinFrazar - [x] MariaDB @GavinFrazar - [x] MongoDB @Tener - [x] CockroachDB @GavinFrazar - [x] Redis @greedy52 - [x] MSSQL @smallinsky - [x] Snowflake @smallinsky - [x] Elasticsearch. @greedy52 - [x] Cassandra/ScyllaDB. @gabrielcorado - [ ] Verify connecting to a database through TLS ALPN SNI local proxy `tsh proxy db` with a GUI client. @smallinsky @GavinFrazar @greedy52 @Tener @gabrielcorado - [x] Verify tsh proxy db with teleport proxy behind ALB. @smallinsky - [x] Application Access @GavinFrazar - [x] Verify app access through proxy running in `multiplex` mode - [x] SSH Access @gabrielcorado - [x] Connect to a OpenSSH server through a local ssh proxy `ssh -o "ForwardAgent yes" -o "ProxyCommand tsh proxy ssh" user@host.example.com` - [x] Connect to a OpenSSH server on leaf-cluster through a local ssh proxy`ssh -o "ForwardAgent yes" -o "ProxyCommand tsh proxy ssh --user=%r --cluster=leaf-cluster %h:%p" user@node.foo.com` - [x] Verify `tsh ssh` access through proxy running in multiplex mode - [x] Kubernetes access: @GavinFrazar - [x] Verify kubernetes access through proxy running in `multiplex` mode ## Desktop Access @ibeckermayer @probakowski @LKozlowski - Direct mode (set `listen_addr`): - [x] Can connect to desktop defined in static `hosts` section. - [x] Can connect to desktop discovered via LDAP - IoT mode (reverse tunnel through proxy): - [x] Can connect to desktop defined in static `hosts` section. - [x] Can connect to desktop discovered via LDAP - [x] Connect multiple `windows_desktop_service`s to the same Teleport cluster, verify that connections to desktops on different AD domains works. (Attempt to connect several times to verify that you are routed to the correct `windows_desktop_service`) - Verify user input - [x] Download [Keyboard Key Info](https://dennisbabkin.com/kbdkeyinfo/) and verify all keys are processed correctly in each supported browser. Known issues: F11 cannot be captured by the browser without [special configuration](https://social.technet.microsoft.com/Forums/en-US/784b2bbe-353f-412e-ac9a-193d81f306b6/remote-desktop-for-mac-f11-key-not-working-on-macbook-pro-touchbar?forum=winRDc) on MacOS. - [x] Left click and right click register as Windows clicks. (Right click on the desktop should show a Windows menu, not a browser context menu) - [x] Vertical and horizontal scroll work. [Horizontal Scroll Test](https://codepen.io/jaemskyle/pen/inbmB) - Locking - [x] Verify that placing a user lock terminates an active desktop session. - [x] Verify that placing a desktop lock terminates an active desktop session. - [x] Verify that placing a role lock terminates an active desktop session. - Labeling - [x] Set `client_idle_timeout` to a small value and verify that idle sessions are terminated (the session should end and an audit event will confirm it was due to idle connection) - [x] All desktops have `teleport.dev/origin` label. - [x] Dynamic desktops have additional `teleport.dev` labels for OS, OS Version, DNS hostname. - [x] Regexp-based host labeling applies across all desktops, regardless of origin. - RBAC - [x] RBAC denies access to a Windows desktop due to labels - [x] RBAC denies access to a Windows desktop with the wrong OS-login. - Clipboard Support - When a user has a role with clipboard sharing enabled and is using a chromium based browser - [x] Going to a desktop when clipboard permissions are in "Ask" mode (aka "prompt") causes the browser to show a prompt while the UI shows a spinner - [x] X-ing out of the prompt (causing the clipboard permission to remain in "Ask" mode) causes the prompt to show up again - [x] Denying clipboard permissions brings up a relevant error alert (with "Clipboard Sharing Disabled" in the top bar) - [x] Allowing clipboard permissions allows you to see the desktop session, with "Clipboard Sharing Enabled" highlighted in the top bar - [x] Copy text from local workstation, paste into remote desktop - [x] Copy text from remote desktop, paste into local workstation - When a user has a role with clipboard sharing enabled and is *not* using a chromium based browser - [x] The UI shows a relevant alert and "Clipboard Sharing Disabled" is highlighted in the top bar - When a user has a role with clipboard sharing *disabled* and is using a chromium and non-chromium based browser (confirm both) - [x] The live session should show disabled in the top bar and copy/paste should not work between your workstation and the remote desktop. - Directory Sharing - On supported, non-chromium based browsers (Firefox/Safari) - [x] Attempting to share directory shows a dismissible "Unsupported Action" dialog - On supported, chromium based browsers (Chrome/Edge) - Begin sharing works - [x] The shared directory icon in the top right of the screen is highlighted when directory sharing is initiated - [x] The shared directory appears as a network drive named "<directory_name> on teleport" - [x] The share directory menu option dissapears from the menu - Navigation - [x] The folders of the shared directory are navigable (move up and down the directory tree) - CRUD - [x] A new text file can be created - [x] The text file can be written to (saved) - [x] The text file can be read (close it, check that its saved on the local machine, then open it again on the remote) - [x] The text file can be deleted - File/Folder movement - In to out (make at least one of these from a non-top-level-directory) - [x] A file from inside the shared directory can be drag-and-dropped outside the shared directory - [x] A folder from inside the shared directory can be drag-and-dropped outside the shared directory (and its contents retained) - [x] A file from inside the shared directory can be cut-pasted outside the shared directory - [x] A folder from inside the shared directory can be cut-pasted outside the shared directory - [x] A file from inside the shared directory can be copy-pasted outside the shared directory - [x] A folder from inside the shared directory can be copy-pasted outside the shared directory - Out to in (make at least one of these overwrite an existing file, and one go into a non-top-level directory) - [x] A file from outside the shared directory can be drag-and-dropped into the shared directory - [x] A folder from outside the shared directory can be drag-and-dropped into the shared directory (and its contents retained) - [x] A file from outside the shared directory can be cut-pasted into the shared directory - [x] A folder from outside the shared directory can be cut-pasted into the shared directory - [x] A file from outside the shared directory can be copy-pasted into the shared directory - [x] A folder from outside the shared directory can be copy-pasted into the shared directory - Within - [x] A file from inside the shared directory cannot be drag-and-dropped to another folder inside the shared directory: a dismissible "Unsupported Action" dialog is shown - [x] A folder from inside the shared directory cannot be drag-and-dropped to another folder inside the shared directory: a dismissible "Unsupported Action" dialog is shown - [x] A file from inside the shared directory cannot be cut-pasted to another folder inside the shared directory: a dismissible "Unsupported Action" dialog is shown - [x] A folder from inside the shared directory cannot be cut-pasted to another folder inside the shared directory: a dismissible "Unsupported Action" dialog is shown - [x] A file from inside the shared directory can be copy-pasted to another folder inside the shared directory - [x] A folder from inside the shared directory can be copy-pasted to another folder inside shared directory (and its contents retained) - RBAC - [x] Give the user one role that explicitly disables directory sharing (`desktop_directory_sharing: false`) and confirm that the option to share a directory doesn't appear in the menu - Per-Session MFA (try webauthn on each of Chrome, Safari, and Firefox; u2f only works with Firefox) - [x] Attempting to start a session no keys registered shows an error message - [x] ~Attempting to start a session with a u2f key registered shows an error message~ (N/A now that u2f support has been removed) - [x] Attempting to start a session with a webauthn registered pops up the "Verify Your Identity" dialog - [x] Hitting "Cancel" shows an error message - [x] Hitting "Verify" causes your browser to prompt you for MFA - [x] Cancelling that browser MFA prompt shows an error - [x] Successful MFA verification allows you to connect - Session Recording - [x] Verify sessions are not recorded if *all* of a user's roles disable recording - [x] Verify sync recording (`mode: node-sync` or `mode: proy-sync`) - [x] Verify async recording (`mode: node` or `mode: proxy`) - [x] Sessions show up in session recordings UI with desktop icon - [x] Sessions can be played back, including play/pause functionality - [x] A session that ends with a TDP error message can be played back, ends by displaying the error message, and the progress bar progresses to the end. - [x] Attempting to play back a session that doesn't exist (i.e. by entering a non-existing session id in the url) shows a relevant error message. - [x] RBAC for sessions: ensure users can only see their own recordings when using the RBAC rule from our [docs](../../docs/pages/access-controls/reference.mdx#rbac-for-sessions) - Audit Events (check these after performing the above tests) - [x] `windows.desktop.session.start` (`TDP00I`) emitted on start - [x] `windows.desktop.session.start` (`TDP00W`) emitted when session fails to start (due to RBAC, for example) - [x] `windows.desktop.session.end` (`TDP01I`) emitted on end - [x] `desktop.clipboard.send` (`TDP02I`) emitted for local copy -> remote paste - [x] `desktop.clipboard.receive` (`TDP03I`) emitted for remote copy -> local paste ## Binaries compatibility @fheinecke - Verify that teleport/tsh/tctl/tbot run on: - [ ] CentOS 7 - [ ] CentOS 8 - [ ] Ubuntu 18.04 - [ ] Ubuntu 20.04 - [ ] Debian 9 - Verify tsh runs on: - [ ] Windows 10 - [ ] MacOS ## Machine ID @timothyb89 ### SSH With a default Teleport instance configured with a SSH node: - [x] Verify you are able to create a new bot user with `tctl bots add robot --roles=access`. Follow the instructions provided in the output to start `tbot` - [x] Verify you are able to connect to the SSH node using openssh with the generated `ssh_config` in the destination directory - [x] Verify that after the renewal period (default 20m, but this can be reduced via configuration), that newly generated certificates are placed in the destination directory - [x] Verify that sending both `SIGUSR1` and `SIGHUP` to a running tbot process causes a renewal and new certificates to be generated - [x] Verify that you are able to make a connection to the SSH node using the `ssh_config` provided by `tbot` after each phase of a manual CA rotation. Ensure the above tests are completed for both: - [x] Directly connecting to the auth server - [x] Connecting to the auth server via the proxy reverse tunnel ### DB Access With a default Postgres DB instance, a Teleport instance configured with DB access and a bot user configured: - [x] Verify you are able to connect to and interact with a database using `tbot db` while `tbot start` is running ## Host users creation @lxea [Host users creation docs](https://github.com/gravitational/teleport/pull/13056) [Host users creation RFD](https://github.com/gravitational/teleport/pull/11077) <!--- TODO(lxea): replace links with actual docs once merged [Host users creation docs](../../docs/pages/server-access/guides/host-user-creation.mdx) [Host users creation RFD](../../rfd/0057-automatic-user-provisioning.md) --> - Verify host users creation functionality - [x] non-existing users are created automatically - [x] users are added to groups - [x] non existing configured groups are created - [x] created users are added to the `teleport-system` group - [x] users are cleaned up after their session ends - [x] cleanup occurs if a program was left running after session ends - [x] sudoers file creation is successful - [x] Invalid sudoers files are _not_ created - [x] existing host users are not modified - [x] setting `disable_create_host_user: true` stops user creation from occurring ## CA rotations @espadolini - Verify the CA rotation functionality itself (by checking in the backend or with `tctl get cert_authority`) - [x] `standby` phase: only `active_keys`, no `additional_trusted_keys` - [x] `init` phase: `active_keys` and `additional_trusted_keys` - [x] `update_clients` and `update_servers` phases: the certs from the `init` phase are swapped - [x] `standby` phase: only the new certs remain in `active_keys`, nothing in `additional_trusted_keys` - [x] `rollback` phase (second pass, after completing a regular rotation): same content as in the `init` phase - [x] `standby` phase after `rollback`: same content as in the previous `standby` phase - Verify functionality in all phases (clients might have to log in again in lieu of waiting for credentials to expire between phases) - [x] SSH session in tsh from a previous phase - [x] SSH session in web UI from a previous phase - [x] New SSH session with tsh - [x] New SSH session with web UI - [x] New SSH session in a child cluster on the same major version - [x] New SSH session in a child cluster on the previous major version - [x] New SSH session from a parent cluster - [x] Application access through a browser - [x] Application access through curl with `tsh app login` - [x] `kubectl get po` after `tsh kube login` - [ ] Database access (no configuration change should be necessary if the database CA isn't rotated, other Teleport functionality should not be affected if only the database CA is rotated) ## EC2 Discovery @lxea [EC2 Discovery docs](https://goteleport.com/docs/ver/11.0/server-access/guides/ec2-discovery/) - Verify EC2 instance discovery - [x] Only EC2 instances matching given AWS tags have the installer executed on them - [x] Only the IAM permissions mentioned in the discovery docs are required for operation - [x] Custom scripts specified in different matchers are executed - [x] Custom SSM documents specified in different matchers are executed - [x] New EC2 instances with matching AWS tags are discovered and added to the teleport cluster - [x] Large numbers of EC2 instances (51+) are all successfully added to the cluster - [x] Nodes that have been discovered do not have the install script run on the node multiple times ## Resources [Quick GitHub/SAML/OIDC Setup Tips] <!--- reference style links --> [Quick GitHub/SAML/OIDC Setup Tips]: https://gravitational.slab.com/posts/quick-git-hub-saml-oidc-setup-6dfp292a
test
teleport test plan manual testing plan below are the items that should be manually tested with each release of teleport these tests should be run on both a fresh install of the version to be released as well as an upgrade of the previous version of teleport adding nodes to a cluster edwarddowling adding nodes via valid static token adding nodes via valid short lived tokens adding nodes via invalid token fails revoking node invitation labels edwarddowling static labels dynamic labels trusted clusters lxea adding trusted cluster valid static token adding trusted cluster valid short lived token adding trusted cluster invalid token removing trusted cluster rbac atburke make sure that invalid and valid attempts are reflected in audit log successfully connect to node with correct role unsuccessfully connect to a node in a role restricting access by label unsuccessfully connect to a node in a role restricting access by invalid ssh login allow deny role option ssh agent forwarding allow deny role option port forwarding verify that custom pam environment variables are available as expected jakule users codingllama with every user combination try to login and signup with invalid second factor invalid password to see how the system reacts webauthn in the release tsh binary is implemented using for linux macos ask for a statically built pre release binary for realistic tests tsh diag should work in our binary webauthn in windows build is implemented using webauthn dll tsh webauthn diag with security key selected in dialog should work touch id requires a signed tsh ask for a signed pre release binary so you may run the tests windows webauthn requires windows and device capable of windows hello adding users password only adding users otp adding users webauthn macos linux windows adding users via platform authenticator touch id windows hello managing mfa devices add an otp device with tsh mfa add add a webauthn device with tsh mfa add macos linux windows add platform authenticator device with tsh mfa add touch id windows hello list mfa devices with tsh mfa ls remove an otp device with tsh mfa rm remove a webauthn device with tsh mfa rm attempt removing the last mfa device on the user with second factor on in auth service should fail with second factor optional in auth service should succeed login password only login with mfa add an otp a webauthn and a touch id device with tsh mfa add login via otp login via webauthn macos linux windows login via platform authenticator touch id windows hello login via webauthn using an device devices must be registered in a previous version of teleport using teleport set auth service authentication second factor restart the server and then register an device tsh mfa add upgrade the install to the current teleport version one major at a time and try to login using the device as your second factor it should work deleting users sso camscale login oidc login saml login github backends joerger teleport runs with etcd teleport runs with dynamodb teleport runs with sqlite teleport runs with firestore session recording strideynet session recording can be disabled sessions can be recorded at the node sessions in remote clusters are recorded in remote clusters sessions can be recorded at the proxy sessions on remote clusters are recorded in the local cluster enable disable host key checking audit log capnspacehook failed login attempts are recorded interactive sessions have the correct server id server id is the id of the node in session recording node mode server id is the id of the proxy in session recording proxy mode node proxy id may be found at var lib teleport host uuid in the corresponding machine node ids may also be queried via tctl nodes ls exec commands are recorded scp commands are recorded subsystem results are recorded subsystem testing may be achieved using both and assuming the proxy is proxy example com and is a node running openssh sshd you may use the following command to trigger a subsystem audit log shell sftp o proxycommand ssh o forwardagent yes p r proxy example com s proxy h p root interact with a cluster using tsh mdwn these commands should ideally be tested for recording and non recording modes as they are implemented in a different ways tsh ssh tsh ssh tsh ssh a tsh ssh a tsh ssh ls tsh ssh ls tsh join tsh join tsh play tsh play tsh scp tsh scp tsh ssh l tsh ssh l tsh ls tsh clusters interact with a cluster using ssh tobiaszheller make sure to test both recording and regular proxy modes ssh ssh ssh a ssh a ssh ls ssh ls scp scp ssh l ssh l verify proxy jump functionality joerger log into leaf cluster via root shut down the root proxy and verify proxy jump works tls routing disabled tsh ssh j ssh j tls routing enabled tsh ssh j tsh proxy ssh j interact with a cluster using the web ui capnspacehook connect to a teleport node connect to a openssh node check agent forwarding is correct based on role and proxy mode user accounting jakule verify that active interactive sessions are tracked in var run utmp on linux verify that interactive sessions are logged in var log wtmp on linux combinations nklaassen for some manual testing many combinations need to be tested for example for interactive sessions the combinations are below connect to a openssh node in a local cluster using openssh connect to a openssh node in a local cluster using teleport connect to a openssh node in a local cluster using the web ui connect to a teleport node in a local cluster using openssh connect to a teleport node in a local cluster using teleport connect to a teleport node in a local cluster using the web ui connect to a openssh node in a remote cluster using openssh connect to a openssh node in a remote cluster using teleport connect to a openssh node in a remote cluster using the web ui connect to a teleport node in a remote cluster using openssh connect to a teleport node in a remote cluster using teleport connect to a teleport node in a remote cluster using the web ui teleport with eks gke tigrato deploy teleport on a single eks cluster deploy teleport on two eks clusters and connect them via trusted cluster feature deploy teleport proxy outside of gke cluster fronting connections to it use to generate a kubeconfig deploy teleport proxy outside of eks cluster fronting connections to it use to generate a kubeconfig teleport with multiple kubernetes clusters antonam note you can use gke or eks or minikube to run kubernetes clusters minikube is the only caveat it s not reachable publicly so don t run a proxy there deploy combo auth proxy kubernetes service outside of a kubernetes cluster using a kubeconfig login with tsh login check that tsh kube ls has your cluster run kubectl get nodes kubectl exec it some pod sh verify that the audit log recorded the above request and session deploy combo auth proxy kubernetes service inside of a kubernetes cluster login with tsh login check that tsh kube ls has your cluster run kubectl get nodes kubectl exec it some pod sh verify that the audit log recorded the above request and session deploy combo auth proxy service outside of the kubernetes cluster and kubernetes service inside of a kubernetes cluster connected over a reverse tunnel login with tsh login check that tsh kube ls has your cluster run kubectl get nodes kubectl exec it some pod sh verify that the audit log recorded the above request and session deploy a second kubernetes service inside of another kubernetes cluster connected over a reverse tunnel login with tsh login check that tsh kube ls has both clusters switch to a second cluster using tsh kube login run kubectl get nodes kubectl exec it some pod sh on the new cluster verify that the audit log recorded the above request and session deploy combo auth proxy kubernetes service outside of a kubernetes cluster using a kubeconfig with multiple clusters in it login with tsh login check that tsh kube ls has all clusters test kubernetes screen in the web ui tab is located on left side nav on dashboard verify that all kubes registered are shown with correct name and labels verify that clicking on a rows connect button renders a dialogue on manual instructions with step login value matching the rows name column verify searching for name or labels in the search bar works verify you can sort by name column kubernetes auto discovery tigrato test kubernetes auto discovery verify that azure aks clusters are discovered and enrolled for different azure auth configs local accounts only azure ad azure rbac verify that aws eks clusters are discovered and enrolled kubernetes secret storage tigrato kubernetes secret storage for agent s identity install teleport agent with a short lived token validate if the teleport is installed as a kubernetes statefulset restart the agent after token ttl expires to see if it reuses the same identity upgrade from teleport agent with storage validate if agent identity was read from storage and stored in the secret without generating a new one validate if the teleport is still running as a statefulset resource and if it contains the new env variables upgrade from teleport agent without storage failing due to validate if the agent identity is created and stored in the secret using the long lived token validate if the teleport kubernetes deployment was correctly converted into a statefulset and if the old deployment object was removed after a successful upgrade force cluster ca rotation test kubernetes exec via websockets teleport with fips mode alistanis perform trusted clusters web and ssh sanity check with all teleport components deployed in fips mode acme alistanis teleport can fetch tls certificate automatically using acme protocol migrations jakule migrate trusted clusters from to migrate auth server on main cluster then rest of the servers on main cluster ssh should work for both main and old clusters migrate auth server on remote cluster then rest of the remote cluster ssh should work command templates when interacting with a cluster the following command templates are useful openssh when connecting to the recording proxy o forwardagent yes is required ssh o proxycommand ssh o forwardagent yes p r proxy example com s proxy h p node example com the above command only forwards the agent to the proxy to forward the agent to the target node o forwardagent yes needs to be passed twice ssh o forwardagent yes o proxycommand ssh o forwardagent yes p r proxy example com s proxy h p node example com when connecting to a remote cluster using openssh the subsystem request is updated with the name of the remote cluster ssh o proxycommand ssh o forwardagent yes p r proxy example com s proxy h p foo com node foo com teleport when connecting to a openssh node remember p needs to be passed tsh proxy proxy example com user insecure ssh p node example com an agent can be forwarded to the target node with a tsh proxy proxy example com user insecure ssh a p node example com the cluster flag is used to connect to a node in a remote cluster tsh proxy proxy example com user insecure ssh cluster foo com p node foo com teleport with sso providers g suite install instructions work antonam g suite screenshots are up to date azure active directory ad install instructions work alistanis azure active directory ad screenshots are up to date activedirectory adfs install instructions work alistanis active directory adfs screenshots are up to date okta install instructions work camscale okta screenshots are up to date onelogin install instructions work hugoshaka onelogin screenshots are up to date gitlab install instructions work capnspacehook gitlab screenshots are up to date oidc install instructions work camscale oidc screenshots are up to date all providers with guides in docs are covered in this test plan tctl sso family of commands tener for help with setting up sso connectors check out the tctl sso configure helps to construct a valid connector definition tctl sso configure github creates valid connector definitions tctl sso configure oidc creates valid connector definitions tctl sso configure saml creates valid connector definitions tctl sso test test a provided connector definition which can be loaded from file or piped in with tctl sso configure or tctl get with secrets valid connectors are accepted invalid are rejected with sensible error messages connectors can be tested with tctl sso test github saml oidc google workspace non google idp teleport plugins hugoshaka test receiving a message via teleport slackbot test receiving a new jira ticket via teleport jira aws node joining nklaassen on instance with describeinstances permissions for local account teleport test go test integration run on instance with any attached role teleport test go test integration run testiamnodejoin join method in iot mode with node and auth in different aws accounts iam join method in iot mode with node and auth in different aws accounts passwordless codingllama passwordless requires tsh compiled with for most operations apart from touch id ask for a statically built tsh binary for realistic tests touch id requires a properly built and signed tsh binary ask for a pre release binary so you may run the tests this sections complements users managing mfa devices tsh binaries for each operating system linux macos and windows must be tested separately for items diagnostics commands should pass all tests tsh diag macos linux tsh touchid diag macos only tsh webauthnwin diag windows only registration register a passworldess key tsh mfa add choose webauthn and passwordless macos linux windows register a platform authenticator touch id credential tsh mfa add choose touchid windows hello credential tsh mfa add choose webauthn and passwordless login passwordless login using tsh login auth passwordless macos linux windows passwordless login using platform authenticator tsh login auth passwordless touch id windows hello tsh login auth passwordless mfa mode cross platform uses macos linux windows tsh login auth passwordless mfa mode platform uses platform authenticator touch id windows hello tsh login auth passwordless mfa mode auto prefers platform authenticator touch id windows hello passwordless disable switch works auth service authentication passwordless false cluster in passwordless mode defaults to passwordless auth service authentication connector name passwordless cluster in passwordless mode allows mfa login tsh login auth local touch id support commands tsh touchid ls works tsh touchid rm works careful may lock you out hardware key support joerger hardware key support is an enterprise feature and is not available for oss you will need a yubikey to test this feature this feature has additional build requirements so it should be tested with a pre release build from drone eg these tests should be carried out sequentially tsh tests should be carried out on linux macos and windows tsh login as user with login and no hardware key requirement request a role with role role options require session mfa hardware key tsh login request roles hardware key required assuming the role should force automatic re login with yubikey tsh ssh requires yubikey to be connected for re login prompts for per session mfa request a role with role role options require session mfa hardware key touch tsh login request roles hardware key touch required assuming the role should force automatic re login with yubikey prompts for touch if not cached last touch within seconds tsh ssh requires yubikey to be connected for re login prompts for touch if not cached tsh logout and tsh login as the user with no hardware key requirement upgrade auth settings to auth service authentication require session mfa hardware key using the existing login session tsh ls should force automatic re login with yubikey tsh ssh requires yubikey to be connected for re login prompts for per session mfa upgrade auth settings to auth service authentication require session mfa hardware key touch using the existing login session tsh ls should force automatic re login with yubikey prompts for touch if not cached tsh ssh requires yubikey to be connected for re login prompts for touch if not cached performance rosstimothy fspmarshall perform all tests on the following configurations with default networking configuration with proxy peering enabled with tls routing enabled cluster with direct dial nodes etcd dynamodb firestore cluster with reverse tunnel nodes etcd dynamodb firestore cluster with trusted clusters etcd dynamodb firestore soak test run minute soak test with a mix of interactive non interactive sessions for both direct and reverse tunnel nodes shell tsh bench duration user direct dial node ls tsh bench i duration user direct dial node ps uax tsh bench duration user reverse tunnel node ls tsh bench i duration user reverse tunnel node ps uax observe prometheus metrics for goroutines open files ram cpu timers and make sure there are no leaks verify that prometheus metrics are accurate concurrent session test cluster with reverse tunnel nodes run a concurrent session test that will spawn interactive sessions per node in the cluster shell tsh bench sessions max user ls tsh bench sessions max web user ls verify that all sessions are able to be established verify that tsh and the web ui are still functional teleport with cloud providers aws hugoshaka deploy teleport to aws using dynamodb deploy teleport enterprise to aws using ha setup gcp antonam deploy teleport to gcp using cloud firestore cloud storage deploy teleport to gke google kubernetes engine deploy teleport enterprise to gcp ibm atburke deploy teleport to ibm cloud using ibm database for etcd ibm object store deploy teleport to ibm cloud kubernetes deploy teleport enterprise to ibm cloud application access mdwn run an application within local cluster verify the debug application debug app true works verify an application can be configured with command line flags verify an application can be configured from file configuration verify that applications are available at auto generated addresses name rootproxypublicaddr and well as publicaddr run an application within a trusted cluster verify that applications are available at auto generated addresses name rootproxypublicaddr verify audit records app session start and app session chunk events are created in the audit log app session chunk points to a minute session archive with multiple app session request events inside tsh play can fetch and print a session chunk archive verify jwt using verify rbac verify with tsh app login verify aws console access can log into aws web console through the web ui can interact with aws using tsh aws commands verify dynamic registration can register a new app using tctl create can update registered app using tctl create f can delete registered app using tctl rm test applications screen in the web ui tab is located on left side nav on dashboard verify that all apps registered are shown verify that clicking on the app icon takes you to another tab verify using the bash command produced from add application dialogue works refresh app screen to see it registered database access smallinsky db access team connect to a database within a local cluster self hosted postgres tener self hosted mysql gavinfrazar self hosted mariadb gavinfrazar self hosted mongodb tener self hosted cockroachdb gavinfrazar self hosted redis self hosted redis cluster self hosted mssql smallinsky aws aurora postgres gabrielcorado aws aurora mysql tener aws redshift gabrielcorado aws elasticache aws memorydb gcp cloud sql postgres gabrielcorado gcp cloud sql mysql tener snowflake smallinsky azure cache for redis smallinsky elasticsearch cassandra scylladb gabrielcorado connect to a database within a remote cluster via a trusted cluster self hosted postgres tener self hosted mysql gavinfrazar self hosted mariadb gavinfrazar self hosted mongodb tener self hosted cockroachdb gavinfrazar self hosted redis self hosted redis cluster self hosted mssql smallinsky aws aurora postgres gabrielcorado aws aurora mysql tener aws redshift gabrielcorado aws elasticache aws memorydb gcp cloud sql postgres gabrielcorado gcp cloud sql mysql tener snowflake smallinsky azure cache for redis smallinsky elasticsearch cassandra scylladb gabrielcorado verify audit events gavinfrazar db session start is emitted when you connect db session end is emitted when you disconnect db session query is emitted when you execute a sql query verify rbac tsh db ls shows only databases matching role s db labels gabrielcorado can only connect as users from db users gabrielcorado postgres only can only connect to databases from db names gabrielcorado db session start is emitted when connection attempt is denied mongodb only can only execute commands in databases from db names gabrielcorado db session query is emitted when command fails due to permissions can configure per session mfa gavinfrazar mfa tap is required on each tsh db connect verify dynamic registration gavinfrazar can register a new database using tctl create can update registered database using tctl create f can delete registered database using tctl rm verify discovery aws can detect and register rds instances gavinfrazar can detect and register aurora clusters and their reader and custom endpoints gabrielcorado can detect and register redshift clusters smallinsky can detect and register elasticache redis clusters can detect and register memorydb clusters azure can detect and register mysql and postgres instances gavinfrazar can detect and register azure cache for redis servers smallinsky verify teleport managed users password rotation auto auth on connection etc can detect and manage elasticache users can detect and manage memorydb users test databases screen in the web ui tab is located on left side nav on dashboard tener verify that all dbs registered are shown with correct name description type and labels verify that clicking on a rows connect button renders a dialogue on manual instructions with step login value matching the rows name column verify searching for all columns in the search bar works verify you can sort by all columns except labels other mysql server version reported by teleport is correct smallinsky tls routing smallinsky verify that teleport proxy configuration starts only a single listener smallinsky version teleport proxy service enabled yes public addr web listen addr run teleport proxy in multiplex mode auth service proxy listener mode multiplex smallinsky trusted cluster setup trusted clusters using single port setup web proxy addr tunnel addr kind trusted cluster spec web proxy addr root example com tunnel addr root example com database access verify that tsh db connect works through proxy running in multiplex mode postgres tener mysql gavinfrazar mariadb gavinfrazar mongodb tener cockroachdb gavinfrazar redis mssql smallinsky snowflake smallinsky elasticsearch cassandra scylladb gabrielcorado verify connecting to a database through tls alpn sni local proxy tsh proxy db with a gui client smallinsky gavinfrazar tener gabrielcorado verify tsh proxy db with teleport proxy behind alb smallinsky application access gavinfrazar verify app access through proxy running in multiplex mode ssh access gabrielcorado connect to a openssh server through a local ssh proxy ssh o forwardagent yes o proxycommand tsh proxy ssh user host example com connect to a openssh server on leaf cluster through a local ssh proxy ssh o forwardagent yes o proxycommand tsh proxy ssh user r cluster leaf cluster h p user node foo com verify tsh ssh access through proxy running in multiplex mode kubernetes access gavinfrazar verify kubernetes access through proxy running in multiplex mode desktop access ibeckermayer probakowski lkozlowski direct mode set listen addr can connect to desktop defined in static hosts section can connect to desktop discovered via ldap iot mode reverse tunnel through proxy can connect to desktop defined in static hosts section can connect to desktop discovered via ldap connect multiple windows desktop service s to the same teleport cluster verify that connections to desktops on different ad domains works attempt to connect several times to verify that you are routed to the correct windows desktop service verify user input download and verify all keys are processed correctly in each supported browser known issues cannot be captured by the browser without on macos left click and right click register as windows clicks right click on the desktop should show a windows menu not a browser context menu vertical and horizontal scroll work locking verify that placing a user lock terminates an active desktop session verify that placing a desktop lock terminates an active desktop session verify that placing a role lock terminates an active desktop session labeling set client idle timeout to a small value and verify that idle sessions are terminated the session should end and an audit event will confirm it was due to idle connection all desktops have teleport dev origin label dynamic desktops have additional teleport dev labels for os os version dns hostname regexp based host labeling applies across all desktops regardless of origin rbac rbac denies access to a windows desktop due to labels rbac denies access to a windows desktop with the wrong os login clipboard support when a user has a role with clipboard sharing enabled and is using a chromium based browser going to a desktop when clipboard permissions are in ask mode aka prompt causes the browser to show a prompt while the ui shows a spinner x ing out of the prompt causing the clipboard permission to remain in ask mode causes the prompt to show up again denying clipboard permissions brings up a relevant error alert with clipboard sharing disabled in the top bar allowing clipboard permissions allows you to see the desktop session with clipboard sharing enabled highlighted in the top bar copy text from local workstation paste into remote desktop copy text from remote desktop paste into local workstation when a user has a role with clipboard sharing enabled and is not using a chromium based browser the ui shows a relevant alert and clipboard sharing disabled is highlighted in the top bar when a user has a role with clipboard sharing disabled and is using a chromium and non chromium based browser confirm both the live session should show disabled in the top bar and copy paste should not work between your workstation and the remote desktop directory sharing on supported non chromium based browsers firefox safari attempting to share directory shows a dismissible unsupported action dialog on supported chromium based browsers chrome edge begin sharing works the shared directory icon in the top right of the screen is highlighted when directory sharing is initiated the shared directory appears as a network drive named on teleport the share directory menu option dissapears from the menu navigation the folders of the shared directory are navigable move up and down the directory tree crud a new text file can be created the text file can be written to saved the text file can be read close it check that its saved on the local machine then open it again on the remote the text file can be deleted file folder movement in to out make at least one of these from a non top level directory a file from inside the shared directory can be drag and dropped outside the shared directory a folder from inside the shared directory can be drag and dropped outside the shared directory and its contents retained a file from inside the shared directory can be cut pasted outside the shared directory a folder from inside the shared directory can be cut pasted outside the shared directory a file from inside the shared directory can be copy pasted outside the shared directory a folder from inside the shared directory can be copy pasted outside the shared directory out to in make at least one of these overwrite an existing file and one go into a non top level directory a file from outside the shared directory can be drag and dropped into the shared directory a folder from outside the shared directory can be drag and dropped into the shared directory and its contents retained a file from outside the shared directory can be cut pasted into the shared directory a folder from outside the shared directory can be cut pasted into the shared directory a file from outside the shared directory can be copy pasted into the shared directory a folder from outside the shared directory can be copy pasted into the shared directory within a file from inside the shared directory cannot be drag and dropped to another folder inside the shared directory a dismissible unsupported action dialog is shown a folder from inside the shared directory cannot be drag and dropped to another folder inside the shared directory a dismissible unsupported action dialog is shown a file from inside the shared directory cannot be cut pasted to another folder inside the shared directory a dismissible unsupported action dialog is shown a folder from inside the shared directory cannot be cut pasted to another folder inside the shared directory a dismissible unsupported action dialog is shown a file from inside the shared directory can be copy pasted to another folder inside the shared directory a folder from inside the shared directory can be copy pasted to another folder inside shared directory and its contents retained rbac give the user one role that explicitly disables directory sharing desktop directory sharing false and confirm that the option to share a directory doesn t appear in the menu per session mfa try webauthn on each of chrome safari and firefox only works with firefox attempting to start a session no keys registered shows an error message attempting to start a session with a key registered shows an error message n a now that support has been removed attempting to start a session with a webauthn registered pops up the verify your identity dialog hitting cancel shows an error message hitting verify causes your browser to prompt you for mfa cancelling that browser mfa prompt shows an error successful mfa verification allows you to connect session recording verify sessions are not recorded if all of a user s roles disable recording verify sync recording mode node sync or mode proy sync verify async recording mode node or mode proxy sessions show up in session recordings ui with desktop icon sessions can be played back including play pause functionality a session that ends with a tdp error message can be played back ends by displaying the error message and the progress bar progresses to the end attempting to play back a session that doesn t exist i e by entering a non existing session id in the url shows a relevant error message rbac for sessions ensure users can only see their own recordings when using the rbac rule from our docs pages access controls reference mdx rbac for sessions audit events check these after performing the above tests windows desktop session start emitted on start windows desktop session start emitted when session fails to start due to rbac for example windows desktop session end emitted on end desktop clipboard send emitted for local copy remote paste desktop clipboard receive emitted for remote copy local paste binaries compatibility fheinecke verify that teleport tsh tctl tbot run on centos centos ubuntu ubuntu debian verify tsh runs on windows macos machine id ssh with a default teleport instance configured with a ssh node verify you are able to create a new bot user with tctl bots add robot roles access follow the instructions provided in the output to start tbot verify you are able to connect to the ssh node using openssh with the generated ssh config in the destination directory verify that after the renewal period default but this can be reduced via configuration that newly generated certificates are placed in the destination directory verify that sending both and sighup to a running tbot process causes a renewal and new certificates to be generated verify that you are able to make a connection to the ssh node using the ssh config provided by tbot after each phase of a manual ca rotation ensure the above tests are completed for both directly connecting to the auth server connecting to the auth server via the proxy reverse tunnel db access with a default postgres db instance a teleport instance configured with db access and a bot user configured verify you are able to connect to and interact with a database using tbot db while tbot start is running host users creation lxea todo lxea replace links with actual docs once merged docs pages server access guides host user creation mdx rfd automatic user provisioning md verify host users creation functionality non existing users are created automatically users are added to groups non existing configured groups are created created users are added to the teleport system group users are cleaned up after their session ends cleanup occurs if a program was left running after session ends sudoers file creation is successful invalid sudoers files are not created existing host users are not modified setting disable create host user true stops user creation from occurring ca rotations espadolini verify the ca rotation functionality itself by checking in the backend or with tctl get cert authority standby phase only active keys no additional trusted keys init phase active keys and additional trusted keys update clients and update servers phases the certs from the init phase are swapped standby phase only the new certs remain in active keys nothing in additional trusted keys rollback phase second pass after completing a regular rotation same content as in the init phase standby phase after rollback same content as in the previous standby phase verify functionality in all phases clients might have to log in again in lieu of waiting for credentials to expire between phases ssh session in tsh from a previous phase ssh session in web ui from a previous phase new ssh session with tsh new ssh session with web ui new ssh session in a child cluster on the same major version new ssh session in a child cluster on the previous major version new ssh session from a parent cluster application access through a browser application access through curl with tsh app login kubectl get po after tsh kube login database access no configuration change should be necessary if the database ca isn t rotated other teleport functionality should not be affected if only the database ca is rotated discovery lxea verify instance discovery only instances matching given aws tags have the installer executed on them only the iam permissions mentioned in the discovery docs are required for operation custom scripts specified in different matchers are executed custom ssm documents specified in different matchers are executed new instances with matching aws tags are discovered and added to the teleport cluster large numbers of instances are all successfully added to the cluster nodes that have been discovered do not have the install script run on the node multiple times resources reference style links
1
125,653
26,708,621,696
IssuesEvent
2023-01-27 20:46:28
PickupPlayerApp/AppDevelopment
https://api.github.com/repos/PickupPlayerApp/AppDevelopment
opened
Finish Started Dart Pages in effort to learn Dart
Code work
Kennedy Will be finishing Tylers work to understand Dart better and learn the content of this work.
1.0
Finish Started Dart Pages in effort to learn Dart - Kennedy Will be finishing Tylers work to understand Dart better and learn the content of this work.
non_test
finish started dart pages in effort to learn dart kennedy will be finishing tylers work to understand dart better and learn the content of this work
0
349,310
31,791,973,173
IssuesEvent
2023-09-13 04:36:22
getodk/central
https://api.github.com/repos/getodk/central
closed
Support Single Sign-on with OpenID Connect
enhancement backend frontend ops needs testing
Creating an issue to track work on SSO: - [Release criteria](https://docs.google.com/document/d/1lEIp5te0-SkA9xEPwqQdf3Kp6ciTqY8UxGBosoKPK5Q/edit#heading=h.a02nafhkjvhn) - [Backend PR](https://github.com/getodk/central-backend/pull/910) - [Frontend PR](https://github.com/getodk/central-frontend/pull/819) - [`central` PR](https://github.com/getodk/central/pull/458)
1.0
Support Single Sign-on with OpenID Connect - Creating an issue to track work on SSO: - [Release criteria](https://docs.google.com/document/d/1lEIp5te0-SkA9xEPwqQdf3Kp6ciTqY8UxGBosoKPK5Q/edit#heading=h.a02nafhkjvhn) - [Backend PR](https://github.com/getodk/central-backend/pull/910) - [Frontend PR](https://github.com/getodk/central-frontend/pull/819) - [`central` PR](https://github.com/getodk/central/pull/458)
test
support single sign on with openid connect creating an issue to track work on sso
1
50,684
3,006,556,910
IssuesEvent
2015-07-27 11:13:57
Itseez/opencv
https://api.github.com/repos/Itseez/opencv
opened
OpenCV Manager on Galaxy Note 4
affected: 2.4 auto-transferred bug category: android priority: low
Transferred from http://code.opencv.org/issues/4232 ``` || Russell Priebe on 2015-03-08 23:40 || Priority: Low || Affected: 2.4.0 - 2.4.8 || Category: android || Tracker: Bug || Difficulty: || PR: || Platform: ARM / Android ``` OpenCV Manager on Galaxy Note 4 ----------- ``` I am developing an app that uses OpenCV. I just upgraded to a samsung galaxy note 4 (KitKat 4.4.4). I have installed OpenCV manager (a couple of times). When I run the OpenCV Manager, under "Device Information" it says "Hardware not Detected". When I run my app I get a dialog that says Package Not Found OpenCV libary package was not found. Try to install it? If I select "Yes" it takes me to the Play Store where it says "Item not Found". If there is anything I can do to help debug, please let me know. ``` History ------- ##### Russell Priebe on 2015-03-09 05:06 ``` affected version shown above is incorrect. It should read 2.4.9.0 rev 1 ```
1.0
OpenCV Manager on Galaxy Note 4 - Transferred from http://code.opencv.org/issues/4232 ``` || Russell Priebe on 2015-03-08 23:40 || Priority: Low || Affected: 2.4.0 - 2.4.8 || Category: android || Tracker: Bug || Difficulty: || PR: || Platform: ARM / Android ``` OpenCV Manager on Galaxy Note 4 ----------- ``` I am developing an app that uses OpenCV. I just upgraded to a samsung galaxy note 4 (KitKat 4.4.4). I have installed OpenCV manager (a couple of times). When I run the OpenCV Manager, under "Device Information" it says "Hardware not Detected". When I run my app I get a dialog that says Package Not Found OpenCV libary package was not found. Try to install it? If I select "Yes" it takes me to the Play Store where it says "Item not Found". If there is anything I can do to help debug, please let me know. ``` History ------- ##### Russell Priebe on 2015-03-09 05:06 ``` affected version shown above is incorrect. It should read 2.4.9.0 rev 1 ```
non_test
opencv manager on galaxy note transferred from russell priebe on priority low affected category android tracker bug difficulty pr platform arm android opencv manager on galaxy note i am developing an app that uses opencv i just upgraded to a samsung galaxy note kitkat i have installed opencv manager a couple of times when i run the opencv manager under device information it says hardware not detected when i run my app i get a dialog that says package not found opencv libary package was not found try to install it if i select yes it takes me to the play store where it says item not found if there is anything i can do to help debug please let me know history russell priebe on affected version shown above is incorrect it should read rev
0
71,329
23,542,390,781
IssuesEvent
2022-08-20 15:54:06
scoutplan/scoutplan
https://api.github.com/repos/scoutplan/scoutplan
closed
[Scoutplan Production/production] NoMethodError: undefined method `payment_amount' for #<Event id: nil, unit_id: 32098, title: nil, description: nil, starts_at: "2022-09-17 10:00:00.000000000 -0400", ends_at: "2022-09-17 16:00:00.000000000 -0400", location: nil, created_at: nil, updated_at: nil, requires_rsvp: false, max_total_attendees: nil, rsvp_opens_at: nil, event_category_id: nil, series_parent_id: nil, status: "draft", repeats_until: nil, address: nil, short_description: nil, departs_from: nil, inc...
defect
## Backtrace line 80 of [PROJECT_ROOT]/app/views/events/_form.slim: block in _app_views_events__form_slim___341976583372312877_18728000 line 16 of [PROJECT_ROOT]/app/views/events/_form.slim: _app_views_events__form_slim___341976583372312877_18728000 line 2 of [PROJECT_ROOT]/app/views/events/new.html.slim: block in _app_views_events_new_html_slim___149310029422152573_18727900 [View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/87549915)
1.0
[Scoutplan Production/production] NoMethodError: undefined method `payment_amount' for #<Event id: nil, unit_id: 32098, title: nil, description: nil, starts_at: "2022-09-17 10:00:00.000000000 -0400", ends_at: "2022-09-17 16:00:00.000000000 -0400", location: nil, created_at: nil, updated_at: nil, requires_rsvp: false, max_total_attendees: nil, rsvp_opens_at: nil, event_category_id: nil, series_parent_id: nil, status: "draft", repeats_until: nil, address: nil, short_description: nil, departs_from: nil, inc... - ## Backtrace line 80 of [PROJECT_ROOT]/app/views/events/_form.slim: block in _app_views_events__form_slim___341976583372312877_18728000 line 16 of [PROJECT_ROOT]/app/views/events/_form.slim: _app_views_events__form_slim___341976583372312877_18728000 line 2 of [PROJECT_ROOT]/app/views/events/new.html.slim: block in _app_views_events_new_html_slim___149310029422152573_18727900 [View full backtrace and more info at honeybadger.io](https://app.honeybadger.io/projects/97676/faults/87549915)
non_test
nomethoderror undefined method payment amount for event id nil unit id title nil description nil starts at ends at location nil created at nil updated at nil requires rsvp false max total attendees nil rsvp opens at nil event category id nil series parent id nil status draft repeats until nil address nil short description nil departs from nil inc backtrace line of app views events form slim block in app views events form slim line of app views events form slim app views events form slim line of app views events new html slim block in app views events new html slim
0
281,730
24,414,899,446
IssuesEvent
2022-10-05 15:07:00
yugabyte/yugabyte-db
https://api.github.com/repos/yugabyte/yugabyte-db
closed
[DocDB] TSAN race in SealTxnTest.Update
kind/bug kind/failing-test area/docdb priority/medium
Jira Link: [DB-3057](https://yugabyte.atlassian.net/browse/DB-3057) ### Description ``` WARNING: ThreadSanitizer: data race (pid=1915) Write of size 8 at 0x7b7000084ce0 by thread T79 (mutexes: write M833020073977527028): #0 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::reset(yb::tablet::TransactionParticipant*) /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:1682:20 (libtablet.so+0x4704b2) #1 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::~unique_ptr() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:1638:19 (libtablet.so+0x4410fb) #2 yb::tablet::Tablet::~Tablet() ${BUILD_ROOT}/../../src/yb/tablet/tablet.cc:521:1 (libtablet.so+0x41b3b3) #3 void std::__1::destroy_at<yb::tablet::Tablet>(yb::tablet::Tablet*) /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/__memory/base.h:118:13 (libtablet_test_util.so+0x609b1) #4 void std::__1::allocator_traits<std::__1::allocator<yb::tablet::Tablet> >::destroy<yb::tablet::Tablet, void, void>(std::__1::allocator<yb::tablet::Tablet>&, yb::tablet::Tablet*) /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/__memory/allocator_traits.h:315:9 (libtablet_test_util.so+0x60979) #5 std::__1::__shared_ptr_emplace<yb::tablet::Tablet, std::__1::allocator<yb::tablet::Tablet> >::__on_zero_shared() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:2637:9 (libtablet_test_util.so+0x6082d) #6 std::__1::__shared_count::__release_shared() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:2501:9 (libintegration-tests.so+0x1b8985) #7 std::__1::__shared_weak_count::__release_shared() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:2543:27 (libintegration-tests.so+0x1b8929) #8 std::__1::shared_ptr<yb::tablet::Tablet>::~shared_ptr() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:3233:19 (libintegration-tests.so+0x274228) #9 std::__1::shared_ptr<yb::tablet::Tablet>::reset() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:3328:5 (libtablet.so+0x4e4355) #10 yb::tablet::TabletPeer::CompleteShutdown(yb::StronglyTypedBool<yb::tablet::DisableFlushOnShutdown_Tag>) ${BUILD_ROOT}/../../src/yb/tablet/tablet_peer.cc:533:13 (libtablet.so+0x4d6e67) #11 yb::tablet::TabletPeer::Shutdown(yb::StronglyTypedBool<yb::tablet::ShouldAbortActiveTransactions_Tag>, yb::StronglyTypedBool<yb::tablet::DisableFlushOnShutdown_Tag>) ${BUILD_ROOT}/../../src/yb/tablet/tablet_peer.cc:585:5 (libtablet.so+0x4d76e2) #12 yb::tserver::TSTabletManager::DeleteTablet(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, yb::tablet::TabletDataState, yb::StronglyTypedBool<yb::tablet::ShouldAbortActiveTransactions_Tag>, boost::optional<long> const&, bool, boost::optional<yb::tserver::TabletServerErrorPB_Code>*) ${BUILD_ROOT}/../../src/yb/tserver/ts_tablet_manager.cc:1270:3 (libtserver.so+0x504728) #13 yb::tserver::TabletServiceAdminImpl::DeleteTablet(yb::tserver::DeleteTabletRequestPB const*, yb::tserver::DeleteTabletResponsePB*, yb::rpc::RpcContext) ${BUILD_ROOT}/../../src/yb/tserver/tablet_service.cc:1398:41 (libtserver.so+0x4a21f6) ... Previous read of size 8 at 0x7b7000084ce0 by main thread: #0 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::get() const /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:1657:19 (seal-txn-test+0x313be4) #1 yb::tablet::Tablet::transaction_participant() const ${BUILD_ROOT}/../../src/yb/tablet/tablet.h:550:37 (seal-txn-test+0x312160) #2 yb::AssertRunningTransactionsCountLessOrEqualTo(yb::MiniCluster*, unsigned long) ${BUILD_ROOT}/../../src/yb/integration-tests/mini_cluster_utils.cc:71:42 (libintegration-tests.so+0x28c058) #3 yb::AssertNoRunningTransactions(yb::MiniCluster*) ${BUILD_ROOT}/../../src/yb/integration-tests/mini_cluster_utils.cc:92:3 (libintegration-tests.so+0x28c53b) #4 yb::client::TransactionTestBase<yb::MiniCluster>::AssertNoRunningTransactions() ${BUILD_ROOT}/../../src/yb/client/txn-test-base.cc:307:3 (libql-dml-test-base.so+0x8d925) #5 yb::client::SealTxnTest_Update_Test::TestBody() ${BUILD_ROOT}/../../src/yb/client/seal-txn-test.cc:151:3 (seal-txn-test+0x30e5fd) ```
1.0
[DocDB] TSAN race in SealTxnTest.Update - Jira Link: [DB-3057](https://yugabyte.atlassian.net/browse/DB-3057) ### Description ``` WARNING: ThreadSanitizer: data race (pid=1915) Write of size 8 at 0x7b7000084ce0 by thread T79 (mutexes: write M833020073977527028): #0 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::reset(yb::tablet::TransactionParticipant*) /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:1682:20 (libtablet.so+0x4704b2) #1 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::~unique_ptr() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:1638:19 (libtablet.so+0x4410fb) #2 yb::tablet::Tablet::~Tablet() ${BUILD_ROOT}/../../src/yb/tablet/tablet.cc:521:1 (libtablet.so+0x41b3b3) #3 void std::__1::destroy_at<yb::tablet::Tablet>(yb::tablet::Tablet*) /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/__memory/base.h:118:13 (libtablet_test_util.so+0x609b1) #4 void std::__1::allocator_traits<std::__1::allocator<yb::tablet::Tablet> >::destroy<yb::tablet::Tablet, void, void>(std::__1::allocator<yb::tablet::Tablet>&, yb::tablet::Tablet*) /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/__memory/allocator_traits.h:315:9 (libtablet_test_util.so+0x60979) #5 std::__1::__shared_ptr_emplace<yb::tablet::Tablet, std::__1::allocator<yb::tablet::Tablet> >::__on_zero_shared() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:2637:9 (libtablet_test_util.so+0x6082d) #6 std::__1::__shared_count::__release_shared() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:2501:9 (libintegration-tests.so+0x1b8985) #7 std::__1::__shared_weak_count::__release_shared() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:2543:27 (libintegration-tests.so+0x1b8929) #8 std::__1::shared_ptr<yb::tablet::Tablet>::~shared_ptr() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:3233:19 (libintegration-tests.so+0x274228) #9 std::__1::shared_ptr<yb::tablet::Tablet>::reset() /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:3328:5 (libtablet.so+0x4e4355) #10 yb::tablet::TabletPeer::CompleteShutdown(yb::StronglyTypedBool<yb::tablet::DisableFlushOnShutdown_Tag>) ${BUILD_ROOT}/../../src/yb/tablet/tablet_peer.cc:533:13 (libtablet.so+0x4d6e67) #11 yb::tablet::TabletPeer::Shutdown(yb::StronglyTypedBool<yb::tablet::ShouldAbortActiveTransactions_Tag>, yb::StronglyTypedBool<yb::tablet::DisableFlushOnShutdown_Tag>) ${BUILD_ROOT}/../../src/yb/tablet/tablet_peer.cc:585:5 (libtablet.so+0x4d76e2) #12 yb::tserver::TSTabletManager::DeleteTablet(std::__1::basic_string<char, std::__1::char_traits<char>, std::__1::allocator<char> > const&, yb::tablet::TabletDataState, yb::StronglyTypedBool<yb::tablet::ShouldAbortActiveTransactions_Tag>, boost::optional<long> const&, bool, boost::optional<yb::tserver::TabletServerErrorPB_Code>*) ${BUILD_ROOT}/../../src/yb/tserver/ts_tablet_manager.cc:1270:3 (libtserver.so+0x504728) #13 yb::tserver::TabletServiceAdminImpl::DeleteTablet(yb::tserver::DeleteTabletRequestPB const*, yb::tserver::DeleteTabletResponsePB*, yb::rpc::RpcContext) ${BUILD_ROOT}/../../src/yb/tserver/tablet_service.cc:1398:41 (libtserver.so+0x4a21f6) ... Previous read of size 8 at 0x7b7000084ce0 by main thread: #0 std::__1::unique_ptr<yb::tablet::TransactionParticipant, std::__1::default_delete<yb::tablet::TransactionParticipant> >::get() const /opt/yb-build/thirdparty/yugabyte-db-thirdparty-v20220630123401-af96d73e39-centos7-x86_64-clang12/installed/tsan/libcxx/include/c++/v1/memory:1657:19 (seal-txn-test+0x313be4) #1 yb::tablet::Tablet::transaction_participant() const ${BUILD_ROOT}/../../src/yb/tablet/tablet.h:550:37 (seal-txn-test+0x312160) #2 yb::AssertRunningTransactionsCountLessOrEqualTo(yb::MiniCluster*, unsigned long) ${BUILD_ROOT}/../../src/yb/integration-tests/mini_cluster_utils.cc:71:42 (libintegration-tests.so+0x28c058) #3 yb::AssertNoRunningTransactions(yb::MiniCluster*) ${BUILD_ROOT}/../../src/yb/integration-tests/mini_cluster_utils.cc:92:3 (libintegration-tests.so+0x28c53b) #4 yb::client::TransactionTestBase<yb::MiniCluster>::AssertNoRunningTransactions() ${BUILD_ROOT}/../../src/yb/client/txn-test-base.cc:307:3 (libql-dml-test-base.so+0x8d925) #5 yb::client::SealTxnTest_Update_Test::TestBody() ${BUILD_ROOT}/../../src/yb/client/seal-txn-test.cc:151:3 (seal-txn-test+0x30e5fd) ```
test
tsan race in sealtxntest update jira link description warning threadsanitizer data race pid write of size at by thread mutexes write std unique ptr reset yb tablet transactionparticipant opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory libtablet so std unique ptr unique ptr opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory libtablet so yb tablet tablet tablet build root src yb tablet tablet cc libtablet so void std destroy at yb tablet tablet opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory base h libtablet test util so void std allocator traits destroy std allocator yb tablet tablet opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory allocator traits h libtablet test util so std shared ptr emplace on zero shared opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory libtablet test util so std shared count release shared opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory libintegration tests so std shared weak count release shared opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory libintegration tests so std shared ptr shared ptr opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory libintegration tests so std shared ptr reset opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory libtablet so yb tablet tabletpeer completeshutdown yb stronglytypedbool build root src yb tablet tablet peer cc libtablet so yb tablet tabletpeer shutdown yb stronglytypedbool yb stronglytypedbool build root src yb tablet tablet peer cc libtablet so yb tserver tstabletmanager deletetablet std basic string std allocator const yb tablet tabletdatastate yb stronglytypedbool boost optional const bool boost optional build root src yb tserver ts tablet manager cc libtserver so yb tserver tabletserviceadminimpl deletetablet yb tserver deletetabletrequestpb const yb tserver deletetabletresponsepb yb rpc rpccontext build root src yb tserver tablet service cc libtserver so previous read of size at by main thread std unique ptr get const opt yb build thirdparty yugabyte db thirdparty installed tsan libcxx include c memory seal txn test yb tablet tablet transaction participant const build root src yb tablet tablet h seal txn test yb assertrunningtransactionscountlessorequalto yb minicluster unsigned long build root src yb integration tests mini cluster utils cc libintegration tests so yb assertnorunningtransactions yb minicluster build root src yb integration tests mini cluster utils cc libintegration tests so yb client transactiontestbase assertnorunningtransactions build root src yb client txn test base cc libql dml test base so yb client sealtxntest update test testbody build root src yb client seal txn test cc seal txn test
1